Unnamed: 0
int64
3
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
2
430
labels
stringlengths
4
347
body
stringlengths
5
237k
index
stringclasses
7 values
text_combine
stringlengths
96
237k
label
stringclasses
2 values
text
stringlengths
96
219k
binary_label
int64
0
1
108,934
16,822,763,437
IssuesEvent
2021-06-17 14:50:53
idonthaveafifaaddiction/flink
https://api.github.com/repos/idonthaveafifaaddiction/flink
opened
WS-2019-0424 (Medium) detected in elliptic-6.4.1.tgz
security vulnerability
## WS-2019-0424 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>elliptic-6.4.1.tgz</b></p></summary> <p>EC cryptography</p> <p>Library home page: <a href="https://registry.npmjs.org/elliptic/-/elliptic-6.4.1.tgz">https://registry.npmjs.org/elliptic/-/elliptic-6.4.1.tgz</a></p> <p>Path to dependency file: flink/flink-runtime-web/web-dashboard/package.json</p> <p>Path to vulnerable library: flink/flink-runtime-web/web-dashboard/node_modules/elliptic/package.json</p> <p> Dependency Hierarchy: - build-angular-0.13.6.tgz (Root Library) - webpack-4.29.0.tgz - node-libs-browser-2.2.0.tgz - crypto-browserify-3.12.0.tgz - browserify-sign-4.0.4.tgz - :x: **elliptic-6.4.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/idonthaveafifaaddiction/flink/commit/d77b18bba5da590fb2e8e8aa13f2dcb0674d52be">d77b18bba5da590fb2e8e8aa13f2dcb0674d52be</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> all versions of elliptic are vulnerable to Timing Attack through side-channels. <p>Publish Date: 2019-11-13 <p>URL: <a href=https://github.com/indutny/elliptic/commit/ec735edde187a43693197f6fa3667ceade751a3a>WS-2019-0424</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Adjacent - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"elliptic","packageVersion":"6.4.1","packageFilePaths":["/flink-runtime-web/web-dashboard/package.json"],"isTransitiveDependency":true,"dependencyTree":"@angular-devkit/build-angular:0.13.6;webpack:4.29.0;node-libs-browser:2.2.0;crypto-browserify:3.12.0;browserify-sign:4.0.4;elliptic:6.4.1","isMinimumFixVersionAvailable":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2019-0424","vulnerabilityDetails":"all versions of elliptic are vulnerable to Timing Attack through side-channels.","vulnerabilityUrl":"https://github.com/indutny/elliptic/commit/ec735edde187a43693197f6fa3667ceade751a3a","cvss3Severity":"medium","cvss3Score":"5.9","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Adjacent","I":"High"},"extraData":{}}</REMEDIATE> -->
True
WS-2019-0424 (Medium) detected in elliptic-6.4.1.tgz - ## WS-2019-0424 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>elliptic-6.4.1.tgz</b></p></summary> <p>EC cryptography</p> <p>Library home page: <a href="https://registry.npmjs.org/elliptic/-/elliptic-6.4.1.tgz">https://registry.npmjs.org/elliptic/-/elliptic-6.4.1.tgz</a></p> <p>Path to dependency file: flink/flink-runtime-web/web-dashboard/package.json</p> <p>Path to vulnerable library: flink/flink-runtime-web/web-dashboard/node_modules/elliptic/package.json</p> <p> Dependency Hierarchy: - build-angular-0.13.6.tgz (Root Library) - webpack-4.29.0.tgz - node-libs-browser-2.2.0.tgz - crypto-browserify-3.12.0.tgz - browserify-sign-4.0.4.tgz - :x: **elliptic-6.4.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/idonthaveafifaaddiction/flink/commit/d77b18bba5da590fb2e8e8aa13f2dcb0674d52be">d77b18bba5da590fb2e8e8aa13f2dcb0674d52be</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> all versions of elliptic are vulnerable to Timing Attack through side-channels. <p>Publish Date: 2019-11-13 <p>URL: <a href=https://github.com/indutny/elliptic/commit/ec735edde187a43693197f6fa3667ceade751a3a>WS-2019-0424</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Adjacent - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"elliptic","packageVersion":"6.4.1","packageFilePaths":["/flink-runtime-web/web-dashboard/package.json"],"isTransitiveDependency":true,"dependencyTree":"@angular-devkit/build-angular:0.13.6;webpack:4.29.0;node-libs-browser:2.2.0;crypto-browserify:3.12.0;browserify-sign:4.0.4;elliptic:6.4.1","isMinimumFixVersionAvailable":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2019-0424","vulnerabilityDetails":"all versions of elliptic are vulnerable to Timing Attack through side-channels.","vulnerabilityUrl":"https://github.com/indutny/elliptic/commit/ec735edde187a43693197f6fa3667ceade751a3a","cvss3Severity":"medium","cvss3Score":"5.9","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Adjacent","I":"High"},"extraData":{}}</REMEDIATE> -->
non_comp
ws medium detected in elliptic tgz ws medium severity vulnerability vulnerable library elliptic tgz ec cryptography library home page a href path to dependency file flink flink runtime web web dashboard package json path to vulnerable library flink flink runtime web web dashboard node modules elliptic package json dependency hierarchy build angular tgz root library webpack tgz node libs browser tgz crypto browserify tgz browserify sign tgz x elliptic tgz vulnerable library found in head commit a href found in base branch master vulnerability details all versions of elliptic are vulnerable to timing attack through side channels publish date url a href cvss score details base score metrics exploitability metrics attack vector adjacent attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact high availability impact none for more information on scores click a href isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree angular devkit build angular webpack node libs browser crypto browserify browserify sign elliptic isminimumfixversionavailable false basebranches vulnerabilityidentifier ws vulnerabilitydetails all versions of elliptic are vulnerable to timing attack through side channels vulnerabilityurl
0
16,821
4,098,918,269
IssuesEvent
2016-06-03 10:19:39
digitalmethodsinitiative/dmi-tcat
https://api.github.com/repos/digitalmethodsinitiative/dmi-tcat
opened
Auto-installer needs to install ntpdate
documentation enhancement
Twitter returns a crypted 401 Authorization Denied whenever the TCAT system clocktime is out-of-sync with the Twitter servers, around 5-10 minutes will be enough. The auto-installer should install and configure ntp so this will not occur. Also, add this as a FAQ to the Wiki.
1.0
Auto-installer needs to install ntpdate - Twitter returns a crypted 401 Authorization Denied whenever the TCAT system clocktime is out-of-sync with the Twitter servers, around 5-10 minutes will be enough. The auto-installer should install and configure ntp so this will not occur. Also, add this as a FAQ to the Wiki.
non_comp
auto installer needs to install ntpdate twitter returns a crypted authorization denied whenever the tcat system clocktime is out of sync with the twitter servers around minutes will be enough the auto installer should install and configure ntp so this will not occur also add this as a faq to the wiki
0
18,417
25,485,400,978
IssuesEvent
2022-11-26 10:16:51
eloquentarduino/EloquentTinyML
https://api.github.com/repos/eloquentarduino/EloquentTinyML
closed
Didn't find op for builtin opcode 'CONV_2D' version '5' Failed to get registration from op code d
compatibility
@eloquentarduino I am trying to take an inference for a neural network on ESP 32 DEV KIT but it's throwing an error that is not understandable. Please find my code below: #include <EloquentTinyML.h> #include <eloquent_tinyml/tensorflow.h> #include "test_model.h" #define NUMBER_OF_INPUTS 7200 #define NUMBER_OF_OUTPUTS 4 #define TENSOR_ARENA_SIZE 2*1024 Eloquent::TinyML::TensorFlow::TensorFlow<NUMBER_OF_INPUTS, NUMBER_OF_OUTPUTS, TENSOR_ARENA_SIZE> ml; void setup() { Serial.begin(115200); ml.begin(test_model); } void loop() { float x_test[7200] = {0.000000, 0.000139, ........ , 0.999722, 0.999861}; float y_pred[4] = {0, 0, 0, 0}; float predicted = ml.predict(x_test, y_pred); Serial.print("\t predicted: "); Serial.println(predicted); delay(10000); } below is my model summary: ![image](https://user-images.githubusercontent.com/74509896/202165526-6dbf7888-b900-407c-8d35-84d6856cd437.png)
True
Didn't find op for builtin opcode 'CONV_2D' version '5' Failed to get registration from op code d - @eloquentarduino I am trying to take an inference for a neural network on ESP 32 DEV KIT but it's throwing an error that is not understandable. Please find my code below: #include <EloquentTinyML.h> #include <eloquent_tinyml/tensorflow.h> #include "test_model.h" #define NUMBER_OF_INPUTS 7200 #define NUMBER_OF_OUTPUTS 4 #define TENSOR_ARENA_SIZE 2*1024 Eloquent::TinyML::TensorFlow::TensorFlow<NUMBER_OF_INPUTS, NUMBER_OF_OUTPUTS, TENSOR_ARENA_SIZE> ml; void setup() { Serial.begin(115200); ml.begin(test_model); } void loop() { float x_test[7200] = {0.000000, 0.000139, ........ , 0.999722, 0.999861}; float y_pred[4] = {0, 0, 0, 0}; float predicted = ml.predict(x_test, y_pred); Serial.print("\t predicted: "); Serial.println(predicted); delay(10000); } below is my model summary: ![image](https://user-images.githubusercontent.com/74509896/202165526-6dbf7888-b900-407c-8d35-84d6856cd437.png)
comp
didn t find op for builtin opcode conv version failed to get registration from op code d eloquentarduino i am trying to take an inference for a neural network on esp dev kit but it s throwing an error that is not understandable please find my code below include include include test model h define number of inputs define number of outputs define tensor arena size eloquent tinyml tensorflow tensorflow ml void setup serial begin ml begin test model void loop float x test float y pred float predicted ml predict x test y pred serial print t predicted serial println predicted delay below is my model summary
1
13,106
15,394,077,924
IssuesEvent
2021-03-03 17:27:42
docker/compose-cli
https://api.github.com/repos/docker/compose-cli
closed
Compose: support --build-arg key=val
compatibility compose
build option: `--build-arg key=val Set build-time variables for services.`
True
Compose: support --build-arg key=val - build option: `--build-arg key=val Set build-time variables for services.`
comp
compose support build arg key val build option build arg key val set build time variables for services
1
269,368
28,960,116,323
IssuesEvent
2023-05-10 01:16:18
dpteam/RK3188_TABLET
https://api.github.com/repos/dpteam/RK3188_TABLET
reopened
CVE-2019-10220 (High) detected in linuxv3.0
Mend: dependency security vulnerability
## CVE-2019-10220 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv3.0</b></p></summary> <p> <p>Linux kernel source tree</p> <p>Library home page: <a href=https://github.com/verygreen/linux.git>https://github.com/verygreen/linux.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/dpteam/RK3188_TABLET/commit/0c501f5a0fd72c7b2ac82904235363bd44fd8f9e">0c501f5a0fd72c7b2ac82904235363bd44fd8f9e</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/readdir.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/readdir.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/readdir.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> Linux kernel CIFS implementation, version 4.9.0 is vulnerable to a relative paths injection in directory entry lists. <p>Publish Date: 2019-11-27 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-10220>CVE-2019-10220</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-10220">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-10220</a></p> <p>Release Date: 2019-11-27</p> <p>Fix Resolution: v5.4-rc2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-10220 (High) detected in linuxv3.0 - ## CVE-2019-10220 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv3.0</b></p></summary> <p> <p>Linux kernel source tree</p> <p>Library home page: <a href=https://github.com/verygreen/linux.git>https://github.com/verygreen/linux.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/dpteam/RK3188_TABLET/commit/0c501f5a0fd72c7b2ac82904235363bd44fd8f9e">0c501f5a0fd72c7b2ac82904235363bd44fd8f9e</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/readdir.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/readdir.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/readdir.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> Linux kernel CIFS implementation, version 4.9.0 is vulnerable to a relative paths injection in directory entry lists. <p>Publish Date: 2019-11-27 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-10220>CVE-2019-10220</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-10220">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-10220</a></p> <p>Release Date: 2019-11-27</p> <p>Fix Resolution: v5.4-rc2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_comp
cve high detected in cve high severity vulnerability vulnerable library linux kernel source tree library home page a href found in head commit a href found in base branch master vulnerable source files fs readdir c fs readdir c fs readdir c vulnerability details linux kernel cifs implementation version is vulnerable to a relative paths injection in directory entry lists publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
7,956
10,135,788,178
IssuesEvent
2019-08-02 11:07:40
storybookjs/storybook
https://api.github.com/repos/storybookjs/storybook
closed
React-Jss injectSheet overriden by Material ui V4
compatibility with other tools cra inactive question / support
Hi, React-jss injectSheet is injected before Material ui styles so it's overridden . it's only happen in Storybook and not in production code. this is my configuration: - "@storybook/addon-actions": "5.1.1" - "@storybook/react": "5.1.1" - "react-scripts": "3.0.1" - "react-jss": "8.6.1" - "@material-ui/core": "4.1.0" I'm not working with hooks but on classes so I'm using "injectSheet" that wrap my components. ![jss](https://user-images.githubusercontent.com/26362217/59254212-732e9280-8c38-11e9-913e-369b0cb8cf98.PNG) thanks
True
React-Jss injectSheet overriden by Material ui V4 - Hi, React-jss injectSheet is injected before Material ui styles so it's overridden . it's only happen in Storybook and not in production code. this is my configuration: - "@storybook/addon-actions": "5.1.1" - "@storybook/react": "5.1.1" - "react-scripts": "3.0.1" - "react-jss": "8.6.1" - "@material-ui/core": "4.1.0" I'm not working with hooks but on classes so I'm using "injectSheet" that wrap my components. ![jss](https://user-images.githubusercontent.com/26362217/59254212-732e9280-8c38-11e9-913e-369b0cb8cf98.PNG) thanks
comp
react jss injectsheet overriden by material ui hi react jss injectsheet is injected before material ui styles so it s overridden it s only happen in storybook and not in production code this is my configuration storybook addon actions storybook react react scripts react jss material ui core i m not working with hooks but on classes so i m using injectsheet that wrap my components thanks
1
120,761
25,859,119,640
IssuesEvent
2022-12-13 15:43:52
pnp/pnpjs
https://api.github.com/repos/pnp/pnpjs
closed
How to delete a term from term store using PnP JS
type: enhancement area: code status: complete
- [ ] Enhancement - [ ] Bug - [ X ] Question - [ ] Documentation gap/issue Please specify what version of the library you are using: [ 2.12.0 ] Please specify what version(s) of SharePoint you are targeting: [ SharePoint Online ] Hello, How can we delete a term from a term store using PnP JS? I took a look at https://pnp.github.io/pnpjs/sp/taxonomy/ and found nothing, Can you help? If it is not possible, what is the alternative? Call SharePoint REST API? Call old SOAP web services from javascript? Thanks
1.0
How to delete a term from term store using PnP JS - - [ ] Enhancement - [ ] Bug - [ X ] Question - [ ] Documentation gap/issue Please specify what version of the library you are using: [ 2.12.0 ] Please specify what version(s) of SharePoint you are targeting: [ SharePoint Online ] Hello, How can we delete a term from a term store using PnP JS? I took a look at https://pnp.github.io/pnpjs/sp/taxonomy/ and found nothing, Can you help? If it is not possible, what is the alternative? Call SharePoint REST API? Call old SOAP web services from javascript? Thanks
non_comp
how to delete a term from term store using pnp js enhancement bug question documentation gap issue please specify what version of the library you are using please specify what version s of sharepoint you are targeting hello how can we delete a term from a term store using pnp js i took a look at and found nothing can you help if it is not possible what is the alternative call sharepoint rest api call old soap web services from javascript thanks
0
11,530
13,526,943,000
IssuesEvent
2020-09-15 14:50:03
zymex22/Project-RimFactory-Revived
https://api.github.com/repos/zymex22/Project-RimFactory-Revived
opened
SeedsPlease incompatibility
Incompatibility
# Describe the bug The cultivators are incompatible with the `SeedsPlease` Mod. They ignore the System where seeds are required in order to plant. Link to the Mod: https://steamcommunity.com/sharedfiles/filedetails/?id=935732834 Patches if possible are required for: - [ ] Drone Cultivator This might be already resolved with the Drone Overhaul #80 (Check Required) - [ ] Mechanites Cultivator --- Reported on steam by `Extrien`
True
SeedsPlease incompatibility - # Describe the bug The cultivators are incompatible with the `SeedsPlease` Mod. They ignore the System where seeds are required in order to plant. Link to the Mod: https://steamcommunity.com/sharedfiles/filedetails/?id=935732834 Patches if possible are required for: - [ ] Drone Cultivator This might be already resolved with the Drone Overhaul #80 (Check Required) - [ ] Mechanites Cultivator --- Reported on steam by `Extrien`
comp
seedsplease incompatibility describe the bug the cultivators are incompatible with the seedsplease mod they ignore the system where seeds are required in order to plant link to the mod patches if possible are required for drone cultivator this might be already resolved with the drone overhaul check required mechanites cultivator reported on steam by extrien
1
5,818
8,269,563,062
IssuesEvent
2018-09-15 07:40:29
arcticicestudio/igloo
https://api.github.com/repos/arcticicestudio/igloo
opened
taskopen workaround support for macOS
os-macos scope-compatibility snowblock-taskwarrior snowblock-timewarrior type-feature
> Epic: #131 > Related to #110 The management of installed [Perl modules][cpan-doc-modules] on macOS is not as simple and well thought through like with package managers on Linux systems, e.g. via [pacman][archw-pacman] on [Arch Linux][archlinux]. There are problems when is comes to configuring the runtime path the modules have been installed to even when using the most popular module manager called [cpanminus][]. This causes the [Taskwarrior][] plugin [taskopen][] fail to load because the Perl core module `JSON` can't be found and loaded. Next to this the As a workaround a custom script should be implemented to create and open a attached note of an task: 1. Use the [`_get`][tw-doc-api-_get] function of the [Taskwarrior DOM API][tw-doc-dom-api] to extract any stored piece of information of an task. This allows to receive the [UUID of an task][tw-doc-ids]. 2. Create a custom `on` (open note) [Taskwarrior alias][tw-doc-alias] to run the implemented custom script via the `execute` command. The logic of the script should follow the same like taskopen uses for default notes: * Use the [UUID of an task][tw-doc-ids] as the note filename. * Simply pass the file to an editor (in this case [Atom][]) which will… * …create a new file if it doesn't exist yet. * …open the file if it already exists. Note that **this script is not limited to macOS** but can also be used for any other Linux host! It is only necessary to use it on macOS due to the problems described above. [archlinux]: https://archlinux.org [archw-pacman]: https://wiki.archlinux.org/index.php/Pacman [atom]: https://atom.io [cpanminus]: https://github.com/miyagawa/cpanminus [cpan-doc-modules]: http://www.cpan.org/modules [taskopen]: https://github.com/ValiValpas/taskopen [taskwarrior]: https://taskwarrior.org [tw-doc-alias]: https://taskwarrior.org/docs/terminology.html#alias [tw-doc-api-_get]: https://taskwarrior.org/docs/commands/_get.html [tw-doc-dom-api]: https://taskwarrior.org/docs/dom.html [tw-doc-ids]: https://taskwarrior.org/docs/ids.html
True
taskopen workaround support for macOS - > Epic: #131 > Related to #110 The management of installed [Perl modules][cpan-doc-modules] on macOS is not as simple and well thought through like with package managers on Linux systems, e.g. via [pacman][archw-pacman] on [Arch Linux][archlinux]. There are problems when is comes to configuring the runtime path the modules have been installed to even when using the most popular module manager called [cpanminus][]. This causes the [Taskwarrior][] plugin [taskopen][] fail to load because the Perl core module `JSON` can't be found and loaded. Next to this the As a workaround a custom script should be implemented to create and open a attached note of an task: 1. Use the [`_get`][tw-doc-api-_get] function of the [Taskwarrior DOM API][tw-doc-dom-api] to extract any stored piece of information of an task. This allows to receive the [UUID of an task][tw-doc-ids]. 2. Create a custom `on` (open note) [Taskwarrior alias][tw-doc-alias] to run the implemented custom script via the `execute` command. The logic of the script should follow the same like taskopen uses for default notes: * Use the [UUID of an task][tw-doc-ids] as the note filename. * Simply pass the file to an editor (in this case [Atom][]) which will… * …create a new file if it doesn't exist yet. * …open the file if it already exists. Note that **this script is not limited to macOS** but can also be used for any other Linux host! It is only necessary to use it on macOS due to the problems described above. [archlinux]: https://archlinux.org [archw-pacman]: https://wiki.archlinux.org/index.php/Pacman [atom]: https://atom.io [cpanminus]: https://github.com/miyagawa/cpanminus [cpan-doc-modules]: http://www.cpan.org/modules [taskopen]: https://github.com/ValiValpas/taskopen [taskwarrior]: https://taskwarrior.org [tw-doc-alias]: https://taskwarrior.org/docs/terminology.html#alias [tw-doc-api-_get]: https://taskwarrior.org/docs/commands/_get.html [tw-doc-dom-api]: https://taskwarrior.org/docs/dom.html [tw-doc-ids]: https://taskwarrior.org/docs/ids.html
comp
taskopen workaround support for macos epic related to the management of installed on macos is not as simple and well thought through like with package managers on linux systems e g via on there are problems when is comes to configuring the runtime path the modules have been installed to even when using the most popular module manager called this causes the plugin fail to load because the perl core module json can t be found and loaded next to this the as a workaround a custom script should be implemented to create and open a attached note of an task use the function of the to extract any stored piece of information of an task this allows to receive the create a custom on open note to run the implemented custom script via the execute command the logic of the script should follow the same like taskopen uses for default notes use the as the note filename simply pass the file to an editor in this case which will… …create a new file if it doesn t exist yet …open the file if it already exists note that this script is not limited to macos but can also be used for any other linux host it is only necessary to use it on macos due to the problems described above
1
9,993
2,616,018,384
IssuesEvent
2015-03-02 01:00:04
jasonhall/bwapi
https://api.github.com/repos/jasonhall/bwapi
closed
auto-join Local PC game, second instance crashes on some config
auto-migrated Component-Logic Milestone-Release Priority-Critical Type-Defect Usability
``` What steps will reproduce the problem? 1.in bwapi.ini: set game = BWAPI 2. set map = maps\AI\tournament 1\dragoons.scm (from AIIDE 2010 tournament 1) 3. set game_type = USE_MAP_SETTINGS 4. set both ai's to skynet (uses revision 3769) 5. set auto_manu = LAN 6. launch the games (one after another with a small delay) What is the expected output? What do you see instead? Expected to automatically start a match using the provided map Instead, both instances create a separate game called BWAPI. What version of the product are you using? On what operating system? BWAPI 3.7.3 under Windows 7 SP1 Please provide any additional information below. When leaving 'game' blank - it doesn't seem to affect the algorithm When leaving 'map' blank - the game reaches the "select a game" menu, and after selecting a map (in the first instance) the second instance automatically attempts to join it, but without success (the second client freezes) ``` Original issue reported on code.google.com by `Leonid.T...@gmail.com` on 2 Jun 2012 at 10:40 * Merged into: #443
1.0
auto-join Local PC game, second instance crashes on some config - ``` What steps will reproduce the problem? 1.in bwapi.ini: set game = BWAPI 2. set map = maps\AI\tournament 1\dragoons.scm (from AIIDE 2010 tournament 1) 3. set game_type = USE_MAP_SETTINGS 4. set both ai's to skynet (uses revision 3769) 5. set auto_manu = LAN 6. launch the games (one after another with a small delay) What is the expected output? What do you see instead? Expected to automatically start a match using the provided map Instead, both instances create a separate game called BWAPI. What version of the product are you using? On what operating system? BWAPI 3.7.3 under Windows 7 SP1 Please provide any additional information below. When leaving 'game' blank - it doesn't seem to affect the algorithm When leaving 'map' blank - the game reaches the "select a game" menu, and after selecting a map (in the first instance) the second instance automatically attempts to join it, but without success (the second client freezes) ``` Original issue reported on code.google.com by `Leonid.T...@gmail.com` on 2 Jun 2012 at 10:40 * Merged into: #443
non_comp
auto join local pc game second instance crashes on some config what steps will reproduce the problem in bwapi ini set game bwapi set map maps ai tournament dragoons scm from aiide tournament set game type use map settings set both ai s to skynet uses revision set auto manu lan launch the games one after another with a small delay what is the expected output what do you see instead expected to automatically start a match using the provided map instead both instances create a separate game called bwapi what version of the product are you using on what operating system bwapi under windows please provide any additional information below when leaving game blank it doesn t seem to affect the algorithm when leaving map blank the game reaches the select a game menu and after selecting a map in the first instance the second instance automatically attempts to join it but without success the second client freezes original issue reported on code google com by leonid t gmail com on jun at merged into
0
41,376
5,354,144,458
IssuesEvent
2017-02-20 09:01:15
dzhw/metadatamanagement
https://api.github.com/repos/dzhw/metadatamanagement
closed
Rework question management
category:questionmanagement points:13 prio:1 status:testing type:backlog item
In order to hide all ids from the user we need to implement the following changes ## Domain Object - [x] add instrumentNumber to domain object (including search document) - [x] add successorNumbers to domain object (including search document) - [x] remove surveyId - [x] cross check with current domain model ## Validation - [x] change validation of question.id ~~- [ ] change validation of question.number~~ - [x] add validation for instrumentNumber - [x] remove validation of instrumentId - [x] cross check with current documentation in wiki ## Post-Validation - [x] remove surveyId validation ## Import The import has to be reworked: - [x] Import by instrument directory - [x] Generate questionId from projectId, instrumentNumber and question.number - [x] Generate instrumentId from projectId, and instrumentNumber - [x] Allow upload of images having only the number in the file name (e.g. 1.png) ## GUI - [x] Change client urls for question details to /de/studies/{projectId}/instruments/{instrumentNumber}/questions/{questionNumber} including search results cards
1.0
Rework question management - In order to hide all ids from the user we need to implement the following changes ## Domain Object - [x] add instrumentNumber to domain object (including search document) - [x] add successorNumbers to domain object (including search document) - [x] remove surveyId - [x] cross check with current domain model ## Validation - [x] change validation of question.id ~~- [ ] change validation of question.number~~ - [x] add validation for instrumentNumber - [x] remove validation of instrumentId - [x] cross check with current documentation in wiki ## Post-Validation - [x] remove surveyId validation ## Import The import has to be reworked: - [x] Import by instrument directory - [x] Generate questionId from projectId, instrumentNumber and question.number - [x] Generate instrumentId from projectId, and instrumentNumber - [x] Allow upload of images having only the number in the file name (e.g. 1.png) ## GUI - [x] Change client urls for question details to /de/studies/{projectId}/instruments/{instrumentNumber}/questions/{questionNumber} including search results cards
non_comp
rework question management in order to hide all ids from the user we need to implement the following changes domain object add instrumentnumber to domain object including search document add successornumbers to domain object including search document remove surveyid cross check with current domain model validation change validation of question id change validation of question number add validation for instrumentnumber remove validation of instrumentid cross check with current documentation in wiki post validation remove surveyid validation import the import has to be reworked import by instrument directory generate questionid from projectid instrumentnumber and question number generate instrumentid from projectid and instrumentnumber allow upload of images having only the number in the file name e g png gui change client urls for question details to de studies projectid instruments instrumentnumber questions questionnumber including search results cards
0
264,593
28,209,200,903
IssuesEvent
2023-04-05 01:38:55
Trinadh465/linux_4.19.72_CVE-2023-42896
https://api.github.com/repos/Trinadh465/linux_4.19.72_CVE-2023-42896
closed
CVE-2021-0448 (Medium) detected in linuxlinux-4.19.279 - autoclosed
Mend: dependency security vulnerability
## CVE-2021-0448 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.279</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/Trinadh465/linux_4.19.72_CVE-2023-42896/commit/bab817ff3ba663d8c0b4d4ec9cb6c7069f775527">bab817ff3ba663d8c0b4d4ec9cb6c7069f775527</a></p> <p>Found in base branch: <b>main</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/netfilter/nf_conntrack_netlink.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A security vulnerability was found in Linux Kernel before 4.4.239, 4.9.239, 4.14.201, 4.19.150, 5.4.70, and 5.8.13. Missing range check for l3/l4 protonum in netfilter.c <p>Publish Date: 2020-11-07 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-0448>CVE-2021-0448</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2021-0448">https://www.linuxkernelcves.com/cves/CVE-2021-0448</a></p> <p>Release Date: 2020-11-07</p> <p>Fix Resolution: v4.4.239,v4.9.239,v4.14.201,v4.19.150,v5.4.70,v5.8.13</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-0448 (Medium) detected in linuxlinux-4.19.279 - autoclosed - ## CVE-2021-0448 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.279</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/Trinadh465/linux_4.19.72_CVE-2023-42896/commit/bab817ff3ba663d8c0b4d4ec9cb6c7069f775527">bab817ff3ba663d8c0b4d4ec9cb6c7069f775527</a></p> <p>Found in base branch: <b>main</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/netfilter/nf_conntrack_netlink.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A security vulnerability was found in Linux Kernel before 4.4.239, 4.9.239, 4.14.201, 4.19.150, 5.4.70, and 5.8.13. Missing range check for l3/l4 protonum in netfilter.c <p>Publish Date: 2020-11-07 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-0448>CVE-2021-0448</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2021-0448">https://www.linuxkernelcves.com/cves/CVE-2021-0448</a></p> <p>Release Date: 2020-11-07</p> <p>Fix Resolution: v4.4.239,v4.9.239,v4.14.201,v4.19.150,v5.4.70,v5.8.13</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_comp
cve medium detected in linuxlinux autoclosed cve medium severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch main vulnerable source files net netfilter nf conntrack netlink c vulnerability details a security vulnerability was found in linux kernel before and missing range check for protonum in netfilter c publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
11,463
13,447,301,844
IssuesEvent
2020-09-08 14:04:48
higan-emu/higan
https://api.github.com/repos/higan-emu/higan
closed
CGB: Sprite palette incorrect in Pokemon Crystal
bug compatibility
When playing Pokemon Crystal, the player sprite (and other character sprites walking around the map) appear in a random palette that changes as you move from room to room, or visit different menus. Here the player should have a black outline, instead of being a red blob: ![higan-ui 2020-08-22 19-02-35](https://user-images.githubusercontent.com/1450918/90952887-6f3c0880-e4aa-11ea-86ac-3c37dbb6a4ac.png) This doesn't appear on FitzRoy's game bug list in #47, so it may be a regression. We should pin that down.
True
CGB: Sprite palette incorrect in Pokemon Crystal - When playing Pokemon Crystal, the player sprite (and other character sprites walking around the map) appear in a random palette that changes as you move from room to room, or visit different menus. Here the player should have a black outline, instead of being a red blob: ![higan-ui 2020-08-22 19-02-35](https://user-images.githubusercontent.com/1450918/90952887-6f3c0880-e4aa-11ea-86ac-3c37dbb6a4ac.png) This doesn't appear on FitzRoy's game bug list in #47, so it may be a regression. We should pin that down.
comp
cgb sprite palette incorrect in pokemon crystal when playing pokemon crystal the player sprite and other character sprites walking around the map appear in a random palette that changes as you move from room to room or visit different menus here the player should have a black outline instead of being a red blob this doesn t appear on fitzroy s game bug list in so it may be a regression we should pin that down
1
14,731
18,091,678,120
IssuesEvent
2021-09-22 02:51:01
gambitph/Stackable
https://api.github.com/repos/gambitph/Stackable
opened
Container goes full width when Block background is on
bug [version] V3 v2 compatibility
<!-- Before posting, make sure that: 1. you are running the latest version of Stackable, and 2. you have searched whether your issue has already been reported --> **Describe the bug** All v2 blocks with block background goes full width once the block background is turned on. **To Reproduce** Steps to reproduce the behavior: 1. Add a v2 container block with Basic layout 2. Turn on the block background. Save. 3. See bug in Frontend **Screenshots** In v3 build: <img width="1541" alt="Screen Shot 2021-09-22 at 10 47 38 AM" src="https://user-images.githubusercontent.com/28699204/134275804-63f86d39-dd2b-447f-8891-c8aa3d02c71c.png"> **Expected behavior** It should not go full width (by default) when block background is turned on: <img width="1541" alt="Screen Shot 2021-09-22 at 10 47 47 AM" src="https://user-images.githubusercontent.com/28699204/134275823-64f79b51-8b0a-4b16-bb91-b729cc933d5b.png">
True
Container goes full width when Block background is on - <!-- Before posting, make sure that: 1. you are running the latest version of Stackable, and 2. you have searched whether your issue has already been reported --> **Describe the bug** All v2 blocks with block background goes full width once the block background is turned on. **To Reproduce** Steps to reproduce the behavior: 1. Add a v2 container block with Basic layout 2. Turn on the block background. Save. 3. See bug in Frontend **Screenshots** In v3 build: <img width="1541" alt="Screen Shot 2021-09-22 at 10 47 38 AM" src="https://user-images.githubusercontent.com/28699204/134275804-63f86d39-dd2b-447f-8891-c8aa3d02c71c.png"> **Expected behavior** It should not go full width (by default) when block background is turned on: <img width="1541" alt="Screen Shot 2021-09-22 at 10 47 47 AM" src="https://user-images.githubusercontent.com/28699204/134275823-64f79b51-8b0a-4b16-bb91-b729cc933d5b.png">
comp
container goes full width when block background is on before posting make sure that you are running the latest version of stackable and you have searched whether your issue has already been reported describe the bug all blocks with block background goes full width once the block background is turned on to reproduce steps to reproduce the behavior add a container block with basic layout turn on the block background save see bug in frontend screenshots in build img width alt screen shot at am src expected behavior it should not go full width by default when block background is turned on img width alt screen shot at am src
1
137,149
18,752,644,700
IssuesEvent
2021-11-05 05:43:42
madhans23/linux-4.15
https://api.github.com/repos/madhans23/linux-4.15
opened
CVE-2020-27171 (Medium) detected in linuxv4.15
security vulnerability
## CVE-2020-27171 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv4.15</b></p></summary> <p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/brodo/linux.git>https://git.kernel.org/pub/scm/linux/kernel/git/brodo/linux.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/madhans23/linux-4.15/commit/d96ee498864d1a0b6222cfb17d64ca8196014940">d96ee498864d1a0b6222cfb17d64ca8196014940</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in the Linux kernel before 5.11.8. kernel/bpf/verifier.c has an off-by-one error (with a resultant integer underflow) affecting out-of-bounds speculation on pointer arithmetic, leading to side-channel attacks that defeat Spectre mitigations and obtain sensitive information from kernel memory, aka CID-10d2bb2e6b1d. <p>Publish Date: 2021-03-20 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-27171>CVE-2020-27171</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.0</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Change files</p> <p>Origin: <a href="https://github.com/torvalds/linux/commit/10d2bb2e6b1d8c4576c56a748f697dbeb8388899">https://github.com/torvalds/linux/commit/10d2bb2e6b1d8c4576c56a748f697dbeb8388899</a></p> <p>Release Date: 2021-03-17</p> <p>Fix Resolution: Replace or update the following file: verifier.c</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-27171 (Medium) detected in linuxv4.15 - ## CVE-2020-27171 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv4.15</b></p></summary> <p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/brodo/linux.git>https://git.kernel.org/pub/scm/linux/kernel/git/brodo/linux.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/madhans23/linux-4.15/commit/d96ee498864d1a0b6222cfb17d64ca8196014940">d96ee498864d1a0b6222cfb17d64ca8196014940</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in the Linux kernel before 5.11.8. kernel/bpf/verifier.c has an off-by-one error (with a resultant integer underflow) affecting out-of-bounds speculation on pointer arithmetic, leading to side-channel attacks that defeat Spectre mitigations and obtain sensitive information from kernel memory, aka CID-10d2bb2e6b1d. <p>Publish Date: 2021-03-20 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-27171>CVE-2020-27171</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.0</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Change files</p> <p>Origin: <a href="https://github.com/torvalds/linux/commit/10d2bb2e6b1d8c4576c56a748f697dbeb8388899">https://github.com/torvalds/linux/commit/10d2bb2e6b1d8c4576c56a748f697dbeb8388899</a></p> <p>Release Date: 2021-03-17</p> <p>Fix Resolution: Replace or update the following file: verifier.c</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_comp
cve medium detected in cve medium severity vulnerability vulnerable library library home page a href found in head commit a href found in base branch master vulnerable source files vulnerability details an issue was discovered in the linux kernel before kernel bpf verifier c has an off by one error with a resultant integer underflow affecting out of bounds speculation on pointer arithmetic leading to side channel attacks that defeat spectre mitigations and obtain sensitive information from kernel memory aka cid publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact high for more information on scores click a href suggested fix type change files origin a href release date fix resolution replace or update the following file verifier c step up your open source security game with whitesource
0
2,132
2,603,976,845
IssuesEvent
2015-02-24 19:01:43
chrsmith/nishazi6
https://api.github.com/repos/chrsmith/nishazi6
opened
沈阳龟头长疙瘩怎么办
auto-migrated Priority-Medium Type-Defect
``` 沈阳龟头长疙瘩怎么办〓沈陽軍區政治部醫院性病〓TEL:024-3 1023308〓成立于1946年,68年專注于性傳播疾病的研究和治療。� ��于沈陽市沈河區二緯路32號。是一所與新中國同建立共輝煌� ��歷史悠久、設備精良、技術權威、專家云集,是預防、保健 、醫療、科研康復為一體的綜合性醫院。是國家首批公立甲�� �部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、� ��南大學等知名高等院校的教學醫院。曾被中國人民解放軍空 軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體�� �等功。 ``` ----- Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:23
1.0
沈阳龟头长疙瘩怎么办 - ``` 沈阳龟头长疙瘩怎么办〓沈陽軍區政治部醫院性病〓TEL:024-3 1023308〓成立于1946年,68年專注于性傳播疾病的研究和治療。� ��于沈陽市沈河區二緯路32號。是一所與新中國同建立共輝煌� ��歷史悠久、設備精良、技術權威、專家云集,是預防、保健 、醫療、科研康復為一體的綜合性醫院。是國家首批公立甲�� �部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、� ��南大學等知名高等院校的教學醫院。曾被中國人民解放軍空 軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體�� �等功。 ``` ----- Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:23
non_comp
沈阳龟头长疙瘩怎么办 沈阳龟头长疙瘩怎么办〓沈陽軍區政治部醫院性病〓tel: 〓 , 。� �� 。是一所與新中國同建立共輝煌� ��歷史悠久、設備精良、技術權威、專家云集,是預防、保健 、醫療、科研康復為一體的綜合性醫院。是國家首批公立甲�� �部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、� ��南大學等知名高等院校的教學醫院。曾被中國人民解放軍空 軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體�� �等功。 original issue reported on code google com by gmail com on jun at
0
106,194
4,264,373,797
IssuesEvent
2016-07-12 06:49:46
openshift/origin
https://api.github.com/repos/openshift/origin
closed
chown Operation not permitted only when using persistent volume
component/storage kind/question priority/P2
I can able to deployed and use [gogs docker image](https://hub.docker.com/r/gogs/gogs/) in OpenShift Origin as ephemeral. But when using persistent volume I get `chown Operation not permitted` ##### Version oc v1.3.0-alpha.2 kubernetes v1.3.0-alpha.1-331-g0522e63 ##### Steps To Reproduce 1. Create two persistent volume matching the claim in below template 2. `oadm policy add-scc-to-user anyuid -z default` 3. `chcon -Rt svirt_sandbox_file_t volume_path` may be not required 4. Use the below template ``` apiVersion: v1 kind: Template metadata: name: gogs-persistent objects: - apiVersion: v1 kind: PersistentVolumeClaim metadata: name: claim-gogs spec: accessModes: - ReadWriteOnce resources: requests: storage: 1Gi - apiVersion: v1 kind: Pod metadata: name: gogs labels: name: gogs spec: containers: - image: gogs/gogs name: gogs ports: - containerPort: 3000 name: gogs - containerPort: 22 name: gogsssh volumeMounts: - name: gogs-persistent-storage mountPath: /data volumes: - name: gogs-persistent-storage persistentVolumeClaim: claimName: claim-gogs - apiVersion: v1 kind: Service metadata: labels: name: gogs name: gogs spec: ports: - port: 3000 name: gogs - port: 22 name: gogsssh selector: name: gogs - apiVersion: v1 kind: PersistentVolumeClaim metadata: name: claim-gogs-mysql spec: accessModes: - ReadWriteOnce resources: requests: storage: 1Gi - apiVersion: v1 kind: Pod metadata: name: mysql labels: name: mysql spec: containers: - image: openshift/mysql-55-centos7 name: mysql env: - name: MYSQL_ROOT_PASSWORD value: toor - name: MYSQL_USER value: gogs_user - name: MYSQL_PASSWORD value: gogs_pass - name: MYSQL_DATABASE value: gogs_db ports: - containerPort: 3306 name: mysql volumeMounts: - name: mysql-persistent-storage mountPath: /var/lib/mysql/data volumes: - name: mysql-persistent-storage persistentVolumeClaim: claimName: claim-gogs-mysql - apiVersion: v1 kind: Service metadata: labels: name: mysql name: mysql spec: ports: - port: 3306 name: mysql selector: name: mysql ``` ##### Current Result `$ oc logs pod gogs` ``` init:socat | Can't bind linked container GOGS to localhost, port 22 already in use init:socat | Linked container KUBERNETES will be binded to localhost on port 53 init:socat | Can't bind linked container GOGS to localhost, port 3000 already in use init:socat | Linked container MYSQL will be binded to localhost on port 3306 init:socat | Linked container KUBERNETES will be binded to localhost on port 443 Jul 7 16:28:20 syslogd started: BusyBox v1.24.2 chown: /data/gogs/data: Operation not permitted chown: /data/gogs/data: Operation not permitted chown: /data/gogs/conf: Operation not permitted chown: /data/gogs/conf: Operation not permitted chown: /data/gogs/log: Operation not permitted chown: /data/gogs/log: Operation not permitted chown: /data/gogs: Operation not permitted chown: /data/gogs: Operation not permitted chown: /data/git/.ssh/environment: Operation not permitted chown: /data/git/.ssh: Operation not permitted chown: /data/git/.ssh: Operation not permitted chown: /data/git: Operation not permitted chown: /data/git: Operation not permitted chown: /data/ssh/ssh_host_rsa_key: Operation not permitted chown: /data/ssh: Operation not permitted chown: /data/ssh: Operation not permitted chown: /data: Operation not permitted chown: /data: Operation not permitted chown: /data/ssh/ssh_host_dsa_key: Operation not permitted chown: /data/ssh/ssh_host_dsa_key.pub: Operation not permitted chown: /data/ssh/ssh_host_ecdsa_key: Operation not permitted chown: /data/ssh/ssh_host_ecdsa_key.pub: Operation not permitted chown: /data/ssh/ssh_host_ed25519_key: Operation not permitted chown: /data/ssh/ssh_host_ed25519_key.pub: Operation not permitted chown: /data/ssh/ssh_host_rsa_key: Operation not permitted chown: /data/ssh/ssh_host_rsa_key.pub: Operation not permitted Jul 7 16:28:22 sshd[49]: Server listening on :: port 22. Jul 7 16:28:22 sshd[49]: Server listening on 0.0.0.0 port 22. chown: /data/git/.ssh/environment: Operation not permitted chown: /data/git/.ssh: Operation not permitted chown: /data/git/.ssh: Operation not permitted chown: /data/git/: Operation not permitted chown: /data/git/: Operation not permitted chmod: /data: Operation not permitted 2016/07/07 16:28:23 [W] Custom config '/data/gogs/conf/app.ini' not found, ignore this if you're running first time 2016/07/07 16:28:23 [T] Custom path: /data/gogs 2016/07/07 16:28:23 [T] Log path: /app/gogs/log 2016/07/07 16:28:23 [I] Gogs: Go Git Service 0.9.35.0702 2016/07/07 16:28:23 [I] Build Time: 2016-07-02 04:16:49 UTC 2016/07/07 16:28:23 [I] Build Git Hash:  2016/07/07 16:28:23 [I] Log Mode: Console(Trace) 2016/07/07 16:28:23 [I] Cache Service Enabled 2016/07/07 16:28:23 [I] Session Service Enabled 2016/07/07 16:28:23 [I] SQLite3 Supported 2016/07/07 16:28:23 [I] Run Mode: Development 2016/07/07 16:28:26 [I] Listen: http://0.0.0.0:3000 ``` ##### Additional Information `$ oadm diagnostics` ``` [Note] Determining if client configuration exists for client/cluster diagnostics Info: Successfully read a client config file at '/root/.kube/config' Info: Using context for cluster-admin access: 'gogs/192-168-0-55:8443/system:admin' [Note] Running diagnostic: ConfigContexts[wordpress/192-168-0-55:8443/system:admin] Description: Validate client config context is complete and has connectivity Info: For client config context 'wordpress/192-168-0-55:8443/system:admin': The server URL is 'https://192.168.0.55:8443' The user authentication is 'system:admin/192-168-0-55:8443' The current project is 'wordpress' Successfully requested project list; has access to project(s): [default gogs myproject openshift openshift-infra] [Note] Running diagnostic: ConfigContexts[gogs/192-168-0-55:8443/developer] Description: Validate client config context is complete and has connectivity Info: For client config context 'gogs/192-168-0-55:8443/developer': The server URL is 'https://192.168.0.55:8443' The user authentication is 'developer/192-168-0-55:8443' The current project is 'gogs' Successfully requested project list; has access to project(s): [gogs myproject] [Note] Running diagnostic: DiagnosticPod Description: Create a pod to run diagnostics from the application standpoint Info: Output from the diagnostic pod (image openshift/origin-deployer:v1.3.0-alpha.2): [Note] Running diagnostic: PodCheckAuth Description: Check that service account credentials authenticate as expected Info: Service account token successfully authenticated to master Info: Service account token was authenticated by the integrated registry. [Note] Running diagnostic: PodCheckDns Description: Check that DNS within a pod works as expected [Note] Summary of diagnostics execution (version v1.3.0-alpha.2): [Note] Completed with no errors or warnings seen. [Note] Running diagnostic: ClusterRegistry Description: Check that there is a working Docker registry [Note] Running diagnostic: ClusterRoleBindings Description: Check that the default ClusterRoleBindings are present and contain the expected subjects [Note] Running diagnostic: ClusterRoles Description: Check that the default ClusterRoles are present and contain the expected permissions ERROR: [CRD1005 from diagnostic ClusterRoles@openshift/origin/pkg/diagnostics/cluster/roles.go:90] clusterrole/cluster-reader has changed and the existing role does not have enough permissions. Use the `oadm policy reconcile-cluster-roles` command to update the role. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["list"], APIGroups:[""], Resources:["minions"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["watch"], APIGroups:[""], Resources:["minions"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["get"], APIGroups:[""], Resources:["minions"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["get"], APIGroups:[""], Resources:["subjectaccessreviews"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["list"], APIGroups:[""], Resources:["subjectaccessreviews"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["watch"], APIGroups:[""], Resources:["subjectaccessreviews"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["watch"], APIGroups:[""], Resources:["builds/clone"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["get"], APIGroups:[""], Resources:["builds/clone"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["list"], APIGroups:[""], Resources:["builds/clone"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["get"], APIGroups:[""], Resources:["buildconfigs/instantiate"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["list"], APIGroups:[""], Resources:["buildconfigs/instantiate"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["watch"], APIGroups:[""], Resources:["buildconfigs/instantiate"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["get"], APIGroups:[""], Resources:["localsubjectaccessreviews"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["list"], APIGroups:[""], Resources:["localsubjectaccessreviews"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["watch"], APIGroups:[""], Resources:["localsubjectaccessreviews"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["get"], APIGroups:[""], Resources:["localresourceaccessreviews"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["list"], APIGroups:[""], Resources:["localresourceaccessreviews"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["watch"], APIGroups:[""], Resources:["localresourceaccessreviews"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["get"], APIGroups:[""], Resources:["resourceaccessreviews"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["list"], APIGroups:[""], Resources:["resourceaccessreviews"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["watch"], APIGroups:[""], Resources:["resourceaccessreviews"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["get"], APIGroups:[""], Resources:["generatedeploymentconfigs"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["list"], APIGroups:[""], Resources:["generatedeploymentconfigs"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["watch"], APIGroups:[""], Resources:["generatedeploymentconfigs"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["get"], APIGroups:[""], Resources:["imagestreamimports"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["list"], APIGroups:[""], Resources:["imagestreamimports"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["watch"], APIGroups:[""], Resources:["imagestreamimports"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["list"], APIGroups:[""], Resources:["imagestreammappings"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["watch"], APIGroups:[""], Resources:["imagestreammappings"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["get"], APIGroups:[""], Resources:["imagestreammappings"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["get"], APIGroups:[""], Resources:["oauthclients"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["list"], APIGroups:[""], Resources:["oauthclients"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["watch"], APIGroups:[""], Resources:["oauthclients"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["list"], APIGroups:[""], Resources:["deploymentconfigrollbacks"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["watch"], APIGroups:[""], Resources:["deploymentconfigrollbacks"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["get"], APIGroups:[""], Resources:["deploymentconfigrollbacks"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["get"], APIGroups:[""], Resources:["buildconfigs/instantiatebinary"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["list"], APIGroups:[""], Resources:["buildconfigs/instantiatebinary"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["watch"], APIGroups:[""], Resources:["buildconfigs/instantiatebinary"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["watch"], APIGroups:[""], Resources:["selfsubjectrulesreviews"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["get"], APIGroups:[""], Resources:["selfsubjectrulesreviews"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["list"], APIGroups:[""], Resources:["selfsubjectrulesreviews"]}. WARN: [CRD1003 from diagnostic ClusterRoles@openshift/origin/pkg/diagnostics/cluster/roles.go:82] clusterrole/admin has changed, but the existing role has more permissions than the new role. Use the `oadm policy reconcile-cluster-roles` command to update the role to reduce permissions. Info: clusterrole/admin has extra permission PolicyRule{Verbs:["patch"], APIGroups:[""], Resources:["deploymentconfigs/rollback"]}. Info: clusterrole/admin has extra permission PolicyRule{Verbs:["update"], APIGroups:[""], Resources:["deploymentconfigs/rollback"]}. Info: clusterrole/admin has extra permission PolicyRule{Verbs:["watch"], APIGroups:[""], Resources:["deploymentconfigs/rollback"]}. Info: clusterrole/admin has extra permission PolicyRule{Verbs:["create"], APIGroups:[""], Resources:["deploymentconfigs/rollback"]}. Info: clusterrole/admin has extra permission PolicyRule{Verbs:["delete"], APIGroups:[""], Resources:["deploymentconfigs/rollback"]}. Info: clusterrole/admin has extra permission PolicyRule{Verbs:["deletecollection"], APIGroups:[""], Resources:["deploymentconfigs/rollback"]}. Info: clusterrole/admin has extra permission PolicyRule{Verbs:["get"], APIGroups:[""], Resources:["deploymentconfigs/rollback"]}. Info: clusterrole/admin has extra permission PolicyRule{Verbs:["list"], APIGroups:[""], Resources:["deploymentconfigs/rollback"]}. WARN: [CRD1003 from diagnostic ClusterRoles@openshift/origin/pkg/diagnostics/cluster/roles.go:82] clusterrole/edit has changed, but the existing role has more permissions than the new role. Use the `oadm policy reconcile-cluster-roles` command to update the role to reduce permissions. Info: clusterrole/edit has extra permission PolicyRule{Verbs:["get"], APIGroups:[""], Resources:["deploymentconfigs/rollback"]}. Info: clusterrole/edit has extra permission PolicyRule{Verbs:["list"], APIGroups:[""], Resources:["deploymentconfigs/rollback"]}. Info: clusterrole/edit has extra permission PolicyRule{Verbs:["patch"], APIGroups:[""], Resources:["deploymentconfigs/rollback"]}. Info: clusterrole/edit has extra permission PolicyRule{Verbs:["update"], APIGroups:[""], Resources:["deploymentconfigs/rollback"]}. Info: clusterrole/edit has extra permission PolicyRule{Verbs:["watch"], APIGroups:[""], Resources:["deploymentconfigs/rollback"]}. Info: clusterrole/edit has extra permission PolicyRule{Verbs:["create"], APIGroups:[""], Resources:["deploymentconfigs/rollback"]}. Info: clusterrole/edit has extra permission PolicyRule{Verbs:["delete"], APIGroups:[""], Resources:["deploymentconfigs/rollback"]}. Info: clusterrole/edit has extra permission PolicyRule{Verbs:["deletecollection"], APIGroups:[""], Resources:["deploymentconfigs/rollback"]}. ERROR: [CRD1005 from diagnostic ClusterRoles@openshift/origin/pkg/diagnostics/cluster/roles.go:90] clusterrole/view has changed and the existing role does not have enough permissions. Use the `oadm policy reconcile-cluster-roles` command to update the role. Info: clusterrole/view is missing permission PolicyRule{Verbs:["get"], APIGroups:[""], Resources:["generatedeploymentconfigs"]}. Info: clusterrole/view is missing permission PolicyRule{Verbs:["list"], APIGroups:[""], Resources:["generatedeploymentconfigs"]}. Info: clusterrole/view is missing permission PolicyRule{Verbs:["watch"], APIGroups:[""], Resources:["generatedeploymentconfigs"]}. Info: clusterrole/view is missing permission PolicyRule{Verbs:["get"], APIGroups:[""], Resources:["deploymentconfigrollbacks"]}. Info: clusterrole/view is missing permission PolicyRule{Verbs:["list"], APIGroups:[""], Resources:["deploymentconfigrollbacks"]}. Info: clusterrole/view is missing permission PolicyRule{Verbs:["watch"], APIGroups:[""], Resources:["deploymentconfigrollbacks"]}. WARN: [CRD1003 from diagnostic ClusterRoles@openshift/origin/pkg/diagnostics/cluster/roles.go:82] clusterrole/system:registry has changed, but the existing role has more permissions than the new role. Use the `oadm policy reconcile-cluster-roles` command to update the role to reduce permissions. Info: clusterrole/system:registry has extra permission PolicyRule{Verbs:["delete"], APIGroups:[""], Resources:["imagestreamtags"]}. [Note] Running diagnostic: ClusterRouterName Description: Check there is a working router [Note] Running diagnostic: MasterNode Description: Check if master is also running node (for Open vSwitch) Info: Found a node with same IP as master: localhost.localdomain [Note] Skipping diagnostic: MetricsApiProxy Description: Check the integrated heapster metrics can be reached via the API proxy Because: The heapster service does not exist in the openshift-infra project at this time, so it is not available for the Horizontal Pod Autoscaler to use as a source of metrics. [Note] Running diagnostic: NodeDefinitions Description: Check node records on master [Note] Skipping diagnostic: ServiceExternalIPs Description: Check for existing services with ExternalIPs that are disallowed by master config Because: No master config file was detected [Note] Summary of diagnostics execution (version v1.3.0-alpha.2): [Note] Warnings seen: 3 [Note] Errors seen: 2 ``` `$ oc get all -o json -n gogs` ``` { "kind": "List", "apiVersion": "v1", "metadata": {}, "items": [ { "kind": "Route", "apiVersion": "v1", "metadata": { "name": "gogs", "namespace": "gogs", "selfLink": "/oapi/v1/namespaces/gogs/routes/gogs", "uid": "d751e18f-445f-11e6-877a-080027d0e0cb", "resourceVersion": "1859", "creationTimestamp": "2016-07-07T16:28:30Z", "labels": { "name": "gogs" }, "annotations": { "openshift.io/host.generated": "true" } }, "spec": { "host": "gogs-gogs.192.168.0.55.xip.io", "to": { "kind": "Service", "name": "gogs" }, "port": { "targetPort": "gogs" } }, "status": { "ingress": [ { "host": "gogs-gogs.192.168.0.55.xip.io", "routerName": "router", "conditions": [ { "type": "Admitted", "status": "True", "lastTransitionTime": "2016-07-07T16:28:30Z" } ] } ] } }, { "kind": "Service", "apiVersion": "v1", "metadata": { "name": "gogs", "namespace": "gogs", "selfLink": "/api/v1/namespaces/gogs/services/gogs", "uid": "c5bda77c-445f-11e6-877a-080027d0e0cb", "resourceVersion": "1820", "creationTimestamp": "2016-07-07T16:28:01Z", "labels": { "name": "gogs" } }, "spec": { "ports": [ { "name": "gogs", "protocol": "TCP", "port": 3000, "targetPort": 3000 }, { "name": "gogsssh", "protocol": "TCP", "port": 22, "targetPort": 22 } ], "selector": { "name": "gogs" }, "portalIP": "172.30.131.203", "clusterIP": "172.30.131.203", "type": "ClusterIP", "sessionAffinity": "None" }, "status": { "loadBalancer": {} } }, { "kind": "Service", "apiVersion": "v1", "metadata": { "name": "mysql", "namespace": "gogs", "selfLink": "/api/v1/namespaces/gogs/services/mysql", "uid": "c5c29ff8-445f-11e6-877a-080027d0e0cb", "resourceVersion": "1824", "creationTimestamp": "2016-07-07T16:28:01Z", "labels": { "name": "mysql" } }, "spec": { "ports": [ { "name": "mysql", "protocol": "TCP", "port": 3306, "targetPort": 3306 } ], "selector": { "name": "mysql" }, "portalIP": "172.30.39.49", "clusterIP": "172.30.39.49", "type": "ClusterIP", "sessionAffinity": "None" }, "status": { "loadBalancer": {} } }, { "kind": "Pod", "apiVersion": "v1", "metadata": { "name": "gogs", "namespace": "gogs", "selfLink": "/api/v1/namespaces/gogs/pods/gogs", "uid": "c5bf6934-445f-11e6-877a-080027d0e0cb", "resourceVersion": "1853", "creationTimestamp": "2016-07-07T16:28:01Z", "labels": { "name": "gogs" }, "annotations": { "openshift.io/scc": "anyuid" } }, "spec": { "volumes": [ { "name": "gogs-persistent-storage", "persistentVolumeClaim": { "claimName": "claim-gogs" } }, { "name": "default-token-qs7i9", "secret": { "secretName": "default-token-qs7i9" } } ], "containers": [ { "name": "gogs", "image": "gogs/gogs", "ports": [ { "name": "gogs", "containerPort": 3000, "protocol": "TCP" }, { "name": "gogsssh", "containerPort": 22, "protocol": "TCP" } ], "resources": {}, "volumeMounts": [ { "name": "gogs-persistent-storage", "mountPath": "/data" }, { "name": "default-token-qs7i9", "readOnly": true, "mountPath": "/var/run/secrets/kubernetes.io/serviceaccount" } ], "terminationMessagePath": "/dev/termination-log", "imagePullPolicy": "Always", "securityContext": { "capabilities": { "drop": [ "MKNOD", "SYS_CHROOT" ] }, "privileged": false, "seLinuxOptions": { "level": "s0:c10,c5" } } } ], "restartPolicy": "Always", "terminationGracePeriodSeconds": 30, "dnsPolicy": "ClusterFirst", "host": "localhost.localdomain", "serviceAccountName": "default", "serviceAccount": "default", "nodeName": "localhost.localdomain", "securityContext": { "seLinuxOptions": { "level": "s0:c10,c5" } }, "imagePullSecrets": [ { "name": "default-dockercfg-3br2j" } ] }, "status": { "phase": "Running", "conditions": [ { "type": "Ready", "status": "True", "lastProbeTime": null, "lastTransitionTime": "2016-07-07T16:28:21Z" } ], "hostIP": "192.168.0.55", "podIP": "172.17.0.2", "startTime": "2016-07-07T16:28:02Z", "containerStatuses": [ { "name": "gogs", "state": { "running": { "startedAt": "2016-07-07T16:28:19Z" } }, "lastState": {}, "ready": true, "restartCount": 0, "image": "gogs/gogs", "imageID": "docker://sha256:b555fb982f4f32a4041bd91e0366567d60f68c729380f8cd1151c942ac88c9df", "containerID": "docker://ba208236f26aed1952352bcee002dc275f5de34c2917dc47a1e6c22744b48346" } ] } }, { "kind": "Pod", "apiVersion": "v1", "metadata": { "name": "mysql", "namespace": "gogs", "selfLink": "/api/v1/namespaces/gogs/pods/mysql", "uid": "c5c0d164-445f-11e6-877a-080027d0e0cb", "resourceVersion": "1856", "creationTimestamp": "2016-07-07T16:28:01Z", "labels": { "name": "mysql" }, "annotations": { "openshift.io/scc": "anyuid" } }, "spec": { "volumes": [ { "name": "mysql-persistent-storage", "persistentVolumeClaim": { "claimName": "claim-gogs-mysql" } }, { "name": "default-token-qs7i9", "secret": { "secretName": "default-token-qs7i9" } } ], "containers": [ { "name": "mysql", "image": "openshift/mysql-55-centos7", "ports": [ { "name": "mysql", "containerPort": 3306, "protocol": "TCP" } ], "env": [ { "name": "MYSQL_ROOT_PASSWORD", "value": "toor" }, { "name": "MYSQL_USER", "value": "gogs_user" }, { "name": "MYSQL_PASSWORD", "value": "gogs_pass" }, { "name": "MYSQL_DATABASE", "value": "gogs_db" } ], "resources": {}, "volumeMounts": [ { "name": "mysql-persistent-storage", "mountPath": "/var/lib/mysql/data" }, { "name": "default-token-qs7i9", "readOnly": true, "mountPath": "/var/run/secrets/kubernetes.io/serviceaccount" } ], "terminationMessagePath": "/dev/termination-log", "imagePullPolicy": "Always", "securityContext": { "capabilities": { "drop": [ "MKNOD", "SYS_CHROOT" ] }, "privileged": false, "seLinuxOptions": { "level": "s0:c10,c5" } } } ], "restartPolicy": "Always", "terminationGracePeriodSeconds": 30, "dnsPolicy": "ClusterFirst", "host": "localhost.localdomain", "serviceAccountName": "default", "serviceAccount": "default", "nodeName": "localhost.localdomain", "securityContext": { "seLinuxOptions": { "level": "s0:c10,c5" } }, "imagePullSecrets": [ { "name": "default-dockercfg-3br2j" } ] }, "status": { "phase": "Running", "conditions": [ { "type": "Ready", "status": "True", "lastProbeTime": null, "lastTransitionTime": "2016-07-07T16:28:22Z" } ], "hostIP": "192.168.0.55", "podIP": "172.17.0.3", "startTime": "2016-07-07T16:28:02Z", "containerStatuses": [ { "name": "mysql", "state": { "running": { "startedAt": "2016-07-07T16:28:21Z" } }, "lastState": {}, "ready": true, "restartCount": 0, "image": "openshift/mysql-55-centos7", "imageID": "docker://sha256:6c380419ea3169be643a3afd1b681213a7a2f40343851e9f70d5014041506966", "containerID": "docker://2351284f38ecf1665256232bbcab49681808ad912f9a06127bba89f74358ad72" } ] } } ] } ```
1.0
chown Operation not permitted only when using persistent volume - I can able to deployed and use [gogs docker image](https://hub.docker.com/r/gogs/gogs/) in OpenShift Origin as ephemeral. But when using persistent volume I get `chown Operation not permitted` ##### Version oc v1.3.0-alpha.2 kubernetes v1.3.0-alpha.1-331-g0522e63 ##### Steps To Reproduce 1. Create two persistent volume matching the claim in below template 2. `oadm policy add-scc-to-user anyuid -z default` 3. `chcon -Rt svirt_sandbox_file_t volume_path` may be not required 4. Use the below template ``` apiVersion: v1 kind: Template metadata: name: gogs-persistent objects: - apiVersion: v1 kind: PersistentVolumeClaim metadata: name: claim-gogs spec: accessModes: - ReadWriteOnce resources: requests: storage: 1Gi - apiVersion: v1 kind: Pod metadata: name: gogs labels: name: gogs spec: containers: - image: gogs/gogs name: gogs ports: - containerPort: 3000 name: gogs - containerPort: 22 name: gogsssh volumeMounts: - name: gogs-persistent-storage mountPath: /data volumes: - name: gogs-persistent-storage persistentVolumeClaim: claimName: claim-gogs - apiVersion: v1 kind: Service metadata: labels: name: gogs name: gogs spec: ports: - port: 3000 name: gogs - port: 22 name: gogsssh selector: name: gogs - apiVersion: v1 kind: PersistentVolumeClaim metadata: name: claim-gogs-mysql spec: accessModes: - ReadWriteOnce resources: requests: storage: 1Gi - apiVersion: v1 kind: Pod metadata: name: mysql labels: name: mysql spec: containers: - image: openshift/mysql-55-centos7 name: mysql env: - name: MYSQL_ROOT_PASSWORD value: toor - name: MYSQL_USER value: gogs_user - name: MYSQL_PASSWORD value: gogs_pass - name: MYSQL_DATABASE value: gogs_db ports: - containerPort: 3306 name: mysql volumeMounts: - name: mysql-persistent-storage mountPath: /var/lib/mysql/data volumes: - name: mysql-persistent-storage persistentVolumeClaim: claimName: claim-gogs-mysql - apiVersion: v1 kind: Service metadata: labels: name: mysql name: mysql spec: ports: - port: 3306 name: mysql selector: name: mysql ``` ##### Current Result `$ oc logs pod gogs` ``` init:socat | Can't bind linked container GOGS to localhost, port 22 already in use init:socat | Linked container KUBERNETES will be binded to localhost on port 53 init:socat | Can't bind linked container GOGS to localhost, port 3000 already in use init:socat | Linked container MYSQL will be binded to localhost on port 3306 init:socat | Linked container KUBERNETES will be binded to localhost on port 443 Jul 7 16:28:20 syslogd started: BusyBox v1.24.2 chown: /data/gogs/data: Operation not permitted chown: /data/gogs/data: Operation not permitted chown: /data/gogs/conf: Operation not permitted chown: /data/gogs/conf: Operation not permitted chown: /data/gogs/log: Operation not permitted chown: /data/gogs/log: Operation not permitted chown: /data/gogs: Operation not permitted chown: /data/gogs: Operation not permitted chown: /data/git/.ssh/environment: Operation not permitted chown: /data/git/.ssh: Operation not permitted chown: /data/git/.ssh: Operation not permitted chown: /data/git: Operation not permitted chown: /data/git: Operation not permitted chown: /data/ssh/ssh_host_rsa_key: Operation not permitted chown: /data/ssh: Operation not permitted chown: /data/ssh: Operation not permitted chown: /data: Operation not permitted chown: /data: Operation not permitted chown: /data/ssh/ssh_host_dsa_key: Operation not permitted chown: /data/ssh/ssh_host_dsa_key.pub: Operation not permitted chown: /data/ssh/ssh_host_ecdsa_key: Operation not permitted chown: /data/ssh/ssh_host_ecdsa_key.pub: Operation not permitted chown: /data/ssh/ssh_host_ed25519_key: Operation not permitted chown: /data/ssh/ssh_host_ed25519_key.pub: Operation not permitted chown: /data/ssh/ssh_host_rsa_key: Operation not permitted chown: /data/ssh/ssh_host_rsa_key.pub: Operation not permitted Jul 7 16:28:22 sshd[49]: Server listening on :: port 22. Jul 7 16:28:22 sshd[49]: Server listening on 0.0.0.0 port 22. chown: /data/git/.ssh/environment: Operation not permitted chown: /data/git/.ssh: Operation not permitted chown: /data/git/.ssh: Operation not permitted chown: /data/git/: Operation not permitted chown: /data/git/: Operation not permitted chmod: /data: Operation not permitted 2016/07/07 16:28:23 [W] Custom config '/data/gogs/conf/app.ini' not found, ignore this if you're running first time 2016/07/07 16:28:23 [T] Custom path: /data/gogs 2016/07/07 16:28:23 [T] Log path: /app/gogs/log 2016/07/07 16:28:23 [I] Gogs: Go Git Service 0.9.35.0702 2016/07/07 16:28:23 [I] Build Time: 2016-07-02 04:16:49 UTC 2016/07/07 16:28:23 [I] Build Git Hash:  2016/07/07 16:28:23 [I] Log Mode: Console(Trace) 2016/07/07 16:28:23 [I] Cache Service Enabled 2016/07/07 16:28:23 [I] Session Service Enabled 2016/07/07 16:28:23 [I] SQLite3 Supported 2016/07/07 16:28:23 [I] Run Mode: Development 2016/07/07 16:28:26 [I] Listen: http://0.0.0.0:3000 ``` ##### Additional Information `$ oadm diagnostics` ``` [Note] Determining if client configuration exists for client/cluster diagnostics Info: Successfully read a client config file at '/root/.kube/config' Info: Using context for cluster-admin access: 'gogs/192-168-0-55:8443/system:admin' [Note] Running diagnostic: ConfigContexts[wordpress/192-168-0-55:8443/system:admin] Description: Validate client config context is complete and has connectivity Info: For client config context 'wordpress/192-168-0-55:8443/system:admin': The server URL is 'https://192.168.0.55:8443' The user authentication is 'system:admin/192-168-0-55:8443' The current project is 'wordpress' Successfully requested project list; has access to project(s): [default gogs myproject openshift openshift-infra] [Note] Running diagnostic: ConfigContexts[gogs/192-168-0-55:8443/developer] Description: Validate client config context is complete and has connectivity Info: For client config context 'gogs/192-168-0-55:8443/developer': The server URL is 'https://192.168.0.55:8443' The user authentication is 'developer/192-168-0-55:8443' The current project is 'gogs' Successfully requested project list; has access to project(s): [gogs myproject] [Note] Running diagnostic: DiagnosticPod Description: Create a pod to run diagnostics from the application standpoint Info: Output from the diagnostic pod (image openshift/origin-deployer:v1.3.0-alpha.2): [Note] Running diagnostic: PodCheckAuth Description: Check that service account credentials authenticate as expected Info: Service account token successfully authenticated to master Info: Service account token was authenticated by the integrated registry. [Note] Running diagnostic: PodCheckDns Description: Check that DNS within a pod works as expected [Note] Summary of diagnostics execution (version v1.3.0-alpha.2): [Note] Completed with no errors or warnings seen. [Note] Running diagnostic: ClusterRegistry Description: Check that there is a working Docker registry [Note] Running diagnostic: ClusterRoleBindings Description: Check that the default ClusterRoleBindings are present and contain the expected subjects [Note] Running diagnostic: ClusterRoles Description: Check that the default ClusterRoles are present and contain the expected permissions ERROR: [CRD1005 from diagnostic ClusterRoles@openshift/origin/pkg/diagnostics/cluster/roles.go:90] clusterrole/cluster-reader has changed and the existing role does not have enough permissions. Use the `oadm policy reconcile-cluster-roles` command to update the role. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["list"], APIGroups:[""], Resources:["minions"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["watch"], APIGroups:[""], Resources:["minions"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["get"], APIGroups:[""], Resources:["minions"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["get"], APIGroups:[""], Resources:["subjectaccessreviews"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["list"], APIGroups:[""], Resources:["subjectaccessreviews"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["watch"], APIGroups:[""], Resources:["subjectaccessreviews"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["watch"], APIGroups:[""], Resources:["builds/clone"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["get"], APIGroups:[""], Resources:["builds/clone"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["list"], APIGroups:[""], Resources:["builds/clone"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["get"], APIGroups:[""], Resources:["buildconfigs/instantiate"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["list"], APIGroups:[""], Resources:["buildconfigs/instantiate"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["watch"], APIGroups:[""], Resources:["buildconfigs/instantiate"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["get"], APIGroups:[""], Resources:["localsubjectaccessreviews"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["list"], APIGroups:[""], Resources:["localsubjectaccessreviews"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["watch"], APIGroups:[""], Resources:["localsubjectaccessreviews"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["get"], APIGroups:[""], Resources:["localresourceaccessreviews"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["list"], APIGroups:[""], Resources:["localresourceaccessreviews"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["watch"], APIGroups:[""], Resources:["localresourceaccessreviews"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["get"], APIGroups:[""], Resources:["resourceaccessreviews"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["list"], APIGroups:[""], Resources:["resourceaccessreviews"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["watch"], APIGroups:[""], Resources:["resourceaccessreviews"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["get"], APIGroups:[""], Resources:["generatedeploymentconfigs"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["list"], APIGroups:[""], Resources:["generatedeploymentconfigs"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["watch"], APIGroups:[""], Resources:["generatedeploymentconfigs"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["get"], APIGroups:[""], Resources:["imagestreamimports"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["list"], APIGroups:[""], Resources:["imagestreamimports"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["watch"], APIGroups:[""], Resources:["imagestreamimports"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["list"], APIGroups:[""], Resources:["imagestreammappings"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["watch"], APIGroups:[""], Resources:["imagestreammappings"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["get"], APIGroups:[""], Resources:["imagestreammappings"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["get"], APIGroups:[""], Resources:["oauthclients"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["list"], APIGroups:[""], Resources:["oauthclients"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["watch"], APIGroups:[""], Resources:["oauthclients"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["list"], APIGroups:[""], Resources:["deploymentconfigrollbacks"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["watch"], APIGroups:[""], Resources:["deploymentconfigrollbacks"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["get"], APIGroups:[""], Resources:["deploymentconfigrollbacks"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["get"], APIGroups:[""], Resources:["buildconfigs/instantiatebinary"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["list"], APIGroups:[""], Resources:["buildconfigs/instantiatebinary"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["watch"], APIGroups:[""], Resources:["buildconfigs/instantiatebinary"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["watch"], APIGroups:[""], Resources:["selfsubjectrulesreviews"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["get"], APIGroups:[""], Resources:["selfsubjectrulesreviews"]}. Info: clusterrole/cluster-reader is missing permission PolicyRule{Verbs:["list"], APIGroups:[""], Resources:["selfsubjectrulesreviews"]}. WARN: [CRD1003 from diagnostic ClusterRoles@openshift/origin/pkg/diagnostics/cluster/roles.go:82] clusterrole/admin has changed, but the existing role has more permissions than the new role. Use the `oadm policy reconcile-cluster-roles` command to update the role to reduce permissions. Info: clusterrole/admin has extra permission PolicyRule{Verbs:["patch"], APIGroups:[""], Resources:["deploymentconfigs/rollback"]}. Info: clusterrole/admin has extra permission PolicyRule{Verbs:["update"], APIGroups:[""], Resources:["deploymentconfigs/rollback"]}. Info: clusterrole/admin has extra permission PolicyRule{Verbs:["watch"], APIGroups:[""], Resources:["deploymentconfigs/rollback"]}. Info: clusterrole/admin has extra permission PolicyRule{Verbs:["create"], APIGroups:[""], Resources:["deploymentconfigs/rollback"]}. Info: clusterrole/admin has extra permission PolicyRule{Verbs:["delete"], APIGroups:[""], Resources:["deploymentconfigs/rollback"]}. Info: clusterrole/admin has extra permission PolicyRule{Verbs:["deletecollection"], APIGroups:[""], Resources:["deploymentconfigs/rollback"]}. Info: clusterrole/admin has extra permission PolicyRule{Verbs:["get"], APIGroups:[""], Resources:["deploymentconfigs/rollback"]}. Info: clusterrole/admin has extra permission PolicyRule{Verbs:["list"], APIGroups:[""], Resources:["deploymentconfigs/rollback"]}. WARN: [CRD1003 from diagnostic ClusterRoles@openshift/origin/pkg/diagnostics/cluster/roles.go:82] clusterrole/edit has changed, but the existing role has more permissions than the new role. Use the `oadm policy reconcile-cluster-roles` command to update the role to reduce permissions. Info: clusterrole/edit has extra permission PolicyRule{Verbs:["get"], APIGroups:[""], Resources:["deploymentconfigs/rollback"]}. Info: clusterrole/edit has extra permission PolicyRule{Verbs:["list"], APIGroups:[""], Resources:["deploymentconfigs/rollback"]}. Info: clusterrole/edit has extra permission PolicyRule{Verbs:["patch"], APIGroups:[""], Resources:["deploymentconfigs/rollback"]}. Info: clusterrole/edit has extra permission PolicyRule{Verbs:["update"], APIGroups:[""], Resources:["deploymentconfigs/rollback"]}. Info: clusterrole/edit has extra permission PolicyRule{Verbs:["watch"], APIGroups:[""], Resources:["deploymentconfigs/rollback"]}. Info: clusterrole/edit has extra permission PolicyRule{Verbs:["create"], APIGroups:[""], Resources:["deploymentconfigs/rollback"]}. Info: clusterrole/edit has extra permission PolicyRule{Verbs:["delete"], APIGroups:[""], Resources:["deploymentconfigs/rollback"]}. Info: clusterrole/edit has extra permission PolicyRule{Verbs:["deletecollection"], APIGroups:[""], Resources:["deploymentconfigs/rollback"]}. ERROR: [CRD1005 from diagnostic ClusterRoles@openshift/origin/pkg/diagnostics/cluster/roles.go:90] clusterrole/view has changed and the existing role does not have enough permissions. Use the `oadm policy reconcile-cluster-roles` command to update the role. Info: clusterrole/view is missing permission PolicyRule{Verbs:["get"], APIGroups:[""], Resources:["generatedeploymentconfigs"]}. Info: clusterrole/view is missing permission PolicyRule{Verbs:["list"], APIGroups:[""], Resources:["generatedeploymentconfigs"]}. Info: clusterrole/view is missing permission PolicyRule{Verbs:["watch"], APIGroups:[""], Resources:["generatedeploymentconfigs"]}. Info: clusterrole/view is missing permission PolicyRule{Verbs:["get"], APIGroups:[""], Resources:["deploymentconfigrollbacks"]}. Info: clusterrole/view is missing permission PolicyRule{Verbs:["list"], APIGroups:[""], Resources:["deploymentconfigrollbacks"]}. Info: clusterrole/view is missing permission PolicyRule{Verbs:["watch"], APIGroups:[""], Resources:["deploymentconfigrollbacks"]}. WARN: [CRD1003 from diagnostic ClusterRoles@openshift/origin/pkg/diagnostics/cluster/roles.go:82] clusterrole/system:registry has changed, but the existing role has more permissions than the new role. Use the `oadm policy reconcile-cluster-roles` command to update the role to reduce permissions. Info: clusterrole/system:registry has extra permission PolicyRule{Verbs:["delete"], APIGroups:[""], Resources:["imagestreamtags"]}. [Note] Running diagnostic: ClusterRouterName Description: Check there is a working router [Note] Running diagnostic: MasterNode Description: Check if master is also running node (for Open vSwitch) Info: Found a node with same IP as master: localhost.localdomain [Note] Skipping diagnostic: MetricsApiProxy Description: Check the integrated heapster metrics can be reached via the API proxy Because: The heapster service does not exist in the openshift-infra project at this time, so it is not available for the Horizontal Pod Autoscaler to use as a source of metrics. [Note] Running diagnostic: NodeDefinitions Description: Check node records on master [Note] Skipping diagnostic: ServiceExternalIPs Description: Check for existing services with ExternalIPs that are disallowed by master config Because: No master config file was detected [Note] Summary of diagnostics execution (version v1.3.0-alpha.2): [Note] Warnings seen: 3 [Note] Errors seen: 2 ``` `$ oc get all -o json -n gogs` ``` { "kind": "List", "apiVersion": "v1", "metadata": {}, "items": [ { "kind": "Route", "apiVersion": "v1", "metadata": { "name": "gogs", "namespace": "gogs", "selfLink": "/oapi/v1/namespaces/gogs/routes/gogs", "uid": "d751e18f-445f-11e6-877a-080027d0e0cb", "resourceVersion": "1859", "creationTimestamp": "2016-07-07T16:28:30Z", "labels": { "name": "gogs" }, "annotations": { "openshift.io/host.generated": "true" } }, "spec": { "host": "gogs-gogs.192.168.0.55.xip.io", "to": { "kind": "Service", "name": "gogs" }, "port": { "targetPort": "gogs" } }, "status": { "ingress": [ { "host": "gogs-gogs.192.168.0.55.xip.io", "routerName": "router", "conditions": [ { "type": "Admitted", "status": "True", "lastTransitionTime": "2016-07-07T16:28:30Z" } ] } ] } }, { "kind": "Service", "apiVersion": "v1", "metadata": { "name": "gogs", "namespace": "gogs", "selfLink": "/api/v1/namespaces/gogs/services/gogs", "uid": "c5bda77c-445f-11e6-877a-080027d0e0cb", "resourceVersion": "1820", "creationTimestamp": "2016-07-07T16:28:01Z", "labels": { "name": "gogs" } }, "spec": { "ports": [ { "name": "gogs", "protocol": "TCP", "port": 3000, "targetPort": 3000 }, { "name": "gogsssh", "protocol": "TCP", "port": 22, "targetPort": 22 } ], "selector": { "name": "gogs" }, "portalIP": "172.30.131.203", "clusterIP": "172.30.131.203", "type": "ClusterIP", "sessionAffinity": "None" }, "status": { "loadBalancer": {} } }, { "kind": "Service", "apiVersion": "v1", "metadata": { "name": "mysql", "namespace": "gogs", "selfLink": "/api/v1/namespaces/gogs/services/mysql", "uid": "c5c29ff8-445f-11e6-877a-080027d0e0cb", "resourceVersion": "1824", "creationTimestamp": "2016-07-07T16:28:01Z", "labels": { "name": "mysql" } }, "spec": { "ports": [ { "name": "mysql", "protocol": "TCP", "port": 3306, "targetPort": 3306 } ], "selector": { "name": "mysql" }, "portalIP": "172.30.39.49", "clusterIP": "172.30.39.49", "type": "ClusterIP", "sessionAffinity": "None" }, "status": { "loadBalancer": {} } }, { "kind": "Pod", "apiVersion": "v1", "metadata": { "name": "gogs", "namespace": "gogs", "selfLink": "/api/v1/namespaces/gogs/pods/gogs", "uid": "c5bf6934-445f-11e6-877a-080027d0e0cb", "resourceVersion": "1853", "creationTimestamp": "2016-07-07T16:28:01Z", "labels": { "name": "gogs" }, "annotations": { "openshift.io/scc": "anyuid" } }, "spec": { "volumes": [ { "name": "gogs-persistent-storage", "persistentVolumeClaim": { "claimName": "claim-gogs" } }, { "name": "default-token-qs7i9", "secret": { "secretName": "default-token-qs7i9" } } ], "containers": [ { "name": "gogs", "image": "gogs/gogs", "ports": [ { "name": "gogs", "containerPort": 3000, "protocol": "TCP" }, { "name": "gogsssh", "containerPort": 22, "protocol": "TCP" } ], "resources": {}, "volumeMounts": [ { "name": "gogs-persistent-storage", "mountPath": "/data" }, { "name": "default-token-qs7i9", "readOnly": true, "mountPath": "/var/run/secrets/kubernetes.io/serviceaccount" } ], "terminationMessagePath": "/dev/termination-log", "imagePullPolicy": "Always", "securityContext": { "capabilities": { "drop": [ "MKNOD", "SYS_CHROOT" ] }, "privileged": false, "seLinuxOptions": { "level": "s0:c10,c5" } } } ], "restartPolicy": "Always", "terminationGracePeriodSeconds": 30, "dnsPolicy": "ClusterFirst", "host": "localhost.localdomain", "serviceAccountName": "default", "serviceAccount": "default", "nodeName": "localhost.localdomain", "securityContext": { "seLinuxOptions": { "level": "s0:c10,c5" } }, "imagePullSecrets": [ { "name": "default-dockercfg-3br2j" } ] }, "status": { "phase": "Running", "conditions": [ { "type": "Ready", "status": "True", "lastProbeTime": null, "lastTransitionTime": "2016-07-07T16:28:21Z" } ], "hostIP": "192.168.0.55", "podIP": "172.17.0.2", "startTime": "2016-07-07T16:28:02Z", "containerStatuses": [ { "name": "gogs", "state": { "running": { "startedAt": "2016-07-07T16:28:19Z" } }, "lastState": {}, "ready": true, "restartCount": 0, "image": "gogs/gogs", "imageID": "docker://sha256:b555fb982f4f32a4041bd91e0366567d60f68c729380f8cd1151c942ac88c9df", "containerID": "docker://ba208236f26aed1952352bcee002dc275f5de34c2917dc47a1e6c22744b48346" } ] } }, { "kind": "Pod", "apiVersion": "v1", "metadata": { "name": "mysql", "namespace": "gogs", "selfLink": "/api/v1/namespaces/gogs/pods/mysql", "uid": "c5c0d164-445f-11e6-877a-080027d0e0cb", "resourceVersion": "1856", "creationTimestamp": "2016-07-07T16:28:01Z", "labels": { "name": "mysql" }, "annotations": { "openshift.io/scc": "anyuid" } }, "spec": { "volumes": [ { "name": "mysql-persistent-storage", "persistentVolumeClaim": { "claimName": "claim-gogs-mysql" } }, { "name": "default-token-qs7i9", "secret": { "secretName": "default-token-qs7i9" } } ], "containers": [ { "name": "mysql", "image": "openshift/mysql-55-centos7", "ports": [ { "name": "mysql", "containerPort": 3306, "protocol": "TCP" } ], "env": [ { "name": "MYSQL_ROOT_PASSWORD", "value": "toor" }, { "name": "MYSQL_USER", "value": "gogs_user" }, { "name": "MYSQL_PASSWORD", "value": "gogs_pass" }, { "name": "MYSQL_DATABASE", "value": "gogs_db" } ], "resources": {}, "volumeMounts": [ { "name": "mysql-persistent-storage", "mountPath": "/var/lib/mysql/data" }, { "name": "default-token-qs7i9", "readOnly": true, "mountPath": "/var/run/secrets/kubernetes.io/serviceaccount" } ], "terminationMessagePath": "/dev/termination-log", "imagePullPolicy": "Always", "securityContext": { "capabilities": { "drop": [ "MKNOD", "SYS_CHROOT" ] }, "privileged": false, "seLinuxOptions": { "level": "s0:c10,c5" } } } ], "restartPolicy": "Always", "terminationGracePeriodSeconds": 30, "dnsPolicy": "ClusterFirst", "host": "localhost.localdomain", "serviceAccountName": "default", "serviceAccount": "default", "nodeName": "localhost.localdomain", "securityContext": { "seLinuxOptions": { "level": "s0:c10,c5" } }, "imagePullSecrets": [ { "name": "default-dockercfg-3br2j" } ] }, "status": { "phase": "Running", "conditions": [ { "type": "Ready", "status": "True", "lastProbeTime": null, "lastTransitionTime": "2016-07-07T16:28:22Z" } ], "hostIP": "192.168.0.55", "podIP": "172.17.0.3", "startTime": "2016-07-07T16:28:02Z", "containerStatuses": [ { "name": "mysql", "state": { "running": { "startedAt": "2016-07-07T16:28:21Z" } }, "lastState": {}, "ready": true, "restartCount": 0, "image": "openshift/mysql-55-centos7", "imageID": "docker://sha256:6c380419ea3169be643a3afd1b681213a7a2f40343851e9f70d5014041506966", "containerID": "docker://2351284f38ecf1665256232bbcab49681808ad912f9a06127bba89f74358ad72" } ] } } ] } ```
non_comp
chown operation not permitted only when using persistent volume i can able to deployed and use in openshift origin as ephemeral but when using persistent volume i get chown operation not permitted version oc alpha kubernetes alpha steps to reproduce create two persistent volume matching the claim in below template oadm policy add scc to user anyuid z default chcon rt svirt sandbox file t volume path may be not required use the below template apiversion kind template metadata name gogs persistent objects apiversion kind persistentvolumeclaim metadata name claim gogs spec accessmodes readwriteonce resources requests storage apiversion kind pod metadata name gogs labels name gogs spec containers image gogs gogs name gogs ports containerport name gogs containerport name gogsssh volumemounts name gogs persistent storage mountpath data volumes name gogs persistent storage persistentvolumeclaim claimname claim gogs apiversion kind service metadata labels name gogs name gogs spec ports port name gogs port name gogsssh selector name gogs apiversion kind persistentvolumeclaim metadata name claim gogs mysql spec accessmodes readwriteonce resources requests storage apiversion kind pod metadata name mysql labels name mysql spec containers image openshift mysql name mysql env name mysql root password value toor name mysql user value gogs user name mysql password value gogs pass name mysql database value gogs db ports containerport name mysql volumemounts name mysql persistent storage mountpath var lib mysql data volumes name mysql persistent storage persistentvolumeclaim claimname claim gogs mysql apiversion kind service metadata labels name mysql name mysql spec ports port name mysql selector name mysql current result oc logs pod gogs init socat can t bind linked container gogs to localhost port already in use init socat linked container kubernetes will be binded to localhost on port init socat can t bind linked container gogs to localhost port already in use init socat linked container mysql will be binded to localhost on port init socat linked container kubernetes will be binded to localhost on port jul syslogd started busybox chown data gogs data operation not permitted chown data gogs data operation not permitted chown data gogs conf operation not permitted chown data gogs conf operation not permitted chown data gogs log operation not permitted chown data gogs log operation not permitted chown data gogs operation not permitted chown data gogs operation not permitted chown data git ssh environment operation not permitted chown data git ssh operation not permitted chown data git ssh operation not permitted chown data git operation not permitted chown data git operation not permitted chown data ssh ssh host rsa key operation not permitted chown data ssh operation not permitted chown data ssh operation not permitted chown data operation not permitted chown data operation not permitted chown data ssh ssh host dsa key operation not permitted chown data ssh ssh host dsa key pub operation not permitted chown data ssh ssh host ecdsa key operation not permitted chown data ssh ssh host ecdsa key pub operation not permitted chown data ssh ssh host key operation not permitted chown data ssh ssh host key pub operation not permitted chown data ssh ssh host rsa key operation not permitted chown data ssh ssh host rsa key pub operation not permitted jul sshd server listening on port jul sshd server listening on port chown data git ssh environment operation not permitted chown data git ssh operation not permitted chown data git ssh operation not permitted chown data git operation not permitted chown data git operation not permitted chmod data operation not permitted  custom config data gogs conf app ini not found ignore this if you re running first time  custom path data gogs  log path app gogs log  gogs go git service   build time utc  build git hash   log mode console trace   cache service enabled  session service enabled  supported  run mode development  listen additional information oadm diagnostics determining if client configuration exists for client cluster diagnostics info successfully read a client config file at root kube config info using context for cluster admin access gogs system admin running diagnostic configcontexts description validate client config context is complete and has connectivity info for client config context wordpress system admin the server url is the user authentication is system admin the current project is wordpress successfully requested project list has access to project s running diagnostic configcontexts description validate client config context is complete and has connectivity info for client config context gogs developer the server url is the user authentication is developer the current project is gogs successfully requested project list has access to project s running diagnostic diagnosticpod description create a pod to run diagnostics from the application standpoint info output from the diagnostic pod image openshift origin deployer alpha running diagnostic podcheckauth description check that service account credentials authenticate as expected info service account token successfully authenticated to master info service account token was authenticated by the integrated registry running diagnostic podcheckdns description check that dns within a pod works as expected summary of diagnostics execution version alpha completed with no errors or warnings seen running diagnostic clusterregistry description check that there is a working docker registry running diagnostic clusterrolebindings description check that the default clusterrolebindings are present and contain the expected subjects running diagnostic clusterroles description check that the default clusterroles are present and contain the expected permissions error clusterrole cluster reader has changed and the existing role does not have enough permissions use the oadm policy reconcile cluster roles command to update the role info clusterrole cluster reader is missing permission policyrule verbs apigroups resources info clusterrole cluster reader is missing permission policyrule verbs apigroups resources info clusterrole cluster reader is missing permission policyrule verbs apigroups resources info clusterrole cluster reader is missing permission policyrule verbs apigroups resources info clusterrole cluster reader is missing permission policyrule verbs apigroups resources info clusterrole cluster reader is missing permission policyrule verbs apigroups resources info clusterrole cluster reader is missing permission policyrule verbs apigroups resources info clusterrole cluster reader is missing permission policyrule verbs apigroups resources info clusterrole cluster reader is missing permission policyrule verbs apigroups resources info clusterrole cluster reader is missing permission policyrule verbs apigroups resources info clusterrole cluster reader is missing permission policyrule verbs apigroups resources info clusterrole cluster reader is missing permission policyrule verbs apigroups resources info clusterrole cluster reader is missing permission policyrule verbs apigroups resources info clusterrole cluster reader is missing permission policyrule verbs apigroups resources info clusterrole cluster reader is missing permission policyrule verbs apigroups resources info clusterrole cluster reader is missing permission policyrule verbs apigroups resources info clusterrole cluster reader is missing permission policyrule verbs apigroups resources info clusterrole cluster reader is missing permission policyrule verbs apigroups resources info clusterrole cluster reader is missing permission policyrule verbs apigroups resources info clusterrole cluster reader is missing permission policyrule verbs apigroups resources info clusterrole cluster reader is missing permission policyrule verbs apigroups resources info clusterrole cluster reader is missing permission policyrule verbs apigroups resources info clusterrole cluster reader is missing permission policyrule verbs apigroups resources info clusterrole cluster reader is missing permission policyrule verbs apigroups resources info clusterrole cluster reader is missing permission policyrule verbs apigroups resources info clusterrole cluster reader is missing permission policyrule verbs apigroups resources info clusterrole cluster reader is missing permission policyrule verbs apigroups resources info clusterrole cluster reader is missing permission policyrule verbs apigroups resources info clusterrole cluster reader is missing permission policyrule verbs apigroups resources info clusterrole cluster reader is missing permission policyrule verbs apigroups resources info clusterrole cluster reader is missing permission policyrule verbs apigroups resources info clusterrole cluster reader is missing permission policyrule verbs apigroups resources info clusterrole cluster reader is missing permission policyrule verbs apigroups resources info clusterrole cluster reader is missing permission policyrule verbs apigroups resources info clusterrole cluster reader is missing permission policyrule verbs apigroups resources info clusterrole cluster reader is missing permission policyrule verbs apigroups resources info clusterrole cluster reader is missing permission policyrule verbs apigroups resources info clusterrole cluster reader is missing permission policyrule verbs apigroups resources info clusterrole cluster reader is missing permission policyrule verbs apigroups resources info clusterrole cluster reader is missing permission policyrule verbs apigroups resources info clusterrole cluster reader is missing permission policyrule verbs apigroups resources info clusterrole cluster reader is missing permission policyrule verbs apigroups resources warn clusterrole admin has changed but the existing role has more permissions than the new role use the oadm policy reconcile cluster roles command to update the role to reduce permissions info clusterrole admin has extra permission policyrule verbs apigroups resources info clusterrole admin has extra permission policyrule verbs apigroups resources info clusterrole admin has extra permission policyrule verbs apigroups resources info clusterrole admin has extra permission policyrule verbs apigroups resources info clusterrole admin has extra permission policyrule verbs apigroups resources info clusterrole admin has extra permission policyrule verbs apigroups resources info clusterrole admin has extra permission policyrule verbs apigroups resources info clusterrole admin has extra permission policyrule verbs apigroups resources warn clusterrole edit has changed but the existing role has more permissions than the new role use the oadm policy reconcile cluster roles command to update the role to reduce permissions info clusterrole edit has extra permission policyrule verbs apigroups resources info clusterrole edit has extra permission policyrule verbs apigroups resources info clusterrole edit has extra permission policyrule verbs apigroups resources info clusterrole edit has extra permission policyrule verbs apigroups resources info clusterrole edit has extra permission policyrule verbs apigroups resources info clusterrole edit has extra permission policyrule verbs apigroups resources info clusterrole edit has extra permission policyrule verbs apigroups resources info clusterrole edit has extra permission policyrule verbs apigroups resources error clusterrole view has changed and the existing role does not have enough permissions use the oadm policy reconcile cluster roles command to update the role info clusterrole view is missing permission policyrule verbs apigroups resources info clusterrole view is missing permission policyrule verbs apigroups resources info clusterrole view is missing permission policyrule verbs apigroups resources info clusterrole view is missing permission policyrule verbs apigroups resources info clusterrole view is missing permission policyrule verbs apigroups resources info clusterrole view is missing permission policyrule verbs apigroups resources warn clusterrole system registry has changed but the existing role has more permissions than the new role use the oadm policy reconcile cluster roles command to update the role to reduce permissions info clusterrole system registry has extra permission policyrule verbs apigroups resources running diagnostic clusterroutername description check there is a working router running diagnostic masternode description check if master is also running node for open vswitch info found a node with same ip as master localhost localdomain skipping diagnostic metricsapiproxy description check the integrated heapster metrics can be reached via the api proxy because the heapster service does not exist in the openshift infra project at this time so it is not available for the horizontal pod autoscaler to use as a source of metrics running diagnostic nodedefinitions description check node records on master skipping diagnostic serviceexternalips description check for existing services with externalips that are disallowed by master config because no master config file was detected summary of diagnostics execution version alpha warnings seen errors seen oc get all o json n gogs kind list apiversion metadata items kind route apiversion metadata name gogs namespace gogs selflink oapi namespaces gogs routes gogs uid resourceversion creationtimestamp labels name gogs annotations openshift io host generated true spec host gogs gogs xip io to kind service name gogs port targetport gogs status ingress host gogs gogs xip io routername router conditions type admitted status true lasttransitiontime kind service apiversion metadata name gogs namespace gogs selflink api namespaces gogs services gogs uid resourceversion creationtimestamp labels name gogs spec ports name gogs protocol tcp port targetport name gogsssh protocol tcp port targetport selector name gogs portalip clusterip type clusterip sessionaffinity none status loadbalancer kind service apiversion metadata name mysql namespace gogs selflink api namespaces gogs services mysql uid resourceversion creationtimestamp labels name mysql spec ports name mysql protocol tcp port targetport selector name mysql portalip clusterip type clusterip sessionaffinity none status loadbalancer kind pod apiversion metadata name gogs namespace gogs selflink api namespaces gogs pods gogs uid resourceversion creationtimestamp labels name gogs annotations openshift io scc anyuid spec volumes name gogs persistent storage persistentvolumeclaim claimname claim gogs name default token secret secretname default token containers name gogs image gogs gogs ports name gogs containerport protocol tcp name gogsssh containerport protocol tcp resources volumemounts name gogs persistent storage mountpath data name default token readonly true mountpath var run secrets kubernetes io serviceaccount terminationmessagepath dev termination log imagepullpolicy always securitycontext capabilities drop mknod sys chroot privileged false selinuxoptions level restartpolicy always terminationgraceperiodseconds dnspolicy clusterfirst host localhost localdomain serviceaccountname default serviceaccount default nodename localhost localdomain securitycontext selinuxoptions level imagepullsecrets name default dockercfg status phase running conditions type ready status true lastprobetime null lasttransitiontime hostip podip starttime containerstatuses name gogs state running startedat laststate ready true restartcount image gogs gogs imageid docker containerid docker kind pod apiversion metadata name mysql namespace gogs selflink api namespaces gogs pods mysql uid resourceversion creationtimestamp labels name mysql annotations openshift io scc anyuid spec volumes name mysql persistent storage persistentvolumeclaim claimname claim gogs mysql name default token secret secretname default token containers name mysql image openshift mysql ports name mysql containerport protocol tcp env name mysql root password value toor name mysql user value gogs user name mysql password value gogs pass name mysql database value gogs db resources volumemounts name mysql persistent storage mountpath var lib mysql data name default token readonly true mountpath var run secrets kubernetes io serviceaccount terminationmessagepath dev termination log imagepullpolicy always securitycontext capabilities drop mknod sys chroot privileged false selinuxoptions level restartpolicy always terminationgraceperiodseconds dnspolicy clusterfirst host localhost localdomain serviceaccountname default serviceaccount default nodename localhost localdomain securitycontext selinuxoptions level imagepullsecrets name default dockercfg status phase running conditions type ready status true lastprobetime null lasttransitiontime hostip podip starttime containerstatuses name mysql state running startedat laststate ready true restartcount image openshift mysql imageid docker containerid docker
0
13,557
16,066,566,143
IssuesEvent
2021-04-23 20:08:50
Vazkii/Botania
https://api.github.com/repos/Vazkii/Botania
closed
Runic Altar makes the world see-through
compatibility
# Version Information Forge version: 36.1.4 Botania version: 1.16.4-414 Modpack: Direwolf20 1.9.0 (the version that released a few days ago), with "Immersive Portals" and "Create Additions) added to it. # Further Information Link to crash log: Not crashing Steps to reproduce: 1. Place Runic Altar 2. Add ingredients 3. Look at Altar What I expected to happen: Shows the animation of the ingredients being charged and a UI indicating when it is done. What happened instead: As expected, plus the entire world goes transparent and I can see underground caves. Before adding last ingredient: ![Before adding last ingredient](https://user-images.githubusercontent.com/4059786/115882217-8efd4500-a41a-11eb-8500-a26f3117565c.png) After adding last ingredient: ![After adding last ingredient](https://user-images.githubusercontent.com/4059786/115882259-96bce980-a41a-11eb-9567-8dc5572deab7.png) At first I thought this might be because I was doing this on a platform above a lake, but I replicated it in a desert in creative as well: ![Desert not looking](https://user-images.githubusercontent.com/4059786/115882738-0e8b1400-a41b-11eb-8d6b-7a3d0b5b8012.png) ![Desert After](https://user-images.githubusercontent.com/4059786/115882733-0df27d80-a41b-11eb-93d6-fe2db344bcf6.png)
True
Runic Altar makes the world see-through - # Version Information Forge version: 36.1.4 Botania version: 1.16.4-414 Modpack: Direwolf20 1.9.0 (the version that released a few days ago), with "Immersive Portals" and "Create Additions) added to it. # Further Information Link to crash log: Not crashing Steps to reproduce: 1. Place Runic Altar 2. Add ingredients 3. Look at Altar What I expected to happen: Shows the animation of the ingredients being charged and a UI indicating when it is done. What happened instead: As expected, plus the entire world goes transparent and I can see underground caves. Before adding last ingredient: ![Before adding last ingredient](https://user-images.githubusercontent.com/4059786/115882217-8efd4500-a41a-11eb-8500-a26f3117565c.png) After adding last ingredient: ![After adding last ingredient](https://user-images.githubusercontent.com/4059786/115882259-96bce980-a41a-11eb-9567-8dc5572deab7.png) At first I thought this might be because I was doing this on a platform above a lake, but I replicated it in a desert in creative as well: ![Desert not looking](https://user-images.githubusercontent.com/4059786/115882738-0e8b1400-a41b-11eb-8d6b-7a3d0b5b8012.png) ![Desert After](https://user-images.githubusercontent.com/4059786/115882733-0df27d80-a41b-11eb-93d6-fe2db344bcf6.png)
comp
runic altar makes the world see through version information forge version botania version modpack the version that released a few days ago with immersive portals and create additions added to it further information link to crash log not crashing steps to reproduce place runic altar add ingredients look at altar what i expected to happen shows the animation of the ingredients being charged and a ui indicating when it is done what happened instead as expected plus the entire world goes transparent and i can see underground caves before adding last ingredient after adding last ingredient at first i thought this might be because i was doing this on a platform above a lake but i replicated it in a desert in creative as well
1
16,800
5,290,798,002
IssuesEvent
2017-02-08 20:49:58
dotnet/coreclr
https://api.github.com/repos/dotnet/coreclr
opened
Optimize default(T) == null at compile time
area-CodeGen optimization
`default(T) == null` in generic code is a popular* pattern to determine if `T` is a nullable or a reference type. Example: ```csharp using System; using System.Runtime.CompilerServices; class Program { [MethodImpl(MethodImplOptions.NoInlining)] static bool IsNotNullableValueType<T>() { return default(T) != null; } private static void Main(string[] args) { Console.WriteLine(IsNotNullableValueType<string>()); Console.WriteLine(IsNotNullableValueType<int>()); Console.WriteLine(IsNotNullableValueType<int?>()); } } ``` Will print `False True False`. While the codegen for the "it's a reference type" case is pretty efficient: ```nasm xor eax,eax test rax,rax setb al movzx eax,al ret ``` For the "it's a valuetype" and "it's a nullable" case we get a useless allocation: ```nasm sub rsp,28h call __NewHelper_System_Private_CoreLib_System_Int32 xor edx,edx mov dword ptr [rax+8],edx test rax,rax seta al movzx eax,al add rsp,28h ret ``` The places that use this pattern would likely benefit from being able to statically evaluate this because it could lead to elimination of entire branches and smaller code (aside from removing a useless allocation). \* There are 17 hits for the `default\(.*\) == null` regular expression in the CoreFX repo and a couple more for the "not equal to null" case.
1.0
Optimize default(T) == null at compile time - `default(T) == null` in generic code is a popular* pattern to determine if `T` is a nullable or a reference type. Example: ```csharp using System; using System.Runtime.CompilerServices; class Program { [MethodImpl(MethodImplOptions.NoInlining)] static bool IsNotNullableValueType<T>() { return default(T) != null; } private static void Main(string[] args) { Console.WriteLine(IsNotNullableValueType<string>()); Console.WriteLine(IsNotNullableValueType<int>()); Console.WriteLine(IsNotNullableValueType<int?>()); } } ``` Will print `False True False`. While the codegen for the "it's a reference type" case is pretty efficient: ```nasm xor eax,eax test rax,rax setb al movzx eax,al ret ``` For the "it's a valuetype" and "it's a nullable" case we get a useless allocation: ```nasm sub rsp,28h call __NewHelper_System_Private_CoreLib_System_Int32 xor edx,edx mov dword ptr [rax+8],edx test rax,rax seta al movzx eax,al add rsp,28h ret ``` The places that use this pattern would likely benefit from being able to statically evaluate this because it could lead to elimination of entire branches and smaller code (aside from removing a useless allocation). \* There are 17 hits for the `default\(.*\) == null` regular expression in the CoreFX repo and a couple more for the "not equal to null" case.
non_comp
optimize default t null at compile time default t null in generic code is a popular pattern to determine if t is a nullable or a reference type example csharp using system using system runtime compilerservices class program static bool isnotnullablevaluetype return default t null private static void main string args console writeline isnotnullablevaluetype console writeline isnotnullablevaluetype console writeline isnotnullablevaluetype will print false true false while the codegen for the it s a reference type case is pretty efficient nasm xor eax eax test rax rax setb al movzx eax al ret for the it s a valuetype and it s a nullable case we get a useless allocation nasm sub rsp call newhelper system private corelib system xor edx edx mov dword ptr edx test rax rax seta al movzx eax al add rsp ret the places that use this pattern would likely benefit from being able to statically evaluate this because it could lead to elimination of entire branches and smaller code aside from removing a useless allocation there are hits for the default null regular expression in the corefx repo and a couple more for the not equal to null case
0
15,123
18,991,988,991
IssuesEvent
2021-11-22 08:32:15
KiwiHawk/SeaBlock
https://api.github.com/repos/KiwiHawk/SeaBlock
opened
Error with Alien Biomes
bug mod compatibility
35.938 Warning Map.cpp:346: Map loading failed: Undefined named noise expression 'control-setting:cold:size:multiplier' Ignore Alien Biomes rather than add incompatibility.
True
Error with Alien Biomes - 35.938 Warning Map.cpp:346: Map loading failed: Undefined named noise expression 'control-setting:cold:size:multiplier' Ignore Alien Biomes rather than add incompatibility.
comp
error with alien biomes warning map cpp map loading failed undefined named noise expression control setting cold size multiplier ignore alien biomes rather than add incompatibility
1
4,761
7,370,877,539
IssuesEvent
2018-03-13 09:56:07
mike42/escpos-php
https://api.github.com/repos/mike42/escpos-php
closed
Printer tested: Bematech-4200-TH
printer-compatibility
Testei hoje e funcionou esse driver na impressora Bematech-4200-TH utilizando somente as impressões modo texto (sem imagens ou outras funcionalidades).
True
Printer tested: Bematech-4200-TH - Testei hoje e funcionou esse driver na impressora Bematech-4200-TH utilizando somente as impressões modo texto (sem imagens ou outras funcionalidades).
comp
printer tested bematech th testei hoje e funcionou esse driver na impressora bematech th utilizando somente as impressões modo texto sem imagens ou outras funcionalidades
1
27,799
12,706,957,284
IssuesEvent
2020-06-23 08:07:10
LiskHQ/lisk-docs
https://api.github.com/repos/LiskHQ/lisk-docs
closed
Create Introduction page
feature service
The Introduction page of the Lisk Service documentation should include: - A general introduction and explanation about the purpose of Lisk Service - An architecture overview - A section with basic usage commands
1.0
Create Introduction page - The Introduction page of the Lisk Service documentation should include: - A general introduction and explanation about the purpose of Lisk Service - An architecture overview - A section with basic usage commands
non_comp
create introduction page the introduction page of the lisk service documentation should include a general introduction and explanation about the purpose of lisk service an architecture overview a section with basic usage commands
0
213,044
16,507,913,864
IssuesEvent
2021-05-25 21:57:24
usbr/MTOM
https://api.github.com/repos/usbr/MTOM
closed
Cheat Sheet - MTOM
documentation
The 1-page “cheat-sheet” for running MTOM that LC put together needs to be reviewed for clarification Assigned to Rich Eastland
1.0
Cheat Sheet - MTOM - The 1-page “cheat-sheet” for running MTOM that LC put together needs to be reviewed for clarification Assigned to Rich Eastland
non_comp
cheat sheet mtom the page “cheat sheet” for running mtom that lc put together needs to be reviewed for clarification assigned to rich eastland
0
1,425
3,955,191,482
IssuesEvent
2016-04-29 19:55:47
facebook/hhvm
https://api.github.com/repos/facebook/hhvm
closed
System wide environment variables.
no isolated repro php5 incompatibility
I'm getting similar behavior to #1650 . Except it is system environment variables set in `/etc/environment` works on command line but fastcgi just not working. I've set the environment variable in fastcgi params and it works, really is just system environment variables not being read. ``` HipHop VM 3.2.0 (rel) Compiler: tags/HHVM-3.2.0-0-g01228273b8cf709aacbd3df1c51b1e690ecebac8 Repo schema: c52ba40f4a246d35a88f1dfc1daf959851ced8aa ```
True
System wide environment variables. - I'm getting similar behavior to #1650 . Except it is system environment variables set in `/etc/environment` works on command line but fastcgi just not working. I've set the environment variable in fastcgi params and it works, really is just system environment variables not being read. ``` HipHop VM 3.2.0 (rel) Compiler: tags/HHVM-3.2.0-0-g01228273b8cf709aacbd3df1c51b1e690ecebac8 Repo schema: c52ba40f4a246d35a88f1dfc1daf959851ced8aa ```
comp
system wide environment variables i m getting similar behavior to except it is system environment variables set in etc environment works on command line but fastcgi just not working i ve set the environment variable in fastcgi params and it works really is just system environment variables not being read hiphop vm rel compiler tags hhvm repo schema
1
87,820
25,224,816,070
IssuesEvent
2022-11-14 15:16:09
apache/pulsar
https://api.github.com/repos/apache/pulsar
closed
Docker is deprecated in Kubernetes -> identify need for action
type/feature component/build triage/week-49 lifecycle/stale
**Is your feature request related to a problem? Please describe.** Docker is **now deprecated** in Kubernetes. By the **end of 2021** "dockershim" which implements CRI support for Docker will be **removed.** > Reason: Why is dockershim being deprecated? > Maintaining dockershim has become a heavy burden on the Kubernetes maintainers. The CRI standard was created to reduce this burden and allow smooth interoperability of different container runtimes. Docker itself doesn't currently implement CRI, thus the problem. see: https://dev.to/inductor/wait-docker-is-deprecated-in-kubernetes-now-what-do-i-do-e4m https://kubernetes.io/blog/2020/12/02/dockershim-faq/ **Describe the solution you'd like** Directly support other container technologies for keeping setup of pulsar as easy as possible. One alternative to look at is probably "containerd" https://github.com/containerd/containerd which has already a good adoption https://github.com/containerd/containerd/blob/master/ADOPTERS.md
1.0
Docker is deprecated in Kubernetes -> identify need for action - **Is your feature request related to a problem? Please describe.** Docker is **now deprecated** in Kubernetes. By the **end of 2021** "dockershim" which implements CRI support for Docker will be **removed.** > Reason: Why is dockershim being deprecated? > Maintaining dockershim has become a heavy burden on the Kubernetes maintainers. The CRI standard was created to reduce this burden and allow smooth interoperability of different container runtimes. Docker itself doesn't currently implement CRI, thus the problem. see: https://dev.to/inductor/wait-docker-is-deprecated-in-kubernetes-now-what-do-i-do-e4m https://kubernetes.io/blog/2020/12/02/dockershim-faq/ **Describe the solution you'd like** Directly support other container technologies for keeping setup of pulsar as easy as possible. One alternative to look at is probably "containerd" https://github.com/containerd/containerd which has already a good adoption https://github.com/containerd/containerd/blob/master/ADOPTERS.md
non_comp
docker is deprecated in kubernetes identify need for action is your feature request related to a problem please describe docker is now deprecated in kubernetes by the end of dockershim which implements cri support for docker will be removed reason why is dockershim being deprecated maintaining dockershim has become a heavy burden on the kubernetes maintainers the cri standard was created to reduce this burden and allow smooth interoperability of different container runtimes docker itself doesn t currently implement cri thus the problem see describe the solution you d like directly support other container technologies for keeping setup of pulsar as easy as possible one alternative to look at is probably containerd which has already a good adoption
0
607,176
18,774,020,966
IssuesEvent
2021-11-07 10:55:00
AY2122S1-CS2103T-W15-3/tp
https://api.github.com/repos/AY2122S1-CS2103T-W15-3/tp
closed
Mark and unmark with duplicate index bug
type.Bug priority.High severity.High
When the user attempts to enter the command `emark 2 2 1`, the app will not mark the events correctly. We should not allow duplicate indexes.
1.0
Mark and unmark with duplicate index bug - When the user attempts to enter the command `emark 2 2 1`, the app will not mark the events correctly. We should not allow duplicate indexes.
non_comp
mark and unmark with duplicate index bug when the user attempts to enter the command emark the app will not mark the events correctly we should not allow duplicate indexes
0
153,905
12,168,269,505
IssuesEvent
2020-04-27 12:24:18
kyma-project/kyma
https://api.github.com/repos/kyma-project/kyma
opened
Switch fake function used in external solution test to real function
area/serverless test-missing
<!-- Thank you for your contribution. Before you submit the issue: 1. Search open and closed issues for duplicates. 2. Read the contributing guidelines. --> **Description** <!-- Provide a clear and concise description of the feature. --> TBD **Reasons** <!-- Explain why we should add this feature. Provide use cases to illustrate its benefits. --> **Attachments** <!-- Attach any files, links, code samples, or screenshots that will convince us to your idea. -->
1.0
Switch fake function used in external solution test to real function - <!-- Thank you for your contribution. Before you submit the issue: 1. Search open and closed issues for duplicates. 2. Read the contributing guidelines. --> **Description** <!-- Provide a clear and concise description of the feature. --> TBD **Reasons** <!-- Explain why we should add this feature. Provide use cases to illustrate its benefits. --> **Attachments** <!-- Attach any files, links, code samples, or screenshots that will convince us to your idea. -->
non_comp
switch fake function used in external solution test to real function thank you for your contribution before you submit the issue search open and closed issues for duplicates read the contributing guidelines description tbd reasons attachments
0
162,254
13,884,967,308
IssuesEvent
2020-10-18 18:09:49
phpmd/phpmd
https://api.github.com/repos/phpmd/phpmd
closed
unique identifier to distinguish from pmd
Documentation Enhancement Good first issue On Hold
The output xml files of PMD and PHPMD tools are almost identical. Without knowing which program generated them, it is programatically tricky to identify which tool the file derived from. Would you consider changing the top level `<pmd></pmd>` tag to `<phpmd></phpmd>`? A unique attribute like `tool="phpmd"` would work just as well. Currently, the best identifier I see is `version="@project.version@"` which is not used by PMD. Why do you use this format? Is it a permanent change?
1.0
unique identifier to distinguish from pmd - The output xml files of PMD and PHPMD tools are almost identical. Without knowing which program generated them, it is programatically tricky to identify which tool the file derived from. Would you consider changing the top level `<pmd></pmd>` tag to `<phpmd></phpmd>`? A unique attribute like `tool="phpmd"` would work just as well. Currently, the best identifier I see is `version="@project.version@"` which is not used by PMD. Why do you use this format? Is it a permanent change?
non_comp
unique identifier to distinguish from pmd the output xml files of pmd and phpmd tools are almost identical without knowing which program generated them it is programatically tricky to identify which tool the file derived from would you consider changing the top level tag to a unique attribute like tool phpmd would work just as well currently the best identifier i see is version project version which is not used by pmd why do you use this format is it a permanent change
0
33,185
27,288,807,143
IssuesEvent
2023-02-23 15:15:41
centerofci/mathesar
https://api.github.com/repos/centerofci/mathesar
closed
Demo server 500 right after login: "mathesar_demo_template" is being accessed by other users
type: bug work: backend work: infrastructure status: started
## Description Response was `500 https://demo.mathesar.org/` with following body: ![2023-02-13_18-14](https://user-images.githubusercontent.com/2715476/218511825-d0aa4e0e-1f0d-4a78-aafa-3080d33258a0.png) Reloaded after ~5mn (waited a long time because was submitting this bug report): view rendered fine, but I had no schemas (while I expected to see all the demo schemas). Another logout and login fixes it. <!-- A clear and concise description of what the bug is. --> ## To Reproduce Logged out of the demo server and logged back in.
1.0
Demo server 500 right after login: "mathesar_demo_template" is being accessed by other users - ## Description Response was `500 https://demo.mathesar.org/` with following body: ![2023-02-13_18-14](https://user-images.githubusercontent.com/2715476/218511825-d0aa4e0e-1f0d-4a78-aafa-3080d33258a0.png) Reloaded after ~5mn (waited a long time because was submitting this bug report): view rendered fine, but I had no schemas (while I expected to see all the demo schemas). Another logout and login fixes it. <!-- A clear and concise description of what the bug is. --> ## To Reproduce Logged out of the demo server and logged back in.
non_comp
demo server right after login mathesar demo template is being accessed by other users description response was with following body reloaded after waited a long time because was submitting this bug report view rendered fine but i had no schemas while i expected to see all the demo schemas another logout and login fixes it to reproduce logged out of the demo server and logged back in
0
9,998
11,985,493,582
IssuesEvent
2020-04-07 17:36:22
Juuxel/Adorn
https://api.github.com/repos/Juuxel/Adorn
closed
Deal with DraylARR's and Vatuu's mistakes
bug mod compatibility
`lil-tater` was first changed to `liltater` and then `lil_tater`
True
Deal with DraylARR's and Vatuu's mistakes - `lil-tater` was first changed to `liltater` and then `lil_tater`
comp
deal with draylarr s and vatuu s mistakes lil tater was first changed to liltater and then lil tater
1
18,996
26,419,195,759
IssuesEvent
2023-01-13 18:39:11
oracle/truffleruby
https://api.github.com/repos/oracle/truffleruby
closed
String#casecmp? fails between ascii and utf-8
compatibility
Comparing an ASCII-8BIT string with UTF-8, for both of which `#ascii_only?` is true, using `String#casecmp?` returns nil rather than a boolean e.g. ```ruby ''.b.casecmp?('') ``` It returns boolean on MRI, and the implementation to return boolean is present in truffleruby but not reached: https://github.com/oracle/truffleruby/blob/vm-22.3.0/src/main/ruby/truffleruby/core/string.rb#L1204-L1213 The check `if ascii_only? && other.ascii_only?` is not reached after `Primitive.encoding_compatible?(encoding, other.encoding)` returns nil.
True
String#casecmp? fails between ascii and utf-8 - Comparing an ASCII-8BIT string with UTF-8, for both of which `#ascii_only?` is true, using `String#casecmp?` returns nil rather than a boolean e.g. ```ruby ''.b.casecmp?('') ``` It returns boolean on MRI, and the implementation to return boolean is present in truffleruby but not reached: https://github.com/oracle/truffleruby/blob/vm-22.3.0/src/main/ruby/truffleruby/core/string.rb#L1204-L1213 The check `if ascii_only? && other.ascii_only?` is not reached after `Primitive.encoding_compatible?(encoding, other.encoding)` returns nil.
comp
string casecmp fails between ascii and utf comparing an ascii string with utf for both of which ascii only is true using string casecmp returns nil rather than a boolean e g ruby b casecmp it returns boolean on mri and the implementation to return boolean is present in truffleruby but not reached the check if ascii only other ascii only is not reached after primitive encoding compatible encoding other encoding returns nil
1
6,271
8,650,230,169
IssuesEvent
2018-11-26 21:51:45
metarhia/impress
https://api.github.com/repos/metarhia/impress
closed
Refactor avoiding mixins
compatibility optimization
Now Impress core uses mixins for everything, it turned to a performance problem, so plugins and objects should be refactored to use: prototypes, factories, functors: - [x] Server - [x] Config - [x] Application - [ ] Plugin - [x] Client - [ ] Connection - [x] Logger
True
Refactor avoiding mixins - Now Impress core uses mixins for everything, it turned to a performance problem, so plugins and objects should be refactored to use: prototypes, factories, functors: - [x] Server - [x] Config - [x] Application - [ ] Plugin - [x] Client - [ ] Connection - [x] Logger
comp
refactor avoiding mixins now impress core uses mixins for everything it turned to a performance problem so plugins and objects should be refactored to use prototypes factories functors server config application plugin client connection logger
1
267,732
28,509,192,190
IssuesEvent
2023-04-19 01:43:16
dpteam/RK3188_TABLET
https://api.github.com/repos/dpteam/RK3188_TABLET
closed
CVE-2017-1000370 (High) detected in linuxv3.0 - autoclosed
Mend: dependency security vulnerability
## CVE-2017-1000370 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv3.0</b></p></summary> <p> <p>Linux kernel source tree</p> <p>Library home page: <a href=https://github.com/verygreen/linux.git>https://github.com/verygreen/linux.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/dpteam/RK3188_TABLET/commit/0c501f5a0fd72c7b2ac82904235363bd44fd8f9e">0c501f5a0fd72c7b2ac82904235363bd44fd8f9e</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (0)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The offset2lib patch as used in the Linux Kernel contains a vulnerability that allows a PIE binary to be execve()'ed with 1GB of arguments or environmental strings then the stack occupies the address 0x80000000 and the PIE binary is mapped above 0x40000000 nullifying the protection of the offset2lib patch. This affects Linux Kernel version 4.11.5 and earlier. This is a different issue than CVE-2017-1000371. This issue appears to be limited to i386 based systems. <p>Publish Date: 2017-06-19 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2017-1000370>CVE-2017-1000370</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-1000370">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-1000370</a></p> <p>Release Date: 2017-06-19</p> <p>Fix Resolution: v4.13-rc1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2017-1000370 (High) detected in linuxv3.0 - autoclosed - ## CVE-2017-1000370 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv3.0</b></p></summary> <p> <p>Linux kernel source tree</p> <p>Library home page: <a href=https://github.com/verygreen/linux.git>https://github.com/verygreen/linux.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/dpteam/RK3188_TABLET/commit/0c501f5a0fd72c7b2ac82904235363bd44fd8f9e">0c501f5a0fd72c7b2ac82904235363bd44fd8f9e</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (0)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The offset2lib patch as used in the Linux Kernel contains a vulnerability that allows a PIE binary to be execve()'ed with 1GB of arguments or environmental strings then the stack occupies the address 0x80000000 and the PIE binary is mapped above 0x40000000 nullifying the protection of the offset2lib patch. This affects Linux Kernel version 4.11.5 and earlier. This is a different issue than CVE-2017-1000371. This issue appears to be limited to i386 based systems. <p>Publish Date: 2017-06-19 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2017-1000370>CVE-2017-1000370</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-1000370">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-1000370</a></p> <p>Release Date: 2017-06-19</p> <p>Fix Resolution: v4.13-rc1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_comp
cve high detected in autoclosed cve high severity vulnerability vulnerable library linux kernel source tree library home page a href found in head commit a href found in base branch master vulnerable source files vulnerability details the patch as used in the linux kernel contains a vulnerability that allows a pie binary to be execve ed with of arguments or environmental strings then the stack occupies the address and the pie binary is mapped above nullifying the protection of the patch this affects linux kernel version and earlier this is a different issue than cve this issue appears to be limited to based systems publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
321,373
27,523,456,765
IssuesEvent
2023-03-06 16:26:09
Automattic/wp-calypso
https://api.github.com/repos/Automattic/wp-calypso
opened
Medium Woo: Verify upgrades from Free trial
Testing Woo Express
Once #74119 and #74120 are complete, we want to verify upgrading from the Woo Express free trial. First create a new trial site by visiting http://calypso.localhost:3000/setup/wooexpress. Then go through the upgrade process by manually adding the Medium plan to the cart and checkin gout `http://calypso.localhost:3000/checkout/:siteSlug/wooexpress-medium-bundle`. Things to verify: TBD
1.0
Medium Woo: Verify upgrades from Free trial - Once #74119 and #74120 are complete, we want to verify upgrading from the Woo Express free trial. First create a new trial site by visiting http://calypso.localhost:3000/setup/wooexpress. Then go through the upgrade process by manually adding the Medium plan to the cart and checkin gout `http://calypso.localhost:3000/checkout/:siteSlug/wooexpress-medium-bundle`. Things to verify: TBD
non_comp
medium woo verify upgrades from free trial once and are complete we want to verify upgrading from the woo express free trial first create a new trial site by visiting then go through the upgrade process by manually adding the medium plan to the cart and checkin gout things to verify tbd
0
17,309
23,884,390,886
IssuesEvent
2022-09-08 06:14:38
elBukkit/MagicPlugin
https://api.github.com/repos/elBukkit/MagicPlugin
reopened
Prevent breaking signs
Compatibility
I have a plugin to lock chests using a sign so only the owner can access it, with spells that break blocks, when the sign is replaced people can access and steal from the chests temporarily until the sign is fixed with the writing that locks it,
True
Prevent breaking signs - I have a plugin to lock chests using a sign so only the owner can access it, with spells that break blocks, when the sign is replaced people can access and steal from the chests temporarily until the sign is fixed with the writing that locks it,
comp
prevent breaking signs i have a plugin to lock chests using a sign so only the owner can access it with spells that break blocks when the sign is replaced people can access and steal from the chests temporarily until the sign is fixed with the writing that locks it
1
134,101
18,421,805,849
IssuesEvent
2021-10-13 16:56:27
daniel-brown-ws-test/verademo
https://api.github.com/repos/daniel-brown-ws-test/verademo
opened
CVE-2019-11358 (Medium) detected in jquery-1.11.2.min.js
security vulnerability
## CVE-2019-11358 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.11.2.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.2/jquery.min.js</a></p> <p>Path to vulnerable library: /app/src/main/webapp/resources/js/jquery-1.11.2.min.js</p> <p> Dependency Hierarchy: - :x: **jquery-1.11.2.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/daniel-brown-ws-test/verademo/commit/919af4911488ae49a0463c8d639240781e4088b0">919af4911488ae49a0463c8d639240781e4088b0</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> jQuery before 3.4.0, as used in Drupal, Backdrop CMS, and other products, mishandles jQuery.extend(true, {}, ...) because of Object.prototype pollution. If an unsanitized source object contained an enumerable __proto__ property, it could extend the native Object.prototype. <p>Publish Date: 2019-04-20 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-11358>CVE-2019-11358</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11358">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11358</a></p> <p>Release Date: 2019-04-20</p> <p>Fix Resolution: 3.4.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.11.2","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"jquery:1.11.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.4.0"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2019-11358","vulnerabilityDetails":"jQuery before 3.4.0, as used in Drupal, Backdrop CMS, and other products, mishandles jQuery.extend(true, {}, ...) because of Object.prototype pollution. If an unsanitized source object contained an enumerable __proto__ property, it could extend the native Object.prototype.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-11358","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
True
CVE-2019-11358 (Medium) detected in jquery-1.11.2.min.js - ## CVE-2019-11358 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.11.2.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.2/jquery.min.js</a></p> <p>Path to vulnerable library: /app/src/main/webapp/resources/js/jquery-1.11.2.min.js</p> <p> Dependency Hierarchy: - :x: **jquery-1.11.2.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/daniel-brown-ws-test/verademo/commit/919af4911488ae49a0463c8d639240781e4088b0">919af4911488ae49a0463c8d639240781e4088b0</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> jQuery before 3.4.0, as used in Drupal, Backdrop CMS, and other products, mishandles jQuery.extend(true, {}, ...) because of Object.prototype pollution. If an unsanitized source object contained an enumerable __proto__ property, it could extend the native Object.prototype. <p>Publish Date: 2019-04-20 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-11358>CVE-2019-11358</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11358">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11358</a></p> <p>Release Date: 2019-04-20</p> <p>Fix Resolution: 3.4.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.11.2","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"jquery:1.11.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.4.0"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2019-11358","vulnerabilityDetails":"jQuery before 3.4.0, as used in Drupal, Backdrop CMS, and other products, mishandles jQuery.extend(true, {}, ...) because of Object.prototype pollution. If an unsanitized source object contained an enumerable __proto__ property, it could extend the native Object.prototype.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-11358","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
non_comp
cve medium detected in jquery min js cve medium severity vulnerability vulnerable library jquery min js javascript library for dom operations library home page a href path to vulnerable library app src main webapp resources js jquery min js dependency hierarchy x jquery min js vulnerable library found in head commit a href found in base branch main vulnerability details jquery before as used in drupal backdrop cms and other products mishandles jquery extend true because of object prototype pollution if an unsanitized source object contained an enumerable proto property it could extend the native object prototype publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree jquery isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails jquery before as used in drupal backdrop cms and other products mishandles jquery extend true because of object prototype pollution if an unsanitized source object contained an enumerable proto property it could extend the native object prototype vulnerabilityurl
0
89,691
11,274,115,183
IssuesEvent
2020-01-14 17:52:26
WordPress/gutenberg
https://api.github.com/repos/WordPress/gutenberg
closed
Rename add_theme_support( 'disable-custom-font-sizes' ) and default it to true
Customization Needs Accessibility Feedback Needs Decision Needs Design Feedback
**Is your feature request related to a problem? Please describe.** Any scenario in which a user needs custom pixel sizes for font-sizing seems fairly niche to me. Is there a specific reason this feature is defaulted to on when there's a good theme-controlled means of enabling a number of variations of font sizing? I propose that the pixel box simply be disabled by default to encourage more use of the primary font-sizing feature. Ideally, I would pair that with adjusting the theme_support property to be both more descriptive and not falling into the weird situation where `true` = disabled. Maybe: `add_theme_support( 'pixel-font-sizes' )`? Further, encouraging custom pixel sizes may complicate implementing more responsive sizing (#11671). **Describe the solution you'd like** 1. Disable custom font sizing in pixels by default. 1. Rename the filter for enabling custom font sizing to make more sense. **Describe alternatives you've considered** The existing default, theme-extensible font sizing feels like it will cover most user requirements and encourage more consistency in site designs _and_ likely lead to smoother theme transitions. Vaguely related: #11671, #9549, #11836
1.0
Rename add_theme_support( 'disable-custom-font-sizes' ) and default it to true - **Is your feature request related to a problem? Please describe.** Any scenario in which a user needs custom pixel sizes for font-sizing seems fairly niche to me. Is there a specific reason this feature is defaulted to on when there's a good theme-controlled means of enabling a number of variations of font sizing? I propose that the pixel box simply be disabled by default to encourage more use of the primary font-sizing feature. Ideally, I would pair that with adjusting the theme_support property to be both more descriptive and not falling into the weird situation where `true` = disabled. Maybe: `add_theme_support( 'pixel-font-sizes' )`? Further, encouraging custom pixel sizes may complicate implementing more responsive sizing (#11671). **Describe the solution you'd like** 1. Disable custom font sizing in pixels by default. 1. Rename the filter for enabling custom font sizing to make more sense. **Describe alternatives you've considered** The existing default, theme-extensible font sizing feels like it will cover most user requirements and encourage more consistency in site designs _and_ likely lead to smoother theme transitions. Vaguely related: #11671, #9549, #11836
non_comp
rename add theme support disable custom font sizes and default it to true is your feature request related to a problem please describe any scenario in which a user needs custom pixel sizes for font sizing seems fairly niche to me is there a specific reason this feature is defaulted to on when there s a good theme controlled means of enabling a number of variations of font sizing i propose that the pixel box simply be disabled by default to encourage more use of the primary font sizing feature ideally i would pair that with adjusting the theme support property to be both more descriptive and not falling into the weird situation where true disabled maybe add theme support pixel font sizes further encouraging custom pixel sizes may complicate implementing more responsive sizing describe the solution you d like disable custom font sizing in pixels by default rename the filter for enabling custom font sizing to make more sense describe alternatives you ve considered the existing default theme extensible font sizing feels like it will cover most user requirements and encourage more consistency in site designs and likely lead to smoother theme transitions vaguely related
0
6,769
9,082,369,266
IssuesEvent
2019-02-17 11:43:23
SpongePowered/SpongeForge
https://api.github.com/repos/SpongePowered/SpongeForge
closed
No entity type is registered for modded entity
status: pr pending system: registry type: mod incompatibility version: 1.12
**I am currently running** - SpongeForge version: 1.12.2-2705-7.1.0-BETA-3355 & 1.12.2-2705-7.1.0-BETA-3358 - Forge version: 14.23.4.2739 - Java version: 1.8.0_171-b10 - Operating System: CentOS <!-- Please include ALL mods/plugins you had installed when your issue happened, you can get a list of your mods and plugins by running "/sponge plugins" and/or "/sponge mods" --> - Plugins/Mods: Plugins (18): Minecraft, Minecraft Coder Pack, SpongeAPI, SpongeForge, AdamantineShield, AntiWDL, BetterChunkLoader, Broadcast, Enjin Minecraft Plugin, GriefPrevention, LuckPerms, MagiBridge, Nucleus, PlayerShopsRPG, Server Backup, SleepVote, Total Economy, WrapperPing Mods (129): Minecraft, Minecraft Coder Pack, Forge Mod Loader, Minecraft Forge, SpongeAPI, SpongeForge, AbyssalCraft, AbyssalCraft Integration, Aether Legacy, AetherWorks, Akashic Tome, Aquaculture, Astral Sorcery, AutoRegLib, Bad Wither No Cookie! Reloaded, Baubles, Better Builder's Wands, BiblioCraft, Bird's Foods, Blockcraftery, Blood Arsenal, Blood Magic: Alchemical Wizardry, Bloodmoon, Bookshelf, Botania, Chameleon, Chameleon Creepers, Chickens, Chisel, Chop Down Updated, CodeChicken Lib, Coralreef, CraftStudio API, CraftTweaker JEI Support, CraftTweaker2, Creeper Confetti, Crossroads, Cucumber Library, Cyclops Core, Diet Hopper, Doomlike Dungeons, DummyCoreUnofficial, Dungeon Mods, Dungeons2!, Electroblob's Wizardry, EluCore, EluLib, Embers, Essentials, EvilCraft, EvilCraft-Compat, Exotic Birds, FoamFix, FoamFixCore, Geolosys, GottschCore, Gravestone Mod, Guide-API, HardLib, Harder Farming, HelpFixer, Huggable Cactuses, Ice and Fire, Immersive Craft, Immersive Engineering, Immersive Petroleum, Infernal Mobs, Inventory Pets, Inventory Tweaks, Iron Backpacks, Iron Chest, Just Enough Buttons, Just Enough Items, Just a Few Fish, KleeSlabs, LLibrary, LagGoggles, Level Up! Reloaded, Lost Souls, Lumen, Mantle, McJtyLib, Mob Totems, Mobultion, Moon's Core, Mowzie's Mobs, Mystical Agriculture, Nature's Compass, OMLib, Open Modular Passive Defense, Overlord, ProjectE, Prospects, Reliquary, Reptile Mod, Roguelike Dungeons, Roots Classic, Runes of Wizardry, Runes of Wizardry - Classic Dusts Pack, ShetiPhian-Core, Silent Lib, Silent's Gems, Silent's Gems: Extra Parts, Snad, Soul Shards - The Old Ways, StepUp, Storage Drawers, SwingThroughGrass, TamModized, TeamLapen Library, Terraqueous, Thaumcrafft Inventory Scanning, Thaumcraft, Thaumic Bases, Thaumic Periphery, Thaumic Tinkerer, The Lost Cities, TorchMaster, Totemic, Underground Biomes, Vampirism, VampirismIntegrations, Village Names, VoidCraft, Waila, Waystones, WorldEdit, Zoo and Wild Animals Mod: Rebuilt, aqua_creepers **Issue Description** When typing `/sponge -g reload` from console, or as player, leads to an exception ``` [18:23:20] [Server thread/ERROR] [Sponge]: Error occurred while executing command 'sponge -g reload' for source EntityPlayerMP['LemADEC'/30309, l='Twinkle', x=302.62, y=65.00, z=406.51]: org.spongepowered.common.entity.SpongeEntityType$1 cannot be cast to org.spongepowered.common.entity.SpongeEntityType java.lang.ClassCastException: org.spongepowered.common.entity.SpongeEntityType$1 cannot be cast to org.spongepowered.common.entity.SpongeEntityType > at net.minecraft.entity.Entity.refreshCache(Entity.java:6237) ~[vg.class:?] ``` https://gist.github.com/LemADEC/1799a9e8370a423e34715aeba6f8e40f
True
No entity type is registered for modded entity - **I am currently running** - SpongeForge version: 1.12.2-2705-7.1.0-BETA-3355 & 1.12.2-2705-7.1.0-BETA-3358 - Forge version: 14.23.4.2739 - Java version: 1.8.0_171-b10 - Operating System: CentOS <!-- Please include ALL mods/plugins you had installed when your issue happened, you can get a list of your mods and plugins by running "/sponge plugins" and/or "/sponge mods" --> - Plugins/Mods: Plugins (18): Minecraft, Minecraft Coder Pack, SpongeAPI, SpongeForge, AdamantineShield, AntiWDL, BetterChunkLoader, Broadcast, Enjin Minecraft Plugin, GriefPrevention, LuckPerms, MagiBridge, Nucleus, PlayerShopsRPG, Server Backup, SleepVote, Total Economy, WrapperPing Mods (129): Minecraft, Minecraft Coder Pack, Forge Mod Loader, Minecraft Forge, SpongeAPI, SpongeForge, AbyssalCraft, AbyssalCraft Integration, Aether Legacy, AetherWorks, Akashic Tome, Aquaculture, Astral Sorcery, AutoRegLib, Bad Wither No Cookie! Reloaded, Baubles, Better Builder's Wands, BiblioCraft, Bird's Foods, Blockcraftery, Blood Arsenal, Blood Magic: Alchemical Wizardry, Bloodmoon, Bookshelf, Botania, Chameleon, Chameleon Creepers, Chickens, Chisel, Chop Down Updated, CodeChicken Lib, Coralreef, CraftStudio API, CraftTweaker JEI Support, CraftTweaker2, Creeper Confetti, Crossroads, Cucumber Library, Cyclops Core, Diet Hopper, Doomlike Dungeons, DummyCoreUnofficial, Dungeon Mods, Dungeons2!, Electroblob's Wizardry, EluCore, EluLib, Embers, Essentials, EvilCraft, EvilCraft-Compat, Exotic Birds, FoamFix, FoamFixCore, Geolosys, GottschCore, Gravestone Mod, Guide-API, HardLib, Harder Farming, HelpFixer, Huggable Cactuses, Ice and Fire, Immersive Craft, Immersive Engineering, Immersive Petroleum, Infernal Mobs, Inventory Pets, Inventory Tweaks, Iron Backpacks, Iron Chest, Just Enough Buttons, Just Enough Items, Just a Few Fish, KleeSlabs, LLibrary, LagGoggles, Level Up! Reloaded, Lost Souls, Lumen, Mantle, McJtyLib, Mob Totems, Mobultion, Moon's Core, Mowzie's Mobs, Mystical Agriculture, Nature's Compass, OMLib, Open Modular Passive Defense, Overlord, ProjectE, Prospects, Reliquary, Reptile Mod, Roguelike Dungeons, Roots Classic, Runes of Wizardry, Runes of Wizardry - Classic Dusts Pack, ShetiPhian-Core, Silent Lib, Silent's Gems, Silent's Gems: Extra Parts, Snad, Soul Shards - The Old Ways, StepUp, Storage Drawers, SwingThroughGrass, TamModized, TeamLapen Library, Terraqueous, Thaumcrafft Inventory Scanning, Thaumcraft, Thaumic Bases, Thaumic Periphery, Thaumic Tinkerer, The Lost Cities, TorchMaster, Totemic, Underground Biomes, Vampirism, VampirismIntegrations, Village Names, VoidCraft, Waila, Waystones, WorldEdit, Zoo and Wild Animals Mod: Rebuilt, aqua_creepers **Issue Description** When typing `/sponge -g reload` from console, or as player, leads to an exception ``` [18:23:20] [Server thread/ERROR] [Sponge]: Error occurred while executing command 'sponge -g reload' for source EntityPlayerMP['LemADEC'/30309, l='Twinkle', x=302.62, y=65.00, z=406.51]: org.spongepowered.common.entity.SpongeEntityType$1 cannot be cast to org.spongepowered.common.entity.SpongeEntityType java.lang.ClassCastException: org.spongepowered.common.entity.SpongeEntityType$1 cannot be cast to org.spongepowered.common.entity.SpongeEntityType > at net.minecraft.entity.Entity.refreshCache(Entity.java:6237) ~[vg.class:?] ``` https://gist.github.com/LemADEC/1799a9e8370a423e34715aeba6f8e40f
comp
no entity type is registered for modded entity i am currently running spongeforge version beta beta forge version java version operating system centos please include all mods plugins you had installed when your issue happened you can get a list of your mods and plugins by running sponge plugins and or sponge mods plugins mods plugins minecraft minecraft coder pack spongeapi spongeforge adamantineshield antiwdl betterchunkloader broadcast enjin minecraft plugin griefprevention luckperms magibridge nucleus playershopsrpg server backup sleepvote total economy wrapperping mods minecraft minecraft coder pack forge mod loader minecraft forge spongeapi spongeforge abyssalcraft abyssalcraft integration aether legacy aetherworks akashic tome aquaculture astral sorcery autoreglib bad wither no cookie reloaded baubles better builder s wands bibliocraft bird s foods blockcraftery blood arsenal blood magic alchemical wizardry bloodmoon bookshelf botania chameleon chameleon creepers chickens chisel chop down updated codechicken lib coralreef craftstudio api crafttweaker jei support creeper confetti crossroads cucumber library cyclops core diet hopper doomlike dungeons dummycoreunofficial dungeon mods electroblob s wizardry elucore elulib embers essentials evilcraft evilcraft compat exotic birds foamfix foamfixcore geolosys gottschcore gravestone mod guide api hardlib harder farming helpfixer huggable cactuses ice and fire immersive craft immersive engineering immersive petroleum infernal mobs inventory pets inventory tweaks iron backpacks iron chest just enough buttons just enough items just a few fish kleeslabs llibrary laggoggles level up reloaded lost souls lumen mantle mcjtylib mob totems mobultion moon s core mowzie s mobs mystical agriculture nature s compass omlib open modular passive defense overlord projecte prospects reliquary reptile mod roguelike dungeons roots classic runes of wizardry runes of wizardry classic dusts pack shetiphian core silent lib silent s gems silent s gems extra parts snad soul shards the old ways stepup storage drawers swingthroughgrass tammodized teamlapen library terraqueous thaumcrafft inventory scanning thaumcraft thaumic bases thaumic periphery thaumic tinkerer the lost cities torchmaster totemic underground biomes vampirism vampirismintegrations village names voidcraft waila waystones worldedit zoo and wild animals mod rebuilt aqua creepers issue description when typing sponge g reload from console or as player leads to an exception error occurred while executing command sponge g reload for source entityplayermp org spongepowered common entity spongeentitytype cannot be cast to org spongepowered common entity spongeentitytype java lang classcastexception org spongepowered common entity spongeentitytype cannot be cast to org spongepowered common entity spongeentitytype at net minecraft entity entity refreshcache entity java
1
170,535
20,883,752,102
IssuesEvent
2022-03-23 01:08:53
snowdensb/dependabot-core
https://api.github.com/repos/snowdensb/dependabot-core
reopened
CVE-2018-16487 (Medium) detected in multiple libraries
security vulnerability
## CVE-2018-16487 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>lodash-2.4.1.tgz</b>, <b>lodash-4.17.4.tgz</b>, <b>lodash-4.17.5.tgz</b>, <b>lodash-1.3.1.tgz</b>, <b>lodash-1.2.0.tgz</b>, <b>lodash-3.9.3.tgz</b>, <b>lodash-0.1.0.tgz</b>, <b>lodash-1.2.1.tgz</b>, <b>lodash-3.10.0.tgz</b>, <b>lodash-3.10.1.tgz</b></p></summary> <p> <details><summary><b>lodash-2.4.1.tgz</b></p></summary> <p>A utility library delivering consistency, customization, performance, & extras.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-2.4.1.tgz">https://registry.npmjs.org/lodash/-/lodash-2.4.1.tgz</a></p> <p>Path to dependency file: /npm_and_yarn/spec/fixtures/projects/npm6/multiple_sources/package.json</p> <p>Path to vulnerable library: /npm_and_yarn/spec/fixtures/projects/npm6/multiple_sources/node_modules/nock/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/npm6/duplicate/node_modules/nock/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/npm7/duplicate_identical/node_modules/nock/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/npm7/minor_version_specified/node_modules/nock/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - fetch-factory-0.1.0.tgz (Root Library) - nock-2.18.2.tgz - :x: **lodash-2.4.1.tgz** (Vulnerable Library) </details> <details><summary><b>lodash-4.17.4.tgz</b></p></summary> <p>Lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.4.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.4.tgz</a></p> <p>Path to dependency file: /npm_and_yarn/spec/fixtures/projects/yarn/lockfile_only_change/package.json</p> <p>Path to vulnerable library: /npm_and_yarn/spec/fixtures/projects/yarn/lockfile_only_change/node_modules/lodash,/npm_and_yarn/spec/fixtures/projects/yarn/no_lockfile_change/node_modules/lodash,/npm_and_yarn/spec/fixtures/projects/yarn/typedoc-plugin-ui-router/node_modules/lodash</p> <p> Dependency Hierarchy: - babel-register-6.24.1.tgz (Root Library) - :x: **lodash-4.17.4.tgz** (Vulnerable Library) </details> <details><summary><b>lodash-4.17.5.tgz</b></p></summary> <p>Lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.5.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.5.tgz</a></p> <p>Path to dependency file: /npm_and_yarn/spec/fixtures/projects/yarn/duplicate_indirect_dependency/package.json</p> <p>Path to vulnerable library: /npm_and_yarn/spec/fixtures/projects/yarn/duplicate_indirect_dependency/node_modules/lodash</p> <p> Dependency Hierarchy: - graphql-cli-3.0.3.tgz (Root Library) - graphql-cli-prepare-1.4.19.tgz - :x: **lodash-4.17.5.tgz** (Vulnerable Library) </details> <details><summary><b>lodash-1.3.1.tgz</b></p></summary> <p>A utility library delivering consistency, customization, performance, and extras.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-1.3.1.tgz">https://registry.npmjs.org/lodash/-/lodash-1.3.1.tgz</a></p> <p>Path to dependency file: /npm_and_yarn/spec/fixtures/projects/npm7/git_dependency/package.json</p> <p>Path to vulnerable library: /npm_and_yarn/spec/fixtures/projects/npm7/git_dependency/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/npm6/git_dependency/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/npm6/git_dependency_version/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/yarn/other_package/node_modules/lodash,/npm_and_yarn/spec/fixtures/projects/yarn/workspaces_bad/node_modules/lodash,/npm_and_yarn/spec/fixtures/projects/yarn/git_dependency_from_subdep/node_modules/lodash,/npm_and_yarn/spec/fixtures/projects/yarn/git_dependency_with_auth/node_modules/lodash,/npm_and_yarn/spec/fixtures/projects/npm6/git_dependency_no_lockfile/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/yarn/git_dependency/node_modules/lodash,/npm_and_yarn/spec/fixtures/projects/yarn/workspaces/node_modules/lodash,/npm_and_yarn/spec/fixtures/projects/yarn/symlinked_dependency/node_modules/lodash,/npm_and_yarn/spec/fixtures/projects/npm6_and_yarn/git_dependency/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/npm6/git_dependency_branch_version/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/yarn/git_dependency_outdated_req/node_modules/lodash,/npm_and_yarn/spec/fixtures/projects/npm7/git_dependency_git_url/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/npm6_and_yarn/git_dependency_git_url/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/yarn/path_dependency/node_modules/lodash,/npm_and_yarn/spec/fixtures/projects/npm7/git_dependency_version/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - :x: **lodash-1.3.1.tgz** (Vulnerable Library) </details> <details><summary><b>lodash-1.2.0.tgz</b></p></summary> <p>A utility library delivering consistency, customization, performance, and extras.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-1.2.0.tgz">https://registry.npmjs.org/lodash/-/lodash-1.2.0.tgz</a></p> <p>Path to dependency file: /npm_and_yarn/spec/fixtures/updated_projects/npm7/workspaces_dev/package.json</p> <p>Path to vulnerable library: /npm_and_yarn/spec/fixtures/updated_projects/npm7/workspaces_dev/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/npm7/workspaces/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/yarn/workspaces/node_modules/lodash,/npm_and_yarn/spec/fixtures/projects/yarn/workspaces_bad/node_modules/lodash</p> <p> Dependency Hierarchy: - :x: **lodash-1.2.0.tgz** (Vulnerable Library) </details> <details><summary><b>lodash-3.9.3.tgz</b></p></summary> <p>The modern build of lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.9.3.tgz">https://registry.npmjs.org/lodash/-/lodash-3.9.3.tgz</a></p> <p>Path to dependency file: /npm_and_yarn/spec/fixtures/projects/yarn/lockfile_only_change/package.json</p> <p>Path to vulnerable library: /npm_and_yarn/spec/fixtures/projects/yarn/lockfile_only_change/node_modules/lodash</p> <p> Dependency Hierarchy: - git-contributors-0.2.3.tgz (Root Library) - :x: **lodash-3.9.3.tgz** (Vulnerable Library) </details> <details><summary><b>lodash-0.1.0.tgz</b></p></summary> <p>A drop-in replacement for Underscore.js that delivers up to 8x performance improvements, bug fixes, and additional features.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-0.1.0.tgz">https://registry.npmjs.org/lodash/-/lodash-0.1.0.tgz</a></p> <p>Path to dependency file: /npm_and_yarn/spec/fixtures/projects/yarn/resolutions/package.json</p> <p>Path to vulnerable library: /npm_and_yarn/spec/fixtures/projects/yarn/resolutions/node_modules/lodash</p> <p> Dependency Hierarchy: - :x: **lodash-0.1.0.tgz** (Vulnerable Library) </details> <details><summary><b>lodash-1.2.1.tgz</b></p></summary> <p>A utility library delivering consistency, customization, performance, and extras.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-1.2.1.tgz">https://registry.npmjs.org/lodash/-/lodash-1.2.1.tgz</a></p> <p>Path to dependency file: /npm_and_yarn/spec/fixtures/projects/npm6_and_yarn/path_dependency/package.json</p> <p>Path to vulnerable library: /npm_and_yarn/spec/fixtures/projects/npm6_and_yarn/path_dependency/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/npm6/path_dependency/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - :x: **lodash-1.2.1.tgz** (Vulnerable Library) </details> <details><summary><b>lodash-3.10.0.tgz</b></p></summary> <p>The modern build of lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.10.0.tgz">https://registry.npmjs.org/lodash/-/lodash-3.10.0.tgz</a></p> <p>Path to dependency file: /npm_and_yarn/spec/fixtures/projects/yarn/resolution_specified/package.json</p> <p>Path to vulnerable library: /npm_and_yarn/spec/fixtures/projects/yarn/resolution_specified/node_modules/lodash</p> <p> Dependency Hierarchy: - :x: **lodash-3.10.0.tgz** (Vulnerable Library) </details> <details><summary><b>lodash-3.10.1.tgz</b></p></summary> <p>The modern build of lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz">https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz</a></p> <p>Path to dependency file: /npm_and_yarn/spec/fixtures/projects/npm7/minor_version_specified/package.json</p> <p>Path to vulnerable library: /npm_and_yarn/spec/fixtures/projects/npm7/minor_version_specified/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/yarn/no_lockfile_change/node_modules/lodash,/npm_and_yarn/spec/fixtures/projects/npm6_and_yarn/npm_subdependency_update/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/yarn/yanked_version/node_modules/lodash,/npm_and_yarn/spec/fixtures/projects/npm7/git_sub_dep_invalid_from/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/npm6/monorepo_dep_multiple_no_lockfile/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/npm7/subdependency_update/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/npm7/tarball_bug/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/npm5_and_yarn/npm_subdependency_update/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/npm6/multiple_sources/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/yarn/aliased_dependency_name/node_modules/lodash,/npm_and_yarn/spec/fixtures/projects/npm7/monorepo_dep_multiple/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/npm7/duplicate_identical/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/yarn/optional_dependencies/node_modules/lodash,/npm_and_yarn/spec/fixtures/projects/yarn/yarnrc_npm_registry/node_modules/lodash,/npm_and_yarn/spec/fixtures/projects/npm7/app_no_version/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/npm6/subdependency_update/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/npm6/tarball_bug/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/yarn/dist_tag/node_modules/lodash,/npm_and_yarn/spec/fixtures/projects/npm5/subdependency_update/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/npm6/peer_dependency_mismatch_no_lockfile/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/npm5/git_sub_dep_invalid/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/npm6/duplicate/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/npm6/optional_dependencies/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/npm6/monorepo_dep_single_no_lockfile/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/yarn/aliased_dependency/node_modules/lodash</p> <p> Dependency Hierarchy: - npm-5.8.0.tgz (Root Library) - cli-table2-0.2.0.tgz - :x: **lodash-3.10.1.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/snowdensb/dependabot-core/commit/ba8cd9078c8ce0cb202767d627706711237abf71">ba8cd9078c8ce0cb202767d627706711237abf71</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A prototype pollution vulnerability was found in lodash <4.17.11 where the functions merge, mergeWith, and defaultsDeep can be tricked into adding or modifying properties of Object.prototype. <p>Publish Date: 2019-02-01 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-16487>CVE-2018-16487</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2018-16487">https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2018-16487</a></p> <p>Release Date: 2019-02-01</p> <p>Fix Resolution (lodash): 4.17.11</p> <p>Direct dependency fix Resolution (babel-register): 6.26.0</p><p>Fix Resolution (lodash): 4.17.11</p> <p>Direct dependency fix Resolution (graphql-cli): 4.0.0-alpha.0</p><p>Fix Resolution (lodash): 4.17.11</p> <p>Direct dependency fix Resolution (npm): 6.2.0</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"2.4.1","packageFilePaths":["/npm_and_yarn/spec/fixtures/projects/npm6/multiple_sources/package.json"],"isTransitiveDependency":true,"dependencyTree":"fetch-factory:0.1.0;nock:2.18.2;lodash:2.4.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.17.11","isBinary":false},{"packageType":"javascript/Node.js","packageName":"babel-register","packageVersion":"6.24.1","packageFilePaths":["/npm_and_yarn/spec/fixtures/projects/yarn/lockfile_only_change/package.json"],"isTransitiveDependency":false,"dependencyTree":"babel-register:6.24.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"6.26.0","isBinary":false},{"packageType":"javascript/Node.js","packageName":"graphql-cli","packageVersion":"3.0.3","packageFilePaths":["/npm_and_yarn/spec/fixtures/projects/yarn/duplicate_indirect_dependency/package.json"],"isTransitiveDependency":false,"dependencyTree":"graphql-cli:3.0.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.0.0-alpha.0","isBinary":false},{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"1.3.1","packageFilePaths":["/npm_and_yarn/spec/fixtures/projects/npm7/git_dependency/package.json"],"isTransitiveDependency":false,"dependencyTree":"lodash:1.3.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.17.11","isBinary":false},{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"1.2.0","packageFilePaths":["/npm_and_yarn/spec/fixtures/updated_projects/npm7/workspaces_dev/package.json"],"isTransitiveDependency":false,"dependencyTree":"lodash:1.2.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.17.11","isBinary":false},{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"3.9.3","packageFilePaths":["/npm_and_yarn/spec/fixtures/projects/yarn/lockfile_only_change/package.json"],"isTransitiveDependency":true,"dependencyTree":"git-contributors:0.2.3;lodash:3.9.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.17.11","isBinary":false},{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"0.1.0","packageFilePaths":["/npm_and_yarn/spec/fixtures/projects/yarn/resolutions/package.json"],"isTransitiveDependency":false,"dependencyTree":"lodash:0.1.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.17.11","isBinary":false},{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"1.2.1","packageFilePaths":["/npm_and_yarn/spec/fixtures/projects/npm6_and_yarn/path_dependency/package.json"],"isTransitiveDependency":false,"dependencyTree":"lodash:1.2.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.17.11","isBinary":false},{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"3.10.0","packageFilePaths":["/npm_and_yarn/spec/fixtures/projects/yarn/resolution_specified/package.json"],"isTransitiveDependency":false,"dependencyTree":"lodash:3.10.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.17.11","isBinary":false},{"packageType":"javascript/Node.js","packageName":"npm","packageVersion":"5.8.0","packageFilePaths":["/npm_and_yarn/spec/fixtures/projects/npm7/minor_version_specified/package.json"],"isTransitiveDependency":false,"dependencyTree":"npm:5.8.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"6.2.0","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2018-16487","vulnerabilityDetails":"A prototype pollution vulnerability was found in lodash \u003c4.17.11 where the functions merge, mergeWith, and defaultsDeep can be tricked into adding or modifying properties of Object.prototype.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-16487","cvss3Severity":"medium","cvss3Score":"5.6","cvss3Metrics":{"A":"Low","AC":"High","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
True
CVE-2018-16487 (Medium) detected in multiple libraries - ## CVE-2018-16487 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>lodash-2.4.1.tgz</b>, <b>lodash-4.17.4.tgz</b>, <b>lodash-4.17.5.tgz</b>, <b>lodash-1.3.1.tgz</b>, <b>lodash-1.2.0.tgz</b>, <b>lodash-3.9.3.tgz</b>, <b>lodash-0.1.0.tgz</b>, <b>lodash-1.2.1.tgz</b>, <b>lodash-3.10.0.tgz</b>, <b>lodash-3.10.1.tgz</b></p></summary> <p> <details><summary><b>lodash-2.4.1.tgz</b></p></summary> <p>A utility library delivering consistency, customization, performance, & extras.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-2.4.1.tgz">https://registry.npmjs.org/lodash/-/lodash-2.4.1.tgz</a></p> <p>Path to dependency file: /npm_and_yarn/spec/fixtures/projects/npm6/multiple_sources/package.json</p> <p>Path to vulnerable library: /npm_and_yarn/spec/fixtures/projects/npm6/multiple_sources/node_modules/nock/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/npm6/duplicate/node_modules/nock/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/npm7/duplicate_identical/node_modules/nock/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/npm7/minor_version_specified/node_modules/nock/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - fetch-factory-0.1.0.tgz (Root Library) - nock-2.18.2.tgz - :x: **lodash-2.4.1.tgz** (Vulnerable Library) </details> <details><summary><b>lodash-4.17.4.tgz</b></p></summary> <p>Lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.4.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.4.tgz</a></p> <p>Path to dependency file: /npm_and_yarn/spec/fixtures/projects/yarn/lockfile_only_change/package.json</p> <p>Path to vulnerable library: /npm_and_yarn/spec/fixtures/projects/yarn/lockfile_only_change/node_modules/lodash,/npm_and_yarn/spec/fixtures/projects/yarn/no_lockfile_change/node_modules/lodash,/npm_and_yarn/spec/fixtures/projects/yarn/typedoc-plugin-ui-router/node_modules/lodash</p> <p> Dependency Hierarchy: - babel-register-6.24.1.tgz (Root Library) - :x: **lodash-4.17.4.tgz** (Vulnerable Library) </details> <details><summary><b>lodash-4.17.5.tgz</b></p></summary> <p>Lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.5.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.5.tgz</a></p> <p>Path to dependency file: /npm_and_yarn/spec/fixtures/projects/yarn/duplicate_indirect_dependency/package.json</p> <p>Path to vulnerable library: /npm_and_yarn/spec/fixtures/projects/yarn/duplicate_indirect_dependency/node_modules/lodash</p> <p> Dependency Hierarchy: - graphql-cli-3.0.3.tgz (Root Library) - graphql-cli-prepare-1.4.19.tgz - :x: **lodash-4.17.5.tgz** (Vulnerable Library) </details> <details><summary><b>lodash-1.3.1.tgz</b></p></summary> <p>A utility library delivering consistency, customization, performance, and extras.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-1.3.1.tgz">https://registry.npmjs.org/lodash/-/lodash-1.3.1.tgz</a></p> <p>Path to dependency file: /npm_and_yarn/spec/fixtures/projects/npm7/git_dependency/package.json</p> <p>Path to vulnerable library: /npm_and_yarn/spec/fixtures/projects/npm7/git_dependency/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/npm6/git_dependency/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/npm6/git_dependency_version/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/yarn/other_package/node_modules/lodash,/npm_and_yarn/spec/fixtures/projects/yarn/workspaces_bad/node_modules/lodash,/npm_and_yarn/spec/fixtures/projects/yarn/git_dependency_from_subdep/node_modules/lodash,/npm_and_yarn/spec/fixtures/projects/yarn/git_dependency_with_auth/node_modules/lodash,/npm_and_yarn/spec/fixtures/projects/npm6/git_dependency_no_lockfile/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/yarn/git_dependency/node_modules/lodash,/npm_and_yarn/spec/fixtures/projects/yarn/workspaces/node_modules/lodash,/npm_and_yarn/spec/fixtures/projects/yarn/symlinked_dependency/node_modules/lodash,/npm_and_yarn/spec/fixtures/projects/npm6_and_yarn/git_dependency/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/npm6/git_dependency_branch_version/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/yarn/git_dependency_outdated_req/node_modules/lodash,/npm_and_yarn/spec/fixtures/projects/npm7/git_dependency_git_url/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/npm6_and_yarn/git_dependency_git_url/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/yarn/path_dependency/node_modules/lodash,/npm_and_yarn/spec/fixtures/projects/npm7/git_dependency_version/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - :x: **lodash-1.3.1.tgz** (Vulnerable Library) </details> <details><summary><b>lodash-1.2.0.tgz</b></p></summary> <p>A utility library delivering consistency, customization, performance, and extras.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-1.2.0.tgz">https://registry.npmjs.org/lodash/-/lodash-1.2.0.tgz</a></p> <p>Path to dependency file: /npm_and_yarn/spec/fixtures/updated_projects/npm7/workspaces_dev/package.json</p> <p>Path to vulnerable library: /npm_and_yarn/spec/fixtures/updated_projects/npm7/workspaces_dev/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/npm7/workspaces/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/yarn/workspaces/node_modules/lodash,/npm_and_yarn/spec/fixtures/projects/yarn/workspaces_bad/node_modules/lodash</p> <p> Dependency Hierarchy: - :x: **lodash-1.2.0.tgz** (Vulnerable Library) </details> <details><summary><b>lodash-3.9.3.tgz</b></p></summary> <p>The modern build of lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.9.3.tgz">https://registry.npmjs.org/lodash/-/lodash-3.9.3.tgz</a></p> <p>Path to dependency file: /npm_and_yarn/spec/fixtures/projects/yarn/lockfile_only_change/package.json</p> <p>Path to vulnerable library: /npm_and_yarn/spec/fixtures/projects/yarn/lockfile_only_change/node_modules/lodash</p> <p> Dependency Hierarchy: - git-contributors-0.2.3.tgz (Root Library) - :x: **lodash-3.9.3.tgz** (Vulnerable Library) </details> <details><summary><b>lodash-0.1.0.tgz</b></p></summary> <p>A drop-in replacement for Underscore.js that delivers up to 8x performance improvements, bug fixes, and additional features.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-0.1.0.tgz">https://registry.npmjs.org/lodash/-/lodash-0.1.0.tgz</a></p> <p>Path to dependency file: /npm_and_yarn/spec/fixtures/projects/yarn/resolutions/package.json</p> <p>Path to vulnerable library: /npm_and_yarn/spec/fixtures/projects/yarn/resolutions/node_modules/lodash</p> <p> Dependency Hierarchy: - :x: **lodash-0.1.0.tgz** (Vulnerable Library) </details> <details><summary><b>lodash-1.2.1.tgz</b></p></summary> <p>A utility library delivering consistency, customization, performance, and extras.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-1.2.1.tgz">https://registry.npmjs.org/lodash/-/lodash-1.2.1.tgz</a></p> <p>Path to dependency file: /npm_and_yarn/spec/fixtures/projects/npm6_and_yarn/path_dependency/package.json</p> <p>Path to vulnerable library: /npm_and_yarn/spec/fixtures/projects/npm6_and_yarn/path_dependency/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/npm6/path_dependency/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - :x: **lodash-1.2.1.tgz** (Vulnerable Library) </details> <details><summary><b>lodash-3.10.0.tgz</b></p></summary> <p>The modern build of lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.10.0.tgz">https://registry.npmjs.org/lodash/-/lodash-3.10.0.tgz</a></p> <p>Path to dependency file: /npm_and_yarn/spec/fixtures/projects/yarn/resolution_specified/package.json</p> <p>Path to vulnerable library: /npm_and_yarn/spec/fixtures/projects/yarn/resolution_specified/node_modules/lodash</p> <p> Dependency Hierarchy: - :x: **lodash-3.10.0.tgz** (Vulnerable Library) </details> <details><summary><b>lodash-3.10.1.tgz</b></p></summary> <p>The modern build of lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz">https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz</a></p> <p>Path to dependency file: /npm_and_yarn/spec/fixtures/projects/npm7/minor_version_specified/package.json</p> <p>Path to vulnerable library: /npm_and_yarn/spec/fixtures/projects/npm7/minor_version_specified/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/yarn/no_lockfile_change/node_modules/lodash,/npm_and_yarn/spec/fixtures/projects/npm6_and_yarn/npm_subdependency_update/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/yarn/yanked_version/node_modules/lodash,/npm_and_yarn/spec/fixtures/projects/npm7/git_sub_dep_invalid_from/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/npm6/monorepo_dep_multiple_no_lockfile/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/npm7/subdependency_update/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/npm7/tarball_bug/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/npm5_and_yarn/npm_subdependency_update/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/npm6/multiple_sources/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/yarn/aliased_dependency_name/node_modules/lodash,/npm_and_yarn/spec/fixtures/projects/npm7/monorepo_dep_multiple/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/npm7/duplicate_identical/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/yarn/optional_dependencies/node_modules/lodash,/npm_and_yarn/spec/fixtures/projects/yarn/yarnrc_npm_registry/node_modules/lodash,/npm_and_yarn/spec/fixtures/projects/npm7/app_no_version/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/npm6/subdependency_update/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/npm6/tarball_bug/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/yarn/dist_tag/node_modules/lodash,/npm_and_yarn/spec/fixtures/projects/npm5/subdependency_update/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/npm6/peer_dependency_mismatch_no_lockfile/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/npm5/git_sub_dep_invalid/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/npm6/duplicate/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/npm6/optional_dependencies/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/npm6/monorepo_dep_single_no_lockfile/node_modules/lodash/package.json,/npm_and_yarn/spec/fixtures/projects/yarn/aliased_dependency/node_modules/lodash</p> <p> Dependency Hierarchy: - npm-5.8.0.tgz (Root Library) - cli-table2-0.2.0.tgz - :x: **lodash-3.10.1.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/snowdensb/dependabot-core/commit/ba8cd9078c8ce0cb202767d627706711237abf71">ba8cd9078c8ce0cb202767d627706711237abf71</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A prototype pollution vulnerability was found in lodash <4.17.11 where the functions merge, mergeWith, and defaultsDeep can be tricked into adding or modifying properties of Object.prototype. <p>Publish Date: 2019-02-01 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-16487>CVE-2018-16487</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2018-16487">https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2018-16487</a></p> <p>Release Date: 2019-02-01</p> <p>Fix Resolution (lodash): 4.17.11</p> <p>Direct dependency fix Resolution (babel-register): 6.26.0</p><p>Fix Resolution (lodash): 4.17.11</p> <p>Direct dependency fix Resolution (graphql-cli): 4.0.0-alpha.0</p><p>Fix Resolution (lodash): 4.17.11</p> <p>Direct dependency fix Resolution (npm): 6.2.0</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"2.4.1","packageFilePaths":["/npm_and_yarn/spec/fixtures/projects/npm6/multiple_sources/package.json"],"isTransitiveDependency":true,"dependencyTree":"fetch-factory:0.1.0;nock:2.18.2;lodash:2.4.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.17.11","isBinary":false},{"packageType":"javascript/Node.js","packageName":"babel-register","packageVersion":"6.24.1","packageFilePaths":["/npm_and_yarn/spec/fixtures/projects/yarn/lockfile_only_change/package.json"],"isTransitiveDependency":false,"dependencyTree":"babel-register:6.24.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"6.26.0","isBinary":false},{"packageType":"javascript/Node.js","packageName":"graphql-cli","packageVersion":"3.0.3","packageFilePaths":["/npm_and_yarn/spec/fixtures/projects/yarn/duplicate_indirect_dependency/package.json"],"isTransitiveDependency":false,"dependencyTree":"graphql-cli:3.0.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.0.0-alpha.0","isBinary":false},{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"1.3.1","packageFilePaths":["/npm_and_yarn/spec/fixtures/projects/npm7/git_dependency/package.json"],"isTransitiveDependency":false,"dependencyTree":"lodash:1.3.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.17.11","isBinary":false},{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"1.2.0","packageFilePaths":["/npm_and_yarn/spec/fixtures/updated_projects/npm7/workspaces_dev/package.json"],"isTransitiveDependency":false,"dependencyTree":"lodash:1.2.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.17.11","isBinary":false},{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"3.9.3","packageFilePaths":["/npm_and_yarn/spec/fixtures/projects/yarn/lockfile_only_change/package.json"],"isTransitiveDependency":true,"dependencyTree":"git-contributors:0.2.3;lodash:3.9.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.17.11","isBinary":false},{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"0.1.0","packageFilePaths":["/npm_and_yarn/spec/fixtures/projects/yarn/resolutions/package.json"],"isTransitiveDependency":false,"dependencyTree":"lodash:0.1.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.17.11","isBinary":false},{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"1.2.1","packageFilePaths":["/npm_and_yarn/spec/fixtures/projects/npm6_and_yarn/path_dependency/package.json"],"isTransitiveDependency":false,"dependencyTree":"lodash:1.2.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.17.11","isBinary":false},{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"3.10.0","packageFilePaths":["/npm_and_yarn/spec/fixtures/projects/yarn/resolution_specified/package.json"],"isTransitiveDependency":false,"dependencyTree":"lodash:3.10.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.17.11","isBinary":false},{"packageType":"javascript/Node.js","packageName":"npm","packageVersion":"5.8.0","packageFilePaths":["/npm_and_yarn/spec/fixtures/projects/npm7/minor_version_specified/package.json"],"isTransitiveDependency":false,"dependencyTree":"npm:5.8.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"6.2.0","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2018-16487","vulnerabilityDetails":"A prototype pollution vulnerability was found in lodash \u003c4.17.11 where the functions merge, mergeWith, and defaultsDeep can be tricked into adding or modifying properties of Object.prototype.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-16487","cvss3Severity":"medium","cvss3Score":"5.6","cvss3Metrics":{"A":"Low","AC":"High","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
non_comp
cve medium detected in multiple libraries cve medium severity vulnerability vulnerable libraries lodash tgz lodash tgz lodash tgz lodash tgz lodash tgz lodash tgz lodash tgz lodash tgz lodash tgz lodash tgz lodash tgz a utility library delivering consistency customization performance extras library home page a href path to dependency file npm and yarn spec fixtures projects multiple sources package json path to vulnerable library npm and yarn spec fixtures projects multiple sources node modules nock node modules lodash package json npm and yarn spec fixtures projects duplicate node modules nock node modules lodash package json npm and yarn spec fixtures projects duplicate identical node modules nock node modules lodash package json npm and yarn spec fixtures projects minor version specified node modules nock node modules lodash package json dependency hierarchy fetch factory tgz root library nock tgz x lodash tgz vulnerable library lodash tgz lodash modular utilities library home page a href path to dependency file npm and yarn spec fixtures projects yarn lockfile only change package json path to vulnerable library npm and yarn spec fixtures projects yarn lockfile only change node modules lodash npm and yarn spec fixtures projects yarn no lockfile change node modules lodash npm and yarn spec fixtures projects yarn typedoc plugin ui router node modules lodash dependency hierarchy babel register tgz root library x lodash tgz vulnerable library lodash tgz lodash modular utilities library home page a href path to dependency file npm and yarn spec fixtures projects yarn duplicate indirect dependency package json path to vulnerable library npm and yarn spec fixtures projects yarn duplicate indirect dependency node modules lodash dependency hierarchy graphql cli tgz root library graphql cli prepare tgz x lodash tgz vulnerable library lodash tgz a utility library delivering consistency customization performance and extras library home page a href path to dependency file npm and yarn spec fixtures projects git dependency package json path to vulnerable library npm and yarn spec fixtures projects git dependency node modules lodash package json npm and yarn spec fixtures projects git dependency node modules lodash package json npm and yarn spec fixtures projects git dependency version node modules lodash package json npm and yarn spec fixtures projects yarn other package node modules lodash npm and yarn spec fixtures projects yarn workspaces bad node modules lodash npm and yarn spec fixtures projects yarn git dependency from subdep node modules lodash npm and yarn spec fixtures projects yarn git dependency with auth node modules lodash npm and yarn spec fixtures projects git dependency no lockfile node modules lodash package json npm and yarn spec fixtures projects yarn git dependency node modules lodash npm and yarn spec fixtures projects yarn workspaces node modules lodash npm and yarn spec fixtures projects yarn symlinked dependency node modules lodash npm and yarn spec fixtures projects and yarn git dependency node modules lodash package json npm and yarn spec fixtures projects git dependency branch version node modules lodash package json npm and yarn spec fixtures projects yarn git dependency outdated req node modules lodash npm and yarn spec fixtures projects git dependency git url node modules lodash package json npm and yarn spec fixtures projects and yarn git dependency git url node modules lodash package json npm and yarn spec fixtures projects yarn path dependency node modules lodash npm and yarn spec fixtures projects git dependency version node modules lodash package json dependency hierarchy x lodash tgz vulnerable library lodash tgz a utility library delivering consistency customization performance and extras library home page a href path to dependency file npm and yarn spec fixtures updated projects workspaces dev package json path to vulnerable library npm and yarn spec fixtures updated projects workspaces dev node modules lodash package json npm and yarn spec fixtures projects workspaces node modules lodash package json npm and yarn spec fixtures projects yarn workspaces node modules lodash npm and yarn spec fixtures projects yarn workspaces bad node modules lodash dependency hierarchy x lodash tgz vulnerable library lodash tgz the modern build of lodash modular utilities library home page a href path to dependency file npm and yarn spec fixtures projects yarn lockfile only change package json path to vulnerable library npm and yarn spec fixtures projects yarn lockfile only change node modules lodash dependency hierarchy git contributors tgz root library x lodash tgz vulnerable library lodash tgz a drop in replacement for underscore js that delivers up to performance improvements bug fixes and additional features library home page a href path to dependency file npm and yarn spec fixtures projects yarn resolutions package json path to vulnerable library npm and yarn spec fixtures projects yarn resolutions node modules lodash dependency hierarchy x lodash tgz vulnerable library lodash tgz a utility library delivering consistency customization performance and extras library home page a href path to dependency file npm and yarn spec fixtures projects and yarn path dependency package json path to vulnerable library npm and yarn spec fixtures projects and yarn path dependency node modules lodash package json npm and yarn spec fixtures projects path dependency node modules lodash package json dependency hierarchy x lodash tgz vulnerable library lodash tgz the modern build of lodash modular utilities library home page a href path to dependency file npm and yarn spec fixtures projects yarn resolution specified package json path to vulnerable library npm and yarn spec fixtures projects yarn resolution specified node modules lodash dependency hierarchy x lodash tgz vulnerable library lodash tgz the modern build of lodash modular utilities library home page a href path to dependency file npm and yarn spec fixtures projects minor version specified package json path to vulnerable library npm and yarn spec fixtures projects minor version specified node modules lodash package json npm and yarn spec fixtures projects yarn no lockfile change node modules lodash npm and yarn spec fixtures projects and yarn npm subdependency update node modules lodash package json npm and yarn spec fixtures projects yarn yanked version node modules lodash npm and yarn spec fixtures projects git sub dep invalid from node modules lodash package json npm and yarn spec fixtures projects monorepo dep multiple no lockfile node modules lodash package json npm and yarn spec fixtures projects subdependency update node modules lodash package json npm and yarn spec fixtures projects tarball bug node modules lodash package json npm and yarn spec fixtures projects and yarn npm subdependency update node modules lodash package json npm and yarn spec fixtures projects multiple sources node modules lodash package json npm and yarn spec fixtures projects yarn aliased dependency name node modules lodash npm and yarn spec fixtures projects monorepo dep multiple node modules lodash package json npm and yarn spec fixtures projects duplicate identical node modules lodash package json npm and yarn spec fixtures projects yarn optional dependencies node modules lodash npm and yarn spec fixtures projects yarn yarnrc npm registry node modules lodash npm and yarn spec fixtures projects app no version node modules lodash package json npm and yarn spec fixtures projects subdependency update node modules lodash package json npm and yarn spec fixtures projects tarball bug node modules lodash package json npm and yarn spec fixtures projects yarn dist tag node modules lodash npm and yarn spec fixtures projects subdependency update node modules lodash package json npm and yarn spec fixtures projects peer dependency mismatch no lockfile node modules lodash package json npm and yarn spec fixtures projects git sub dep invalid node modules lodash package json npm and yarn spec fixtures projects duplicate node modules lodash package json npm and yarn spec fixtures projects optional dependencies node modules lodash package json npm and yarn spec fixtures projects monorepo dep single no lockfile node modules lodash package json npm and yarn spec fixtures projects yarn aliased dependency node modules lodash dependency hierarchy npm tgz root library cli tgz x lodash tgz vulnerable library found in head commit a href found in base branch main vulnerability details a prototype pollution vulnerability was found in lodash where the functions merge mergewith and defaultsdeep can be tricked into adding or modifying properties of object prototype publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution lodash direct dependency fix resolution babel register fix resolution lodash direct dependency fix resolution graphql cli alpha fix resolution lodash direct dependency fix resolution npm rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree fetch factory nock lodash isminimumfixversionavailable true minimumfixversion isbinary false packagetype javascript node js packagename babel register packageversion packagefilepaths istransitivedependency false dependencytree babel register isminimumfixversionavailable true minimumfixversion isbinary false packagetype javascript node js packagename graphql cli packageversion packagefilepaths istransitivedependency false dependencytree graphql cli isminimumfixversionavailable true minimumfixversion alpha isbinary false packagetype javascript node js packagename lodash packageversion packagefilepaths istransitivedependency false dependencytree lodash isminimumfixversionavailable true minimumfixversion isbinary false packagetype javascript node js packagename lodash packageversion packagefilepaths istransitivedependency false dependencytree lodash isminimumfixversionavailable true minimumfixversion isbinary false packagetype javascript node js packagename lodash packageversion packagefilepaths istransitivedependency true dependencytree git contributors lodash isminimumfixversionavailable true minimumfixversion isbinary false packagetype javascript node js packagename lodash packageversion packagefilepaths istransitivedependency false dependencytree lodash isminimumfixversionavailable true minimumfixversion isbinary false packagetype javascript node js packagename lodash packageversion packagefilepaths istransitivedependency false dependencytree lodash isminimumfixversionavailable true minimumfixversion isbinary false packagetype javascript node js packagename lodash packageversion packagefilepaths istransitivedependency false dependencytree lodash isminimumfixversionavailable true minimumfixversion isbinary false packagetype javascript node js packagename npm packageversion packagefilepaths istransitivedependency false dependencytree npm isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails a prototype pollution vulnerability was found in lodash where the functions merge mergewith and defaultsdeep can be tricked into adding or modifying properties of object prototype vulnerabilityurl
0
168,073
20,741,163,772
IssuesEvent
2022-03-14 17:49:38
opensearch-project/OpenSearch-Dashboards
https://api.github.com/repos/opensearch-project/OpenSearch-Dashboards
closed
CVE-2021-3807 (High) detected in ansi-regex-3.0.0.tgz, ansi-regex-4.1.0.tgz
v2.0.0 security vulnerability high severity cve
## CVE-2021-3807 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>ansi-regex-3.0.0.tgz</b>, <b>ansi-regex-4.1.0.tgz</b></p></summary> <p> <details><summary><b>ansi-regex-3.0.0.tgz</b></p></summary> <p>Regular expression for matching ANSI escape codes</p> <p>Library home page: <a href="https://registry.npmjs.org/ansi-regex/-/ansi-regex-3.0.0.tgz">https://registry.npmjs.org/ansi-regex/-/ansi-regex-3.0.0.tgz</a></p> <p> Dependency Hierarchy: - has-ansi-3.0.0.tgz (Root Library) - :x: **ansi-regex-3.0.0.tgz** (Vulnerable Library) </details> <details><summary><b>ansi-regex-4.1.0.tgz</b></p></summary> <p>Regular expression for matching ANSI escape codes</p> <p>Library home page: <a href="https://registry.npmjs.org/ansi-regex/-/ansi-regex-4.1.0.tgz">https://registry.npmjs.org/ansi-regex/-/ansi-regex-4.1.0.tgz</a></p> <p> Dependency Hierarchy: - github-checks-reporter-0.0.20-b3.tgz (Root Library) - strip-ansi-5.2.0.tgz - :x: **ansi-regex-4.1.0.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/opensearch-project/OpenSearch-Dashboards/commit/4fd064970b66ce555f48c22dfab6ed965d0e260a">4fd064970b66ce555f48c22dfab6ed965d0e260a</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> ansi-regex is vulnerable to Inefficient Regular Expression Complexity <p>Publish Date: 2021-09-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3807>CVE-2021-3807</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://huntr.dev/bounties/5b3cf33b-ede0-4398-9974-800876dfd994/">https://huntr.dev/bounties/5b3cf33b-ede0-4398-9974-800876dfd994/</a></p> <p>Release Date: 2021-09-17</p> <p>Fix Resolution (ansi-regex): 5.0.1</p> <p>Direct dependency fix Resolution (has-ansi): 5.0.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"has-ansi","packageVersion":"3.0.0","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"has-ansi:3.0.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"5.0.0","isBinary":true},{"packageType":"javascript/Node.js","packageName":"ansi-regex","packageVersion":"4.1.0","packageFilePaths":[],"isTransitiveDependency":true,"dependencyTree":"@elastic/github-checks-reporter:0.0.20-b3;strip-ansi:5.2.0;ansi-regex:4.1.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"ansi-regex - 5.0.1,6.0.1","isBinary":true}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2021-3807","vulnerabilityDetails":"ansi-regex is vulnerable to Inefficient Regular Expression Complexity","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3807","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2021-3807 (High) detected in ansi-regex-3.0.0.tgz, ansi-regex-4.1.0.tgz - ## CVE-2021-3807 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>ansi-regex-3.0.0.tgz</b>, <b>ansi-regex-4.1.0.tgz</b></p></summary> <p> <details><summary><b>ansi-regex-3.0.0.tgz</b></p></summary> <p>Regular expression for matching ANSI escape codes</p> <p>Library home page: <a href="https://registry.npmjs.org/ansi-regex/-/ansi-regex-3.0.0.tgz">https://registry.npmjs.org/ansi-regex/-/ansi-regex-3.0.0.tgz</a></p> <p> Dependency Hierarchy: - has-ansi-3.0.0.tgz (Root Library) - :x: **ansi-regex-3.0.0.tgz** (Vulnerable Library) </details> <details><summary><b>ansi-regex-4.1.0.tgz</b></p></summary> <p>Regular expression for matching ANSI escape codes</p> <p>Library home page: <a href="https://registry.npmjs.org/ansi-regex/-/ansi-regex-4.1.0.tgz">https://registry.npmjs.org/ansi-regex/-/ansi-regex-4.1.0.tgz</a></p> <p> Dependency Hierarchy: - github-checks-reporter-0.0.20-b3.tgz (Root Library) - strip-ansi-5.2.0.tgz - :x: **ansi-regex-4.1.0.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/opensearch-project/OpenSearch-Dashboards/commit/4fd064970b66ce555f48c22dfab6ed965d0e260a">4fd064970b66ce555f48c22dfab6ed965d0e260a</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> ansi-regex is vulnerable to Inefficient Regular Expression Complexity <p>Publish Date: 2021-09-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3807>CVE-2021-3807</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://huntr.dev/bounties/5b3cf33b-ede0-4398-9974-800876dfd994/">https://huntr.dev/bounties/5b3cf33b-ede0-4398-9974-800876dfd994/</a></p> <p>Release Date: 2021-09-17</p> <p>Fix Resolution (ansi-regex): 5.0.1</p> <p>Direct dependency fix Resolution (has-ansi): 5.0.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"has-ansi","packageVersion":"3.0.0","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"has-ansi:3.0.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"5.0.0","isBinary":true},{"packageType":"javascript/Node.js","packageName":"ansi-regex","packageVersion":"4.1.0","packageFilePaths":[],"isTransitiveDependency":true,"dependencyTree":"@elastic/github-checks-reporter:0.0.20-b3;strip-ansi:5.2.0;ansi-regex:4.1.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"ansi-regex - 5.0.1,6.0.1","isBinary":true}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2021-3807","vulnerabilityDetails":"ansi-regex is vulnerable to Inefficient Regular Expression Complexity","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3807","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_comp
cve high detected in ansi regex tgz ansi regex tgz cve high severity vulnerability vulnerable libraries ansi regex tgz ansi regex tgz ansi regex tgz regular expression for matching ansi escape codes library home page a href dependency hierarchy has ansi tgz root library x ansi regex tgz vulnerable library ansi regex tgz regular expression for matching ansi escape codes library home page a href dependency hierarchy github checks reporter tgz root library strip ansi tgz x ansi regex tgz vulnerable library found in head commit a href found in base branch main vulnerability details ansi regex is vulnerable to inefficient regular expression complexity publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution ansi regex direct dependency fix resolution has ansi isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree has ansi isminimumfixversionavailable true minimumfixversion isbinary true packagetype javascript node js packagename ansi regex packageversion packagefilepaths istransitivedependency true dependencytree elastic github checks reporter strip ansi ansi regex isminimumfixversionavailable true minimumfixversion ansi regex isbinary true basebranches vulnerabilityidentifier cve vulnerabilitydetails ansi regex is vulnerable to inefficient regular expression complexity vulnerabilityurl
0
15,029
18,869,717,035
IssuesEvent
2021-11-13 01:13:37
Craluminum2413/Craluminum-Mods
https://api.github.com/repos/Craluminum2413/Craluminum-Mods
opened
Could you add compatibility with the more molds mod
compatibility
Simple enough, I dont like grinding. here is the mod in question https://mods.vintagestory.at/moremolds
True
Could you add compatibility with the more molds mod - Simple enough, I dont like grinding. here is the mod in question https://mods.vintagestory.at/moremolds
comp
could you add compatibility with the more molds mod simple enough i dont like grinding here is the mod in question
1
9,890
11,893,565,473
IssuesEvent
2020-03-29 12:11:32
Creators-of-Create/Create
https://api.github.com/repos/Creators-of-Create/Create
closed
Minecolonies Schematicannon compat
compatibility feature
I was curious is It possible to use Schematicannon (Which I love <3) in exchange of Builder (this lazy thief) from Minecolonies. But Minecolonies schematics are in .blueprint format instead of .nbt and I have no idea how to convert them. Builder have to place last block to complete the building, so It should work.
True
Minecolonies Schematicannon compat - I was curious is It possible to use Schematicannon (Which I love <3) in exchange of Builder (this lazy thief) from Minecolonies. But Minecolonies schematics are in .blueprint format instead of .nbt and I have no idea how to convert them. Builder have to place last block to complete the building, so It should work.
comp
minecolonies schematicannon compat i was curious is it possible to use schematicannon which i love in exchange of builder this lazy thief from minecolonies but minecolonies schematics are in blueprint format instead of nbt and i have no idea how to convert them builder have to place last block to complete the building so it should work
1
15,397
19,658,760,160
IssuesEvent
2022-01-10 15:04:24
e2nIEE/pandapower
https://api.github.com/repos/e2nIEE/pandapower
closed
pandapower doesn't seem to work with Numpy 1.20 (latest)
compatibility
_First off, thank you for this amazing tool. Pandapower is great!!_ Here's my issue: After upgrading Numpy to 1.20, I can no longer run powerflow (`pp.runpp(net)`). To try to prove that the issue is with numpy, I rolled back to numpy 1.19.0, and it worked like a charm. In fact, when I run `pandapower.test.run_all_tests()`, it mentions a failure related to numpy deprecations. Opinion: It's possibly because of how datatypes are used in pandapower. For example, as mentioned in the following link, `np.int` can no longer be used; rather, you should just use `int`. [Numpy 1.20 Deprecations](https://numpy.org/devdocs/release/1.20.0-notes.html#deprecations). If I'm missing something very basic or obvious, please let me know.
True
pandapower doesn't seem to work with Numpy 1.20 (latest) - _First off, thank you for this amazing tool. Pandapower is great!!_ Here's my issue: After upgrading Numpy to 1.20, I can no longer run powerflow (`pp.runpp(net)`). To try to prove that the issue is with numpy, I rolled back to numpy 1.19.0, and it worked like a charm. In fact, when I run `pandapower.test.run_all_tests()`, it mentions a failure related to numpy deprecations. Opinion: It's possibly because of how datatypes are used in pandapower. For example, as mentioned in the following link, `np.int` can no longer be used; rather, you should just use `int`. [Numpy 1.20 Deprecations](https://numpy.org/devdocs/release/1.20.0-notes.html#deprecations). If I'm missing something very basic or obvious, please let me know.
comp
pandapower doesn t seem to work with numpy latest first off thank you for this amazing tool pandapower is great here s my issue after upgrading numpy to i can no longer run powerflow pp runpp net to try to prove that the issue is with numpy i rolled back to numpy and it worked like a charm in fact when i run pandapower test run all tests it mentions a failure related to numpy deprecations opinion it s possibly because of how datatypes are used in pandapower for example as mentioned in the following link np int can no longer be used rather you should just use int if i m missing something very basic or obvious please let me know
1
824,884
31,234,248,165
IssuesEvent
2023-08-20 04:16:40
space-wizards/RobustToolbox
https://api.github.com/repos/space-wizards/RobustToolbox
opened
Deleted / Exists cleanup
Issue: Needs Cleanup Difficulty: 3-Hard Priority: 2-Before Release
1. So Exists checks if the metadatacomp exists at all. 2. Deleted checks if it exists and the entity is not flagged as deleted. This is kinda whacky and ideally they'd be combined. Whenever an entity is deleted it should just be removed and hence no need for EntityDeleted. If / when we get command buffers then you'd just store a hashset of queuedeleted entities and also check that if you really wanted or something.
1.0
Deleted / Exists cleanup - 1. So Exists checks if the metadatacomp exists at all. 2. Deleted checks if it exists and the entity is not flagged as deleted. This is kinda whacky and ideally they'd be combined. Whenever an entity is deleted it should just be removed and hence no need for EntityDeleted. If / when we get command buffers then you'd just store a hashset of queuedeleted entities and also check that if you really wanted or something.
non_comp
deleted exists cleanup so exists checks if the metadatacomp exists at all deleted checks if it exists and the entity is not flagged as deleted this is kinda whacky and ideally they d be combined whenever an entity is deleted it should just be removed and hence no need for entitydeleted if when we get command buffers then you d just store a hashset of queuedeleted entities and also check that if you really wanted or something
0
21,600
2,641,717,230
IssuesEvent
2015-03-11 19:19:51
chrsmith/html5rocks
https://api.github.com/repos/chrsmith/html5rocks
closed
Client side storage explained
Milestone-2 Priority-Medium Tutorial Type-Enhancement
Original [issue 71](https://code.google.com/p/html5rocks/issues/detail?id=71) created by chrsmith on 2010-07-28T00:11:03.000Z: vli: cover localstorage , indexdb, web database, difference , use cases explains tools and best practices to determine when to use what.
1.0
Client side storage explained - Original [issue 71](https://code.google.com/p/html5rocks/issues/detail?id=71) created by chrsmith on 2010-07-28T00:11:03.000Z: vli: cover localstorage , indexdb, web database, difference , use cases explains tools and best practices to determine when to use what.
non_comp
client side storage explained original created by chrsmith on vli cover localstorage indexdb web database difference use cases explains tools and best practices to determine when to use what
0
548,169
16,059,245,773
IssuesEvent
2021-04-23 10:03:57
docker-mailserver/docker-mailserver
https://api.github.com/repos/docker-mailserver/docker-mailserver
closed
Settings to use service
kind/question meta/help wanted meta/needs triage priority/low
# Domain name and client settings Hello, a beginner question: how to use docker mailserver with node mailer to send mails and what are the settings to make on the domain name (redirection / MX)? Thank you!
1.0
Settings to use service - # Domain name and client settings Hello, a beginner question: how to use docker mailserver with node mailer to send mails and what are the settings to make on the domain name (redirection / MX)? Thank you!
non_comp
settings to use service domain name and client settings hello a beginner question how to use docker mailserver with node mailer to send mails and what are the settings to make on the domain name redirection mx thank you
0
175,456
21,313,548,578
IssuesEvent
2022-04-16 00:08:07
elastic/kibana
https://api.github.com/repos/elastic/kibana
closed
[Security Solution] Session view icon is overlapping and not working under event renderer on alerts page
bug triage_needed impact:high Team: SecuritySolution v8.2.0
**Describe the bug** Session view icon is overlapping and not working under event renderer on alerts page **Build Details:** ``` Version : 8.2.0 SNAPSHOT Build : 51940 Commit : cbab89d8a43311d11dfaac3d9145a2be8841a857 ``` **preconditions** 1. Alerts should be exist with session view **Steps to Reproduce** 1. Navigate to Alerts page 2. Click on Event rendered view on alert table 3. Observe that Session view icon is overlapping and not working under event renderer on alerts page **Actual Result** Session view icon is overlapping and not working under event renderer on alerts page **Expected Result** Session view icon should not be overlapped and should be working under event renderer on alerts page **Screen-Shot** ![image](https://user-images.githubusercontent.com/61860752/163563790-75bb789f-b398-47b2-a645-dd1c7cffa03c.png)
True
[Security Solution] Session view icon is overlapping and not working under event renderer on alerts page - **Describe the bug** Session view icon is overlapping and not working under event renderer on alerts page **Build Details:** ``` Version : 8.2.0 SNAPSHOT Build : 51940 Commit : cbab89d8a43311d11dfaac3d9145a2be8841a857 ``` **preconditions** 1. Alerts should be exist with session view **Steps to Reproduce** 1. Navigate to Alerts page 2. Click on Event rendered view on alert table 3. Observe that Session view icon is overlapping and not working under event renderer on alerts page **Actual Result** Session view icon is overlapping and not working under event renderer on alerts page **Expected Result** Session view icon should not be overlapped and should be working under event renderer on alerts page **Screen-Shot** ![image](https://user-images.githubusercontent.com/61860752/163563790-75bb789f-b398-47b2-a645-dd1c7cffa03c.png)
non_comp
session view icon is overlapping and not working under event renderer on alerts page describe the bug session view icon is overlapping and not working under event renderer on alerts page build details version snapshot build commit preconditions alerts should be exist with session view steps to reproduce navigate to alerts page click on event rendered view on alert table observe that session view icon is overlapping and not working under event renderer on alerts page actual result session view icon is overlapping and not working under event renderer on alerts page expected result session view icon should not be overlapped and should be working under event renderer on alerts page screen shot
0
148,492
11,854,212,360
IssuesEvent
2020-03-25 00:06:12
QubesOS/updates-status
https://api.github.com/repos/QubesOS/updates-status
closed
linux-kernel v5.4.25-1 (r4.1)
r4.1-dom0-cur-test
Update of linux-kernel to v5.4.25-1 for Qubes r4.1, see comments below for details. Built from: https://github.com/QubesOS/qubes-linux-kernel/commit/64dcb6ced3a017cc582258a893986416c604c309 [Changes since previous version](https://github.com/QubesOS/qubes-linux-kernel/compare/v4.19.100-1...v5.4.25-1): QubesOS/qubes-linux-kernel@64dcb6c Update to kernel-5.4.25 QubesOS/qubes-linux-kernel@26db5f5 Merge remote-tracking branch 'origin/pr/183' into stable-5.4 QubesOS/qubes-linux-kernel@645b1bf Makefile: remove extra tab QubesOS/qubes-linux-kernel@0bcc904 update-sources: clean version modification in case of failure QubesOS/qubes-linux-kernel@21828df Makefile: set default BRANCH variable to master QubesOS/qubes-linux-kernel@56dbda2 Add scripts for kernel updates QubesOS/qubes-linux-kernel@43854b5 gitignore: update wireguard filenames QubesOS/qubes-linux-kernel@320b3d9 gitignore: ignore pkgs QubesOS/qubes-linux-kernel@1e11574 makefile: clean unused targets QubesOS/qubes-linux-kernel@8522362 Update to kernel-5.4.24 QubesOS/qubes-linux-kernel@a0992a8 Merge remote-tracking branch 'origin/pr/179' into stable-5.4 QubesOS/qubes-linux-kernel@104eaad Update to kernel-5.4.23 QubesOS/qubes-linux-kernel@3fcf2ad Update to kernel-5.4.22 QubesOS/qubes-linux-kernel@ea8d8f4 Update to kernel-5.4.21 QubesOS/qubes-linux-kernel@d03bb55 Increase modules.img sparse size QubesOS/qubes-linux-kernel@a776d31 get-latest-config: use keys under builder-rpm QubesOS/qubes-linux-kernel@b065f33 Drop -latest suffix for LTS kernel QubesOS/qubes-linux-kernel@af7128d Update to kernel-5.4.5 QubesOS/qubes-linux-kernel@b5170f8 update WireGuard from 20191212 to 20191219 QubesOS/qubes-linux-kernel@1252f79 Update to kernel-5.4.3 QubesOS/qubes-linux-kernel@9f28598 update WireGuard from 20191127 to 20191212 QubesOS/qubes-linux-kernel@1554c07 Add BR: python3-devel for pathfix.py script QubesOS/qubes-linux-kernel@8ee5779 Update to kernel-5.4.2 QubesOS/qubes-linux-kernel@366e681 get latest config: exit with clean and disable getting rc by default QubesOS/qubes-linux-kernel@f841402 Merge remote-tracking branch 'origin/pr/128' QubesOS/qubes-linux-kernel@8aca753 (re-)sign all the modules after processing debuginfo QubesOS/qubes-linux-kernel@a5e7326 update WireGuard from 20191012 to 20191127 QubesOS/qubes-linux-kernel@5d68948 version 5.4.1-2 QubesOS/qubes-linux-kernel@1cc6951 config-qubes: allow signed modules as Fedora set it QubesOS/qubes-linux-kernel@22b6866 Update to version 5.4.1 QubesOS/qubes-linux-kernel@7b1c020 Update Fedora keys for config updater script (add 32, remove 30 and 31) QubesOS/qubes-linux-kernel@9fc4062 Merge remote-tracking branch 'origin/pr/119' QubesOS/qubes-linux-kernel@fb9b485 spec: mangle /usr/bin/python shebangs to /usr/bin/python3 QubesOS/qubes-linux-kernel@ddf7602 Update to kernel-5.3.12 QubesOS/qubes-linux-kernel@18d5d66 Update to kernel-5.3.11 QubesOS/qubes-linux-kernel@ea53bfe Merge remote-tracking branch 'origin/pr/113' QubesOS/qubes-linux-kernel@71964cd Update to kernel-5.3.9 QubesOS/qubes-linux-kernel@cddf334 update WireGuard from 20190913 to 20191012 QubesOS/qubes-linux-kernel@beccd92 Update to kernel-5.3.8 QubesOS/qubes-linux-kernel@17b7186 Make makefile and spec compatible with rc versions QubesOS/qubes-linux-kernel@24ee349 Add Linus's signing key QubesOS/qubes-linux-kernel@87ba67c Update to kernel-5.3.7 QubesOS/qubes-linux-kernel@9cf5108 Update to kernel-5.3.6 QubesOS/qubes-linux-kernel@8f82b34 Update to kernel-5.3.2 QubesOS/qubes-linux-kernel@383118a change WireGuard version from 20190702 to 20190913 QubesOS/qubes-linux-kernel@9a3ca00 Update to kernel-5.2.16 QubesOS/qubes-linux-kernel@c7cfdfe Merge remote-tracking branch 'origin/pr/91' QubesOS/qubes-linux-kernel@1d50db3 rpm: set default kernel version based on package flavor QubesOS/qubes-linux-kernel@9aac829 Update to kernel-5.2.14 QubesOS/qubes-linux-kernel@4e4cdf6 Update to kernel-5.2.13 QubesOS/qubes-linux-kernel@61485da Update to kernel-5.2.11 QubesOS/qubes-linux-kernel@c270f52 Update to kernel-5.2.9 QubesOS/qubes-linux-kernel@8b2aed9 version 5.2.7-1 QubesOS/qubes-linux-kernel@d9f5315 Drop AFTER_LINK patch QubesOS/qubes-linux-kernel@150288e Update to kernel-5.2.3 QubesOS/qubes-linux-kernel@9ec4071 Update WireGuard to 0.0.20190702 QubesOS/qubes-linux-kernel@a40377d Update to kernel-5.1.17 QubesOS/qubes-linux-kernel@bd26af6 Update to kernel-5.1.16 QubesOS/qubes-linux-kernel@11d6f25 Update to kernel-5.1.15 QubesOS/qubes-linux-kernel@1574118 Update to kernel-5.1.13 QubesOS/qubes-linux-kernel@5adbe65 Update to kernel-5.1.9 QubesOS/qubes-linux-kernel@4f870c8 version 5.1.7-2 QubesOS/qubes-linux-kernel@d48c5ba Update to kernel-5.1.7 QubesOS/qubes-linux-kernel@e3a3420 Disable GCC plugins for out of tree kernel modules QubesOS/qubes-linux-kernel@d1ad508 Update to kernel-5.1.6 QubesOS/qubes-linux-kernel@7108c83 Simplify WG_SIG_FILE set QubesOS/qubes-linux-kernel@c6e1265 gitignore: add WireGuard QubesOS/qubes-linux-kernel@7906851 build wireguard module (after gpg check) QubesOS/qubes-linux-kernel@ef094d5 version 5.1.2-1 QubesOS/qubes-linux-kernel@c4c75cb Fix setting default kernel for VM QubesOS/qubes-linux-kernel@98cd4d1 config: disable SELinux QubesOS/qubes-linux-kernel@c68ee34 Update to kernel-5.1.1 QubesOS/qubes-linux-kernel@732780e get-fedora-latest-config: sort result when repoquerying kernel-core QubesOS/qubes-linux-kernel@d08d1bf Update to kernel-5.0.11 QubesOS/qubes-linux-kernel@598d75d Update to kernel-5.0.9 QubesOS/qubes-linux-kernel@a48e051 get-fedora-latest-config: enable only fedora and fedora-updates when downloading the rpm QubesOS/qubes-linux-kernel@5a1a029 get-fedora-latest-config: download relatively to current directory QubesOS/qubes-linux-kernel@a5a4ae6 Support a build without u2mfn module QubesOS/qubes-linux-kernel@1ffeed9 Update to kernel-5.0.7 QubesOS/qubes-linux-kernel@62c107b Update to kernel-5.0.5 QubesOS/qubes-linux-kernel@b16cc10 get-fedora-latest-config: enable only Fedora repos QubesOS/qubes-linux-kernel@f2ddd7a Refactor: handle getting default latest stable Fedora QubesOS/qubes-linux-kernel@9290c06 Add Fedora 31 primary key QubesOS/qubes-linux-kernel@cfc9a4c version 5.0.2-1 QubesOS/qubes-linux-kernel@17bbb7d Merge remote-tracking branch 'origin/pr/28' QubesOS/qubes-linux-kernel@1fe8359 rpm: one more typo fix QubesOS/qubes-linux-kernel@047a30a version 4.20.16-2 QubesOS/qubes-linux-kernel@50b7baa rpm: typo fix QubesOS/qubes-linux-kernel@fd14ac9 rpm: build modules.img at package build time only on new system QubesOS/qubes-linux-kernel@1bedd11 Update patches QubesOS/qubes-linux-kernel@0643a07 Update config with respect to upstream QubesOS/qubes-linux-kernel@b935b16 get-fedora-latest-config: allow fetching config from rawhide repository QubesOS/qubes-linux-kernel@924b8ab get-fedora-latest-config: don't hardcode Fedora version in regex rpm QubesOS/qubes-linux-kernel@fe376be keys: add Fedora 30 primary key QubesOS/qubes-linux-kernel@80fa61e get-sources: handle kernel-5.x QubesOS/qubes-linux-kernel@00bbcc6 version 4.20.16-1 QubesOS/qubes-linux-kernel@c2a2c69 Merge remote-tracking branch 'origin/pr/27' QubesOS/qubes-linux-kernel@61b9336 plymouth: ignore serial console hvc0 in UEFI QubesOS/qubes-linux-kernel@a44160b Ensure the rebuild of grub.cfg with plymouth.ignore-serial-consoles QubesOS/qubes-linux-kernel@e69b02b plymouth-ignore-serial-consoles: adjusting the method from marmarek suggestion QubesOS/qubes-linux-kernel@af67412 plymouth: ignore serial console hvc0 QubesOS/qubes-linux-kernel@c8f70c1 Update to kernel-4.20.14 QubesOS/qubes-linux-kernel@9cfa9a9 Include default-kernelopts-common.txt with kernel-specific default options QubesOS/qubes-linux-kernel@43235dd Improve reproducibility of initramfs QubesOS/qubes-linux-kernel@eb4c9ed Build modules.img at module build time QubesOS/qubes-linux-kernel@c2f7109 version 4.20.3-1 QubesOS/qubes-linux-kernel@c168934 Update to kernel-4.20.3 and config from Fedora's config 4.19.15-300 Referenced issues: QubesOS/qubes-issues#4736 QubesOS/qubes-issues#4839 QubesOS/qubes-issues#3849 QubesOS/qubes-issues#4280 QubesOS/qubes-issues#5497 QubesOS/qubes-issues#2844 QubesOS/qubes-issues#5309 If you're release manager, you can issue GPG-inline signed command: * `Upload linux-kernel 64dcb6ced3a017cc582258a893986416c604c309 r4.1 current repo` (available 7 days from now) * `Upload linux-kernel 64dcb6ced3a017cc582258a893986416c604c309 r4.1 current (dists) repo`, you can choose subset of distributions, like `vm-fc24 vm-fc25` (available 7 days from now) * `Upload linux-kernel 64dcb6ced3a017cc582258a893986416c604c309 r4.1 security-testing repo` Above commands will work only if packages in current-testing repository were built from given commit (i.e. no new version superseded it).
1.0
linux-kernel v5.4.25-1 (r4.1) - Update of linux-kernel to v5.4.25-1 for Qubes r4.1, see comments below for details. Built from: https://github.com/QubesOS/qubes-linux-kernel/commit/64dcb6ced3a017cc582258a893986416c604c309 [Changes since previous version](https://github.com/QubesOS/qubes-linux-kernel/compare/v4.19.100-1...v5.4.25-1): QubesOS/qubes-linux-kernel@64dcb6c Update to kernel-5.4.25 QubesOS/qubes-linux-kernel@26db5f5 Merge remote-tracking branch 'origin/pr/183' into stable-5.4 QubesOS/qubes-linux-kernel@645b1bf Makefile: remove extra tab QubesOS/qubes-linux-kernel@0bcc904 update-sources: clean version modification in case of failure QubesOS/qubes-linux-kernel@21828df Makefile: set default BRANCH variable to master QubesOS/qubes-linux-kernel@56dbda2 Add scripts for kernel updates QubesOS/qubes-linux-kernel@43854b5 gitignore: update wireguard filenames QubesOS/qubes-linux-kernel@320b3d9 gitignore: ignore pkgs QubesOS/qubes-linux-kernel@1e11574 makefile: clean unused targets QubesOS/qubes-linux-kernel@8522362 Update to kernel-5.4.24 QubesOS/qubes-linux-kernel@a0992a8 Merge remote-tracking branch 'origin/pr/179' into stable-5.4 QubesOS/qubes-linux-kernel@104eaad Update to kernel-5.4.23 QubesOS/qubes-linux-kernel@3fcf2ad Update to kernel-5.4.22 QubesOS/qubes-linux-kernel@ea8d8f4 Update to kernel-5.4.21 QubesOS/qubes-linux-kernel@d03bb55 Increase modules.img sparse size QubesOS/qubes-linux-kernel@a776d31 get-latest-config: use keys under builder-rpm QubesOS/qubes-linux-kernel@b065f33 Drop -latest suffix for LTS kernel QubesOS/qubes-linux-kernel@af7128d Update to kernel-5.4.5 QubesOS/qubes-linux-kernel@b5170f8 update WireGuard from 20191212 to 20191219 QubesOS/qubes-linux-kernel@1252f79 Update to kernel-5.4.3 QubesOS/qubes-linux-kernel@9f28598 update WireGuard from 20191127 to 20191212 QubesOS/qubes-linux-kernel@1554c07 Add BR: python3-devel for pathfix.py script QubesOS/qubes-linux-kernel@8ee5779 Update to kernel-5.4.2 QubesOS/qubes-linux-kernel@366e681 get latest config: exit with clean and disable getting rc by default QubesOS/qubes-linux-kernel@f841402 Merge remote-tracking branch 'origin/pr/128' QubesOS/qubes-linux-kernel@8aca753 (re-)sign all the modules after processing debuginfo QubesOS/qubes-linux-kernel@a5e7326 update WireGuard from 20191012 to 20191127 QubesOS/qubes-linux-kernel@5d68948 version 5.4.1-2 QubesOS/qubes-linux-kernel@1cc6951 config-qubes: allow signed modules as Fedora set it QubesOS/qubes-linux-kernel@22b6866 Update to version 5.4.1 QubesOS/qubes-linux-kernel@7b1c020 Update Fedora keys for config updater script (add 32, remove 30 and 31) QubesOS/qubes-linux-kernel@9fc4062 Merge remote-tracking branch 'origin/pr/119' QubesOS/qubes-linux-kernel@fb9b485 spec: mangle /usr/bin/python shebangs to /usr/bin/python3 QubesOS/qubes-linux-kernel@ddf7602 Update to kernel-5.3.12 QubesOS/qubes-linux-kernel@18d5d66 Update to kernel-5.3.11 QubesOS/qubes-linux-kernel@ea53bfe Merge remote-tracking branch 'origin/pr/113' QubesOS/qubes-linux-kernel@71964cd Update to kernel-5.3.9 QubesOS/qubes-linux-kernel@cddf334 update WireGuard from 20190913 to 20191012 QubesOS/qubes-linux-kernel@beccd92 Update to kernel-5.3.8 QubesOS/qubes-linux-kernel@17b7186 Make makefile and spec compatible with rc versions QubesOS/qubes-linux-kernel@24ee349 Add Linus's signing key QubesOS/qubes-linux-kernel@87ba67c Update to kernel-5.3.7 QubesOS/qubes-linux-kernel@9cf5108 Update to kernel-5.3.6 QubesOS/qubes-linux-kernel@8f82b34 Update to kernel-5.3.2 QubesOS/qubes-linux-kernel@383118a change WireGuard version from 20190702 to 20190913 QubesOS/qubes-linux-kernel@9a3ca00 Update to kernel-5.2.16 QubesOS/qubes-linux-kernel@c7cfdfe Merge remote-tracking branch 'origin/pr/91' QubesOS/qubes-linux-kernel@1d50db3 rpm: set default kernel version based on package flavor QubesOS/qubes-linux-kernel@9aac829 Update to kernel-5.2.14 QubesOS/qubes-linux-kernel@4e4cdf6 Update to kernel-5.2.13 QubesOS/qubes-linux-kernel@61485da Update to kernel-5.2.11 QubesOS/qubes-linux-kernel@c270f52 Update to kernel-5.2.9 QubesOS/qubes-linux-kernel@8b2aed9 version 5.2.7-1 QubesOS/qubes-linux-kernel@d9f5315 Drop AFTER_LINK patch QubesOS/qubes-linux-kernel@150288e Update to kernel-5.2.3 QubesOS/qubes-linux-kernel@9ec4071 Update WireGuard to 0.0.20190702 QubesOS/qubes-linux-kernel@a40377d Update to kernel-5.1.17 QubesOS/qubes-linux-kernel@bd26af6 Update to kernel-5.1.16 QubesOS/qubes-linux-kernel@11d6f25 Update to kernel-5.1.15 QubesOS/qubes-linux-kernel@1574118 Update to kernel-5.1.13 QubesOS/qubes-linux-kernel@5adbe65 Update to kernel-5.1.9 QubesOS/qubes-linux-kernel@4f870c8 version 5.1.7-2 QubesOS/qubes-linux-kernel@d48c5ba Update to kernel-5.1.7 QubesOS/qubes-linux-kernel@e3a3420 Disable GCC plugins for out of tree kernel modules QubesOS/qubes-linux-kernel@d1ad508 Update to kernel-5.1.6 QubesOS/qubes-linux-kernel@7108c83 Simplify WG_SIG_FILE set QubesOS/qubes-linux-kernel@c6e1265 gitignore: add WireGuard QubesOS/qubes-linux-kernel@7906851 build wireguard module (after gpg check) QubesOS/qubes-linux-kernel@ef094d5 version 5.1.2-1 QubesOS/qubes-linux-kernel@c4c75cb Fix setting default kernel for VM QubesOS/qubes-linux-kernel@98cd4d1 config: disable SELinux QubesOS/qubes-linux-kernel@c68ee34 Update to kernel-5.1.1 QubesOS/qubes-linux-kernel@732780e get-fedora-latest-config: sort result when repoquerying kernel-core QubesOS/qubes-linux-kernel@d08d1bf Update to kernel-5.0.11 QubesOS/qubes-linux-kernel@598d75d Update to kernel-5.0.9 QubesOS/qubes-linux-kernel@a48e051 get-fedora-latest-config: enable only fedora and fedora-updates when downloading the rpm QubesOS/qubes-linux-kernel@5a1a029 get-fedora-latest-config: download relatively to current directory QubesOS/qubes-linux-kernel@a5a4ae6 Support a build without u2mfn module QubesOS/qubes-linux-kernel@1ffeed9 Update to kernel-5.0.7 QubesOS/qubes-linux-kernel@62c107b Update to kernel-5.0.5 QubesOS/qubes-linux-kernel@b16cc10 get-fedora-latest-config: enable only Fedora repos QubesOS/qubes-linux-kernel@f2ddd7a Refactor: handle getting default latest stable Fedora QubesOS/qubes-linux-kernel@9290c06 Add Fedora 31 primary key QubesOS/qubes-linux-kernel@cfc9a4c version 5.0.2-1 QubesOS/qubes-linux-kernel@17bbb7d Merge remote-tracking branch 'origin/pr/28' QubesOS/qubes-linux-kernel@1fe8359 rpm: one more typo fix QubesOS/qubes-linux-kernel@047a30a version 4.20.16-2 QubesOS/qubes-linux-kernel@50b7baa rpm: typo fix QubesOS/qubes-linux-kernel@fd14ac9 rpm: build modules.img at package build time only on new system QubesOS/qubes-linux-kernel@1bedd11 Update patches QubesOS/qubes-linux-kernel@0643a07 Update config with respect to upstream QubesOS/qubes-linux-kernel@b935b16 get-fedora-latest-config: allow fetching config from rawhide repository QubesOS/qubes-linux-kernel@924b8ab get-fedora-latest-config: don't hardcode Fedora version in regex rpm QubesOS/qubes-linux-kernel@fe376be keys: add Fedora 30 primary key QubesOS/qubes-linux-kernel@80fa61e get-sources: handle kernel-5.x QubesOS/qubes-linux-kernel@00bbcc6 version 4.20.16-1 QubesOS/qubes-linux-kernel@c2a2c69 Merge remote-tracking branch 'origin/pr/27' QubesOS/qubes-linux-kernel@61b9336 plymouth: ignore serial console hvc0 in UEFI QubesOS/qubes-linux-kernel@a44160b Ensure the rebuild of grub.cfg with plymouth.ignore-serial-consoles QubesOS/qubes-linux-kernel@e69b02b plymouth-ignore-serial-consoles: adjusting the method from marmarek suggestion QubesOS/qubes-linux-kernel@af67412 plymouth: ignore serial console hvc0 QubesOS/qubes-linux-kernel@c8f70c1 Update to kernel-4.20.14 QubesOS/qubes-linux-kernel@9cfa9a9 Include default-kernelopts-common.txt with kernel-specific default options QubesOS/qubes-linux-kernel@43235dd Improve reproducibility of initramfs QubesOS/qubes-linux-kernel@eb4c9ed Build modules.img at module build time QubesOS/qubes-linux-kernel@c2f7109 version 4.20.3-1 QubesOS/qubes-linux-kernel@c168934 Update to kernel-4.20.3 and config from Fedora's config 4.19.15-300 Referenced issues: QubesOS/qubes-issues#4736 QubesOS/qubes-issues#4839 QubesOS/qubes-issues#3849 QubesOS/qubes-issues#4280 QubesOS/qubes-issues#5497 QubesOS/qubes-issues#2844 QubesOS/qubes-issues#5309 If you're release manager, you can issue GPG-inline signed command: * `Upload linux-kernel 64dcb6ced3a017cc582258a893986416c604c309 r4.1 current repo` (available 7 days from now) * `Upload linux-kernel 64dcb6ced3a017cc582258a893986416c604c309 r4.1 current (dists) repo`, you can choose subset of distributions, like `vm-fc24 vm-fc25` (available 7 days from now) * `Upload linux-kernel 64dcb6ced3a017cc582258a893986416c604c309 r4.1 security-testing repo` Above commands will work only if packages in current-testing repository were built from given commit (i.e. no new version superseded it).
non_comp
linux kernel update of linux kernel to for qubes see comments below for details built from qubesos qubes linux kernel update to kernel qubesos qubes linux kernel merge remote tracking branch origin pr into stable qubesos qubes linux kernel makefile remove extra tab qubesos qubes linux kernel update sources clean version modification in case of failure qubesos qubes linux kernel makefile set default branch variable to master qubesos qubes linux kernel add scripts for kernel updates qubesos qubes linux kernel gitignore update wireguard filenames qubesos qubes linux kernel gitignore ignore pkgs qubesos qubes linux kernel makefile clean unused targets qubesos qubes linux kernel update to kernel qubesos qubes linux kernel merge remote tracking branch origin pr into stable qubesos qubes linux kernel update to kernel qubesos qubes linux kernel update to kernel qubesos qubes linux kernel update to kernel qubesos qubes linux kernel increase modules img sparse size qubesos qubes linux kernel get latest config use keys under builder rpm qubesos qubes linux kernel drop latest suffix for lts kernel qubesos qubes linux kernel update to kernel qubesos qubes linux kernel update wireguard from to qubesos qubes linux kernel update to kernel qubesos qubes linux kernel update wireguard from to qubesos qubes linux kernel add br devel for pathfix py script qubesos qubes linux kernel update to kernel qubesos qubes linux kernel get latest config exit with clean and disable getting rc by default qubesos qubes linux kernel merge remote tracking branch origin pr qubesos qubes linux kernel re sign all the modules after processing debuginfo qubesos qubes linux kernel update wireguard from to qubesos qubes linux kernel version qubesos qubes linux kernel config qubes allow signed modules as fedora set it qubesos qubes linux kernel update to version qubesos qubes linux kernel update fedora keys for config updater script add remove and qubesos qubes linux kernel merge remote tracking branch origin pr qubesos qubes linux kernel spec mangle usr bin python shebangs to usr bin qubesos qubes linux kernel update to kernel qubesos qubes linux kernel update to kernel qubesos qubes linux kernel merge remote tracking branch origin pr qubesos qubes linux kernel update to kernel qubesos qubes linux kernel update wireguard from to qubesos qubes linux kernel update to kernel qubesos qubes linux kernel make makefile and spec compatible with rc versions qubesos qubes linux kernel add linus s signing key qubesos qubes linux kernel update to kernel qubesos qubes linux kernel update to kernel qubesos qubes linux kernel update to kernel qubesos qubes linux kernel change wireguard version from to qubesos qubes linux kernel update to kernel qubesos qubes linux kernel merge remote tracking branch origin pr qubesos qubes linux kernel rpm set default kernel version based on package flavor qubesos qubes linux kernel update to kernel qubesos qubes linux kernel update to kernel qubesos qubes linux kernel update to kernel qubesos qubes linux kernel update to kernel qubesos qubes linux kernel version qubesos qubes linux kernel drop after link patch qubesos qubes linux kernel update to kernel qubesos qubes linux kernel update wireguard to qubesos qubes linux kernel update to kernel qubesos qubes linux kernel update to kernel qubesos qubes linux kernel update to kernel qubesos qubes linux kernel update to kernel qubesos qubes linux kernel update to kernel qubesos qubes linux kernel version qubesos qubes linux kernel update to kernel qubesos qubes linux kernel disable gcc plugins for out of tree kernel modules qubesos qubes linux kernel update to kernel qubesos qubes linux kernel simplify wg sig file set qubesos qubes linux kernel gitignore add wireguard qubesos qubes linux kernel build wireguard module after gpg check qubesos qubes linux kernel version qubesos qubes linux kernel fix setting default kernel for vm qubesos qubes linux kernel config disable selinux qubesos qubes linux kernel update to kernel qubesos qubes linux kernel get fedora latest config sort result when repoquerying kernel core qubesos qubes linux kernel update to kernel qubesos qubes linux kernel update to kernel qubesos qubes linux kernel get fedora latest config enable only fedora and fedora updates when downloading the rpm qubesos qubes linux kernel get fedora latest config download relatively to current directory qubesos qubes linux kernel support a build without module qubesos qubes linux kernel update to kernel qubesos qubes linux kernel update to kernel qubesos qubes linux kernel get fedora latest config enable only fedora repos qubesos qubes linux kernel refactor handle getting default latest stable fedora qubesos qubes linux kernel add fedora primary key qubesos qubes linux kernel version qubesos qubes linux kernel merge remote tracking branch origin pr qubesos qubes linux kernel rpm one more typo fix qubesos qubes linux kernel version qubesos qubes linux kernel rpm typo fix qubesos qubes linux kernel rpm build modules img at package build time only on new system qubesos qubes linux kernel update patches qubesos qubes linux kernel update config with respect to upstream qubesos qubes linux kernel get fedora latest config allow fetching config from rawhide repository qubesos qubes linux kernel get fedora latest config don t hardcode fedora version in regex rpm qubesos qubes linux kernel keys add fedora primary key qubesos qubes linux kernel get sources handle kernel x qubesos qubes linux kernel version qubesos qubes linux kernel merge remote tracking branch origin pr qubesos qubes linux kernel plymouth ignore serial console in uefi qubesos qubes linux kernel ensure the rebuild of grub cfg with plymouth ignore serial consoles qubesos qubes linux kernel plymouth ignore serial consoles adjusting the method from marmarek suggestion qubesos qubes linux kernel plymouth ignore serial console qubesos qubes linux kernel update to kernel qubesos qubes linux kernel include default kernelopts common txt with kernel specific default options qubesos qubes linux kernel improve reproducibility of initramfs qubesos qubes linux kernel build modules img at module build time qubesos qubes linux kernel version qubesos qubes linux kernel update to kernel and config from fedora s config referenced issues qubesos qubes issues qubesos qubes issues qubesos qubes issues qubesos qubes issues qubesos qubes issues qubesos qubes issues qubesos qubes issues if you re release manager you can issue gpg inline signed command upload linux kernel current repo available days from now upload linux kernel current dists repo you can choose subset of distributions like vm vm available days from now upload linux kernel security testing repo above commands will work only if packages in current testing repository were built from given commit i e no new version superseded it
0
630
3,059,998,355
IssuesEvent
2015-08-14 18:07:06
facebook/hhvm
https://api.github.com/repos/facebook/hhvm
closed
Segmentation fault if memcache host doesn't resolve
crash php5 incompatibility
HHVM will crash after the object is reused or free'd after a Memcache::add call where the server being used has a hostname that doesn't resolve. For example: ```php <?php $memcache = new Memcache(); $memcache->connect( "junkhost", 123 ); $memcache->add( "foo", "bar", 0, 5 ); echo $memcache->get( "foo" ) . "\n"; // crashes echo "Done\n"; ?> ``` ``` Name: /usr/bin/hhvm Type: Segmentation fault Runtime: hhvm Version: tags/HHVM-3.4.0-0-g817b3a07fc4e509ce15635dbc87778e5b3496663 DebuggerCount: 0 Arguments: hhvm_memcache_bug_test.php ThreadType: CLI # 0 ?? at hhvm:0 # 1 free at /tmp/tmp.1cIyH3Bll4/jemalloc-3.6.0/src/jemalloc.c:2109 # 2 freeaddrinfo at /lib/x86_64-linux-gnu/libc.so.6:0 # 3 memcached_behavior_get_distribution_hash at /usr/lib/libmemcached.so.6:0 # 4 memcached_flush_buffers at /usr/lib/libmemcached.so.6:0 # 5 memcached_get_by_key at /usr/lib/libmemcached.so.6:0 # 6 memcached_get at /usr/lib/libmemcached.so.6:0 # 7 ?? at hhvm:0 # 8 void HPHP::Native::callFunc<false, false>(HPHP::Func const*, void*, HPHP::TypedValue*, HPHP::TypedValue&) at hhvm:0 # 9 HPHP::TypedValue* HPHP::Native::methodWrapper<false, false>(HPHP::ActRec*) at hhvm:0 ``` PHP, for comparison, emits the following, but does not crash: ``` PHP Warning: Memcache::connect(): php_network_getaddresses: getaddrinfo failed: Temporary failure in name resolution in /home/steam/hhvm_memcache_bug_test.php on line 4 PHP Notice: Memcache::connect(): Server junkhost (tcp 123, udp 0) failed with: php_network_getaddresses: getaddrinfo failed: Temporary failure in name resolution (0) in /home/steam/hhvm_memcache_bug_test.php on line 4 PHP Warning: Memcache::connect(): Can't connect to junkhost:123, php_network_getaddresses: getaddrinfo failed: Temporary failure in name resolution (0) in /home/steam/hhvm_memcache_bug_test.php on line 4 ```
True
Segmentation fault if memcache host doesn't resolve - HHVM will crash after the object is reused or free'd after a Memcache::add call where the server being used has a hostname that doesn't resolve. For example: ```php <?php $memcache = new Memcache(); $memcache->connect( "junkhost", 123 ); $memcache->add( "foo", "bar", 0, 5 ); echo $memcache->get( "foo" ) . "\n"; // crashes echo "Done\n"; ?> ``` ``` Name: /usr/bin/hhvm Type: Segmentation fault Runtime: hhvm Version: tags/HHVM-3.4.0-0-g817b3a07fc4e509ce15635dbc87778e5b3496663 DebuggerCount: 0 Arguments: hhvm_memcache_bug_test.php ThreadType: CLI # 0 ?? at hhvm:0 # 1 free at /tmp/tmp.1cIyH3Bll4/jemalloc-3.6.0/src/jemalloc.c:2109 # 2 freeaddrinfo at /lib/x86_64-linux-gnu/libc.so.6:0 # 3 memcached_behavior_get_distribution_hash at /usr/lib/libmemcached.so.6:0 # 4 memcached_flush_buffers at /usr/lib/libmemcached.so.6:0 # 5 memcached_get_by_key at /usr/lib/libmemcached.so.6:0 # 6 memcached_get at /usr/lib/libmemcached.so.6:0 # 7 ?? at hhvm:0 # 8 void HPHP::Native::callFunc<false, false>(HPHP::Func const*, void*, HPHP::TypedValue*, HPHP::TypedValue&) at hhvm:0 # 9 HPHP::TypedValue* HPHP::Native::methodWrapper<false, false>(HPHP::ActRec*) at hhvm:0 ``` PHP, for comparison, emits the following, but does not crash: ``` PHP Warning: Memcache::connect(): php_network_getaddresses: getaddrinfo failed: Temporary failure in name resolution in /home/steam/hhvm_memcache_bug_test.php on line 4 PHP Notice: Memcache::connect(): Server junkhost (tcp 123, udp 0) failed with: php_network_getaddresses: getaddrinfo failed: Temporary failure in name resolution (0) in /home/steam/hhvm_memcache_bug_test.php on line 4 PHP Warning: Memcache::connect(): Can't connect to junkhost:123, php_network_getaddresses: getaddrinfo failed: Temporary failure in name resolution (0) in /home/steam/hhvm_memcache_bug_test.php on line 4 ```
comp
segmentation fault if memcache host doesn t resolve hhvm will crash after the object is reused or free d after a memcache add call where the server being used has a hostname that doesn t resolve for example php php memcache new memcache memcache connect junkhost memcache add foo bar echo memcache get foo n crashes echo done n name usr bin hhvm type segmentation fault runtime hhvm version tags hhvm debuggercount arguments hhvm memcache bug test php threadtype cli at hhvm free at tmp tmp jemalloc src jemalloc c freeaddrinfo at lib linux gnu libc so memcached behavior get distribution hash at usr lib libmemcached so memcached flush buffers at usr lib libmemcached so memcached get by key at usr lib libmemcached so memcached get at usr lib libmemcached so at hhvm void hphp native callfunc hphp func const void hphp typedvalue hphp typedvalue at hhvm hphp typedvalue hphp native methodwrapper hphp actrec at hhvm php for comparison emits the following but does not crash php warning memcache connect php network getaddresses getaddrinfo failed temporary failure in name resolution in home steam hhvm memcache bug test php on line php notice memcache connect server junkhost tcp udp failed with php network getaddresses getaddrinfo failed temporary failure in name resolution in home steam hhvm memcache bug test php on line php warning memcache connect can t connect to junkhost php network getaddresses getaddrinfo failed temporary failure in name resolution in home steam hhvm memcache bug test php on line
1
16,289
21,930,693,535
IssuesEvent
2022-05-23 09:28:01
handsontable/handsontable
https://api.github.com/repos/handsontable/handsontable
opened
Replace void 0 with undefined
Type: Improvement suggestion Core: compatibility
### Description Handsontable used `void 0` since the beginning, for its greater compatibility than `undefined` with the browsers *before* ES5. In the current release, we no longer have to support legacy browsers, so we can replace all occurrences of `void 0 `with `undefined`. ### Steps to reproduce <!--- Provide steps to reproduce this issue --> 1. 2. 3. ### Demo <!--- Provide a link to a live example on JSFiddle or CodePen or fill the following demo with your settings --> https://jsfiddle.net/handsoncode/8ffpsqt6/ ### Your environment * Handsontable version: * Browser name and version: * Operating system:
True
Replace void 0 with undefined - ### Description Handsontable used `void 0` since the beginning, for its greater compatibility than `undefined` with the browsers *before* ES5. In the current release, we no longer have to support legacy browsers, so we can replace all occurrences of `void 0 `with `undefined`. ### Steps to reproduce <!--- Provide steps to reproduce this issue --> 1. 2. 3. ### Demo <!--- Provide a link to a live example on JSFiddle or CodePen or fill the following demo with your settings --> https://jsfiddle.net/handsoncode/8ffpsqt6/ ### Your environment * Handsontable version: * Browser name and version: * Operating system:
comp
replace void with undefined description handsontable used void since the beginning for its greater compatibility than undefined with the browsers before in the current release we no longer have to support legacy browsers so we can replace all occurrences of void with undefined steps to reproduce demo your environment handsontable version browser name and version operating system
1
92,176
11,614,431,557
IssuesEvent
2020-02-26 12:35:05
pandas-dev/pandas
https://api.github.com/repos/pandas-dev/pandas
closed
API/ENH: Support/document/test fold argument in Timestamp
API Design Timeseries Timezones
Python 3.6 added a `fold` argument in `datetime.datetime` to disambiguate DST transition times that occur twice (in wall time). https://docs.python.org/3/library/datetime.html#datetime-objects. Technically `Timestamp` will accept the argument in 3.6, but it's not formally documented or tested. Additionally since we will still be supporting 3.5 after dropping 2.7, we can add/handle a fold argument directly in the `Timestamp` constructor as well.
1.0
API/ENH: Support/document/test fold argument in Timestamp - Python 3.6 added a `fold` argument in `datetime.datetime` to disambiguate DST transition times that occur twice (in wall time). https://docs.python.org/3/library/datetime.html#datetime-objects. Technically `Timestamp` will accept the argument in 3.6, but it's not formally documented or tested. Additionally since we will still be supporting 3.5 after dropping 2.7, we can add/handle a fold argument directly in the `Timestamp` constructor as well.
non_comp
api enh support document test fold argument in timestamp python added a fold argument in datetime datetime to disambiguate dst transition times that occur twice in wall time technically timestamp will accept the argument in but it s not formally documented or tested additionally since we will still be supporting after dropping we can add handle a fold argument directly in the timestamp constructor as well
0
16,995
23,418,158,261
IssuesEvent
2022-08-13 09:09:28
jesus2099/konami-command
https://api.github.com/repos/jesus2099/konami-command
reopened
Quick redirects now bypass browser history?
browser compatibility mb_REDIRECT-WHEN-UNIQUE-RESULT
Apparently (now) quick redirects, bypass the browser history. In my Windows VIvaldi, if my search triggers immediate (or quick) redirect, I can no longer go back to the search result.
True
Quick redirects now bypass browser history? - Apparently (now) quick redirects, bypass the browser history. In my Windows VIvaldi, if my search triggers immediate (or quick) redirect, I can no longer go back to the search result.
comp
quick redirects now bypass browser history apparently now quick redirects bypass the browser history in my windows vivaldi if my search triggers immediate or quick redirect i can no longer go back to the search result
1
219,008
24,424,879,075
IssuesEvent
2022-10-06 01:11:48
dreamboy9/mongo
https://api.github.com/repos/dreamboy9/mongo
opened
WS-2022-0322 (Medium) detected in d3-color-2.0.0.tgz
security vulnerability
## WS-2022-0322 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>d3-color-2.0.0.tgz</b></p></summary> <p>Color spaces! RGB, HSL, Cubehelix, Lab and HCL (Lch).</p> <p>Library home page: <a href="https://registry.npmjs.org/d3-color/-/d3-color-2.0.0.tgz">https://registry.npmjs.org/d3-color/-/d3-color-2.0.0.tgz</a></p> <p>Path to dependency file: /buildscripts/libdeps/graph_visualizer_web_stack/package.json</p> <p>Path to vulnerable library: /buildscripts/libdeps/graph_visualizer_web_stack/node_modules/d3-color/package.json</p> <p> Dependency Hierarchy: - force-graph-1.40.0.tgz (Root Library) - d3-scale-chromatic-2.0.0.tgz - :x: **d3-color-2.0.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/dreamboy9/mongo/commit/60ef70ebd8d46f4c893b3fb90ccf2616f8e21d2b">60ef70ebd8d46f4c893b3fb90ccf2616f8e21d2b</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The d3-color module provides representations for various color spaces in the browser. Versions prior to 3.1.0 are vulnerable to a Regular expression Denial of Service. This issue has been patched in version 3.1.0. There are no known workarounds. <p>Publish Date: 2022-09-29 <p>URL: <a href=https://github.com/d3/d3-color/commit/994d8fd95181484a5a27c5edc919aa625781432d>WS-2022-0322</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-36jr-mh4h-2g58">https://github.com/advisories/GHSA-36jr-mh4h-2g58</a></p> <p>Release Date: 2022-09-29</p> <p>Fix Resolution (d3-color): 3.1.0</p> <p>Direct dependency fix Resolution (force-graph): 1.41.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2022-0322 (Medium) detected in d3-color-2.0.0.tgz - ## WS-2022-0322 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>d3-color-2.0.0.tgz</b></p></summary> <p>Color spaces! RGB, HSL, Cubehelix, Lab and HCL (Lch).</p> <p>Library home page: <a href="https://registry.npmjs.org/d3-color/-/d3-color-2.0.0.tgz">https://registry.npmjs.org/d3-color/-/d3-color-2.0.0.tgz</a></p> <p>Path to dependency file: /buildscripts/libdeps/graph_visualizer_web_stack/package.json</p> <p>Path to vulnerable library: /buildscripts/libdeps/graph_visualizer_web_stack/node_modules/d3-color/package.json</p> <p> Dependency Hierarchy: - force-graph-1.40.0.tgz (Root Library) - d3-scale-chromatic-2.0.0.tgz - :x: **d3-color-2.0.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/dreamboy9/mongo/commit/60ef70ebd8d46f4c893b3fb90ccf2616f8e21d2b">60ef70ebd8d46f4c893b3fb90ccf2616f8e21d2b</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The d3-color module provides representations for various color spaces in the browser. Versions prior to 3.1.0 are vulnerable to a Regular expression Denial of Service. This issue has been patched in version 3.1.0. There are no known workarounds. <p>Publish Date: 2022-09-29 <p>URL: <a href=https://github.com/d3/d3-color/commit/994d8fd95181484a5a27c5edc919aa625781432d>WS-2022-0322</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-36jr-mh4h-2g58">https://github.com/advisories/GHSA-36jr-mh4h-2g58</a></p> <p>Release Date: 2022-09-29</p> <p>Fix Resolution (d3-color): 3.1.0</p> <p>Direct dependency fix Resolution (force-graph): 1.41.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_comp
ws medium detected in color tgz ws medium severity vulnerability vulnerable library color tgz color spaces rgb hsl cubehelix lab and hcl lch library home page a href path to dependency file buildscripts libdeps graph visualizer web stack package json path to vulnerable library buildscripts libdeps graph visualizer web stack node modules color package json dependency hierarchy force graph tgz root library scale chromatic tgz x color tgz vulnerable library found in head commit a href found in base branch master vulnerability details the color module provides representations for various color spaces in the browser versions prior to are vulnerable to a regular expression denial of service this issue has been patched in version there are no known workarounds publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution color direct dependency fix resolution force graph step up your open source security game with mend
0
1,664
3,862,590,921
IssuesEvent
2016-04-08 04:02:46
krpc/krpc
https://api.github.com/repos/krpc/krpc
opened
Camera controls
services-spacecenter
Add RPCs to control the camera: * Camera mode (auto/chase/free etc., map and IVA) * Zoom level * Pan/rotate * Set the body/vessel/node in focus when the camera is in map mode
1.0
Camera controls - Add RPCs to control the camera: * Camera mode (auto/chase/free etc., map and IVA) * Zoom level * Pan/rotate * Set the body/vessel/node in focus when the camera is in map mode
non_comp
camera controls add rpcs to control the camera camera mode auto chase free etc map and iva zoom level pan rotate set the body vessel node in focus when the camera is in map mode
0
20,323
29,811,429,248
IssuesEvent
2023-06-16 15:20:48
okfn-brasil/querido-diario
https://api.github.com/repos/okfn-brasil/querido-diario
closed
[Novo spider base]: Siganet
type: spider incompatible
Ao verificar o PR#746 de @lucasocarvalhos, foi possível ver no rodapé da página inicial consta uma nota de copyright mencionando uma empresa ou produto chamado [Siganet](https://www.siganet.net.br/) Baseado nos resultados das busca em Google por `Siganet site:gov.br diário prefeitura`, temos alguns municípios que parecem utilizar o mesmo sistema: - [ ] 2100907 Araioses, MA (último diário: 2022-10-20) https://transparencia.araioses.ma.gov.br/acessoInformacao/diario/diario - [ ] 2101251 Bacabeira, MA (último diário: 2022-09-28) https://transparencia.bacabeira.ma.gov.br/acessoInformacao/diario/diario - [ ] 2101350 Bacurituba, MA (último diário: 2022-10-20) https://transparencia.bacurituba.ma.gov.br/acessoInformacao/diario/diario - [X] 2101400 Balsas, MA (último diário: 2022-10-21 - PR #746) https://transparencia.balsas.ma.gov.br/acessoInformacao/diario/diario - [ ] 2101806 Benedito Leite, MA (último diário: 2022-10-20) https://transparencia.beneditoleite.ma.gov.br/acessoInformacao/diario/diario - [ ] 2102804 Carolina, MA (último diário: 2022-10-19) https://transparencia.carolina.ma.gov.br/acessoInformacao/diario/diario - [ ] 2103604 Coroatá, MA (último diário: 2022-10-19) https://transparencia.coroata.ma.gov.br/acessoInformacao/diario/diario - [ ] 2104073 Feira Nova do Maranhão, MA (último diário: 2022-10-14) https://transparencia.feiranovadomaranhao.ma.gov.br/acessoInformacao/diario/diario - [ ] 2104107 Fortaleza dos Nogueiras, MA (último diário: 2022-10-21) https://transparencia.fortalezadosnogueiras.ma.gov.br/acessoInformacao/diario/diario - [ ] 2104800 Grajaú, MA (último diário: 2022-10-21) https://transparencia.grajau.ma.gov.br/acessoInformacao/diario/diario - [ ] 2106607 Matões, MA (último diário: 2022-10-20) https://www.transparencia.matoes.ma.gov.br/acessoInformacao/diario/diario - [ ] 2107258 Nova Colinas, MA (último diário: 2022-10-21) https://transparencia.novacolinas.ma.gov.br/acessoInformacao/diario/diario - [ ] 2107407 Olho d'Água das Cunhãs, MA (último diário: 2022-10-20) https://transparencia.olhodaguadascunhas.ma.gov.br/acessoInformacao/diario/diario - [ ] 2108058 Paulino Neves, MA (último diário: 2022-10-17) https://transparencia.paulinoneves.ma.gov.br/acessoInformacao/diario/diario - [ ] 2108306 Penalva, MA (último diário: 2022-10-21) https://transparencia.penalva.ma.gov.br/acessoInformacao/diario/diario - [ ] 2109007 Porto Franco, MA (último diário: 2016-12-11 !!!!!!) https://transparencia.portofranco.ma.gov.br/acessoInformacao/diario/diario - [ ] 2109601 Rosário, MA (último diário: 2022-10-21) https://www.transparencia.rosario.ma.gov.br/acessoInformacao/diario/diario - [ ] 2110039 Santa Luzia do Paruá, MA (último diário: 2022-10-21) https://transparencia.santaluziadoparua.ma.gov.br/acessoInformacao/diario/diario - [ ] 2110658 São Domingos do Azeitão, MA (último diário: 2022-10-21) https://transparencia.saodomingosdoazeitao.ma.gov.br/acessoInformacao/diario/diario - [ ] 2110906 São Francisco do Maranhão, MA (último diário: 2022-10-17) https://transparencia.saofranciscodomaranhao.ma.gov.br/acessoInformacao/diario/diario - [x] 2111250 São José dos Basílios, MA (último diário: 2022-10-20) https://transparencia.saojosedosbasilios.ma.gov.br/acessoInformacao/diario/diario - [ ] 2111573 São Pedro dos Crentes, MA (último diário: 2022-10-18) https://transparencia.saopedrodoscrentes.ma.gov.br/acessoInformacao/diario/diario - [ ] 2112506 Tutóia, MA (último diário: 2022-10-18) https://transparencia.tutoia.ma.gov.br/acessoInformacao/diario/diario - [ ] 2112803 Viana, MA (último diário: 2022-10-21) https://transparencia.viana.ma.gov.br/acessoInformacao/diario/diario Essa issue seria sobre criar um spider base para Siganet ou refatorar o criado para Balsas, MA.
True
[Novo spider base]: Siganet - Ao verificar o PR#746 de @lucasocarvalhos, foi possível ver no rodapé da página inicial consta uma nota de copyright mencionando uma empresa ou produto chamado [Siganet](https://www.siganet.net.br/) Baseado nos resultados das busca em Google por `Siganet site:gov.br diário prefeitura`, temos alguns municípios que parecem utilizar o mesmo sistema: - [ ] 2100907 Araioses, MA (último diário: 2022-10-20) https://transparencia.araioses.ma.gov.br/acessoInformacao/diario/diario - [ ] 2101251 Bacabeira, MA (último diário: 2022-09-28) https://transparencia.bacabeira.ma.gov.br/acessoInformacao/diario/diario - [ ] 2101350 Bacurituba, MA (último diário: 2022-10-20) https://transparencia.bacurituba.ma.gov.br/acessoInformacao/diario/diario - [X] 2101400 Balsas, MA (último diário: 2022-10-21 - PR #746) https://transparencia.balsas.ma.gov.br/acessoInformacao/diario/diario - [ ] 2101806 Benedito Leite, MA (último diário: 2022-10-20) https://transparencia.beneditoleite.ma.gov.br/acessoInformacao/diario/diario - [ ] 2102804 Carolina, MA (último diário: 2022-10-19) https://transparencia.carolina.ma.gov.br/acessoInformacao/diario/diario - [ ] 2103604 Coroatá, MA (último diário: 2022-10-19) https://transparencia.coroata.ma.gov.br/acessoInformacao/diario/diario - [ ] 2104073 Feira Nova do Maranhão, MA (último diário: 2022-10-14) https://transparencia.feiranovadomaranhao.ma.gov.br/acessoInformacao/diario/diario - [ ] 2104107 Fortaleza dos Nogueiras, MA (último diário: 2022-10-21) https://transparencia.fortalezadosnogueiras.ma.gov.br/acessoInformacao/diario/diario - [ ] 2104800 Grajaú, MA (último diário: 2022-10-21) https://transparencia.grajau.ma.gov.br/acessoInformacao/diario/diario - [ ] 2106607 Matões, MA (último diário: 2022-10-20) https://www.transparencia.matoes.ma.gov.br/acessoInformacao/diario/diario - [ ] 2107258 Nova Colinas, MA (último diário: 2022-10-21) https://transparencia.novacolinas.ma.gov.br/acessoInformacao/diario/diario - [ ] 2107407 Olho d'Água das Cunhãs, MA (último diário: 2022-10-20) https://transparencia.olhodaguadascunhas.ma.gov.br/acessoInformacao/diario/diario - [ ] 2108058 Paulino Neves, MA (último diário: 2022-10-17) https://transparencia.paulinoneves.ma.gov.br/acessoInformacao/diario/diario - [ ] 2108306 Penalva, MA (último diário: 2022-10-21) https://transparencia.penalva.ma.gov.br/acessoInformacao/diario/diario - [ ] 2109007 Porto Franco, MA (último diário: 2016-12-11 !!!!!!) https://transparencia.portofranco.ma.gov.br/acessoInformacao/diario/diario - [ ] 2109601 Rosário, MA (último diário: 2022-10-21) https://www.transparencia.rosario.ma.gov.br/acessoInformacao/diario/diario - [ ] 2110039 Santa Luzia do Paruá, MA (último diário: 2022-10-21) https://transparencia.santaluziadoparua.ma.gov.br/acessoInformacao/diario/diario - [ ] 2110658 São Domingos do Azeitão, MA (último diário: 2022-10-21) https://transparencia.saodomingosdoazeitao.ma.gov.br/acessoInformacao/diario/diario - [ ] 2110906 São Francisco do Maranhão, MA (último diário: 2022-10-17) https://transparencia.saofranciscodomaranhao.ma.gov.br/acessoInformacao/diario/diario - [x] 2111250 São José dos Basílios, MA (último diário: 2022-10-20) https://transparencia.saojosedosbasilios.ma.gov.br/acessoInformacao/diario/diario - [ ] 2111573 São Pedro dos Crentes, MA (último diário: 2022-10-18) https://transparencia.saopedrodoscrentes.ma.gov.br/acessoInformacao/diario/diario - [ ] 2112506 Tutóia, MA (último diário: 2022-10-18) https://transparencia.tutoia.ma.gov.br/acessoInformacao/diario/diario - [ ] 2112803 Viana, MA (último diário: 2022-10-21) https://transparencia.viana.ma.gov.br/acessoInformacao/diario/diario Essa issue seria sobre criar um spider base para Siganet ou refatorar o criado para Balsas, MA.
comp
siganet ao verificar o pr de lucasocarvalhos foi possível ver no rodapé da página inicial consta uma nota de copyright mencionando uma empresa ou produto chamado baseado nos resultados das busca em google por siganet site gov br diário prefeitura temos alguns municípios que parecem utilizar o mesmo sistema araioses ma último diário bacabeira ma último diário bacurituba ma último diário balsas ma último diário pr benedito leite ma último diário carolina ma último diário coroatá ma último diário feira nova do maranhão ma último diário fortaleza dos nogueiras ma último diário grajaú ma último diário matões ma último diário nova colinas ma último diário olho d água das cunhãs ma último diário paulino neves ma último diário penalva ma último diário porto franco ma último diário rosário ma último diário santa luzia do paruá ma último diário são domingos do azeitão ma último diário são francisco do maranhão ma último diário são josé dos basílios ma último diário são pedro dos crentes ma último diário tutóia ma último diário viana ma último diário essa issue seria sobre criar um spider base para siganet ou refatorar o criado para balsas ma
1
8,328
10,347,458,235
IssuesEvent
2019-09-04 17:26:52
cobalt-org/liquid-rust
https://api.github.com/repos/cobalt-org/liquid-rust
closed
`compact` needs to accept an optional `property` parameter
enhancement question std-compatibility
See #267, #333, #334 Example test: ``` assert_eq!(v!([]), filters!(Compact, v!([]), v!("a"))); ```
True
`compact` needs to accept an optional `property` parameter - See #267, #333, #334 Example test: ``` assert_eq!(v!([]), filters!(Compact, v!([]), v!("a"))); ```
comp
compact needs to accept an optional property parameter see example test assert eq v filters compact v v a
1
274,593
23,852,101,692
IssuesEvent
2022-09-06 18:58:30
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachtest: kv0/enc=false/nodes=3/cpu=96 failed
C-test-failure O-robot O-roachtest release-blocker T-kv branch-release-22.1
roachtest.kv0/enc=false/nodes=3/cpu=96 [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=6195973&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=6195973&tab=artifacts#/kv0/enc=false/nodes=3/cpu=96) on release-22.1 @ [714fa0ad80c499cbd96ba97c560a9b414c61104f](https://github.com/cockroachdb/cockroach/commits/714fa0ad80c499cbd96ba97c560a9b414c61104f): ``` | 344.0s 0 56548.8 55618.0 2.6 10.0 18.9 39.8 write | 345.0s 0 57164.1 55622.5 2.6 7.3 18.9 33.6 write | 346.0s 0 55803.8 55623.0 2.2 10.0 18.9 30.4 write | 347.0s 0 53686.0 55617.4 2.4 11.5 19.9 29.4 write | 348.0s 0 14877.9 55500.3 2.2 13.6 19.9 3355.4 write | 349.0s 0 78.1 55341.6 1.5 3355.4 3355.4 3355.4 write | 350.0s 0 0.0 55183.4 0.0 0.0 0.0 0.0 write | 351.0s 0 114.0 55026.6 1.4 3355.4 3355.4 3355.4 write | 352.0s 0 0.0 54870.1 0.0 0.0 0.0 0.0 write | 353.0s 0 0.0 54714.7 0.0 0.0 0.0 0.0 write | 354.0s 0 0.0 54560.3 0.0 0.0 0.0 0.0 write Wraps: (4) COMMAND_PROBLEM Wraps: (5) Node 4. Command with error: | `````` | ./workload run kv --init --histograms=perf/stats.json --concurrency=192 --splits=1000 --duration=30m0s --read-percent=0 {pgurl:1-3} | `````` Wraps: (6) exit status 1 Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *cluster.WithCommandDetails (4) errors.Cmd (5) *hintdetail.withDetail (6) *exec.ExitError monitor.go:127,kv.go:154,kv.go:276,test_runner.go:883: monitor failure: monitor task failed: t.Fatal() was called (1) attached stack trace -- stack trace: | main.(*monitorImpl).WaitE | main/pkg/cmd/roachtest/monitor.go:115 | main.(*monitorImpl).Wait | main/pkg/cmd/roachtest/monitor.go:123 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.registerKV.func2 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/kv.go:154 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.registerKV.func3 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/kv.go:276 | main.(*testRunner).runTest.func2 | main/pkg/cmd/roachtest/test_runner.go:883 Wraps: (2) monitor failure Wraps: (3) attached stack trace -- stack trace: | main.(*monitorImpl).wait.func2 | main/pkg/cmd/roachtest/monitor.go:171 Wraps: (4) monitor task failed Wraps: (5) attached stack trace -- stack trace: | main.init | main/pkg/cmd/roachtest/monitor.go:80 | runtime.doInit | GOROOT/src/runtime/proc.go:6498 | runtime.main | GOROOT/src/runtime/proc.go:238 | runtime.goexit | GOROOT/src/runtime/asm_amd64.s:1581 Wraps: (6) t.Fatal() was called Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *withstack.withStack (4) *errutil.withPrefix (5) *withstack.withStack (6) *errutil.leafError ``` <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/kv-triage <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*kv0/enc=false/nodes=3/cpu=96.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-18871
2.0
roachtest: kv0/enc=false/nodes=3/cpu=96 failed - roachtest.kv0/enc=false/nodes=3/cpu=96 [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=6195973&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=6195973&tab=artifacts#/kv0/enc=false/nodes=3/cpu=96) on release-22.1 @ [714fa0ad80c499cbd96ba97c560a9b414c61104f](https://github.com/cockroachdb/cockroach/commits/714fa0ad80c499cbd96ba97c560a9b414c61104f): ``` | 344.0s 0 56548.8 55618.0 2.6 10.0 18.9 39.8 write | 345.0s 0 57164.1 55622.5 2.6 7.3 18.9 33.6 write | 346.0s 0 55803.8 55623.0 2.2 10.0 18.9 30.4 write | 347.0s 0 53686.0 55617.4 2.4 11.5 19.9 29.4 write | 348.0s 0 14877.9 55500.3 2.2 13.6 19.9 3355.4 write | 349.0s 0 78.1 55341.6 1.5 3355.4 3355.4 3355.4 write | 350.0s 0 0.0 55183.4 0.0 0.0 0.0 0.0 write | 351.0s 0 114.0 55026.6 1.4 3355.4 3355.4 3355.4 write | 352.0s 0 0.0 54870.1 0.0 0.0 0.0 0.0 write | 353.0s 0 0.0 54714.7 0.0 0.0 0.0 0.0 write | 354.0s 0 0.0 54560.3 0.0 0.0 0.0 0.0 write Wraps: (4) COMMAND_PROBLEM Wraps: (5) Node 4. Command with error: | `````` | ./workload run kv --init --histograms=perf/stats.json --concurrency=192 --splits=1000 --duration=30m0s --read-percent=0 {pgurl:1-3} | `````` Wraps: (6) exit status 1 Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *cluster.WithCommandDetails (4) errors.Cmd (5) *hintdetail.withDetail (6) *exec.ExitError monitor.go:127,kv.go:154,kv.go:276,test_runner.go:883: monitor failure: monitor task failed: t.Fatal() was called (1) attached stack trace -- stack trace: | main.(*monitorImpl).WaitE | main/pkg/cmd/roachtest/monitor.go:115 | main.(*monitorImpl).Wait | main/pkg/cmd/roachtest/monitor.go:123 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.registerKV.func2 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/kv.go:154 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.registerKV.func3 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/kv.go:276 | main.(*testRunner).runTest.func2 | main/pkg/cmd/roachtest/test_runner.go:883 Wraps: (2) monitor failure Wraps: (3) attached stack trace -- stack trace: | main.(*monitorImpl).wait.func2 | main/pkg/cmd/roachtest/monitor.go:171 Wraps: (4) monitor task failed Wraps: (5) attached stack trace -- stack trace: | main.init | main/pkg/cmd/roachtest/monitor.go:80 | runtime.doInit | GOROOT/src/runtime/proc.go:6498 | runtime.main | GOROOT/src/runtime/proc.go:238 | runtime.goexit | GOROOT/src/runtime/asm_amd64.s:1581 Wraps: (6) t.Fatal() was called Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *withstack.withStack (4) *errutil.withPrefix (5) *withstack.withStack (6) *errutil.leafError ``` <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/kv-triage <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*kv0/enc=false/nodes=3/cpu=96.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-18871
non_comp
roachtest enc false nodes cpu failed roachtest enc false nodes cpu with on release write write write write write write write write write write write wraps command problem wraps node command with error workload run kv init histograms perf stats json concurrency splits duration read percent pgurl wraps exit status error types withstack withstack errutil withprefix cluster withcommanddetails errors cmd hintdetail withdetail exec exiterror monitor go kv go kv go test runner go monitor failure monitor task failed t fatal was called attached stack trace stack trace main monitorimpl waite main pkg cmd roachtest monitor go main monitorimpl wait main pkg cmd roachtest monitor go github com cockroachdb cockroach pkg cmd roachtest tests registerkv github com cockroachdb cockroach pkg cmd roachtest tests kv go github com cockroachdb cockroach pkg cmd roachtest tests registerkv github com cockroachdb cockroach pkg cmd roachtest tests kv go main testrunner runtest main pkg cmd roachtest test runner go wraps monitor failure wraps attached stack trace stack trace main monitorimpl wait main pkg cmd roachtest monitor go wraps monitor task failed wraps attached stack trace stack trace main init main pkg cmd roachtest monitor go runtime doinit goroot src runtime proc go runtime main goroot src runtime proc go runtime goexit goroot src runtime asm s wraps t fatal was called error types withstack withstack errutil withprefix withstack withstack errutil withprefix withstack withstack errutil leaferror help see see cc cockroachdb kv triage jira issue crdb
0
13,179
15,531,276,750
IssuesEvent
2021-03-13 22:48:06
cseelhoff/RimThreaded
https://api.github.com/repos/cseelhoff/RimThreaded
closed
"Shield Generators by Frontier Developments" not working as intended
Bug Mod Incompatibility Needs Validation Not Enough Info Reproducible
**Describe the bug** In using the mod "Shield Generators by Frontier Developments," the shield generators within the mod do not function as intended. Typically, bullets are supposed to be stopped by the shield, and their damage absorbed by the generator. Currently, the shield generated by all 3 generator types stop line of sight into them. This causes pawns to not be able to fire into the shield, nor select any targets within it or through it. Pawns can still shoot out of the shield without issue. This issue does not generate an error. **To Reproduce** Steps to reproduce the behavior: 1. Enter any colony. 2. Place any Shield Generator (make sure it's powered.) 3. Attempt to have a pawn shoot into it. **Mod List** Shield Generators by Frontier Developments
True
"Shield Generators by Frontier Developments" not working as intended - **Describe the bug** In using the mod "Shield Generators by Frontier Developments," the shield generators within the mod do not function as intended. Typically, bullets are supposed to be stopped by the shield, and their damage absorbed by the generator. Currently, the shield generated by all 3 generator types stop line of sight into them. This causes pawns to not be able to fire into the shield, nor select any targets within it or through it. Pawns can still shoot out of the shield without issue. This issue does not generate an error. **To Reproduce** Steps to reproduce the behavior: 1. Enter any colony. 2. Place any Shield Generator (make sure it's powered.) 3. Attempt to have a pawn shoot into it. **Mod List** Shield Generators by Frontier Developments
comp
shield generators by frontier developments not working as intended describe the bug in using the mod shield generators by frontier developments the shield generators within the mod do not function as intended typically bullets are supposed to be stopped by the shield and their damage absorbed by the generator currently the shield generated by all generator types stop line of sight into them this causes pawns to not be able to fire into the shield nor select any targets within it or through it pawns can still shoot out of the shield without issue this issue does not generate an error to reproduce steps to reproduce the behavior enter any colony place any shield generator make sure it s powered attempt to have a pawn shoot into it mod list shield generators by frontier developments
1
1,201
3,697,522,397
IssuesEvent
2016-02-27 18:44:45
sekiguchi-nagisa/ydsh
https://api.github.com/repos/sekiguchi-nagisa/ydsh
opened
support background job
enhancement incompatible change
introduce Async type. Async type has wait method for waiting process termination. string representation of Async object is pid. ``` $ ping -c 192.168.0.1 & (Async) 5678 $ var a = echo hoge & $ $a (Async) 5679 $ (echo huga &).wait() # wait termination ```
True
support background job - introduce Async type. Async type has wait method for waiting process termination. string representation of Async object is pid. ``` $ ping -c 192.168.0.1 & (Async) 5678 $ var a = echo hoge & $ $a (Async) 5679 $ (echo huga &).wait() # wait termination ```
comp
support background job introduce async type async type has wait method for waiting process termination string representation of async object is pid ping c async var a echo hoge a async echo huga wait wait termination
1
432,542
30,287,233,261
IssuesEvent
2023-07-08 21:01:12
gdg-berlin-android/ZeBadge
https://api.github.com/repos/gdg-berlin-android/ZeBadge
opened
What should the buttons on ZeBadge do while connected
📖 documentation ⁉️ question 🤖 app 🪪 badge
Ideate on what the buttons should do while connected to the phone.
1.0
What should the buttons on ZeBadge do while connected - Ideate on what the buttons should do while connected to the phone.
non_comp
what should the buttons on zebadge do while connected ideate on what the buttons should do while connected to the phone
0
333,502
29,669,509,681
IssuesEvent
2023-06-11 08:19:06
cse110-sp23-group23/Zoltar
https://api.github.com/repos/cse110-sp23-group23/Zoltar
opened
Mute button test
End to End Tests
Requirement Description: - Toggle mute button test Any key challenges: - Steps to Implementing: - visuals.test.js Other: -
1.0
Mute button test - Requirement Description: - Toggle mute button test Any key challenges: - Steps to Implementing: - visuals.test.js Other: -
non_comp
mute button test requirement description toggle mute button test any key challenges steps to implementing visuals test js other
0
8,343
10,367,917,704
IssuesEvent
2019-09-07 12:45:46
NEZNAMY/TAB
https://api.github.com/repos/NEZNAMY/TAB
closed
Permissions EX bug
[1] error [2] compatibility [4] wontfix
07.09.2019 - 14:15:59 - [TAB v2.5.1] An error occured when setting placeholders(2) (player=Pericol123, online=true) 07.09.2019 - 14:15:59 - ru.tehkode.permissions.exceptions.PermissionsNotAvailable: Permissions manager is not accessable. Is the PermissionsEx plugin enabled? 07.09.2019 - 14:15:59 - at ru.tehkode.permissions.bukkit.PermissionsEx.getPermissionManager(PermissionsEx.java:405) 07.09.2019 - 14:15:59 - at net.milkbowl.vault.chat.plugins.Chat_PermissionsEx.getUser(Chat_PermissionsEx.java:108) 07.09.2019 - 14:15:59 - at net.milkbowl.vault.chat.plugins.Chat_PermissionsEx.getPlayerPrefix(Chat_PermissionsEx.java:292) 07.09.2019 - 14:15:59 - at net.milkbowl.vault.chat.Chat.getPlayerPrefix(Chat.java:91) 07.09.2019 - 14:15:59 - at me.neznamy.tab.bukkit.Placeholders.replaceSimplePlaceholders(Placeholders.java:151) 07.09.2019 - 14:15:59 - at me.neznamy.tab.bukkit.Placeholders.replace(Placeholders.java:106) 07.09.2019 - 14:15:59 - at me.neznamy.tab.shared.Placeholders.replace(Placeholders.java:123) 07.09.2019 - 14:15:59 - at me.neznamy.tab.shared.Placeholders.replaceMultiple(Placeholders.java:101) 07.09.2019 - 14:15:59 - at me.neznamy.tab.shared.ITabPlayer.updateTeamPrefixSuffix(ITabPlayer.java:348) 07.09.2019 - 14:15:59 - at me.neznamy.tab.shared.ITabPlayer.updateTeam(ITabPlayer.java:110) 07.09.2019 - 14:15:59 - at me.neznamy.tab.shared.NameTag16$1.run(NameTag16.java:22) 07.09.2019 - 14:15:59 - at me.neznamy.tab.shared.Shared$2.run(Shared.java:149) 07.09.2019 - 14:15:59 - at java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) 07.09.2019 - 14:15:59 - at java.util.concurrent.FutureTask.run(Unknown Source) 07.09.2019 - 14:15:59 - at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) 07.09.2019 - 14:15:59 - at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) 07.09.2019 - 14:15:59 - at java.lang.Thread.run(Unknown Source)
True
Permissions EX bug - 07.09.2019 - 14:15:59 - [TAB v2.5.1] An error occured when setting placeholders(2) (player=Pericol123, online=true) 07.09.2019 - 14:15:59 - ru.tehkode.permissions.exceptions.PermissionsNotAvailable: Permissions manager is not accessable. Is the PermissionsEx plugin enabled? 07.09.2019 - 14:15:59 - at ru.tehkode.permissions.bukkit.PermissionsEx.getPermissionManager(PermissionsEx.java:405) 07.09.2019 - 14:15:59 - at net.milkbowl.vault.chat.plugins.Chat_PermissionsEx.getUser(Chat_PermissionsEx.java:108) 07.09.2019 - 14:15:59 - at net.milkbowl.vault.chat.plugins.Chat_PermissionsEx.getPlayerPrefix(Chat_PermissionsEx.java:292) 07.09.2019 - 14:15:59 - at net.milkbowl.vault.chat.Chat.getPlayerPrefix(Chat.java:91) 07.09.2019 - 14:15:59 - at me.neznamy.tab.bukkit.Placeholders.replaceSimplePlaceholders(Placeholders.java:151) 07.09.2019 - 14:15:59 - at me.neznamy.tab.bukkit.Placeholders.replace(Placeholders.java:106) 07.09.2019 - 14:15:59 - at me.neznamy.tab.shared.Placeholders.replace(Placeholders.java:123) 07.09.2019 - 14:15:59 - at me.neznamy.tab.shared.Placeholders.replaceMultiple(Placeholders.java:101) 07.09.2019 - 14:15:59 - at me.neznamy.tab.shared.ITabPlayer.updateTeamPrefixSuffix(ITabPlayer.java:348) 07.09.2019 - 14:15:59 - at me.neznamy.tab.shared.ITabPlayer.updateTeam(ITabPlayer.java:110) 07.09.2019 - 14:15:59 - at me.neznamy.tab.shared.NameTag16$1.run(NameTag16.java:22) 07.09.2019 - 14:15:59 - at me.neznamy.tab.shared.Shared$2.run(Shared.java:149) 07.09.2019 - 14:15:59 - at java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) 07.09.2019 - 14:15:59 - at java.util.concurrent.FutureTask.run(Unknown Source) 07.09.2019 - 14:15:59 - at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) 07.09.2019 - 14:15:59 - at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) 07.09.2019 - 14:15:59 - at java.lang.Thread.run(Unknown Source)
comp
permissions ex bug an error occured when setting placeholders player online true ru tehkode permissions exceptions permissionsnotavailable permissions manager is not accessable is the permissionsex plugin enabled at ru tehkode permissions bukkit permissionsex getpermissionmanager permissionsex java at net milkbowl vault chat plugins chat permissionsex getuser chat permissionsex java at net milkbowl vault chat plugins chat permissionsex getplayerprefix chat permissionsex java at net milkbowl vault chat chat getplayerprefix chat java at me neznamy tab bukkit placeholders replacesimpleplaceholders placeholders java at me neznamy tab bukkit placeholders replace placeholders java at me neznamy tab shared placeholders replace placeholders java at me neznamy tab shared placeholders replacemultiple placeholders java at me neznamy tab shared itabplayer updateteamprefixsuffix itabplayer java at me neznamy tab shared itabplayer updateteam itabplayer java at me neznamy tab shared run java at me neznamy tab shared shared run shared java at java util concurrent executors runnableadapter call unknown source at java util concurrent futuretask run unknown source at java util concurrent threadpoolexecutor runworker unknown source at java util concurrent threadpoolexecutor worker run unknown source at java lang thread run unknown source
1
17,091
23,593,094,471
IssuesEvent
2022-08-23 16:47:43
Frank-Mayer/n5ednd
https://api.github.com/repos/Frank-Mayer/n5ednd
closed
Samsung Internet Dark mode issue
bug wontfix browser compatibility ui
Checkboxes not shown on Samsung Internet when Darkmode is active. Light mode is working fine. ![Screenshot_2022-07-31-12-51-06-075_com sec android app sbrowser](https://user-images.githubusercontent.com/53651857/182022815-fe64b4aa-782d-42a8-b664-6514adb25e8a.jpg)
True
Samsung Internet Dark mode issue - Checkboxes not shown on Samsung Internet when Darkmode is active. Light mode is working fine. ![Screenshot_2022-07-31-12-51-06-075_com sec android app sbrowser](https://user-images.githubusercontent.com/53651857/182022815-fe64b4aa-782d-42a8-b664-6514adb25e8a.jpg)
comp
samsung internet dark mode issue checkboxes not shown on samsung internet when darkmode is active light mode is working fine
1
17,313
23,887,491,062
IssuesEvent
2022-09-08 08:50:53
akuker/RASCSI
https://api.github.com/repos/akuker/RASCSI
closed
Explicit SCSI-1 support: Add support for .hd1 image files
compatibility
rascsi shall support a new image type ".hd1". This type is identical with ".hds" except for the SCSI level and response code format being SCSI-1 instead of SCSI-2. The alternative approach of being able to (having to ...) configure the SCSI and response code levels would force additional configuration settings on the users and would require much more changes in the core, and also in the clients. Most users do not have the expert knowledge required for these settings anyway and would run into issues with regular platforms when playing with these settings. To be done after merging the SASI removal changes. Also see https://github.com/akuker/RASCSI/issues/799.
True
Explicit SCSI-1 support: Add support for .hd1 image files - rascsi shall support a new image type ".hd1". This type is identical with ".hds" except for the SCSI level and response code format being SCSI-1 instead of SCSI-2. The alternative approach of being able to (having to ...) configure the SCSI and response code levels would force additional configuration settings on the users and would require much more changes in the core, and also in the clients. Most users do not have the expert knowledge required for these settings anyway and would run into issues with regular platforms when playing with these settings. To be done after merging the SASI removal changes. Also see https://github.com/akuker/RASCSI/issues/799.
comp
explicit scsi support add support for image files rascsi shall support a new image type this type is identical with hds except for the scsi level and response code format being scsi instead of scsi the alternative approach of being able to having to configure the scsi and response code levels would force additional configuration settings on the users and would require much more changes in the core and also in the clients most users do not have the expert knowledge required for these settings anyway and would run into issues with regular platforms when playing with these settings to be done after merging the sasi removal changes also see
1
58,810
3,091,511,794
IssuesEvent
2015-08-26 13:35:18
pmem/issues
https://api.github.com/repos/pmem/issues
opened
pmemobj: TX_STAGE_ONABORT is not set after fail of allocation/reallocation
Exposure: Medium Priority: 3 medium Type: Bug
Steps to reproduce: 1. Create pool with MIN_POOL_SIZE 2. Start new transaction 3. Use one of functions pmemobj_tx_alloc/zalloc/realloc/zrealloc with size equals range of size_t 4. Check stage Expected result: TX_STAGE_ONABORT stage is set, OID_NULL is returned, errno = 12 Current result: OID_NULL is returned, errno = 12 but TX_STAGE_WORK is set
1.0
pmemobj: TX_STAGE_ONABORT is not set after fail of allocation/reallocation - Steps to reproduce: 1. Create pool with MIN_POOL_SIZE 2. Start new transaction 3. Use one of functions pmemobj_tx_alloc/zalloc/realloc/zrealloc with size equals range of size_t 4. Check stage Expected result: TX_STAGE_ONABORT stage is set, OID_NULL is returned, errno = 12 Current result: OID_NULL is returned, errno = 12 but TX_STAGE_WORK is set
non_comp
pmemobj tx stage onabort is not set after fail of allocation reallocation steps to reproduce create pool with min pool size start new transaction use one of functions pmemobj tx alloc zalloc realloc zrealloc with size equals range of size t check stage expected result tx stage onabort stage is set oid null is returned errno current result oid null is returned errno but tx stage work is set
0
11,426
13,425,730,112
IssuesEvent
2020-09-06 11:29:22
nicoboss/nsz
https://api.github.com/repos/nicoboss/nsz
closed
install failing on kubuntu 20.04
compatibility deployment
some errors: ` ERROR: Command errored out with exit status 1: command: /usr/bin/python3 -c 'import sys, setuptools, tokenize; sys.argv[0] = '"'"'/tmp/pip-install-b1phyb9o/kivy/setup.py'"'"'; __file__='"'"'/tmp/pip-install-b1phyb9o/kivy/setup.py'"'"';f=getattr(tokenize, '"'"'open'"'"', open)(__file__);code=f.read().replace('"'"'\r\n'"'"', '"'"'\n'"'"');f.close();exec(compile(code, __file__, '"'"'exec'"'"'))' egg_info --egg-base /tmp/pip-pip-egg-info-wmee5ot5 cwd: /tmp/pip-install-b1phyb9o/kivy/ Complete output (668 lines): fatal: not a git repository (or any of the parent directories): .git ERROR: Command errored out with exit status 1: command: /usr/bin/python3 -u -c 'import sys, setuptools, tokenize; sys.argv[0] = '"'"'/tmp/pip-wheel-52mhf973/cython/setup.py'"'"'; __file__='"'"'/tmp/pip-wheel-52mhf973/cython/setup.py'"'"';f=getattr(tokenize, '"'"'open'"'"', open)(__file__);code=f.read().replace('"'"'\r\n'"'"', '"'"'\n'"'"');f.close();exec(compile(code, __file__, '"'"'exec'"'"'))' bdist_wheel -d /tmp/pip-wheel-ugtpsi0m cwd: /tmp/pip-wheel-52mhf973/cython/ Complete output (598 lines): Unable to find pgen, not compiling formal grammar. ` creating build/temp.linux-x86_64-3.8/tmp/pip-wheel-52mhf973/cython/Cython/Plex x86_64-linux-gnu-gcc -pthread -Wno-unused-result -Wsign-compare -DNDEBUG -g -fwrapv -O2 -Wall -g -fstack-protector-strong -Wformat -Werror=format-security -g -fwrapv -O2 -g -fstack-protector-strong -Wformat -Werror=format-security -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -I/usr/include/python3.8 -c /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c -o build/temp.linux-x86_64-3.8/tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.o /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c: In function ‘__Pyx_InitCachedConstants’: /usr/include/python3.8/code.h:72:25: warning: passing argument 6 of ‘PyCode_New’ makes pointer from integer without a cast [-Wint-conversion] 72 | #define CO_OPTIMIZED 0x0001 | ^ | | | int /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:333:37: note: in definition of macro ‘__Pyx_PyCode_New’ 333 | PyCode_New(a, 0, k, l, s, f, code, c, n, v, fv, cell, fn, name, fline, lnos) | ^ /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:7241:62: note: in expansion of macro ‘CO_OPTIMIZED’ 7241 | __pyx_codeobj__6 = (PyObject*)__Pyx_PyCode_New(1, 0, 1, 0, CO_OPTIMIZED|CO_NEWLOCALS, __pyx_empty_bytes, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_tuple__5, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_kp_s_Cython_Plex_Scanners_py, __pyx_n_s_read, 113, __pyx_empty_bytes); if (unlikely(!__pyx_codeobj__6)) __PYX_ERR(0, 113, __pyx_L1_error) | ^~~~~~~~~~~~ In file included from /usr/include/python3.8/compile.h:5, from /usr/include/python3.8/Python.h:138, from /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:19: /usr/include/python3.8/code.h:122:28: note: expected ‘PyObject *’ {aka ‘struct _object *’} but argument is of type ‘int’ 122 | PyAPI_FUNC(PyCodeObject *) PyCode_New( | ^~~~~~~~~~ /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:7241:236: warning: passing argument 14 of ‘PyCode_New’ makes integer from pointer without a cast [-Wint-conversion] 7241 | __pyx_codeobj__6 = (PyObject*)__Pyx_PyCode_New(1, 0, 1, 0, CO_OPTIMIZED|CO_NEWLOCALS, __pyx_empty_bytes, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_tuple__5, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_kp_s_Cython_Plex_Scanners_py, __pyx_n_s_read, 113, __pyx_empty_bytes); if (unlikely(!__pyx_codeobj__6)) __PYX_ERR(0, 113, __pyx_L1_error) | ^~~~~~~~~~~~~~ | | | PyObject * {aka struct _object *} /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:333:69: note: in definition of macro ‘__Pyx_PyCode_New’ 333 | PyCode_New(a, 0, k, l, s, f, code, c, n, v, fv, cell, fn, name, fline, lnos) | ^~~~ In file included from /usr/include/python3.8/compile.h:5, from /usr/include/python3.8/Python.h:138, from /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:19: /usr/include/python3.8/code.h:122:28: note: expected ‘int’ but argument is of type ‘PyObject *’ {aka ‘struct _object *’} 122 | PyAPI_FUNC(PyCodeObject *) PyCode_New( | ^~~~~~~~~~ /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:7241:252: warning: passing argument 15 of ‘PyCode_New’ makes pointer from integer without a cast [-Wint-conversion] 7241 | __pyx_codeobj__6 = (PyObject*)__Pyx_PyCode_New(1, 0, 1, 0, CO_OPTIMIZED|CO_NEWLOCALS, __pyx_empty_bytes, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_tuple__5, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_kp_s_Cython_Plex_Scanners_py, __pyx_n_s_read, 113, __pyx_empty_bytes); if (unlikely(!__pyx_codeobj__6)) __PYX_ERR(0, 113, __pyx_L1_error) | ^~~ | | | int /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:333:75: note: in definition of macro ‘__Pyx_PyCode_New’ 333 | PyCode_New(a, 0, k, l, s, f, code, c, n, v, fv, cell, fn, name, fline, lnos) | ^~~~~ In file included from /usr/include/python3.8/compile.h:5, from /usr/include/python3.8/Python.h:138, from /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:19: /usr/include/python3.8/code.h:122:28: note: expected ‘PyObject *’ {aka ‘struct _object *’} but argument is of type ‘int’ 122 | PyAPI_FUNC(PyCodeObject *) PyCode_New( | ^~~~~~~~~~ /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:333:11: error: too many arguments to function ‘PyCode_New’ 333 | PyCode_New(a, 0, k, l, s, f, code, c, n, v, fv, cell, fn, name, fline, lnos) | ^~~~~~~~~~ /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:7241:33: note: in expansion of macro ‘__Pyx_PyCode_New’ 7241 | __pyx_codeobj__6 = (PyObject*)__Pyx_PyCode_New(1, 0, 1, 0, CO_OPTIMIZED|CO_NEWLOCALS, __pyx_empty_bytes, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_tuple__5, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_kp_s_Cython_Plex_Scanners_py, __pyx_n_s_read, 113, __pyx_empty_bytes); if (unlikely(!__pyx_codeobj__6)) __PYX_ERR(0, 113, __pyx_L1_error) | ^~~~~~~~~~~~~~~~ In file included from /usr/include/python3.8/compile.h:5, from /usr/include/python3.8/Python.h:138, from /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:19: /usr/include/python3.8/code.h:122:28: note: declared here 122 | PyAPI_FUNC(PyCodeObject *) PyCode_New( | ^~~~~~~~~~ /usr/include/python3.8/code.h:72:25: warning: passing argument 6 of ‘PyCode_New’ makes pointer from integer without a cast [-Wint-conversion] 72 | #define CO_OPTIMIZED 0x0001 | ^ | | | int /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:333:37: note: in definition of macro ‘__Pyx_PyCode_New’ 333 | PyCode_New(a, 0, k, l, s, f, code, c, n, v, fv, cell, fn, name, fline, lnos) | ^ /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:7253:62: note: in expansion of macro ‘CO_OPTIMIZED’ 7253 | __pyx_codeobj__8 = (PyObject*)__Pyx_PyCode_New(1, 0, 1, 0, CO_OPTIMIZED|CO_NEWLOCALS, __pyx_empty_bytes, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_tuple__7, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_kp_s_Cython_Plex_Scanners_py, __pyx_n_s_position, 297, __pyx_empty_bytes); if (unlikely(!__pyx_codeobj__8)) __PYX_ERR(0, 297, __pyx_L1_error) | ^~~~~~~~~~~~ In file included from /usr/include/python3.8/compile.h:5, from /usr/include/python3.8/Python.h:138, from /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:19: /usr/include/python3.8/code.h:122:28: note: expected ‘PyObject *’ {aka ‘struct _object *’} but argument is of type ‘int’ 122 | PyAPI_FUNC(PyCodeObject *) PyCode_New( | ^~~~~~~~~~ /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:7253:236: warning: passing argument 14 of ‘PyCode_New’ makes integer from pointer without a cast [-Wint-conversion] 7253 | __pyx_codeobj__8 = (PyObject*)__Pyx_PyCode_New(1, 0, 1, 0, CO_OPTIMIZED|CO_NEWLOCALS, __pyx_empty_bytes, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_tuple__7, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_kp_s_Cython_Plex_Scanners_py, __pyx_n_s_position, 297, __pyx_empty_bytes); if (unlikely(!__pyx_codeobj__8)) __PYX_ERR(0, 297, __pyx_L1_error) | ^~~~~~~~~~~~~~~~~~ | | | PyObject * {aka struct _object *} /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:333:69: note: in definition of macro ‘__Pyx_PyCode_New’ 333 | PyCode_New(a, 0, k, l, s, f, code, c, n, v, fv, cell, fn, name, fline, lnos) | ^~~~ In file included from /usr/include/python3.8/compile.h:5, from /usr/include/python3.8/Python.h:138, from /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:19: /usr/include/python3.8/code.h:122:28: note: expected ‘int’ but argument is of type ‘PyObject *’ {aka ‘struct _object *’} 122 | PyAPI_FUNC(PyCodeObject *) PyCode_New( | ^~~~~~~~~~ /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:7253:256: warning: passing argument 15 of ‘PyCode_New’ makes pointer from integer without a cast [-Wint-conversion] 7253 | __pyx_codeobj__8 = (PyObject*)__Pyx_PyCode_New(1, 0, 1, 0, CO_OPTIMIZED|CO_NEWLOCALS, __pyx_empty_bytes, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_tuple__7, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_kp_s_Cython_Plex_Scanners_py, __pyx_n_s_position, 297, __pyx_empty_bytes); if (unlikely(!__pyx_codeobj__8)) __PYX_ERR(0, 297, __pyx_L1_error) | ^~~ | | | int /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:333:75: note: in definition of macro ‘__Pyx_PyCode_New’ 333 | PyCode_New(a, 0, k, l, s, f, code, c, n, v, fv, cell, fn, name, fline, lnos) | ^~~~~ In file included from /usr/include/python3.8/compile.h:5, from /usr/include/python3.8/Python.h:138, from /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:19: /usr/include/python3.8/code.h:122:28: note: expected ‘PyObject *’ {aka ‘struct _object *’} but argument is of type ‘int’ 122 | PyAPI_FUNC(PyCodeObject *) PyCode_New( | ^~~~~~~~~~ /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:333:11: error: too many arguments to function ‘PyCode_New’ 333 | PyCode_New(a, 0, k, l, s, f, code, c, n, v, fv, cell, fn, name, fline, lnos) | ^~~~~~~~~~ /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:7253:33: note: in expansion of macro ‘__Pyx_PyCode_New’ 7253 | __pyx_codeobj__8 = (PyObject*)__Pyx_PyCode_New(1, 0, 1, 0, CO_OPTIMIZED|CO_NEWLOCALS, __pyx_empty_bytes, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_tuple__7, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_kp_s_Cython_Plex_Scanners_py, __pyx_n_s_position, 297, __pyx_empty_bytes); if (unlikely(!__pyx_codeobj__8)) __PYX_ERR(0, 297, __pyx_L1_error) | ^~~~~~~~~~~~~~~~ In file included from /usr/include/python3.8/compile.h:5, from /usr/include/python3.8/Python.h:138, from /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:19: /usr/include/python3.8/code.h:122:28: note: declared here 122 | PyAPI_FUNC(PyCodeObject *) PyCode_New( | ^~~~~~~~~~ /usr/include/python3.8/code.h:72:25: warning: passing argument 6 of ‘PyCode_New’ makes pointer from integer without a cast [-Wint-conversion] 72 | #define CO_OPTIMIZED 0x0001 | ^ | | | int /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:333:37: note: in definition of macro ‘__Pyx_PyCode_New’ 333 | PyCode_New(a, 0, k, l, s, f, code, c, n, v, fv, cell, fn, name, fline, lnos) | ^ /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:7265:63: note: in expansion of macro ‘CO_OPTIMIZED’ 7265 | __pyx_codeobj__10 = (PyObject*)__Pyx_PyCode_New(1, 0, 1, 0, CO_OPTIMIZED|CO_NEWLOCALS, __pyx_empty_bytes, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_tuple__9, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_kp_s_Cython_Plex_Scanners_py, __pyx_n_s_get_position, 308, __pyx_empty_bytes); if (unlikely(!__pyx_codeobj__10)) __PYX_ERR(0, 308, __pyx_L1_error) | ^~~~~~~~~~~~ In file included from /usr/include/python3.8/compile.h:5, from /usr/include/python3.8/Python.h:138, from /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:19: /usr/include/python3.8/code.h:122:28: note: expected ‘PyObject *’ {aka ‘struct _object *’} but argument is of type ‘int’ 122 | PyAPI_FUNC(PyCodeObject *) PyCode_New( | ^~~~~~~~~~ /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:7265:237: warning: passing argument 14 of ‘PyCode_New’ makes integer from pointer without a cast [-Wint-conversion] 7265 | __pyx_codeobj__10 = (PyObject*)__Pyx_PyCode_New(1, 0, 1, 0, CO_OPTIMIZED|CO_NEWLOCALS, __pyx_empty_bytes, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_tuple__9, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_kp_s_Cython_Plex_Scanners_py, __pyx_n_s_get_position, 308, __pyx_empty_bytes); if (unlikely(!__pyx_codeobj__10)) __PYX_ERR(0, 308, __pyx_L1_error) | ^~~~~~~~~~~~~~~~~~~~~~ | | | PyObject * {aka struct _object *} /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:333:69: note: in definition of macro ‘__Pyx_PyCode_New’ 333 | PyCode_New(a, 0, k, l, s, f, code, c, n, v, fv, cell, fn, name, fline, lnos) | ^~~~ In file included from /usr/include/python3.8/compile.h:5, from /usr/include/python3.8/Python.h:138, from /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:19: /usr/include/python3.8/code.h:122:28: note: expected ‘int’ but argument is of type ‘PyObject *’ {aka ‘struct _object *’} 122 | PyAPI_FUNC(PyCodeObject *) PyCode_New( | ^~~~~~~~~~ /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:7265:261: warning: passing argument 15 of ‘PyCode_New’ makes pointer from integer without a cast [-Wint-conversion] 7265 | __pyx_codeobj__10 = (PyObject*)__Pyx_PyCode_New(1, 0, 1, 0, CO_OPTIMIZED|CO_NEWLOCALS, __pyx_empty_bytes, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_tuple__9, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_kp_s_Cython_Plex_Scanners_py, __pyx_n_s_get_position, 308, __pyx_empty_bytes); if (unlikely(!__pyx_codeobj__10)) __PYX_ERR(0, 308, __pyx_L1_error) | ^~~ | | | int /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:333:75: note: in definition of macro ‘__Pyx_PyCode_New’ 333 | PyCode_New(a, 0, k, l, s, f, code, c, n, v, fv, cell, fn, name, fline, lnos) | ^~~~~ In file included from /usr/include/python3.8/compile.h:5, from /usr/include/python3.8/Python.h:138, from /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:19: /usr/include/python3.8/code.h:122:28: note: expected ‘PyObject *’ {aka ‘struct _object *’} but argument is of type ‘int’ 122 | PyAPI_FUNC(PyCodeObject *) PyCode_New( | ^~~~~~~~~~ /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:333:11: error: too many arguments to function ‘PyCode_New’ 333 | PyCode_New(a, 0, k, l, s, f, code, c, n, v, fv, cell, fn, name, fline, lnos) | ^~~~~~~~~~ /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:7265:34: note: in expansion of macro ‘__Pyx_PyCode_New’ 7265 | __pyx_codeobj__10 = (PyObject*)__Pyx_PyCode_New(1, 0, 1, 0, CO_OPTIMIZED|CO_NEWLOCALS, __pyx_empty_bytes, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_tuple__9, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_kp_s_Cython_Plex_Scanners_py, __pyx_n_s_get_position, 308, __pyx_empty_bytes); if (unlikely(!__pyx_codeobj__10)) __PYX_ERR(0, 308, __pyx_L1_error) | ^~~~~~~~~~~~~~~~ In file included from /usr/include/python3.8/compile.h:5, from /usr/include/python3.8/Python.h:138, from /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:19: /usr/include/python3.8/code.h:122:28: note: declared here 122 | PyAPI_FUNC(PyCodeObject *) PyCode_New( | ^~~~~~~~~~ /usr/include/python3.8/code.h:72:25: warning: passing argument 6 of ‘PyCode_New’ makes pointer from integer without a cast [-Wint-conversion] 72 | #define CO_OPTIMIZED 0x0001 | ^ | | | int /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:333:37: note: in definition of macro ‘__Pyx_PyCode_New’ 333 | PyCode_New(a, 0, k, l, s, f, code, c, n, v, fv, cell, fn, name, fline, lnos) | ^ /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:7277:63: note: in expansion of macro ‘CO_OPTIMIZED’ 7277 | __pyx_codeobj__12 = (PyObject*)__Pyx_PyCode_New(1, 0, 1, 0, CO_OPTIMIZED|CO_NEWLOCALS, __pyx_empty_bytes, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_tuple__11, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_kp_s_Cython_Plex_Scanners_py, __pyx_n_s_eof, 334, __pyx_empty_bytes); if (unlikely(!__pyx_codeobj__12)) __PYX_ERR(0, 334, __pyx_L1_error) | ^~~~~~~~~~~~ In file included from /usr/include/python3.8/compile.h:5, from /usr/include/python3.8/Python.h:138, from /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:19: /usr/include/python3.8/code.h:122:28: note: expected ‘PyObject *’ {aka ‘struct _object *’} but argument is of type ‘int’ 122 | PyAPI_FUNC(PyCodeObject *) PyCode_New( | ^~~~~~~~~~ /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:7277:238: warning: passing argument 14 of ‘PyCode_New’ makes integer from pointer without a cast [-Wint-conversion] 7277 | __pyx_codeobj__12 = (PyObject*)__Pyx_PyCode_New(1, 0, 1, 0, CO_OPTIMIZED|CO_NEWLOCALS, __pyx_empty_bytes, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_tuple__11, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_kp_s_Cython_Plex_Scanners_py, __pyx_n_s_eof, 334, __pyx_empty_bytes); if (unlikely(!__pyx_codeobj__12)) __PYX_ERR(0, 334, __pyx_L1_error) | ^~~~~~~~~~~~~ | | | PyObject * {aka struct _object *} /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:333:69: note: in definition of macro ‘__Pyx_PyCode_New’ 333 | PyCode_New(a, 0, k, l, s, f, code, c, n, v, fv, cell, fn, name, fline, lnos) | ^~~~ In file included from /usr/include/python3.8/compile.h:5, from /usr/include/python3.8/Python.h:138, from /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:19: /usr/include/python3.8/code.h:122:28: note: expected ‘int’ but argument is of type ‘PyObject *’ {aka ‘struct _object *’} 122 | PyAPI_FUNC(PyCodeObject *) PyCode_New( | ^~~~~~~~~~ /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:7277:253: warning: passing argument 15 of ‘PyCode_New’ makes pointer from integer without a cast [-Wint-conversion] 7277 | __pyx_codeobj__12 = (PyObject*)__Pyx_PyCode_New(1, 0, 1, 0, CO_OPTIMIZED|CO_NEWLOCALS, __pyx_empty_bytes, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_tuple__11, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_kp_s_Cython_Plex_Scanners_py, __pyx_n_s_eof, 334, __pyx_empty_bytes); if (unlikely(!__pyx_codeobj__12)) __PYX_ERR(0, 334, __pyx_L1_error) | ^~~ | | | int /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:333:75: note: in definition of macro ‘__Pyx_PyCode_New’ 333 | PyCode_New(a, 0, k, l, s, f, code, c, n, v, fv, cell, fn, name, fline, lnos) | ^~~~~ In file included from /usr/include/python3.8/compile.h:5, from /usr/include/python3.8/Python.h:138, from /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:19: /usr/include/python3.8/code.h:122:28: note: expected ‘PyObject *’ {aka ‘struct _object *’} but argument is of type ‘int’ 122 | PyAPI_FUNC(PyCodeObject *) PyCode_New( | ^~~~~~~~~~ /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:333:11: error: too many arguments to function ‘PyCode_New’ 333 | PyCode_New(a, 0, k, l, s, f, code, c, n, v, fv, cell, fn, name, fline, lnos) | ^~~~~~~~~~ /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:7277:34: note: in expansion of macro ‘__Pyx_PyCode_New’ 7277 | __pyx_codeobj__12 = (PyObject*)__Pyx_PyCode_New(1, 0, 1, 0, CO_OPTIMIZED|CO_NEWLOCALS, __pyx_empty_bytes, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_tuple__11, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_kp_s_Cython_Plex_Scanners_py, __pyx_n_s_eof, 334, __pyx_empty_bytes); if (unlikely(!__pyx_codeobj__12)) __PYX_ERR(0, 334, __pyx_L1_error) | ^~~~~~~~~~~~~~~~ In file included from /usr/include/python3.8/compile.h:5, from /usr/include/python3.8/Python.h:138, from /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:19: /usr/include/python3.8/code.h:122:28: note: declared here 122 | PyAPI_FUNC(PyCodeObject *) PyCode_New( | ^~~~~~~~~~ /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c: In function ‘__Pyx_CreateCodeObjectForTraceback’: /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:10051:9: warning: passing argument 14 of ‘PyCode_New’ makes integer from pointer without a cast [-Wint-conversion] 10051 | py_funcname, /*PyObject *name,*/ | ^~~~~~~~~~~ | | | PyObject * {aka struct _object *} /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:333:69: note: in definition of macro ‘__Pyx_PyCode_New’ 333 | PyCode_New(a, 0, k, l, s, f, code, c, n, v, fv, cell, fn, name, fline, lnos) | ^~~~ In file included from /usr/include/python3.8/compile.h:5, from /usr/include/python3.8/Python.h:138, from /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:19: /usr/include/python3.8/code.h:122:28: note: expected ‘int’ but argument is of type ‘PyObject *’ {aka ‘struct _object *’} 122 | PyAPI_FUNC(PyCodeObject *) PyCode_New( | ^~~~~~~~~~ /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:10052:9: warning: passing argument 15 of ‘PyCode_New’ makes pointer from integer without a cast [-Wint-conversion] 10052 | py_line, | ^~~~~~~ | | | int /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:333:75: note: in definition of macro ‘__Pyx_PyCode_New’ 333 | PyCode_New(a, 0, k, l, s, f, code, c, n, v, fv, cell, fn, name, fline, lnos) | ^~~~~ In file included from /usr/include/python3.8/compile.h:5, from /usr/include/python3.8/Python.h:138, from /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:19: /usr/include/python3.8/code.h:122:28: note: expected ‘PyObject *’ {aka ‘struct _object *’} but argument is of type ‘int’ 122 | PyAPI_FUNC(PyCodeObject *) PyCode_New( | ^~~~~~~~~~ /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:333:11: error: too many arguments to function ‘PyCode_New’ 333 | PyCode_New(a, 0, k, l, s, f, code, c, n, v, fv, cell, fn, name, fline, lnos) | ^~~~~~~~~~ /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:10038:15: note: in expansion of macro ‘__Pyx_PyCode_New’ 10038 | py_code = __Pyx_PyCode_New( | ^~~~~~~~~~~~~~~~ In file included from /usr/include/python3.8/compile.h:5, from /usr/include/python3.8/Python.h:138, from /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:19: /usr/include/python3.8/code.h:122:28: note: declared here 122 | PyAPI_FUNC(PyCodeObject *) PyCode_New( | ^~~~~~~~~~ error: command 'x86_64-linux-gnu-gcc' failed with exit status 1 ---------------------------------------- ERROR: Failed building wheel for cython ERROR: Failed to build one or more wheels Traceback (most recent call last): File "/root/.local/lib/python3.8/site-packages/setuptools/installer.py", line 128, in fetch_build_egg subprocess.check_call(cmd) File "/usr/lib/python3.8/subprocess.py", line 364, in check_call raise CalledProcessError(retcode, cmd) subprocess.CalledProcessError: Command '['/usr/bin/python3', '-m', 'pip', '--disable-pip-version-check', 'wheel', '--no-deps', '-w', '/tmp/tmpvll9vibn', '--quiet', '--find-links', 'https://github.com/kivy-garden/garden/archive/master.zip', 'cython!=0.27,!=0.27.2,<=0.29.10,>=0.24']' returned non-zero exit status 1. During handling of the above exception, another exception occurred: Traceback (most recent call last): File "<string>", line 1, in <module> File "/tmp/pip-install-b1phyb9o/kivy/setup.py", line 1073, in <module> setup( File "/root/.local/lib/python3.8/site-packages/setuptools/__init__.py", line 143, in setup _install_setup_requires(attrs) File "/root/.local/lib/python3.8/site-packages/setuptools/__init__.py", line 138, in _install_setup_requires dist.fetch_build_eggs(dist.setup_requires) File "/root/.local/lib/python3.8/site-packages/setuptools/dist.py", line 695, in fetch_build_eggs resolved_dists = pkg_resources.working_set.resolve( File "/root/.local/lib/python3.8/site-packages/pkg_resources/__init__.py", line 781, in resolve dist = best[req.key] = env.best_match( File "/root/.local/lib/python3.8/site-packages/pkg_resources/__init__.py", line 1066, in best_match return self.obtain(req, installer) File "/root/.local/lib/python3.8/site-packages/pkg_resources/__init__.py", line 1078, in obtain return installer(requirement) File "/root/.local/lib/python3.8/site-packages/setuptools/dist.py", line 754, in fetch_build_egg return fetch_build_egg(self, req) File "/root/.local/lib/python3.8/site-packages/setuptools/installer.py", line 130, in fetch_build_egg raise DistutilsError(str(e)) distutils.errors.DistutilsError: Command '['/usr/bin/python3', '-m', 'pip', '--disable-pip-version-check', 'wheel', '--no-deps', '-w', '/tmp/tmpvll9vibn', '--quiet', '--find-links', 'https://github.com/kivy-garden/garden/archive/master.zip', 'cython!=0.27,!=0.27.2,<=0.29.10,>=0.24']' returned non-zero exit status 1. Using setuptools Environ change use_x11 -> True Using this graphics system: OpenGL GStreamer found via pkg-config SDL2 found via pkg-config SDL2: found SDL header at /usr/include/SDL2/SDL.h SDL2: found SDL_mixer header at /usr/include/SDL2/SDL_mixer.h SDL2: found SDL_ttf header at /usr/include/SDL2/SDL_ttf.h SDL2: found SDL_image header at /usr/include/SDL2/SDL_image.h Pango: pangoft2 found via pkg-config {'libraries': ['pangoft2-1.0', 'pango-1.0', 'gobject-2.0', 'glib-2.0', 'harfbuzz', 'fontconfig', 'freetype'], 'include_dirs': ['/tmp/pip-install-b1phyb9o/kivy/kivy/include', '/usr/include/pango-1.0', '/usr/include/fribidi', '/usr/include/cairo', '/usr/include/pixman-1', '/usr/include/harfbuzz', '/usr/include/glib-2.0', '/usr/lib/x86_64-linux-gnu/glib-2.0/include', '/usr/include/uuid', '/usr/include/freetype2', '/usr/include/libpng16'], 'library_dirs': [], 'extra_link_args': [], 'extra_compile_args': [], 'depends': ['lib/pango/pangoft2.pxi', 'lib/pango/pangoft2.h']} ERROR: Dependency for context.pyx not resolved: config.pxi ERROR: Dependency for compiler.pyx not resolved: config.pxi ERROR: Dependency for context_instructions.pyx not resolved: config.pxi ERROR: Dependency for fbo.pyx not resolved: config.pxi ERROR: Dependency for gl_instructions.pyx not resolved: config.pxi ERROR: Dependency for instructions.pyx not resolved: config.pxi ERROR: Dependency for opengl.pyx not resolved: config.pxi ERROR: Dependency for opengl_utils.pyx not resolved: config.pxi ERROR: Dependency for shader.pyx not resolved: config.pxi ERROR: Dependency for stencil_instructions.pyx not resolved: config.pxi ERROR: Dependency for scissor_instructions.pyx not resolved: config.pxi ERROR: Dependency for texture.pyx not resolved: config.pxi ERROR: Dependency for vbo.pyx not resolved: config.pxi ERROR: Dependency for vertex.pyx not resolved: config.pxi ERROR: Dependency for vertex_instructions.pyx not resolved: config.pxi ERROR: Dependency for cgl.pyx not resolved: config.pxi ERROR: Dependency for cgl_mock.pyx not resolved: config.pxi ERROR: Dependency for cgl_gl.pyx not resolved: config.pxi ERROR: Dependency for cgl_glew.pyx not resolved: config.pxi ERROR: Dependency for cgl_sdl2.pyx not resolved: config.pxi ERROR: Dependency for svg.pyx not resolved: config.pxi ---------------------------------------- ERROR: Command errored out with exit status 1: python setup.py egg_info Check the logs for full command output.
True
install failing on kubuntu 20.04 - some errors: ` ERROR: Command errored out with exit status 1: command: /usr/bin/python3 -c 'import sys, setuptools, tokenize; sys.argv[0] = '"'"'/tmp/pip-install-b1phyb9o/kivy/setup.py'"'"'; __file__='"'"'/tmp/pip-install-b1phyb9o/kivy/setup.py'"'"';f=getattr(tokenize, '"'"'open'"'"', open)(__file__);code=f.read().replace('"'"'\r\n'"'"', '"'"'\n'"'"');f.close();exec(compile(code, __file__, '"'"'exec'"'"'))' egg_info --egg-base /tmp/pip-pip-egg-info-wmee5ot5 cwd: /tmp/pip-install-b1phyb9o/kivy/ Complete output (668 lines): fatal: not a git repository (or any of the parent directories): .git ERROR: Command errored out with exit status 1: command: /usr/bin/python3 -u -c 'import sys, setuptools, tokenize; sys.argv[0] = '"'"'/tmp/pip-wheel-52mhf973/cython/setup.py'"'"'; __file__='"'"'/tmp/pip-wheel-52mhf973/cython/setup.py'"'"';f=getattr(tokenize, '"'"'open'"'"', open)(__file__);code=f.read().replace('"'"'\r\n'"'"', '"'"'\n'"'"');f.close();exec(compile(code, __file__, '"'"'exec'"'"'))' bdist_wheel -d /tmp/pip-wheel-ugtpsi0m cwd: /tmp/pip-wheel-52mhf973/cython/ Complete output (598 lines): Unable to find pgen, not compiling formal grammar. ` creating build/temp.linux-x86_64-3.8/tmp/pip-wheel-52mhf973/cython/Cython/Plex x86_64-linux-gnu-gcc -pthread -Wno-unused-result -Wsign-compare -DNDEBUG -g -fwrapv -O2 -Wall -g -fstack-protector-strong -Wformat -Werror=format-security -g -fwrapv -O2 -g -fstack-protector-strong -Wformat -Werror=format-security -Wdate-time -D_FORTIFY_SOURCE=2 -fPIC -I/usr/include/python3.8 -c /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c -o build/temp.linux-x86_64-3.8/tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.o /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c: In function ‘__Pyx_InitCachedConstants’: /usr/include/python3.8/code.h:72:25: warning: passing argument 6 of ‘PyCode_New’ makes pointer from integer without a cast [-Wint-conversion] 72 | #define CO_OPTIMIZED 0x0001 | ^ | | | int /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:333:37: note: in definition of macro ‘__Pyx_PyCode_New’ 333 | PyCode_New(a, 0, k, l, s, f, code, c, n, v, fv, cell, fn, name, fline, lnos) | ^ /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:7241:62: note: in expansion of macro ‘CO_OPTIMIZED’ 7241 | __pyx_codeobj__6 = (PyObject*)__Pyx_PyCode_New(1, 0, 1, 0, CO_OPTIMIZED|CO_NEWLOCALS, __pyx_empty_bytes, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_tuple__5, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_kp_s_Cython_Plex_Scanners_py, __pyx_n_s_read, 113, __pyx_empty_bytes); if (unlikely(!__pyx_codeobj__6)) __PYX_ERR(0, 113, __pyx_L1_error) | ^~~~~~~~~~~~ In file included from /usr/include/python3.8/compile.h:5, from /usr/include/python3.8/Python.h:138, from /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:19: /usr/include/python3.8/code.h:122:28: note: expected ‘PyObject *’ {aka ‘struct _object *’} but argument is of type ‘int’ 122 | PyAPI_FUNC(PyCodeObject *) PyCode_New( | ^~~~~~~~~~ /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:7241:236: warning: passing argument 14 of ‘PyCode_New’ makes integer from pointer without a cast [-Wint-conversion] 7241 | __pyx_codeobj__6 = (PyObject*)__Pyx_PyCode_New(1, 0, 1, 0, CO_OPTIMIZED|CO_NEWLOCALS, __pyx_empty_bytes, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_tuple__5, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_kp_s_Cython_Plex_Scanners_py, __pyx_n_s_read, 113, __pyx_empty_bytes); if (unlikely(!__pyx_codeobj__6)) __PYX_ERR(0, 113, __pyx_L1_error) | ^~~~~~~~~~~~~~ | | | PyObject * {aka struct _object *} /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:333:69: note: in definition of macro ‘__Pyx_PyCode_New’ 333 | PyCode_New(a, 0, k, l, s, f, code, c, n, v, fv, cell, fn, name, fline, lnos) | ^~~~ In file included from /usr/include/python3.8/compile.h:5, from /usr/include/python3.8/Python.h:138, from /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:19: /usr/include/python3.8/code.h:122:28: note: expected ‘int’ but argument is of type ‘PyObject *’ {aka ‘struct _object *’} 122 | PyAPI_FUNC(PyCodeObject *) PyCode_New( | ^~~~~~~~~~ /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:7241:252: warning: passing argument 15 of ‘PyCode_New’ makes pointer from integer without a cast [-Wint-conversion] 7241 | __pyx_codeobj__6 = (PyObject*)__Pyx_PyCode_New(1, 0, 1, 0, CO_OPTIMIZED|CO_NEWLOCALS, __pyx_empty_bytes, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_tuple__5, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_kp_s_Cython_Plex_Scanners_py, __pyx_n_s_read, 113, __pyx_empty_bytes); if (unlikely(!__pyx_codeobj__6)) __PYX_ERR(0, 113, __pyx_L1_error) | ^~~ | | | int /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:333:75: note: in definition of macro ‘__Pyx_PyCode_New’ 333 | PyCode_New(a, 0, k, l, s, f, code, c, n, v, fv, cell, fn, name, fline, lnos) | ^~~~~ In file included from /usr/include/python3.8/compile.h:5, from /usr/include/python3.8/Python.h:138, from /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:19: /usr/include/python3.8/code.h:122:28: note: expected ‘PyObject *’ {aka ‘struct _object *’} but argument is of type ‘int’ 122 | PyAPI_FUNC(PyCodeObject *) PyCode_New( | ^~~~~~~~~~ /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:333:11: error: too many arguments to function ‘PyCode_New’ 333 | PyCode_New(a, 0, k, l, s, f, code, c, n, v, fv, cell, fn, name, fline, lnos) | ^~~~~~~~~~ /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:7241:33: note: in expansion of macro ‘__Pyx_PyCode_New’ 7241 | __pyx_codeobj__6 = (PyObject*)__Pyx_PyCode_New(1, 0, 1, 0, CO_OPTIMIZED|CO_NEWLOCALS, __pyx_empty_bytes, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_tuple__5, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_kp_s_Cython_Plex_Scanners_py, __pyx_n_s_read, 113, __pyx_empty_bytes); if (unlikely(!__pyx_codeobj__6)) __PYX_ERR(0, 113, __pyx_L1_error) | ^~~~~~~~~~~~~~~~ In file included from /usr/include/python3.8/compile.h:5, from /usr/include/python3.8/Python.h:138, from /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:19: /usr/include/python3.8/code.h:122:28: note: declared here 122 | PyAPI_FUNC(PyCodeObject *) PyCode_New( | ^~~~~~~~~~ /usr/include/python3.8/code.h:72:25: warning: passing argument 6 of ‘PyCode_New’ makes pointer from integer without a cast [-Wint-conversion] 72 | #define CO_OPTIMIZED 0x0001 | ^ | | | int /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:333:37: note: in definition of macro ‘__Pyx_PyCode_New’ 333 | PyCode_New(a, 0, k, l, s, f, code, c, n, v, fv, cell, fn, name, fline, lnos) | ^ /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:7253:62: note: in expansion of macro ‘CO_OPTIMIZED’ 7253 | __pyx_codeobj__8 = (PyObject*)__Pyx_PyCode_New(1, 0, 1, 0, CO_OPTIMIZED|CO_NEWLOCALS, __pyx_empty_bytes, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_tuple__7, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_kp_s_Cython_Plex_Scanners_py, __pyx_n_s_position, 297, __pyx_empty_bytes); if (unlikely(!__pyx_codeobj__8)) __PYX_ERR(0, 297, __pyx_L1_error) | ^~~~~~~~~~~~ In file included from /usr/include/python3.8/compile.h:5, from /usr/include/python3.8/Python.h:138, from /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:19: /usr/include/python3.8/code.h:122:28: note: expected ‘PyObject *’ {aka ‘struct _object *’} but argument is of type ‘int’ 122 | PyAPI_FUNC(PyCodeObject *) PyCode_New( | ^~~~~~~~~~ /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:7253:236: warning: passing argument 14 of ‘PyCode_New’ makes integer from pointer without a cast [-Wint-conversion] 7253 | __pyx_codeobj__8 = (PyObject*)__Pyx_PyCode_New(1, 0, 1, 0, CO_OPTIMIZED|CO_NEWLOCALS, __pyx_empty_bytes, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_tuple__7, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_kp_s_Cython_Plex_Scanners_py, __pyx_n_s_position, 297, __pyx_empty_bytes); if (unlikely(!__pyx_codeobj__8)) __PYX_ERR(0, 297, __pyx_L1_error) | ^~~~~~~~~~~~~~~~~~ | | | PyObject * {aka struct _object *} /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:333:69: note: in definition of macro ‘__Pyx_PyCode_New’ 333 | PyCode_New(a, 0, k, l, s, f, code, c, n, v, fv, cell, fn, name, fline, lnos) | ^~~~ In file included from /usr/include/python3.8/compile.h:5, from /usr/include/python3.8/Python.h:138, from /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:19: /usr/include/python3.8/code.h:122:28: note: expected ‘int’ but argument is of type ‘PyObject *’ {aka ‘struct _object *’} 122 | PyAPI_FUNC(PyCodeObject *) PyCode_New( | ^~~~~~~~~~ /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:7253:256: warning: passing argument 15 of ‘PyCode_New’ makes pointer from integer without a cast [-Wint-conversion] 7253 | __pyx_codeobj__8 = (PyObject*)__Pyx_PyCode_New(1, 0, 1, 0, CO_OPTIMIZED|CO_NEWLOCALS, __pyx_empty_bytes, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_tuple__7, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_kp_s_Cython_Plex_Scanners_py, __pyx_n_s_position, 297, __pyx_empty_bytes); if (unlikely(!__pyx_codeobj__8)) __PYX_ERR(0, 297, __pyx_L1_error) | ^~~ | | | int /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:333:75: note: in definition of macro ‘__Pyx_PyCode_New’ 333 | PyCode_New(a, 0, k, l, s, f, code, c, n, v, fv, cell, fn, name, fline, lnos) | ^~~~~ In file included from /usr/include/python3.8/compile.h:5, from /usr/include/python3.8/Python.h:138, from /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:19: /usr/include/python3.8/code.h:122:28: note: expected ‘PyObject *’ {aka ‘struct _object *’} but argument is of type ‘int’ 122 | PyAPI_FUNC(PyCodeObject *) PyCode_New( | ^~~~~~~~~~ /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:333:11: error: too many arguments to function ‘PyCode_New’ 333 | PyCode_New(a, 0, k, l, s, f, code, c, n, v, fv, cell, fn, name, fline, lnos) | ^~~~~~~~~~ /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:7253:33: note: in expansion of macro ‘__Pyx_PyCode_New’ 7253 | __pyx_codeobj__8 = (PyObject*)__Pyx_PyCode_New(1, 0, 1, 0, CO_OPTIMIZED|CO_NEWLOCALS, __pyx_empty_bytes, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_tuple__7, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_kp_s_Cython_Plex_Scanners_py, __pyx_n_s_position, 297, __pyx_empty_bytes); if (unlikely(!__pyx_codeobj__8)) __PYX_ERR(0, 297, __pyx_L1_error) | ^~~~~~~~~~~~~~~~ In file included from /usr/include/python3.8/compile.h:5, from /usr/include/python3.8/Python.h:138, from /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:19: /usr/include/python3.8/code.h:122:28: note: declared here 122 | PyAPI_FUNC(PyCodeObject *) PyCode_New( | ^~~~~~~~~~ /usr/include/python3.8/code.h:72:25: warning: passing argument 6 of ‘PyCode_New’ makes pointer from integer without a cast [-Wint-conversion] 72 | #define CO_OPTIMIZED 0x0001 | ^ | | | int /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:333:37: note: in definition of macro ‘__Pyx_PyCode_New’ 333 | PyCode_New(a, 0, k, l, s, f, code, c, n, v, fv, cell, fn, name, fline, lnos) | ^ /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:7265:63: note: in expansion of macro ‘CO_OPTIMIZED’ 7265 | __pyx_codeobj__10 = (PyObject*)__Pyx_PyCode_New(1, 0, 1, 0, CO_OPTIMIZED|CO_NEWLOCALS, __pyx_empty_bytes, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_tuple__9, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_kp_s_Cython_Plex_Scanners_py, __pyx_n_s_get_position, 308, __pyx_empty_bytes); if (unlikely(!__pyx_codeobj__10)) __PYX_ERR(0, 308, __pyx_L1_error) | ^~~~~~~~~~~~ In file included from /usr/include/python3.8/compile.h:5, from /usr/include/python3.8/Python.h:138, from /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:19: /usr/include/python3.8/code.h:122:28: note: expected ‘PyObject *’ {aka ‘struct _object *’} but argument is of type ‘int’ 122 | PyAPI_FUNC(PyCodeObject *) PyCode_New( | ^~~~~~~~~~ /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:7265:237: warning: passing argument 14 of ‘PyCode_New’ makes integer from pointer without a cast [-Wint-conversion] 7265 | __pyx_codeobj__10 = (PyObject*)__Pyx_PyCode_New(1, 0, 1, 0, CO_OPTIMIZED|CO_NEWLOCALS, __pyx_empty_bytes, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_tuple__9, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_kp_s_Cython_Plex_Scanners_py, __pyx_n_s_get_position, 308, __pyx_empty_bytes); if (unlikely(!__pyx_codeobj__10)) __PYX_ERR(0, 308, __pyx_L1_error) | ^~~~~~~~~~~~~~~~~~~~~~ | | | PyObject * {aka struct _object *} /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:333:69: note: in definition of macro ‘__Pyx_PyCode_New’ 333 | PyCode_New(a, 0, k, l, s, f, code, c, n, v, fv, cell, fn, name, fline, lnos) | ^~~~ In file included from /usr/include/python3.8/compile.h:5, from /usr/include/python3.8/Python.h:138, from /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:19: /usr/include/python3.8/code.h:122:28: note: expected ‘int’ but argument is of type ‘PyObject *’ {aka ‘struct _object *’} 122 | PyAPI_FUNC(PyCodeObject *) PyCode_New( | ^~~~~~~~~~ /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:7265:261: warning: passing argument 15 of ‘PyCode_New’ makes pointer from integer without a cast [-Wint-conversion] 7265 | __pyx_codeobj__10 = (PyObject*)__Pyx_PyCode_New(1, 0, 1, 0, CO_OPTIMIZED|CO_NEWLOCALS, __pyx_empty_bytes, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_tuple__9, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_kp_s_Cython_Plex_Scanners_py, __pyx_n_s_get_position, 308, __pyx_empty_bytes); if (unlikely(!__pyx_codeobj__10)) __PYX_ERR(0, 308, __pyx_L1_error) | ^~~ | | | int /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:333:75: note: in definition of macro ‘__Pyx_PyCode_New’ 333 | PyCode_New(a, 0, k, l, s, f, code, c, n, v, fv, cell, fn, name, fline, lnos) | ^~~~~ In file included from /usr/include/python3.8/compile.h:5, from /usr/include/python3.8/Python.h:138, from /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:19: /usr/include/python3.8/code.h:122:28: note: expected ‘PyObject *’ {aka ‘struct _object *’} but argument is of type ‘int’ 122 | PyAPI_FUNC(PyCodeObject *) PyCode_New( | ^~~~~~~~~~ /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:333:11: error: too many arguments to function ‘PyCode_New’ 333 | PyCode_New(a, 0, k, l, s, f, code, c, n, v, fv, cell, fn, name, fline, lnos) | ^~~~~~~~~~ /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:7265:34: note: in expansion of macro ‘__Pyx_PyCode_New’ 7265 | __pyx_codeobj__10 = (PyObject*)__Pyx_PyCode_New(1, 0, 1, 0, CO_OPTIMIZED|CO_NEWLOCALS, __pyx_empty_bytes, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_tuple__9, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_kp_s_Cython_Plex_Scanners_py, __pyx_n_s_get_position, 308, __pyx_empty_bytes); if (unlikely(!__pyx_codeobj__10)) __PYX_ERR(0, 308, __pyx_L1_error) | ^~~~~~~~~~~~~~~~ In file included from /usr/include/python3.8/compile.h:5, from /usr/include/python3.8/Python.h:138, from /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:19: /usr/include/python3.8/code.h:122:28: note: declared here 122 | PyAPI_FUNC(PyCodeObject *) PyCode_New( | ^~~~~~~~~~ /usr/include/python3.8/code.h:72:25: warning: passing argument 6 of ‘PyCode_New’ makes pointer from integer without a cast [-Wint-conversion] 72 | #define CO_OPTIMIZED 0x0001 | ^ | | | int /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:333:37: note: in definition of macro ‘__Pyx_PyCode_New’ 333 | PyCode_New(a, 0, k, l, s, f, code, c, n, v, fv, cell, fn, name, fline, lnos) | ^ /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:7277:63: note: in expansion of macro ‘CO_OPTIMIZED’ 7277 | __pyx_codeobj__12 = (PyObject*)__Pyx_PyCode_New(1, 0, 1, 0, CO_OPTIMIZED|CO_NEWLOCALS, __pyx_empty_bytes, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_tuple__11, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_kp_s_Cython_Plex_Scanners_py, __pyx_n_s_eof, 334, __pyx_empty_bytes); if (unlikely(!__pyx_codeobj__12)) __PYX_ERR(0, 334, __pyx_L1_error) | ^~~~~~~~~~~~ In file included from /usr/include/python3.8/compile.h:5, from /usr/include/python3.8/Python.h:138, from /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:19: /usr/include/python3.8/code.h:122:28: note: expected ‘PyObject *’ {aka ‘struct _object *’} but argument is of type ‘int’ 122 | PyAPI_FUNC(PyCodeObject *) PyCode_New( | ^~~~~~~~~~ /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:7277:238: warning: passing argument 14 of ‘PyCode_New’ makes integer from pointer without a cast [-Wint-conversion] 7277 | __pyx_codeobj__12 = (PyObject*)__Pyx_PyCode_New(1, 0, 1, 0, CO_OPTIMIZED|CO_NEWLOCALS, __pyx_empty_bytes, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_tuple__11, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_kp_s_Cython_Plex_Scanners_py, __pyx_n_s_eof, 334, __pyx_empty_bytes); if (unlikely(!__pyx_codeobj__12)) __PYX_ERR(0, 334, __pyx_L1_error) | ^~~~~~~~~~~~~ | | | PyObject * {aka struct _object *} /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:333:69: note: in definition of macro ‘__Pyx_PyCode_New’ 333 | PyCode_New(a, 0, k, l, s, f, code, c, n, v, fv, cell, fn, name, fline, lnos) | ^~~~ In file included from /usr/include/python3.8/compile.h:5, from /usr/include/python3.8/Python.h:138, from /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:19: /usr/include/python3.8/code.h:122:28: note: expected ‘int’ but argument is of type ‘PyObject *’ {aka ‘struct _object *’} 122 | PyAPI_FUNC(PyCodeObject *) PyCode_New( | ^~~~~~~~~~ /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:7277:253: warning: passing argument 15 of ‘PyCode_New’ makes pointer from integer without a cast [-Wint-conversion] 7277 | __pyx_codeobj__12 = (PyObject*)__Pyx_PyCode_New(1, 0, 1, 0, CO_OPTIMIZED|CO_NEWLOCALS, __pyx_empty_bytes, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_tuple__11, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_kp_s_Cython_Plex_Scanners_py, __pyx_n_s_eof, 334, __pyx_empty_bytes); if (unlikely(!__pyx_codeobj__12)) __PYX_ERR(0, 334, __pyx_L1_error) | ^~~ | | | int /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:333:75: note: in definition of macro ‘__Pyx_PyCode_New’ 333 | PyCode_New(a, 0, k, l, s, f, code, c, n, v, fv, cell, fn, name, fline, lnos) | ^~~~~ In file included from /usr/include/python3.8/compile.h:5, from /usr/include/python3.8/Python.h:138, from /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:19: /usr/include/python3.8/code.h:122:28: note: expected ‘PyObject *’ {aka ‘struct _object *’} but argument is of type ‘int’ 122 | PyAPI_FUNC(PyCodeObject *) PyCode_New( | ^~~~~~~~~~ /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:333:11: error: too many arguments to function ‘PyCode_New’ 333 | PyCode_New(a, 0, k, l, s, f, code, c, n, v, fv, cell, fn, name, fline, lnos) | ^~~~~~~~~~ /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:7277:34: note: in expansion of macro ‘__Pyx_PyCode_New’ 7277 | __pyx_codeobj__12 = (PyObject*)__Pyx_PyCode_New(1, 0, 1, 0, CO_OPTIMIZED|CO_NEWLOCALS, __pyx_empty_bytes, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_tuple__11, __pyx_empty_tuple, __pyx_empty_tuple, __pyx_kp_s_Cython_Plex_Scanners_py, __pyx_n_s_eof, 334, __pyx_empty_bytes); if (unlikely(!__pyx_codeobj__12)) __PYX_ERR(0, 334, __pyx_L1_error) | ^~~~~~~~~~~~~~~~ In file included from /usr/include/python3.8/compile.h:5, from /usr/include/python3.8/Python.h:138, from /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:19: /usr/include/python3.8/code.h:122:28: note: declared here 122 | PyAPI_FUNC(PyCodeObject *) PyCode_New( | ^~~~~~~~~~ /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c: In function ‘__Pyx_CreateCodeObjectForTraceback’: /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:10051:9: warning: passing argument 14 of ‘PyCode_New’ makes integer from pointer without a cast [-Wint-conversion] 10051 | py_funcname, /*PyObject *name,*/ | ^~~~~~~~~~~ | | | PyObject * {aka struct _object *} /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:333:69: note: in definition of macro ‘__Pyx_PyCode_New’ 333 | PyCode_New(a, 0, k, l, s, f, code, c, n, v, fv, cell, fn, name, fline, lnos) | ^~~~ In file included from /usr/include/python3.8/compile.h:5, from /usr/include/python3.8/Python.h:138, from /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:19: /usr/include/python3.8/code.h:122:28: note: expected ‘int’ but argument is of type ‘PyObject *’ {aka ‘struct _object *’} 122 | PyAPI_FUNC(PyCodeObject *) PyCode_New( | ^~~~~~~~~~ /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:10052:9: warning: passing argument 15 of ‘PyCode_New’ makes pointer from integer without a cast [-Wint-conversion] 10052 | py_line, | ^~~~~~~ | | | int /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:333:75: note: in definition of macro ‘__Pyx_PyCode_New’ 333 | PyCode_New(a, 0, k, l, s, f, code, c, n, v, fv, cell, fn, name, fline, lnos) | ^~~~~ In file included from /usr/include/python3.8/compile.h:5, from /usr/include/python3.8/Python.h:138, from /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:19: /usr/include/python3.8/code.h:122:28: note: expected ‘PyObject *’ {aka ‘struct _object *’} but argument is of type ‘int’ 122 | PyAPI_FUNC(PyCodeObject *) PyCode_New( | ^~~~~~~~~~ /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:333:11: error: too many arguments to function ‘PyCode_New’ 333 | PyCode_New(a, 0, k, l, s, f, code, c, n, v, fv, cell, fn, name, fline, lnos) | ^~~~~~~~~~ /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:10038:15: note: in expansion of macro ‘__Pyx_PyCode_New’ 10038 | py_code = __Pyx_PyCode_New( | ^~~~~~~~~~~~~~~~ In file included from /usr/include/python3.8/compile.h:5, from /usr/include/python3.8/Python.h:138, from /tmp/pip-wheel-52mhf973/cython/Cython/Plex/Scanners.c:19: /usr/include/python3.8/code.h:122:28: note: declared here 122 | PyAPI_FUNC(PyCodeObject *) PyCode_New( | ^~~~~~~~~~ error: command 'x86_64-linux-gnu-gcc' failed with exit status 1 ---------------------------------------- ERROR: Failed building wheel for cython ERROR: Failed to build one or more wheels Traceback (most recent call last): File "/root/.local/lib/python3.8/site-packages/setuptools/installer.py", line 128, in fetch_build_egg subprocess.check_call(cmd) File "/usr/lib/python3.8/subprocess.py", line 364, in check_call raise CalledProcessError(retcode, cmd) subprocess.CalledProcessError: Command '['/usr/bin/python3', '-m', 'pip', '--disable-pip-version-check', 'wheel', '--no-deps', '-w', '/tmp/tmpvll9vibn', '--quiet', '--find-links', 'https://github.com/kivy-garden/garden/archive/master.zip', 'cython!=0.27,!=0.27.2,<=0.29.10,>=0.24']' returned non-zero exit status 1. During handling of the above exception, another exception occurred: Traceback (most recent call last): File "<string>", line 1, in <module> File "/tmp/pip-install-b1phyb9o/kivy/setup.py", line 1073, in <module> setup( File "/root/.local/lib/python3.8/site-packages/setuptools/__init__.py", line 143, in setup _install_setup_requires(attrs) File "/root/.local/lib/python3.8/site-packages/setuptools/__init__.py", line 138, in _install_setup_requires dist.fetch_build_eggs(dist.setup_requires) File "/root/.local/lib/python3.8/site-packages/setuptools/dist.py", line 695, in fetch_build_eggs resolved_dists = pkg_resources.working_set.resolve( File "/root/.local/lib/python3.8/site-packages/pkg_resources/__init__.py", line 781, in resolve dist = best[req.key] = env.best_match( File "/root/.local/lib/python3.8/site-packages/pkg_resources/__init__.py", line 1066, in best_match return self.obtain(req, installer) File "/root/.local/lib/python3.8/site-packages/pkg_resources/__init__.py", line 1078, in obtain return installer(requirement) File "/root/.local/lib/python3.8/site-packages/setuptools/dist.py", line 754, in fetch_build_egg return fetch_build_egg(self, req) File "/root/.local/lib/python3.8/site-packages/setuptools/installer.py", line 130, in fetch_build_egg raise DistutilsError(str(e)) distutils.errors.DistutilsError: Command '['/usr/bin/python3', '-m', 'pip', '--disable-pip-version-check', 'wheel', '--no-deps', '-w', '/tmp/tmpvll9vibn', '--quiet', '--find-links', 'https://github.com/kivy-garden/garden/archive/master.zip', 'cython!=0.27,!=0.27.2,<=0.29.10,>=0.24']' returned non-zero exit status 1. Using setuptools Environ change use_x11 -> True Using this graphics system: OpenGL GStreamer found via pkg-config SDL2 found via pkg-config SDL2: found SDL header at /usr/include/SDL2/SDL.h SDL2: found SDL_mixer header at /usr/include/SDL2/SDL_mixer.h SDL2: found SDL_ttf header at /usr/include/SDL2/SDL_ttf.h SDL2: found SDL_image header at /usr/include/SDL2/SDL_image.h Pango: pangoft2 found via pkg-config {'libraries': ['pangoft2-1.0', 'pango-1.0', 'gobject-2.0', 'glib-2.0', 'harfbuzz', 'fontconfig', 'freetype'], 'include_dirs': ['/tmp/pip-install-b1phyb9o/kivy/kivy/include', '/usr/include/pango-1.0', '/usr/include/fribidi', '/usr/include/cairo', '/usr/include/pixman-1', '/usr/include/harfbuzz', '/usr/include/glib-2.0', '/usr/lib/x86_64-linux-gnu/glib-2.0/include', '/usr/include/uuid', '/usr/include/freetype2', '/usr/include/libpng16'], 'library_dirs': [], 'extra_link_args': [], 'extra_compile_args': [], 'depends': ['lib/pango/pangoft2.pxi', 'lib/pango/pangoft2.h']} ERROR: Dependency for context.pyx not resolved: config.pxi ERROR: Dependency for compiler.pyx not resolved: config.pxi ERROR: Dependency for context_instructions.pyx not resolved: config.pxi ERROR: Dependency for fbo.pyx not resolved: config.pxi ERROR: Dependency for gl_instructions.pyx not resolved: config.pxi ERROR: Dependency for instructions.pyx not resolved: config.pxi ERROR: Dependency for opengl.pyx not resolved: config.pxi ERROR: Dependency for opengl_utils.pyx not resolved: config.pxi ERROR: Dependency for shader.pyx not resolved: config.pxi ERROR: Dependency for stencil_instructions.pyx not resolved: config.pxi ERROR: Dependency for scissor_instructions.pyx not resolved: config.pxi ERROR: Dependency for texture.pyx not resolved: config.pxi ERROR: Dependency for vbo.pyx not resolved: config.pxi ERROR: Dependency for vertex.pyx not resolved: config.pxi ERROR: Dependency for vertex_instructions.pyx not resolved: config.pxi ERROR: Dependency for cgl.pyx not resolved: config.pxi ERROR: Dependency for cgl_mock.pyx not resolved: config.pxi ERROR: Dependency for cgl_gl.pyx not resolved: config.pxi ERROR: Dependency for cgl_glew.pyx not resolved: config.pxi ERROR: Dependency for cgl_sdl2.pyx not resolved: config.pxi ERROR: Dependency for svg.pyx not resolved: config.pxi ---------------------------------------- ERROR: Command errored out with exit status 1: python setup.py egg_info Check the logs for full command output.
comp
install failing on kubuntu some errors error command errored out with exit status command usr bin c import sys setuptools tokenize sys argv tmp pip install kivy setup py file tmp pip install kivy setup py f getattr tokenize open open file code f read replace r n n f close exec compile code file exec egg info egg base tmp pip pip egg info cwd tmp pip install kivy complete output lines fatal not a git repository or any of the parent directories git error command errored out with exit status command usr bin u c import sys setuptools tokenize sys argv tmp pip wheel cython setup py file tmp pip wheel cython setup py f getattr tokenize open open file code f read replace r n n f close exec compile code file exec bdist wheel d tmp pip wheel cwd tmp pip wheel cython complete output lines unable to find pgen not compiling formal grammar creating build temp linux tmp pip wheel cython cython plex linux gnu gcc pthread wno unused result wsign compare dndebug g fwrapv wall g fstack protector strong wformat werror format security g fwrapv g fstack protector strong wformat werror format security wdate time d fortify source fpic i usr include c tmp pip wheel cython cython plex scanners c o build temp linux tmp pip wheel cython cython plex scanners o tmp pip wheel cython cython plex scanners c in function ‘ pyx initcachedconstants’ usr include code h warning passing argument of ‘pycode new’ makes pointer from integer without a cast define co optimized int tmp pip wheel cython cython plex scanners c note in definition of macro ‘ pyx pycode new’ pycode new a k l s f code c n v fv cell fn name fline lnos tmp pip wheel cython cython plex scanners c note in expansion of macro ‘co optimized’ pyx codeobj pyobject pyx pycode new co optimized co newlocals pyx empty bytes pyx empty tuple pyx empty tuple pyx tuple pyx empty tuple pyx empty tuple pyx kp s cython plex scanners py pyx n s read pyx empty bytes if unlikely pyx codeobj pyx err pyx error in file included from usr include compile h from usr include python h from tmp pip wheel cython cython plex scanners c usr include code h note expected ‘pyobject ’ aka ‘struct object ’ but argument is of type ‘int’ pyapi func pycodeobject pycode new tmp pip wheel cython cython plex scanners c warning passing argument of ‘pycode new’ makes integer from pointer without a cast pyx codeobj pyobject pyx pycode new co optimized co newlocals pyx empty bytes pyx empty tuple pyx empty tuple pyx tuple pyx empty tuple pyx empty tuple pyx kp s cython plex scanners py pyx n s read pyx empty bytes if unlikely pyx codeobj pyx err pyx error pyobject aka struct object tmp pip wheel cython cython plex scanners c note in definition of macro ‘ pyx pycode new’ pycode new a k l s f code c n v fv cell fn name fline lnos in file included from usr include compile h from usr include python h from tmp pip wheel cython cython plex scanners c usr include code h note expected ‘int’ but argument is of type ‘pyobject ’ aka ‘struct object ’ pyapi func pycodeobject pycode new tmp pip wheel cython cython plex scanners c warning passing argument of ‘pycode new’ makes pointer from integer without a cast pyx codeobj pyobject pyx pycode new co optimized co newlocals pyx empty bytes pyx empty tuple pyx empty tuple pyx tuple pyx empty tuple pyx empty tuple pyx kp s cython plex scanners py pyx n s read pyx empty bytes if unlikely pyx codeobj pyx err pyx error int tmp pip wheel cython cython plex scanners c note in definition of macro ‘ pyx pycode new’ pycode new a k l s f code c n v fv cell fn name fline lnos in file included from usr include compile h from usr include python h from tmp pip wheel cython cython plex scanners c usr include code h note expected ‘pyobject ’ aka ‘struct object ’ but argument is of type ‘int’ pyapi func pycodeobject pycode new tmp pip wheel cython cython plex scanners c error too many arguments to function ‘pycode new’ pycode new a k l s f code c n v fv cell fn name fline lnos tmp pip wheel cython cython plex scanners c note in expansion of macro ‘ pyx pycode new’ pyx codeobj pyobject pyx pycode new co optimized co newlocals pyx empty bytes pyx empty tuple pyx empty tuple pyx tuple pyx empty tuple pyx empty tuple pyx kp s cython plex scanners py pyx n s read pyx empty bytes if unlikely pyx codeobj pyx err pyx error in file included from usr include compile h from usr include python h from tmp pip wheel cython cython plex scanners c usr include code h note declared here pyapi func pycodeobject pycode new usr include code h warning passing argument of ‘pycode new’ makes pointer from integer without a cast define co optimized int tmp pip wheel cython cython plex scanners c note in definition of macro ‘ pyx pycode new’ pycode new a k l s f code c n v fv cell fn name fline lnos tmp pip wheel cython cython plex scanners c note in expansion of macro ‘co optimized’ pyx codeobj pyobject pyx pycode new co optimized co newlocals pyx empty bytes pyx empty tuple pyx empty tuple pyx tuple pyx empty tuple pyx empty tuple pyx kp s cython plex scanners py pyx n s position pyx empty bytes if unlikely pyx codeobj pyx err pyx error in file included from usr include compile h from usr include python h from tmp pip wheel cython cython plex scanners c usr include code h note expected ‘pyobject ’ aka ‘struct object ’ but argument is of type ‘int’ pyapi func pycodeobject pycode new tmp pip wheel cython cython plex scanners c warning passing argument of ‘pycode new’ makes integer from pointer without a cast pyx codeobj pyobject pyx pycode new co optimized co newlocals pyx empty bytes pyx empty tuple pyx empty tuple pyx tuple pyx empty tuple pyx empty tuple pyx kp s cython plex scanners py pyx n s position pyx empty bytes if unlikely pyx codeobj pyx err pyx error pyobject aka struct object tmp pip wheel cython cython plex scanners c note in definition of macro ‘ pyx pycode new’ pycode new a k l s f code c n v fv cell fn name fline lnos in file included from usr include compile h from usr include python h from tmp pip wheel cython cython plex scanners c usr include code h note expected ‘int’ but argument is of type ‘pyobject ’ aka ‘struct object ’ pyapi func pycodeobject pycode new tmp pip wheel cython cython plex scanners c warning passing argument of ‘pycode new’ makes pointer from integer without a cast pyx codeobj pyobject pyx pycode new co optimized co newlocals pyx empty bytes pyx empty tuple pyx empty tuple pyx tuple pyx empty tuple pyx empty tuple pyx kp s cython plex scanners py pyx n s position pyx empty bytes if unlikely pyx codeobj pyx err pyx error int tmp pip wheel cython cython plex scanners c note in definition of macro ‘ pyx pycode new’ pycode new a k l s f code c n v fv cell fn name fline lnos in file included from usr include compile h from usr include python h from tmp pip wheel cython cython plex scanners c usr include code h note expected ‘pyobject ’ aka ‘struct object ’ but argument is of type ‘int’ pyapi func pycodeobject pycode new tmp pip wheel cython cython plex scanners c error too many arguments to function ‘pycode new’ pycode new a k l s f code c n v fv cell fn name fline lnos tmp pip wheel cython cython plex scanners c note in expansion of macro ‘ pyx pycode new’ pyx codeobj pyobject pyx pycode new co optimized co newlocals pyx empty bytes pyx empty tuple pyx empty tuple pyx tuple pyx empty tuple pyx empty tuple pyx kp s cython plex scanners py pyx n s position pyx empty bytes if unlikely pyx codeobj pyx err pyx error in file included from usr include compile h from usr include python h from tmp pip wheel cython cython plex scanners c usr include code h note declared here pyapi func pycodeobject pycode new usr include code h warning passing argument of ‘pycode new’ makes pointer from integer without a cast define co optimized int tmp pip wheel cython cython plex scanners c note in definition of macro ‘ pyx pycode new’ pycode new a k l s f code c n v fv cell fn name fline lnos tmp pip wheel cython cython plex scanners c note in expansion of macro ‘co optimized’ pyx codeobj pyobject pyx pycode new co optimized co newlocals pyx empty bytes pyx empty tuple pyx empty tuple pyx tuple pyx empty tuple pyx empty tuple pyx kp s cython plex scanners py pyx n s get position pyx empty bytes if unlikely pyx codeobj pyx err pyx error in file included from usr include compile h from usr include python h from tmp pip wheel cython cython plex scanners c usr include code h note expected ‘pyobject ’ aka ‘struct object ’ but argument is of type ‘int’ pyapi func pycodeobject pycode new tmp pip wheel cython cython plex scanners c warning passing argument of ‘pycode new’ makes integer from pointer without a cast pyx codeobj pyobject pyx pycode new co optimized co newlocals pyx empty bytes pyx empty tuple pyx empty tuple pyx tuple pyx empty tuple pyx empty tuple pyx kp s cython plex scanners py pyx n s get position pyx empty bytes if unlikely pyx codeobj pyx err pyx error pyobject aka struct object tmp pip wheel cython cython plex scanners c note in definition of macro ‘ pyx pycode new’ pycode new a k l s f code c n v fv cell fn name fline lnos in file included from usr include compile h from usr include python h from tmp pip wheel cython cython plex scanners c usr include code h note expected ‘int’ but argument is of type ‘pyobject ’ aka ‘struct object ’ pyapi func pycodeobject pycode new tmp pip wheel cython cython plex scanners c warning passing argument of ‘pycode new’ makes pointer from integer without a cast pyx codeobj pyobject pyx pycode new co optimized co newlocals pyx empty bytes pyx empty tuple pyx empty tuple pyx tuple pyx empty tuple pyx empty tuple pyx kp s cython plex scanners py pyx n s get position pyx empty bytes if unlikely pyx codeobj pyx err pyx error int tmp pip wheel cython cython plex scanners c note in definition of macro ‘ pyx pycode new’ pycode new a k l s f code c n v fv cell fn name fline lnos in file included from usr include compile h from usr include python h from tmp pip wheel cython cython plex scanners c usr include code h note expected ‘pyobject ’ aka ‘struct object ’ but argument is of type ‘int’ pyapi func pycodeobject pycode new tmp pip wheel cython cython plex scanners c error too many arguments to function ‘pycode new’ pycode new a k l s f code c n v fv cell fn name fline lnos tmp pip wheel cython cython plex scanners c note in expansion of macro ‘ pyx pycode new’ pyx codeobj pyobject pyx pycode new co optimized co newlocals pyx empty bytes pyx empty tuple pyx empty tuple pyx tuple pyx empty tuple pyx empty tuple pyx kp s cython plex scanners py pyx n s get position pyx empty bytes if unlikely pyx codeobj pyx err pyx error in file included from usr include compile h from usr include python h from tmp pip wheel cython cython plex scanners c usr include code h note declared here pyapi func pycodeobject pycode new usr include code h warning passing argument of ‘pycode new’ makes pointer from integer without a cast define co optimized int tmp pip wheel cython cython plex scanners c note in definition of macro ‘ pyx pycode new’ pycode new a k l s f code c n v fv cell fn name fline lnos tmp pip wheel cython cython plex scanners c note in expansion of macro ‘co optimized’ pyx codeobj pyobject pyx pycode new co optimized co newlocals pyx empty bytes pyx empty tuple pyx empty tuple pyx tuple pyx empty tuple pyx empty tuple pyx kp s cython plex scanners py pyx n s eof pyx empty bytes if unlikely pyx codeobj pyx err pyx error in file included from usr include compile h from usr include python h from tmp pip wheel cython cython plex scanners c usr include code h note expected ‘pyobject ’ aka ‘struct object ’ but argument is of type ‘int’ pyapi func pycodeobject pycode new tmp pip wheel cython cython plex scanners c warning passing argument of ‘pycode new’ makes integer from pointer without a cast pyx codeobj pyobject pyx pycode new co optimized co newlocals pyx empty bytes pyx empty tuple pyx empty tuple pyx tuple pyx empty tuple pyx empty tuple pyx kp s cython plex scanners py pyx n s eof pyx empty bytes if unlikely pyx codeobj pyx err pyx error pyobject aka struct object tmp pip wheel cython cython plex scanners c note in definition of macro ‘ pyx pycode new’ pycode new a k l s f code c n v fv cell fn name fline lnos in file included from usr include compile h from usr include python h from tmp pip wheel cython cython plex scanners c usr include code h note expected ‘int’ but argument is of type ‘pyobject ’ aka ‘struct object ’ pyapi func pycodeobject pycode new tmp pip wheel cython cython plex scanners c warning passing argument of ‘pycode new’ makes pointer from integer without a cast pyx codeobj pyobject pyx pycode new co optimized co newlocals pyx empty bytes pyx empty tuple pyx empty tuple pyx tuple pyx empty tuple pyx empty tuple pyx kp s cython plex scanners py pyx n s eof pyx empty bytes if unlikely pyx codeobj pyx err pyx error int tmp pip wheel cython cython plex scanners c note in definition of macro ‘ pyx pycode new’ pycode new a k l s f code c n v fv cell fn name fline lnos in file included from usr include compile h from usr include python h from tmp pip wheel cython cython plex scanners c usr include code h note expected ‘pyobject ’ aka ‘struct object ’ but argument is of type ‘int’ pyapi func pycodeobject pycode new tmp pip wheel cython cython plex scanners c error too many arguments to function ‘pycode new’ pycode new a k l s f code c n v fv cell fn name fline lnos tmp pip wheel cython cython plex scanners c note in expansion of macro ‘ pyx pycode new’ pyx codeobj pyobject pyx pycode new co optimized co newlocals pyx empty bytes pyx empty tuple pyx empty tuple pyx tuple pyx empty tuple pyx empty tuple pyx kp s cython plex scanners py pyx n s eof pyx empty bytes if unlikely pyx codeobj pyx err pyx error in file included from usr include compile h from usr include python h from tmp pip wheel cython cython plex scanners c usr include code h note declared here pyapi func pycodeobject pycode new tmp pip wheel cython cython plex scanners c in function ‘ pyx createcodeobjectfortraceback’ tmp pip wheel cython cython plex scanners c warning passing argument of ‘pycode new’ makes integer from pointer without a cast py funcname pyobject name pyobject aka struct object tmp pip wheel cython cython plex scanners c note in definition of macro ‘ pyx pycode new’ pycode new a k l s f code c n v fv cell fn name fline lnos in file included from usr include compile h from usr include python h from tmp pip wheel cython cython plex scanners c usr include code h note expected ‘int’ but argument is of type ‘pyobject ’ aka ‘struct object ’ pyapi func pycodeobject pycode new tmp pip wheel cython cython plex scanners c warning passing argument of ‘pycode new’ makes pointer from integer without a cast py line int tmp pip wheel cython cython plex scanners c note in definition of macro ‘ pyx pycode new’ pycode new a k l s f code c n v fv cell fn name fline lnos in file included from usr include compile h from usr include python h from tmp pip wheel cython cython plex scanners c usr include code h note expected ‘pyobject ’ aka ‘struct object ’ but argument is of type ‘int’ pyapi func pycodeobject pycode new tmp pip wheel cython cython plex scanners c error too many arguments to function ‘pycode new’ pycode new a k l s f code c n v fv cell fn name fline lnos tmp pip wheel cython cython plex scanners c note in expansion of macro ‘ pyx pycode new’ py code pyx pycode new in file included from usr include compile h from usr include python h from tmp pip wheel cython cython plex scanners c usr include code h note declared here pyapi func pycodeobject pycode new error command linux gnu gcc failed with exit status error failed building wheel for cython error failed to build one or more wheels traceback most recent call last file root local lib site packages setuptools installer py line in fetch build egg subprocess check call cmd file usr lib subprocess py line in check call raise calledprocesserror retcode cmd subprocess calledprocesserror command returned non zero exit status during handling of the above exception another exception occurred traceback most recent call last file line in file tmp pip install kivy setup py line in setup file root local lib site packages setuptools init py line in setup install setup requires attrs file root local lib site packages setuptools init py line in install setup requires dist fetch build eggs dist setup requires file root local lib site packages setuptools dist py line in fetch build eggs resolved dists pkg resources working set resolve file root local lib site packages pkg resources init py line in resolve dist best env best match file root local lib site packages pkg resources init py line in best match return self obtain req installer file root local lib site packages pkg resources init py line in obtain return installer requirement file root local lib site packages setuptools dist py line in fetch build egg return fetch build egg self req file root local lib site packages setuptools installer py line in fetch build egg raise distutilserror str e distutils errors distutilserror command returned non zero exit status using setuptools environ change use true using this graphics system opengl gstreamer found via pkg config found via pkg config found sdl header at usr include sdl h found sdl mixer header at usr include sdl mixer h found sdl ttf header at usr include sdl ttf h found sdl image header at usr include sdl image h pango found via pkg config libraries include dirs library dirs extra link args extra compile args depends error dependency for context pyx not resolved config pxi error dependency for compiler pyx not resolved config pxi error dependency for context instructions pyx not resolved config pxi error dependency for fbo pyx not resolved config pxi error dependency for gl instructions pyx not resolved config pxi error dependency for instructions pyx not resolved config pxi error dependency for opengl pyx not resolved config pxi error dependency for opengl utils pyx not resolved config pxi error dependency for shader pyx not resolved config pxi error dependency for stencil instructions pyx not resolved config pxi error dependency for scissor instructions pyx not resolved config pxi error dependency for texture pyx not resolved config pxi error dependency for vbo pyx not resolved config pxi error dependency for vertex pyx not resolved config pxi error dependency for vertex instructions pyx not resolved config pxi error dependency for cgl pyx not resolved config pxi error dependency for cgl mock pyx not resolved config pxi error dependency for cgl gl pyx not resolved config pxi error dependency for cgl glew pyx not resolved config pxi error dependency for cgl pyx not resolved config pxi error dependency for svg pyx not resolved config pxi error command errored out with exit status python setup py egg info check the logs for full command output
1
12,238
9,659,837,393
IssuesEvent
2019-05-20 14:17:13
dart-lang/sdk
https://api.github.com/repos/dart-lang/sdk
closed
Flush the analysis server cache after each bot run
area-infrastructure
We believe that the build bots keep the analysis cache in ~/.dartServer/.analysis-driver and suspect that this cache is not getting deleted on the Flutter/Dart head-head-head bot. As a result, any cached errors are continued to be reported until a file change that invalidates them., leading to problems. We need to have the cache deleted to avoid this problem See the comments for https://dart-review.googlesource.com/c/sdk/+/84041 for some background.
1.0
Flush the analysis server cache after each bot run - We believe that the build bots keep the analysis cache in ~/.dartServer/.analysis-driver and suspect that this cache is not getting deleted on the Flutter/Dart head-head-head bot. As a result, any cached errors are continued to be reported until a file change that invalidates them., leading to problems. We need to have the cache deleted to avoid this problem See the comments for https://dart-review.googlesource.com/c/sdk/+/84041 for some background.
non_comp
flush the analysis server cache after each bot run we believe that the build bots keep the analysis cache in dartserver analysis driver and suspect that this cache is not getting deleted on the flutter dart head head head bot as a result any cached errors are continued to be reported until a file change that invalidates them leading to problems we need to have the cache deleted to avoid this problem see the comments for for some background
0
89,479
15,829,500,474
IssuesEvent
2021-04-06 11:16:28
VivekBuzruk/Hygieia
https://api.github.com/repos/VivekBuzruk/Hygieia
closed
CVE-2020-15366 (Medium) detected in ajv-4.11.8.tgz - autoclosed
security vulnerability
## CVE-2020-15366 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ajv-4.11.8.tgz</b></p></summary> <p>Another JSON Schema Validator</p> <p>Library home page: <a href="https://registry.npmjs.org/ajv/-/ajv-4.11.8.tgz">https://registry.npmjs.org/ajv/-/ajv-4.11.8.tgz</a></p> <p>Path to dependency file: Hygieia/UI-protractor-tests/package.json</p> <p>Path to vulnerable library: Hygieia/UI-protractor-tests/node_modules/ajv/package.json</p> <p> Dependency Hierarchy: - chai-as-promised-7.0.0.tgz (Root Library) - eslint-3.19.0.tgz - table-3.8.3.tgz - :x: **ajv-4.11.8.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://api.github.com/repos/VivekBuzruk/Hygieia/commits/3c4f119e4343cf7fa276bb4756361b926902248e">3c4f119e4343cf7fa276bb4756361b926902248e</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in ajv.validate() in Ajv (aka Another JSON Schema Validator) 6.12.2. A carefully crafted JSON schema could be provided that allows execution of other code by prototype pollution. (While untrusted schemas are recommended against, the worst case of an untrusted schema should be a denial of service, not execution of code.) <p>Publish Date: 2020-07-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-15366>CVE-2020-15366</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/ajv-validator/ajv/releases/tag/v6.12.3">https://github.com/ajv-validator/ajv/releases/tag/v6.12.3</a></p> <p>Release Date: 2020-07-15</p> <p>Fix Resolution: ajv - 6.12.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-15366 (Medium) detected in ajv-4.11.8.tgz - autoclosed - ## CVE-2020-15366 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ajv-4.11.8.tgz</b></p></summary> <p>Another JSON Schema Validator</p> <p>Library home page: <a href="https://registry.npmjs.org/ajv/-/ajv-4.11.8.tgz">https://registry.npmjs.org/ajv/-/ajv-4.11.8.tgz</a></p> <p>Path to dependency file: Hygieia/UI-protractor-tests/package.json</p> <p>Path to vulnerable library: Hygieia/UI-protractor-tests/node_modules/ajv/package.json</p> <p> Dependency Hierarchy: - chai-as-promised-7.0.0.tgz (Root Library) - eslint-3.19.0.tgz - table-3.8.3.tgz - :x: **ajv-4.11.8.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://api.github.com/repos/VivekBuzruk/Hygieia/commits/3c4f119e4343cf7fa276bb4756361b926902248e">3c4f119e4343cf7fa276bb4756361b926902248e</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in ajv.validate() in Ajv (aka Another JSON Schema Validator) 6.12.2. A carefully crafted JSON schema could be provided that allows execution of other code by prototype pollution. (While untrusted schemas are recommended against, the worst case of an untrusted schema should be a denial of service, not execution of code.) <p>Publish Date: 2020-07-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-15366>CVE-2020-15366</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/ajv-validator/ajv/releases/tag/v6.12.3">https://github.com/ajv-validator/ajv/releases/tag/v6.12.3</a></p> <p>Release Date: 2020-07-15</p> <p>Fix Resolution: ajv - 6.12.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_comp
cve medium detected in ajv tgz autoclosed cve medium severity vulnerability vulnerable library ajv tgz another json schema validator library home page a href path to dependency file hygieia ui protractor tests package json path to vulnerable library hygieia ui protractor tests node modules ajv package json dependency hierarchy chai as promised tgz root library eslint tgz table tgz x ajv tgz vulnerable library found in head commit a href found in base branch master vulnerability details an issue was discovered in ajv validate in ajv aka another json schema validator a carefully crafted json schema could be provided that allows execution of other code by prototype pollution while untrusted schemas are recommended against the worst case of an untrusted schema should be a denial of service not execution of code publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution ajv step up your open source security game with whitesource
0
730,765
25,188,232,549
IssuesEvent
2022-11-11 20:30:52
jrbrawner/CS478
https://api.github.com/repos/jrbrawner/CS478
closed
Create and Implement Search functionality
UserStory Priority 2 3
AS a USER I want to be able to search for users, and attributes like graduation year, major, etc So that I can filter my content for what I am interested in
1.0
Create and Implement Search functionality - AS a USER I want to be able to search for users, and attributes like graduation year, major, etc So that I can filter my content for what I am interested in
non_comp
create and implement search functionality as a user i want to be able to search for users and attributes like graduation year major etc so that i can filter my content for what i am interested in
0
6,982
9,257,248,503
IssuesEvent
2019-03-17 03:55:41
darlinghq/darling
https://api.github.com/repos/darlinghq/darling
closed
Xcode 9 and 10 need some stubs
Application Compatibility
This prevents running Xcode 9 CLI tools or any other recent Swift version.
True
Xcode 9 and 10 need some stubs - This prevents running Xcode 9 CLI tools or any other recent Swift version.
comp
xcode and need some stubs this prevents running xcode cli tools or any other recent swift version
1
19,659
27,298,916,886
IssuesEvent
2023-02-23 23:09:50
scikit-learn/scikit-learn
https://api.github.com/repos/scikit-learn/scikit-learn
closed
Support nullable pandas dtypes in `confusion_matrix`
New Feature Pandas compatibility
### Describe the workflow you want to enable I would like to be able to pass the nullable pandas dtypes ("Int64", "Float64", "boolean") into sklearn's `confusion_matrix` function. Because the dtypes become object dtype when converted to numpy arrays we get `ValueError: Classification metrics can't handle a mix of unknown and binary targets`: Repro with sklearn 1.2.1: ```python import pandas as pd import pytest from sklearn.metrics import confusion_matrix for dtype in ["Int64", "Float64", "boolean"]: y_true = pd.Series([1, 0, 0, 1, 0, 1, 1, 0, 1], dtype=dtype) y_predicted = pd.Series([0, 0, 1, 1, 0, 1, 1, 1, 1], dtype="int64") with pytest.raises(ValueError, match="Classification metrics can't handle a mix of unknown and binary targets"): confusion_matrix(y_true, y_predicted) ``` ### Describe your proposed solution We should get the same behavior as when int64, float64, and bool dtypes are used, which is no error: ```python import pandas as pd from sklearn.metrics import confusion_matrix for dtype in ["int64", "float64", "bool"]: y_true = pd.Series([1, 0, 0, 1, 0, 1, 1, 0, 1], dtype=dtype) y_predicted = pd.Series([0, 0, 1, 1, 0, 1, 1, 1, 1], dtype="int64") confusion_matrix(y_true, y_predicted) ``` ### Describe alternatives you've considered, if relevant Our current workaround is to convert the data to numpy arrays with the corresponding dtype that works prior to passing it into `confusion_matrix` ### Additional context _No response_
True
Support nullable pandas dtypes in `confusion_matrix` - ### Describe the workflow you want to enable I would like to be able to pass the nullable pandas dtypes ("Int64", "Float64", "boolean") into sklearn's `confusion_matrix` function. Because the dtypes become object dtype when converted to numpy arrays we get `ValueError: Classification metrics can't handle a mix of unknown and binary targets`: Repro with sklearn 1.2.1: ```python import pandas as pd import pytest from sklearn.metrics import confusion_matrix for dtype in ["Int64", "Float64", "boolean"]: y_true = pd.Series([1, 0, 0, 1, 0, 1, 1, 0, 1], dtype=dtype) y_predicted = pd.Series([0, 0, 1, 1, 0, 1, 1, 1, 1], dtype="int64") with pytest.raises(ValueError, match="Classification metrics can't handle a mix of unknown and binary targets"): confusion_matrix(y_true, y_predicted) ``` ### Describe your proposed solution We should get the same behavior as when int64, float64, and bool dtypes are used, which is no error: ```python import pandas as pd from sklearn.metrics import confusion_matrix for dtype in ["int64", "float64", "bool"]: y_true = pd.Series([1, 0, 0, 1, 0, 1, 1, 0, 1], dtype=dtype) y_predicted = pd.Series([0, 0, 1, 1, 0, 1, 1, 1, 1], dtype="int64") confusion_matrix(y_true, y_predicted) ``` ### Describe alternatives you've considered, if relevant Our current workaround is to convert the data to numpy arrays with the corresponding dtype that works prior to passing it into `confusion_matrix` ### Additional context _No response_
comp
support nullable pandas dtypes in confusion matrix describe the workflow you want to enable i would like to be able to pass the nullable pandas dtypes boolean into sklearn s confusion matrix function because the dtypes become object dtype when converted to numpy arrays we get valueerror classification metrics can t handle a mix of unknown and binary targets repro with sklearn python import pandas as pd import pytest from sklearn metrics import confusion matrix for dtype in y true pd series dtype dtype y predicted pd series dtype with pytest raises valueerror match classification metrics can t handle a mix of unknown and binary targets confusion matrix y true y predicted describe your proposed solution we should get the same behavior as when and bool dtypes are used which is no error python import pandas as pd from sklearn metrics import confusion matrix for dtype in y true pd series dtype dtype y predicted pd series dtype confusion matrix y true y predicted describe alternatives you ve considered if relevant our current workaround is to convert the data to numpy arrays with the corresponding dtype that works prior to passing it into confusion matrix additional context no response
1
191,410
22,215,771,526
IssuesEvent
2022-06-08 01:22:05
ShaikUsaf/linux-3.0.35
https://api.github.com/repos/ShaikUsaf/linux-3.0.35
opened
CVE-2015-8966 (High) detected in linux-stable-rtv3.8.6
security vulnerability
## CVE-2015-8966 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv3.8.6</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/arch/arm/kernel/sys_oabi-compat.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/arch/arm/kernel/sys_oabi-compat.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/arch/arm/kernel/sys_oabi-compat.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> arch/arm/kernel/sys_oabi-compat.c in the Linux kernel before 4.4 allows local users to gain privileges via a crafted (1) F_OFD_GETLK, (2) F_OFD_SETLK, or (3) F_OFD_SETLKW command in an fcntl64 system call. <p>Publish Date: 2016-12-08 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-8966>CVE-2015-8966</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2015-8966">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2015-8966</a></p> <p>Release Date: 2016-12-08</p> <p>Fix Resolution: v4.4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2015-8966 (High) detected in linux-stable-rtv3.8.6 - ## CVE-2015-8966 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv3.8.6</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/arch/arm/kernel/sys_oabi-compat.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/arch/arm/kernel/sys_oabi-compat.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/arch/arm/kernel/sys_oabi-compat.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> arch/arm/kernel/sys_oabi-compat.c in the Linux kernel before 4.4 allows local users to gain privileges via a crafted (1) F_OFD_GETLK, (2) F_OFD_SETLK, or (3) F_OFD_SETLKW command in an fcntl64 system call. <p>Publish Date: 2016-12-08 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-8966>CVE-2015-8966</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2015-8966">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2015-8966</a></p> <p>Release Date: 2016-12-08</p> <p>Fix Resolution: v4.4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_comp
cve high detected in linux stable cve high severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in base branch master vulnerable source files arch arm kernel sys oabi compat c arch arm kernel sys oabi compat c arch arm kernel sys oabi compat c vulnerability details arch arm kernel sys oabi compat c in the linux kernel before allows local users to gain privileges via a crafted f ofd getlk f ofd setlk or f ofd setlkw command in an system call publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
16,068
21,394,258,028
IssuesEvent
2022-04-21 10:02:06
PluginBugs/Issues-ItemsAdder
https://api.github.com/repos/PluginBugs/Issues-ItemsAdder
closed
Mimic: New API implementation registering mechanism
Compatibility with other plugin
Hm. It looks like you've not migrated to the new mechanism of API imlementations registering. I see the following message on ItemsAdapter enable: ``` [Mimic] Service ru.endlesscode.mimic.items.BukkitItemsRegistry with id 'ia' registered in deprecated way. Please ask the ItemsAdder authors (LoneDev) to migrate to the new service registration API introduced in Mimic v0.7: https://github.com/EndlessCodeGroup/Mimic/releases/tag/v0.7 ``` _Originally posted by @osipxd in https://github.com/PluginBugs/Issues-ItemsAdder/issues/1503#issuecomment-1101743625_
True
Mimic: New API implementation registering mechanism - Hm. It looks like you've not migrated to the new mechanism of API imlementations registering. I see the following message on ItemsAdapter enable: ``` [Mimic] Service ru.endlesscode.mimic.items.BukkitItemsRegistry with id 'ia' registered in deprecated way. Please ask the ItemsAdder authors (LoneDev) to migrate to the new service registration API introduced in Mimic v0.7: https://github.com/EndlessCodeGroup/Mimic/releases/tag/v0.7 ``` _Originally posted by @osipxd in https://github.com/PluginBugs/Issues-ItemsAdder/issues/1503#issuecomment-1101743625_
comp
mimic new api implementation registering mechanism hm it looks like you ve not migrated to the new mechanism of api imlementations registering i see the following message on itemsadapter enable service ru endlesscode mimic items bukkititemsregistry with id ia registered in deprecated way please ask the itemsadder authors lonedev to migrate to the new service registration api introduced in mimic originally posted by osipxd in
1
17,363
23,936,555,649
IssuesEvent
2022-09-11 10:10:38
elementor/elementor
https://api.github.com/repos/elementor/elementor
closed
Post Widget - pause on hover not working
compatibility/3rd_party
I am seeing an issue with the Posts widget when using a custom template loop (ele custom skins plugin) and in slider mode that the posts do not pause on hover when function is active. ![image](https://user-images.githubusercontent.com/62887371/107665584-7ee02100-6c85-11eb-861a-d7c855956d88.png) Video of widget not pausing on hover - https://www.loom.com/share/54021cbb6a53405a8f0c73be65291ea3 <!-- ## BEFORE POSTING YOUR ISSUE - Please create GitHub issues only for bugs and feature requests. GitHub issues ARE NOT FOR SUPPORT! - If you have questions or need general support, Please use: https://wordpress.org/support/plugin/elementor - For help and support from the Elementor community, see: https://www.facebook.com/groups/Elementors/ - To read more about Elementor, check out our documentation: https://docs.elementor.com - Developers docs are located at https://developers.elementor.com/ ===== Guidelines ==== - Search this repository (top of the page) for the issue, and make sure it has not been fixed or reported before. - Make sure you are using the most updated versions of Elementor, WordPress & your theme. - Deactivate all of your plugins. If this solves the problem, gradually activate your plugins one by one, until you spot the problematic plugin. - Change your WordPress theme to WordPress TwentySixteen (or other default WordPress theme). - If you're requesting a new feature, explain why you'd like it to be added. Try to add as much detail as you can, and be specific. --> ## Prerequisites <!-- MARK COMPLETED ITEMS WITH AN [x] --> - [x ] I have searched for similar issues in both open and closed tickets and cannot find a duplicate. - [ x] The issue still exists against the latest stable version of Elementor. ## Description <!-- Describe which problem you've encountered. What caused the issue, and what did you expect to happen. Attach screenshots and related links to help us understand the issue in more detail. Please be as descriptive as possible; issues lacking the below details, or for any other reason than to report a bug, may be closed without action. --> ## Steps to reproduce <!-- For bug reports, list all the steps needed to reproduce your issue, so we can replicate it ourselves. --> ## Isolating the problem <!-- MARK COMPLETED ITEMS WITH AN [x] --> - [ x] This bug happens with only Elementor plugin active (and Elementor Pro). - [ x] This bug happens with a default WordPress theme active. - [ x] I can reproduce this bug consistently using the steps above. ## Environment <details> <summary>System Info</summary> ``` <!-- For bug reports, let us know about your system environment: Copy and paste the system info report from Elementor => System info, and paste it here or in http://pastebin.com/ --> ``` </details>
True
Post Widget - pause on hover not working - I am seeing an issue with the Posts widget when using a custom template loop (ele custom skins plugin) and in slider mode that the posts do not pause on hover when function is active. ![image](https://user-images.githubusercontent.com/62887371/107665584-7ee02100-6c85-11eb-861a-d7c855956d88.png) Video of widget not pausing on hover - https://www.loom.com/share/54021cbb6a53405a8f0c73be65291ea3 <!-- ## BEFORE POSTING YOUR ISSUE - Please create GitHub issues only for bugs and feature requests. GitHub issues ARE NOT FOR SUPPORT! - If you have questions or need general support, Please use: https://wordpress.org/support/plugin/elementor - For help and support from the Elementor community, see: https://www.facebook.com/groups/Elementors/ - To read more about Elementor, check out our documentation: https://docs.elementor.com - Developers docs are located at https://developers.elementor.com/ ===== Guidelines ==== - Search this repository (top of the page) for the issue, and make sure it has not been fixed or reported before. - Make sure you are using the most updated versions of Elementor, WordPress & your theme. - Deactivate all of your plugins. If this solves the problem, gradually activate your plugins one by one, until you spot the problematic plugin. - Change your WordPress theme to WordPress TwentySixteen (or other default WordPress theme). - If you're requesting a new feature, explain why you'd like it to be added. Try to add as much detail as you can, and be specific. --> ## Prerequisites <!-- MARK COMPLETED ITEMS WITH AN [x] --> - [x ] I have searched for similar issues in both open and closed tickets and cannot find a duplicate. - [ x] The issue still exists against the latest stable version of Elementor. ## Description <!-- Describe which problem you've encountered. What caused the issue, and what did you expect to happen. Attach screenshots and related links to help us understand the issue in more detail. Please be as descriptive as possible; issues lacking the below details, or for any other reason than to report a bug, may be closed without action. --> ## Steps to reproduce <!-- For bug reports, list all the steps needed to reproduce your issue, so we can replicate it ourselves. --> ## Isolating the problem <!-- MARK COMPLETED ITEMS WITH AN [x] --> - [ x] This bug happens with only Elementor plugin active (and Elementor Pro). - [ x] This bug happens with a default WordPress theme active. - [ x] I can reproduce this bug consistently using the steps above. ## Environment <details> <summary>System Info</summary> ``` <!-- For bug reports, let us know about your system environment: Copy and paste the system info report from Elementor => System info, and paste it here or in http://pastebin.com/ --> ``` </details>
comp
post widget pause on hover not working i am seeing an issue with the posts widget when using a custom template loop ele custom skins plugin and in slider mode that the posts do not pause on hover when function is active video of widget not pausing on hover before posting your issue please create github issues only for bugs and feature requests github issues are not for support if you have questions or need general support please use for help and support from the elementor community see to read more about elementor check out our documentation developers docs are located at guidelines search this repository top of the page for the issue and make sure it has not been fixed or reported before make sure you are using the most updated versions of elementor wordpress your theme deactivate all of your plugins if this solves the problem gradually activate your plugins one by one until you spot the problematic plugin change your wordpress theme to wordpress twentysixteen or other default wordpress theme if you re requesting a new feature explain why you d like it to be added try to add as much detail as you can and be specific prerequisites i have searched for similar issues in both open and closed tickets and cannot find a duplicate the issue still exists against the latest stable version of elementor description describe which problem you ve encountered what caused the issue and what did you expect to happen attach screenshots and related links to help us understand the issue in more detail please be as descriptive as possible issues lacking the below details or for any other reason than to report a bug may be closed without action steps to reproduce for bug reports list all the steps needed to reproduce your issue so we can replicate it ourselves isolating the problem this bug happens with only elementor plugin active and elementor pro this bug happens with a default wordpress theme active i can reproduce this bug consistently using the steps above environment system info system info and paste it here or in
1
12,088
14,247,569,957
IssuesEvent
2020-11-19 11:38:57
the1812/Bilibili-Evolved
https://api.github.com/repos/the1812/Bilibili-Evolved
closed
关于鼠标中键点击B站链接 不加载脚本的问题
bug compatibility
习惯了 鼠标中间先连续在后台新标签点开多个要看的视频 然后一个一个的标签看过来 发现只要用中间点击的 都不会加载网页 请问有什么方法可以兼容鼠标中间
True
关于鼠标中键点击B站链接 不加载脚本的问题 - 习惯了 鼠标中间先连续在后台新标签点开多个要看的视频 然后一个一个的标签看过来 发现只要用中间点击的 都不会加载网页 请问有什么方法可以兼容鼠标中间
comp
关于鼠标中键点击b站链接 不加载脚本的问题 习惯了 鼠标中间先连续在后台新标签点开多个要看的视频 然后一个一个的标签看过来 发现只要用中间点击的 都不会加载网页 请问有什么方法可以兼容鼠标中间
1
12,807
15,078,413,568
IssuesEvent
2021-02-05 08:41:03
vesoft-inc/nebula-graph
https://api.github.com/repos/vesoft-inc/nebula-graph
closed
[openCypher compatibility] Can't use DESCENDING after ORDER BY
incompatible
nGQL: MATCH (v:player) RETURN v.age AS Age, v.name AS Name | ORDER BY Age DESCENDING, Name ASC | LIMIT 10 ![image](https://user-images.githubusercontent.com/69501902/104308522-c56a2080-550b-11eb-9a09-ce9cb34799b3.png) openCypher: ![image](https://user-images.githubusercontent.com/69501902/104404630-158dc500-5596-11eb-9493-d7b3a18aa687.png)
True
[openCypher compatibility] Can't use DESCENDING after ORDER BY - nGQL: MATCH (v:player) RETURN v.age AS Age, v.name AS Name | ORDER BY Age DESCENDING, Name ASC | LIMIT 10 ![image](https://user-images.githubusercontent.com/69501902/104308522-c56a2080-550b-11eb-9a09-ce9cb34799b3.png) openCypher: ![image](https://user-images.githubusercontent.com/69501902/104404630-158dc500-5596-11eb-9493-d7b3a18aa687.png)
comp
can t use descending after order by ngql match v player return v age as age v name as name order by age descending name asc limit opencypher
1
17,828
24,599,895,325
IssuesEvent
2022-10-14 11:34:18
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
closed
More windows for filtering and spectral analysis
triaged module: scipy compatibility
### 🚀 The feature, motivation and pitch #### Introduction This feature request looks to add more windows for filtering and spectral analysis. I've noticed that we're supporting only the following windows: - `barlett_window` - `blackman_window` - `hamming_window` - `hann_window` - `kaiser_window` But there are many more windows that are pretty useful. For instance, the Dolph-Chebyshev window which optimizes for the minimum main lobe width for a given side-lobe specification. So, I'm proposing we add the following windows: - [ ] `chebyshev_window` - [ ] `taylor_window` - [ ] `cosine_window` - [ ] `gaussian_window` - [ ] `exponential_window` - [ ] `kbd_window` - [ ] `nutall_window` We could add more to the list. #### Motivation One of the things I came across is filtering signals when trying to create custom datasets. The window function that I tried to use was Dolph-Chebyshev, but PyTorch does not support it at the moment. Other functions such as Kaiser-Bessel Derived window function is broadly used in audio signal processing. ### Alternatives The solutions are quite straightforward. From the list of proposed windows, the most complex one is the Dolph-Chebyshev window which requires an inverse FFT. But this is easy to apply as the IFFT implementation exists in PyTorch already. This window would require creating CUDA and CPU kernels. I already have an implementation that I could share in a draft PR if this feature is accepted. The other window functions are quite simple and easy to implement, namely, the cosine window, exponential window, and gaussian window. I also have implementations for these. A sneak peek of the CPU `chebyshev_window_kernel`: This calculates the window coefficients in the frequency domain. The IFFT would be applied after the call to the kernel's finished. ```cpp static void chebyshev_window_kernel(TensorIteratorBase& iter, int64_t window_length, double attenuation) { AT_DISPATCH_FLOATING_TYPES_AND(kBFloat16, iter.dtype(), "chebyshev_window_cpu", [&](){ const int64_t n = window_length - 1; const scalar_t beta = static_cast<scalar_t>(std::cosh(1.0 / n * std::acosh(std::pow(10, attenuation / 20.0)))); cpu_kernel(iter, [=](scalar_t a){ auto x = beta * static_cast<scalar_t>(std::cos(c10::pi<double> * a / window_length)); return static_cast<scalar_t>(chebyshev_polynomial_t_forward(x, n) / std::pow(10, attenuation / 20.0)); }); }); } ``` And for the `cosine_window` function: Does not need a kernel, I believe. ```cpp Tensor cosine_window( int64_t window_length, bool periodic, c10::optional<ScalarType> dtype_opt, c10::optional<Layout> layout, c10::optional<Device> device, c10::optional<bool> pin_memory) { // See [Note: hacky wrapper removal for TensorOptions] ScalarType dtype = c10::dtype_or_default(dtype_opt); TensorOptions options = TensorOptions().dtype(dtype).layout(layout).device(device).pinned_memory(pin_memory); window_function_checks("cosine_window", options, window_length); if (window_length == 0) { return at::empty({0}, options); } if (window_length == 1) { return native::ones({1}, dtype, layout, device, pin_memory); } if (periodic) { window_length += 1; } auto window = native::arange(window_length, dtype, layout, device, pin_memory) .add(0.5).mul_(c10::pi<double>).mul_(static_cast<double>(1.0 / window_length)).sin_(); return periodic ? window.narrow(0, 0, window_length - 1) : window; } ``` The other window functions would follow the same or similar approach. ### Additional context #### References - Window functions: - https://sites.google.com/site/stevedtran/course/intro-to-digital-signal-processing/notes2/windowing/type-of-windowing - Kaiser-Bessel Derived Window: - Bosi, Marina, and Richard E. Goldberg. Introduction to Digital Audio Coding and Standards. Dordrecht: Kluwer, 2003. - Dolph-Chebyshev Window: - https://ccrma.stanford.edu/~jos/sasp/Dolph_Chebyshev_Window.html - Taylor Window: - Armin Doerry, "Catalog of Window Taper Functions for Sidelobe Control", 2017 https://www.researchgate.net/profile/Armin_Doerry/publication/316281181_Catalog_of_Window_Taper_Functions_for_Sidelobe_Control/links/58f92cb2a6fdccb121c9d54d/Catalog-of-Window-Taper-Functions-for-Sidelobe-Control.pdf - Nutall Window: - A. Nuttall, “Some windows with very good sidelobe behavior,” IEEE Transactions on Acoustics, Speech, and Signal Processing, vol. 29, no. 1, pp. 84-91, Feb 1981. [DOI:10.1109/TASSP.1981.1163506](https://doi.org/10.1109/TASSP.1981.1163506)
True
More windows for filtering and spectral analysis - ### 🚀 The feature, motivation and pitch #### Introduction This feature request looks to add more windows for filtering and spectral analysis. I've noticed that we're supporting only the following windows: - `barlett_window` - `blackman_window` - `hamming_window` - `hann_window` - `kaiser_window` But there are many more windows that are pretty useful. For instance, the Dolph-Chebyshev window which optimizes for the minimum main lobe width for a given side-lobe specification. So, I'm proposing we add the following windows: - [ ] `chebyshev_window` - [ ] `taylor_window` - [ ] `cosine_window` - [ ] `gaussian_window` - [ ] `exponential_window` - [ ] `kbd_window` - [ ] `nutall_window` We could add more to the list. #### Motivation One of the things I came across is filtering signals when trying to create custom datasets. The window function that I tried to use was Dolph-Chebyshev, but PyTorch does not support it at the moment. Other functions such as Kaiser-Bessel Derived window function is broadly used in audio signal processing. ### Alternatives The solutions are quite straightforward. From the list of proposed windows, the most complex one is the Dolph-Chebyshev window which requires an inverse FFT. But this is easy to apply as the IFFT implementation exists in PyTorch already. This window would require creating CUDA and CPU kernels. I already have an implementation that I could share in a draft PR if this feature is accepted. The other window functions are quite simple and easy to implement, namely, the cosine window, exponential window, and gaussian window. I also have implementations for these. A sneak peek of the CPU `chebyshev_window_kernel`: This calculates the window coefficients in the frequency domain. The IFFT would be applied after the call to the kernel's finished. ```cpp static void chebyshev_window_kernel(TensorIteratorBase& iter, int64_t window_length, double attenuation) { AT_DISPATCH_FLOATING_TYPES_AND(kBFloat16, iter.dtype(), "chebyshev_window_cpu", [&](){ const int64_t n = window_length - 1; const scalar_t beta = static_cast<scalar_t>(std::cosh(1.0 / n * std::acosh(std::pow(10, attenuation / 20.0)))); cpu_kernel(iter, [=](scalar_t a){ auto x = beta * static_cast<scalar_t>(std::cos(c10::pi<double> * a / window_length)); return static_cast<scalar_t>(chebyshev_polynomial_t_forward(x, n) / std::pow(10, attenuation / 20.0)); }); }); } ``` And for the `cosine_window` function: Does not need a kernel, I believe. ```cpp Tensor cosine_window( int64_t window_length, bool periodic, c10::optional<ScalarType> dtype_opt, c10::optional<Layout> layout, c10::optional<Device> device, c10::optional<bool> pin_memory) { // See [Note: hacky wrapper removal for TensorOptions] ScalarType dtype = c10::dtype_or_default(dtype_opt); TensorOptions options = TensorOptions().dtype(dtype).layout(layout).device(device).pinned_memory(pin_memory); window_function_checks("cosine_window", options, window_length); if (window_length == 0) { return at::empty({0}, options); } if (window_length == 1) { return native::ones({1}, dtype, layout, device, pin_memory); } if (periodic) { window_length += 1; } auto window = native::arange(window_length, dtype, layout, device, pin_memory) .add(0.5).mul_(c10::pi<double>).mul_(static_cast<double>(1.0 / window_length)).sin_(); return periodic ? window.narrow(0, 0, window_length - 1) : window; } ``` The other window functions would follow the same or similar approach. ### Additional context #### References - Window functions: - https://sites.google.com/site/stevedtran/course/intro-to-digital-signal-processing/notes2/windowing/type-of-windowing - Kaiser-Bessel Derived Window: - Bosi, Marina, and Richard E. Goldberg. Introduction to Digital Audio Coding and Standards. Dordrecht: Kluwer, 2003. - Dolph-Chebyshev Window: - https://ccrma.stanford.edu/~jos/sasp/Dolph_Chebyshev_Window.html - Taylor Window: - Armin Doerry, "Catalog of Window Taper Functions for Sidelobe Control", 2017 https://www.researchgate.net/profile/Armin_Doerry/publication/316281181_Catalog_of_Window_Taper_Functions_for_Sidelobe_Control/links/58f92cb2a6fdccb121c9d54d/Catalog-of-Window-Taper-Functions-for-Sidelobe-Control.pdf - Nutall Window: - A. Nuttall, “Some windows with very good sidelobe behavior,” IEEE Transactions on Acoustics, Speech, and Signal Processing, vol. 29, no. 1, pp. 84-91, Feb 1981. [DOI:10.1109/TASSP.1981.1163506](https://doi.org/10.1109/TASSP.1981.1163506)
comp
more windows for filtering and spectral analysis 🚀 the feature motivation and pitch introduction this feature request looks to add more windows for filtering and spectral analysis i ve noticed that we re supporting only the following windows barlett window blackman window hamming window hann window kaiser window but there are many more windows that are pretty useful for instance the dolph chebyshev window which optimizes for the minimum main lobe width for a given side lobe specification so i m proposing we add the following windows chebyshev window taylor window cosine window gaussian window exponential window kbd window nutall window we could add more to the list motivation one of the things i came across is filtering signals when trying to create custom datasets the window function that i tried to use was dolph chebyshev but pytorch does not support it at the moment other functions such as kaiser bessel derived window function is broadly used in audio signal processing alternatives the solutions are quite straightforward from the list of proposed windows the most complex one is the dolph chebyshev window which requires an inverse fft but this is easy to apply as the ifft implementation exists in pytorch already this window would require creating cuda and cpu kernels i already have an implementation that i could share in a draft pr if this feature is accepted the other window functions are quite simple and easy to implement namely the cosine window exponential window and gaussian window i also have implementations for these a sneak peek of the cpu chebyshev window kernel this calculates the window coefficients in the frequency domain the ifft would be applied after the call to the kernel s finished cpp static void chebyshev window kernel tensoriteratorbase iter t window length double attenuation at dispatch floating types and iter dtype chebyshev window cpu const t n window length const scalar t beta static cast std cosh n std acosh std pow attenuation cpu kernel iter scalar t a auto x beta static cast std cos pi a window length return static cast chebyshev polynomial t forward x n std pow attenuation and for the cosine window function does not need a kernel i believe cpp tensor cosine window t window length bool periodic optional dtype opt optional layout optional device optional pin memory see scalartype dtype dtype or default dtype opt tensoroptions options tensoroptions dtype dtype layout layout device device pinned memory pin memory window function checks cosine window options window length if window length return at empty options if window length return native ones dtype layout device pin memory if periodic window length auto window native arange window length dtype layout device pin memory add mul pi mul static cast window length sin return periodic window narrow window length window the other window functions would follow the same or similar approach additional context references window functions kaiser bessel derived window bosi marina and richard e goldberg introduction to digital audio coding and standards dordrecht kluwer dolph chebyshev window taylor window armin doerry catalog of window taper functions for sidelobe control nutall window a nuttall “some windows with very good sidelobe behavior ” ieee transactions on acoustics speech and signal processing vol no pp feb
1
306,282
23,153,368,426
IssuesEvent
2022-07-29 10:30:14
glamod/glamod_landQC
https://api.github.com/repos/glamod/glamod_landQC
closed
Update documentation
documentation
Update documentation to ensure that the steps are clear to non-expert users.
1.0
Update documentation - Update documentation to ensure that the steps are clear to non-expert users.
non_comp
update documentation update documentation to ensure that the steps are clear to non expert users
0
11,627
13,675,463,735
IssuesEvent
2020-09-29 12:43:19
Scholar-6/brillder
https://api.github.com/repos/Scholar-6/brillder
closed
Homepage icons still too small on 2560x1600 laptop
Adjustments Device/Browser Compatibility
<img width="1918" alt="Screenshot 2020-09-23 at 13 35 17" src="https://user-images.githubusercontent.com/59654112/94007356-a0f20900-fda1-11ea-9e2f-038966f4ad08.png"> - [ ] reduce icons by 10% for standard HD resolution - [ ] match to new sizing on 2560x1600 laptop
True
Homepage icons still too small on 2560x1600 laptop - <img width="1918" alt="Screenshot 2020-09-23 at 13 35 17" src="https://user-images.githubusercontent.com/59654112/94007356-a0f20900-fda1-11ea-9e2f-038966f4ad08.png"> - [ ] reduce icons by 10% for standard HD resolution - [ ] match to new sizing on 2560x1600 laptop
comp
homepage icons still too small on laptop img width alt screenshot at src reduce icons by for standard hd resolution match to new sizing on laptop
1
17,449
24,064,581,311
IssuesEvent
2022-09-17 09:41:11
ValveSoftware/Proton
https://api.github.com/repos/ValveSoftware/Proton
closed
Bail or Jail Demo (1881400)
Game compatibility - Unofficial
# Compatibility Report - Name of the game with compatibility issues: Bail or Jail Demo - Steam AppID of the game: 1881400 - URL: https://store.steampowered.com/app/1715980/Bail_or_Jail/ ## System Information - GPU: GTX 970 - Driver/LLVM version: 515.49.14 - Kernel version: 5.19 - Link to full system information report as [Gist](https://gist.github.com/): https://gist.github.com/Riesi/87e8ece0a060b053b94755f8485640e7 - Proton version: 7.0-4 or Bleeding Edge ## I confirm: - [X] that I haven't found an existing compatibility report for this game. - [X] that I have checked whether there are updates for my system available. [steam-1881400_crash.log](https://github.com/ValveSoftware/Proton/files/9583390/steam-1881400_crash.log) [steam-1881400_nocrash.log](https://github.com/ValveSoftware/Proton/files/9583391/steam-1881400_nocrash.log) ## Symptoms <!-- What's the problem? --> The game crashes after the first few greeting screens when opening the Terms of Service. Reproduction of the crash should take a few seconds. The "nocrash" log is from exiting the game before clicking the Terms of Service button. ## Reproduction 1. Start the demo and go through the initial greetings dialog screens until it asks to show you the Terms of Service. 2. Click the Terms of Service button 3. Game crashes
True
Bail or Jail Demo (1881400) - # Compatibility Report - Name of the game with compatibility issues: Bail or Jail Demo - Steam AppID of the game: 1881400 - URL: https://store.steampowered.com/app/1715980/Bail_or_Jail/ ## System Information - GPU: GTX 970 - Driver/LLVM version: 515.49.14 - Kernel version: 5.19 - Link to full system information report as [Gist](https://gist.github.com/): https://gist.github.com/Riesi/87e8ece0a060b053b94755f8485640e7 - Proton version: 7.0-4 or Bleeding Edge ## I confirm: - [X] that I haven't found an existing compatibility report for this game. - [X] that I have checked whether there are updates for my system available. [steam-1881400_crash.log](https://github.com/ValveSoftware/Proton/files/9583390/steam-1881400_crash.log) [steam-1881400_nocrash.log](https://github.com/ValveSoftware/Proton/files/9583391/steam-1881400_nocrash.log) ## Symptoms <!-- What's the problem? --> The game crashes after the first few greeting screens when opening the Terms of Service. Reproduction of the crash should take a few seconds. The "nocrash" log is from exiting the game before clicking the Terms of Service button. ## Reproduction 1. Start the demo and go through the initial greetings dialog screens until it asks to show you the Terms of Service. 2. Click the Terms of Service button 3. Game crashes
comp
bail or jail demo compatibility report name of the game with compatibility issues bail or jail demo steam appid of the game url system information gpu gtx driver llvm version kernel version link to full system information report as proton version or bleeding edge i confirm that i haven t found an existing compatibility report for this game that i have checked whether there are updates for my system available symptoms the game crashes after the first few greeting screens when opening the terms of service reproduction of the crash should take a few seconds the nocrash log is from exiting the game before clicking the terms of service button reproduction start the demo and go through the initial greetings dialog screens until it asks to show you the terms of service click the terms of service button game crashes
1
73,503
9,668,319,241
IssuesEvent
2019-05-21 14:55:32
gajus/eslint-plugin-jsdoc
https://api.github.com/repos/gajus/eslint-plugin-jsdoc
closed
Create a CONTRIBUTING.md file
Documentation
It would be a lot easier for someone to contribute if there was a document that explained some basic aspects of the project. 1. How to test changes locally (npm link) 1. How to build the project 1. Coding standards 1. Are tests needed? If so, how to write them. 1. Requirements for PR Of course, most of these can just be links to other resources, like `npm link`. The point of this document would allow someone who knows javascript, but doesn't necessarily know this project, how to contribute. This document is not a guide of how to write rules for eslint.
1.0
Create a CONTRIBUTING.md file - It would be a lot easier for someone to contribute if there was a document that explained some basic aspects of the project. 1. How to test changes locally (npm link) 1. How to build the project 1. Coding standards 1. Are tests needed? If so, how to write them. 1. Requirements for PR Of course, most of these can just be links to other resources, like `npm link`. The point of this document would allow someone who knows javascript, but doesn't necessarily know this project, how to contribute. This document is not a guide of how to write rules for eslint.
non_comp
create a contributing md file it would be a lot easier for someone to contribute if there was a document that explained some basic aspects of the project how to test changes locally npm link how to build the project coding standards are tests needed if so how to write them requirements for pr of course most of these can just be links to other resources like npm link the point of this document would allow someone who knows javascript but doesn t necessarily know this project how to contribute this document is not a guide of how to write rules for eslint
0
16,845
23,187,062,828
IssuesEvent
2022-08-01 09:17:52
myparcelnl/magento
https://api.github.com/repos/myparcelnl/magento
closed
Add PHP 8.1 compatibility
Next version compatibility released
### Plugin/theme name PHP ### Pricing Free ### What problem does this feature solve? So we can upgrade to Magento 2.4.4 ### What should the solution look like? PHP 8.1 compatibility ### Relevant log output Here's the output of the PHPCompatibility code sniffer for PHP 8.0 and 8.1: ```shell vendor/bin/phpcs -p vendor/myparcelnl/magento --standard=vendor/phpcompatibility/php-compatibility/PHPCompatibility --extensions=php,phtml --runtime-set testVersion 8.0 ............................................................ 60 / 95 (63%) ................................... 95 / 95 (100%) Time: 2.65 secs; Memory: 14MB vendor/bin/phpcs -p vendor/myparcelnl/magento --standard=vendor/phpcompatibility/php-compatibility/PHPCompatibility --extensions=php,phtml --runtime-set testVersion 8.1 ............................................................ 60 / 95 (63%) ................................... 95 / 95 (100%) Time: 2.65 secs; Memory: 14MB ``` So looks like updating the constraints and we're good to go! PR here; https://github.com/myparcelnl/magento/pull/688
True
Add PHP 8.1 compatibility - ### Plugin/theme name PHP ### Pricing Free ### What problem does this feature solve? So we can upgrade to Magento 2.4.4 ### What should the solution look like? PHP 8.1 compatibility ### Relevant log output Here's the output of the PHPCompatibility code sniffer for PHP 8.0 and 8.1: ```shell vendor/bin/phpcs -p vendor/myparcelnl/magento --standard=vendor/phpcompatibility/php-compatibility/PHPCompatibility --extensions=php,phtml --runtime-set testVersion 8.0 ............................................................ 60 / 95 (63%) ................................... 95 / 95 (100%) Time: 2.65 secs; Memory: 14MB vendor/bin/phpcs -p vendor/myparcelnl/magento --standard=vendor/phpcompatibility/php-compatibility/PHPCompatibility --extensions=php,phtml --runtime-set testVersion 8.1 ............................................................ 60 / 95 (63%) ................................... 95 / 95 (100%) Time: 2.65 secs; Memory: 14MB ``` So looks like updating the constraints and we're good to go! PR here; https://github.com/myparcelnl/magento/pull/688
comp
add php compatibility plugin theme name php pricing free what problem does this feature solve so we can upgrade to magento what should the solution look like php compatibility relevant log output here s the output of the phpcompatibility code sniffer for php and shell vendor bin phpcs p vendor myparcelnl magento standard vendor phpcompatibility php compatibility phpcompatibility extensions php phtml runtime set testversion time secs memory vendor bin phpcs p vendor myparcelnl magento standard vendor phpcompatibility php compatibility phpcompatibility extensions php phtml runtime set testversion time secs memory so looks like updating the constraints and we re good to go pr here
1
10,637
12,549,873,812
IssuesEvent
2020-06-06 08:53:39
widelands/widelands
https://api.github.com/repos/widelands/widelands
closed
Segfault loading save game
bug crashes or hangs saveloading & compatibility seafaring urgent
When loading a game saved using commit 3b49fc9 I'm getting a segfault; with both a manual save and an autosave. If I load previous saves, which used a build from commits 2ce3222 and prior, the saves load OK. I've verified this by reading the `preload` file in the saves. Attached are the map, "aD 25.wgf" (crashing save), "aD 24.wgf" (successful save) and the stdout/err from a debug build of commit 3b49fc9. [save-crash.zip](https://github.com/widelands/widelands/files/4441922/save-crash.zip)
True
Segfault loading save game - When loading a game saved using commit 3b49fc9 I'm getting a segfault; with both a manual save and an autosave. If I load previous saves, which used a build from commits 2ce3222 and prior, the saves load OK. I've verified this by reading the `preload` file in the saves. Attached are the map, "aD 25.wgf" (crashing save), "aD 24.wgf" (successful save) and the stdout/err from a debug build of commit 3b49fc9. [save-crash.zip](https://github.com/widelands/widelands/files/4441922/save-crash.zip)
comp
segfault loading save game when loading a game saved using commit i m getting a segfault with both a manual save and an autosave if i load previous saves which used a build from commits and prior the saves load ok i ve verified this by reading the preload file in the saves attached are the map ad wgf crashing save ad wgf successful save and the stdout err from a debug build of commit
1
11,571
13,610,659,372
IssuesEvent
2020-09-23 07:43:30
MetaMask/metamask-mobile
https://api.github.com/repos/MetaMask/metamask-mobile
closed
bancor.network not accessible
bug dapp-compatibility
**Describe the bug** When trying to load https://bancor.network it will load eternally **To Reproduce** Open tab and type for the url: https://bancor.network **Expected behavior** Dowload the webapp and have Metamask usable on it :-) **Smartphone (please complete the following information):** - Device: Nokia 6.1 - OS: Android one / Android 10 ------------------------------------------------------------- **Severity** Very severe it renders the user incapable to use the app for using the bancor network. - Is this visible to all users? All users who might want to use the bancor network
True
bancor.network not accessible - **Describe the bug** When trying to load https://bancor.network it will load eternally **To Reproduce** Open tab and type for the url: https://bancor.network **Expected behavior** Dowload the webapp and have Metamask usable on it :-) **Smartphone (please complete the following information):** - Device: Nokia 6.1 - OS: Android one / Android 10 ------------------------------------------------------------- **Severity** Very severe it renders the user incapable to use the app for using the bancor network. - Is this visible to all users? All users who might want to use the bancor network
comp
bancor network not accessible describe the bug when trying to load it will load eternally to reproduce open tab and type for the url expected behavior dowload the webapp and have metamask usable on it smartphone please complete the following information device nokia os android one android severity very severe it renders the user incapable to use the app for using the bancor network is this visible to all users all users who might want to use the bancor network
1
134,502
5,229,247,204
IssuesEvent
2017-01-29 00:49:48
abentele/Fraise
https://api.github.com/repos/abentele/Fraise
closed
Automatic updates (maybe distribute using the App Store)
feature request priority:high
automatic updates didn't work for a long time => automatic updates removed (temporary) with release 3.7.4. If we would distribute using the Mac App Store, we wouldn't need any proprietary update feature. To be clarified if it would comply with the license of Fraise. Any automatic update feature would require an active maintainer. Update using Sparkle while distributing the binaries on github: - https://sparkle-project.org/ - https://gist.github.com/huangyq23/668e6d6fcccf714e802a
1.0
Automatic updates (maybe distribute using the App Store) - automatic updates didn't work for a long time => automatic updates removed (temporary) with release 3.7.4. If we would distribute using the Mac App Store, we wouldn't need any proprietary update feature. To be clarified if it would comply with the license of Fraise. Any automatic update feature would require an active maintainer. Update using Sparkle while distributing the binaries on github: - https://sparkle-project.org/ - https://gist.github.com/huangyq23/668e6d6fcccf714e802a
non_comp
automatic updates maybe distribute using the app store automatic updates didn t work for a long time automatic updates removed temporary with release if we would distribute using the mac app store we wouldn t need any proprietary update feature to be clarified if it would comply with the license of fraise any automatic update feature would require an active maintainer update using sparkle while distributing the binaries on github
0
18,802
26,164,495,573
IssuesEvent
2023-01-01 04:04:06
jackeys/Power-Armor-to-the-People
https://api.github.com/repos/jackeys/Power-Armor-to-the-People
opened
Patch for ECO
compatibility
- [ ] Add additional attach slots to effects - [ ] Loose mods with swapping ability - [ ] Available for other legendary slots - [ ] Make sure swapping in legendary armor effects works
True
Patch for ECO - - [ ] Add additional attach slots to effects - [ ] Loose mods with swapping ability - [ ] Available for other legendary slots - [ ] Make sure swapping in legendary armor effects works
comp
patch for eco add additional attach slots to effects loose mods with swapping ability available for other legendary slots make sure swapping in legendary armor effects works
1
119,473
10,054,324,421
IssuesEvent
2019-07-22 00:38:27
wesnoth/wesnoth
https://api.github.com/repos/wesnoth/wesnoth
closed
Add the asymmetric theme?
Enhancement Ready for testing UI
Since gloccusv posted his [asymmetric theme](https://forums.wesnoth.org/viewtopic.php?f=6&t=41065&start=15), I've been using a modified version of it ([code](http://sprunge.us/uLV6Pj), [screenshot](https://forums.wesnoth.org/download/file.php?id=83887&mode=view)). My version works best on master (because it uses some of the features from #3852). I've considered packaging it [as an add-on](https://forums.wesnoth.org/viewtopic.php?f=21&t=50213) but I wonder if it'll be easier to just add it to mainline? *edit* That code patch is just what I'm using right now in my personal branch. I am **not** proposing to just apply that to master as-is; if the concept is acceptable, I'll clean the patch up before merging it. - [ ] On 2560x1440 the left bar is 1106 pixels high, not full length.
1.0
Add the asymmetric theme? - Since gloccusv posted his [asymmetric theme](https://forums.wesnoth.org/viewtopic.php?f=6&t=41065&start=15), I've been using a modified version of it ([code](http://sprunge.us/uLV6Pj), [screenshot](https://forums.wesnoth.org/download/file.php?id=83887&mode=view)). My version works best on master (because it uses some of the features from #3852). I've considered packaging it [as an add-on](https://forums.wesnoth.org/viewtopic.php?f=21&t=50213) but I wonder if it'll be easier to just add it to mainline? *edit* That code patch is just what I'm using right now in my personal branch. I am **not** proposing to just apply that to master as-is; if the concept is acceptable, I'll clean the patch up before merging it. - [ ] On 2560x1440 the left bar is 1106 pixels high, not full length.
non_comp
add the asymmetric theme since gloccusv posted his i ve been using a modified version of it my version works best on master because it uses some of the features from i ve considered packaging it but i wonder if it ll be easier to just add it to mainline edit that code patch is just what i m using right now in my personal branch i am not proposing to just apply that to master as is if the concept is acceptable i ll clean the patch up before merging it on the left bar is pixels high not full length
0
16,039
21,314,067,948
IssuesEvent
2022-04-16 02:00:01
jtothebell/fake-08
https://api.github.com/repos/jtothebell/fake-08
closed
custom pause menus don't seem to work
bug compatibility
Noticed this on Run Gun Bot and PICORACER2048. Not really sure what causes the issues on the latter though.
True
custom pause menus don't seem to work - Noticed this on Run Gun Bot and PICORACER2048. Not really sure what causes the issues on the latter though.
comp
custom pause menus don t seem to work noticed this on run gun bot and not really sure what causes the issues on the latter though
1
16,324
21,950,250,545
IssuesEvent
2022-05-24 07:14:20
jOOQ/jOOQ
https://api.github.com/repos/jOOQ/jOOQ
closed
DSLContext.fetch(ResultSet, Class...) and similar methods should be lenient regarding row size
T: Enhancement C: Functionality P: Medium R: Wontfix T: Incompatible change E: All Editions
The current implementations of `DSLContext.fetch(ResultSet, Class...)`, etc. throw an exception if there are too many varargs arguments compared to the number of columns in the `ResultSet`. There is no problem passing fewer arguments, though, so there's a certain irregularity in the behaviour. It would be quite convenient to be able to pass more arguments than available from the `ResultSet` in some cases, e.g. when querying JDBC `DatabaseMetaData` result sets that may not be implemented completely by JDBC drivers. Affected methods include - [ ] `DSLContext.fetch(ResultSet, X...)` - [ ] `DSLContext.fetchOne(ResultSet, X...)` - [ ] `DSLContext.fetchSingle(ResultSet, X...)` - [ ] `DSLContext.fetchOptional(ResultSet, X...)` - [ ] `DSLContext.fetchLazy(ResultSet, X...)` - [ ] `DSLContext.fetchAsync(ResultSet, X...)` - [ ] `DSLContext.fetchAsync(Executor, ResultSet, X...)` - [ ] `DSLContext.fetchStream(ResultSet, X...)` In the absence of an actual field in the `ResultSet`, a dummy name would need to be assumed, unless the name is given by the `Field<?>...` overload ### Risk Users may rely on the current API throwing exceptions. Perhaps, this new behaviour should be introduced only when activating a `Settings`?
True
DSLContext.fetch(ResultSet, Class...) and similar methods should be lenient regarding row size - The current implementations of `DSLContext.fetch(ResultSet, Class...)`, etc. throw an exception if there are too many varargs arguments compared to the number of columns in the `ResultSet`. There is no problem passing fewer arguments, though, so there's a certain irregularity in the behaviour. It would be quite convenient to be able to pass more arguments than available from the `ResultSet` in some cases, e.g. when querying JDBC `DatabaseMetaData` result sets that may not be implemented completely by JDBC drivers. Affected methods include - [ ] `DSLContext.fetch(ResultSet, X...)` - [ ] `DSLContext.fetchOne(ResultSet, X...)` - [ ] `DSLContext.fetchSingle(ResultSet, X...)` - [ ] `DSLContext.fetchOptional(ResultSet, X...)` - [ ] `DSLContext.fetchLazy(ResultSet, X...)` - [ ] `DSLContext.fetchAsync(ResultSet, X...)` - [ ] `DSLContext.fetchAsync(Executor, ResultSet, X...)` - [ ] `DSLContext.fetchStream(ResultSet, X...)` In the absence of an actual field in the `ResultSet`, a dummy name would need to be assumed, unless the name is given by the `Field<?>...` overload ### Risk Users may rely on the current API throwing exceptions. Perhaps, this new behaviour should be introduced only when activating a `Settings`?
comp
dslcontext fetch resultset class and similar methods should be lenient regarding row size the current implementations of dslcontext fetch resultset class etc throw an exception if there are too many varargs arguments compared to the number of columns in the resultset there is no problem passing fewer arguments though so there s a certain irregularity in the behaviour it would be quite convenient to be able to pass more arguments than available from the resultset in some cases e g when querying jdbc databasemetadata result sets that may not be implemented completely by jdbc drivers affected methods include dslcontext fetch resultset x dslcontext fetchone resultset x dslcontext fetchsingle resultset x dslcontext fetchoptional resultset x dslcontext fetchlazy resultset x dslcontext fetchasync resultset x dslcontext fetchasync executor resultset x dslcontext fetchstream resultset x in the absence of an actual field in the resultset a dummy name would need to be assumed unless the name is given by the field overload risk users may rely on the current api throwing exceptions perhaps this new behaviour should be introduced only when activating a settings
1
15,713
19,848,804,290
IssuesEvent
2022-01-21 09:56:43
ooi-data/CE06ISSP-SP001-08-FLORTJ000-recovered_cspp-flort_sample
https://api.github.com/repos/ooi-data/CE06ISSP-SP001-08-FLORTJ000-recovered_cspp-flort_sample
opened
🛑 Processing failed: ValueError
process
## Overview `ValueError` found in `processing_task` task during run ended on 2022-01-21T09:56:43.128543. ## Details Flow name: `CE06ISSP-SP001-08-FLORTJ000-recovered_cspp-flort_sample` Task name: `processing_task` Error type: `ValueError` Error message: not enough values to unpack (expected 3, got 0) <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing final_path = finalize_data_stream( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream append_to_zarr(mod_ds, final_store, enc, logger=logger) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr _append_zarr(store, mod_ds) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr existing_arr.append(var_data.values) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values return _as_array_or_item(self._data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item data = np.asarray(data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__ x = self.compute() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute (result,) = compute(self, traverse=False, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute results = schedule(dsk, keys, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get results = get_async( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async raise_exception(exc, tb) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise raise exc File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task result = _execute_task(task, data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task return func(*(_execute_task(a, cache) for a in args)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter c = np.asarray(c) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__ self._ensure_cached() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached self.array = NumpyIndexingAdapter(np.asarray(self.array)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__ return array[key.tuple] File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__ return self.get_basic_selection(selection, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection return self._get_basic_selection_nd(selection=selection, out=out, File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd return self._get_selection(indexer=indexer, out=out, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection lchunk_coords, lchunk_selection, lout_selection = zip(*indexer) ValueError: not enough values to unpack (expected 3, got 0) ``` </details>
1.0
🛑 Processing failed: ValueError - ## Overview `ValueError` found in `processing_task` task during run ended on 2022-01-21T09:56:43.128543. ## Details Flow name: `CE06ISSP-SP001-08-FLORTJ000-recovered_cspp-flort_sample` Task name: `processing_task` Error type: `ValueError` Error message: not enough values to unpack (expected 3, got 0) <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing final_path = finalize_data_stream( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream append_to_zarr(mod_ds, final_store, enc, logger=logger) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr _append_zarr(store, mod_ds) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr existing_arr.append(var_data.values) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values return _as_array_or_item(self._data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item data = np.asarray(data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__ x = self.compute() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute (result,) = compute(self, traverse=False, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute results = schedule(dsk, keys, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get results = get_async( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async raise_exception(exc, tb) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise raise exc File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task result = _execute_task(task, data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task return func(*(_execute_task(a, cache) for a in args)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter c = np.asarray(c) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__ self._ensure_cached() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached self.array = NumpyIndexingAdapter(np.asarray(self.array)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__ return array[key.tuple] File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__ return self.get_basic_selection(selection, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection return self._get_basic_selection_nd(selection=selection, out=out, File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd return self._get_selection(indexer=indexer, out=out, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection lchunk_coords, lchunk_selection, lout_selection = zip(*indexer) ValueError: not enough values to unpack (expected 3, got 0) ``` </details>
non_comp
🛑 processing failed valueerror overview valueerror found in processing task task during run ended on details flow name recovered cspp flort sample task name processing task error type valueerror error message not enough values to unpack expected got traceback traceback most recent call last file srv conda envs notebook lib site packages ooi harvester processor pipeline py line in processing final path finalize data stream file srv conda envs notebook lib site packages ooi harvester processor init py line in finalize data stream append to zarr mod ds final store enc logger logger file srv conda envs notebook lib site packages ooi harvester processor init py line in append to zarr append zarr store mod ds file srv conda envs notebook lib site packages ooi harvester processor utils py line in append zarr existing arr append var data values file srv conda envs notebook lib site packages xarray core variable py line in values return as array or item self data file srv conda envs notebook lib site packages xarray core variable py line in as array or item data np asarray data file srv conda envs notebook lib site packages dask array core py line in array x self compute file srv conda envs notebook lib site packages dask base py line in compute result compute self traverse false kwargs file srv conda envs notebook lib site packages dask base py line in compute results schedule dsk keys kwargs file srv conda envs notebook lib site packages dask threaded py line in get results get async file srv conda envs notebook lib site packages dask local py line in get async raise exception exc tb file srv conda envs notebook lib site packages dask local py line in reraise raise exc file srv conda envs notebook lib site packages dask local py line in execute task result execute task task data file srv conda envs notebook lib site packages dask core py line in execute task return func execute task a cache for a in args file srv conda envs notebook lib site packages dask array core py line in getter c np asarray c file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array self ensure cached file srv conda envs notebook lib site packages xarray core indexing py line in ensure cached self array numpyindexingadapter np asarray self array file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray backends zarr py line in getitem return array file srv conda envs notebook lib site packages zarr core py line in getitem return self get basic selection selection fields fields file srv conda envs notebook lib site packages zarr core py line in get basic selection return self get basic selection nd selection selection out out file srv conda envs notebook lib site packages zarr core py line in get basic selection nd return self get selection indexer indexer out out fields fields file srv conda envs notebook lib site packages zarr core py line in get selection lchunk coords lchunk selection lout selection zip indexer valueerror not enough values to unpack expected got
0
49,120
12,290,298,595
IssuesEvent
2020-05-10 02:58:38
tensorflow/tensorflow
https://api.github.com/repos/tensorflow/tensorflow
opened
bazel build; encountered error while reading extension file 'swift/repositories.bzl': no such package '@build_bazel_rules_swift//swif
type:build/install
**System information** - OS Platform and Distribution: Ubuntu 18.04 - TensorFlow installed from: source - TensorFlow version: 1.14.0 - Python version:3.7.7 - Installed using virtualenv? pip? conda?: conda - Bazel version: 0.24.1 - trying to build tensorflow cpu not gpu I followed the intructions and based my versions found here: https://www.tensorflow.org/install/source#tested_build_configurations **Describe the problem** When i run the command ``` bazel build --config=opt //tensorflow/tools/pip_package:build_pip_package ``` it gives the following error: ``` Starting local Bazel server and connecting to it... ERROR: error loading package '': Encountered error while reading extension file 'swift/repositories.bzl': no such package '@build_bazel_rules_swift//swift': java.io.IOException: Error downloading [https://github.com/bazelbuild/rules_swift/releases/download/0.9.0/rules_swift.0.9.0.tar.gz] to /home/domingo_cm/.cache/bazel/_bazel_domingo_cm/f08b2f8b39197d5a57f7b557c17f0caf/external/build_bazel_rules_swift/rules_swift.0.9.0.tar.gz: Unknown host: github.com ERROR: error loading package '': Encountered error while reading extension file 'swift/repositories.bzl': no such package '@build_bazel_rules_swift//swift': java.io.IOException: Error downloading [https://github.com/bazelbuild/rules_swift/releases/download/0.9.0/rules_swift.0.9.0.tar.gz] to /home/domingo_cm/.cache/bazel/_bazel_domingo_cm/f08b2f8b39197d5a57f7b557c17f0caf/external/build_bazel_rules_swift/rules_swift.0.9.0.tar.gz: Unknown host: github.com INFO: Elapsed time: 14.145s INFO: 0 processes. FAILED: Build did NOT complete successfully (0 packages loaded) ``` **Provide the exact sequence of commands / steps that you executed before running into the problem** 1. git clone https://github.com/tensorflow/tensorflow.git 2. cd tensorflow 3. git checkout r1.14 4. ./configure 5. (assuing bazel is installed ) bazel build --config=opt //tensorflow/tools/pip_package:build_pip_package **Any other info / logs** I am running under a corporate proxy network if it gives any importance, git version is 2.17.1
1.0
bazel build; encountered error while reading extension file 'swift/repositories.bzl': no such package '@build_bazel_rules_swift//swif - **System information** - OS Platform and Distribution: Ubuntu 18.04 - TensorFlow installed from: source - TensorFlow version: 1.14.0 - Python version:3.7.7 - Installed using virtualenv? pip? conda?: conda - Bazel version: 0.24.1 - trying to build tensorflow cpu not gpu I followed the intructions and based my versions found here: https://www.tensorflow.org/install/source#tested_build_configurations **Describe the problem** When i run the command ``` bazel build --config=opt //tensorflow/tools/pip_package:build_pip_package ``` it gives the following error: ``` Starting local Bazel server and connecting to it... ERROR: error loading package '': Encountered error while reading extension file 'swift/repositories.bzl': no such package '@build_bazel_rules_swift//swift': java.io.IOException: Error downloading [https://github.com/bazelbuild/rules_swift/releases/download/0.9.0/rules_swift.0.9.0.tar.gz] to /home/domingo_cm/.cache/bazel/_bazel_domingo_cm/f08b2f8b39197d5a57f7b557c17f0caf/external/build_bazel_rules_swift/rules_swift.0.9.0.tar.gz: Unknown host: github.com ERROR: error loading package '': Encountered error while reading extension file 'swift/repositories.bzl': no such package '@build_bazel_rules_swift//swift': java.io.IOException: Error downloading [https://github.com/bazelbuild/rules_swift/releases/download/0.9.0/rules_swift.0.9.0.tar.gz] to /home/domingo_cm/.cache/bazel/_bazel_domingo_cm/f08b2f8b39197d5a57f7b557c17f0caf/external/build_bazel_rules_swift/rules_swift.0.9.0.tar.gz: Unknown host: github.com INFO: Elapsed time: 14.145s INFO: 0 processes. FAILED: Build did NOT complete successfully (0 packages loaded) ``` **Provide the exact sequence of commands / steps that you executed before running into the problem** 1. git clone https://github.com/tensorflow/tensorflow.git 2. cd tensorflow 3. git checkout r1.14 4. ./configure 5. (assuing bazel is installed ) bazel build --config=opt //tensorflow/tools/pip_package:build_pip_package **Any other info / logs** I am running under a corporate proxy network if it gives any importance, git version is 2.17.1
non_comp
bazel build encountered error while reading extension file swift repositories bzl no such package build bazel rules swift swif system information os platform and distribution ubuntu tensorflow installed from source tensorflow version python version installed using virtualenv pip conda conda bazel version trying to build tensorflow cpu not gpu i followed the intructions and based my versions found here describe the problem when i run the command bazel build config opt tensorflow tools pip package build pip package it gives the following error starting local bazel server and connecting to it error error loading package encountered error while reading extension file swift repositories bzl no such package build bazel rules swift swift java io ioexception error downloading to home domingo cm cache bazel bazel domingo cm external build bazel rules swift rules swift tar gz unknown host github com error error loading package encountered error while reading extension file swift repositories bzl no such package build bazel rules swift swift java io ioexception error downloading to home domingo cm cache bazel bazel domingo cm external build bazel rules swift rules swift tar gz unknown host github com info elapsed time info processes failed build did not complete successfully packages loaded provide the exact sequence of commands steps that you executed before running into the problem git clone cd tensorflow git checkout configure assuing bazel is installed bazel build config opt tensorflow tools pip package build pip package any other info logs i am running under a corporate proxy network if it gives any importance git version is
0
94,608
19,564,143,560
IssuesEvent
2022-01-03 20:49:18
HookCycle/BITcc
https://api.github.com/repos/HookCycle/BITcc
closed
ChatBox
code
means of communication between the student and the educator who is mentoring your TCC, or any other teacher who is available to be your TCC mentor
1.0
ChatBox - means of communication between the student and the educator who is mentoring your TCC, or any other teacher who is available to be your TCC mentor
non_comp
chatbox means of communication between the student and the educator who is mentoring your tcc or any other teacher who is available to be your tcc mentor
0
304,111
9,321,276,463
IssuesEvent
2019-03-27 03:07:53
k8smeetup/website-tasks
https://api.github.com/repos/k8smeetup/website-tasks
closed
/docs/reference/kubernetes-api/labels-annotations-taints.md
doc/core finished lang/zh priority/P0 version/1.12
Path:`/docs/reference/kubernetes-api/labels-annotations-taints.md` [Source code](https://github.com/kubernetes/website/tree/release-1.12/content/en//docs/reference/kubernetes-api/labels-annotations-taints.md)
1.0
/docs/reference/kubernetes-api/labels-annotations-taints.md - Path:`/docs/reference/kubernetes-api/labels-annotations-taints.md` [Source code](https://github.com/kubernetes/website/tree/release-1.12/content/en//docs/reference/kubernetes-api/labels-annotations-taints.md)
non_comp
docs reference kubernetes api labels annotations taints md path: docs reference kubernetes api labels annotations taints md
0
4,916
7,529,469,100
IssuesEvent
2018-04-14 05:15:58
mob-sakai/UIEffect
https://api.github.com/repos/mob-sakai/UIEffect
closed
Change: Color effect use vertex color to reduce parameters.
backwards-incompatible unreleased
Color effect use vertex color to reduce parameters and to avoid conflict with other components. Add `Multiply (default)` color effect.
True
Change: Color effect use vertex color to reduce parameters. - Color effect use vertex color to reduce parameters and to avoid conflict with other components. Add `Multiply (default)` color effect.
comp
change color effect use vertex color to reduce parameters color effect use vertex color to reduce parameters and to avoid conflict with other components add multiply default color effect
1
13,612
16,134,960,479
IssuesEvent
2021-04-29 10:33:38
jtothebell/fake-08
https://api.github.com/repos/jtothebell/fake-08
closed
Sound on Wii U
bug compatibility
Sound on the Wii U is currently turned off because it sounds mixed up. The sounds sound like the right pitches, but out of order. Since Wii U is an SDL 2 platform, my first suspicion is that it is a problem with the Wii U being Big Endian, but I've been focusing on other issues and haven't dug too deep.
True
Sound on Wii U - Sound on the Wii U is currently turned off because it sounds mixed up. The sounds sound like the right pitches, but out of order. Since Wii U is an SDL 2 platform, my first suspicion is that it is a problem with the Wii U being Big Endian, but I've been focusing on other issues and haven't dug too deep.
comp
sound on wii u sound on the wii u is currently turned off because it sounds mixed up the sounds sound like the right pitches but out of order since wii u is an sdl platform my first suspicion is that it is a problem with the wii u being big endian but i ve been focusing on other issues and haven t dug too deep
1
18,434
25,496,865,187
IssuesEvent
2022-11-27 19:39:00
openlayers/openlayers
https://api.github.com/repos/openlayers/openlayers
opened
Remove @api annotations from webgl utility classes
backwards incompatible
In the 7.0 release, we meant to remove all the @api annotations from things related to webgl rendering (see https://github.com/openlayers/openlayers/pull/13461#issuecomment-1085076127), but we overlooked the following: * `ol/webgl/Buffer.js` * `ol/webgl/PostProcessingPass.js` * `ol/webgl/RenderTarget.js`
True
Remove @api annotations from webgl utility classes - In the 7.0 release, we meant to remove all the @api annotations from things related to webgl rendering (see https://github.com/openlayers/openlayers/pull/13461#issuecomment-1085076127), but we overlooked the following: * `ol/webgl/Buffer.js` * `ol/webgl/PostProcessingPass.js` * `ol/webgl/RenderTarget.js`
comp
remove api annotations from webgl utility classes in the release we meant to remove all the api annotations from things related to webgl rendering see but we overlooked the following ol webgl buffer js ol webgl postprocessingpass js ol webgl rendertarget js
1
387,783
11,470,376,486
IssuesEvent
2020-02-09 02:32:45
strongloop/loopback-sdk-angular
https://api.github.com/repos/strongloop/loopback-sdk-angular
closed
methodNames with "dashes" cause syntax error
bug good first issue help wanted major needs-priority
Our project, uses api names that have dashes in them. An example is: `/sliding-doors/`. When the sdk generator runs, it pulls the names from the rest'ful API, instead of the actual model name (maybe), causing there to be a syntax error in the resulting .js file. The error lies here: https://github.com/strongloop/loopback-sdk-angular/blob/master/lib/services.template.ejs#L260 Notice the bad template syntax here... ``` R.<%- methodName %> = function() { var TargetResource = $injector.get(<%-: targetClass | q %>); var action = TargetResource[<%-: action.name | q %>]; return action.apply(R, arguments); }; ``` It should look something like... ``` R.['<%- methodName %>'] = function() { ``` But, methodName is also a complex string with "dots" in it, so a better solution is needed
1.0
methodNames with "dashes" cause syntax error - Our project, uses api names that have dashes in them. An example is: `/sliding-doors/`. When the sdk generator runs, it pulls the names from the rest'ful API, instead of the actual model name (maybe), causing there to be a syntax error in the resulting .js file. The error lies here: https://github.com/strongloop/loopback-sdk-angular/blob/master/lib/services.template.ejs#L260 Notice the bad template syntax here... ``` R.<%- methodName %> = function() { var TargetResource = $injector.get(<%-: targetClass | q %>); var action = TargetResource[<%-: action.name | q %>]; return action.apply(R, arguments); }; ``` It should look something like... ``` R.['<%- methodName %>'] = function() { ``` But, methodName is also a complex string with "dots" in it, so a better solution is needed
non_comp
methodnames with dashes cause syntax error our project uses api names that have dashes in them an example is sliding doors when the sdk generator runs it pulls the names from the rest ful api instead of the actual model name maybe causing there to be a syntax error in the resulting js file the error lies here notice the bad template syntax here r function var targetresource injector get var action targetresource return action apply r arguments it should look something like r function but methodname is also a complex string with dots in it so a better solution is needed
0
110,656
16,985,730,458
IssuesEvent
2021-06-30 14:11:58
turkdevops/prettier
https://api.github.com/repos/turkdevops/prettier
opened
CVE-2021-27515 (Medium) detected in url-parse-1.4.7.tgz
security vulnerability
## CVE-2021-27515 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>url-parse-1.4.7.tgz</b></p></summary> <p>Small footprint URL parser that works seamlessly across Node.js and browser environments</p> <p>Library home page: <a href="https://registry.npmjs.org/url-parse/-/url-parse-1.4.7.tgz">https://registry.npmjs.org/url-parse/-/url-parse-1.4.7.tgz</a></p> <p>Path to dependency file: prettier/website/package.json</p> <p>Path to vulnerable library: prettier/website/node_modules/url-parse</p> <p> Dependency Hierarchy: - docusaurus-1.14.6.tgz (Root Library) - react-dev-utils-9.1.0.tgz - sockjs-client-1.4.0.tgz - :x: **url-parse-1.4.7.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/turkdevops/prettier/commit/23d97cf18fada2c5a6f280543bbf99493ca0e60f">23d97cf18fada2c5a6f280543bbf99493ca0e60f</a></p> <p>Found in base branch: <b>patch-release</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> url-parse before 1.5.0 mishandles certain uses of backslash such as http:\/ and interprets the URI as a relative path. <p>Publish Date: 2021-02-22 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-27515>CVE-2021-27515</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-27515">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-27515</a></p> <p>Release Date: 2021-02-22</p> <p>Fix Resolution: 1.5.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-27515 (Medium) detected in url-parse-1.4.7.tgz - ## CVE-2021-27515 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>url-parse-1.4.7.tgz</b></p></summary> <p>Small footprint URL parser that works seamlessly across Node.js and browser environments</p> <p>Library home page: <a href="https://registry.npmjs.org/url-parse/-/url-parse-1.4.7.tgz">https://registry.npmjs.org/url-parse/-/url-parse-1.4.7.tgz</a></p> <p>Path to dependency file: prettier/website/package.json</p> <p>Path to vulnerable library: prettier/website/node_modules/url-parse</p> <p> Dependency Hierarchy: - docusaurus-1.14.6.tgz (Root Library) - react-dev-utils-9.1.0.tgz - sockjs-client-1.4.0.tgz - :x: **url-parse-1.4.7.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/turkdevops/prettier/commit/23d97cf18fada2c5a6f280543bbf99493ca0e60f">23d97cf18fada2c5a6f280543bbf99493ca0e60f</a></p> <p>Found in base branch: <b>patch-release</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> url-parse before 1.5.0 mishandles certain uses of backslash such as http:\/ and interprets the URI as a relative path. <p>Publish Date: 2021-02-22 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-27515>CVE-2021-27515</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-27515">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-27515</a></p> <p>Release Date: 2021-02-22</p> <p>Fix Resolution: 1.5.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_comp
cve medium detected in url parse tgz cve medium severity vulnerability vulnerable library url parse tgz small footprint url parser that works seamlessly across node js and browser environments library home page a href path to dependency file prettier website package json path to vulnerable library prettier website node modules url parse dependency hierarchy docusaurus tgz root library react dev utils tgz sockjs client tgz x url parse tgz vulnerable library found in head commit a href found in base branch patch release vulnerability details url parse before mishandles certain uses of backslash such as http and interprets the uri as a relative path publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
52,993
22,586,906,403
IssuesEvent
2022-06-28 15:58:34
cityofaustin/atd-data-tech
https://api.github.com/repos/cityofaustin/atd-data-tech
closed
[BUG] "Source" field in funding line items is required but there's no indication for user
Workgroup: AMD Type: Bug Report Impact: 3-Minor Service: Dev Product: Moped Type: Snackoo 🍫
This will be resolved by not making "Source" required. Not urgent at all because Ivonne & Renée have an easy workaround in using the "Other" option in the dropdown. --- Original request from Ivonne: > Describe the problem. check mark to save info doesnt come up -- funding tab (screenshot attached) > Website Address https://mobility.austin.gov/moped/projects/174?tab=funding Internet Browser: Chrome > Requested By Ivonne N. [Attachment](https://api.knack.com/v1/applications/595d00ebd315cc4cb98daff4/download/asset/6286bbeef13bbb001ec3acd9/mopedfundingtab.jpg) (118.28kb) Request ID: DTS22-104040
1.0
[BUG] "Source" field in funding line items is required but there's no indication for user - This will be resolved by not making "Source" required. Not urgent at all because Ivonne & Renée have an easy workaround in using the "Other" option in the dropdown. --- Original request from Ivonne: > Describe the problem. check mark to save info doesnt come up -- funding tab (screenshot attached) > Website Address https://mobility.austin.gov/moped/projects/174?tab=funding Internet Browser: Chrome > Requested By Ivonne N. [Attachment](https://api.knack.com/v1/applications/595d00ebd315cc4cb98daff4/download/asset/6286bbeef13bbb001ec3acd9/mopedfundingtab.jpg) (118.28kb) Request ID: DTS22-104040
non_comp
source field in funding line items is required but there s no indication for user this will be resolved by not making source required not urgent at all because ivonne renée have an easy workaround in using the other option in the dropdown original request from ivonne describe the problem check mark to save info doesnt come up funding tab screenshot attached website address internet browser chrome requested by ivonne n request id
0
7,873
10,090,564,110
IssuesEvent
2019-07-26 12:02:36
pdepend/pdepend
https://api.github.com/repos/pdepend/pdepend
closed
Unexpected token error calling invokable property
Bug Compatiblity php5.x
This seems to throw an "unexpected token" exception when you are calling an invokable object that is a property of a class. In the sample files below, the error is thrown from the `index` method of the `Controller` class where the call is being made. It seems to be because the syntax isn't being handled properly. Other tooling such as PHPStan, PHPStorm, and PHP CS don't see a token problem and the code executes properly within PHP itself. <details> <summary>InjectedThing.php</summary> ```php <?php namespace App; class InjectedThing { public function __invoke(string $name, string $place): void { echo "Hello $place, $name"; } } ``` </details> <details> <summary>Controller.php</summary> ```php <?php namespace App; class Controller { /** @var \App\InjectedThing $thing */ private $thing; public function __construct() { $this->thing = new InjectedThing(); } public function index(): void { ($this->thing)('Garbee', 'Universe'); } } ``` </details>
True
Unexpected token error calling invokable property - This seems to throw an "unexpected token" exception when you are calling an invokable object that is a property of a class. In the sample files below, the error is thrown from the `index` method of the `Controller` class where the call is being made. It seems to be because the syntax isn't being handled properly. Other tooling such as PHPStan, PHPStorm, and PHP CS don't see a token problem and the code executes properly within PHP itself. <details> <summary>InjectedThing.php</summary> ```php <?php namespace App; class InjectedThing { public function __invoke(string $name, string $place): void { echo "Hello $place, $name"; } } ``` </details> <details> <summary>Controller.php</summary> ```php <?php namespace App; class Controller { /** @var \App\InjectedThing $thing */ private $thing; public function __construct() { $this->thing = new InjectedThing(); } public function index(): void { ($this->thing)('Garbee', 'Universe'); } } ``` </details>
comp
unexpected token error calling invokable property this seems to throw an unexpected token exception when you are calling an invokable object that is a property of a class in the sample files below the error is thrown from the index method of the controller class where the call is being made it seems to be because the syntax isn t being handled properly other tooling such as phpstan phpstorm and php cs don t see a token problem and the code executes properly within php itself injectedthing php php php namespace app class injectedthing public function invoke string name string place void echo hello place name controller php php php namespace app class controller var app injectedthing thing private thing public function construct this thing new injectedthing public function index void this thing garbee universe
1