Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
1k
labels
stringlengths
4
1.38k
body
stringlengths
1
262k
index
stringclasses
16 values
text_combine
stringlengths
96
262k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
298,550
9,200,507,932
IssuesEvent
2019-03-07 17:13:09
qissue-bot/QGIS
https://api.github.com/repos/qissue-bot/QGIS
closed
Run actions directly from the map canvas
Category: Map Canvas Component: Easy fix? Component: Pull Request or Patch supplied Component: Resolution Priority: Low Project: QGIS Application Status: Closed Tracker: Feature request
--- Author Name: **Gavin Macaulay -** (Gavin Macaulay -) Original Redmine Issue: 119, https://issues.qgis.org/issues/119 Original Assignee: Giuseppe Sucameli --- It would be nice to be able to run actions by clicking on the map canvas aka identify but instead getting a list of the various actions that could be run on the identified features.
1.0
Run actions directly from the map canvas - --- Author Name: **Gavin Macaulay -** (Gavin Macaulay -) Original Redmine Issue: 119, https://issues.qgis.org/issues/119 Original Assignee: Giuseppe Sucameli --- It would be nice to be able to run actions by clicking on the map canvas aka identify but instead getting a list of the various actions that could be run on the identified features.
priority
run actions directly from the map canvas author name gavin macaulay gavin macaulay original redmine issue original assignee giuseppe sucameli it would be nice to be able to run actions by clicking on the map canvas aka identify but instead getting a list of the various actions that could be run on the identified features
1
777,702
27,291,473,770
IssuesEvent
2023-02-23 16:54:27
department-of-veterans-affairs/abd-vro
https://api.github.com/repos/department-of-veterans-affairs/abd-vro
closed
Remove rrd_mas_disability_tracking and rrd_mas_notification feature toggles
Engineer Low-priority
**Notes about work** - Follow-up to #576
1.0
Remove rrd_mas_disability_tracking and rrd_mas_notification feature toggles - **Notes about work** - Follow-up to #576
priority
remove rrd mas disability tracking and rrd mas notification feature toggles notes about work follow up to
1
150,401
5,766,046,925
IssuesEvent
2017-04-27 05:31:55
bosik/diacomp
https://api.github.com/repos/bosik/diacomp
opened
Possibility to select meal display format in diary
app:android priority:minor type:idea user-mentioned
E.g.: - most carbohydrated food only (current) - summ of carbohydrates (like "85 g"), maybe with BU ("85 g / 7 BU")
1.0
Possibility to select meal display format in diary - E.g.: - most carbohydrated food only (current) - summ of carbohydrates (like "85 g"), maybe with BU ("85 g / 7 BU")
priority
possibility to select meal display format in diary e g most carbohydrated food only current summ of carbohydrates like g maybe with bu g bu
1
55,940
14,074,855,260
IssuesEvent
2020-11-04 08:06:38
teena24/WebGoat
https://api.github.com/repos/teena24/WebGoat
opened
CVE-2019-10746 (High) detected in mixin-deep-1.3.1.tgz
security vulnerability
## CVE-2019-10746 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mixin-deep-1.3.1.tgz</b></p></summary> <p>Deeply mix the properties of objects into the first object. Like merge-deep, but doesn't clone.</p> <p>Library home page: <a href="https://registry.npmjs.org/mixin-deep/-/mixin-deep-1.3.1.tgz">https://registry.npmjs.org/mixin-deep/-/mixin-deep-1.3.1.tgz</a></p> <p>Path to dependency file: WebGoat/docs/package.json</p> <p>Path to vulnerable library: WebGoat/docs/node_modules/mixin-deep/package.json</p> <p> Dependency Hierarchy: - browser-sync-2.26.3.tgz (Root Library) - chokidar-2.0.4.tgz - braces-2.3.2.tgz - snapdragon-0.8.2.tgz - base-0.11.2.tgz - :x: **mixin-deep-1.3.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/teena24/WebGoat/commit/b8a568f6e08fcde3c08370e69ce7236fef395ad5">b8a568f6e08fcde3c08370e69ce7236fef395ad5</a></p> <p>Found in base branch: <b>develop</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> mixin-deep is vulnerable to Prototype Pollution in versions before 1.3.2 and version 2.0.0. The function mixin-deep could be tricked into adding or modifying properties of Object.prototype using a constructor payload. <p>Publish Date: 2019-08-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10746>CVE-2019-10746</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/jonschlinkert/mixin-deep/commit/8f464c8ce9761a8c9c2b3457eaeee9d404fa7af9">https://github.com/jonschlinkert/mixin-deep/commit/8f464c8ce9761a8c9c2b3457eaeee9d404fa7af9</a></p> <p>Release Date: 2019-07-11</p> <p>Fix Resolution: 1.3.2,2.0.1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"mixin-deep","packageVersion":"1.3.1","isTransitiveDependency":true,"dependencyTree":"browser-sync:2.26.3;chokidar:2.0.4;braces:2.3.2;snapdragon:0.8.2;base:0.11.2;mixin-deep:1.3.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.3.2,2.0.1"}],"vulnerabilityIdentifier":"CVE-2019-10746","vulnerabilityDetails":"mixin-deep is vulnerable to Prototype Pollution in versions before 1.3.2 and version 2.0.0. The function mixin-deep could be tricked into adding or modifying properties of Object.prototype using a constructor payload.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10746","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2019-10746 (High) detected in mixin-deep-1.3.1.tgz - ## CVE-2019-10746 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mixin-deep-1.3.1.tgz</b></p></summary> <p>Deeply mix the properties of objects into the first object. Like merge-deep, but doesn't clone.</p> <p>Library home page: <a href="https://registry.npmjs.org/mixin-deep/-/mixin-deep-1.3.1.tgz">https://registry.npmjs.org/mixin-deep/-/mixin-deep-1.3.1.tgz</a></p> <p>Path to dependency file: WebGoat/docs/package.json</p> <p>Path to vulnerable library: WebGoat/docs/node_modules/mixin-deep/package.json</p> <p> Dependency Hierarchy: - browser-sync-2.26.3.tgz (Root Library) - chokidar-2.0.4.tgz - braces-2.3.2.tgz - snapdragon-0.8.2.tgz - base-0.11.2.tgz - :x: **mixin-deep-1.3.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/teena24/WebGoat/commit/b8a568f6e08fcde3c08370e69ce7236fef395ad5">b8a568f6e08fcde3c08370e69ce7236fef395ad5</a></p> <p>Found in base branch: <b>develop</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> mixin-deep is vulnerable to Prototype Pollution in versions before 1.3.2 and version 2.0.0. The function mixin-deep could be tricked into adding or modifying properties of Object.prototype using a constructor payload. <p>Publish Date: 2019-08-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10746>CVE-2019-10746</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/jonschlinkert/mixin-deep/commit/8f464c8ce9761a8c9c2b3457eaeee9d404fa7af9">https://github.com/jonschlinkert/mixin-deep/commit/8f464c8ce9761a8c9c2b3457eaeee9d404fa7af9</a></p> <p>Release Date: 2019-07-11</p> <p>Fix Resolution: 1.3.2,2.0.1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"mixin-deep","packageVersion":"1.3.1","isTransitiveDependency":true,"dependencyTree":"browser-sync:2.26.3;chokidar:2.0.4;braces:2.3.2;snapdragon:0.8.2;base:0.11.2;mixin-deep:1.3.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.3.2,2.0.1"}],"vulnerabilityIdentifier":"CVE-2019-10746","vulnerabilityDetails":"mixin-deep is vulnerable to Prototype Pollution in versions before 1.3.2 and version 2.0.0. The function mixin-deep could be tricked into adding or modifying properties of Object.prototype using a constructor payload.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10746","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_priority
cve high detected in mixin deep tgz cve high severity vulnerability vulnerable library mixin deep tgz deeply mix the properties of objects into the first object like merge deep but doesn t clone library home page a href path to dependency file webgoat docs package json path to vulnerable library webgoat docs node modules mixin deep package json dependency hierarchy browser sync tgz root library chokidar tgz braces tgz snapdragon tgz base tgz x mixin deep tgz vulnerable library found in head commit a href found in base branch develop vulnerability details mixin deep is vulnerable to prototype pollution in versions before and version the function mixin deep could be tricked into adding or modifying properties of object prototype using a constructor payload publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails mixin deep is vulnerable to prototype pollution in versions before and version the function mixin deep could be tricked into adding or modifying properties of object prototype using a constructor payload vulnerabilityurl
0
667,087
22,408,766,467
IssuesEvent
2022-06-18 11:46:39
dec0dOS/zero-ui
https://api.github.com/repos/dec0dOS/zero-ui
closed
feat: ref to zyclonite/zerotier in docker-compose
new-feature beginner-friendly priority-medium
# Feature Request **Describe the Feature Request** as https://hub.docker.com/r/dec0dos/zerotier-controller is quite outdated, perhaps its easier to just refer to latest `zyclonite/zerotier` **Describe Preferred Solution** refer to in https://github.com/dec0dOS/zero-ui/blob/main/docker-compose.yml#L5 to `zyclonite/zerotier`
1.0
feat: ref to zyclonite/zerotier in docker-compose - # Feature Request **Describe the Feature Request** as https://hub.docker.com/r/dec0dos/zerotier-controller is quite outdated, perhaps its easier to just refer to latest `zyclonite/zerotier` **Describe Preferred Solution** refer to in https://github.com/dec0dOS/zero-ui/blob/main/docker-compose.yml#L5 to `zyclonite/zerotier`
priority
feat ref to zyclonite zerotier in docker compose feature request describe the feature request as is quite outdated perhaps its easier to just refer to latest zyclonite zerotier describe preferred solution refer to in to zyclonite zerotier
1
272,695
29,795,072,415
IssuesEvent
2023-06-16 01:08:43
billmcchesney1/pacbot
https://api.github.com/repos/billmcchesney1/pacbot
closed
CVE-2022-37603 (High) detected in loader-utils-1.1.0.tgz - autoclosed
Mend: dependency security vulnerability
## CVE-2022-37603 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>loader-utils-1.1.0.tgz</b></p></summary> <p>utils for webpack loaders</p> <p>Library home page: <a href="https://registry.npmjs.org/loader-utils/-/loader-utils-1.1.0.tgz">https://registry.npmjs.org/loader-utils/-/loader-utils-1.1.0.tgz</a></p> <p>Path to dependency file: /webapp/package.json</p> <p>Path to vulnerable library: /webapp/node_modules/loader-utils/package.json</p> <p> Dependency Hierarchy: - cli-1.6.8.tgz (Root Library) - :x: **loader-utils-1.1.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/billmcchesney1/pacbot/commit/acf9a0620c1a37cee4f2896d71e1c3731c5c7b06">acf9a0620c1a37cee4f2896d71e1c3731c5c7b06</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> A Regular expression denial of service (ReDoS) flaw was found in Function interpolateName in interpolateName.js in webpack loader-utils 2.0.0 via the url variable in interpolateName.js. <p>Publish Date: 2022-10-14 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-37603>CVE-2022-37603</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-3rfm-jhwj-7488">https://github.com/advisories/GHSA-3rfm-jhwj-7488</a></p> <p>Release Date: 2022-10-14</p> <p>Fix Resolution: loader-utils - 1.4.2,2.0.4,3.2.1</p> </p> </details> <p></p>
True
CVE-2022-37603 (High) detected in loader-utils-1.1.0.tgz - autoclosed - ## CVE-2022-37603 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>loader-utils-1.1.0.tgz</b></p></summary> <p>utils for webpack loaders</p> <p>Library home page: <a href="https://registry.npmjs.org/loader-utils/-/loader-utils-1.1.0.tgz">https://registry.npmjs.org/loader-utils/-/loader-utils-1.1.0.tgz</a></p> <p>Path to dependency file: /webapp/package.json</p> <p>Path to vulnerable library: /webapp/node_modules/loader-utils/package.json</p> <p> Dependency Hierarchy: - cli-1.6.8.tgz (Root Library) - :x: **loader-utils-1.1.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/billmcchesney1/pacbot/commit/acf9a0620c1a37cee4f2896d71e1c3731c5c7b06">acf9a0620c1a37cee4f2896d71e1c3731c5c7b06</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> A Regular expression denial of service (ReDoS) flaw was found in Function interpolateName in interpolateName.js in webpack loader-utils 2.0.0 via the url variable in interpolateName.js. <p>Publish Date: 2022-10-14 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-37603>CVE-2022-37603</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-3rfm-jhwj-7488">https://github.com/advisories/GHSA-3rfm-jhwj-7488</a></p> <p>Release Date: 2022-10-14</p> <p>Fix Resolution: loader-utils - 1.4.2,2.0.4,3.2.1</p> </p> </details> <p></p>
non_priority
cve high detected in loader utils tgz autoclosed cve high severity vulnerability vulnerable library loader utils tgz utils for webpack loaders library home page a href path to dependency file webapp package json path to vulnerable library webapp node modules loader utils package json dependency hierarchy cli tgz root library x loader utils tgz vulnerable library found in head commit a href found in base branch master vulnerability details a regular expression denial of service redos flaw was found in function interpolatename in interpolatename js in webpack loader utils via the url variable in interpolatename js publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution loader utils
0
139,323
11,257,988,309
IssuesEvent
2020-01-13 02:22:03
NMGRL/pychron
https://api.github.com/repos/NMGRL/pychron
closed
When plotting air blanks in the Fit Blanks pipeline I get this error
Bug Testing Required
active branch=develop active analyses=ba-01-N-2244,ba-01-N-2245,ba-01-N-2246,ba-01-N-2247,ba-01-N-2248,ba-01-N-2249,ba-01-N-2250,ba-01-N-2251,ba-01-N-2252,ba-01-N-2253,ba-01-N-2254,ba-01-N-2255,ba-01-N-2256,ba-01-N-2257,ba-01-N-2258,ba-01-N-2259,ba-01-N-2260,ba-01-N-2261,ba-01-N-2262,ba-01-N-2263,ba-01-N-2264,ba-01-N-2265,ba-01-N-2266,ba-01-N-2267,ba-01-N-2268,ba-01-N-2269,ba-01-N-2270,ba-01-N-2271,ba-01-N-2272,ba-01-N-2273,ba-01-N-2274 description="" Traceback ``` Traceback (most recent call last): File "/Users/matthewheizler/miniconda3/envs/pychron3/lib/python3.5/site-packages/traits/trait_notifiers.py", line 519, in _dispatch_change_event self.dispatch( handler, *args ) File "/Users/matthewheizler/miniconda3/envs/pychron3/lib/python3.5/site-packages/traits/trait_notifiers.py", line 482, in dispatch handler( *args ) File "/Users/matthewheizler/.pychron.0/pychron/pychron/pipeline/tasks/task.py", line 649, in _handle_run_needed self.run() File "/Users/matthewheizler/.pychron.0/pychron/pychron/pipeline/tasks/task.py", line 349, in run self._run_pipeline() File "/Users/matthewheizler/.pychron.0/pychron/pychron/pipeline/tasks/task.py", line 536, in _run_pipeline self._run('run pipeline', 'run_pipeline') File "/Users/matthewheizler/.pychron.0/pychron/pychron/pipeline/tasks/task.py", line 515, in _run if not getattr(self.engine, func)(): File "/Users/matthewheizler/.pychron.0/pychron/pychron/pipeline/engine.py", line 723, in run_pipeline node.run(state) File "/Users/matthewheizler/.pychron.0/pychron/pychron/pipeline/nodes/fit.py", line 118, in run self.editor.force_update(force=True) File "/Users/matthewheizler/.pychron.0/pychron/pychron/pipeline/plot/editors/figure_editor.py", line 85, in force_update model.refresh(force=force) File "/Users/matthewheizler/.pychron.0/pychron/pychron/pipeline/plot/models/figure_model.py", line 43, in refresh p.make_graph() File "/Users/matthewheizler/.pychron.0/pychron/pychron/pipeline/plot/panels/figure_panel.py", line 118, in make_graph fig.build(plots) File "/Users/matthewheizler/.pychron.0/pychron/pychron/pipeline/plot/plotter/arar_figure.py", line 141, in build self._setup_plot(i, p, po) TypeError: _setup_plot() missing 1 required positional argument: 'ytitle' ```
1.0
When plotting air blanks in the Fit Blanks pipeline I get this error - active branch=develop active analyses=ba-01-N-2244,ba-01-N-2245,ba-01-N-2246,ba-01-N-2247,ba-01-N-2248,ba-01-N-2249,ba-01-N-2250,ba-01-N-2251,ba-01-N-2252,ba-01-N-2253,ba-01-N-2254,ba-01-N-2255,ba-01-N-2256,ba-01-N-2257,ba-01-N-2258,ba-01-N-2259,ba-01-N-2260,ba-01-N-2261,ba-01-N-2262,ba-01-N-2263,ba-01-N-2264,ba-01-N-2265,ba-01-N-2266,ba-01-N-2267,ba-01-N-2268,ba-01-N-2269,ba-01-N-2270,ba-01-N-2271,ba-01-N-2272,ba-01-N-2273,ba-01-N-2274 description="" Traceback ``` Traceback (most recent call last): File "/Users/matthewheizler/miniconda3/envs/pychron3/lib/python3.5/site-packages/traits/trait_notifiers.py", line 519, in _dispatch_change_event self.dispatch( handler, *args ) File "/Users/matthewheizler/miniconda3/envs/pychron3/lib/python3.5/site-packages/traits/trait_notifiers.py", line 482, in dispatch handler( *args ) File "/Users/matthewheizler/.pychron.0/pychron/pychron/pipeline/tasks/task.py", line 649, in _handle_run_needed self.run() File "/Users/matthewheizler/.pychron.0/pychron/pychron/pipeline/tasks/task.py", line 349, in run self._run_pipeline() File "/Users/matthewheizler/.pychron.0/pychron/pychron/pipeline/tasks/task.py", line 536, in _run_pipeline self._run('run pipeline', 'run_pipeline') File "/Users/matthewheizler/.pychron.0/pychron/pychron/pipeline/tasks/task.py", line 515, in _run if not getattr(self.engine, func)(): File "/Users/matthewheizler/.pychron.0/pychron/pychron/pipeline/engine.py", line 723, in run_pipeline node.run(state) File "/Users/matthewheizler/.pychron.0/pychron/pychron/pipeline/nodes/fit.py", line 118, in run self.editor.force_update(force=True) File "/Users/matthewheizler/.pychron.0/pychron/pychron/pipeline/plot/editors/figure_editor.py", line 85, in force_update model.refresh(force=force) File "/Users/matthewheizler/.pychron.0/pychron/pychron/pipeline/plot/models/figure_model.py", line 43, in refresh p.make_graph() File "/Users/matthewheizler/.pychron.0/pychron/pychron/pipeline/plot/panels/figure_panel.py", line 118, in make_graph fig.build(plots) File "/Users/matthewheizler/.pychron.0/pychron/pychron/pipeline/plot/plotter/arar_figure.py", line 141, in build self._setup_plot(i, p, po) TypeError: _setup_plot() missing 1 required positional argument: 'ytitle' ```
non_priority
when plotting air blanks in the fit blanks pipeline i get this error active branch develop active analyses ba n ba n ba n ba n ba n ba n ba n ba n ba n ba n ba n ba n ba n ba n ba n ba n ba n ba n ba n ba n ba n ba n ba n ba n ba n ba n ba n ba n ba n ba n ba n description traceback traceback most recent call last file users matthewheizler envs lib site packages traits trait notifiers py line in dispatch change event self dispatch handler args file users matthewheizler envs lib site packages traits trait notifiers py line in dispatch handler args file users matthewheizler pychron pychron pychron pipeline tasks task py line in handle run needed self run file users matthewheizler pychron pychron pychron pipeline tasks task py line in run self run pipeline file users matthewheizler pychron pychron pychron pipeline tasks task py line in run pipeline self run run pipeline run pipeline file users matthewheizler pychron pychron pychron pipeline tasks task py line in run if not getattr self engine func file users matthewheizler pychron pychron pychron pipeline engine py line in run pipeline node run state file users matthewheizler pychron pychron pychron pipeline nodes fit py line in run self editor force update force true file users matthewheizler pychron pychron pychron pipeline plot editors figure editor py line in force update model refresh force force file users matthewheizler pychron pychron pychron pipeline plot models figure model py line in refresh p make graph file users matthewheizler pychron pychron pychron pipeline plot panels figure panel py line in make graph fig build plots file users matthewheizler pychron pychron pychron pipeline plot plotter arar figure py line in build self setup plot i p po typeerror setup plot missing required positional argument ytitle
0
742,677
25,865,827,147
IssuesEvent
2022-12-13 20:48:16
Alberto-X3/Alberto-X3-V3
https://api.github.com/repos/Alberto-X3/Alberto-X3-V3
opened
[EXT REQ] File Case
extension request priority: high
### Extension Name file_case ### What are the benefits of having such an extension? Please describe. clearing out the "Aktenfälle" from inside the staff rows ### Describe the solution you'd like maybe entries inside the database and a management system for admins ### Checklist - [X] I have searched the open extension requests for ``duplicates``. - [X] I have searched the closed extension requests for ``wontadd``'s. ### Additional Information _No response_
1.0
[EXT REQ] File Case - ### Extension Name file_case ### What are the benefits of having such an extension? Please describe. clearing out the "Aktenfälle" from inside the staff rows ### Describe the solution you'd like maybe entries inside the database and a management system for admins ### Checklist - [X] I have searched the open extension requests for ``duplicates``. - [X] I have searched the closed extension requests for ``wontadd``'s. ### Additional Information _No response_
priority
file case extension name file case what are the benefits of having such an extension please describe clearing out the aktenfälle from inside the staff rows describe the solution you d like maybe entries inside the database and a management system for admins checklist i have searched the open extension requests for duplicates i have searched the closed extension requests for wontadd s additional information no response
1
6,037
2,610,219,807
IssuesEvent
2015-02-26 19:09:51
chrsmith/somefinders
https://api.github.com/repos/chrsmith/somefinders
opened
вязание спицами для полных женщин схемы.txt
auto-migrated Priority-Medium Type-Defect
``` '''Гермоген Емельянов''' День добрый никак не могу найти .вязание спицами для полных женщин схемы.txt. как то выкладывали уже '''Владелин Нестеров''' Вот хороший сайт где можно скачать http://bit.ly/1csH2v4 '''Герман Белов''' Просит ввести номер мобилы!Не опасно ли это? '''Варлам Соловьёв''' Неа все ок у меня ничего не списало '''Гектор Мельников''' Не это не влияет на баланс Информация о файле: вязание спицами для полных женщин схемы.txt Загружен: В этом месяце Скачан раз: 999 Рейтинг: 1073 Средняя скорость скачивания: 713 Похожих файлов: 17 ``` ----- Original issue reported on code.google.com by `kondense...@gmail.com` on 17 Dec 2013 at 5:12
1.0
вязание спицами для полных женщин схемы.txt - ``` '''Гермоген Емельянов''' День добрый никак не могу найти .вязание спицами для полных женщин схемы.txt. как то выкладывали уже '''Владелин Нестеров''' Вот хороший сайт где можно скачать http://bit.ly/1csH2v4 '''Герман Белов''' Просит ввести номер мобилы!Не опасно ли это? '''Варлам Соловьёв''' Неа все ок у меня ничего не списало '''Гектор Мельников''' Не это не влияет на баланс Информация о файле: вязание спицами для полных женщин схемы.txt Загружен: В этом месяце Скачан раз: 999 Рейтинг: 1073 Средняя скорость скачивания: 713 Похожих файлов: 17 ``` ----- Original issue reported on code.google.com by `kondense...@gmail.com` on 17 Dec 2013 at 5:12
non_priority
вязание спицами для полных женщин схемы txt гермоген емельянов день добрый никак не могу найти вязание спицами для полных женщин схемы txt как то выкладывали уже владелин нестеров вот хороший сайт где можно скачать герман белов просит ввести номер мобилы не опасно ли это варлам соловьёв неа все ок у меня ничего не списало гектор мельников не это не влияет на баланс информация о файле вязание спицами для полных женщин схемы txt загружен в этом месяце скачан раз рейтинг средняя скорость скачивания похожих файлов original issue reported on code google com by kondense gmail com on dec at
0
55,811
6,492,930,234
IssuesEvent
2017-08-21 15:06:51
minishift/minishift
https://api.github.com/repos/minishift/minishift
closed
Update "executing" steps in integration tests to accept commands with quotes
component/integration-test kind/task priority/minor
Currently nearly all steps which execute a command have problem with commands containing quotes. ```golang s.Step(`([^"]*) of command "minishift ([^"]*)" (is equal|is not equal) to "([^"]*)"$`, commandReturnEquals) ``` Most commands for minishift do not use quotes, but there are some exceptions, for example `minishift openshift config set` where following argument must be a JSON, which uses quotes. And this will not be matched properly by adequate steps in test suite. This can be probably fixed by using matching group and its surroundings similar to this: ```"minishift (.*)"```, which is already added in one place into tests by @hferentschik.
1.0
Update "executing" steps in integration tests to accept commands with quotes - Currently nearly all steps which execute a command have problem with commands containing quotes. ```golang s.Step(`([^"]*) of command "minishift ([^"]*)" (is equal|is not equal) to "([^"]*)"$`, commandReturnEquals) ``` Most commands for minishift do not use quotes, but there are some exceptions, for example `minishift openshift config set` where following argument must be a JSON, which uses quotes. And this will not be matched properly by adequate steps in test suite. This can be probably fixed by using matching group and its surroundings similar to this: ```"minishift (.*)"```, which is already added in one place into tests by @hferentschik.
non_priority
update executing steps in integration tests to accept commands with quotes currently nearly all steps which execute a command have problem with commands containing quotes golang s step of command minishift is equal is not equal to commandreturnequals most commands for minishift do not use quotes but there are some exceptions for example minishift openshift config set where following argument must be a json which uses quotes and this will not be matched properly by adequate steps in test suite this can be probably fixed by using matching group and its surroundings similar to this minishift which is already added in one place into tests by hferentschik
0
440,373
12,698,801,885
IssuesEvent
2020-06-22 13:58:20
qutebrowser/qutebrowser
https://api.github.com/repos/qutebrowser/qutebrowser
opened
Use classes rather than functions for completion models
component: completion component: extensions component: style / refactoring priority: 1 - middle
We currently use completion functions which then return a configured completion model, and pass the function to `@cmdutils.register`: ```python @cmdutils.argument('topic', completion=miscmodels.helptopic) ``` However, that feels a bit awkward - we can't "configure" the function from the decorator, so there are some places where we have various completion functions then using a private "utility" function - also we define deletion functions inside those completion functions. What if those were a class instead? Then we could use subclassing where appropriate, and pass arguments from the decorator to the completion class. Also deletion functions could then just be a method of that class. cc @rcorre
1.0
Use classes rather than functions for completion models - We currently use completion functions which then return a configured completion model, and pass the function to `@cmdutils.register`: ```python @cmdutils.argument('topic', completion=miscmodels.helptopic) ``` However, that feels a bit awkward - we can't "configure" the function from the decorator, so there are some places where we have various completion functions then using a private "utility" function - also we define deletion functions inside those completion functions. What if those were a class instead? Then we could use subclassing where appropriate, and pass arguments from the decorator to the completion class. Also deletion functions could then just be a method of that class. cc @rcorre
priority
use classes rather than functions for completion models we currently use completion functions which then return a configured completion model and pass the function to cmdutils register python cmdutils argument topic completion miscmodels helptopic however that feels a bit awkward we can t configure the function from the decorator so there are some places where we have various completion functions then using a private utility function also we define deletion functions inside those completion functions what if those were a class instead then we could use subclassing where appropriate and pass arguments from the decorator to the completion class also deletion functions could then just be a method of that class cc rcorre
1
326,414
9,956,122,230
IssuesEvent
2019-07-05 13:06:03
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
mail.google.com - site is not usable
browser-firefox engine-gecko priority-critical type-connection-error-unknown
<!-- @browser: Firefox 68.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; WOW64; rv:68.0) Gecko/20100101 Firefox/68.0 --> <!-- @reported_with: desktop-reporter --> **URL**: https://mail.google.com/mail/u/0/?q=iluve#sent **Browser / Version**: Firefox 68.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes **Problem type**: Site is not usable **Description**: does not connect via firefox **Steps to Reproduce**: i tried opening my mails using Firefox and it refused indicating "corrupted content error" [![Screenshot Description](https://webcompat.com/uploads/2019/7/709a3a0f-ca4c-4f25-8f10-b0ae7cd2d605-thumb.jpeg)](https://webcompat.com/uploads/2019/7/709a3a0f-ca4c-4f25-8f10-b0ae7cd2d605.jpeg) <details> <summary>Browser Configuration</summary> <ul> <li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190627143605</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: beta</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
mail.google.com - site is not usable - <!-- @browser: Firefox 68.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; WOW64; rv:68.0) Gecko/20100101 Firefox/68.0 --> <!-- @reported_with: desktop-reporter --> **URL**: https://mail.google.com/mail/u/0/?q=iluve#sent **Browser / Version**: Firefox 68.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes **Problem type**: Site is not usable **Description**: does not connect via firefox **Steps to Reproduce**: i tried opening my mails using Firefox and it refused indicating "corrupted content error" [![Screenshot Description](https://webcompat.com/uploads/2019/7/709a3a0f-ca4c-4f25-8f10-b0ae7cd2d605-thumb.jpeg)](https://webcompat.com/uploads/2019/7/709a3a0f-ca4c-4f25-8f10-b0ae7cd2d605.jpeg) <details> <summary>Browser Configuration</summary> <ul> <li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190627143605</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: beta</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
priority
mail google com site is not usable url browser version firefox operating system windows tested another browser yes problem type site is not usable description does not connect via firefox steps to reproduce i tried opening my mails using firefox and it refused indicating corrupted content error browser configuration mixed active content blocked false image mem shared true buildid tracking content blocked false gfx webrender blob images true hastouchscreen false mixed passive content blocked false gfx webrender enabled false gfx webrender all false channel beta from with ❤️
1
93,081
3,882,303,516
IssuesEvent
2016-04-13 09:22:21
quintel/etmoses
https://api.github.com/repos/quintel/etmoses
closed
Selecting a week not always shows this week
Bug Priority
This shows the first week of [this LES](https://beta-moses.energytransitionmodel.com/testing_grounds/479) while the second week is selected: ![banners_and_alerts_and_etmoses](https://cloud.githubusercontent.com/assets/1303760/14488714/d022ef02-0169-11e6-9c45-2dfdb0affdcd.png)
1.0
Selecting a week not always shows this week - This shows the first week of [this LES](https://beta-moses.energytransitionmodel.com/testing_grounds/479) while the second week is selected: ![banners_and_alerts_and_etmoses](https://cloud.githubusercontent.com/assets/1303760/14488714/d022ef02-0169-11e6-9c45-2dfdb0affdcd.png)
priority
selecting a week not always shows this week this shows the first week of while the second week is selected
1
19,505
10,361,350,738
IssuesEvent
2019-09-06 09:48:52
hisptz/hris-maintenance-app
https://api.github.com/repos/hisptz/hris-maintenance-app
opened
CVE-2018-20834 (High) detected in tar-2.2.2.tgz
security vulnerability
## CVE-2018-20834 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tar-2.2.2.tgz</b></p></summary> <p>tar for node</p> <p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-2.2.2.tgz">https://registry.npmjs.org/tar/-/tar-2.2.2.tgz</a></p> <p>Path to dependency file: /hris-maintenance-app/package.json</p> <p>Path to vulnerable library: /tmp/git/hris-maintenance-app/node_modules/tar/package.json</p> <p> Dependency Hierarchy: - build-angular-0.13.9.tgz (Root Library) - node-sass-4.12.0.tgz - node-gyp-3.8.0.tgz - :x: **tar-2.2.2.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/hisptz/hris-maintenance-app/commit/9b8ba96bae108a99362590311a55849d18fc964a">9b8ba96bae108a99362590311a55849d18fc964a</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A vulnerability was found in node-tar before version 4.4.2. An Arbitrary File Overwrite issue exists when extracting a tarball containing a hardlink to a file that already exists on the system, in conjunction with a later plain file with the same name as the hardlink. This plain file content replaces the existing file content. <p>Publish Date: 2019-04-30 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20834>CVE-2018-20834</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://hackerone.com/reports/344595">https://hackerone.com/reports/344595</a></p> <p>Release Date: 2019-04-30</p> <p>Fix Resolution: v4.4.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-20834 (High) detected in tar-2.2.2.tgz - ## CVE-2018-20834 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tar-2.2.2.tgz</b></p></summary> <p>tar for node</p> <p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-2.2.2.tgz">https://registry.npmjs.org/tar/-/tar-2.2.2.tgz</a></p> <p>Path to dependency file: /hris-maintenance-app/package.json</p> <p>Path to vulnerable library: /tmp/git/hris-maintenance-app/node_modules/tar/package.json</p> <p> Dependency Hierarchy: - build-angular-0.13.9.tgz (Root Library) - node-sass-4.12.0.tgz - node-gyp-3.8.0.tgz - :x: **tar-2.2.2.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/hisptz/hris-maintenance-app/commit/9b8ba96bae108a99362590311a55849d18fc964a">9b8ba96bae108a99362590311a55849d18fc964a</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A vulnerability was found in node-tar before version 4.4.2. An Arbitrary File Overwrite issue exists when extracting a tarball containing a hardlink to a file that already exists on the system, in conjunction with a later plain file with the same name as the hardlink. This plain file content replaces the existing file content. <p>Publish Date: 2019-04-30 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20834>CVE-2018-20834</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://hackerone.com/reports/344595">https://hackerone.com/reports/344595</a></p> <p>Release Date: 2019-04-30</p> <p>Fix Resolution: v4.4.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve high detected in tar tgz cve high severity vulnerability vulnerable library tar tgz tar for node library home page a href path to dependency file hris maintenance app package json path to vulnerable library tmp git hris maintenance app node modules tar package json dependency hierarchy build angular tgz root library node sass tgz node gyp tgz x tar tgz vulnerable library found in head commit a href vulnerability details a vulnerability was found in node tar before version an arbitrary file overwrite issue exists when extracting a tarball containing a hardlink to a file that already exists on the system in conjunction with a later plain file with the same name as the hardlink this plain file content replaces the existing file content publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
139,577
5,378,651,123
IssuesEvent
2017-02-23 15:27:17
mantidproject/mantid
https://api.github.com/repos/mantidproject/mantid
closed
User file not detected in batch reduction mode when extension is missing in ISIS SANS GUI
Component: SANS Priority: High
Steve wrote: >I’ve just discovered a bug in the batch reduction. >If you specify a user file in the batch file it is supposed to use that in preference to the one at the top of the GUI. This was an enhancement that was put in for 3.4.1. >It’s no longer working. The reduction is using the main GUI user file all the time. >In the attached screen shot, the bottom graph was generated using the user file _MAIN that was specified at the top of the GUI **and** in the a batch file _MAIN. The data is indeed the main detector. The top graph, however, was generated with user file _MAIN still at the top of the GUI but with _HAB in the batch file. But the data is still the main detector. You’ll also notice that it has appended _rear to the figure legends instead of _front. >The workaround for now is to NOT specify user files in batch files. The functionality is actually fine. There are two issues: 1. The user file which is specified in the batch file did not have a file extension (which is .txt). Mantid's `FileFinder` needs the full file name in order to find the user file. We should search for `.txt` files, i.e. add the ending. 2. The suffix generation in batch mode is a quite odd. If the input variable `combineDet` of `BatchReduce` is `rear` it will append `rear` to all files, if it is `front` it will append `front` and so on. This is clearly meant for SANS2D. In addition this is not changed when the detector selection of changes due to a new user file. We need to hedge for that.
1.0
User file not detected in batch reduction mode when extension is missing in ISIS SANS GUI - Steve wrote: >I’ve just discovered a bug in the batch reduction. >If you specify a user file in the batch file it is supposed to use that in preference to the one at the top of the GUI. This was an enhancement that was put in for 3.4.1. >It’s no longer working. The reduction is using the main GUI user file all the time. >In the attached screen shot, the bottom graph was generated using the user file _MAIN that was specified at the top of the GUI **and** in the a batch file _MAIN. The data is indeed the main detector. The top graph, however, was generated with user file _MAIN still at the top of the GUI but with _HAB in the batch file. But the data is still the main detector. You’ll also notice that it has appended _rear to the figure legends instead of _front. >The workaround for now is to NOT specify user files in batch files. The functionality is actually fine. There are two issues: 1. The user file which is specified in the batch file did not have a file extension (which is .txt). Mantid's `FileFinder` needs the full file name in order to find the user file. We should search for `.txt` files, i.e. add the ending. 2. The suffix generation in batch mode is a quite odd. If the input variable `combineDet` of `BatchReduce` is `rear` it will append `rear` to all files, if it is `front` it will append `front` and so on. This is clearly meant for SANS2D. In addition this is not changed when the detector selection of changes due to a new user file. We need to hedge for that.
priority
user file not detected in batch reduction mode when extension is missing in isis sans gui steve wrote i’ve just discovered a bug in the batch reduction if you specify a user file in the batch file it is supposed to use that in preference to the one at the top of the gui this was an enhancement that was put in for it’s no longer working the reduction is using the main gui user file all the time in the attached screen shot the bottom graph was generated using the user file main that was specified at the top of the gui and in the a batch file main the data is indeed the main detector the top graph however was generated with user file main still at the top of the gui but with hab in the batch file but the data is still the main detector you’ll also notice that it has appended rear to the figure legends instead of front the workaround for now is to not specify user files in batch files the functionality is actually fine there are two issues the user file which is specified in the batch file did not have a file extension which is txt mantid s filefinder needs the full file name in order to find the user file we should search for txt files i e add the ending the suffix generation in batch mode is a quite odd if the input variable combinedet of batchreduce is rear it will append rear to all files if it is front it will append front and so on this is clearly meant for in addition this is not changed when the detector selection of changes due to a new user file we need to hedge for that
1
371,906
25,972,276,417
IssuesEvent
2022-12-19 12:14:33
kostaleonard/populare-iac
https://api.github.com/repos/kostaleonard/populare-iac
closed
Update system diagram
documentation
In #51, we moved the EKS cluster into private subnets. Update the system diagram to reflect that change.
1.0
Update system diagram - In #51, we moved the EKS cluster into private subnets. Update the system diagram to reflect that change.
non_priority
update system diagram in we moved the eks cluster into private subnets update the system diagram to reflect that change
0
276,669
8,607,262,893
IssuesEvent
2018-11-17 20:30:04
tideland/go
https://api.github.com/repos/tideland/go
opened
audit/asserts: Add assert for WaitGroup
Priority: Normal Status: Available Type: Enhancement
Waiting for a `sync.WaitGroup` to be done or a timeout (assert fails).
1.0
audit/asserts: Add assert for WaitGroup - Waiting for a `sync.WaitGroup` to be done or a timeout (assert fails).
priority
audit asserts add assert for waitgroup waiting for a sync waitgroup to be done or a timeout assert fails
1
96,663
8,628,413,697
IssuesEvent
2018-11-21 17:19:39
chamilo/chamilo-lms
https://api.github.com/repos/chamilo/chamilo-lms
closed
Course toolbar in wrong position
Bug Requires testing/validation
### Current behavior / Resultado actual / Résultat actuel ![20181120bandeau02](https://user-images.githubusercontent.com/158935/48764988-86384000-ecb1-11e8-917f-05e4cb5665c7.JPG) ### Expected behavior / Resultado esperado / Résultat attendu ![20181120bandeau01](https://user-images.githubusercontent.com/158935/48764993-8c2e2100-ecb1-11e8-9cf5-972dfc7aa21b.JPG) ### Steps to reproduce / Pasos para reproducir / Étapes pour reproduire Active option "show_toolshortcuts" in settings ### Chamilo Version / Versión de Chamilo / Version de Chamilo 1.11.x
1.0
Course toolbar in wrong position - ### Current behavior / Resultado actual / Résultat actuel ![20181120bandeau02](https://user-images.githubusercontent.com/158935/48764988-86384000-ecb1-11e8-917f-05e4cb5665c7.JPG) ### Expected behavior / Resultado esperado / Résultat attendu ![20181120bandeau01](https://user-images.githubusercontent.com/158935/48764993-8c2e2100-ecb1-11e8-9cf5-972dfc7aa21b.JPG) ### Steps to reproduce / Pasos para reproducir / Étapes pour reproduire Active option "show_toolshortcuts" in settings ### Chamilo Version / Versión de Chamilo / Version de Chamilo 1.11.x
non_priority
course toolbar in wrong position current behavior resultado actual résultat actuel expected behavior resultado esperado résultat attendu steps to reproduce pasos para reproducir étapes pour reproduire active option show toolshortcuts in settings chamilo version versión de chamilo version de chamilo x
0
791,998
27,884,465,351
IssuesEvent
2023-03-21 22:19:56
googleapis/repo-automation-bots
https://api.github.com/repos/googleapis/repo-automation-bots
opened
Can OwlBot service
type: feature request priority: p3
**Is your feature request related to a problem? Please describe.** Currently the OwlBot service creates a new pull request with the "owl-bot-update-lock" label when the post processor image is updated. When the pull requests are piled up in the pull request queue, only the latest pull request needs library owners' attention and the rests are ok to close. <img width="982" alt="Screenshot 2023-03-21 at 6 08 03 PM" src="https://user-images.githubusercontent.com/28604/226753169-5b46ba19-3398-4b40-af18-3ff051419c9a.png"> This fact is not clear to library owners and they often encounter file conflicts in pull requests: <img width="918" alt="Screenshot 2023-03-21 at 6 14 44 PM" src="https://user-images.githubusercontent.com/28604/226753930-c6034615-9cff-468d-a727-160c6918ecdf.png"> Now the library owner, who might not know the mechanism above, manually modify the pull request to find the latest value of the post processor sha256 value and would end up in another conflict in another pull request. (All they had to do was just to merge the latest pull request and close the rest...) **Describe the solution you'd like** Can OwlBot service close existing "owl-bot-update-lock"-labelled pull requests when it creates a new one? **Describe alternatives you've considered** A clear and concise description of any alternative solutions or features you've considered. - Can OwlBot update one existing "owl-bot-update-lock"-labelled pull request when it creates a new one? - Can OwlBot add description in pull request that only the latest pull requests among "owl-bot-update-lock" needs to be merged? **Additional context** Add any other context or screenshots about the feature request here.
1.0
Can OwlBot service - **Is your feature request related to a problem? Please describe.** Currently the OwlBot service creates a new pull request with the "owl-bot-update-lock" label when the post processor image is updated. When the pull requests are piled up in the pull request queue, only the latest pull request needs library owners' attention and the rests are ok to close. <img width="982" alt="Screenshot 2023-03-21 at 6 08 03 PM" src="https://user-images.githubusercontent.com/28604/226753169-5b46ba19-3398-4b40-af18-3ff051419c9a.png"> This fact is not clear to library owners and they often encounter file conflicts in pull requests: <img width="918" alt="Screenshot 2023-03-21 at 6 14 44 PM" src="https://user-images.githubusercontent.com/28604/226753930-c6034615-9cff-468d-a727-160c6918ecdf.png"> Now the library owner, who might not know the mechanism above, manually modify the pull request to find the latest value of the post processor sha256 value and would end up in another conflict in another pull request. (All they had to do was just to merge the latest pull request and close the rest...) **Describe the solution you'd like** Can OwlBot service close existing "owl-bot-update-lock"-labelled pull requests when it creates a new one? **Describe alternatives you've considered** A clear and concise description of any alternative solutions or features you've considered. - Can OwlBot update one existing "owl-bot-update-lock"-labelled pull request when it creates a new one? - Can OwlBot add description in pull request that only the latest pull requests among "owl-bot-update-lock" needs to be merged? **Additional context** Add any other context or screenshots about the feature request here.
priority
can owlbot service is your feature request related to a problem please describe currently the owlbot service creates a new pull request with the owl bot update lock label when the post processor image is updated when the pull requests are piled up in the pull request queue only the latest pull request needs library owners attention and the rests are ok to close img width alt screenshot at pm src this fact is not clear to library owners and they often encounter file conflicts in pull requests img width alt screenshot at pm src now the library owner who might not know the mechanism above manually modify the pull request to find the latest value of the post processor value and would end up in another conflict in another pull request all they had to do was just to merge the latest pull request and close the rest describe the solution you d like can owlbot service close existing owl bot update lock labelled pull requests when it creates a new one describe alternatives you ve considered a clear and concise description of any alternative solutions or features you ve considered can owlbot update one existing owl bot update lock labelled pull request when it creates a new one can owlbot add description in pull request that only the latest pull requests among owl bot update lock needs to be merged additional context add any other context or screenshots about the feature request here
1
282,696
21,315,869,425
IssuesEvent
2022-04-16 09:03:13
kktai1512/pe
https://api.github.com/repos/kktai1512/pe
opened
View documentation not updated
type.DocumentationBug severity.Low
![image.png](https://raw.githubusercontent.com/kktai1512/pe/main/files/6e68b7c1-205a-4d74-bd20-ebe4c5fe9e3c.png) "View " documentation did not update to include other details. <!--session: 1650096894385-df1b1ebe-b59c-4972-98ac-794d61b6fc74--> <!--Version: Web v3.4.2-->
1.0
View documentation not updated - ![image.png](https://raw.githubusercontent.com/kktai1512/pe/main/files/6e68b7c1-205a-4d74-bd20-ebe4c5fe9e3c.png) "View " documentation did not update to include other details. <!--session: 1650096894385-df1b1ebe-b59c-4972-98ac-794d61b6fc74--> <!--Version: Web v3.4.2-->
non_priority
view documentation not updated view documentation did not update to include other details
0
88,224
17,498,878,163
IssuesEvent
2021-08-10 06:45:51
Kerita/kerita.github.io
https://api.github.com/repos/Kerita/kerita.github.io
opened
Redux createStore 参数传递的精妙之处 | Kerita
Gitalk /08/10/read-source-code/redux-createStore/
https://kerita.me/08/10/read-source-code/redux-createStore/#more Redux createStore 是 Redux 初始化函数,在参数传递时有个有意思的实现。 判断 preloadState 是否为函数,如果为函数将其作为 enhancer 处理,并重新赋值 preloadState 为 undefined。
1.0
Redux createStore 参数传递的精妙之处 | Kerita - https://kerita.me/08/10/read-source-code/redux-createStore/#more Redux createStore 是 Redux 初始化函数,在参数传递时有个有意思的实现。 判断 preloadState 是否为函数,如果为函数将其作为 enhancer 处理,并重新赋值 preloadState 为 undefined。
non_priority
redux createstore 参数传递的精妙之处 kerita redux createstore 是 redux 初始化函数,在参数传递时有个有意思的实现。 判断 preloadstate 是否为函数,如果为函数将其作为 enhancer 处理,并重新赋值 preloadstate 为 undefined。
0
63,482
8,681,183,149
IssuesEvent
2018-12-01 17:53:21
Azure/azure-cli
https://api.github.com/repos/Azure/azure-cli
closed
--location name formatting
Documentation
It may seem straightforward, "westus" and "West US", but initially I actually wasn't sure what to use, being from a different location; it may be helpful to put a note on the location-name-formatting. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 58c81c4d-10d2-fc04-e707-fcf914f4d540 * Version Independent ID: 8598f788-bd33-a4c8-ed66-593e0fde75d5 * Content: [az group](https://docs.microsoft.com/en-us/cli/azure/group?view=azure-cli-latest#az-group-create) * Content Source: [src/command_modules/azure-cli-resource/azure/cli/command_modules/resource/_help.py](https://github.com/Azure/azure-cli/blob/master/src/command_modules/azure-cli-resource/azure/cli/command_modules/resource/_help.py) * Service: **azure-resource-manager** * GitHub Login: @rloutlaw * Microsoft Alias: **routlaw**
1.0
--location name formatting - It may seem straightforward, "westus" and "West US", but initially I actually wasn't sure what to use, being from a different location; it may be helpful to put a note on the location-name-formatting. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 58c81c4d-10d2-fc04-e707-fcf914f4d540 * Version Independent ID: 8598f788-bd33-a4c8-ed66-593e0fde75d5 * Content: [az group](https://docs.microsoft.com/en-us/cli/azure/group?view=azure-cli-latest#az-group-create) * Content Source: [src/command_modules/azure-cli-resource/azure/cli/command_modules/resource/_help.py](https://github.com/Azure/azure-cli/blob/master/src/command_modules/azure-cli-resource/azure/cli/command_modules/resource/_help.py) * Service: **azure-resource-manager** * GitHub Login: @rloutlaw * Microsoft Alias: **routlaw**
non_priority
location name formatting it may seem straightforward westus and west us but initially i actually wasn t sure what to use being from a different location it may be helpful to put a note on the location name formatting document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service azure resource manager github login rloutlaw microsoft alias routlaw
0
197,291
14,917,159,024
IssuesEvent
2021-01-22 19:23:45
rancher/rancher
https://api.github.com/repos/rancher/rancher
closed
[2.4] KDM Release - k8s Jan patches
[zube]: QA Working status/to-test
Going with v2.4.13 release: | Kubernetes Version | Rancher Versions | | --------- | ------ | | 1.18.15 | 2.4.5+ | | 1.17.17 | 2.4.4+ |
1.0
[2.4] KDM Release - k8s Jan patches - Going with v2.4.13 release: | Kubernetes Version | Rancher Versions | | --------- | ------ | | 1.18.15 | 2.4.5+ | | 1.17.17 | 2.4.4+ |
non_priority
kdm release jan patches going with release kubernetes version rancher versions
0
436,765
12,553,399,413
IssuesEvent
2020-06-06 21:50:04
space-wizards/RobustToolbox
https://api.github.com/repos/space-wizards/RobustToolbox
closed
CurTime jumps around if you modify server tick rate on a live server.
Priority: 3-low Type: Bug
This is because it's calculated purely as `ticks passed / tickrate`. Changing tickrate changes total time. This is problematic because `CurTime` is used as an absolute time measure server side. If it were to change many things like cooldowns could get stuck or act incorrectly. To reproduce: 1. Spawn a bike horn 2. `cvar net.tickrate 10` in the server. 3. Honk the horn 4. `cvar net.tickrate 60` in the server. 5. Your bike horn cooldown is now stuck on a huge value.
1.0
CurTime jumps around if you modify server tick rate on a live server. - This is because it's calculated purely as `ticks passed / tickrate`. Changing tickrate changes total time. This is problematic because `CurTime` is used as an absolute time measure server side. If it were to change many things like cooldowns could get stuck or act incorrectly. To reproduce: 1. Spawn a bike horn 2. `cvar net.tickrate 10` in the server. 3. Honk the horn 4. `cvar net.tickrate 60` in the server. 5. Your bike horn cooldown is now stuck on a huge value.
priority
curtime jumps around if you modify server tick rate on a live server this is because it s calculated purely as ticks passed tickrate changing tickrate changes total time this is problematic because curtime is used as an absolute time measure server side if it were to change many things like cooldowns could get stuck or act incorrectly to reproduce spawn a bike horn cvar net tickrate in the server honk the horn cvar net tickrate in the server your bike horn cooldown is now stuck on a huge value
1
173,951
14,442,655,473
IssuesEvent
2020-12-07 18:28:19
vuejs/jp.vuejs.org
https://api.github.com/repos/vuejs/jp.vuejs.org
closed
[Doc]: Revert "Add Vue mastery Promotion Banner for Black Friday 2020 (#2733…
assigned documentation
本家のドキュメントに更新がありました:page_facing_up: Original:https://github.com/vuejs/vuejs.org/commit/18a7d8ef13d7f82b34330444f896f117bd86055e
1.0
[Doc]: Revert "Add Vue mastery Promotion Banner for Black Friday 2020 (#2733… - 本家のドキュメントに更新がありました:page_facing_up: Original:https://github.com/vuejs/vuejs.org/commit/18a7d8ef13d7f82b34330444f896f117bd86055e
non_priority
revert add vue mastery promotion banner for black friday … 本家のドキュメントに更新がありました page facing up original
0
35,703
7,979,569,620
IssuesEvent
2018-07-17 22:19:38
Microsoft/TypeScript
https://api.github.com/repos/Microsoft/TypeScript
closed
TSServer should warn about supicious project setups
API Suggestion VS Code Tracked help wanted
**TypeScript Version:** 1.8.x The tsserver should warn when a file ends up in a virtual project although inside the workspace there is a tsconfig.json file that would cover the file but doesn't include it. This points very likely to a misconfigured project. The diagnostic maker is best generated on a file opened from the virtual project.
1.0
TSServer should warn about supicious project setups - **TypeScript Version:** 1.8.x The tsserver should warn when a file ends up in a virtual project although inside the workspace there is a tsconfig.json file that would cover the file but doesn't include it. This points very likely to a misconfigured project. The diagnostic maker is best generated on a file opened from the virtual project.
non_priority
tsserver should warn about supicious project setups typescript version x the tsserver should warn when a file ends up in a virtual project although inside the workspace there is a tsconfig json file that would cover the file but doesn t include it this points very likely to a misconfigured project the diagnostic maker is best generated on a file opened from the virtual project
0
158,787
6,035,380,045
IssuesEvent
2017-06-09 13:46:14
GeekyAnts/NativeBase
https://api.github.com/repos/GeekyAnts/NativeBase
closed
How can I change style of Toast
2 high priority need doc
I don't find any option to change the style (e.g opacity, backgroundColor,...) of Toast from the document http://docs.nativebase.io/Components.html#Toast So how can I do that ?
1.0
How can I change style of Toast - I don't find any option to change the style (e.g opacity, backgroundColor,...) of Toast from the document http://docs.nativebase.io/Components.html#Toast So how can I do that ?
priority
how can i change style of toast i don t find any option to change the style e g opacity backgroundcolor of toast from the document so how can i do that
1
811,462
30,289,092,231
IssuesEvent
2023-07-09 03:28:23
googleapis/api-linter
https://api.github.com/repos/googleapis/api-linter
closed
AIP-121: require re-using proto message across CRUDL
type: feature request priority: p3
AIP-121 states: > If the request to or the response from a standard method (or a custom method in the same service) is the resource or contains the resource, the resource schema for that resource across all methods must be the same.
1.0
AIP-121: require re-using proto message across CRUDL - AIP-121 states: > If the request to or the response from a standard method (or a custom method in the same service) is the resource or contains the resource, the resource schema for that resource across all methods must be the same.
priority
aip require re using proto message across crudl aip states if the request to or the response from a standard method or a custom method in the same service is the resource or contains the resource the resource schema for that resource across all methods must be the same
1
390,760
11,563,089,331
IssuesEvent
2020-02-20 04:50:49
google/ground-platform
https://api.github.com/repos/google/ground-platform
closed
[Router] Set up app router with stubs for key pages
priority: p1 type: feature request web
- [ ] `/projects`: Landing page - [ ] `/projects/project-id`: Main map/project UI - [ ] `/projects/project-id#f=feature-id`: Feature id selected - [ ] `/projects/project-id#l=layer-id`: Customize layer The first implementation can point to placeholder pages that simply print the selected ids. How-to: https://angular.io/guide/router#query-parameters-and-fragments Format of URLs is open for discussion, especially `#key=value` vs `#/a/b/c`.
1.0
[Router] Set up app router with stubs for key pages - - [ ] `/projects`: Landing page - [ ] `/projects/project-id`: Main map/project UI - [ ] `/projects/project-id#f=feature-id`: Feature id selected - [ ] `/projects/project-id#l=layer-id`: Customize layer The first implementation can point to placeholder pages that simply print the selected ids. How-to: https://angular.io/guide/router#query-parameters-and-fragments Format of URLs is open for discussion, especially `#key=value` vs `#/a/b/c`.
priority
set up app router with stubs for key pages projects landing page projects project id main map project ui projects project id f feature id feature id selected projects project id l layer id customize layer the first implementation can point to placeholder pages that simply print the selected ids how to format of urls is open for discussion especially key value vs a b c
1
144,648
5,543,746,017
IssuesEvent
2017-03-22 17:34:38
multidadosti-erp/multidadosti-addons
https://api.github.com/repos/multidadosti-erp/multidadosti-addons
opened
Remover menus o timesheet relacionados a RH
Category: Frontend Priority: Medium Stage: Backlog Type: Sub-task
Remover os seguintes menus: - [ ] "Minha planilhas de hora atual" - [ ] "Minha planilha de horas" - [ ] "Para Aprovar" - [ ] "Timesheet / Attendance"
1.0
Remover menus o timesheet relacionados a RH - Remover os seguintes menus: - [ ] "Minha planilhas de hora atual" - [ ] "Minha planilha de horas" - [ ] "Para Aprovar" - [ ] "Timesheet / Attendance"
priority
remover menus o timesheet relacionados a rh remover os seguintes menus minha planilhas de hora atual minha planilha de horas para aprovar timesheet attendance
1
277,400
24,068,082,939
IssuesEvent
2022-09-17 19:42:00
valory-xyz/open-autonomy
https://api.github.com/repos/valory-xyz/open-autonomy
closed
Fetch and production image building tests needs to be updated
test
### Subject of the issue The fetch test has not been updated because of a merge error. #1102 attempts to reintroduce the changes, but the windows tests now fail on CI: https://github.com/valory-xyz/open-autonomy/runs/7451351055?check_suite_focus=true The issue is also present for the `test_build_prod` test. If the hashes get updated, we fail with: ``` unable to stream build output: The command '/bin/sh -c INSTALL=1 VALORY_APPLICATION=${AEA_AGENT} bash start.sh' returned a non-zero code ``` Example on CI: https://github.com/valory-xyz/open-autonomy/runs/7462591408?check_suite_focus=true
1.0
Fetch and production image building tests needs to be updated - ### Subject of the issue The fetch test has not been updated because of a merge error. #1102 attempts to reintroduce the changes, but the windows tests now fail on CI: https://github.com/valory-xyz/open-autonomy/runs/7451351055?check_suite_focus=true The issue is also present for the `test_build_prod` test. If the hashes get updated, we fail with: ``` unable to stream build output: The command '/bin/sh -c INSTALL=1 VALORY_APPLICATION=${AEA_AGENT} bash start.sh' returned a non-zero code ``` Example on CI: https://github.com/valory-xyz/open-autonomy/runs/7462591408?check_suite_focus=true
non_priority
fetch and production image building tests needs to be updated subject of the issue the fetch test has not been updated because of a merge error attempts to reintroduce the changes but the windows tests now fail on ci the issue is also present for the test build prod test if the hashes get updated we fail with unable to stream build output the command bin sh c install valory application aea agent bash start sh returned a non zero code example on ci
0
473,285
13,639,701,839
IssuesEvent
2020-09-25 11:29:39
OpenNebula/one
https://api.github.com/repos/OpenNebula/one
closed
Quotas misbehavior when resuming a running VM
Category: Core & System Priority: High Sponsored Status: Accepted
**Description** When a VM in `RUNNING` state is resumed, `oned` through an error complaining about the resume action not being available for VMs in `RUNNING` state, but it will increase the number of running VMs in quotas. **To Reproduce** Execute `onevm resume <vm_id>` for a VM in `RUNNING` state. **Expected behavior** Quotas doesn't get updated as the VM is already in `RUNNING` state. **Details** - Affected Component: oned - Version: 5.12 - development **Additional context** Add any other context about the problem here. <!--////////////////////////////////////////////--> <!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM --> <!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS --> <!-- PROGRESS WILL BE REFLECTED HERE --> <!--////////////////////////////////////////////--> ## Progress Status - [ ] Branch created - [ ] Code committed to development branch - [ ] Testing - QA - [ ] Documentation - [ ] Release notes - resolved issues, compatibility, known issues - [ ] Code committed to upstream release/hotfix branches - [ ] Documentation committed to upstream release/hotfix branches
1.0
Quotas misbehavior when resuming a running VM - **Description** When a VM in `RUNNING` state is resumed, `oned` through an error complaining about the resume action not being available for VMs in `RUNNING` state, but it will increase the number of running VMs in quotas. **To Reproduce** Execute `onevm resume <vm_id>` for a VM in `RUNNING` state. **Expected behavior** Quotas doesn't get updated as the VM is already in `RUNNING` state. **Details** - Affected Component: oned - Version: 5.12 - development **Additional context** Add any other context about the problem here. <!--////////////////////////////////////////////--> <!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM --> <!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS --> <!-- PROGRESS WILL BE REFLECTED HERE --> <!--////////////////////////////////////////////--> ## Progress Status - [ ] Branch created - [ ] Code committed to development branch - [ ] Testing - QA - [ ] Documentation - [ ] Release notes - resolved issues, compatibility, known issues - [ ] Code committed to upstream release/hotfix branches - [ ] Documentation committed to upstream release/hotfix branches
priority
quotas misbehavior when resuming a running vm description when a vm in running state is resumed oned through an error complaining about the resume action not being available for vms in running state but it will increase the number of running vms in quotas to reproduce execute onevm resume for a vm in running state expected behavior quotas doesn t get updated as the vm is already in running state details affected component oned version development additional context add any other context about the problem here progress status branch created code committed to development branch testing qa documentation release notes resolved issues compatibility known issues code committed to upstream release hotfix branches documentation committed to upstream release hotfix branches
1
780,966
27,416,358,378
IssuesEvent
2023-03-01 14:02:37
beckerbenj/eatGADS
https://api.github.com/repos/beckerbenj/eatGADS
opened
multichar2fac() and positive values tagged as missings
bug high priority
If a missing tag is used for a low positive integer this will often run into an error... These should also be set to `NA`?
1.0
multichar2fac() and positive values tagged as missings - If a missing tag is used for a low positive integer this will often run into an error... These should also be set to `NA`?
priority
and positive values tagged as missings if a missing tag is used for a low positive integer this will often run into an error these should also be set to na
1
368,168
10,867,089,645
IssuesEvent
2019-11-14 22:51:54
kubernetes-sigs/cluster-api-provider-vsphere
https://api.github.com/repos/kubernetes-sigs/cluster-api-provider-vsphere
closed
Feature: create an automated solution to get node images created, stored, and referenced
kind/feature lifecycle/rotten priority/important-soon
Today, we require users to create a VM template from a cloud init image. The image must be stored in the vSphere datastore, and the name must be referenced in the machines.yaml file. This is a big barrier of entry for users of cluster api on vSphere. Design a better solution that can automate most of this.
1.0
Feature: create an automated solution to get node images created, stored, and referenced - Today, we require users to create a VM template from a cloud init image. The image must be stored in the vSphere datastore, and the name must be referenced in the machines.yaml file. This is a big barrier of entry for users of cluster api on vSphere. Design a better solution that can automate most of this.
priority
feature create an automated solution to get node images created stored and referenced today we require users to create a vm template from a cloud init image the image must be stored in the vsphere datastore and the name must be referenced in the machines yaml file this is a big barrier of entry for users of cluster api on vsphere design a better solution that can automate most of this
1
14,954
2,832,229,053
IssuesEvent
2015-05-25 05:46:33
ibus/ibus
https://api.github.com/repos/ibus/ibus
closed
ibus无法正确运行。
Priority-Medium Type-Defect
``` 我在Gentoo下,写了ibus和ibus-pinyin的ebuilds用来安装,但在启动ibus时出现如 下信息: 请问能否告知时什么地方出了问题,谢谢。 --------------8<---------------------8<----------------------------- start ibus-daemon Traceback (most recent call last): File "/usr/share/ibus/daemon/ibusdaemon.py", line 76, in __launch_auto_load_engines engines = self.__ibus.config_get_value("general/preload_engines", None) File "/usr/share/ibus/daemon/bus.py", line 366, in config_get_value return self.__config.get_value(key, **kargs) File "/usr/share/ibus/daemon/config.py", line 104, in get_value raise e IBusException: key = "/desktop/ibus/general/preload_engines" does not exist start ibus-x11 start ibus-panel ----------------8<-----------------8<--------------------------------- ``` Original issue reported on code.google.com by `oahong` on 2008-08-05 11:01:09
1.0
ibus无法正确运行。 - ``` 我在Gentoo下,写了ibus和ibus-pinyin的ebuilds用来安装,但在启动ibus时出现如 下信息: 请问能否告知时什么地方出了问题,谢谢。 --------------8<---------------------8<----------------------------- start ibus-daemon Traceback (most recent call last): File "/usr/share/ibus/daemon/ibusdaemon.py", line 76, in __launch_auto_load_engines engines = self.__ibus.config_get_value("general/preload_engines", None) File "/usr/share/ibus/daemon/bus.py", line 366, in config_get_value return self.__config.get_value(key, **kargs) File "/usr/share/ibus/daemon/config.py", line 104, in get_value raise e IBusException: key = "/desktop/ibus/general/preload_engines" does not exist start ibus-x11 start ibus-panel ----------------8<-----------------8<--------------------------------- ``` Original issue reported on code.google.com by `oahong` on 2008-08-05 11:01:09
non_priority
ibus无法正确运行。 我在gentoo下,写了ibus和ibus pinyin的ebuilds用来安装,但在启动ibus时出现如 下信息: 请问能否告知时什么地方出了问题,谢谢。 start ibus daemon traceback most recent call last file usr share ibus daemon ibusdaemon py line in launch auto load engines engines self ibus config get value general preload engines none file usr share ibus daemon bus py line in config get value return self config get value key kargs file usr share ibus daemon config py line in get value raise e ibusexception key desktop ibus general preload engines does not exist start ibus start ibus panel original issue reported on code google com by oahong on
0
248,157
20,999,969,029
IssuesEvent
2022-03-29 16:29:25
PalisadoesFoundation/talawa-admin
https://api.github.com/repos/PalisadoesFoundation/talawa-admin
closed
Code Coverage: Create tests for pluginReducer.ts
good first issue test points 01
The Talawa-Admin code base needs to be 100% reliable. This means we need to have 100% test code coverage. Tests need to be written for file `src/state/reducers/pluginReducer.ts` - When complete, all methods, classes and/or functions in the file will need to be tested. - These tests must be placed in a single file with the name `src/state/reducers/pluginReducer.test.ts`. You may need to create the appropriate directory structure to do this. ### IMPORTANT: Please refer to the parent issue on how to implement these tests correctly: - #241 ### PR Acceptance Criteria - When complete this file must show **100%** coverage when merged into the code base. This will be clearly visible when you submit your PR. - [The current code coverage for the file can be found here](https://codecov.io/gh/PalisadoesFoundation/talawa-admin/tree/develop/src/state/reducers/). If the file isn't found in this directory, or there is a 404 error, then tests have not been created. - Create a code coverage account for your repo's preferred branch to generate the values when you do your commits for that branch to your repo. - _**NOTE:**_ Make sure you select the correct branch when you do the setup, or else the reporting will fail. You must also create a `CODECOV` GitHub secret for your repo as part of the process. - The PR will show a report for the code coverage for the file you have added. You can also use that as a guide.
1.0
Code Coverage: Create tests for pluginReducer.ts - The Talawa-Admin code base needs to be 100% reliable. This means we need to have 100% test code coverage. Tests need to be written for file `src/state/reducers/pluginReducer.ts` - When complete, all methods, classes and/or functions in the file will need to be tested. - These tests must be placed in a single file with the name `src/state/reducers/pluginReducer.test.ts`. You may need to create the appropriate directory structure to do this. ### IMPORTANT: Please refer to the parent issue on how to implement these tests correctly: - #241 ### PR Acceptance Criteria - When complete this file must show **100%** coverage when merged into the code base. This will be clearly visible when you submit your PR. - [The current code coverage for the file can be found here](https://codecov.io/gh/PalisadoesFoundation/talawa-admin/tree/develop/src/state/reducers/). If the file isn't found in this directory, or there is a 404 error, then tests have not been created. - Create a code coverage account for your repo's preferred branch to generate the values when you do your commits for that branch to your repo. - _**NOTE:**_ Make sure you select the correct branch when you do the setup, or else the reporting will fail. You must also create a `CODECOV` GitHub secret for your repo as part of the process. - The PR will show a report for the code coverage for the file you have added. You can also use that as a guide.
non_priority
code coverage create tests for pluginreducer ts the talawa admin code base needs to be reliable this means we need to have test code coverage tests need to be written for file src state reducers pluginreducer ts when complete all methods classes and or functions in the file will need to be tested these tests must be placed in a single file with the name src state reducers pluginreducer test ts you may need to create the appropriate directory structure to do this important please refer to the parent issue on how to implement these tests correctly pr acceptance criteria when complete this file must show coverage when merged into the code base this will be clearly visible when you submit your pr if the file isn t found in this directory or there is a error then tests have not been created create a code coverage account for your repo s preferred branch to generate the values when you do your commits for that branch to your repo note make sure you select the correct branch when you do the setup or else the reporting will fail you must also create a codecov github secret for your repo as part of the process the pr will show a report for the code coverage for the file you have added you can also use that as a guide
0
98,969
11,102,213,047
IssuesEvent
2019-12-16 23:18:20
vccolombo/nossoQL
https://api.github.com/repos/vccolombo/nossoQL
closed
Documentação
documentation good first issue
Precisa ser feita uma documentação para todos os comandos implementados. Opções são criar um arquivo só para a documentação, ou jogar tudo no README. Nesse momento acredito que tanto faz, mas no futuro um arquivo só para a documentação será melhor. Adicionalmente, precisa ser adicionadas instruções de como instalar o programa (passo a passo para fazer funcionar)
1.0
Documentação - Precisa ser feita uma documentação para todos os comandos implementados. Opções são criar um arquivo só para a documentação, ou jogar tudo no README. Nesse momento acredito que tanto faz, mas no futuro um arquivo só para a documentação será melhor. Adicionalmente, precisa ser adicionadas instruções de como instalar o programa (passo a passo para fazer funcionar)
non_priority
documentação precisa ser feita uma documentação para todos os comandos implementados opções são criar um arquivo só para a documentação ou jogar tudo no readme nesse momento acredito que tanto faz mas no futuro um arquivo só para a documentação será melhor adicionalmente precisa ser adicionadas instruções de como instalar o programa passo a passo para fazer funcionar
0
95,889
12,058,176,612
IssuesEvent
2020-04-15 17:00:29
cloudfour/cloudfour.com-patterns
https://api.github.com/repos/cloudfour/cloudfour.com-patterns
closed
Thumbnail patterns
size:2 🎨 design
We have several different styles of thumbnails, some are within a contained shape and some have borders. These may need to be different patterns, especially the calendar. <img width="615" alt="Thumbnails" src="https://user-images.githubusercontent.com/42841342/75062496-c2031c00-5497-11ea-968f-ef2fbfe19bac.png">
1.0
Thumbnail patterns - We have several different styles of thumbnails, some are within a contained shape and some have borders. These may need to be different patterns, especially the calendar. <img width="615" alt="Thumbnails" src="https://user-images.githubusercontent.com/42841342/75062496-c2031c00-5497-11ea-968f-ef2fbfe19bac.png">
non_priority
thumbnail patterns we have several different styles of thumbnails some are within a contained shape and some have borders these may need to be different patterns especially the calendar img width alt thumbnails src
0
129,193
17,755,112,583
IssuesEvent
2021-08-28 15:51:28
bitlap/bitlap
https://api.github.com/repos/bitlap/bitlap
reopened
Implementation of `BitlapCli`
good first issue design client
Implementation of `BitlapCli`, it hasn't started yet. * Based on `sqlLine` project * Working with `BitlapDriver` #36
1.0
Implementation of `BitlapCli` - Implementation of `BitlapCli`, it hasn't started yet. * Based on `sqlLine` project * Working with `BitlapDriver` #36
non_priority
implementation of bitlapcli implementation of bitlapcli it hasn t started yet based on sqlline project working with bitlapdriver
0
7,569
2,603,619,194
IssuesEvent
2015-02-24 17:00:36
tsgrp/hpi
https://api.github.com/repos/tsgrp/hpi
closed
Cascading Queries - Issue with new typeahead?
High Priority issue
Seems to be an issue now on @joehof's current client. Assuming that this is also a problem for trunk as well. Did something in the new typeahead break cascading queries? @joehof - feel free to add more details here if needed.
1.0
Cascading Queries - Issue with new typeahead? - Seems to be an issue now on @joehof's current client. Assuming that this is also a problem for trunk as well. Did something in the new typeahead break cascading queries? @joehof - feel free to add more details here if needed.
priority
cascading queries issue with new typeahead seems to be an issue now on joehof s current client assuming that this is also a problem for trunk as well did something in the new typeahead break cascading queries joehof feel free to add more details here if needed
1
493,683
14,236,952,037
IssuesEvent
2020-11-18 16:36:38
geosolutions-it/MapStore2
https://api.github.com/repos/geosolutions-it/MapStore2
closed
Data Gaps for Charts
Accepted C117-UNESCO-2019-GEONODE New Feature Priority: High
## Description <!-- A few sentences describing new feature --> <!-- screenshot, video, or link to mockup/prototype are welcome --> Using PlotyJS, this does not require any work, except the skipping aggregation. **What kind of improvement you want to add?** (check one with "x", remove the others) - [ ] Minor changes to existing features - [ ] Code style update (formatting, local variables) - [ ] Refactoring (no functional changes, no api changes) - [ ] Build related changes - [ ] CI related changes - [X] Other... Please describe: Chart improvement ## Other useful information See also #6055.
1.0
Data Gaps for Charts - ## Description <!-- A few sentences describing new feature --> <!-- screenshot, video, or link to mockup/prototype are welcome --> Using PlotyJS, this does not require any work, except the skipping aggregation. **What kind of improvement you want to add?** (check one with "x", remove the others) - [ ] Minor changes to existing features - [ ] Code style update (formatting, local variables) - [ ] Refactoring (no functional changes, no api changes) - [ ] Build related changes - [ ] CI related changes - [X] Other... Please describe: Chart improvement ## Other useful information See also #6055.
priority
data gaps for charts description using plotyjs this does not require any work except the skipping aggregation what kind of improvement you want to add check one with x remove the others minor changes to existing features code style update formatting local variables refactoring no functional changes no api changes build related changes ci related changes other please describe chart improvement other useful information see also
1
62,023
17,023,833,999
IssuesEvent
2021-07-03 04:05:18
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
osm2pgsql creates duplicate db entries when using extracts as source
Component: osm2pgsql Priority: critical Resolution: invalid Type: defect
**[Submitted to the original trac issue database at 9.51pm, Friday, 19th October 2012]** When creating a database using area extracts as input the database contains duplicate entries. Reproduced with version 28761 and using burkina_faso geofabrik extract. Format of input file does not matter. Tested with both xml and pbf input. Create test DB: ``` createdb -E UTF8 -T template0 -O osm testpbf psql -d testpbf -c "CREATE EXTENSION postgis;" psql -d testpbf -c "CREATE EXTENSION hstore;" echo "ALTER TABLE geometry_columns OWNER TO osm; ALTER TABLE spatial_ref_sys OWNER TO osm;" | psql -d testpbf ``` import using PBF as source: ``` ./osm2pgsql -v --create --slim --cache 800 --hstore-all --database testreadpbf --username osm --style default.style --number-processes 1 --disable-parallel-indexing -r pbf ../burkina_faso.osm.pbf ``` Checking for duplicate ways returns 110 rows: ``` SELECT osm_id,count(1) FROM planet_osm_line GROUP BY osm_id HAVING count(osm_id) > 1 ORDER BY count(osm_id) DESC ``` For example relation 192783 is contained 36 times in the db. A grep on the input file confirms it's only one time there.
1.0
osm2pgsql creates duplicate db entries when using extracts as source - **[Submitted to the original trac issue database at 9.51pm, Friday, 19th October 2012]** When creating a database using area extracts as input the database contains duplicate entries. Reproduced with version 28761 and using burkina_faso geofabrik extract. Format of input file does not matter. Tested with both xml and pbf input. Create test DB: ``` createdb -E UTF8 -T template0 -O osm testpbf psql -d testpbf -c "CREATE EXTENSION postgis;" psql -d testpbf -c "CREATE EXTENSION hstore;" echo "ALTER TABLE geometry_columns OWNER TO osm; ALTER TABLE spatial_ref_sys OWNER TO osm;" | psql -d testpbf ``` import using PBF as source: ``` ./osm2pgsql -v --create --slim --cache 800 --hstore-all --database testreadpbf --username osm --style default.style --number-processes 1 --disable-parallel-indexing -r pbf ../burkina_faso.osm.pbf ``` Checking for duplicate ways returns 110 rows: ``` SELECT osm_id,count(1) FROM planet_osm_line GROUP BY osm_id HAVING count(osm_id) > 1 ORDER BY count(osm_id) DESC ``` For example relation 192783 is contained 36 times in the db. A grep on the input file confirms it's only one time there.
non_priority
creates duplicate db entries when using extracts as source when creating a database using area extracts as input the database contains duplicate entries reproduced with version and using burkina faso geofabrik extract format of input file does not matter tested with both xml and pbf input create test db createdb e t o osm testpbf psql d testpbf c create extension postgis psql d testpbf c create extension hstore echo alter table geometry columns owner to osm alter table spatial ref sys owner to osm psql d testpbf import using pbf as source v create slim cache hstore all database testreadpbf username osm style default style number processes disable parallel indexing r pbf burkina faso osm pbf checking for duplicate ways returns rows select osm id count from planet osm line group by osm id having count osm id order by count osm id desc for example relation is contained times in the db a grep on the input file confirms it s only one time there
0
6,771
3,900,229,621
IssuesEvent
2016-04-18 04:04:37
mapbox/mapbox-gl-native
https://api.github.com/repos/mapbox/mapbox-gl-native
closed
[iOS] [Mason] duplicate symbols on static library
build iOS
From a clean master branch pull, I run "make ipackage". It ends correctly (lib mapbox compiled and linked) but with warnings about duplicated symbols from geojsonvt. When used, this library has duplicated symbols preventing our app to link. Further more, Mason keeps old geojson libs / headers on our build machine, so it even more confused about duplicated symbols. Mapbox lib error output: ``` Building static library... /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/libtool: warning for architecture: armv7 same member name (tile.o) in output file used for input files: mason_packages/ios-9.2/geojsonvt/3.1.0/lib/libgeojsonvt.a(tile.o) and: gyp/build/Release-iphoneos/libmbgl-core.a(tile.o) due to use of basename, truncation and blank padding /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/libtool: warning for architecture: armv7 same member name (transform.o) in output file used for input files: mason_packages/ios-9.2/geojsonvt/3.1.0/lib/libgeojsonvt.a(transform.o) and: gyp/build/Release-iphoneos/libmbgl-core.a(transform.o) due to use of basename, truncation and blank padding /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/libtool: warning for architecture: armv7s same member name (tile.o) in output file used for input files: mason_packages/ios-9.2/geojsonvt/3.1.0/lib/libgeojsonvt.a(tile.o) and: gyp/build/Release-iphoneos/libmbgl-core.a(tile.o) due to use of basename, truncation and blank padding /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/libtool: warning for architecture: armv7s same member name (transform.o) in output file used for input files: mason_packages/ios-9.2/geojsonvt/3.1.0/lib/libgeojsonvt.a(transform.o) and: gyp/build/Release-iphoneos/libmbgl-core.a(transform.o) due to use of basename, truncation and blank padding /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/libtool: warning for architecture: i386 same member name (tile.o) in output file used for input files: mason_packages/ios-9.2/geojsonvt/3.1.0/lib/libgeojsonvt.a(tile.o) and: gyp/build/Release-iphonesimulator/libmbgl-core.a(tile.o) due to use of basename, truncation and blank padding /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/libtool: warning for architecture: i386 same member name (transform.o) in output file used for input files: mason_packages/ios-9.2/geojsonvt/3.1.0/lib/libgeojsonvt.a(transform.o) and: gyp/build/Release-iphonesimulator/libmbgl-core.a(transform.o) due to use of basename, truncation and blank padding /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/libtool: warning for architecture: x86_64 same member name (tile.o) in output file used for input files: mason_packages/ios-9.2/geojsonvt/3.1.0/lib/libgeojsonvt.a(tile.o) and: gyp/build/Release-iphonesimulator/libmbgl-core.a(tile.o) due to use of basename, truncation and blank padding /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/libtool: warning for architecture: x86_64 same member name (transform.o) in output file used for input files: mason_packages/ios-9.2/geojsonvt/3.1.0/lib/libgeojsonvt.a(transform.o) and: gyp/build/Release-iphonesimulator/libmbgl-core.a(transform.o) due to use of basename, truncation and blank padding /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/libtool: warning for architecture: arm64 same member name (tile.o) in output file used for input files: mason_packages/ios-9.2/geojsonvt/3.1.0/lib/libgeojsonvt.a(tile.o) and: gyp/build/Release-iphoneos/libmbgl-core.a(tile.o) due to use of basename, truncation and blank padding /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/libtool: warning for architecture: arm64 same member name (transform.o) in output file used for input files: mason_packages/ios-9.2/geojsonvt/3.1.0/lib/libgeojsonvt.a(transform.o) and: gyp/build/Release-iphoneos/libmbgl-core.a(transform.o) due to use of basename, truncation and blank padding Created build/ios/pkg/static/libMapbox.a ```
1.0
[iOS] [Mason] duplicate symbols on static library - From a clean master branch pull, I run "make ipackage". It ends correctly (lib mapbox compiled and linked) but with warnings about duplicated symbols from geojsonvt. When used, this library has duplicated symbols preventing our app to link. Further more, Mason keeps old geojson libs / headers on our build machine, so it even more confused about duplicated symbols. Mapbox lib error output: ``` Building static library... /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/libtool: warning for architecture: armv7 same member name (tile.o) in output file used for input files: mason_packages/ios-9.2/geojsonvt/3.1.0/lib/libgeojsonvt.a(tile.o) and: gyp/build/Release-iphoneos/libmbgl-core.a(tile.o) due to use of basename, truncation and blank padding /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/libtool: warning for architecture: armv7 same member name (transform.o) in output file used for input files: mason_packages/ios-9.2/geojsonvt/3.1.0/lib/libgeojsonvt.a(transform.o) and: gyp/build/Release-iphoneos/libmbgl-core.a(transform.o) due to use of basename, truncation and blank padding /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/libtool: warning for architecture: armv7s same member name (tile.o) in output file used for input files: mason_packages/ios-9.2/geojsonvt/3.1.0/lib/libgeojsonvt.a(tile.o) and: gyp/build/Release-iphoneos/libmbgl-core.a(tile.o) due to use of basename, truncation and blank padding /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/libtool: warning for architecture: armv7s same member name (transform.o) in output file used for input files: mason_packages/ios-9.2/geojsonvt/3.1.0/lib/libgeojsonvt.a(transform.o) and: gyp/build/Release-iphoneos/libmbgl-core.a(transform.o) due to use of basename, truncation and blank padding /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/libtool: warning for architecture: i386 same member name (tile.o) in output file used for input files: mason_packages/ios-9.2/geojsonvt/3.1.0/lib/libgeojsonvt.a(tile.o) and: gyp/build/Release-iphonesimulator/libmbgl-core.a(tile.o) due to use of basename, truncation and blank padding /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/libtool: warning for architecture: i386 same member name (transform.o) in output file used for input files: mason_packages/ios-9.2/geojsonvt/3.1.0/lib/libgeojsonvt.a(transform.o) and: gyp/build/Release-iphonesimulator/libmbgl-core.a(transform.o) due to use of basename, truncation and blank padding /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/libtool: warning for architecture: x86_64 same member name (tile.o) in output file used for input files: mason_packages/ios-9.2/geojsonvt/3.1.0/lib/libgeojsonvt.a(tile.o) and: gyp/build/Release-iphonesimulator/libmbgl-core.a(tile.o) due to use of basename, truncation and blank padding /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/libtool: warning for architecture: x86_64 same member name (transform.o) in output file used for input files: mason_packages/ios-9.2/geojsonvt/3.1.0/lib/libgeojsonvt.a(transform.o) and: gyp/build/Release-iphonesimulator/libmbgl-core.a(transform.o) due to use of basename, truncation and blank padding /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/libtool: warning for architecture: arm64 same member name (tile.o) in output file used for input files: mason_packages/ios-9.2/geojsonvt/3.1.0/lib/libgeojsonvt.a(tile.o) and: gyp/build/Release-iphoneos/libmbgl-core.a(tile.o) due to use of basename, truncation and blank padding /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/libtool: warning for architecture: arm64 same member name (transform.o) in output file used for input files: mason_packages/ios-9.2/geojsonvt/3.1.0/lib/libgeojsonvt.a(transform.o) and: gyp/build/Release-iphoneos/libmbgl-core.a(transform.o) due to use of basename, truncation and blank padding Created build/ios/pkg/static/libMapbox.a ```
non_priority
duplicate symbols on static library from a clean master branch pull i run make ipackage it ends correctly lib mapbox compiled and linked but with warnings about duplicated symbols from geojsonvt when used this library has duplicated symbols preventing our app to link further more mason keeps old geojson libs headers on our build machine so it even more confused about duplicated symbols mapbox lib error output building static library applications xcode app contents developer toolchains xcodedefault xctoolchain usr bin libtool warning for architecture same member name tile o in output file used for input files mason packages ios geojsonvt lib libgeojsonvt a tile o and gyp build release iphoneos libmbgl core a tile o due to use of basename truncation and blank padding applications xcode app contents developer toolchains xcodedefault xctoolchain usr bin libtool warning for architecture same member name transform o in output file used for input files mason packages ios geojsonvt lib libgeojsonvt a transform o and gyp build release iphoneos libmbgl core a transform o due to use of basename truncation and blank padding applications xcode app contents developer toolchains xcodedefault xctoolchain usr bin libtool warning for architecture same member name tile o in output file used for input files mason packages ios geojsonvt lib libgeojsonvt a tile o and gyp build release iphoneos libmbgl core a tile o due to use of basename truncation and blank padding applications xcode app contents developer toolchains xcodedefault xctoolchain usr bin libtool warning for architecture same member name transform o in output file used for input files mason packages ios geojsonvt lib libgeojsonvt a transform o and gyp build release iphoneos libmbgl core a transform o due to use of basename truncation and blank padding applications xcode app contents developer toolchains xcodedefault xctoolchain usr bin libtool warning for architecture same member name tile o in output file used for input files mason packages ios geojsonvt lib libgeojsonvt a tile o and gyp build release iphonesimulator libmbgl core a tile o due to use of basename truncation and blank padding applications xcode app contents developer toolchains xcodedefault xctoolchain usr bin libtool warning for architecture same member name transform o in output file used for input files mason packages ios geojsonvt lib libgeojsonvt a transform o and gyp build release iphonesimulator libmbgl core a transform o due to use of basename truncation and blank padding applications xcode app contents developer toolchains xcodedefault xctoolchain usr bin libtool warning for architecture same member name tile o in output file used for input files mason packages ios geojsonvt lib libgeojsonvt a tile o and gyp build release iphonesimulator libmbgl core a tile o due to use of basename truncation and blank padding applications xcode app contents developer toolchains xcodedefault xctoolchain usr bin libtool warning for architecture same member name transform o in output file used for input files mason packages ios geojsonvt lib libgeojsonvt a transform o and gyp build release iphonesimulator libmbgl core a transform o due to use of basename truncation and blank padding applications xcode app contents developer toolchains xcodedefault xctoolchain usr bin libtool warning for architecture same member name tile o in output file used for input files mason packages ios geojsonvt lib libgeojsonvt a tile o and gyp build release iphoneos libmbgl core a tile o due to use of basename truncation and blank padding applications xcode app contents developer toolchains xcodedefault xctoolchain usr bin libtool warning for architecture same member name transform o in output file used for input files mason packages ios geojsonvt lib libgeojsonvt a transform o and gyp build release iphoneos libmbgl core a transform o due to use of basename truncation and blank padding created build ios pkg static libmapbox a
0
42,775
2,873,962,118
IssuesEvent
2015-06-08 19:53:33
Glavin001/atom-beautify
https://api.github.com/repos/Glavin001/atom-beautify
closed
Windows unit testing with AppVeyor
high priority published windows
See https://ci.appveyor.com/ - [x] Language packages - [x] CLI beautifiers - [x] `uncrustify` - [x] `autopep8` - [x] `htmlbeautifier` - [x] `php` - [x] `php-cs-fixer` - [x] `rubocop` - [x] `emacs`
1.0
Windows unit testing with AppVeyor - See https://ci.appveyor.com/ - [x] Language packages - [x] CLI beautifiers - [x] `uncrustify` - [x] `autopep8` - [x] `htmlbeautifier` - [x] `php` - [x] `php-cs-fixer` - [x] `rubocop` - [x] `emacs`
priority
windows unit testing with appveyor see language packages cli beautifiers uncrustify htmlbeautifier php php cs fixer rubocop emacs
1
71,242
15,189,219,842
IssuesEvent
2021-02-15 16:07:03
NixOS/nixpkgs
https://api.github.com/repos/NixOS/nixpkgs
closed
Vulnerability roundup 93: terraform-0.13.2: 1 advisory [9.8]
1.severity: security
[search](https://search.nix.gsc.io/?q=terraform&i=fosho&repos=NixOS-nixpkgs), [files](https://github.com/NixOS/nixpkgs/search?utf8=%E2%9C%93&q=terraform+in%3Apath&type=Code) * [ ] [CVE-2018-9057](https://nvd.nist.gov/vuln/detail/CVE-2018-9057) CVSSv3=9.8 (nixos-20.09) Scanned versions: nixos-20.09: ad3a5d5092e. Cc @Chili-Man Cc @babariviere Cc @kalbasit Cc @marsam Cc @peterhoeg Cc @zimbatm
True
Vulnerability roundup 93: terraform-0.13.2: 1 advisory [9.8] - [search](https://search.nix.gsc.io/?q=terraform&i=fosho&repos=NixOS-nixpkgs), [files](https://github.com/NixOS/nixpkgs/search?utf8=%E2%9C%93&q=terraform+in%3Apath&type=Code) * [ ] [CVE-2018-9057](https://nvd.nist.gov/vuln/detail/CVE-2018-9057) CVSSv3=9.8 (nixos-20.09) Scanned versions: nixos-20.09: ad3a5d5092e. Cc @Chili-Man Cc @babariviere Cc @kalbasit Cc @marsam Cc @peterhoeg Cc @zimbatm
non_priority
vulnerability roundup terraform advisory nixos scanned versions nixos cc chili man cc babariviere cc kalbasit cc marsam cc peterhoeg cc zimbatm
0
2,883
5,071,605,406
IssuesEvent
2016-12-26 14:47:07
TechnionYP5777/SmartCity-Accessibility
https://api.github.com/repos/TechnionYP5777/SmartCity-Accessibility
closed
Should we replace JxMaps with GMapsFx?
Map services question
This was raised as a question after the last presentation. [GMapsFX](https://github.com/rterp/GMapsFX) Things that seems to be in favor of switching : - GMapsFX works with JavaFX natively, a thing that JxMaps claims to do as well but doesn't seem to actually work (failure on trying to import the javafx instead of swing library of JxMaps). Swing is an outdated GUI library and we really should be working with javafx. - It is open source which is very convenient to work with. - After actually working a bit with JxMaps (and as been raised by most people in the project) it seems to be very poorly built and not very convenient to use. While the above is true this would require alot of changes as we already started working with jxmaps. Most of all this seems to effect @yaelAmitay so your opinion will be deciding on this point. Thoughts? @Kolikant @koralchapnik @ArthurSap @assaflu
1.0
Should we replace JxMaps with GMapsFx? - This was raised as a question after the last presentation. [GMapsFX](https://github.com/rterp/GMapsFX) Things that seems to be in favor of switching : - GMapsFX works with JavaFX natively, a thing that JxMaps claims to do as well but doesn't seem to actually work (failure on trying to import the javafx instead of swing library of JxMaps). Swing is an outdated GUI library and we really should be working with javafx. - It is open source which is very convenient to work with. - After actually working a bit with JxMaps (and as been raised by most people in the project) it seems to be very poorly built and not very convenient to use. While the above is true this would require alot of changes as we already started working with jxmaps. Most of all this seems to effect @yaelAmitay so your opinion will be deciding on this point. Thoughts? @Kolikant @koralchapnik @ArthurSap @assaflu
non_priority
should we replace jxmaps with gmapsfx this was raised as a question after the last presentation things that seems to be in favor of switching gmapsfx works with javafx natively a thing that jxmaps claims to do as well but doesn t seem to actually work failure on trying to import the javafx instead of swing library of jxmaps swing is an outdated gui library and we really should be working with javafx it is open source which is very convenient to work with after actually working a bit with jxmaps and as been raised by most people in the project it seems to be very poorly built and not very convenient to use while the above is true this would require alot of changes as we already started working with jxmaps most of all this seems to effect yaelamitay so your opinion will be deciding on this point thoughts kolikant koralchapnik arthursap assaflu
0
322,507
9,818,740,800
IssuesEvent
2019-06-13 20:03:46
googlefonts/noto-fonts
https://api.github.com/repos/googlefonts/noto-fonts
closed
TIbetan sequence <U+0F41, U+0F9A> is not stacked.
Android Priority-Medium Script-Tibetan
Spun off from googlei18n/nototools#38. Below is the result of shaping U+0F41, U+0F9A with Noto Sans Tibetan + harfbuzz: ![ཁྚ](https://cloud.githubusercontent.com/assets/1527880/13630850/c62bf9d8-e5e1-11e5-810d-37a4faa048da.png) @kmansourMT : please, take a look at it. It's from [Chris Fynn’s list](https://sites.google.com/site/chrisfynn2/home/tibetanscriptfonts/standardization/precomposedtibetan-parta/precomposed-tibetan---part-a-page-4)
1.0
TIbetan sequence <U+0F41, U+0F9A> is not stacked. - Spun off from googlei18n/nototools#38. Below is the result of shaping U+0F41, U+0F9A with Noto Sans Tibetan + harfbuzz: ![ཁྚ](https://cloud.githubusercontent.com/assets/1527880/13630850/c62bf9d8-e5e1-11e5-810d-37a4faa048da.png) @kmansourMT : please, take a look at it. It's from [Chris Fynn’s list](https://sites.google.com/site/chrisfynn2/home/tibetanscriptfonts/standardization/precomposedtibetan-parta/precomposed-tibetan---part-a-page-4)
priority
tibetan sequence is not stacked spun off from nototools below is the result of shaping u u with noto sans tibetan harfbuzz kmansourmt please take a look at it it s from
1
55,718
6,489,367,224
IssuesEvent
2017-08-21 01:13:32
FireFly-WoW/FireFly-IssueTracker
https://api.github.com/repos/FireFly-WoW/FireFly-IssueTracker
closed
Dungeon: Halls of Lightning (Normal): Hallway with Statues leading out of Volkhan's room
Priority: Medium Status: Tested / Working - Requires Optimizations Target: Core
**Description:** In this hallway when you pass one of the thresholds a bunch of statues is supposed to come alive. Issue 1: When you pass first threshold, too many statues will come to life. = wipe for group Issue 2: When you pass next thresholds nothing happen. - A bunch of statues is supposed to come alive. Issue 3: Statutes are not supposed to be targetable. ![wowscrnshot_041617_165110](https://cloud.githubusercontent.com/assets/27430839/25069518/81763164-22c6-11e7-86d6-39c30d759cc7.jpg)
1.0
Dungeon: Halls of Lightning (Normal): Hallway with Statues leading out of Volkhan's room - **Description:** In this hallway when you pass one of the thresholds a bunch of statues is supposed to come alive. Issue 1: When you pass first threshold, too many statues will come to life. = wipe for group Issue 2: When you pass next thresholds nothing happen. - A bunch of statues is supposed to come alive. Issue 3: Statutes are not supposed to be targetable. ![wowscrnshot_041617_165110](https://cloud.githubusercontent.com/assets/27430839/25069518/81763164-22c6-11e7-86d6-39c30d759cc7.jpg)
non_priority
dungeon halls of lightning normal hallway with statues leading out of volkhan s room description in this hallway when you pass one of the thresholds a bunch of statues is supposed to come alive issue when you pass first threshold too many statues will come to life wipe for group issue when you pass next thresholds nothing happen a bunch of statues is supposed to come alive issue statutes are not supposed to be targetable
0
272,694
23,696,552,669
IssuesEvent
2022-08-29 15:05:34
Joystream/pioneer
https://api.github.com/repos/Joystream/pioneer
closed
Create free membership/Add Membership modals. "Member Name" field validation.
bug low-prio qa-task community-dev qa-tested-ready-for-prod qa-effort-15min
Tested on https://dao-git-carthage-joystream.vercel.app/ "Member Name" field doesn't allow spaces, so user cant input "John Doe" <img width="797" alt="Снимок экрана 2022-07-30 в 13 20 39" src="https://user-images.githubusercontent.com/46903215/181906065-de8feb2f-c5d6-4d7a-9abc-1f17ed396f58.png">
1.0
Create free membership/Add Membership modals. "Member Name" field validation. - Tested on https://dao-git-carthage-joystream.vercel.app/ "Member Name" field doesn't allow spaces, so user cant input "John Doe" <img width="797" alt="Снимок экрана 2022-07-30 в 13 20 39" src="https://user-images.githubusercontent.com/46903215/181906065-de8feb2f-c5d6-4d7a-9abc-1f17ed396f58.png">
non_priority
create free membership add membership modals member name field validation tested on member name field doesn t allow spaces so user cant input john doe img width alt снимок экрана в src
0
192,727
22,215,996,315
IssuesEvent
2022-06-08 01:45:08
praneethpanasala/linux
https://api.github.com/repos/praneethpanasala/linux
reopened
CVE-2019-19072 (Medium) detected in linuxlinux-4.19.6
security vulnerability
## CVE-2019-19072 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.6</b></p></summary> <p> <p>Apache Software Foundation (ASF)</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://api.github.com/repos/praneethpanasala/linux/commits/d80c4f847c91020292cb280132b15e2ea147f1a3">d80c4f847c91020292cb280132b15e2ea147f1a3</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/trace/trace_events_filter.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/trace/trace_events_filter.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A memory leak in the predicate_parse() function in kernel/trace/trace_events_filter.c in the Linux kernel through 5.3.11 allows attackers to cause a denial of service (memory consumption), aka CID-96c5c6e6a5b6. <p>Publish Date: 2019-11-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-19072>CVE-2019-19072</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19072">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19072</a></p> <p>Release Date: 2020-08-24</p> <p>Fix Resolution: v5.4-rc1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-19072 (Medium) detected in linuxlinux-4.19.6 - ## CVE-2019-19072 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.6</b></p></summary> <p> <p>Apache Software Foundation (ASF)</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://api.github.com/repos/praneethpanasala/linux/commits/d80c4f847c91020292cb280132b15e2ea147f1a3">d80c4f847c91020292cb280132b15e2ea147f1a3</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/trace/trace_events_filter.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/trace/trace_events_filter.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A memory leak in the predicate_parse() function in kernel/trace/trace_events_filter.c in the Linux kernel through 5.3.11 allows attackers to cause a denial of service (memory consumption), aka CID-96c5c6e6a5b6. <p>Publish Date: 2019-11-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-19072>CVE-2019-19072</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19072">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19072</a></p> <p>Release Date: 2020-08-24</p> <p>Fix Resolution: v5.4-rc1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve medium detected in linuxlinux cve medium severity vulnerability vulnerable library linuxlinux apache software foundation asf library home page a href found in head commit a href found in base branch master vulnerable source files kernel trace trace events filter c kernel trace trace events filter c vulnerability details a memory leak in the predicate parse function in kernel trace trace events filter c in the linux kernel through allows attackers to cause a denial of service memory consumption aka cid publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required high user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
720,324
24,788,627,011
IssuesEvent
2022-10-24 12:02:30
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
login.live.com - desktop site instead of mobile site
priority-critical browser-focus-geckoview engine-gecko
<!-- @browser: Firefox Mobile 106.0 --> <!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:106.0) Gecko/106.0 Firefox/106.0 --> <!-- @reported_with: android-components-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/112761 --> <!-- @extra_labels: browser-focus-geckoview --> **URL**: https://login.live.com/login.srf?wa=wsignin1.0&rpsnv=13&ct=1666489150&rver=7.0.6738.0&wp=MBI_SSL&wreply=https:%2F%2Faccount.microsoft.com%2Fauth%2Fcomplete-signin%3Fru%3Dhttps%253A%252F%252Faccount.microsoft.com%252F%253Frefd%253Dwww.microsoft.com&lc=1033&id=292666&lw=1&fl=easi2 **Browser / Version**: Firefox Mobile 106.0 **Operating System**: Android 10 **Tested Another Browser**: Yes Chrome **Problem type**: Desktop site instead of mobile site **Description**: Desktop site instead of mobile site **Steps to Reproduce**: Finally found a browser that seems to take me to actual sites I'm looking for instead of redirects or malware when I use desktop mode. Maybe? I hope so! Been dealing with this WAY TOO LONG!! Think it just caused a nervous breakdown. I need someone to help me with all this mxl and dex and open source code before it kills me. <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20221010181815</li><li>channel: release</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2022/10/39d38849-5f69-450d-bf2c-56b573dd76c3) _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
login.live.com - desktop site instead of mobile site - <!-- @browser: Firefox Mobile 106.0 --> <!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:106.0) Gecko/106.0 Firefox/106.0 --> <!-- @reported_with: android-components-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/112761 --> <!-- @extra_labels: browser-focus-geckoview --> **URL**: https://login.live.com/login.srf?wa=wsignin1.0&rpsnv=13&ct=1666489150&rver=7.0.6738.0&wp=MBI_SSL&wreply=https:%2F%2Faccount.microsoft.com%2Fauth%2Fcomplete-signin%3Fru%3Dhttps%253A%252F%252Faccount.microsoft.com%252F%253Frefd%253Dwww.microsoft.com&lc=1033&id=292666&lw=1&fl=easi2 **Browser / Version**: Firefox Mobile 106.0 **Operating System**: Android 10 **Tested Another Browser**: Yes Chrome **Problem type**: Desktop site instead of mobile site **Description**: Desktop site instead of mobile site **Steps to Reproduce**: Finally found a browser that seems to take me to actual sites I'm looking for instead of redirects or malware when I use desktop mode. Maybe? I hope so! Been dealing with this WAY TOO LONG!! Think it just caused a nervous breakdown. I need someone to help me with all this mxl and dex and open source code before it kills me. <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20221010181815</li><li>channel: release</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2022/10/39d38849-5f69-450d-bf2c-56b573dd76c3) _From [webcompat.com](https://webcompat.com/) with ❤️_
priority
login live com desktop site instead of mobile site url browser version firefox mobile operating system android tested another browser yes chrome problem type desktop site instead of mobile site description desktop site instead of mobile site steps to reproduce finally found a browser that seems to take me to actual sites i m looking for instead of redirects or malware when i use desktop mode maybe i hope so been dealing with this way too long think it just caused a nervous breakdown i need someone to help me with all this mxl and dex and open source code before it kills me browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel release hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
1
243,541
18,716,916,461
IssuesEvent
2021-11-03 06:50:38
synqs/pennylane-ls
https://api.github.com/repos/synqs/pennylane-ls
closed
Example for atomic mixtures
documentation
Make a notebook that examplifies our experiments on spin-changing collisions.
1.0
Example for atomic mixtures - Make a notebook that examplifies our experiments on spin-changing collisions.
non_priority
example for atomic mixtures make a notebook that examplifies our experiments on spin changing collisions
0
436,588
12,550,984,655
IssuesEvent
2020-06-06 13:10:45
googleapis/elixir-google-api
https://api.github.com/repos/googleapis/elixir-google-api
opened
Synthesis failed for LibraryAgent
autosynth failure priority: p1 type: bug
Hello! Autosynth couldn't regenerate LibraryAgent. :broken_heart: Here's the output from running `synth.py`: ``` 2020-06-06 06:10:03,550 autosynth [INFO] > logs will be written to: /tmpfs/src/github/synthtool/logs/googleapis/elixir-google-api 2020-06-06 06:10:04,368 autosynth [DEBUG] > Running: git config --global core.excludesfile /home/kbuilder/.autosynth-gitignore 2020-06-06 06:10:04,372 autosynth [DEBUG] > Running: git config user.name yoshi-automation 2020-06-06 06:10:04,375 autosynth [DEBUG] > Running: git config user.email yoshi-automation@google.com 2020-06-06 06:10:04,378 autosynth [DEBUG] > Running: git config push.default simple 2020-06-06 06:10:04,381 autosynth [DEBUG] > Running: git branch -f autosynth-libraryagent 2020-06-06 06:10:04,385 autosynth [DEBUG] > Running: git checkout autosynth-libraryagent Switched to branch 'autosynth-libraryagent' 2020-06-06 06:10:04,793 autosynth [INFO] > Running synthtool 2020-06-06 06:10:04,793 autosynth [INFO] > ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/library_agent/synth.metadata', 'synth.py', '--'] 2020-06-06 06:10:04,795 autosynth [DEBUG] > Running: /tmpfs/src/github/synthtool/env/bin/python3 -m synthtool --metadata clients/library_agent/synth.metadata synth.py -- LibraryAgent tee: /tmpfs/src/github/synthtool/logs/googleapis/elixir-google-api: Is a directory 2020-06-06 06:10:05,027 synthtool [DEBUG] > Executing /home/kbuilder/.cache/synthtool/elixir-google-api/synth.py. On branch autosynth-libraryagent nothing to commit, working tree clean 2020-06-06 06:10:06,923 synthtool [DEBUG] > Running: docker run --rm -v/tmpfs/tmp/tmppy6d0ggc/repo:/workspace -v/var/run/docker.sock:/var/run/docker.sock -e USER_GROUP=1000:1000 -w /workspace gcr.io/cloud-devrel-public-resources/elixir19 scripts/generate_client.sh LibraryAgent DEBUG:synthtool:Running: docker run --rm -v/tmpfs/tmp/tmppy6d0ggc/repo:/workspace -v/var/run/docker.sock:/var/run/docker.sock -e USER_GROUP=1000:1000 -w /workspace gcr.io/cloud-devrel-public-resources/elixir19 scripts/generate_client.sh LibraryAgent /workspace /workspace The mix.lock file was generated with a newer version of Hex. Update your client by running `mix local.hex` to avoid losing data. Resolving Hex dependencies... Dependency resolution completed: Unchanged: certifi 2.5.1 google_api_discovery 0.7.0 google_gax 0.3.2 hackney 1.15.2 idna 6.0.0 jason 1.2.1 metrics 1.0.1 mime 1.3.1 mimerl 1.2.0 oauth2 0.9.4 parse_trans 3.3.0 poison 3.1.0 ssl_verify_fun 1.1.5 temp 0.4.7 tesla 1.3.3 unicode_util_compat 0.4.1 * Getting google_api_discovery (Hex package) * Getting tesla (Hex package) * Getting oauth2 (Hex package) * Getting temp (Hex package) * Getting jason (Hex package) * Getting poison (Hex package) * Getting hackney (Hex package) * Getting certifi (Hex package) * Getting idna (Hex package) * Getting metrics (Hex package) * Getting mimerl (Hex package) * Getting ssl_verify_fun (Hex package) * Getting unicode_util_compat (Hex package) * Getting parse_trans (Hex package) * Getting mime (Hex package) * Getting google_gax (Hex package) The mix.lock file was generated with a newer version of Hex. Update your client by running `mix local.hex` to avoid losing data. ==> temp Compiling 3 files (.ex) Generated temp app ===> Compiling parse_trans ===> Compiling mimerl ===> Compiling metrics ===> Compiling unicode_util_compat ===> Compiling idna ==> jason Compiling 8 files (.ex) Generated jason app warning: String.strip/1 is deprecated. Use String.trim/1 instead /workspace/deps/poison/mix.exs:4 ==> poison Compiling 4 files (.ex) warning: Integer.to_char_list/2 is deprecated. Use Integer.to_charlist/2 instead lib/poison/encoder.ex:173 Generated poison app ==> ssl_verify_fun Compiling 7 files (.erl) Generated ssl_verify_fun app ===> Compiling certifi ===> Compiling hackney ==> oauth2 Compiling 13 files (.ex) Generated oauth2 app ==> mime Compiling 2 files (.ex) Generated mime app ==> tesla Compiling 26 files (.ex) Generated tesla app ==> google_gax Compiling 5 files (.ex) Generated google_gax app ==> google_api_discovery Compiling 21 files (.ex) Generated google_api_discovery app ==> google_apis Compiling 27 files (.ex) warning: System.cwd/0 is deprecated. Use File.cwd/0 instead lib/google_apis/publisher.ex:24 Generated google_apis app 13:10:39.529 [info] FETCHING: https://libraryagent.googleapis.com/$discovery/GOOGLE_REST_SIMPLE_URI?version=v1 13:10:39.924 [info] FETCHING: https://libraryagent.googleapis.com/$discovery/rest?version=v1 13:10:40.246 [info] FOUND: https://libraryagent.googleapis.com/$discovery/rest?version=v1 Revision check: old=20200511, new=20200604, generating=true Creating leading directories Writing GoogleExampleLibraryagentV1Book to clients/library_agent/lib/google_api/library_agent/v1/model/google_example_libraryagent_v1_book.ex. Writing GoogleExampleLibraryagentV1ListBooksResponse to clients/library_agent/lib/google_api/library_agent/v1/model/google_example_libraryagent_v1_list_books_response.ex. Writing GoogleExampleLibraryagentV1ListShelvesResponse to clients/library_agent/lib/google_api/library_agent/v1/model/google_example_libraryagent_v1_list_shelves_response.ex. Writing GoogleExampleLibraryagentV1Shelf to clients/library_agent/lib/google_api/library_agent/v1/model/google_example_libraryagent_v1_shelf.ex. Writing Shelves to clients/library_agent/lib/google_api/library_agent/v1/api/shelves.ex. Writing connection.ex. Writing metadata.ex. Writing mix.exs Writing README.md Writing LICENSE Writing .gitignore Writing config/config.exs Writing test/test_helper.exs 13:10:40.609 [info] Found only discovery_revision and/or formatting changes. Not significant enough for a PR. fixing file permissions 2020-06-06 06:10:43,824 synthtool [DEBUG] > Wrote metadata to clients/library_agent/synth.metadata. DEBUG:synthtool:Wrote metadata to clients/library_agent/synth.metadata. 2020-06-06 06:10:43,850 autosynth [DEBUG] > Running: git clean -fdx Removing __pycache__/ Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 615, in <module> main() File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 476, in main return _inner_main(temp_dir) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 555, in _inner_main ).synthesize(base_synth_log_path) File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 121, in synthesize with open(log_file_path, "rt") as fp: IsADirectoryError: [Errno 21] Is a directory: '/tmpfs/src/github/synthtool/logs/googleapis/elixir-google-api' ``` Google internal developers can see the full log [here](http://sponge2/results/invocations/567e2ab8-5e4f-4fb0-8bae-9d0cc90aa1af/targets/github%2Fsynthtool;config=default/tests;query=elixir-google-api;failed=false).
1.0
Synthesis failed for LibraryAgent - Hello! Autosynth couldn't regenerate LibraryAgent. :broken_heart: Here's the output from running `synth.py`: ``` 2020-06-06 06:10:03,550 autosynth [INFO] > logs will be written to: /tmpfs/src/github/synthtool/logs/googleapis/elixir-google-api 2020-06-06 06:10:04,368 autosynth [DEBUG] > Running: git config --global core.excludesfile /home/kbuilder/.autosynth-gitignore 2020-06-06 06:10:04,372 autosynth [DEBUG] > Running: git config user.name yoshi-automation 2020-06-06 06:10:04,375 autosynth [DEBUG] > Running: git config user.email yoshi-automation@google.com 2020-06-06 06:10:04,378 autosynth [DEBUG] > Running: git config push.default simple 2020-06-06 06:10:04,381 autosynth [DEBUG] > Running: git branch -f autosynth-libraryagent 2020-06-06 06:10:04,385 autosynth [DEBUG] > Running: git checkout autosynth-libraryagent Switched to branch 'autosynth-libraryagent' 2020-06-06 06:10:04,793 autosynth [INFO] > Running synthtool 2020-06-06 06:10:04,793 autosynth [INFO] > ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/library_agent/synth.metadata', 'synth.py', '--'] 2020-06-06 06:10:04,795 autosynth [DEBUG] > Running: /tmpfs/src/github/synthtool/env/bin/python3 -m synthtool --metadata clients/library_agent/synth.metadata synth.py -- LibraryAgent tee: /tmpfs/src/github/synthtool/logs/googleapis/elixir-google-api: Is a directory 2020-06-06 06:10:05,027 synthtool [DEBUG] > Executing /home/kbuilder/.cache/synthtool/elixir-google-api/synth.py. On branch autosynth-libraryagent nothing to commit, working tree clean 2020-06-06 06:10:06,923 synthtool [DEBUG] > Running: docker run --rm -v/tmpfs/tmp/tmppy6d0ggc/repo:/workspace -v/var/run/docker.sock:/var/run/docker.sock -e USER_GROUP=1000:1000 -w /workspace gcr.io/cloud-devrel-public-resources/elixir19 scripts/generate_client.sh LibraryAgent DEBUG:synthtool:Running: docker run --rm -v/tmpfs/tmp/tmppy6d0ggc/repo:/workspace -v/var/run/docker.sock:/var/run/docker.sock -e USER_GROUP=1000:1000 -w /workspace gcr.io/cloud-devrel-public-resources/elixir19 scripts/generate_client.sh LibraryAgent /workspace /workspace The mix.lock file was generated with a newer version of Hex. Update your client by running `mix local.hex` to avoid losing data. Resolving Hex dependencies... Dependency resolution completed: Unchanged: certifi 2.5.1 google_api_discovery 0.7.0 google_gax 0.3.2 hackney 1.15.2 idna 6.0.0 jason 1.2.1 metrics 1.0.1 mime 1.3.1 mimerl 1.2.0 oauth2 0.9.4 parse_trans 3.3.0 poison 3.1.0 ssl_verify_fun 1.1.5 temp 0.4.7 tesla 1.3.3 unicode_util_compat 0.4.1 * Getting google_api_discovery (Hex package) * Getting tesla (Hex package) * Getting oauth2 (Hex package) * Getting temp (Hex package) * Getting jason (Hex package) * Getting poison (Hex package) * Getting hackney (Hex package) * Getting certifi (Hex package) * Getting idna (Hex package) * Getting metrics (Hex package) * Getting mimerl (Hex package) * Getting ssl_verify_fun (Hex package) * Getting unicode_util_compat (Hex package) * Getting parse_trans (Hex package) * Getting mime (Hex package) * Getting google_gax (Hex package) The mix.lock file was generated with a newer version of Hex. Update your client by running `mix local.hex` to avoid losing data. ==> temp Compiling 3 files (.ex) Generated temp app ===> Compiling parse_trans ===> Compiling mimerl ===> Compiling metrics ===> Compiling unicode_util_compat ===> Compiling idna ==> jason Compiling 8 files (.ex) Generated jason app warning: String.strip/1 is deprecated. Use String.trim/1 instead /workspace/deps/poison/mix.exs:4 ==> poison Compiling 4 files (.ex) warning: Integer.to_char_list/2 is deprecated. Use Integer.to_charlist/2 instead lib/poison/encoder.ex:173 Generated poison app ==> ssl_verify_fun Compiling 7 files (.erl) Generated ssl_verify_fun app ===> Compiling certifi ===> Compiling hackney ==> oauth2 Compiling 13 files (.ex) Generated oauth2 app ==> mime Compiling 2 files (.ex) Generated mime app ==> tesla Compiling 26 files (.ex) Generated tesla app ==> google_gax Compiling 5 files (.ex) Generated google_gax app ==> google_api_discovery Compiling 21 files (.ex) Generated google_api_discovery app ==> google_apis Compiling 27 files (.ex) warning: System.cwd/0 is deprecated. Use File.cwd/0 instead lib/google_apis/publisher.ex:24 Generated google_apis app 13:10:39.529 [info] FETCHING: https://libraryagent.googleapis.com/$discovery/GOOGLE_REST_SIMPLE_URI?version=v1 13:10:39.924 [info] FETCHING: https://libraryagent.googleapis.com/$discovery/rest?version=v1 13:10:40.246 [info] FOUND: https://libraryagent.googleapis.com/$discovery/rest?version=v1 Revision check: old=20200511, new=20200604, generating=true Creating leading directories Writing GoogleExampleLibraryagentV1Book to clients/library_agent/lib/google_api/library_agent/v1/model/google_example_libraryagent_v1_book.ex. Writing GoogleExampleLibraryagentV1ListBooksResponse to clients/library_agent/lib/google_api/library_agent/v1/model/google_example_libraryagent_v1_list_books_response.ex. Writing GoogleExampleLibraryagentV1ListShelvesResponse to clients/library_agent/lib/google_api/library_agent/v1/model/google_example_libraryagent_v1_list_shelves_response.ex. Writing GoogleExampleLibraryagentV1Shelf to clients/library_agent/lib/google_api/library_agent/v1/model/google_example_libraryagent_v1_shelf.ex. Writing Shelves to clients/library_agent/lib/google_api/library_agent/v1/api/shelves.ex. Writing connection.ex. Writing metadata.ex. Writing mix.exs Writing README.md Writing LICENSE Writing .gitignore Writing config/config.exs Writing test/test_helper.exs 13:10:40.609 [info] Found only discovery_revision and/or formatting changes. Not significant enough for a PR. fixing file permissions 2020-06-06 06:10:43,824 synthtool [DEBUG] > Wrote metadata to clients/library_agent/synth.metadata. DEBUG:synthtool:Wrote metadata to clients/library_agent/synth.metadata. 2020-06-06 06:10:43,850 autosynth [DEBUG] > Running: git clean -fdx Removing __pycache__/ Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 615, in <module> main() File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 476, in main return _inner_main(temp_dir) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 555, in _inner_main ).synthesize(base_synth_log_path) File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 121, in synthesize with open(log_file_path, "rt") as fp: IsADirectoryError: [Errno 21] Is a directory: '/tmpfs/src/github/synthtool/logs/googleapis/elixir-google-api' ``` Google internal developers can see the full log [here](http://sponge2/results/invocations/567e2ab8-5e4f-4fb0-8bae-9d0cc90aa1af/targets/github%2Fsynthtool;config=default/tests;query=elixir-google-api;failed=false).
priority
synthesis failed for libraryagent hello autosynth couldn t regenerate libraryagent broken heart here s the output from running synth py autosynth logs will be written to tmpfs src github synthtool logs googleapis elixir google api autosynth running git config global core excludesfile home kbuilder autosynth gitignore autosynth running git config user name yoshi automation autosynth running git config user email yoshi automation google com autosynth running git config push default simple autosynth running git branch f autosynth libraryagent autosynth running git checkout autosynth libraryagent switched to branch autosynth libraryagent autosynth running synthtool autosynth autosynth running tmpfs src github synthtool env bin m synthtool metadata clients library agent synth metadata synth py libraryagent tee tmpfs src github synthtool logs googleapis elixir google api is a directory synthtool executing home kbuilder cache synthtool elixir google api synth py on branch autosynth libraryagent nothing to commit working tree clean synthtool running docker run rm v tmpfs tmp repo workspace v var run docker sock var run docker sock e user group w workspace gcr io cloud devrel public resources scripts generate client sh libraryagent debug synthtool running docker run rm v tmpfs tmp repo workspace v var run docker sock var run docker sock e user group w workspace gcr io cloud devrel public resources scripts generate client sh libraryagent workspace workspace  mix lock file was generated with a newer version of hex update your client by running mix local hex to avoid losing data  resolving hex dependencies dependency resolution completed unchanged certifi google api discovery google gax hackney idna jason metrics mime mimerl parse trans poison ssl verify fun temp tesla unicode util compat getting google api discovery hex package getting tesla hex package getting hex package getting temp hex package getting jason hex package getting poison hex package getting hackney hex package getting certifi hex package getting idna hex package getting metrics hex package getting mimerl hex package getting ssl verify fun hex package getting unicode util compat hex package getting parse trans hex package getting mime hex package getting google gax hex package  mix lock file was generated with a newer version of hex update your client by running mix local hex to avoid losing data  temp compiling files ex generated temp app compiling parse trans compiling mimerl compiling metrics compiling unicode util compat compiling idna jason compiling files ex generated jason app warning string strip is deprecated use string trim instead workspace deps poison mix exs poison compiling files ex warning integer to char list is deprecated use integer to charlist instead lib poison encoder ex generated poison app ssl verify fun compiling files erl generated ssl verify fun app compiling certifi compiling hackney compiling files ex generated app mime compiling files ex generated mime app tesla compiling files ex generated tesla app google gax compiling files ex generated google gax app google api discovery compiling files ex generated google api discovery app google apis compiling files ex warning system cwd is deprecated use file cwd instead lib google apis publisher ex generated google apis app fetching fetching found revision check old new generating true creating leading directories writing to clients library agent lib google api library agent model google example libraryagent book ex writing to clients library agent lib google api library agent model google example libraryagent list books response ex writing to clients library agent lib google api library agent model google example libraryagent list shelves response ex writing to clients library agent lib google api library agent model google example libraryagent shelf ex writing shelves to clients library agent lib google api library agent api shelves ex writing connection ex writing metadata ex writing mix exs writing readme md writing license writing gitignore writing config config exs writing test test helper exs found only discovery revision and or formatting changes not significant enough for a pr fixing file permissions synthtool wrote metadata to clients library agent synth metadata debug synthtool wrote metadata to clients library agent synth metadata autosynth running git clean fdx removing pycache traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool autosynth synth py line in main file tmpfs src github synthtool autosynth synth py line in main return inner main temp dir file tmpfs src github synthtool autosynth synth py line in inner main synthesize base synth log path file tmpfs src github synthtool autosynth synthesizer py line in synthesize with open log file path rt as fp isadirectoryerror is a directory tmpfs src github synthtool logs googleapis elixir google api google internal developers can see the full log
1
744,614
25,949,981,387
IssuesEvent
2022-12-17 13:08:44
docker-mailserver/docker-mailserver
https://api.github.com/repos/docker-mailserver/docker-mailserver
opened
[BUG] Building DMS fails with DOVECOT_COMMUNITY_REPO=1
kind/bug meta/needs triage priority/medium
### Miscellaneous first checks - [X] I checked that all ports are open and not blocked by my ISP / hosting provider. - [X] I know that SSL errors are likely the result of a wrong setup on the user side and not caused by DMS itself. I'm confident my setup is correct. ### Affected Component(s) Building DMS ### What happened and when does this occur? ```Markdown #18 31.67 [ TRACE ] Using Dovecot community repository #18 31.67 gpg: directory '/root/.gnupg' created #18 31.67 gpg: keybox '/root/.gnupg/pubring.kbx' created #18 31.68 % Total % Received % Xferd Average Speed Time Time Time Current #18 31.68 Dload Upload Total Spent Left Speed 100 3129 100 3129 0 0 41171 0 --:--:-- --:--:-- --:--:-- 41720 #18 31.75 gpg: /root/.gnupg/trustdb.gpg: trustdb created #18 31.75 gpg: key 18A348AEED409DA1: public key "Dovecot Community Edition <dovecot-ce@dovecot.org>" imported #18 31.76 gpg: Total number processed: 1 #18 31.76 gpg: imported: 1 #18 31.76 [ TRACE ] Updating Dovecot package signatures #18 31.84 Hit:1 http://deb.debian.org/debian bullseye InRelease #18 31.85 Hit:2 http://deb.debian.org/debian-security bullseye-security InRelease #18 31.86 Hit:3 http://deb.debian.org/debian bullseye-updates InRelease #18 31.91 Get:4 https://repo.dovecot.org/ce-2.3-latest/debian/bullseye bullseye InRelease [1747 B] #18 32.17 Get:5 https://repo.dovecot.org/ce-2.3-latest/debian/bullseye bullseye/main amd64 Packages [6759 B] #18 32.18 Fetched 8506 B in 0s (21.2 kB/s) #18 32.18 Reading package lists... #18 32.48 [ DEBUG ] Installing Dovecot #18 32.49 Reading package lists... #18 32.77 Building dependency tree... #18 32.87 Reading state information... #18 32.90 Some packages could not be installed. This may mean that you have #18 32.90 requested an impossible situation or if you are using the unstable #18 32.90 distribution that some required packages have not yet been created #18 32.90 or been moved out of Incoming. #18 32.90 The following information may help to resolve the situation: #18 32.90 #18 32.90 The following packages have unmet dependencies: #18 32.94 dovecot-fts-xapian : Depends: dovecot-abi-2.3.abiv13 #18 32.94 E: Unable to correct problems, you have held broken packages. #18 ERROR: executor failed running [/bin/bash -e -o pipefail -c /bin/bash /build/packages.sh]: exit code: 100 ``` ``` ### What did you expect to happen? ```Markdown Build should succeed. ``` ### How do we replicate the issue? ```Markdown `docker build -t test --build-arg DOVECOT_COMMUNITY_REPO=1 .` ``` ### DMS version edge ### What operating system is DMS running on? Linux ### Which operating system version? Debian 11 ### What instruction set architecture is DMS running on? x86_64 / AMD64 ### What container orchestration tool are you using? Docker ### docker-compose.yml _No response_ ### Relevant log output _No response_ ### Other relevant information _No response_ ### What level of experience do you have with Docker and mail servers? - [ ] I am inexperienced with docker - [ ] I am rather experienced with docker - [ ] I am inexperienced with mail servers - [ ] I am rather experienced with mail servers - [ ] I am uncomfortable with the CLI - [ ] I am rather comfortable with the CLI ### Code of conduct - [X] I have read this project's [Code of Conduct](https://github.com/docker-mailserver/docker-mailserver/blob/master/CODE_OF_CONDUCT.md) and I agree - [X] I have read the [README](https://github.com/docker-mailserver/docker-mailserver/blob/master/README.md) and the [documentation](https://docker-mailserver.github.io/docker-mailserver/edge/) and I searched the [issue tracker](https://github.com/docker-mailserver/docker-mailserver/issues?q=is%3Aissue) but could not find a solution ### Improvements to this form? _No response_
1.0
[BUG] Building DMS fails with DOVECOT_COMMUNITY_REPO=1 - ### Miscellaneous first checks - [X] I checked that all ports are open and not blocked by my ISP / hosting provider. - [X] I know that SSL errors are likely the result of a wrong setup on the user side and not caused by DMS itself. I'm confident my setup is correct. ### Affected Component(s) Building DMS ### What happened and when does this occur? ```Markdown #18 31.67 [ TRACE ] Using Dovecot community repository #18 31.67 gpg: directory '/root/.gnupg' created #18 31.67 gpg: keybox '/root/.gnupg/pubring.kbx' created #18 31.68 % Total % Received % Xferd Average Speed Time Time Time Current #18 31.68 Dload Upload Total Spent Left Speed 100 3129 100 3129 0 0 41171 0 --:--:-- --:--:-- --:--:-- 41720 #18 31.75 gpg: /root/.gnupg/trustdb.gpg: trustdb created #18 31.75 gpg: key 18A348AEED409DA1: public key "Dovecot Community Edition <dovecot-ce@dovecot.org>" imported #18 31.76 gpg: Total number processed: 1 #18 31.76 gpg: imported: 1 #18 31.76 [ TRACE ] Updating Dovecot package signatures #18 31.84 Hit:1 http://deb.debian.org/debian bullseye InRelease #18 31.85 Hit:2 http://deb.debian.org/debian-security bullseye-security InRelease #18 31.86 Hit:3 http://deb.debian.org/debian bullseye-updates InRelease #18 31.91 Get:4 https://repo.dovecot.org/ce-2.3-latest/debian/bullseye bullseye InRelease [1747 B] #18 32.17 Get:5 https://repo.dovecot.org/ce-2.3-latest/debian/bullseye bullseye/main amd64 Packages [6759 B] #18 32.18 Fetched 8506 B in 0s (21.2 kB/s) #18 32.18 Reading package lists... #18 32.48 [ DEBUG ] Installing Dovecot #18 32.49 Reading package lists... #18 32.77 Building dependency tree... #18 32.87 Reading state information... #18 32.90 Some packages could not be installed. This may mean that you have #18 32.90 requested an impossible situation or if you are using the unstable #18 32.90 distribution that some required packages have not yet been created #18 32.90 or been moved out of Incoming. #18 32.90 The following information may help to resolve the situation: #18 32.90 #18 32.90 The following packages have unmet dependencies: #18 32.94 dovecot-fts-xapian : Depends: dovecot-abi-2.3.abiv13 #18 32.94 E: Unable to correct problems, you have held broken packages. #18 ERROR: executor failed running [/bin/bash -e -o pipefail -c /bin/bash /build/packages.sh]: exit code: 100 ``` ``` ### What did you expect to happen? ```Markdown Build should succeed. ``` ### How do we replicate the issue? ```Markdown `docker build -t test --build-arg DOVECOT_COMMUNITY_REPO=1 .` ``` ### DMS version edge ### What operating system is DMS running on? Linux ### Which operating system version? Debian 11 ### What instruction set architecture is DMS running on? x86_64 / AMD64 ### What container orchestration tool are you using? Docker ### docker-compose.yml _No response_ ### Relevant log output _No response_ ### Other relevant information _No response_ ### What level of experience do you have with Docker and mail servers? - [ ] I am inexperienced with docker - [ ] I am rather experienced with docker - [ ] I am inexperienced with mail servers - [ ] I am rather experienced with mail servers - [ ] I am uncomfortable with the CLI - [ ] I am rather comfortable with the CLI ### Code of conduct - [X] I have read this project's [Code of Conduct](https://github.com/docker-mailserver/docker-mailserver/blob/master/CODE_OF_CONDUCT.md) and I agree - [X] I have read the [README](https://github.com/docker-mailserver/docker-mailserver/blob/master/README.md) and the [documentation](https://docker-mailserver.github.io/docker-mailserver/edge/) and I searched the [issue tracker](https://github.com/docker-mailserver/docker-mailserver/issues?q=is%3Aissue) but could not find a solution ### Improvements to this form? _No response_
priority
building dms fails with dovecot community repo miscellaneous first checks i checked that all ports are open and not blocked by my isp hosting provider i know that ssl errors are likely the result of a wrong setup on the user side and not caused by dms itself i m confident my setup is correct affected component s building dms what happened and when does this occur markdown using dovecot community repository gpg directory root gnupg created gpg keybox root gnupg pubring kbx created total received xferd average speed time time time current dload upload total spent left speed gpg root gnupg trustdb gpg trustdb created gpg key public key dovecot community edition imported gpg total number processed gpg imported updating dovecot package signatures hit bullseye inrelease hit bullseye security inrelease hit bullseye updates inrelease get bullseye inrelease get bullseye main packages fetched b in kb s reading package lists installing dovecot reading package lists building dependency tree reading state information some packages could not be installed this may mean that you have requested an impossible situation or if you are using the unstable distribution that some required packages have not yet been created or been moved out of incoming the following information may help to resolve the situation the following packages have unmet dependencies dovecot fts xapian depends dovecot abi e unable to correct problems you have held broken packages error executor failed running exit code what did you expect to happen markdown build should succeed how do we replicate the issue markdown docker build t test build arg dovecot community repo dms version edge what operating system is dms running on linux which operating system version debian what instruction set architecture is dms running on what container orchestration tool are you using docker docker compose yml no response relevant log output no response other relevant information no response what level of experience do you have with docker and mail servers i am inexperienced with docker i am rather experienced with docker i am inexperienced with mail servers i am rather experienced with mail servers i am uncomfortable with the cli i am rather comfortable with the cli code of conduct i have read this project s and i agree i have read the and the and i searched the but could not find a solution improvements to this form no response
1
674,232
23,043,672,978
IssuesEvent
2022-07-23 14:56:28
amplication/amplication
https://api.github.com/repos/amplication/amplication
closed
Client: Search bar position
type: bug type: style priority: low
**Describe the bug** The search bar should be next to the page title and not aligned to the right. **To Reproduce** Please change the order of the HTML elements to: ![image](https://user-images.githubusercontent.com/65485193/101747291-4df03e00-3ad4-11eb-8372-173b8a6e1e12.png) And delete the 'search-field' alignment left. **Screenshots** See below how it should be: ![image](https://user-images.githubusercontent.com/65485193/101742194-19c84d80-3ad3-11eb-912a-5c8ec1b76fcc.png)
1.0
Client: Search bar position - **Describe the bug** The search bar should be next to the page title and not aligned to the right. **To Reproduce** Please change the order of the HTML elements to: ![image](https://user-images.githubusercontent.com/65485193/101747291-4df03e00-3ad4-11eb-8372-173b8a6e1e12.png) And delete the 'search-field' alignment left. **Screenshots** See below how it should be: ![image](https://user-images.githubusercontent.com/65485193/101742194-19c84d80-3ad3-11eb-912a-5c8ec1b76fcc.png)
priority
client search bar position describe the bug the search bar should be next to the page title and not aligned to the right to reproduce please change the order of the html elements to and delete the search field alignment left screenshots see below how it should be
1
476,902
13,751,959,098
IssuesEvent
2020-10-06 13:58:17
fossasia/open-event-frontend
https://api.github.com/repos/fossasia/open-event-frontend
opened
Admin Sales Page: Prpvide buttons, filters and date selection
Priority: High enhancement
The admin sales page (https://eventyay.com/admin/sales) now shows events and their sales data. However, it is difficult to identify relevant events. Admins need to click through different pages to see which events are selling tickets. Also all events are listed by default - live, draft, past events, which possibly creates a load on the server. Please enhance the usability therefore as follows: 1. Show buttons: Events with * Paid tickets (default view) * Sold Tickets * Free Tickets * All Ticket Types 2. Dropdown filter: Events with * [ ] Live Events (pre-selected default view) * [ ] Draft Events * [ ] Past Events 3. Add a filter to choose a date period to show events from/to date 4. Enable sorting for each category (completed, placed, pending orders) according to: * number of tickets sold * sales number ![Screenshot from 2020-10-06 15-55-56](https://user-images.githubusercontent.com/1583873/95211265-a4977e00-07ec-11eb-9983-0fb4f3fc2f94.png)
1.0
Admin Sales Page: Prpvide buttons, filters and date selection - The admin sales page (https://eventyay.com/admin/sales) now shows events and their sales data. However, it is difficult to identify relevant events. Admins need to click through different pages to see which events are selling tickets. Also all events are listed by default - live, draft, past events, which possibly creates a load on the server. Please enhance the usability therefore as follows: 1. Show buttons: Events with * Paid tickets (default view) * Sold Tickets * Free Tickets * All Ticket Types 2. Dropdown filter: Events with * [ ] Live Events (pre-selected default view) * [ ] Draft Events * [ ] Past Events 3. Add a filter to choose a date period to show events from/to date 4. Enable sorting for each category (completed, placed, pending orders) according to: * number of tickets sold * sales number ![Screenshot from 2020-10-06 15-55-56](https://user-images.githubusercontent.com/1583873/95211265-a4977e00-07ec-11eb-9983-0fb4f3fc2f94.png)
priority
admin sales page prpvide buttons filters and date selection the admin sales page now shows events and their sales data however it is difficult to identify relevant events admins need to click through different pages to see which events are selling tickets also all events are listed by default live draft past events which possibly creates a load on the server please enhance the usability therefore as follows show buttons events with paid tickets default view sold tickets free tickets all ticket types dropdown filter events with live events pre selected default view draft events past events add a filter to choose a date period to show events from to date enable sorting for each category completed placed pending orders according to number of tickets sold sales number
1
231,968
25,556,882,752
IssuesEvent
2022-11-30 07:37:10
joshbnewton31080/WebGoat
https://api.github.com/repos/joshbnewton31080/WebGoat
opened
CVE-2022-41854 (Medium) detected in snakeyaml-1.28.jar
security vulnerability
## CVE-2022-41854 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>snakeyaml-1.28.jar</b></p></summary> <p>YAML 1.1 parser and emitter for Java</p> <p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p> <p>Path to dependency file: /webwolf/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.28/snakeyaml-1.28.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.28/snakeyaml-1.28.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.28/snakeyaml-1.28.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-validation-2.5.4.jar (Root Library) - spring-boot-starter-2.5.4.jar - :x: **snakeyaml-1.28.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/joshbnewton31080/WebGoat/commit/aca81bb5cfc690f9b2fbb585f70bf573be067d17">aca81bb5cfc690f9b2fbb585f70bf573be067d17</a></p> <p>Found in base branch: <b>develop</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Those using Snakeyaml to parse untrusted YAML files may be vulnerable to Denial of Service attacks (DOS). If the parser is running on user supplied input, an attacker may supply content that causes the parser to crash by stack overflow. This effect may support a denial of service attack. <p>Publish Date: 2022-11-11 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-41854>CVE-2022-41854</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p>
True
CVE-2022-41854 (Medium) detected in snakeyaml-1.28.jar - ## CVE-2022-41854 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>snakeyaml-1.28.jar</b></p></summary> <p>YAML 1.1 parser and emitter for Java</p> <p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p> <p>Path to dependency file: /webwolf/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.28/snakeyaml-1.28.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.28/snakeyaml-1.28.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.28/snakeyaml-1.28.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-validation-2.5.4.jar (Root Library) - spring-boot-starter-2.5.4.jar - :x: **snakeyaml-1.28.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/joshbnewton31080/WebGoat/commit/aca81bb5cfc690f9b2fbb585f70bf573be067d17">aca81bb5cfc690f9b2fbb585f70bf573be067d17</a></p> <p>Found in base branch: <b>develop</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Those using Snakeyaml to parse untrusted YAML files may be vulnerable to Denial of Service attacks (DOS). If the parser is running on user supplied input, an attacker may supply content that causes the parser to crash by stack overflow. This effect may support a denial of service attack. <p>Publish Date: 2022-11-11 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-41854>CVE-2022-41854</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p>
non_priority
cve medium detected in snakeyaml jar cve medium severity vulnerability vulnerable library snakeyaml jar yaml parser and emitter for java library home page a href path to dependency file webwolf pom xml path to vulnerable library home wss scanner repository org yaml snakeyaml snakeyaml jar home wss scanner repository org yaml snakeyaml snakeyaml jar home wss scanner repository org yaml snakeyaml snakeyaml jar dependency hierarchy spring boot starter validation jar root library spring boot starter jar x snakeyaml jar vulnerable library found in head commit a href found in base branch develop vulnerability details those using snakeyaml to parse untrusted yaml files may be vulnerable to denial of service attacks dos if the parser is running on user supplied input an attacker may supply content that causes the parser to crash by stack overflow this effect may support a denial of service attack publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href
0
50,787
3,006,617,545
IssuesEvent
2015-07-27 11:41:49
Itseez/opencv
https://api.github.com/repos/Itseez/opencv
opened
Opencv python module "cv" always reporting "Runtime error, module compiled against API version 9 but this version of numpy is 4"
affected: master auto-transferred bug category: none priority: normal
Transferred from http://code.opencv.org/issues/4479 ``` || Kun Huang on 2015-07-10 03:01 || Priority: Normal || Affected: branch 'master' (3.0-dev) || Category: None || Tracker: Bug || Difficulty: || PR: || Platform: x64 / Linux ``` Opencv python module "cv" always reporting "Runtime error, module compiled against API version 9 but this version of numpy is 4" ----------- ``` I have compiled opencv python module with numpy 1.9.0, and when I'm writing code importing cv2, it reports such an error. I'm quite sure my numpy version is 9, not 4. I recompiled opencv, it report this error again. How can I fix this error ? ``` History ------- ##### Alexander Alekhin on 2015-07-10 11:02 ``` This message is emitted by numpy development package. Looks like OpenCV uses wrong numpy development files or Python uses wrong numpy runtime library. 1) Could you check numpy version used in Python: <pre> import numpy print(numpy.__version__) print(numpy.__file__) </pre> 2) Try to check used numpy development files: <pre> import cv2 print cv2.getBuildInformation() </pre> Find line with numpy and "include" dir. a) Paths for "numpy.__file__" and include dir should be similar. b) You can also check "numpy/_numpyconfig.h" file here for lines "NPY_ABI_VERSION"/"NPY_API_VERSION" - Priority changed from High to Normal ```
1.0
Opencv python module "cv" always reporting "Runtime error, module compiled against API version 9 but this version of numpy is 4" - Transferred from http://code.opencv.org/issues/4479 ``` || Kun Huang on 2015-07-10 03:01 || Priority: Normal || Affected: branch 'master' (3.0-dev) || Category: None || Tracker: Bug || Difficulty: || PR: || Platform: x64 / Linux ``` Opencv python module "cv" always reporting "Runtime error, module compiled against API version 9 but this version of numpy is 4" ----------- ``` I have compiled opencv python module with numpy 1.9.0, and when I'm writing code importing cv2, it reports such an error. I'm quite sure my numpy version is 9, not 4. I recompiled opencv, it report this error again. How can I fix this error ? ``` History ------- ##### Alexander Alekhin on 2015-07-10 11:02 ``` This message is emitted by numpy development package. Looks like OpenCV uses wrong numpy development files or Python uses wrong numpy runtime library. 1) Could you check numpy version used in Python: <pre> import numpy print(numpy.__version__) print(numpy.__file__) </pre> 2) Try to check used numpy development files: <pre> import cv2 print cv2.getBuildInformation() </pre> Find line with numpy and "include" dir. a) Paths for "numpy.__file__" and include dir should be similar. b) You can also check "numpy/_numpyconfig.h" file here for lines "NPY_ABI_VERSION"/"NPY_API_VERSION" - Priority changed from High to Normal ```
priority
opencv python module cv always reporting runtime error module compiled against api version but this version of numpy is transferred from kun huang on priority normal affected branch master dev category none tracker bug difficulty pr platform linux opencv python module cv always reporting runtime error module compiled against api version but this version of numpy is i have compiled opencv python module with numpy and when i m writing code importing it reports such an error i m quite sure my numpy version is not i recompiled opencv it report this error again how can i fix this error history alexander alekhin on this message is emitted by numpy development package looks like opencv uses wrong numpy development files or python uses wrong numpy runtime library could you check numpy version used in python import numpy print numpy version print numpy file try to check used numpy development files import print getbuildinformation find line with numpy and include dir a paths for numpy file and include dir should be similar b you can also check numpy numpyconfig h file here for lines npy abi version npy api version priority changed from high to normal
1
153,162
24,081,794,558
IssuesEvent
2022-09-19 07:21:44
tarantool/doc
https://api.github.com/repos/tarantool/doc
closed
Make the Getting Started section more intuitive
design getting started 3sp
**Epics:** #2051 – Restructuring Tarantool docs #2052 – Integrate docs and trainings # To do Let's consolidate Getting Started pages in a unified Getting Started section. The audience of Getting Started is beginners who don't yet know much about Cartridge, app server functionality, etc. So it seems better to consolidate the Getting Started pages than to attach a separate Getting Started guide to every section. # Feedback 1. @ igormunkin: "Слушайте, а помните полгода назад еще был нормальный Getting Started без Cartridge: а он куда-то делся или я ослеп?" "А вот тот, где там box.insert с группами был, его видимо уже не найти?" "Нашёл! Но было сложно." https://www.tarantool.io/en/doc/latest/getting_started/getting_started_db/#creating-db-locally 2. This page (Pokemon Go guide) is apart from the other Getting Started pages: https://www.tarantool.io/en/doc/latest/book/app_server/creating_app/ 3. See this discussion: https://github.com/tarantool/doc/discussions/2810
1.0
Make the Getting Started section more intuitive - **Epics:** #2051 – Restructuring Tarantool docs #2052 – Integrate docs and trainings # To do Let's consolidate Getting Started pages in a unified Getting Started section. The audience of Getting Started is beginners who don't yet know much about Cartridge, app server functionality, etc. So it seems better to consolidate the Getting Started pages than to attach a separate Getting Started guide to every section. # Feedback 1. @ igormunkin: "Слушайте, а помните полгода назад еще был нормальный Getting Started без Cartridge: а он куда-то делся или я ослеп?" "А вот тот, где там box.insert с группами был, его видимо уже не найти?" "Нашёл! Но было сложно." https://www.tarantool.io/en/doc/latest/getting_started/getting_started_db/#creating-db-locally 2. This page (Pokemon Go guide) is apart from the other Getting Started pages: https://www.tarantool.io/en/doc/latest/book/app_server/creating_app/ 3. See this discussion: https://github.com/tarantool/doc/discussions/2810
non_priority
make the getting started section more intuitive epics – restructuring tarantool docs – integrate docs and trainings to do let s consolidate getting started pages in a unified getting started section the audience of getting started is beginners who don t yet know much about cartridge app server functionality etc so it seems better to consolidate the getting started pages than to attach a separate getting started guide to every section feedback igormunkin слушайте а помните полгода назад еще был нормальный getting started без cartridge а он куда то делся или я ослеп а вот тот где там box insert с группами был его видимо уже не найти нашёл но было сложно this page pokemon go guide is apart from the other getting started pages see this discussion
0
162,071
25,476,304,277
IssuesEvent
2022-11-25 14:47:00
frontend-horse/holiday-show-page-2022
https://api.github.com/repos/frontend-horse/holiday-show-page-2022
opened
feat(Website): Design Sponsor page
design
Designed on this [Figma File](https://www.figma.com/file/bfVsh47d0hhlsAfya2eBun/Holiday-Snowtacular-2022?node-id=0%3A1&t=UIQOqEbpmtd6JwNq-1). ## Info to include: - Company Name - Logo - Description - Link - Video Embed
1.0
feat(Website): Design Sponsor page - Designed on this [Figma File](https://www.figma.com/file/bfVsh47d0hhlsAfya2eBun/Holiday-Snowtacular-2022?node-id=0%3A1&t=UIQOqEbpmtd6JwNq-1). ## Info to include: - Company Name - Logo - Description - Link - Video Embed
non_priority
feat website design sponsor page designed on this info to include company name logo description link video embed
0
339,259
10,245,279,278
IssuesEvent
2019-08-20 12:31:04
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.manualeduso.it - Text overlaps the picture
browser-firefox-mobile engine-gecko priority-normal severity-minor type-css type-tracking-protection-basic
<!-- @browser: Firefox Mobile 68.0 --> <!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:68.0) Gecko/68.0 Firefox/68.0 --> <!-- @reported_with: mobile-reporter --> <!-- @extra_labels: type-tracking-protection-basic --> **URL**: https://www.manualeduso.it/imetec/eco-extreme-silence-pro/manuale?p=8 **Browser / Version**: Firefox Mobile 68.0 **Operating System**: Android **Tested Another Browser**: Yes **Problem type**: Design is broken **Description**: The page show incorrectly and I am not able to see the content correctly **Steps to Reproduce**: I just entered the site and this is not showed right. In Chrome it works well [![Screenshot Description](https://webcompat.com/uploads/2019/8/728ad9c0-1e40-4310-b18c-12187f13f9f8-thumb.jpeg)](https://webcompat.com/uploads/2019/8/728ad9c0-1e40-4310-b18c-12187f13f9f8.jpeg) <details> <summary>Browser Configuration</summary> <ul> <li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190812183651</li><li>tracking content blocked: true (basic)</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: true</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: beta</li> </ul> <p>Console Messages:</p> <pre> [u'[JavaScript Warning: "The resource at https://pagead2.googlesyndication.com/pagead/js/adsbygoogle.js was blocked because content blocking is enabled." {file: "https://www.manualeduso.it/imetec/eco-extreme-silence-pro/manuale" line: 0}]', u'[JavaScript Warning: "The resource at https://www.google-analytics.com/analytics.js was blocked because content blocking is enabled." {file: "https://www.manualeduso.it/imetec/eco-extreme-silence-pro/manuale" line: 0}]', u'[JavaScript Warning: "The resource at https://pagead2.googlesyndication.com/pagead/js/adsbygoogle.js was blocked because content blocking is enabled." {file: "https://www.manualeduso.it/imetec/eco-extreme-silence-pro/manuale" line: 0}]', u'[JavaScript Warning: "The resource at https://pagead2.googlesyndication.com/pagead/js/adsbygoogle.js was blocked because content blocking is enabled." {file: "https://www.manualeduso.it/imetec/eco-extreme-silence-pro/manuale" line: 0}]', u'[JavaScript Warning: "The resource at https://pagead2.googlesyndication.com/pagead/js/adsbygoogle.js was blocked because content blocking is enabled." {file: "https://www.manualeduso.it/imetec/eco-extreme-silence-pro/manuale" line: 0}]', u'[JavaScript Warning: "Caricamento non riuscito per lo <script> con sorgente https://pagead2.googlesyndication.com/pagead/js/adsbygoogle.js." {file: "https://www.manualeduso.it/imetec/eco-extreme-silence-pro/manuale" line: 191}]', u'[JavaScript Warning: "Caricamento non riuscito per lo <script> con sorgente https://pagead2.googlesyndication.com/pagead/js/adsbygoogle.js." {file: "https://www.manualeduso.it/imetec/eco-extreme-silence-pro/manuale" line: 292}]', u'[JavaScript Warning: "Caricamento non riuscito per lo <script> con sorgente https://pagead2.googlesyndication.com/pagead/js/adsbygoogle.js." {file: "https://www.manualeduso.it/imetec/eco-extreme-silence-pro/manuale" line: 381}]', u'[JavaScript Warning: "Caricamento non riuscito per lo <script> con sorgente https://pagead2.googlesyndication.com/pagead/js/adsbygoogle.js." {file: "https://www.manualeduso.it/imetec/eco-extreme-silence-pro/manuale" line: 1225}]'] </pre> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.manualeduso.it - Text overlaps the picture - <!-- @browser: Firefox Mobile 68.0 --> <!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:68.0) Gecko/68.0 Firefox/68.0 --> <!-- @reported_with: mobile-reporter --> <!-- @extra_labels: type-tracking-protection-basic --> **URL**: https://www.manualeduso.it/imetec/eco-extreme-silence-pro/manuale?p=8 **Browser / Version**: Firefox Mobile 68.0 **Operating System**: Android **Tested Another Browser**: Yes **Problem type**: Design is broken **Description**: The page show incorrectly and I am not able to see the content correctly **Steps to Reproduce**: I just entered the site and this is not showed right. In Chrome it works well [![Screenshot Description](https://webcompat.com/uploads/2019/8/728ad9c0-1e40-4310-b18c-12187f13f9f8-thumb.jpeg)](https://webcompat.com/uploads/2019/8/728ad9c0-1e40-4310-b18c-12187f13f9f8.jpeg) <details> <summary>Browser Configuration</summary> <ul> <li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190812183651</li><li>tracking content blocked: true (basic)</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: true</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: beta</li> </ul> <p>Console Messages:</p> <pre> [u'[JavaScript Warning: "The resource at https://pagead2.googlesyndication.com/pagead/js/adsbygoogle.js was blocked because content blocking is enabled." {file: "https://www.manualeduso.it/imetec/eco-extreme-silence-pro/manuale" line: 0}]', u'[JavaScript Warning: "The resource at https://www.google-analytics.com/analytics.js was blocked because content blocking is enabled." {file: "https://www.manualeduso.it/imetec/eco-extreme-silence-pro/manuale" line: 0}]', u'[JavaScript Warning: "The resource at https://pagead2.googlesyndication.com/pagead/js/adsbygoogle.js was blocked because content blocking is enabled." {file: "https://www.manualeduso.it/imetec/eco-extreme-silence-pro/manuale" line: 0}]', u'[JavaScript Warning: "The resource at https://pagead2.googlesyndication.com/pagead/js/adsbygoogle.js was blocked because content blocking is enabled." {file: "https://www.manualeduso.it/imetec/eco-extreme-silence-pro/manuale" line: 0}]', u'[JavaScript Warning: "The resource at https://pagead2.googlesyndication.com/pagead/js/adsbygoogle.js was blocked because content blocking is enabled." {file: "https://www.manualeduso.it/imetec/eco-extreme-silence-pro/manuale" line: 0}]', u'[JavaScript Warning: "Caricamento non riuscito per lo <script> con sorgente https://pagead2.googlesyndication.com/pagead/js/adsbygoogle.js." {file: "https://www.manualeduso.it/imetec/eco-extreme-silence-pro/manuale" line: 191}]', u'[JavaScript Warning: "Caricamento non riuscito per lo <script> con sorgente https://pagead2.googlesyndication.com/pagead/js/adsbygoogle.js." {file: "https://www.manualeduso.it/imetec/eco-extreme-silence-pro/manuale" line: 292}]', u'[JavaScript Warning: "Caricamento non riuscito per lo <script> con sorgente https://pagead2.googlesyndication.com/pagead/js/adsbygoogle.js." {file: "https://www.manualeduso.it/imetec/eco-extreme-silence-pro/manuale" line: 381}]', u'[JavaScript Warning: "Caricamento non riuscito per lo <script> con sorgente https://pagead2.googlesyndication.com/pagead/js/adsbygoogle.js." {file: "https://www.manualeduso.it/imetec/eco-extreme-silence-pro/manuale" line: 1225}]'] </pre> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
priority
text overlaps the picture url browser version firefox mobile operating system android tested another browser yes problem type design is broken description the page show incorrectly and i am not able to see the content correctly steps to reproduce i just entered the site and this is not showed right in chrome it works well browser configuration mixed active content blocked false image mem shared true buildid tracking content blocked true basic gfx webrender blob images true hastouchscreen true mixed passive content blocked false gfx webrender enabled false gfx webrender all false channel beta console messages u u u u u u u u from with ❤️
1
789,222
27,783,647,025
IssuesEvent
2023-03-17 00:00:08
grpc/grpc
https://api.github.com/repos/grpc/grpc
closed
AddListeningPort blocking after update to 1.52.0
kind/bug lang/c++ priority/P2 disposition/requires reporter action untriaged
I just updated grpc from 1.41.0 to 1.52.0 and tried to start my software again, but it hangs indefinitely on the function `AddListeningPort` when starting the server at 0.0.0.0:50051. ```c++ std::string server_address(fmt::format("0.0.0.0:{}", serverPort)); service = std::make_unique<grpc::station::StationServiceServer>(transferStationManager); grpc::ServerBuilder builder; builder.AddListeningPort(server_address, grpc::InsecureServerCredentials()); builder.RegisterService(service.get()); server = builder.BuildAndStart(); serviceServerActive = true; ``` ### What version of gRPC and what language are you using? 1.52.0 ### What operating system (Linux, Windows,...) and version? ubuntu 20.04 ### What runtime / compiler are you using (e.g. python version or version of gcc) gcc version 9.4.0 ### What did you do? I updated from 1.41.0 to 1.52.0. ### What did you expect to see? The same running application as before. ### What did you see instead? gRPC server is not starting up.
1.0
AddListeningPort blocking after update to 1.52.0 - I just updated grpc from 1.41.0 to 1.52.0 and tried to start my software again, but it hangs indefinitely on the function `AddListeningPort` when starting the server at 0.0.0.0:50051. ```c++ std::string server_address(fmt::format("0.0.0.0:{}", serverPort)); service = std::make_unique<grpc::station::StationServiceServer>(transferStationManager); grpc::ServerBuilder builder; builder.AddListeningPort(server_address, grpc::InsecureServerCredentials()); builder.RegisterService(service.get()); server = builder.BuildAndStart(); serviceServerActive = true; ``` ### What version of gRPC and what language are you using? 1.52.0 ### What operating system (Linux, Windows,...) and version? ubuntu 20.04 ### What runtime / compiler are you using (e.g. python version or version of gcc) gcc version 9.4.0 ### What did you do? I updated from 1.41.0 to 1.52.0. ### What did you expect to see? The same running application as before. ### What did you see instead? gRPC server is not starting up.
priority
addlisteningport blocking after update to i just updated grpc from to and tried to start my software again but it hangs indefinitely on the function addlisteningport when starting the server at c std string server address fmt format serverport service std make unique transferstationmanager grpc serverbuilder builder builder addlisteningport server address grpc insecureservercredentials builder registerservice service get server builder buildandstart serviceserveractive true what version of grpc and what language are you using what operating system linux windows and version ubuntu what runtime compiler are you using e g python version or version of gcc gcc version what did you do i updated from to what did you expect to see the same running application as before what did you see instead grpc server is not starting up
1
250,840
27,112,252,440
IssuesEvent
2023-02-15 16:04:05
MetaMask/snaps-monorepo
https://api.github.com/repos/MetaMask/snaps-monorepo
closed
Threat Modeling - snap_dialog - Spoofing
type-security threat-modeling
Reviewing the Threat Model for `snap_dialog` we found that: > An attacker can craft a dialog from Snap B to look like coming from Snap A. Opening this issue to discuss eventual controls to this threat.
True
Threat Modeling - snap_dialog - Spoofing - Reviewing the Threat Model for `snap_dialog` we found that: > An attacker can craft a dialog from Snap B to look like coming from Snap A. Opening this issue to discuss eventual controls to this threat.
non_priority
threat modeling snap dialog spoofing reviewing the threat model for snap dialog we found that an attacker can craft a dialog from snap b to look like coming from snap a opening this issue to discuss eventual controls to this threat
0
251,644
8,019,798,984
IssuesEvent
2018-07-26 00:53:20
metabase/metabase
https://api.github.com/repos/metabase/metabase
closed
Metabase should not default port in SQL Server database connection - prevents use of dynamic ports
Bug Database/SQLServer Priority/P2
In version 0.26, the SQL Server driver was changed from the jTDS driver to Microsoft's JDBC driver (#3821). Since then, there have been problems connecting to SQL Server for various reasons. The history of these problems are in #6056 . One of the problems is related to dynamic ports which I will detail here. In SQL Server, the default instance listens on TCP port 1433, named instances use dynamic ports. When using dynamic ports, an available port is selected whenever the SQL Server service is started or restarted - this means the port number can't be known ahead of time. If the instance uses dynamic ports and a port is specified in the connection string, the jTDS driver ignores the supplied port and gets the port from the SQL Server Browser service (which runs on port 1434). That is not true for the Microsoft driver - if a port is specified, that value is used. If a port is not specified, the driver asks the SQL Server Browser service and the appropriate port is used. Metabase is defaulting the port to 1433 even when a port is not supplied in the Admin Panel. This means there is no way to connect to SQL Server using dynamic ports. Dynamic ports are not just reserved for named instances, they are preferred from a security stand point. Metabase should not set a default value for the port, the user should have to set it in the Admin Panel. If a value is not entered, the port should not be in the connection string. To confirm this, I created a Java program that connects to SQL Server using the Microsoft driver. I experimented with fixed and dynamic ports. In summary, I experienced the following: Works with Fixed port 1433, port in connection string Works with Fixed port 1433, port not in connection string Works with Fixed port 16000, port in connection string Fails with Fixed port 16000, port not in connection string Fails with Dynamic ports, port in connection string as 1433 Works with Dynamic ports, port not in connection string I tried these same settings in Metabase and got the same result except for the last test. Even though a port was not supplied in the Admin Panel, I still got an error. The field has a grayed out "1433" which is apparently being used in the connection. Here is a helpful article on how to make ports fixed or dynamic in SQL Server: https://docs.microsoft.com/en-us/sql/database-engine/configure-windows/configure-a-server-to-listen-on-a-specific-tcp-port?view=sql-server-2017 This discourse discussion is related: http://discourse.metabase.com/t/adding-sql-server-with-named-instance-as-source/1334/6 Your browser and the version: Firefox 59.0.2 (64-bit) Your operating system: Windows 7 Your databases: SQL Server 2012 Metabase version: 0.24.1 (works), 0.26 (doesn't work), 0.28.6 (doesn't work) Metabase hosting environment: jar on Win 7 - SQL Server, jar, and browser all on same local host Metabase internal database: H2 (default)
1.0
Metabase should not default port in SQL Server database connection - prevents use of dynamic ports - In version 0.26, the SQL Server driver was changed from the jTDS driver to Microsoft's JDBC driver (#3821). Since then, there have been problems connecting to SQL Server for various reasons. The history of these problems are in #6056 . One of the problems is related to dynamic ports which I will detail here. In SQL Server, the default instance listens on TCP port 1433, named instances use dynamic ports. When using dynamic ports, an available port is selected whenever the SQL Server service is started or restarted - this means the port number can't be known ahead of time. If the instance uses dynamic ports and a port is specified in the connection string, the jTDS driver ignores the supplied port and gets the port from the SQL Server Browser service (which runs on port 1434). That is not true for the Microsoft driver - if a port is specified, that value is used. If a port is not specified, the driver asks the SQL Server Browser service and the appropriate port is used. Metabase is defaulting the port to 1433 even when a port is not supplied in the Admin Panel. This means there is no way to connect to SQL Server using dynamic ports. Dynamic ports are not just reserved for named instances, they are preferred from a security stand point. Metabase should not set a default value for the port, the user should have to set it in the Admin Panel. If a value is not entered, the port should not be in the connection string. To confirm this, I created a Java program that connects to SQL Server using the Microsoft driver. I experimented with fixed and dynamic ports. In summary, I experienced the following: Works with Fixed port 1433, port in connection string Works with Fixed port 1433, port not in connection string Works with Fixed port 16000, port in connection string Fails with Fixed port 16000, port not in connection string Fails with Dynamic ports, port in connection string as 1433 Works with Dynamic ports, port not in connection string I tried these same settings in Metabase and got the same result except for the last test. Even though a port was not supplied in the Admin Panel, I still got an error. The field has a grayed out "1433" which is apparently being used in the connection. Here is a helpful article on how to make ports fixed or dynamic in SQL Server: https://docs.microsoft.com/en-us/sql/database-engine/configure-windows/configure-a-server-to-listen-on-a-specific-tcp-port?view=sql-server-2017 This discourse discussion is related: http://discourse.metabase.com/t/adding-sql-server-with-named-instance-as-source/1334/6 Your browser and the version: Firefox 59.0.2 (64-bit) Your operating system: Windows 7 Your databases: SQL Server 2012 Metabase version: 0.24.1 (works), 0.26 (doesn't work), 0.28.6 (doesn't work) Metabase hosting environment: jar on Win 7 - SQL Server, jar, and browser all on same local host Metabase internal database: H2 (default)
priority
metabase should not default port in sql server database connection prevents use of dynamic ports in version the sql server driver was changed from the jtds driver to microsoft s jdbc driver since then there have been problems connecting to sql server for various reasons the history of these problems are in one of the problems is related to dynamic ports which i will detail here in sql server the default instance listens on tcp port named instances use dynamic ports when using dynamic ports an available port is selected whenever the sql server service is started or restarted this means the port number can t be known ahead of time if the instance uses dynamic ports and a port is specified in the connection string the jtds driver ignores the supplied port and gets the port from the sql server browser service which runs on port that is not true for the microsoft driver if a port is specified that value is used if a port is not specified the driver asks the sql server browser service and the appropriate port is used metabase is defaulting the port to even when a port is not supplied in the admin panel this means there is no way to connect to sql server using dynamic ports dynamic ports are not just reserved for named instances they are preferred from a security stand point metabase should not set a default value for the port the user should have to set it in the admin panel if a value is not entered the port should not be in the connection string to confirm this i created a java program that connects to sql server using the microsoft driver i experimented with fixed and dynamic ports in summary i experienced the following works with fixed port port in connection string works with fixed port port not in connection string works with fixed port port in connection string fails with fixed port port not in connection string fails with dynamic ports port in connection string as works with dynamic ports port not in connection string i tried these same settings in metabase and got the same result except for the last test even though a port was not supplied in the admin panel i still got an error the field has a grayed out which is apparently being used in the connection here is a helpful article on how to make ports fixed or dynamic in sql server this discourse discussion is related your browser and the version firefox bit your operating system windows your databases sql server metabase version works doesn t work doesn t work metabase hosting environment jar on win sql server jar and browser all on same local host metabase internal database default
1
219,443
24,493,673,711
IssuesEvent
2022-10-10 06:29:38
vasind/try-ember
https://api.github.com/repos/vasind/try-ember
closed
CVE-2020-28498 (Medium) detected in elliptic-6.5.2.tgz
security vulnerability no-issue-activity
## CVE-2020-28498 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>elliptic-6.5.2.tgz</b></p></summary> <p>EC cryptography</p> <p>Library home page: <a href="https://registry.npmjs.org/elliptic/-/elliptic-6.5.2.tgz">https://registry.npmjs.org/elliptic/-/elliptic-6.5.2.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/elliptic/package.json</p> <p> Dependency Hierarchy: - ember-intl-3.5.2.tgz (Root Library) - ember-auto-import-1.5.3.tgz - webpack-4.28.4.tgz - node-libs-browser-2.2.1.tgz - crypto-browserify-3.12.0.tgz - browserify-sign-4.0.4.tgz - :x: **elliptic-6.5.2.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package elliptic before 6.5.4 are vulnerable to Cryptographic Issues via the secp256k1 implementation in elliptic/ec/key.js. There is no check to confirm that the public key point passed into the derive function actually exists on the secp256k1 curve. This results in the potential for the private key used in this implementation to be revealed after a number of ECDH operations are performed. <p>Publish Date: 2021-02-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28498>CVE-2020-28498</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Changed - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28498">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28498</a></p> <p>Release Date: 2021-02-02</p> <p>Fix Resolution (elliptic): 6.5.4</p> <p>Direct dependency fix Resolution (ember-intl): 4.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-28498 (Medium) detected in elliptic-6.5.2.tgz - ## CVE-2020-28498 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>elliptic-6.5.2.tgz</b></p></summary> <p>EC cryptography</p> <p>Library home page: <a href="https://registry.npmjs.org/elliptic/-/elliptic-6.5.2.tgz">https://registry.npmjs.org/elliptic/-/elliptic-6.5.2.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/elliptic/package.json</p> <p> Dependency Hierarchy: - ember-intl-3.5.2.tgz (Root Library) - ember-auto-import-1.5.3.tgz - webpack-4.28.4.tgz - node-libs-browser-2.2.1.tgz - crypto-browserify-3.12.0.tgz - browserify-sign-4.0.4.tgz - :x: **elliptic-6.5.2.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package elliptic before 6.5.4 are vulnerable to Cryptographic Issues via the secp256k1 implementation in elliptic/ec/key.js. There is no check to confirm that the public key point passed into the derive function actually exists on the secp256k1 curve. This results in the potential for the private key used in this implementation to be revealed after a number of ECDH operations are performed. <p>Publish Date: 2021-02-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28498>CVE-2020-28498</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Changed - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28498">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28498</a></p> <p>Release Date: 2021-02-02</p> <p>Fix Resolution (elliptic): 6.5.4</p> <p>Direct dependency fix Resolution (ember-intl): 4.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve medium detected in elliptic tgz cve medium severity vulnerability vulnerable library elliptic tgz ec cryptography library home page a href path to dependency file package json path to vulnerable library node modules elliptic package json dependency hierarchy ember intl tgz root library ember auto import tgz webpack tgz node libs browser tgz crypto browserify tgz browserify sign tgz x elliptic tgz vulnerable library vulnerability details the package elliptic before are vulnerable to cryptographic issues via the implementation in elliptic ec key js there is no check to confirm that the public key point passed into the derive function actually exists on the curve this results in the potential for the private key used in this implementation to be revealed after a number of ecdh operations are performed publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope changed impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution elliptic direct dependency fix resolution ember intl step up your open source security game with mend
0
623,818
19,680,634,095
IssuesEvent
2022-01-11 16:25:36
googleapis/google-p12-pem
https://api.github.com/repos/googleapis/google-p12-pem
closed
Build broken due to update of "colors" package
type: feature request priority: p3
The release #384 containing a security fix (https://github.com/googleapis/google-p12-pem/pull/382) appears to be blocked by CI due to the sabotaged colors package having updated (https://github.com/Marak/colors.js/issues/285) I notice that this repository doesn't have a `package-lock.json` or `yarn.lock` file in it - should one be added to help avoid this kind of unexpected breakage in future?
1.0
Build broken due to update of "colors" package - The release #384 containing a security fix (https://github.com/googleapis/google-p12-pem/pull/382) appears to be blocked by CI due to the sabotaged colors package having updated (https://github.com/Marak/colors.js/issues/285) I notice that this repository doesn't have a `package-lock.json` or `yarn.lock` file in it - should one be added to help avoid this kind of unexpected breakage in future?
priority
build broken due to update of colors package the release containing a security fix appears to be blocked by ci due to the sabotaged colors package having updated i notice that this repository doesn t have a package lock json or yarn lock file in it should one be added to help avoid this kind of unexpected breakage in future
1
43,878
13,040,723,838
IssuesEvent
2020-07-28 19:01:58
LevyForchh/clusterfuzz
https://api.github.com/repos/LevyForchh/clusterfuzz
opened
CVE-2016-1000111 (Medium) detected in Twisted-14.0.0.tar.bz2
security vulnerability
## CVE-2016-1000111 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Twisted-14.0.0.tar.bz2</b></p></summary> <p>An asynchronous networking framework written in Python</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/76/38/cf8f81c1d7d84fec922d67f0d92bfa9fee59145d875d7263ceefa2bbbaf4/Twisted-14.0.0.tar.bz2">https://files.pythonhosted.org/packages/76/38/cf8f81c1d7d84fec922d67f0d92bfa9fee59145d875d7263ceefa2bbbaf4/Twisted-14.0.0.tar.bz2</a></p> <p>Path to dependency file: /tmp/ws-ua_20200728185744_PPILLS/archiveExtraction_CNYBMU/20200728185744/ws-scm_depth_0/clusterfuzz/resources/platform/linux/peach/peach_mutator/peach_mutator/requirements.txt</p> <p>Path to vulnerable library: _depth_0/clusterfuzz/resources/platform/linux/peach/peach_mutator/peach_mutator/requirements.txt,/clusterfuzz/src/local/butler/scripts,/clusterfuzz/src/python/bot/untrusted_runner/build,_depth_0/clusterfuzz/resources/platform/linux/peach/peach_mutator/peach_mutator/third_party/peach/requirements.txt,/clusterfuzz/src/appengine/handlers/cron/project,/clusterfuzz/src/python/bot/tasks</p> <p> Dependency Hierarchy: - :x: **Twisted-14.0.0.tar.bz2** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/LevyForchh/clusterfuzz/commit/aefe4b0859891117218fba5984e5c3e753ea9597">aefe4b0859891117218fba5984e5c3e753ea9597</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Twisted before 16.3.1 does not attempt to address RFC 3875 section 4.1.18 namespace conflicts and therefore does not protect CGI applications from the presence of untrusted client data in the HTTP_PROXY environment variable, which might allow remote attackers to redirect a CGI application's outbound HTTP traffic to an arbitrary proxy server via a crafted Proxy header in an HTTP request, aka an "httpoxy" issue. <p>Publish Date: 2020-03-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-1000111>CVE-2016-1000111</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Python","packageName":"Twisted","packageVersion":"14.0.0","isTransitiveDependency":false,"dependencyTree":"Twisted:14.0.0","isMinimumFixVersionAvailable":false}],"vulnerabilityIdentifier":"CVE-2016-1000111","vulnerabilityDetails":"Twisted before 16.3.1 does not attempt to address RFC 3875 section 4.1.18 namespace conflicts and therefore does not protect CGI applications from the presence of untrusted client data in the HTTP_PROXY environment variable, which might allow remote attackers to redirect a CGI application\u0027s outbound HTTP traffic to an arbitrary proxy server via a crafted Proxy header in an HTTP request, aka an \"httpoxy\" issue.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-1000111","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
True
CVE-2016-1000111 (Medium) detected in Twisted-14.0.0.tar.bz2 - ## CVE-2016-1000111 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Twisted-14.0.0.tar.bz2</b></p></summary> <p>An asynchronous networking framework written in Python</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/76/38/cf8f81c1d7d84fec922d67f0d92bfa9fee59145d875d7263ceefa2bbbaf4/Twisted-14.0.0.tar.bz2">https://files.pythonhosted.org/packages/76/38/cf8f81c1d7d84fec922d67f0d92bfa9fee59145d875d7263ceefa2bbbaf4/Twisted-14.0.0.tar.bz2</a></p> <p>Path to dependency file: /tmp/ws-ua_20200728185744_PPILLS/archiveExtraction_CNYBMU/20200728185744/ws-scm_depth_0/clusterfuzz/resources/platform/linux/peach/peach_mutator/peach_mutator/requirements.txt</p> <p>Path to vulnerable library: _depth_0/clusterfuzz/resources/platform/linux/peach/peach_mutator/peach_mutator/requirements.txt,/clusterfuzz/src/local/butler/scripts,/clusterfuzz/src/python/bot/untrusted_runner/build,_depth_0/clusterfuzz/resources/platform/linux/peach/peach_mutator/peach_mutator/third_party/peach/requirements.txt,/clusterfuzz/src/appengine/handlers/cron/project,/clusterfuzz/src/python/bot/tasks</p> <p> Dependency Hierarchy: - :x: **Twisted-14.0.0.tar.bz2** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/LevyForchh/clusterfuzz/commit/aefe4b0859891117218fba5984e5c3e753ea9597">aefe4b0859891117218fba5984e5c3e753ea9597</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Twisted before 16.3.1 does not attempt to address RFC 3875 section 4.1.18 namespace conflicts and therefore does not protect CGI applications from the presence of untrusted client data in the HTTP_PROXY environment variable, which might allow remote attackers to redirect a CGI application's outbound HTTP traffic to an arbitrary proxy server via a crafted Proxy header in an HTTP request, aka an "httpoxy" issue. <p>Publish Date: 2020-03-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-1000111>CVE-2016-1000111</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Python","packageName":"Twisted","packageVersion":"14.0.0","isTransitiveDependency":false,"dependencyTree":"Twisted:14.0.0","isMinimumFixVersionAvailable":false}],"vulnerabilityIdentifier":"CVE-2016-1000111","vulnerabilityDetails":"Twisted before 16.3.1 does not attempt to address RFC 3875 section 4.1.18 namespace conflicts and therefore does not protect CGI applications from the presence of untrusted client data in the HTTP_PROXY environment variable, which might allow remote attackers to redirect a CGI application\u0027s outbound HTTP traffic to an arbitrary proxy server via a crafted Proxy header in an HTTP request, aka an \"httpoxy\" issue.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-1000111","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
non_priority
cve medium detected in twisted tar cve medium severity vulnerability vulnerable library twisted tar an asynchronous networking framework written in python library home page a href path to dependency file tmp ws ua ppills archiveextraction cnybmu ws scm depth clusterfuzz resources platform linux peach peach mutator peach mutator requirements txt path to vulnerable library depth clusterfuzz resources platform linux peach peach mutator peach mutator requirements txt clusterfuzz src local butler scripts clusterfuzz src python bot untrusted runner build depth clusterfuzz resources platform linux peach peach mutator peach mutator third party peach requirements txt clusterfuzz src appengine handlers cron project clusterfuzz src python bot tasks dependency hierarchy x twisted tar vulnerable library found in head commit a href vulnerability details twisted before does not attempt to address rfc section namespace conflicts and therefore does not protect cgi applications from the presence of untrusted client data in the http proxy environment variable which might allow remote attackers to redirect a cgi application s outbound http traffic to an arbitrary proxy server via a crafted proxy header in an http request aka an httpoxy issue publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails twisted before does not attempt to address rfc section namespace conflicts and therefore does not protect cgi applications from the presence of untrusted client data in the http proxy environment variable which might allow remote attackers to redirect a cgi application outbound http traffic to an arbitrary proxy server via a crafted proxy header in an http request aka an httpoxy issue vulnerabilityurl
0
942
4,699,072,489
IssuesEvent
2016-10-12 14:44:33
Guite/MostGenerator
https://api.github.com/repos/Guite/MostGenerator
closed
eliminate Api's that are called via ModUtil
architecture zk 1.4.x
for example here: https://github.com/zikula/core/blob/4b8a4b51c5b0e4863b57672dbf18124bd65a0c02/src/system/RoutesModule/Controller/Base/RouteController.php#L246 this should be eliminated as ModUtil is incompatible with Core-2.0 refactor to a helper class or service.
1.0
eliminate Api's that are called via ModUtil - for example here: https://github.com/zikula/core/blob/4b8a4b51c5b0e4863b57672dbf18124bd65a0c02/src/system/RoutesModule/Controller/Base/RouteController.php#L246 this should be eliminated as ModUtil is incompatible with Core-2.0 refactor to a helper class or service.
non_priority
eliminate api s that are called via modutil for example here this should be eliminated as modutil is incompatible with core refactor to a helper class or service
0
280,755
8,686,268,547
IssuesEvent
2018-12-03 10:20:08
cosmos/voyager
https://api.github.com/repos/cosmos/voyager
opened
add e2e test case to check if tally increased after vote
blocked ✋ governance-1 :ballot_box: high priority
Description: <!-- Steps to reproduce, logs, and screenshots are helpful for us to resolve the bug --> Left a `TODO` on #1675, blocked by the latest SDK release
1.0
add e2e test case to check if tally increased after vote - Description: <!-- Steps to reproduce, logs, and screenshots are helpful for us to resolve the bug --> Left a `TODO` on #1675, blocked by the latest SDK release
priority
add test case to check if tally increased after vote description left a todo on blocked by the latest sdk release
1
586,268
17,573,931,127
IssuesEvent
2021-08-15 08:20:56
googleapis/python-spanner
https://api.github.com/repos/googleapis/python-spanner
closed
tests.system.test_instance_api: test_create_instance failed
api: spanner type: bug priority: p1 flakybot: issue
This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: bdd5f8b201d1b442837d4fca1d631fe171e276b9 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/dee9f019-ea44-420a-8fc8-e42e0e5755b9), [Sponge](http://sponge2/dee9f019-ea44-420a-8fc8-e42e0e5755b9) status: failed <details><summary>Test output</summary><br><pre>target = functools.partial(<bound method PollingFuture._done_or_raise of <google.api_core.operation.Operation object at 0x7f6d812aae80>>) predicate = <function if_exception_type.<locals>.if_exception_type_predicate at 0x7f6d8182d3a0> sleep_generator = <generator object exponential_sleep_generator at 0x7f6d8145ef90> deadline = 60, on_error = None def retry_target(target, predicate, sleep_generator, deadline, on_error=None): """Call a function and retry if it fails. This is the lowest-level retry helper. Generally, you'll use the higher-level retry helper :class:`Retry`. Args: target(Callable): The function to call and retry. This must be a nullary function - apply arguments with `functools.partial`. predicate (Callable[Exception]): A callable used to determine if an exception raised by the target should be considered retryable. It should return True to retry or False otherwise. sleep_generator (Iterable[float]): An infinite iterator that determines how long to sleep between retries. deadline (float): How long to keep retrying the target. The last sleep period is shortened as necessary, so that the last retry runs at ``deadline`` (and not considerably beyond it). on_error (Callable[Exception]): A function to call while processing a retryable exception. Any error raised by this function will *not* be caught. Returns: Any: the return value of the target function. Raises: google.api_core.RetryError: If the deadline is exceeded while retrying. ValueError: If the sleep generator stops yielding values. Exception: If the target raises a method that isn't retryable. """ if deadline is not None: deadline_datetime = datetime_helpers.utcnow() + datetime.timedelta( seconds=deadline ) else: deadline_datetime = None last_exc = None for sleep in sleep_generator: try: > return target() .nox/system-3-8/lib/python3.8/site-packages/google/api_core/retry.py:189: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <google.api_core.operation.Operation object at 0x7f6d812aae80> retry = <google.api_core.retry.Retry object at 0x7f6d81831430> def _done_or_raise(self, retry=DEFAULT_RETRY): """Check if the future is done and raise if it's not.""" kwargs = {} if retry is DEFAULT_RETRY else {"retry": retry} if not self.done(**kwargs): > raise _OperationNotComplete() E google.api_core.future.polling._OperationNotComplete .nox/system-3-8/lib/python3.8/site-packages/google/api_core/future/polling.py:87: _OperationNotComplete The above exception was the direct cause of the following exception: self = <google.api_core.operation.Operation object at 0x7f6d812aae80> timeout = 60, retry = <google.api_core.retry.Retry object at 0x7f6d81831430> def _blocking_poll(self, timeout=None, retry=DEFAULT_RETRY): """Poll and wait for the Future to be resolved. Args: timeout (int): How long (in seconds) to wait for the operation to complete. If None, wait indefinitely. """ if self._result_set: return retry_ = self._retry.with_deadline(timeout) try: kwargs = {} if retry is DEFAULT_RETRY else {"retry": retry} > retry_(self._done_or_raise)(**kwargs) .nox/system-3-8/lib/python3.8/site-packages/google/api_core/future/polling.py:108: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ args = (), kwargs = {} target = functools.partial(<bound method PollingFuture._done_or_raise of <google.api_core.operation.Operation object at 0x7f6d812aae80>>) sleep_generator = <generator object exponential_sleep_generator at 0x7f6d8145ef90> @general_helpers.wraps(func) def retry_wrapped_func(*args, **kwargs): """A wrapper that calls target function with retry.""" target = functools.partial(func, *args, **kwargs) sleep_generator = exponential_sleep_generator( self._initial, self._maximum, multiplier=self._multiplier ) > return retry_target( target, self._predicate, sleep_generator, self._deadline, on_error=on_error, ) .nox/system-3-8/lib/python3.8/site-packages/google/api_core/retry.py:286: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ target = functools.partial(<bound method PollingFuture._done_or_raise of <google.api_core.operation.Operation object at 0x7f6d812aae80>>) predicate = <function if_exception_type.<locals>.if_exception_type_predicate at 0x7f6d8182d3a0> sleep_generator = <generator object exponential_sleep_generator at 0x7f6d8145ef90> deadline = 60, on_error = None def retry_target(target, predicate, sleep_generator, deadline, on_error=None): """Call a function and retry if it fails. This is the lowest-level retry helper. Generally, you'll use the higher-level retry helper :class:`Retry`. Args: target(Callable): The function to call and retry. This must be a nullary function - apply arguments with `functools.partial`. predicate (Callable[Exception]): A callable used to determine if an exception raised by the target should be considered retryable. It should return True to retry or False otherwise. sleep_generator (Iterable[float]): An infinite iterator that determines how long to sleep between retries. deadline (float): How long to keep retrying the target. The last sleep period is shortened as necessary, so that the last retry runs at ``deadline`` (and not considerably beyond it). on_error (Callable[Exception]): A function to call while processing a retryable exception. Any error raised by this function will *not* be caught. Returns: Any: the return value of the target function. Raises: google.api_core.RetryError: If the deadline is exceeded while retrying. ValueError: If the sleep generator stops yielding values. Exception: If the target raises a method that isn't retryable. """ if deadline is not None: deadline_datetime = datetime_helpers.utcnow() + datetime.timedelta( seconds=deadline ) else: deadline_datetime = None last_exc = None for sleep in sleep_generator: try: return target() # pylint: disable=broad-except # This function explicitly must deal with broad exceptions. except Exception as exc: if not predicate(exc): raise last_exc = exc if on_error is not None: on_error(exc) now = datetime_helpers.utcnow() if deadline_datetime is not None: if deadline_datetime <= now: > six.raise_from( exceptions.RetryError( "Deadline of {:.1f}s exceeded while calling {}".format( deadline, target ), last_exc, ), last_exc, ) .nox/system-3-8/lib/python3.8/site-packages/google/api_core/retry.py:204: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ value = None, from_value = _OperationNotComplete() > ??? E google.api_core.exceptions.RetryError: Deadline of 60.0s exceeded while calling functools.partial(<bound method PollingFuture._done_or_raise of <google.api_core.operation.Operation object at 0x7f6d812aae80>>), last exception: <string>:3: RetryError During handling of the above exception, another exception occurred: if_create_instance = None spanner_client = <google.cloud.spanner_v1.client.Client object at 0x7f6d811926d0> instance_config = name: "projects/precise-truck-742/instanceConfigs/regional-us-central1" display_name: "us-central1" replicas { locat... replicas { location: "us-central1" type_: READ_WRITE } replicas { location: "us-central1" type_: READ_WRITE } instances_to_delete = [<google.cloud.spanner_v1.instance.Instance object at 0x7f6d812aa0d0>] operation_timeout = 60 def test_create_instance( if_create_instance, spanner_client, instance_config, instances_to_delete, operation_timeout, ): alt_instance_id = _helpers.unique_id("new") instance = spanner_client.instance(alt_instance_id, instance_config.name) operation = instance.create() # Make sure this instance gets deleted after the test case. instances_to_delete.append(instance) # We want to make sure the operation completes. > operation.result(operation_timeout) # raises on failure / timeout. tests/system/test_instance_api.py:73: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ .nox/system-3-8/lib/python3.8/site-packages/google/api_core/future/polling.py:130: in result self._blocking_poll(timeout=timeout, **kwargs) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <google.api_core.operation.Operation object at 0x7f6d812aae80> timeout = 60, retry = <google.api_core.retry.Retry object at 0x7f6d81831430> def _blocking_poll(self, timeout=None, retry=DEFAULT_RETRY): """Poll and wait for the Future to be resolved. Args: timeout (int): How long (in seconds) to wait for the operation to complete. If None, wait indefinitely. """ if self._result_set: return retry_ = self._retry.with_deadline(timeout) try: kwargs = {} if retry is DEFAULT_RETRY else {"retry": retry} retry_(self._done_or_raise)(**kwargs) except exceptions.RetryError: > raise concurrent.futures.TimeoutError( "Operation did not complete within the designated " "timeout." ) E concurrent.futures._base.TimeoutError: Operation did not complete within the designated timeout. .nox/system-3-8/lib/python3.8/site-packages/google/api_core/future/polling.py:110: TimeoutError</pre></details>
1.0
tests.system.test_instance_api: test_create_instance failed - This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: bdd5f8b201d1b442837d4fca1d631fe171e276b9 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/dee9f019-ea44-420a-8fc8-e42e0e5755b9), [Sponge](http://sponge2/dee9f019-ea44-420a-8fc8-e42e0e5755b9) status: failed <details><summary>Test output</summary><br><pre>target = functools.partial(<bound method PollingFuture._done_or_raise of <google.api_core.operation.Operation object at 0x7f6d812aae80>>) predicate = <function if_exception_type.<locals>.if_exception_type_predicate at 0x7f6d8182d3a0> sleep_generator = <generator object exponential_sleep_generator at 0x7f6d8145ef90> deadline = 60, on_error = None def retry_target(target, predicate, sleep_generator, deadline, on_error=None): """Call a function and retry if it fails. This is the lowest-level retry helper. Generally, you'll use the higher-level retry helper :class:`Retry`. Args: target(Callable): The function to call and retry. This must be a nullary function - apply arguments with `functools.partial`. predicate (Callable[Exception]): A callable used to determine if an exception raised by the target should be considered retryable. It should return True to retry or False otherwise. sleep_generator (Iterable[float]): An infinite iterator that determines how long to sleep between retries. deadline (float): How long to keep retrying the target. The last sleep period is shortened as necessary, so that the last retry runs at ``deadline`` (and not considerably beyond it). on_error (Callable[Exception]): A function to call while processing a retryable exception. Any error raised by this function will *not* be caught. Returns: Any: the return value of the target function. Raises: google.api_core.RetryError: If the deadline is exceeded while retrying. ValueError: If the sleep generator stops yielding values. Exception: If the target raises a method that isn't retryable. """ if deadline is not None: deadline_datetime = datetime_helpers.utcnow() + datetime.timedelta( seconds=deadline ) else: deadline_datetime = None last_exc = None for sleep in sleep_generator: try: > return target() .nox/system-3-8/lib/python3.8/site-packages/google/api_core/retry.py:189: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <google.api_core.operation.Operation object at 0x7f6d812aae80> retry = <google.api_core.retry.Retry object at 0x7f6d81831430> def _done_or_raise(self, retry=DEFAULT_RETRY): """Check if the future is done and raise if it's not.""" kwargs = {} if retry is DEFAULT_RETRY else {"retry": retry} if not self.done(**kwargs): > raise _OperationNotComplete() E google.api_core.future.polling._OperationNotComplete .nox/system-3-8/lib/python3.8/site-packages/google/api_core/future/polling.py:87: _OperationNotComplete The above exception was the direct cause of the following exception: self = <google.api_core.operation.Operation object at 0x7f6d812aae80> timeout = 60, retry = <google.api_core.retry.Retry object at 0x7f6d81831430> def _blocking_poll(self, timeout=None, retry=DEFAULT_RETRY): """Poll and wait for the Future to be resolved. Args: timeout (int): How long (in seconds) to wait for the operation to complete. If None, wait indefinitely. """ if self._result_set: return retry_ = self._retry.with_deadline(timeout) try: kwargs = {} if retry is DEFAULT_RETRY else {"retry": retry} > retry_(self._done_or_raise)(**kwargs) .nox/system-3-8/lib/python3.8/site-packages/google/api_core/future/polling.py:108: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ args = (), kwargs = {} target = functools.partial(<bound method PollingFuture._done_or_raise of <google.api_core.operation.Operation object at 0x7f6d812aae80>>) sleep_generator = <generator object exponential_sleep_generator at 0x7f6d8145ef90> @general_helpers.wraps(func) def retry_wrapped_func(*args, **kwargs): """A wrapper that calls target function with retry.""" target = functools.partial(func, *args, **kwargs) sleep_generator = exponential_sleep_generator( self._initial, self._maximum, multiplier=self._multiplier ) > return retry_target( target, self._predicate, sleep_generator, self._deadline, on_error=on_error, ) .nox/system-3-8/lib/python3.8/site-packages/google/api_core/retry.py:286: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ target = functools.partial(<bound method PollingFuture._done_or_raise of <google.api_core.operation.Operation object at 0x7f6d812aae80>>) predicate = <function if_exception_type.<locals>.if_exception_type_predicate at 0x7f6d8182d3a0> sleep_generator = <generator object exponential_sleep_generator at 0x7f6d8145ef90> deadline = 60, on_error = None def retry_target(target, predicate, sleep_generator, deadline, on_error=None): """Call a function and retry if it fails. This is the lowest-level retry helper. Generally, you'll use the higher-level retry helper :class:`Retry`. Args: target(Callable): The function to call and retry. This must be a nullary function - apply arguments with `functools.partial`. predicate (Callable[Exception]): A callable used to determine if an exception raised by the target should be considered retryable. It should return True to retry or False otherwise. sleep_generator (Iterable[float]): An infinite iterator that determines how long to sleep between retries. deadline (float): How long to keep retrying the target. The last sleep period is shortened as necessary, so that the last retry runs at ``deadline`` (and not considerably beyond it). on_error (Callable[Exception]): A function to call while processing a retryable exception. Any error raised by this function will *not* be caught. Returns: Any: the return value of the target function. Raises: google.api_core.RetryError: If the deadline is exceeded while retrying. ValueError: If the sleep generator stops yielding values. Exception: If the target raises a method that isn't retryable. """ if deadline is not None: deadline_datetime = datetime_helpers.utcnow() + datetime.timedelta( seconds=deadline ) else: deadline_datetime = None last_exc = None for sleep in sleep_generator: try: return target() # pylint: disable=broad-except # This function explicitly must deal with broad exceptions. except Exception as exc: if not predicate(exc): raise last_exc = exc if on_error is not None: on_error(exc) now = datetime_helpers.utcnow() if deadline_datetime is not None: if deadline_datetime <= now: > six.raise_from( exceptions.RetryError( "Deadline of {:.1f}s exceeded while calling {}".format( deadline, target ), last_exc, ), last_exc, ) .nox/system-3-8/lib/python3.8/site-packages/google/api_core/retry.py:204: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ value = None, from_value = _OperationNotComplete() > ??? E google.api_core.exceptions.RetryError: Deadline of 60.0s exceeded while calling functools.partial(<bound method PollingFuture._done_or_raise of <google.api_core.operation.Operation object at 0x7f6d812aae80>>), last exception: <string>:3: RetryError During handling of the above exception, another exception occurred: if_create_instance = None spanner_client = <google.cloud.spanner_v1.client.Client object at 0x7f6d811926d0> instance_config = name: "projects/precise-truck-742/instanceConfigs/regional-us-central1" display_name: "us-central1" replicas { locat... replicas { location: "us-central1" type_: READ_WRITE } replicas { location: "us-central1" type_: READ_WRITE } instances_to_delete = [<google.cloud.spanner_v1.instance.Instance object at 0x7f6d812aa0d0>] operation_timeout = 60 def test_create_instance( if_create_instance, spanner_client, instance_config, instances_to_delete, operation_timeout, ): alt_instance_id = _helpers.unique_id("new") instance = spanner_client.instance(alt_instance_id, instance_config.name) operation = instance.create() # Make sure this instance gets deleted after the test case. instances_to_delete.append(instance) # We want to make sure the operation completes. > operation.result(operation_timeout) # raises on failure / timeout. tests/system/test_instance_api.py:73: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ .nox/system-3-8/lib/python3.8/site-packages/google/api_core/future/polling.py:130: in result self._blocking_poll(timeout=timeout, **kwargs) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <google.api_core.operation.Operation object at 0x7f6d812aae80> timeout = 60, retry = <google.api_core.retry.Retry object at 0x7f6d81831430> def _blocking_poll(self, timeout=None, retry=DEFAULT_RETRY): """Poll and wait for the Future to be resolved. Args: timeout (int): How long (in seconds) to wait for the operation to complete. If None, wait indefinitely. """ if self._result_set: return retry_ = self._retry.with_deadline(timeout) try: kwargs = {} if retry is DEFAULT_RETRY else {"retry": retry} retry_(self._done_or_raise)(**kwargs) except exceptions.RetryError: > raise concurrent.futures.TimeoutError( "Operation did not complete within the designated " "timeout." ) E concurrent.futures._base.TimeoutError: Operation did not complete within the designated timeout. .nox/system-3-8/lib/python3.8/site-packages/google/api_core/future/polling.py:110: TimeoutError</pre></details>
priority
tests system test instance api test create instance failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output target functools partial predicate if exception type predicate at sleep generator deadline on error none def retry target target predicate sleep generator deadline on error none call a function and retry if it fails this is the lowest level retry helper generally you ll use the higher level retry helper class retry args target callable the function to call and retry this must be a nullary function apply arguments with functools partial predicate callable a callable used to determine if an exception raised by the target should be considered retryable it should return true to retry or false otherwise sleep generator iterable an infinite iterator that determines how long to sleep between retries deadline float how long to keep retrying the target the last sleep period is shortened as necessary so that the last retry runs at deadline and not considerably beyond it on error callable a function to call while processing a retryable exception any error raised by this function will not be caught returns any the return value of the target function raises google api core retryerror if the deadline is exceeded while retrying valueerror if the sleep generator stops yielding values exception if the target raises a method that isn t retryable if deadline is not none deadline datetime datetime helpers utcnow datetime timedelta seconds deadline else deadline datetime none last exc none for sleep in sleep generator try return target nox system lib site packages google api core retry py self retry def done or raise self retry default retry check if the future is done and raise if it s not kwargs if retry is default retry else retry retry if not self done kwargs raise operationnotcomplete e google api core future polling operationnotcomplete nox system lib site packages google api core future polling py operationnotcomplete the above exception was the direct cause of the following exception self timeout retry def blocking poll self timeout none retry default retry poll and wait for the future to be resolved args timeout int how long in seconds to wait for the operation to complete if none wait indefinitely if self result set return retry self retry with deadline timeout try kwargs if retry is default retry else retry retry retry self done or raise kwargs nox system lib site packages google api core future polling py args kwargs target functools partial sleep generator general helpers wraps func def retry wrapped func args kwargs a wrapper that calls target function with retry target functools partial func args kwargs sleep generator exponential sleep generator self initial self maximum multiplier self multiplier return retry target target self predicate sleep generator self deadline on error on error nox system lib site packages google api core retry py target functools partial predicate if exception type predicate at sleep generator deadline on error none def retry target target predicate sleep generator deadline on error none call a function and retry if it fails this is the lowest level retry helper generally you ll use the higher level retry helper class retry args target callable the function to call and retry this must be a nullary function apply arguments with functools partial predicate callable a callable used to determine if an exception raised by the target should be considered retryable it should return true to retry or false otherwise sleep generator iterable an infinite iterator that determines how long to sleep between retries deadline float how long to keep retrying the target the last sleep period is shortened as necessary so that the last retry runs at deadline and not considerably beyond it on error callable a function to call while processing a retryable exception any error raised by this function will not be caught returns any the return value of the target function raises google api core retryerror if the deadline is exceeded while retrying valueerror if the sleep generator stops yielding values exception if the target raises a method that isn t retryable if deadline is not none deadline datetime datetime helpers utcnow datetime timedelta seconds deadline else deadline datetime none last exc none for sleep in sleep generator try return target pylint disable broad except this function explicitly must deal with broad exceptions except exception as exc if not predicate exc raise last exc exc if on error is not none on error exc now datetime helpers utcnow if deadline datetime is not none if deadline datetime now six raise from exceptions retryerror deadline of s exceeded while calling format deadline target last exc last exc nox system lib site packages google api core retry py value none from value operationnotcomplete e google api core exceptions retryerror deadline of exceeded while calling functools partial last exception retryerror during handling of the above exception another exception occurred if create instance none spanner client instance config name projects precise truck instanceconfigs regional us display name us replicas locat replicas location us type read write replicas location us type read write instances to delete operation timeout def test create instance if create instance spanner client instance config instances to delete operation timeout alt instance id helpers unique id new instance spanner client instance alt instance id instance config name operation instance create make sure this instance gets deleted after the test case instances to delete append instance we want to make sure the operation completes operation result operation timeout raises on failure timeout tests system test instance api py nox system lib site packages google api core future polling py in result self blocking poll timeout timeout kwargs self timeout retry def blocking poll self timeout none retry default retry poll and wait for the future to be resolved args timeout int how long in seconds to wait for the operation to complete if none wait indefinitely if self result set return retry self retry with deadline timeout try kwargs if retry is default retry else retry retry retry self done or raise kwargs except exceptions retryerror raise concurrent futures timeouterror operation did not complete within the designated timeout e concurrent futures base timeouterror operation did not complete within the designated timeout nox system lib site packages google api core future polling py timeouterror
1
589,400
17,695,962,798
IssuesEvent
2021-08-24 15:19:04
google/ground-android
https://api.github.com/repos/google/ground-android
opened
[Nav Drawer] Profile image appears higher than expected
type: bug priority: p3
The (G) appears to overlap with the top system bar, and is far from the user name. Ideally it would be positioned closer to the user's name. Some experimentation may be necessary to find the correct position, but I believe a 16dp gap between the two may be sufficient. ![image](https://user-images.githubusercontent.com/228050/130643382-3503b346-a8d8-4a21-a9b7-9d16d8b539f8.png) @shobhitagarwal1612 PTAL?
1.0
[Nav Drawer] Profile image appears higher than expected - The (G) appears to overlap with the top system bar, and is far from the user name. Ideally it would be positioned closer to the user's name. Some experimentation may be necessary to find the correct position, but I believe a 16dp gap between the two may be sufficient. ![image](https://user-images.githubusercontent.com/228050/130643382-3503b346-a8d8-4a21-a9b7-9d16d8b539f8.png) @shobhitagarwal1612 PTAL?
priority
profile image appears higher than expected the g appears to overlap with the top system bar and is far from the user name ideally it would be positioned closer to the user s name some experimentation may be necessary to find the correct position but i believe a gap between the two may be sufficient ptal
1
32,856
8,961,827,709
IssuesEvent
2019-01-28 10:42:19
j1nx/MycroftOS
https://api.github.com/repos/j1nx/MycroftOS
closed
Mycroft: PIP --binary-only and addition of piwheels
bug buildroot
### Information We do not have gcc on buildroot, so we have to make sure python-pip only installs binaries and related to that we need to add piwheels.org as secondairy repo for that. ### Bug description - Installing a mycroft skill that installs dependencies from requirments.txt should only use binaries as sources can not be compiled To increase the change of a binary candidate being found, piwheels should be added to the repositories. Any python dependency that does not have suitable binaries should be undentified at a later stage and pre-installed to the target at buildroot buildtime. (similar as all the mycroft dependencies)
1.0
Mycroft: PIP --binary-only and addition of piwheels - ### Information We do not have gcc on buildroot, so we have to make sure python-pip only installs binaries and related to that we need to add piwheels.org as secondairy repo for that. ### Bug description - Installing a mycroft skill that installs dependencies from requirments.txt should only use binaries as sources can not be compiled To increase the change of a binary candidate being found, piwheels should be added to the repositories. Any python dependency that does not have suitable binaries should be undentified at a later stage and pre-installed to the target at buildroot buildtime. (similar as all the mycroft dependencies)
non_priority
mycroft pip binary only and addition of piwheels information we do not have gcc on buildroot so we have to make sure python pip only installs binaries and related to that we need to add piwheels org as secondairy repo for that bug description installing a mycroft skill that installs dependencies from requirments txt should only use binaries as sources can not be compiled to increase the change of a binary candidate being found piwheels should be added to the repositories any python dependency that does not have suitable binaries should be undentified at a later stage and pre installed to the target at buildroot buildtime similar as all the mycroft dependencies
0
592,328
17,875,612,426
IssuesEvent
2021-09-07 02:53:39
google/flax
https://api.github.com/repos/google/flax
closed
Rename `master` branch to `main`
Priority: P2 - medium
In line with https://github.com/github/renaming it'd be good to rename the default branch to "main".
1.0
Rename `master` branch to `main` - In line with https://github.com/github/renaming it'd be good to rename the default branch to "main".
priority
rename master branch to main in line with it d be good to rename the default branch to main
1
345,249
10,360,622,165
IssuesEvent
2019-09-06 08:04:42
grpc/grpc
https://api.github.com/repos/grpc/grpc
closed
NodeJS map is an object
disposition/stale kind/bug lang/node lang/node/protobuf.js priority/P2
Hello, when receiving a map<string,string> field the value is an Object. If I convert a map back and forth with protobufs I get a custom protobufjs [Map](https://github.com/dcodeIO/protobuf.js/blob/master/src/ProtoBuf/Map.js) object. Is this something I can control or where can I find the implementation for this?
1.0
NodeJS map is an object - Hello, when receiving a map<string,string> field the value is an Object. If I convert a map back and forth with protobufs I get a custom protobufjs [Map](https://github.com/dcodeIO/protobuf.js/blob/master/src/ProtoBuf/Map.js) object. Is this something I can control or where can I find the implementation for this?
priority
nodejs map is an object hello when receiving a map field the value is an object if i convert a map back and forth with protobufs i get a custom protobufjs object is this something i can control or where can i find the implementation for this
1
142,401
5,474,799,553
IssuesEvent
2017-03-11 03:34:17
CS2103JAN2017-T16-B2/main
https://api.github.com/repos/CS2103JAN2017-T16-B2/main
closed
As a user I want to have words that I am typing to be auto-completed by the system
priority.medium type.story
so that I can execute commands at a faster rate
1.0
As a user I want to have words that I am typing to be auto-completed by the system - so that I can execute commands at a faster rate
priority
as a user i want to have words that i am typing to be auto completed by the system so that i can execute commands at a faster rate
1
140,269
5,399,521,960
IssuesEvent
2017-02-27 19:39:49
Polymer/polymer-analyzer
https://api.github.com/repos/Polymer/polymer-analyzer
closed
Vanilla and Polymer 2 scanners should cooperate
Priority: High Status: Available Type: Bug
Right now scanning with both scanners would lead to two element features. The scanners instead should both contribute to a single element feature. We'll remove the PolymerElement feature and the Polymer scanner will add metadata as necessary.
1.0
Vanilla and Polymer 2 scanners should cooperate - Right now scanning with both scanners would lead to two element features. The scanners instead should both contribute to a single element feature. We'll remove the PolymerElement feature and the Polymer scanner will add metadata as necessary.
priority
vanilla and polymer scanners should cooperate right now scanning with both scanners would lead to two element features the scanners instead should both contribute to a single element feature we ll remove the polymerelement feature and the polymer scanner will add metadata as necessary
1
420,171
12,234,150,383
IssuesEvent
2020-05-04 12:57:52
CDH-Studio/UpSkill
https://api.github.com/repos/CDH-Studio/UpSkill
opened
Add descriptions to "Extra Info" icons on Edit profile components
Low Priority UI
- Extra Info icons at: - Employee Data -> Temporary Role - Language Proficiency -> I have been graded... - Talent -> Competencies, Skills, I am Available to Mentor, Mentorship Skills - Personal Growth Interests -> Developmental Goals, Willing to Relocate to Each icon should have a french/english (language dependant on toggle in nav bar) description about the component or feature it is next to.
1.0
Add descriptions to "Extra Info" icons on Edit profile components - - Extra Info icons at: - Employee Data -> Temporary Role - Language Proficiency -> I have been graded... - Talent -> Competencies, Skills, I am Available to Mentor, Mentorship Skills - Personal Growth Interests -> Developmental Goals, Willing to Relocate to Each icon should have a french/english (language dependant on toggle in nav bar) description about the component or feature it is next to.
priority
add descriptions to extra info icons on edit profile components extra info icons at employee data temporary role language proficiency i have been graded talent competencies skills i am available to mentor mentorship skills personal growth interests developmental goals willing to relocate to each icon should have a french english language dependant on toggle in nav bar description about the component or feature it is next to
1
251,513
8,016,363,408
IssuesEvent
2018-07-25 13:09:31
RagtagOpen/nomad
https://api.github.com/repos/RagtagOpen/nomad
closed
Hidden destination should also hide the carpools associated with the destination on the Find page
bug priority qa requested
The "Find a Ride" still displays carpools for destinations that are hidden. Steps: * create destination * create carpool using that destination * load "Find a Ride" page * note the destination is there * change destination to hidden * load "Find a Ride" page * note the destination is there Actual Result: A carpool with a hidden destination is still visible to a user (driver, rider) Expected Result: 1) A carpool with a hidden destination would not be visible on the "Find a Ride" page 2) Only carpools with visible destinations would be visible on the "Find a Ride" page
1.0
Hidden destination should also hide the carpools associated with the destination on the Find page - The "Find a Ride" still displays carpools for destinations that are hidden. Steps: * create destination * create carpool using that destination * load "Find a Ride" page * note the destination is there * change destination to hidden * load "Find a Ride" page * note the destination is there Actual Result: A carpool with a hidden destination is still visible to a user (driver, rider) Expected Result: 1) A carpool with a hidden destination would not be visible on the "Find a Ride" page 2) Only carpools with visible destinations would be visible on the "Find a Ride" page
priority
hidden destination should also hide the carpools associated with the destination on the find page the find a ride still displays carpools for destinations that are hidden steps create destination create carpool using that destination load find a ride page note the destination is there change destination to hidden load find a ride page note the destination is there actual result a carpool with a hidden destination is still visible to a user driver rider expected result a carpool with a hidden destination would not be visible on the find a ride page only carpools with visible destinations would be visible on the find a ride page
1
806,141
29,802,904,489
IssuesEvent
2023-06-16 09:26:33
OpenNebula/one
https://api.github.com/repos/OpenNebula/one
opened
Clean metrics in Prometheus collectors
Type: Bug Category: Drivers - Monitor Status: Accepted Priority: Normal
/!\ To report a **security issue** please follow this procedure: [https://github.com/OpenNebula/one/wiki/Vulnerability-Management-Process] **Description** Collectors use a client library for Prometheus based on a Ruby rack application. This application uses an internal store (synchronized datastore) that index metrics by label in a Hash object. There is no interface to clean entries on the Hash, so the information of past VMs are reported always till a restart of the collector. **To Reproduce** Create delete a VM, information is reported always **Expected behavior** Remove metrics for past VMs **Details** - Affected Component: Prometheus - Hypervisor: all (KVM specially) - Version: 6.6 <!--////////////////////////////////////////////--> <!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM --> <!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS --> <!-- PROGRESS WILL BE REFLECTED HERE --> <!--////////////////////////////////////////////--> ## Progress Status - [X] Code committed - [ ] Testing - QA - [ ] Documentation (Release notes - resolved issues, compatibility, known issues)
1.0
Clean metrics in Prometheus collectors - /!\ To report a **security issue** please follow this procedure: [https://github.com/OpenNebula/one/wiki/Vulnerability-Management-Process] **Description** Collectors use a client library for Prometheus based on a Ruby rack application. This application uses an internal store (synchronized datastore) that index metrics by label in a Hash object. There is no interface to clean entries on the Hash, so the information of past VMs are reported always till a restart of the collector. **To Reproduce** Create delete a VM, information is reported always **Expected behavior** Remove metrics for past VMs **Details** - Affected Component: Prometheus - Hypervisor: all (KVM specially) - Version: 6.6 <!--////////////////////////////////////////////--> <!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM --> <!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS --> <!-- PROGRESS WILL BE REFLECTED HERE --> <!--////////////////////////////////////////////--> ## Progress Status - [X] Code committed - [ ] Testing - QA - [ ] Documentation (Release notes - resolved issues, compatibility, known issues)
priority
clean metrics in prometheus collectors to report a security issue please follow this procedure description collectors use a client library for prometheus based on a ruby rack application this application uses an internal store synchronized datastore that index metrics by label in a hash object there is no interface to clean entries on the hash so the information of past vms are reported always till a restart of the collector to reproduce create delete a vm information is reported always expected behavior remove metrics for past vms details affected component prometheus hypervisor all kvm specially version progress status code committed testing qa documentation release notes resolved issues compatibility known issues
1
625,204
19,721,702,359
IssuesEvent
2022-01-13 15:56:59
Azure/azure-sdk-tools
https://api.github.com/repos/Azure/azure-sdk-tools
opened
[APIView] Debugger instance is not authenticating users
bug APIView Central-EngSys APIView Priority 1
APIView is not running properly as debugging instance using visual studio due to secure cookie change. Samesite mode is none when running as localhost to debug visual studio and this is causing issue in latest browsers. We should set samesite mode to Lax.
1.0
[APIView] Debugger instance is not authenticating users - APIView is not running properly as debugging instance using visual studio due to secure cookie change. Samesite mode is none when running as localhost to debug visual studio and this is causing issue in latest browsers. We should set samesite mode to Lax.
priority
debugger instance is not authenticating users apiview is not running properly as debugging instance using visual studio due to secure cookie change samesite mode is none when running as localhost to debug visual studio and this is causing issue in latest browsers we should set samesite mode to lax
1
219,724
7,345,463,796
IssuesEvent
2018-03-07 17:29:25
webpack-contrib/uglifyjs-webpack-plugin
https://api.github.com/repos/webpack-contrib/uglifyjs-webpack-plugin
closed
`extractComments` option uses url relative to `output.path` instead of its source file
priority: 4 (important) semver: Patch severity: 3 (broken) type: Bug
<!-- Please don't delete this template or we'll close your issue --> <!-- Before creating an issue please make sure you are using the latest version of webpack. --> <!-- Also consider trying the webpack@beta version, maybe it's already fixed. --> **Do you want to request a *feature* or report a *bug*?** I'm not sure, it's probably a FR <!-- Please ask questions on StackOverflow or the webpack Gitter (https://gitter.im/webpack/webpack). --> <!-- Issues which contain questions or support requests will be closed. --> **What is the current behavior?** When using the `UglifyjsWebpackPlugin` with the following options... ```js const uglifyConfig = { sourceMap: true, extractComments: /(?:^!|@(?:license|preserve))/i } ``` ...the file is written in the correct folder but the path that is written into the file is relative to the base directory (I guess it's the `output.path` option) instead of the file. As an example I'm getting: ``` /*! For license information please see chunks/polyfill-url-search-params/index.123hash.js.LICENSE */ ``` In production, this path is wrong because the relative directory to the `LICENSE` file, so it would be located in: `http://example.com/chunks/polyfill-url-search-params/chunks/polyfill-url-search-params/index.123hash.js.LICENSE` **If the current behavior is a bug, please provide the steps to reproduce.** <!-- A great way to do this is to provide your configuration via a GitHub gist. --> <!-- Best provide a minimal reproduceable repo --> <!-- If your issue is caused by a plugin or loader file the issue on the plugin/loader repo --> **What is the expected behavior?** I was expecting the path to be relative to the file it was extracted out of, in the same way the sourcemap comment is relative to it's base file. So I'd expect this comment: ``` /*! For license information please see index.9aa3a7a3fab51a749ce0.js.LICENSE */ ``` One option would be to provide the source file path so I could do a manual path rewrite with a function for the `filename` option. **If this is a feature request, what is motivation or use case for changing the behavior?** As all the files are hosted in the same directory (`file.123hash.js`, `file.123hash.js.map`, `file.123hash.js.LICENCE`) it's necessary that the url that is found within the file is also relative to the path it is found in. **Please mention other relevant information such as the browser version, Node.js version, webpack version and Operating System.** Version: webpack 3.8.1 Wepback config: ``` const config = { entry: { app: '/Repository-Root/some-cwd/src/app/main/index.js', vendor: ['vendor-libs'], }, output: { path: '/Repository-Root/some-cwd/dist', filename: `entry/[name]/index.js`, chunkFilename: `chunks/[name]/index.[chunkhash].js`, devtoolModuleFilenameTemplate: 'source-webpack:///[resourcePath]', devtoolFallbackModuleFilenameTemplate: 'source-webpack:///[resourcePath]?[hash]' }, plugins: [ new webpack.optimize.CommonsChunkPlugin({ name: 'vendor', minChunks: Infinity }), new webpack.optimize.CommonsChunkPlugin({ name: 'runtime' }), new webpack.optimize.UglifyJsPlugin(uglifyConfig) ], resolve: { modules: [ 'node_modules', '/Repository-Root/some-cwd/src/app' ], }, } ``` --- This issue was moved from webpack/webpack#6059 by @evilebottnawi. Orginal issue was by @Nirazul.
1.0
`extractComments` option uses url relative to `output.path` instead of its source file - <!-- Please don't delete this template or we'll close your issue --> <!-- Before creating an issue please make sure you are using the latest version of webpack. --> <!-- Also consider trying the webpack@beta version, maybe it's already fixed. --> **Do you want to request a *feature* or report a *bug*?** I'm not sure, it's probably a FR <!-- Please ask questions on StackOverflow or the webpack Gitter (https://gitter.im/webpack/webpack). --> <!-- Issues which contain questions or support requests will be closed. --> **What is the current behavior?** When using the `UglifyjsWebpackPlugin` with the following options... ```js const uglifyConfig = { sourceMap: true, extractComments: /(?:^!|@(?:license|preserve))/i } ``` ...the file is written in the correct folder but the path that is written into the file is relative to the base directory (I guess it's the `output.path` option) instead of the file. As an example I'm getting: ``` /*! For license information please see chunks/polyfill-url-search-params/index.123hash.js.LICENSE */ ``` In production, this path is wrong because the relative directory to the `LICENSE` file, so it would be located in: `http://example.com/chunks/polyfill-url-search-params/chunks/polyfill-url-search-params/index.123hash.js.LICENSE` **If the current behavior is a bug, please provide the steps to reproduce.** <!-- A great way to do this is to provide your configuration via a GitHub gist. --> <!-- Best provide a minimal reproduceable repo --> <!-- If your issue is caused by a plugin or loader file the issue on the plugin/loader repo --> **What is the expected behavior?** I was expecting the path to be relative to the file it was extracted out of, in the same way the sourcemap comment is relative to it's base file. So I'd expect this comment: ``` /*! For license information please see index.9aa3a7a3fab51a749ce0.js.LICENSE */ ``` One option would be to provide the source file path so I could do a manual path rewrite with a function for the `filename` option. **If this is a feature request, what is motivation or use case for changing the behavior?** As all the files are hosted in the same directory (`file.123hash.js`, `file.123hash.js.map`, `file.123hash.js.LICENCE`) it's necessary that the url that is found within the file is also relative to the path it is found in. **Please mention other relevant information such as the browser version, Node.js version, webpack version and Operating System.** Version: webpack 3.8.1 Wepback config: ``` const config = { entry: { app: '/Repository-Root/some-cwd/src/app/main/index.js', vendor: ['vendor-libs'], }, output: { path: '/Repository-Root/some-cwd/dist', filename: `entry/[name]/index.js`, chunkFilename: `chunks/[name]/index.[chunkhash].js`, devtoolModuleFilenameTemplate: 'source-webpack:///[resourcePath]', devtoolFallbackModuleFilenameTemplate: 'source-webpack:///[resourcePath]?[hash]' }, plugins: [ new webpack.optimize.CommonsChunkPlugin({ name: 'vendor', minChunks: Infinity }), new webpack.optimize.CommonsChunkPlugin({ name: 'runtime' }), new webpack.optimize.UglifyJsPlugin(uglifyConfig) ], resolve: { modules: [ 'node_modules', '/Repository-Root/some-cwd/src/app' ], }, } ``` --- This issue was moved from webpack/webpack#6059 by @evilebottnawi. Orginal issue was by @Nirazul.
priority
extractcomments option uses url relative to output path instead of its source file do you want to request a feature or report a bug i m not sure it s probably a fr what is the current behavior when using the uglifyjswebpackplugin with the following options js const uglifyconfig sourcemap true extractcomments license preserve i the file is written in the correct folder but the path that is written into the file is relative to the base directory i guess it s the output path option instead of the file as an example i m getting for license information please see chunks polyfill url search params index js license in production this path is wrong because the relative directory to the license file so it would be located in if the current behavior is a bug please provide the steps to reproduce what is the expected behavior i was expecting the path to be relative to the file it was extracted out of in the same way the sourcemap comment is relative to it s base file so i d expect this comment for license information please see index js license one option would be to provide the source file path so i could do a manual path rewrite with a function for the filename option if this is a feature request what is motivation or use case for changing the behavior as all the files are hosted in the same directory file js file js map file js licence it s necessary that the url that is found within the file is also relative to the path it is found in please mention other relevant information such as the browser version node js version webpack version and operating system version webpack wepback config const config entry app repository root some cwd src app main index js vendor output path repository root some cwd dist filename entry index js chunkfilename chunks index js devtoolmodulefilenametemplate source webpack devtoolfallbackmodulefilenametemplate source webpack plugins new webpack optimize commonschunkplugin name vendor minchunks infinity new webpack optimize commonschunkplugin name runtime new webpack optimize uglifyjsplugin uglifyconfig resolve modules node modules repository root some cwd src app this issue was moved from webpack webpack by evilebottnawi orginal issue was by nirazul
1
414,660
12,109,855,325
IssuesEvent
2020-04-21 09:26:52
solgenomics/sgn
https://api.github.com/repos/solgenomics/sgn
closed
cassavabase genotype queries
Priority: Critical Priority: High
Cassavabase is getting overwhelmed by genotype data query of ~ >300 clones. From the wizard it throws ‘502 Bad Gateway' and for solGS submissions, it silently dies. Memory is filling up and swap space maxed out!
2.0
cassavabase genotype queries - Cassavabase is getting overwhelmed by genotype data query of ~ >300 clones. From the wizard it throws ‘502 Bad Gateway' and for solGS submissions, it silently dies. Memory is filling up and swap space maxed out!
priority
cassavabase genotype queries cassavabase is getting overwhelmed by genotype data query of clones from the wizard it throws ‘ bad gateway and for solgs submissions it silently dies memory is filling up and swap space maxed out
1
73,572
14,105,058,400
IssuesEvent
2020-11-06 12:54:50
informalsystems/ibc-rs
https://api.github.com/repos/informalsystems/ibc-rs
opened
Clippy practices
code-hygiene good first issue
<!-- < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < ☺ v ✰ Thanks for opening an issue! ✰ v Before smashing the submit button please review the template. v Word of caution: poorly thought-out proposals may be rejected v without deliberation ☺ > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > --> ## Crate `ibc` ## Summary We have inconsistent practices regarding the use of clippy exceptions, in particular to `too_many_arguments` and `large_enum_variant`. We should decide on a standard. ## Problem Definition We sometimes annotate enums with `#[allow(clippy::large_enum_variant)]` to avoid clippy warnings: https://github.com/informalsystems/ibc-rs/blob/12cb1d6187e878eee0e40295fe0b804ddef0b846/modules/src/ics02_client/msgs.rs#L13-L18 And sometimes we box large enum variants for the same reason (to avoid clippy warnings): https://github.com/informalsystems/ibc-rs/blob/12cb1d6187e878eee0e40295fe0b804ddef0b846/modules/src/ics03_connection/msgs.rs#L29-L34 The `too_many_arguments` check also poses issues, for example this method has 11 arguments, but clippy issues warnings for anything larger than 7: https://github.com/informalsystems/ibc-rs/blob/12cb1d6187e878eee0e40295fe0b804ddef0b846/modules/src/ics07_tendermint/client_state.rs#L34-L47 ## Improvement proposal: - [ ] perhaps we can eliminate methods that have excessive number of arguments - [ ] use a repo-wide clippy configuration with larger thresholds than the defaults ## Original discussion I wonder if we can have a clippy [config to parametrize](https://rust-lang.github.io/rust-clippy/master/index.html#too_many_arguments) the "too_many_arguments" threshold (default is 7) for this crate, since we're using this _allow_ quite often. We are also cutting some corners with allowing `large_enum_variant`. _Originally posted by @adizere in https://github.com/informalsystems/ibc-rs/pull/355#discussion_r516654739_ In future, I think this deserves a discussion about clippy in general, if we want to have a repo-wide [configuration](https://github.com/rust-lang/rust-clippy#configuration) where we adjust the clippy parameters ... so we can remove the allow/clippy annotations altogether. I can open an issue to track this if that seems interesting. _Originally posted by @adizere in https://github.com/informalsystems/ibc-rs/pull/355#discussion_r516832986_ ____ #### For Admin Use - [X] Not duplicate issue - [X] Appropriate labels applied - [X] Appropriate milestone (priority) applied - [ ] Appropriate contributors tagged - [ ] Contributor assigned/self-assigned
1.0
Clippy practices - <!-- < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < ☺ v ✰ Thanks for opening an issue! ✰ v Before smashing the submit button please review the template. v Word of caution: poorly thought-out proposals may be rejected v without deliberation ☺ > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > --> ## Crate `ibc` ## Summary We have inconsistent practices regarding the use of clippy exceptions, in particular to `too_many_arguments` and `large_enum_variant`. We should decide on a standard. ## Problem Definition We sometimes annotate enums with `#[allow(clippy::large_enum_variant)]` to avoid clippy warnings: https://github.com/informalsystems/ibc-rs/blob/12cb1d6187e878eee0e40295fe0b804ddef0b846/modules/src/ics02_client/msgs.rs#L13-L18 And sometimes we box large enum variants for the same reason (to avoid clippy warnings): https://github.com/informalsystems/ibc-rs/blob/12cb1d6187e878eee0e40295fe0b804ddef0b846/modules/src/ics03_connection/msgs.rs#L29-L34 The `too_many_arguments` check also poses issues, for example this method has 11 arguments, but clippy issues warnings for anything larger than 7: https://github.com/informalsystems/ibc-rs/blob/12cb1d6187e878eee0e40295fe0b804ddef0b846/modules/src/ics07_tendermint/client_state.rs#L34-L47 ## Improvement proposal: - [ ] perhaps we can eliminate methods that have excessive number of arguments - [ ] use a repo-wide clippy configuration with larger thresholds than the defaults ## Original discussion I wonder if we can have a clippy [config to parametrize](https://rust-lang.github.io/rust-clippy/master/index.html#too_many_arguments) the "too_many_arguments" threshold (default is 7) for this crate, since we're using this _allow_ quite often. We are also cutting some corners with allowing `large_enum_variant`. _Originally posted by @adizere in https://github.com/informalsystems/ibc-rs/pull/355#discussion_r516654739_ In future, I think this deserves a discussion about clippy in general, if we want to have a repo-wide [configuration](https://github.com/rust-lang/rust-clippy#configuration) where we adjust the clippy parameters ... so we can remove the allow/clippy annotations altogether. I can open an issue to track this if that seems interesting. _Originally posted by @adizere in https://github.com/informalsystems/ibc-rs/pull/355#discussion_r516832986_ ____ #### For Admin Use - [X] Not duplicate issue - [X] Appropriate labels applied - [X] Appropriate milestone (priority) applied - [ ] Appropriate contributors tagged - [ ] Contributor assigned/self-assigned
non_priority
clippy practices ☺ v ✰ thanks for opening an issue ✰ v before smashing the submit button please review the template v word of caution poorly thought out proposals may be rejected v without deliberation ☺ crate ibc summary we have inconsistent practices regarding the use of clippy exceptions in particular to too many arguments and large enum variant we should decide on a standard problem definition we sometimes annotate enums with to avoid clippy warnings and sometimes we box large enum variants for the same reason to avoid clippy warnings the too many arguments check also poses issues for example this method has arguments but clippy issues warnings for anything larger than improvement proposal perhaps we can eliminate methods that have excessive number of arguments use a repo wide clippy configuration with larger thresholds than the defaults original discussion i wonder if we can have a clippy the too many arguments threshold default is for this crate since we re using this allow quite often we are also cutting some corners with allowing large enum variant originally posted by adizere in in future i think this deserves a discussion about clippy in general if we want to have a repo wide where we adjust the clippy parameters so we can remove the allow clippy annotations altogether i can open an issue to track this if that seems interesting originally posted by adizere in for admin use not duplicate issue appropriate labels applied appropriate milestone priority applied appropriate contributors tagged contributor assigned self assigned
0
248,489
7,931,775,465
IssuesEvent
2018-07-07 05:04:42
wso2/testgrid
https://api.github.com/repos/wso2/testgrid
opened
Testgrid distribution should have Jenkins and tinkerer built-in
Priority/High Type/Improvement
**Description:** ATM, we provide the testgrid core libraries and its runtime engine, jenkins, separately. This has lead to some confusion where unnecessary internal information has been exposed to first-time users. For example, users should not have to run generate-testplan and then iterate run-testplan. This is actually be part of the testgrid pipeline script, and the users should be executing that instead. Secondly, there's a manual process involved to configure the tinkerer webapp. We need to analyze the effort required to automate the tinkerer configuration, and the ability to run it locally. Right now, the tinkerer agents cannot connect to tinkerer webapp when we run testgrid locally with aws deployments. **Affected Product Version:** m34 **OS, DB, other environment details and versions:** local testgrid + local tinkerer webapp + aws product deployment **Steps to reproduce:** **Related Issues:** <!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. -->
1.0
Testgrid distribution should have Jenkins and tinkerer built-in - **Description:** ATM, we provide the testgrid core libraries and its runtime engine, jenkins, separately. This has lead to some confusion where unnecessary internal information has been exposed to first-time users. For example, users should not have to run generate-testplan and then iterate run-testplan. This is actually be part of the testgrid pipeline script, and the users should be executing that instead. Secondly, there's a manual process involved to configure the tinkerer webapp. We need to analyze the effort required to automate the tinkerer configuration, and the ability to run it locally. Right now, the tinkerer agents cannot connect to tinkerer webapp when we run testgrid locally with aws deployments. **Affected Product Version:** m34 **OS, DB, other environment details and versions:** local testgrid + local tinkerer webapp + aws product deployment **Steps to reproduce:** **Related Issues:** <!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. -->
priority
testgrid distribution should have jenkins and tinkerer built in description atm we provide the testgrid core libraries and its runtime engine jenkins separately this has lead to some confusion where unnecessary internal information has been exposed to first time users for example users should not have to run generate testplan and then iterate run testplan this is actually be part of the testgrid pipeline script and the users should be executing that instead secondly there s a manual process involved to configure the tinkerer webapp we need to analyze the effort required to automate the tinkerer configuration and the ability to run it locally right now the tinkerer agents cannot connect to tinkerer webapp when we run testgrid locally with aws deployments affected product version os db other environment details and versions local testgrid local tinkerer webapp aws product deployment steps to reproduce related issues
1
190,465
15,239,308,704
IssuesEvent
2021-02-19 04:08:29
chef/chef-workstation
https://api.github.com/repos/chef/chef-workstation
closed
CONTENT: Our getting start page includes many one time steps
Aspect: Documentation Documentation
Our getting started docs tell folks to generate a new chef-repo and create a new org on their chef server as part of the Workstation getting started process. This is only something an organization does once and it potentially shouldn't even be on this page. At the very least we need to make sure to note that you only do this if it hasn't been done already. If you already have one you should be cloning it to your system. https://docs.chef.io/workstation/getting_started/#create-an-organization https://docs.chef.io/workstation/getting_started/#create-an-organization
2.0
CONTENT: Our getting start page includes many one time steps - Our getting started docs tell folks to generate a new chef-repo and create a new org on their chef server as part of the Workstation getting started process. This is only something an organization does once and it potentially shouldn't even be on this page. At the very least we need to make sure to note that you only do this if it hasn't been done already. If you already have one you should be cloning it to your system. https://docs.chef.io/workstation/getting_started/#create-an-organization https://docs.chef.io/workstation/getting_started/#create-an-organization
non_priority
content our getting start page includes many one time steps our getting started docs tell folks to generate a new chef repo and create a new org on their chef server as part of the workstation getting started process this is only something an organization does once and it potentially shouldn t even be on this page at the very least we need to make sure to note that you only do this if it hasn t been done already if you already have one you should be cloning it to your system
0
714,248
24,555,504,663
IssuesEvent
2022-10-12 15:34:01
brave/brave-browser
https://api.github.com/repos/brave/brave-browser
closed
Update `FrequencyCapping` study defaults to match the Griffin seed
enhancement priority/P3 needs-discussion dev-concern feature/ads OS/Desktop
``` { "name": "BraveAds.FrequencyCappingStudy", "experiments": [ { "name": "ExcludeAdIfWithinTimeWindow=0h", "probability_weight": 100, "parameters": [ { "name": "exclude_ad_if_dismissed_within_time_window", "value": "0h" }, { "name": "exclude_ad_if_transferred_within_time_window", "value": "0h" } ], "feature_association": { "enable_feature": ["FrequencyCapping"] } }, { "name": "Default", "probability_weight": 0 } ], "filter": { "min_version": "91.1.26.60", "channel": ["NIGHTLY", "BETA", "RELEASE"], "platform": ["WINDOWS", "MAC", "LINUX", "ANDROID"], "country": ["US", "AU", "CA", "FR", "DE", "IE", "JP", "NZ", "GB"] } }, ```
1.0
Update `FrequencyCapping` study defaults to match the Griffin seed - ``` { "name": "BraveAds.FrequencyCappingStudy", "experiments": [ { "name": "ExcludeAdIfWithinTimeWindow=0h", "probability_weight": 100, "parameters": [ { "name": "exclude_ad_if_dismissed_within_time_window", "value": "0h" }, { "name": "exclude_ad_if_transferred_within_time_window", "value": "0h" } ], "feature_association": { "enable_feature": ["FrequencyCapping"] } }, { "name": "Default", "probability_weight": 0 } ], "filter": { "min_version": "91.1.26.60", "channel": ["NIGHTLY", "BETA", "RELEASE"], "platform": ["WINDOWS", "MAC", "LINUX", "ANDROID"], "country": ["US", "AU", "CA", "FR", "DE", "IE", "JP", "NZ", "GB"] } }, ```
priority
update frequencycapping study defaults to match the griffin seed name braveads frequencycappingstudy experiments name excludeadifwithintimewindow probability weight parameters name exclude ad if dismissed within time window value name exclude ad if transferred within time window value feature association enable feature name default probability weight filter min version channel platform country
1
9,081
2,607,906,090
IssuesEvent
2015-02-26 00:15:41
chrsmithdemos/zen-coding
https://api.github.com/repos/chrsmithdemos/zen-coding
opened
is there a plan for jEdit plugin?
auto-migrated Priority-Medium Type-EditorSupport
``` It would be great if a Zencoding was available for jEdit. I know there is a port of 0.5.1 version but couldn't get it to work. The coolest thing is that this way I could get Zencoding for a crossplatform really customizable editor :) ``` ----- Original issue reported on code.google.com by `cyryl...@gmail.com` on 30 Oct 2010 at 2:22
1.0
is there a plan for jEdit plugin? - ``` It would be great if a Zencoding was available for jEdit. I know there is a port of 0.5.1 version but couldn't get it to work. The coolest thing is that this way I could get Zencoding for a crossplatform really customizable editor :) ``` ----- Original issue reported on code.google.com by `cyryl...@gmail.com` on 30 Oct 2010 at 2:22
priority
is there a plan for jedit plugin it would be great if a zencoding was available for jedit i know there is a port of version but couldn t get it to work the coolest thing is that this way i could get zencoding for a crossplatform really customizable editor original issue reported on code google com by cyryl gmail com on oct at
1
66,380
14,768,188,715
IssuesEvent
2021-01-10 10:50:57
YauheniPo/Elements_Test_Framework
https://api.github.com/repos/YauheniPo/Elements_Test_Framework
opened
CVE-2020-9548 (High) detected in jackson-databind-2.9.8.jar
security vulnerability
## CVE-2020-9548 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.8.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: Elements_Test_Framework/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar</p> <p> Dependency Hierarchy: - allure-testng-2.10.0.jar (Root Library) - allure-java-commons-2.10.0.jar - :x: **jackson-databind-2.9.8.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/YauheniPo/Elements_Test_Framework/commit/b6525683a2173ae823218926b6f8d80d10c5d61f">b6525683a2173ae823218926b6f8d80d10c5d61f</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to br.com.anteros.dbcp.AnterosDBCPConfig (aka anteros-core). <p>Publish Date: 2020-03-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9548>CVE-2020-9548</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-9548">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-9548</a></p> <p>Release Date: 2020-03-02</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.7.9.7,2.8.11.6,2.9.10.4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-9548 (High) detected in jackson-databind-2.9.8.jar - ## CVE-2020-9548 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.8.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: Elements_Test_Framework/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar</p> <p> Dependency Hierarchy: - allure-testng-2.10.0.jar (Root Library) - allure-java-commons-2.10.0.jar - :x: **jackson-databind-2.9.8.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/YauheniPo/Elements_Test_Framework/commit/b6525683a2173ae823218926b6f8d80d10c5d61f">b6525683a2173ae823218926b6f8d80d10c5d61f</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to br.com.anteros.dbcp.AnterosDBCPConfig (aka anteros-core). <p>Publish Date: 2020-03-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9548>CVE-2020-9548</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-9548">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-9548</a></p> <p>Release Date: 2020-03-02</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.7.9.7,2.8.11.6,2.9.10.4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file elements test framework pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy allure testng jar root library allure java commons jar x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to br com anteros dbcp anterosdbcpconfig aka anteros core publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind step up your open source security game with whitesource
0
155,460
13,625,645,449
IssuesEvent
2020-09-24 09:48:12
drafthub/drafthub
https://api.github.com/repos/drafthub/drafthub
closed
missing docstring for `core.views`
documentation good first issue help wanted
Here are some missings docstring reported by pylint. ``` $ docker-compose exec web python check.py lint | grep docstring | grep core | grep views drafthub/core/views.py:1:0: C0114: Missing module docstring (missing-module-docstring) drafthub/core/views.py:15:0: C0115: Missing class docstring (missing-class-docstring) drafthub/core/views.py:25:0: C0115: Missing class docstring (missing-class-docstring) drafthub/core/views.py:46:0: C0115: Missing class docstring (missing-class-docstring) drafthub/core/views.py:59:0: C0115: Missing class docstring (missing-class-docstring) drafthub/core/views.py:64:0: C0115: Missing class docstring (missing-class-docstring) drafthub/core/views.py:89:4: C0116: Missing function or method docstring (missing-function-docstring) drafthub/core/views.py:110:4: C0116: Missing function or method docstring (missing-function-docstring) drafthub/core/views.py:223:0: C0115: Missing class docstring (missing-class-docstring) ``` See how you can contribute: [CONTRIBUTING.md](https://github.com/drafthub/drafthub/blob/master/CONTRIBUTING.md)
1.0
missing docstring for `core.views` - Here are some missings docstring reported by pylint. ``` $ docker-compose exec web python check.py lint | grep docstring | grep core | grep views drafthub/core/views.py:1:0: C0114: Missing module docstring (missing-module-docstring) drafthub/core/views.py:15:0: C0115: Missing class docstring (missing-class-docstring) drafthub/core/views.py:25:0: C0115: Missing class docstring (missing-class-docstring) drafthub/core/views.py:46:0: C0115: Missing class docstring (missing-class-docstring) drafthub/core/views.py:59:0: C0115: Missing class docstring (missing-class-docstring) drafthub/core/views.py:64:0: C0115: Missing class docstring (missing-class-docstring) drafthub/core/views.py:89:4: C0116: Missing function or method docstring (missing-function-docstring) drafthub/core/views.py:110:4: C0116: Missing function or method docstring (missing-function-docstring) drafthub/core/views.py:223:0: C0115: Missing class docstring (missing-class-docstring) ``` See how you can contribute: [CONTRIBUTING.md](https://github.com/drafthub/drafthub/blob/master/CONTRIBUTING.md)
non_priority
missing docstring for core views here are some missings docstring reported by pylint docker compose exec web python check py lint grep docstring grep core grep views drafthub core views py missing module docstring missing module docstring drafthub core views py missing class docstring missing class docstring drafthub core views py missing class docstring missing class docstring drafthub core views py missing class docstring missing class docstring drafthub core views py missing class docstring missing class docstring drafthub core views py missing class docstring missing class docstring drafthub core views py missing function or method docstring missing function docstring drafthub core views py missing function or method docstring missing function docstring drafthub core views py missing class docstring missing class docstring see how you can contribute
0
508,509
14,701,612,752
IssuesEvent
2021-01-04 12:11:52
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
t.bilibili.com - desktop site instead of mobile site
browser-firefox-mobile engine-gecko priority-critical
<!-- @browser: Firefox Mobile 81.0 --> <!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:81.0) Gecko/81.0 Firefox/81.0 --> <!-- @reported_with: desktop-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/64749 --> **URL**: https://t.bilibili.com/?tab=512,4097,4098,4099,4100,4101 **Browser / Version**: Firefox Mobile 81.0 **Operating System**: Android **Tested Another Browser**: Yes Chrome **Problem type**: Desktop site instead of mobile site **Description**: Desktop site instead of mobile site **Steps to Reproduce**: <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200804091327</li><li>channel: nightly</li><li>hasTouchScreen: true</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
t.bilibili.com - desktop site instead of mobile site - <!-- @browser: Firefox Mobile 81.0 --> <!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:81.0) Gecko/81.0 Firefox/81.0 --> <!-- @reported_with: desktop-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/64749 --> **URL**: https://t.bilibili.com/?tab=512,4097,4098,4099,4100,4101 **Browser / Version**: Firefox Mobile 81.0 **Operating System**: Android **Tested Another Browser**: Yes Chrome **Problem type**: Desktop site instead of mobile site **Description**: Desktop site instead of mobile site **Steps to Reproduce**: <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200804091327</li><li>channel: nightly</li><li>hasTouchScreen: true</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
priority
t bilibili com desktop site instead of mobile site url browser version firefox mobile operating system android tested another browser yes chrome problem type desktop site instead of mobile site description desktop site instead of mobile site steps to reproduce browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel nightly hastouchscreen true from with ❤️
1
182,190
14,108,256,906
IssuesEvent
2020-11-06 17:31:14
cyberark/conjur-api-python3
https://api.github.com/repos/cyberark/conjur-api-python3
opened
Builds of master that are also a tag pointer fail in subsequent builds
component/conjur-sdk good first issue good-first-issue kind/bug kind/failing-test priority/low triage/needs-info
Because we publish tagged commits on main branch and nightly builds try to rebuild the same commit each night, failures in publishing cause problems for builds following the first one. AC: - [ ] Subsequent builds of tagged commits on master do not break the build - [ ] Failures in publishing still fail the first publishing attempt ## Steps to Reproduce Steps to reproduce the behavior: 1. Rebuild a tagged build on master ## Expected Results Success ## Actual Results (including error logs, if applicable) [Build failure](https://jenkins.conjur.net/job/cyberark--conjur-api-python3/job/master/) ## Reproducible * [x] Always * [ ] Sometimes * [ ] Non-Reproducible ## Version/Tag number v0.1.1 ## Environment setup Jenkins ## Additional Information N/A
1.0
Builds of master that are also a tag pointer fail in subsequent builds - Because we publish tagged commits on main branch and nightly builds try to rebuild the same commit each night, failures in publishing cause problems for builds following the first one. AC: - [ ] Subsequent builds of tagged commits on master do not break the build - [ ] Failures in publishing still fail the first publishing attempt ## Steps to Reproduce Steps to reproduce the behavior: 1. Rebuild a tagged build on master ## Expected Results Success ## Actual Results (including error logs, if applicable) [Build failure](https://jenkins.conjur.net/job/cyberark--conjur-api-python3/job/master/) ## Reproducible * [x] Always * [ ] Sometimes * [ ] Non-Reproducible ## Version/Tag number v0.1.1 ## Environment setup Jenkins ## Additional Information N/A
non_priority
builds of master that are also a tag pointer fail in subsequent builds because we publish tagged commits on main branch and nightly builds try to rebuild the same commit each night failures in publishing cause problems for builds following the first one ac subsequent builds of tagged commits on master do not break the build failures in publishing still fail the first publishing attempt steps to reproduce steps to reproduce the behavior rebuild a tagged build on master expected results success actual results including error logs if applicable reproducible always sometimes non reproducible version tag number environment setup jenkins additional information n a
0
111,466
4,473,298,223
IssuesEvent
2016-08-26 03:01:15
inf3rno/e3
https://api.github.com/repos/inf3rno/e3
closed
Implement lazy loading by non-v8 and unsupported adapters
bug priority:normal
Lazy loading was removed by removing the defineLoaders utility function. There was some bug related to it, but I wasn't able to figure out what it was. Maybe the lib was just unstable, and that caused the problem.
1.0
Implement lazy loading by non-v8 and unsupported adapters - Lazy loading was removed by removing the defineLoaders utility function. There was some bug related to it, but I wasn't able to figure out what it was. Maybe the lib was just unstable, and that caused the problem.
priority
implement lazy loading by non and unsupported adapters lazy loading was removed by removing the defineloaders utility function there was some bug related to it but i wasn t able to figure out what it was maybe the lib was just unstable and that caused the problem
1
213,848
7,260,889,496
IssuesEvent
2018-02-18 15:18:20
facelessuser/Rummage
https://api.github.com/repos/facelessuser/Rummage
closed
File pattern tweaks
Enhancement Priority - Medium
- File patterns, if nothing is assigned should be get the default `*` or `.*` for wildcard and regex respectively. Folder excludes will continue to ignore an empty input. - Currently wildcard patterns starting with `-` are kind of exceptions, but can't be used by themselves. This can be a little confusing. To fix this, for file patterns, if we have one or more `-` patterns, but no normal pattern, the normal pattern will be the default `*` pattern, so the `-` pattern will be applied to that giving what a user would expect, which is all files except whatever comes after `-`. For excludes, it will be the same same way, if only `-` patterns are specified, all folders will be excluded except what comes after `-`.
1.0
File pattern tweaks - - File patterns, if nothing is assigned should be get the default `*` or `.*` for wildcard and regex respectively. Folder excludes will continue to ignore an empty input. - Currently wildcard patterns starting with `-` are kind of exceptions, but can't be used by themselves. This can be a little confusing. To fix this, for file patterns, if we have one or more `-` patterns, but no normal pattern, the normal pattern will be the default `*` pattern, so the `-` pattern will be applied to that giving what a user would expect, which is all files except whatever comes after `-`. For excludes, it will be the same same way, if only `-` patterns are specified, all folders will be excluded except what comes after `-`.
priority
file pattern tweaks file patterns if nothing is assigned should be get the default or for wildcard and regex respectively folder excludes will continue to ignore an empty input currently wildcard patterns starting with are kind of exceptions but can t be used by themselves this can be a little confusing to fix this for file patterns if we have one or more patterns but no normal pattern the normal pattern will be the default pattern so the pattern will be applied to that giving what a user would expect which is all files except whatever comes after for excludes it will be the same same way if only patterns are specified all folders will be excluded except what comes after
1
43,563
13,020,407,872
IssuesEvent
2020-07-27 02:59:24
LightC0der/arunbhandari.github.io
https://api.github.com/repos/LightC0der/arunbhandari.github.io
opened
CVE-2020-11022 (Medium) detected in jquery-1.9.1.min.js
security vulnerability
## CVE-2020-11022 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.9.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.9.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.9.1/jquery.min.js</a></p> <p>Path to vulnerable library: /arunbhandari.github.io/assets/js/vendor/jquery-1.9.1.min.js</p> <p> Dependency Hierarchy: - :x: **jquery-1.9.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/LightC0der/arunbhandari.github.io/commit/241096b6dd14739925eca764bd8ab9a25a8003c6">241096b6dd14739925eca764bd8ab9a25a8003c6</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022>CVE-2020-11022</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/">https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jQuery - 3.5.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-11022 (Medium) detected in jquery-1.9.1.min.js - ## CVE-2020-11022 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.9.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.9.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.9.1/jquery.min.js</a></p> <p>Path to vulnerable library: /arunbhandari.github.io/assets/js/vendor/jquery-1.9.1.min.js</p> <p> Dependency Hierarchy: - :x: **jquery-1.9.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/LightC0der/arunbhandari.github.io/commit/241096b6dd14739925eca764bd8ab9a25a8003c6">241096b6dd14739925eca764bd8ab9a25a8003c6</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022>CVE-2020-11022</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/">https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jQuery - 3.5.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve medium detected in jquery min js cve medium severity vulnerability vulnerable library jquery min js javascript library for dom operations library home page a href path to vulnerable library arunbhandari github io assets js vendor jquery min js dependency hierarchy x jquery min js vulnerable library found in head commit a href vulnerability details in jquery versions greater than or equal to and before passing html from untrusted sources even after sanitizing it to one of jquery s dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with whitesource
0
285,436
24,666,334,224
IssuesEvent
2022-10-18 10:33:39
hzi-braunschweig/SORMAS-Project
https://api.github.com/repos/hzi-braunschweig/SORMAS-Project
closed
Implement environment default language into env data json and automation framework
testing task e2e-tests
The aim to have a defined default language field for each environment would have the following benefits: 1. Set rollback language for tests marked as @LanguageRisk 2. Create UI Users without hardcoded data for language 3. Getting environment expected default language without checking current language content 4. Easier to understand what is the expected language of an environment if documentation is not in place. Add new filed into json environments file Update framework to read and map data.
2.0
Implement environment default language into env data json and automation framework - The aim to have a defined default language field for each environment would have the following benefits: 1. Set rollback language for tests marked as @LanguageRisk 2. Create UI Users without hardcoded data for language 3. Getting environment expected default language without checking current language content 4. Easier to understand what is the expected language of an environment if documentation is not in place. Add new filed into json environments file Update framework to read and map data.
non_priority
implement environment default language into env data json and automation framework the aim to have a defined default language field for each environment would have the following benefits set rollback language for tests marked as languagerisk create ui users without hardcoded data for language getting environment expected default language without checking current language content easier to understand what is the expected language of an environment if documentation is not in place add new filed into json environments file update framework to read and map data
0
130,104
10,596,627,527
IssuesEvent
2019-10-09 21:44:36
rancher/rancher
https://api.github.com/repos/rancher/rancher
opened
flaky test: test_disable_template_revision
kind/bug-qa kind/flaky-test team/ca
Drone failure seen: https://drone-pr.rancher.io/rancher/rancher/4710/1/2 ``` ==================================== ERRORS ==================================== 754 _____________ ERROR at teardown of test_disable_template_revision ______________ 755 [gw1] linux -- Python 3.7.3 /go/src/github.com/rancher/rancher/tests/integration/.tox/py37/bin/python3.7 756 757 tp = <class 'rancher.ApiError'>, value = None, tb = None 758 759 def reraise(tp, value, tb=None): 760 try: 761 if value is None: 762 value = tp() 763 if value.__traceback__ is not tb: 764 raise value.with_traceback(tb) 765 > raise value 766 767 ../.tox/py37/lib/python3.7/site-packages/six.py:693: self = <rancher.Client object at 0x7faa7881cef0> 786 text = '{"baseType":"error","code":"InvalidAction","message":"Cannot delete the clusterTemplate until Clusters referring it are removed","status":422,"type":"err or"}\n' 787 788 def _error(self, text): 789 > raise ApiError(self._unmarshall(text)) 790 E rancher.ApiError: (ApiError(...), "InvalidAction : Cannot delete the clusterTemplate until Clusters referring it are removed\n\t{'baseType': 'error', 'co de': 'InvalidAction', 'message': 'Cannot delete the clusterTemplate until Clusters referring it are removed', 'status': 422, 'type': 'error'}") ```
1.0
flaky test: test_disable_template_revision - Drone failure seen: https://drone-pr.rancher.io/rancher/rancher/4710/1/2 ``` ==================================== ERRORS ==================================== 754 _____________ ERROR at teardown of test_disable_template_revision ______________ 755 [gw1] linux -- Python 3.7.3 /go/src/github.com/rancher/rancher/tests/integration/.tox/py37/bin/python3.7 756 757 tp = <class 'rancher.ApiError'>, value = None, tb = None 758 759 def reraise(tp, value, tb=None): 760 try: 761 if value is None: 762 value = tp() 763 if value.__traceback__ is not tb: 764 raise value.with_traceback(tb) 765 > raise value 766 767 ../.tox/py37/lib/python3.7/site-packages/six.py:693: self = <rancher.Client object at 0x7faa7881cef0> 786 text = '{"baseType":"error","code":"InvalidAction","message":"Cannot delete the clusterTemplate until Clusters referring it are removed","status":422,"type":"err or"}\n' 787 788 def _error(self, text): 789 > raise ApiError(self._unmarshall(text)) 790 E rancher.ApiError: (ApiError(...), "InvalidAction : Cannot delete the clusterTemplate until Clusters referring it are removed\n\t{'baseType': 'error', 'co de': 'InvalidAction', 'message': 'Cannot delete the clusterTemplate until Clusters referring it are removed', 'status': 422, 'type': 'error'}") ```
non_priority
flaky test test disable template revision drone failure seen errors error at teardown of test disable template revision linux python go src github com rancher rancher tests integration tox bin tp value none tb none def reraise tp value tb none try if value is none value tp if value traceback is not tb raise value with traceback tb raise value tox lib site packages six py self text basetype error code invalidaction message cannot delete the clustertemplate until clusters referring it are removed status type err or n def error self text raise apierror self unmarshall text e rancher apierror apierror invalidaction cannot delete the clustertemplate until clusters referring it are removed n t basetype error co de invalidaction message cannot delete the clustertemplate until clusters referring it are removed status type error
0
301,523
9,221,300,488
IssuesEvent
2019-03-11 19:37:29
infor-design/enterprise
https://api.github.com/repos/infor-design/enterprise
closed
Locale: Add support to set language and locale independently
[5] priority: high type: enhancement :sparkles:
**Background** A customer requirement is the ability to separate locale formatting from language. Ming.le provides these values as separate parameters, and we need the IDS components to support the same. The language and locale need to be set on a high level, when initializing the IDS Locale object, and not as options when initializing specific components. Example: Use English translations for all IDS components, but Swedish formatting (dates, numbers, etc.) **Problem description** There is currently no way to set language and locale to different values when initializing the IDS Locale object, all formatting is done based on the set language. **Proposed solution** Make it possible to set language and locale independently of each other. No code change should be required for Homepages widgets / content that consume the actual IDS components, only for the IDS Locale initialization.
1.0
Locale: Add support to set language and locale independently - **Background** A customer requirement is the ability to separate locale formatting from language. Ming.le provides these values as separate parameters, and we need the IDS components to support the same. The language and locale need to be set on a high level, when initializing the IDS Locale object, and not as options when initializing specific components. Example: Use English translations for all IDS components, but Swedish formatting (dates, numbers, etc.) **Problem description** There is currently no way to set language and locale to different values when initializing the IDS Locale object, all formatting is done based on the set language. **Proposed solution** Make it possible to set language and locale independently of each other. No code change should be required for Homepages widgets / content that consume the actual IDS components, only for the IDS Locale initialization.
priority
locale add support to set language and locale independently background a customer requirement is the ability to separate locale formatting from language ming le provides these values as separate parameters and we need the ids components to support the same the language and locale need to be set on a high level when initializing the ids locale object and not as options when initializing specific components example use english translations for all ids components but swedish formatting dates numbers etc problem description there is currently no way to set language and locale to different values when initializing the ids locale object all formatting is done based on the set language proposed solution make it possible to set language and locale independently of each other no code change should be required for homepages widgets content that consume the actual ids components only for the ids locale initialization
1
29,018
8,250,413,780
IssuesEvent
2018-09-12 02:37:35
godotengine/godot
https://api.github.com/repos/godotengine/godot
closed
Build fails if sub-directory under 'modules/mono/glue/cs_files' contains special characters
bug topic:buildsystem topic:mono
**Godot version:** <!-- Specify commit hash if non-official. --> `master` / c93888ae **OS/device including version:** Manjaro Linux 17.1 **Issue description:** The current master branch fails to build due to compilation errors in _cs_compressed.gen.h_ as shown below: ``` In file included from modules/mono/editor/bindings_generator.cpp:43: modules/mono/editor/../glue/cs_compressed.gen.h:132:18: error: default initialization of an object of const type 'const int' static const int _cs_Godot API_Properties_AssemblyInfo_compressed_size = 627; ^ = 0 modules/mono/editor/../glue/cs_compressed.gen.h:132:27: error: expected ';' after top level declarator static const int _cs_Godot API_Properties_AssemblyInfo_compressed_size = 627; ^ ; modules/mono/editor/../glue/cs_compressed.gen.h:133:18: error: redefinition of '_cs_Godot' static const int _cs_Godot API_Properties_AssemblyInfo_uncompressed_size = 1355; ^ modules/mono/editor/../glue/cs_compressed.gen.h:132:18: note: previous definition is here static const int _cs_Godot API_Properties_AssemblyInfo_compressed_size = 627; ^ modules/mono/editor/../glue/cs_compressed.gen.h:133:27: error: expected ';' after top level declarator static const int _cs_Godot API_Properties_AssemblyInfo_uncompressed_size = 1355; ^ ; modules/mono/editor/../glue/cs_compressed.gen.h:134:28: error: redefinition of '_cs_Godot' with a different type: 'const unsigned char' vs 'const int' static const unsigned char _cs_Godot API_Properties_AssemblyInfo_compressed[] = { 120, 156, 133, 84, 75, 110, 26, 65, 16, 221, 115, 138, 18, 27, 19, 43, 96, 76, 140, 228, 207, 10, ... ^ modules/mono/editor/../glue/cs_compressed.gen.h:132:18: note: previous definition is here static const int _cs_Godot API_Properties_AssemblyInfo_compressed_size = 627; ^ modules/mono/editor/../glue/cs_compressed.gen.h:134:37: error: expected ';' after top level declarator static const unsigned char _cs_Godot API_Properties_AssemblyInfo_compressed[] = { 120, 156, 133, 84, 75, 110, 26, 65, 16, 221, 115, 138, 18, 27, 19, 43, 96, 76, 140, 228, 207, 10, ... ^ ; modules/mono/editor/../glue/cs_compressed.gen.h:187:82: error: expected ')' r_files.insert("Godot API/Properties/AssemblyInfo.cs", CompressedFile(_cs_Godot API_Properties_AssemblyInfo_compressed_size, _cs_Godot API_Properties_AssemblyInfo_uncompres... ^ modules/mono/editor/../glue/cs_compressed.gen.h:187:71: note: to match this '(' r_files.insert("Godot API/Properties/AssemblyInfo.cs", CompressedFile(_cs_Godot API_Properties_AssemblyInfo_compressed_size, _cs_Godot API_Properties_AssemblyInfo_uncompres... ^ modules/mono/editor/../glue/cs_compressed.gen.h:187:57: error: no matching conversion for functional-style cast from 'const int' to 'CompressedFile' r_files.insert("Godot API/Properties/AssemblyInfo.cs", CompressedFile(_cs_Godot API_Properties_AssemblyInfo_compressed_size, _cs_Godot API_Properties_AssemblyInfo_uncompres... ^~~~~~~~~~~~~~~~~~~~~~~~ modules/mono/editor/../glue/cs_compressed.gen.h:138:8: note: candidate constructor (the implicit copy constructor) not viable: no known conversion from 'const int' to 'const CompressedFile' for 1st argument struct CompressedFile ^ modules/mono/editor/../glue/cs_compressed.gen.h:138:8: note: candidate constructor (the implicit move constructor) not viable: no known conversion from 'const int' to 'CompressedFile' for 1st argument struct CompressedFile ^ modules/mono/editor/../glue/cs_compressed.gen.h:151:2: note: candidate constructor not viable: requires 0 arguments, but 1 was provided CompressedFile() {} ^ modules/mono/editor/../glue/cs_compressed.gen.h:144:2: note: candidate constructor not viable: requires 3 arguments, but 1 was provided CompressedFile(int p_comp_size, int p_uncomp_size, const unsigned char* p_data) ^ [ 29%] 8 errors generated. ``` (@neikeq Could it be possible that it was caused by 72873004332b?)
1.0
Build fails if sub-directory under 'modules/mono/glue/cs_files' contains special characters - **Godot version:** <!-- Specify commit hash if non-official. --> `master` / c93888ae **OS/device including version:** Manjaro Linux 17.1 **Issue description:** The current master branch fails to build due to compilation errors in _cs_compressed.gen.h_ as shown below: ``` In file included from modules/mono/editor/bindings_generator.cpp:43: modules/mono/editor/../glue/cs_compressed.gen.h:132:18: error: default initialization of an object of const type 'const int' static const int _cs_Godot API_Properties_AssemblyInfo_compressed_size = 627; ^ = 0 modules/mono/editor/../glue/cs_compressed.gen.h:132:27: error: expected ';' after top level declarator static const int _cs_Godot API_Properties_AssemblyInfo_compressed_size = 627; ^ ; modules/mono/editor/../glue/cs_compressed.gen.h:133:18: error: redefinition of '_cs_Godot' static const int _cs_Godot API_Properties_AssemblyInfo_uncompressed_size = 1355; ^ modules/mono/editor/../glue/cs_compressed.gen.h:132:18: note: previous definition is here static const int _cs_Godot API_Properties_AssemblyInfo_compressed_size = 627; ^ modules/mono/editor/../glue/cs_compressed.gen.h:133:27: error: expected ';' after top level declarator static const int _cs_Godot API_Properties_AssemblyInfo_uncompressed_size = 1355; ^ ; modules/mono/editor/../glue/cs_compressed.gen.h:134:28: error: redefinition of '_cs_Godot' with a different type: 'const unsigned char' vs 'const int' static const unsigned char _cs_Godot API_Properties_AssemblyInfo_compressed[] = { 120, 156, 133, 84, 75, 110, 26, 65, 16, 221, 115, 138, 18, 27, 19, 43, 96, 76, 140, 228, 207, 10, ... ^ modules/mono/editor/../glue/cs_compressed.gen.h:132:18: note: previous definition is here static const int _cs_Godot API_Properties_AssemblyInfo_compressed_size = 627; ^ modules/mono/editor/../glue/cs_compressed.gen.h:134:37: error: expected ';' after top level declarator static const unsigned char _cs_Godot API_Properties_AssemblyInfo_compressed[] = { 120, 156, 133, 84, 75, 110, 26, 65, 16, 221, 115, 138, 18, 27, 19, 43, 96, 76, 140, 228, 207, 10, ... ^ ; modules/mono/editor/../glue/cs_compressed.gen.h:187:82: error: expected ')' r_files.insert("Godot API/Properties/AssemblyInfo.cs", CompressedFile(_cs_Godot API_Properties_AssemblyInfo_compressed_size, _cs_Godot API_Properties_AssemblyInfo_uncompres... ^ modules/mono/editor/../glue/cs_compressed.gen.h:187:71: note: to match this '(' r_files.insert("Godot API/Properties/AssemblyInfo.cs", CompressedFile(_cs_Godot API_Properties_AssemblyInfo_compressed_size, _cs_Godot API_Properties_AssemblyInfo_uncompres... ^ modules/mono/editor/../glue/cs_compressed.gen.h:187:57: error: no matching conversion for functional-style cast from 'const int' to 'CompressedFile' r_files.insert("Godot API/Properties/AssemblyInfo.cs", CompressedFile(_cs_Godot API_Properties_AssemblyInfo_compressed_size, _cs_Godot API_Properties_AssemblyInfo_uncompres... ^~~~~~~~~~~~~~~~~~~~~~~~ modules/mono/editor/../glue/cs_compressed.gen.h:138:8: note: candidate constructor (the implicit copy constructor) not viable: no known conversion from 'const int' to 'const CompressedFile' for 1st argument struct CompressedFile ^ modules/mono/editor/../glue/cs_compressed.gen.h:138:8: note: candidate constructor (the implicit move constructor) not viable: no known conversion from 'const int' to 'CompressedFile' for 1st argument struct CompressedFile ^ modules/mono/editor/../glue/cs_compressed.gen.h:151:2: note: candidate constructor not viable: requires 0 arguments, but 1 was provided CompressedFile() {} ^ modules/mono/editor/../glue/cs_compressed.gen.h:144:2: note: candidate constructor not viable: requires 3 arguments, but 1 was provided CompressedFile(int p_comp_size, int p_uncomp_size, const unsigned char* p_data) ^ [ 29%] 8 errors generated. ``` (@neikeq Could it be possible that it was caused by 72873004332b?)
non_priority
build fails if sub directory under modules mono glue cs files contains special characters godot version master os device including version manjaro linux issue description the current master branch fails to build due to compilation errors in cs compressed gen h as shown below in file included from modules mono editor bindings generator cpp modules mono editor glue cs compressed gen h error default initialization of an object of const type const int static const int cs godot api properties assemblyinfo compressed size modules mono editor glue cs compressed gen h error expected after top level declarator static const int cs godot api properties assemblyinfo compressed size modules mono editor glue cs compressed gen h error redefinition of cs godot static const int cs godot api properties assemblyinfo uncompressed size modules mono editor glue cs compressed gen h note previous definition is here static const int cs godot api properties assemblyinfo compressed size modules mono editor glue cs compressed gen h error expected after top level declarator static const int cs godot api properties assemblyinfo uncompressed size modules mono editor glue cs compressed gen h error redefinition of cs godot with a different type const unsigned char vs const int static const unsigned char cs godot api properties assemblyinfo compressed modules mono editor glue cs compressed gen h note previous definition is here static const int cs godot api properties assemblyinfo compressed size modules mono editor glue cs compressed gen h error expected after top level declarator static const unsigned char cs godot api properties assemblyinfo compressed modules mono editor glue cs compressed gen h error expected r files insert godot api properties assemblyinfo cs compressedfile cs godot api properties assemblyinfo compressed size cs godot api properties assemblyinfo uncompres modules mono editor glue cs compressed gen h note to match this r files insert godot api properties assemblyinfo cs compressedfile cs godot api properties assemblyinfo compressed size cs godot api properties assemblyinfo uncompres modules mono editor glue cs compressed gen h error no matching conversion for functional style cast from const int to compressedfile r files insert godot api properties assemblyinfo cs compressedfile cs godot api properties assemblyinfo compressed size cs godot api properties assemblyinfo uncompres modules mono editor glue cs compressed gen h note candidate constructor the implicit copy constructor not viable no known conversion from const int to const compressedfile for argument struct compressedfile modules mono editor glue cs compressed gen h note candidate constructor the implicit move constructor not viable no known conversion from const int to compressedfile for argument struct compressedfile modules mono editor glue cs compressed gen h note candidate constructor not viable requires arguments but was provided compressedfile modules mono editor glue cs compressed gen h note candidate constructor not viable requires arguments but was provided compressedfile int p comp size int p uncomp size const unsigned char p data errors generated neikeq could it be possible that it was caused by
0
222,724
7,435,534,884
IssuesEvent
2018-03-26 14:26:50
AnSyn/ansyn
https://api.github.com/repos/AnSyn/ansyn
opened
New provider - bugs
Bug Priority: High
1. imagery perspective is incorrect 2. tile loading failure is very common - please see if you can do anything about it
1.0
New provider - bugs - 1. imagery perspective is incorrect 2. tile loading failure is very common - please see if you can do anything about it
priority
new provider bugs imagery perspective is incorrect tile loading failure is very common please see if you can do anything about it
1
31,531
2,733,664,925
IssuesEvent
2015-04-17 15:13:25
GoogleCloudPlatform/kubernetes
https://api.github.com/repos/GoogleCloudPlatform/kubernetes
opened
Setup prometheus server and dashboard
area/performance priority/P2 team/cluster
@wojtek-t @mwielgus It'd be nice to setup prometheus server so that we can query/graph metrics that we export. This doesn't block other work, but would make it easier to debug performance issues.
1.0
Setup prometheus server and dashboard - @wojtek-t @mwielgus It'd be nice to setup prometheus server so that we can query/graph metrics that we export. This doesn't block other work, but would make it easier to debug performance issues.
priority
setup prometheus server and dashboard wojtek t mwielgus it d be nice to setup prometheus server so that we can query graph metrics that we export this doesn t block other work but would make it easier to debug performance issues
1