Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
6,138
9,009,231,675
IssuesEvent
2019-02-05 08:20:31
cypress-io/cypress
https://api.github.com/repos/cypress-io/cypress
closed
CLI unit tests often break - and it is flake
CI: circle process: tests
When building CLI on Circle the same 5 unit tests break, but when rerunning them pass example build (renovate PR) https://circleci.com/gh/cypress-io/cypress/43807 from workflow https://circleci.com/workflow-run/842085ba-85aa-4b7d-a20f-2b3dd58d4409 <img width="1343" alt="screen shot 2018-12-27 at 10 15 32 am" src="https://user-images.githubusercontent.com/2212006/50484903-d85f3600-09c0-11e9-8ac7-0605c4eae6f7.png"> all 5 errors come from `lib/tasks/verify` Seems the error message gets confused - a completely different value of snapshot, like a previous test is not cleared correctly ... ``` -Error: No version of Cypress is installed in: /cache/Cypress/1.2.3/Cypress.app +Error: The cypress npm package is installed, but the Cypress binary is missing. ``` The other error shows ``` TypeError: os.platform.returns is not a function ``` but `os.platform` is stubbed in `spec_helper.coffee` which should run before each test ``` beforeEach(function () { sinon.stub(os, 'platform') sinon.stub(os, 'release') sinon.stub(util, 'getOsVersionAsync').resolves('Foo-OsVersion') }) ```
1.0
CLI unit tests often break - and it is flake - When building CLI on Circle the same 5 unit tests break, but when rerunning them pass example build (renovate PR) https://circleci.com/gh/cypress-io/cypress/43807 from workflow https://circleci.com/workflow-run/842085ba-85aa-4b7d-a20f-2b3dd58d4409 <img width="1343" alt="screen shot 2018-12-27 at 10 15 32 am" src="https://user-images.githubusercontent.com/2212006/50484903-d85f3600-09c0-11e9-8ac7-0605c4eae6f7.png"> all 5 errors come from `lib/tasks/verify` Seems the error message gets confused - a completely different value of snapshot, like a previous test is not cleared correctly ... ``` -Error: No version of Cypress is installed in: /cache/Cypress/1.2.3/Cypress.app +Error: The cypress npm package is installed, but the Cypress binary is missing. ``` The other error shows ``` TypeError: os.platform.returns is not a function ``` but `os.platform` is stubbed in `spec_helper.coffee` which should run before each test ``` beforeEach(function () { sinon.stub(os, 'platform') sinon.stub(os, 'release') sinon.stub(util, 'getOsVersionAsync').resolves('Foo-OsVersion') }) ```
process
cli unit tests often break and it is flake when building cli on circle the same unit tests break but when rerunning them pass example build renovate pr from workflow img width alt screen shot at am src all errors come from lib tasks verify seems the error message gets confused a completely different value of snapshot like a previous test is not cleared correctly error no version of cypress is installed in cache cypress cypress app error the cypress npm package is installed but the cypress binary is missing the other error shows typeerror os platform returns is not a function but os platform is stubbed in spec helper coffee which should run before each test beforeeach function sinon stub os platform sinon stub os release sinon stub util getosversionasync resolves foo osversion
1
289,406
8,870,079,497
IssuesEvent
2019-01-11 08:21:49
Radarr/Radarr
https://api.github.com/repos/Radarr/Radarr
reopened
Memory Leak
bug confirmed help wanted priority:medium under investigation
**Description:** Radarr appears to have a memory leak. I have started getting warnings in the past few weeks about used SWAP space, and I had narrowed it down to being one of my Docker containers. Whenever I restarted all my containers, the usage dropped substantially. Today I went through and found that Radarr was using over half that space (2GB). I removed and rebuilt the container, and that dropped to 0. **Radarr Version:** I am running Radarr through the LSIO Docker container. Version 0.2.0.696 Mono Version 5.0.1.1 (2017-02/5077205 Thu May 25 09:19:18 UTC 2017) Let me know any information I can provide to help diagnose this issue.
1.0
Memory Leak - **Description:** Radarr appears to have a memory leak. I have started getting warnings in the past few weeks about used SWAP space, and I had narrowed it down to being one of my Docker containers. Whenever I restarted all my containers, the usage dropped substantially. Today I went through and found that Radarr was using over half that space (2GB). I removed and rebuilt the container, and that dropped to 0. **Radarr Version:** I am running Radarr through the LSIO Docker container. Version 0.2.0.696 Mono Version 5.0.1.1 (2017-02/5077205 Thu May 25 09:19:18 UTC 2017) Let me know any information I can provide to help diagnose this issue.
non_process
memory leak description radarr appears to have a memory leak i have started getting warnings in the past few weeks about used swap space and i had narrowed it down to being one of my docker containers whenever i restarted all my containers the usage dropped substantially today i went through and found that radarr was using over half that space i removed and rebuilt the container and that dropped to radarr version i am running radarr through the lsio docker container version mono version thu may utc let me know any information i can provide to help diagnose this issue
0
18,451
4,273,043,220
IssuesEvent
2016-07-13 16:07:18
hyperspy/hyperspy
https://api.github.com/repos/hyperspy/hyperspy
reopened
Archlinux installation instructions are old
documentation fix-submitted
As the title says, the archlinux installation instructions are old / deprecated (written for python2 as well). As it stands now, they should be at least removed from the user guide, preferably updated with a working set of steps.
1.0
Archlinux installation instructions are old - As the title says, the archlinux installation instructions are old / deprecated (written for python2 as well). As it stands now, they should be at least removed from the user guide, preferably updated with a working set of steps.
non_process
archlinux installation instructions are old as the title says the archlinux installation instructions are old deprecated written for as well as it stands now they should be at least removed from the user guide preferably updated with a working set of steps
0
22,016
30,521,492,976
IssuesEvent
2023-07-19 08:23:43
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
[Mirror] Please add mirrors for Zulu JDK 11.0.20 LTS
P2 type: process team-OSS mirror request
### Please list the URLs of the archives you'd like to mirror: * Linux x64: https://cdn.azul.com/zulu/bin/zulu11.66.15-ca-jdk11.0.20-linux_x64.tar.gz * Linux aarch64: https://cdn.azul.com/zulu/bin/zulu11.66.15-ca-jdk11.0.20-linux_aarch64.tar.gz * MacOS x64: https://cdn.azul.com/zulu/bin/zulu11.66.15-ca-jdk11.0.20-macosx_x64.tar.gz * MacOS aarch64: https://cdn.azul.com/zulu/bin/zulu11.66.15-ca-jdk11.0.20-macosx_aarch64.tar.gz * Windows x64: https://cdn.azul.com/zulu/bin/zulu11.66.15-ca-jdk11.0.20-win_x64.zip
1.0
[Mirror] Please add mirrors for Zulu JDK 11.0.20 LTS - ### Please list the URLs of the archives you'd like to mirror: * Linux x64: https://cdn.azul.com/zulu/bin/zulu11.66.15-ca-jdk11.0.20-linux_x64.tar.gz * Linux aarch64: https://cdn.azul.com/zulu/bin/zulu11.66.15-ca-jdk11.0.20-linux_aarch64.tar.gz * MacOS x64: https://cdn.azul.com/zulu/bin/zulu11.66.15-ca-jdk11.0.20-macosx_x64.tar.gz * MacOS aarch64: https://cdn.azul.com/zulu/bin/zulu11.66.15-ca-jdk11.0.20-macosx_aarch64.tar.gz * Windows x64: https://cdn.azul.com/zulu/bin/zulu11.66.15-ca-jdk11.0.20-win_x64.zip
process
please add mirrors for zulu jdk lts please list the urls of the archives you d like to mirror linux linux macos macos windows
1
32,209
13,781,184,237
IssuesEvent
2020-10-08 15:50:42
cityofaustin/atd-data-tech
https://api.github.com/repos/cityofaustin/atd-data-tech
closed
[BUG] "All Assignments" not loading on Officers' "Sign Up" page
Product: Vision Zero in Action Service: Apps Type: Bug Report
"All Assignments" table never loads. ![Screen Shot 2020-10-07 at 4 55 17 PM](https://user-images.githubusercontent.com/1463708/95392170-e5ab9300-08bd-11eb-99ac-8baea70e943b.png) **URL** https://atd.knack.com/vza#sign-up/ **Device** I'm seeing this in desktop, OSX, Chrome. Matt (Albert) A. is also experiencing it on his iPhone.
1.0
[BUG] "All Assignments" not loading on Officers' "Sign Up" page - "All Assignments" table never loads. ![Screen Shot 2020-10-07 at 4 55 17 PM](https://user-images.githubusercontent.com/1463708/95392170-e5ab9300-08bd-11eb-99ac-8baea70e943b.png) **URL** https://atd.knack.com/vza#sign-up/ **Device** I'm seeing this in desktop, OSX, Chrome. Matt (Albert) A. is also experiencing it on his iPhone.
non_process
all assignments not loading on officers sign up page all assignments table never loads url device i m seeing this in desktop osx chrome matt albert a is also experiencing it on his iphone
0
21,513
29,799,856,152
IssuesEvent
2023-06-16 07:14:23
phuocduong-agilityio/internship-huy-dao
https://api.github.com/repos/phuocduong-agilityio/internship-huy-dao
closed
Implement filter the `Brand` button products and process the filter results.
In-process
### Filter by `Brand` - [x] Add product filtering based on selected brand types - [x] Add `type` for object button images - [x] Using `useState()` to change the status - [x] Filter by 1 or more brands at once ### Design - **See more details:** [Link](https://docs.google.com/document/d/1iTrT9cdKrMCjUc6aDLS514gEk7BUFj-m2SCs2wCy2Lw/edit#heading=h.7t5pbcmau6ih)
1.0
Implement filter the `Brand` button products and process the filter results. - ### Filter by `Brand` - [x] Add product filtering based on selected brand types - [x] Add `type` for object button images - [x] Using `useState()` to change the status - [x] Filter by 1 or more brands at once ### Design - **See more details:** [Link](https://docs.google.com/document/d/1iTrT9cdKrMCjUc6aDLS514gEk7BUFj-m2SCs2wCy2Lw/edit#heading=h.7t5pbcmau6ih)
process
implement filter the brand button products and process the filter results filter by brand add product filtering based on selected brand types add type for object button images using usestate to change the status filter by or more brands at once design see more details
1
128,405
18,048,206,601
IssuesEvent
2021-09-19 09:09:17
jinhogate/pizza_angular
https://api.github.com/repos/jinhogate/pizza_angular
opened
CVE-2019-6284 (Medium) detected in node-sass-4.9.3.tgz
security vulnerability
## CVE-2019-6284 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-sass-4.9.3.tgz</b></p></summary> <p>Wrapper around libsass</p> <p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.9.3.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.9.3.tgz</a></p> <p> Dependency Hierarchy: - build-angular-0.6.8.tgz (Root Library) - :x: **node-sass-4.9.3.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jinhogate/pizza_angular/commit/353ead6609cbe90c767728d4050746689c22c532">353ead6609cbe90c767728d4050746689c22c532</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In LibSass 3.5.5, a heap-based buffer over-read exists in Sass::Prelexer::alternatives in prelexer.hpp. <p>Publish Date: 2019-01-14 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-6284>CVE-2019-6284</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/sass/libsass/releases/tag/3.6.0">https://github.com/sass/libsass/releases/tag/3.6.0</a></p> <p>Release Date: 2020-08-24</p> <p>Fix Resolution: libsass - 3.6.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-6284 (Medium) detected in node-sass-4.9.3.tgz - ## CVE-2019-6284 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-sass-4.9.3.tgz</b></p></summary> <p>Wrapper around libsass</p> <p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.9.3.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.9.3.tgz</a></p> <p> Dependency Hierarchy: - build-angular-0.6.8.tgz (Root Library) - :x: **node-sass-4.9.3.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jinhogate/pizza_angular/commit/353ead6609cbe90c767728d4050746689c22c532">353ead6609cbe90c767728d4050746689c22c532</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In LibSass 3.5.5, a heap-based buffer over-read exists in Sass::Prelexer::alternatives in prelexer.hpp. <p>Publish Date: 2019-01-14 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-6284>CVE-2019-6284</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/sass/libsass/releases/tag/3.6.0">https://github.com/sass/libsass/releases/tag/3.6.0</a></p> <p>Release Date: 2020-08-24</p> <p>Fix Resolution: libsass - 3.6.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in node sass tgz cve medium severity vulnerability vulnerable library node sass tgz wrapper around libsass library home page a href dependency hierarchy build angular tgz root library x node sass tgz vulnerable library found in head commit a href found in base branch master vulnerability details in libsass a heap based buffer over read exists in sass prelexer alternatives in prelexer hpp publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution libsass step up your open source security game with whitesource
0
60,575
17,023,461,775
IssuesEvent
2021-07-03 02:09:11
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
Potlatch moves nodes without touching them
Component: potlatch (flash editor) Priority: major Resolution: wontfix Type: defect
**[Submitted to the original trac issue database at 12.53pm, Saturday, 15th August 2009]** While editing country names using the Multilingual Country-List*, it's becoming apparent that Potlatch moves nodes ever so slightly, without even touching the node. It's apparent because using the JOSM link on that page centers on the exact coordinates it last saw the country node. If it was edited by a Potlatch user in the mean time, the node has moved slightly off that coordinate. I tested this myself by editing a few in Potlatch. The Potlatch link on said page opens Potlatch with the country node selected. I never click it, but just add a new name:xx key and then click on an empty spot on the canvas to save the change. Checking the history then shows that the node coordinates were changed. Example for Russia (http://www.openstreetmap.org/browse/node/424314830/history) 19 64.6863033, 97.7453216 (Potlatch)[[BR]] 18 64.6863003, 97.7453221 (JOSM)[[BR]] [...][[BR]] 12 64.6863003, 97.7453221 (Potlatch)[[BR]] 11 64.6862985, 97.7453237 (JOSM)[[BR]] [...][[BR]] 9 64.6862985, 97.7453237 (Potlatch)[[BR]] 8 64.6862964, 97.7453264 (JOSM)[[BR]] [...][[BR]] 5 64.6862964, 97.7453264 (Potlatch)[[BR]] 4 64.6862927, 97.7453312 (JOSM)[[BR]] But it's happening in every case, for every country node edit at least. * http://cassini.toolserver.org/~mazder/multilingual-country-list/
1.0
Potlatch moves nodes without touching them - **[Submitted to the original trac issue database at 12.53pm, Saturday, 15th August 2009]** While editing country names using the Multilingual Country-List*, it's becoming apparent that Potlatch moves nodes ever so slightly, without even touching the node. It's apparent because using the JOSM link on that page centers on the exact coordinates it last saw the country node. If it was edited by a Potlatch user in the mean time, the node has moved slightly off that coordinate. I tested this myself by editing a few in Potlatch. The Potlatch link on said page opens Potlatch with the country node selected. I never click it, but just add a new name:xx key and then click on an empty spot on the canvas to save the change. Checking the history then shows that the node coordinates were changed. Example for Russia (http://www.openstreetmap.org/browse/node/424314830/history) 19 64.6863033, 97.7453216 (Potlatch)[[BR]] 18 64.6863003, 97.7453221 (JOSM)[[BR]] [...][[BR]] 12 64.6863003, 97.7453221 (Potlatch)[[BR]] 11 64.6862985, 97.7453237 (JOSM)[[BR]] [...][[BR]] 9 64.6862985, 97.7453237 (Potlatch)[[BR]] 8 64.6862964, 97.7453264 (JOSM)[[BR]] [...][[BR]] 5 64.6862964, 97.7453264 (Potlatch)[[BR]] 4 64.6862927, 97.7453312 (JOSM)[[BR]] But it's happening in every case, for every country node edit at least. * http://cassini.toolserver.org/~mazder/multilingual-country-list/
non_process
potlatch moves nodes without touching them while editing country names using the multilingual country list it s becoming apparent that potlatch moves nodes ever so slightly without even touching the node it s apparent because using the josm link on that page centers on the exact coordinates it last saw the country node if it was edited by a potlatch user in the mean time the node has moved slightly off that coordinate i tested this myself by editing a few in potlatch the potlatch link on said page opens potlatch with the country node selected i never click it but just add a new name xx key and then click on an empty spot on the canvas to save the change checking the history then shows that the node coordinates were changed example for russia potlatch josm potlatch josm potlatch josm potlatch josm but it s happening in every case for every country node edit at least
0
18,824
24,721,882,449
IssuesEvent
2022-10-20 11:20:54
hermes-hmc/workflow
https://api.github.com/repos/hermes-hmc/workflow
opened
Implement "new" step: validation
enhancement 2️ process/validate
After processing, the theoretically unified dataset should be validated, to a specific (configurable?) standard of valid metadata. This probably needs a specific entrypoint. In this step, things like adherence to a sensible default should be asserted.
1.0
Implement "new" step: validation - After processing, the theoretically unified dataset should be validated, to a specific (configurable?) standard of valid metadata. This probably needs a specific entrypoint. In this step, things like adherence to a sensible default should be asserted.
process
implement new step validation after processing the theoretically unified dataset should be validated to a specific configurable standard of valid metadata this probably needs a specific entrypoint in this step things like adherence to a sensible default should be asserted
1
234,699
7,725,065,980
IssuesEvent
2018-05-24 16:46:46
test4gloirin/m
https://api.github.com/repos/test4gloirin/m
closed
0001032: update 0.27 is not working on installation with ldap account backend
Addressbook bug high priority
**Reported by pschuele on 5 May 2009 15:26** update 0.27 is not working on installation with ldap account backend - error when creating addressbook_image table with foreign key
1.0
0001032: update 0.27 is not working on installation with ldap account backend - **Reported by pschuele on 5 May 2009 15:26** update 0.27 is not working on installation with ldap account backend - error when creating addressbook_image table with foreign key
non_process
update is not working on installation with ldap account backend reported by pschuele on may update is not working on installation with ldap account backend error when creating addressbook image table with foreign key
0
85,305
15,736,682,396
IssuesEvent
2021-03-30 01:11:53
vlaship/build-docker-image
https://api.github.com/repos/vlaship/build-docker-image
opened
CVE-2019-17563 (High) detected in tomcat-embed-core-9.0.26.jar
security vulnerability
## CVE-2019-17563 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-core-9.0.26.jar</b></p></summary> <p>Core Tomcat implementation</p> <p>Library home page: <a href="https://tomcat.apache.org/">https://tomcat.apache.org/</a></p> <p>Path to dependency file: build-docker-image/build.gradle</p> <p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/org.apache.tomcat.embed/tomcat-embed-core/9.0.26/6312ba542bc58fa9ee789a43516ce4d862548a6b/tomcat-embed-core-9.0.26.jar,/root/.gradle/caches/modules-2/files-2.1/org.apache.tomcat.embed/tomcat-embed-core/9.0.26/6312ba542bc58fa9ee789a43516ce4d862548a6b/tomcat-embed-core-9.0.26.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.1.9.RELEASE.jar (Root Library) - spring-boot-starter-tomcat-2.1.9.RELEASE.jar - tomcat-embed-websocket-9.0.26.jar - :x: **tomcat-embed-core-9.0.26.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> When using FORM authentication with Apache Tomcat 9.0.0.M1 to 9.0.29, 8.5.0 to 8.5.49 and 7.0.0 to 7.0.98 there was a narrow window where an attacker could perform a session fixation attack. The window was considered too narrow for an exploit to be practical but, erring on the side of caution, this issue has been treated as a security vulnerability. <p>Publish Date: 2019-12-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-17563>CVE-2019-17563</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-17563">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-17563</a></p> <p>Release Date: 2019-12-23</p> <p>Fix Resolution: org.apache.tomcat.embed:tomcat-embed-core:7.0.99,8.5.50,9.0.30;org.apache.tomcat:tomcat-catalina:7.0.99,8.5.50,9.0.30</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-17563 (High) detected in tomcat-embed-core-9.0.26.jar - ## CVE-2019-17563 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-core-9.0.26.jar</b></p></summary> <p>Core Tomcat implementation</p> <p>Library home page: <a href="https://tomcat.apache.org/">https://tomcat.apache.org/</a></p> <p>Path to dependency file: build-docker-image/build.gradle</p> <p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/org.apache.tomcat.embed/tomcat-embed-core/9.0.26/6312ba542bc58fa9ee789a43516ce4d862548a6b/tomcat-embed-core-9.0.26.jar,/root/.gradle/caches/modules-2/files-2.1/org.apache.tomcat.embed/tomcat-embed-core/9.0.26/6312ba542bc58fa9ee789a43516ce4d862548a6b/tomcat-embed-core-9.0.26.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.1.9.RELEASE.jar (Root Library) - spring-boot-starter-tomcat-2.1.9.RELEASE.jar - tomcat-embed-websocket-9.0.26.jar - :x: **tomcat-embed-core-9.0.26.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> When using FORM authentication with Apache Tomcat 9.0.0.M1 to 9.0.29, 8.5.0 to 8.5.49 and 7.0.0 to 7.0.98 there was a narrow window where an attacker could perform a session fixation attack. The window was considered too narrow for an exploit to be practical but, erring on the side of caution, this issue has been treated as a security vulnerability. <p>Publish Date: 2019-12-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-17563>CVE-2019-17563</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-17563">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-17563</a></p> <p>Release Date: 2019-12-23</p> <p>Fix Resolution: org.apache.tomcat.embed:tomcat-embed-core:7.0.99,8.5.50,9.0.30;org.apache.tomcat:tomcat-catalina:7.0.99,8.5.50,9.0.30</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in tomcat embed core jar cve high severity vulnerability vulnerable library tomcat embed core jar core tomcat implementation library home page a href path to dependency file build docker image build gradle path to vulnerable library root gradle caches modules files org apache tomcat embed tomcat embed core tomcat embed core jar root gradle caches modules files org apache tomcat embed tomcat embed core tomcat embed core jar dependency hierarchy spring boot starter web release jar root library spring boot starter tomcat release jar tomcat embed websocket jar x tomcat embed core jar vulnerable library vulnerability details when using form authentication with apache tomcat to to and to there was a narrow window where an attacker could perform a session fixation attack the window was considered too narrow for an exploit to be practical but erring on the side of caution this issue has been treated as a security vulnerability publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache tomcat embed tomcat embed core org apache tomcat tomcat catalina step up your open source security game with whitesource
0
259,298
27,621,798,088
IssuesEvent
2023-03-10 01:12:25
nidhi7598/linux-3.0.35_CVE-2018-13405
https://api.github.com/repos/nidhi7598/linux-3.0.35_CVE-2018-13405
opened
CVE-2023-1078 (High) detected in linuxlinux-3.0.40
Mend: dependency security vulnerability
## CVE-2023-1078 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-3.0.40</b></p></summary> <p> <p>Apache Software Foundation (ASF)</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v3.0/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v3.0/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-3.0.35_CVE-2018-13405/commit/662fbf6e1ed61fd353add2f52e2dd27e990364c7">662fbf6e1ed61fd353add2f52e2dd27e990364c7</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> This CVE was marked as RESERVED by NVD. This means that the CVE-ID is reserved for future use by the CVE Numbering Authority (CNA) or a security researcher. Still, the CVE details were not included in the CVE entry. This is because the original requester of the CVE ID assignment has not sent an update to MITRE with the information needed to populate the CVE entry. This page will reflect the classification results once they are available through NVD <p>Publish Date: 2023-02-28 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-1078>CVE-2023-1078</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2023-1078">https://www.linuxkernelcves.com/cves/CVE-2023-1078</a></p> <p>Release Date: 2023-02-28</p> <p>Fix Resolution: v4.19.273,v5.4.232,v5.10.168,v5.15.94,v6.1.12,v6.2-rc8</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2023-1078 (High) detected in linuxlinux-3.0.40 - ## CVE-2023-1078 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-3.0.40</b></p></summary> <p> <p>Apache Software Foundation (ASF)</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v3.0/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v3.0/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-3.0.35_CVE-2018-13405/commit/662fbf6e1ed61fd353add2f52e2dd27e990364c7">662fbf6e1ed61fd353add2f52e2dd27e990364c7</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> This CVE was marked as RESERVED by NVD. This means that the CVE-ID is reserved for future use by the CVE Numbering Authority (CNA) or a security researcher. Still, the CVE details were not included in the CVE entry. This is because the original requester of the CVE ID assignment has not sent an update to MITRE with the information needed to populate the CVE entry. This page will reflect the classification results once they are available through NVD <p>Publish Date: 2023-02-28 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-1078>CVE-2023-1078</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2023-1078">https://www.linuxkernelcves.com/cves/CVE-2023-1078</a></p> <p>Release Date: 2023-02-28</p> <p>Fix Resolution: v4.19.273,v5.4.232,v5.10.168,v5.15.94,v6.1.12,v6.2-rc8</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in linuxlinux cve high severity vulnerability vulnerable library linuxlinux apache software foundation asf library home page a href found in head commit a href found in base branch master vulnerable source files vulnerability details this cve was marked as reserved by nvd this means that the cve id is reserved for future use by the cve numbering authority cna or a security researcher still the cve details were not included in the cve entry this is because the original requester of the cve id assignment has not sent an update to mitre with the information needed to populate the cve entry this page will reflect the classification results once they are available through nvd publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
130,232
18,055,103,696
IssuesEvent
2021-09-20 07:05:41
opensrp/web
https://api.github.com/repos/opensrp/web
closed
User management Design QA issues
Bug Report Design
- [x] Remove gear icon - [x] Set search text 14px - [x] in the admin menu, when a subsection is open but you are in a different section, the subsection is still white. Is this default Ant? Update the submenu to show selection even when in a different section but under the same submenu - [x] Remove the `Required Actions` field
1.0
User management Design QA issues - - [x] Remove gear icon - [x] Set search text 14px - [x] in the admin menu, when a subsection is open but you are in a different section, the subsection is still white. Is this default Ant? Update the submenu to show selection even when in a different section but under the same submenu - [x] Remove the `Required Actions` field
non_process
user management design qa issues remove gear icon set search text in the admin menu when a subsection is open but you are in a different section the subsection is still white is this default ant update the submenu to show selection even when in a different section but under the same submenu remove the required actions field
0
176,733
6,564,450,879
IssuesEvent
2017-09-08 01:40:18
HAS-CRM/IssueTracker
https://api.github.com/repos/HAS-CRM/IssueTracker
opened
Prepare UAT Server for Vietnam Testing
Priority.High Status.Ongoing Type.ChangeRequest
Background: Irene will want to let user at Vietnam try out at CRM Test Server.
1.0
Prepare UAT Server for Vietnam Testing - Background: Irene will want to let user at Vietnam try out at CRM Test Server.
non_process
prepare uat server for vietnam testing background irene will want to let user at vietnam try out at crm test server
0
413,535
27,957,105,737
IssuesEvent
2023-03-24 13:12:19
cloudscape-design/components
https://api.github.com/repos/cloudscape-design/components
opened
[Documentation]: How should help panels integrate with the app layout?
documentation
### Description I'm struggling to understand the documentation relating to the help panels. https://cloudscape.design/patterns/general/help-system/ says "Info links are the triggers that opens the help panel and display the corresponding content." and "An info link should always be anchored to headers." ... but that documentation just points to the documentation for the Link component and doesn't explain: * How the help panel gets opened * How to link the Link component to specific content - it mentions headers but how are they referenced? ### Code of Conduct - [X] I agree to follow this project's [Code of Conduct](https://github.com/cloudscape-design/components/blob/main/CODE_OF_CONDUCT.md) - [X] I checked the [current issues](https://github.com/cloudscape-design/components/issues) for duplicate requests
1.0
[Documentation]: How should help panels integrate with the app layout? - ### Description I'm struggling to understand the documentation relating to the help panels. https://cloudscape.design/patterns/general/help-system/ says "Info links are the triggers that opens the help panel and display the corresponding content." and "An info link should always be anchored to headers." ... but that documentation just points to the documentation for the Link component and doesn't explain: * How the help panel gets opened * How to link the Link component to specific content - it mentions headers but how are they referenced? ### Code of Conduct - [X] I agree to follow this project's [Code of Conduct](https://github.com/cloudscape-design/components/blob/main/CODE_OF_CONDUCT.md) - [X] I checked the [current issues](https://github.com/cloudscape-design/components/issues) for duplicate requests
non_process
how should help panels integrate with the app layout description i m struggling to understand the documentation relating to the help panels says info links are the triggers that opens the help panel and display the corresponding content and an info link should always be anchored to headers but that documentation just points to the documentation for the link component and doesn t explain how the help panel gets opened how to link the link component to specific content it mentions headers but how are they referenced code of conduct i agree to follow this project s i checked the for duplicate requests
0
267,278
20,198,141,900
IssuesEvent
2022-02-11 12:40:08
wagtail/wagtail
https://api.github.com/repos/wagtail/wagtail
reopened
Add banner to docs (for Wagtail Space or in general)
type:Enhancement Documentation
We want to add a banner to the docs to advertise Wagtail Space. We're referencing #4986, and we could just do the same as it, but for the upcoming Wagtail Space. One potential difficulty is getting it into a tag that will show in the stable version of the docs. Another option would be to add a banner page to [wagtail.io](https://wagtail.io/) or some other place we can update without a PR, and have the docs pull that banner if it exists. This would allow us to update the banner without having to put in a PR to Wagtail core, and more importantly would mean that going forward, changing that banner would change it for all versions of the docs from now on. I'm not sure what pitfalls this might have around cross origin or caching, which could be considerations for this option. @vsalvino has offered to do some looking into this!
1.0
Add banner to docs (for Wagtail Space or in general) - We want to add a banner to the docs to advertise Wagtail Space. We're referencing #4986, and we could just do the same as it, but for the upcoming Wagtail Space. One potential difficulty is getting it into a tag that will show in the stable version of the docs. Another option would be to add a banner page to [wagtail.io](https://wagtail.io/) or some other place we can update without a PR, and have the docs pull that banner if it exists. This would allow us to update the banner without having to put in a PR to Wagtail core, and more importantly would mean that going forward, changing that banner would change it for all versions of the docs from now on. I'm not sure what pitfalls this might have around cross origin or caching, which could be considerations for this option. @vsalvino has offered to do some looking into this!
non_process
add banner to docs for wagtail space or in general we want to add a banner to the docs to advertise wagtail space we re referencing and we could just do the same as it but for the upcoming wagtail space one potential difficulty is getting it into a tag that will show in the stable version of the docs another option would be to add a banner page to or some other place we can update without a pr and have the docs pull that banner if it exists this would allow us to update the banner without having to put in a pr to wagtail core and more importantly would mean that going forward changing that banner would change it for all versions of the docs from now on i m not sure what pitfalls this might have around cross origin or caching which could be considerations for this option vsalvino has offered to do some looking into this
0
15,338
19,480,234,642
IssuesEvent
2021-12-25 04:58:19
emily-writes-poems/emily-writes-poems-processing
https://api.github.com/repos/emily-writes-poems/emily-writes-poems-processing
closed
add poem dropdown selector for creating new feature
processing refinement
also rework the function for creating new feature - since we will now provide a list of poem ids to choose from, we don't need to check that the poem exists already... now, update it so the function receives both id and title and uses those to create the feature.
1.0
add poem dropdown selector for creating new feature - also rework the function for creating new feature - since we will now provide a list of poem ids to choose from, we don't need to check that the poem exists already... now, update it so the function receives both id and title and uses those to create the feature.
process
add poem dropdown selector for creating new feature also rework the function for creating new feature since we will now provide a list of poem ids to choose from we don t need to check that the poem exists already now update it so the function receives both id and title and uses those to create the feature
1
15,239
19,161,813,189
IssuesEvent
2021-12-03 01:41:29
streamnative/pulsar-flink
https://api.github.com/repos/streamnative/pulsar-flink
closed
[FEATURE] pulsar-flink-sql-connector_2.11 support flink 1.12.2
type/bug platform/data-processing
**Is your feature request related to a problem? Please describe.** pulsar-flink-sql-connector_2.11 not support flink 1.12.2 when I use flink 1.12.2 and io.streamnative.connectors:pulsar-flink-sql-connector_2.11:1.12.4.2 it will tips Unrecognized field "readername" **Describe the solution you'd like** support 1.12.2
1.0
[FEATURE] pulsar-flink-sql-connector_2.11 support flink 1.12.2 - **Is your feature request related to a problem? Please describe.** pulsar-flink-sql-connector_2.11 not support flink 1.12.2 when I use flink 1.12.2 and io.streamnative.connectors:pulsar-flink-sql-connector_2.11:1.12.4.2 it will tips Unrecognized field "readername" **Describe the solution you'd like** support 1.12.2
process
pulsar flink sql connector support flink is your feature request related to a problem please describe pulsar flink sql connector not support flink when i use flink and io streamnative connectors pulsar flink sql connector it will tips unrecognized field readername describe the solution you d like support
1
5,966
8,787,894,422
IssuesEvent
2018-12-20 20:12:57
AnotherCodeArtist/CEPWare
https://api.github.com/repos/AnotherCodeArtist/CEPWare
closed
EmptyString Error in Flink-Task
WP6 - Complex Event Processing bug
One of the tasks fails when being submitted through the python-script, but works when submitted manually.
1.0
EmptyString Error in Flink-Task - One of the tasks fails when being submitted through the python-script, but works when submitted manually.
process
emptystring error in flink task one of the tasks fails when being submitted through the python script but works when submitted manually
1
10,555
2,622,173,264
IssuesEvent
2015-03-04 00:15:32
byzhang/leveldb
https://api.github.com/repos/byzhang/leveldb
closed
afdfsdfsdfsdf
auto-migrated Priority-Medium Type-Defect
``` sdfsdfsfd ``` Original issue reported on code.google.com by `wpx...@gmail.com` on 13 May 2011 at 10:13
1.0
afdfsdfsdfsdf - ``` sdfsdfsfd ``` Original issue reported on code.google.com by `wpx...@gmail.com` on 13 May 2011 at 10:13
non_process
afdfsdfsdfsdf sdfsdfsfd original issue reported on code google com by wpx gmail com on may at
0
297,365
22,352,473,213
IssuesEvent
2022-06-15 13:13:26
Deltares/HYDROLIB-core
https://api.github.com/repos/Deltares/HYDROLIB-core
opened
Update "First steps" tutorial to notebook and binder
documentation
**What is the need for this task.** The HYDROLIB team strives to create notebooks as tutorial that can be launched in binder **What is the task?** - improve first steps tutorial - create notebook - include binder option to documentation **Additional context** Add any other context or screenshots about the feature request here.
1.0
Update "First steps" tutorial to notebook and binder - **What is the need for this task.** The HYDROLIB team strives to create notebooks as tutorial that can be launched in binder **What is the task?** - improve first steps tutorial - create notebook - include binder option to documentation **Additional context** Add any other context or screenshots about the feature request here.
non_process
update first steps tutorial to notebook and binder what is the need for this task the hydrolib team strives to create notebooks as tutorial that can be launched in binder what is the task improve first steps tutorial create notebook include binder option to documentation additional context add any other context or screenshots about the feature request here
0
8,558
11,731,084,018
IssuesEvent
2020-03-10 23:00:30
AcademySoftwareFoundation/OpenCue
https://api.github.com/repos/AcademySoftwareFoundation/OpenCue
closed
label "good first issue" to issues where applicable
process triaged
ASWF is trying to standardize this across projects. Section being added to TAC README in https://github.com/AcademySoftwareFoundation/tac/pull/45
1.0
label "good first issue" to issues where applicable - ASWF is trying to standardize this across projects. Section being added to TAC README in https://github.com/AcademySoftwareFoundation/tac/pull/45
process
label good first issue to issues where applicable aswf is trying to standardize this across projects section being added to tac readme in
1
15,839
20,028,184,605
IssuesEvent
2022-02-02 00:26:31
googleapis/java-translate
https://api.github.com/repos/googleapis/java-translate
closed
com.example.translate.BatchTranslateTextTests: testBatchTranslateText failed
priority: p2 type: process api: translate flakybot: issue flakybot: flaky
Note: #709 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky. ---- commit: 8c1ae20536844353ffa329473c35e8a054a69d53 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/98a4e9bf-3fc6-49e1-afb6-db7c642449bc), [Sponge](http://sponge2/98a4e9bf-3fc6-49e1-afb6-db7c642449bc) status: failed <details><summary>Test output</summary><br><pre>java.util.concurrent.ExecutionException: com.google.api.gax.rpc.InvalidArgumentException: io.grpc.StatusRuntimeException: INVALID_ARGUMENT: Output dir is in use by another batch translation job. output_uri_prefix: gs://java-docs-samples-testing/BATCH_TRANSLATION_OUTPUT_58ec83b6-dd3a-4214-badf-bd2aaec14128/ at com.google.common.util.concurrent.AbstractFuture.getDoneValue(AbstractFuture.java:566) at com.google.common.util.concurrent.AbstractFuture.get(AbstractFuture.java:445) at com.google.common.util.concurrent.FluentFuture$TrustedFuture.get(FluentFuture.java:95) at com.google.common.util.concurrent.ForwardingFuture.get(ForwardingFuture.java:68) at com.google.api.gax.longrunning.OperationFutureImpl.get(OperationFutureImpl.java:133) at com.example.translate.BatchTranslateText.batchTranslateText(BatchTranslateText.java:92) at com.example.translate.BatchTranslateTextTests.testBatchTranslateText(BatchTranslateTextTests.java:118) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at com.google.cloud.testing.junit4.MultipleAttemptsRule$1.evaluate(MultipleAttemptsRule.java:94) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:364) at org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:272) at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:237) at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:158) at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:428) at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162) at org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:562) at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:548) Caused by: com.google.api.gax.rpc.InvalidArgumentException: io.grpc.StatusRuntimeException: INVALID_ARGUMENT: Output dir is in use by another batch translation job. output_uri_prefix: gs://java-docs-samples-testing/BATCH_TRANSLATION_OUTPUT_58ec83b6-dd3a-4214-badf-bd2aaec14128/ at com.google.api.gax.rpc.ApiExceptionFactory.createException(ApiExceptionFactory.java:49) at com.google.api.gax.grpc.GrpcApiExceptionFactory.create(GrpcApiExceptionFactory.java:72) at com.google.api.gax.grpc.GrpcApiExceptionFactory.create(GrpcApiExceptionFactory.java:60) at com.google.api.gax.grpc.GrpcExceptionCallable$ExceptionTransformingFuture.onFailure(GrpcExceptionCallable.java:97) at com.google.api.core.ApiFutures$1.onFailure(ApiFutures.java:68) at com.google.common.util.concurrent.Futures$CallbackListener.run(Futures.java:1074) at com.google.common.util.concurrent.DirectExecutor.execute(DirectExecutor.java:30) at com.google.common.util.concurrent.AbstractFuture.executeListener(AbstractFuture.java:1213) at com.google.common.util.concurrent.AbstractFuture.complete(AbstractFuture.java:983) at com.google.common.util.concurrent.AbstractFuture.setException(AbstractFuture.java:771) at io.grpc.stub.ClientCalls$GrpcFuture.setException(ClientCalls.java:563) at io.grpc.stub.ClientCalls$UnaryStreamToFuture.onClose(ClientCalls.java:533) at io.grpc.internal.DelayedClientCall$DelayedListener$3.run(DelayedClientCall.java:463) at io.grpc.internal.DelayedClientCall$DelayedListener.delayOrExecute(DelayedClientCall.java:427) at io.grpc.internal.DelayedClientCall$DelayedListener.onClose(DelayedClientCall.java:460) at io.grpc.internal.ClientCallImpl.closeObserver(ClientCallImpl.java:557) at io.grpc.internal.ClientCallImpl.access$300(ClientCallImpl.java:69) at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInternal(ClientCallImpl.java:738) at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInContext(ClientCallImpl.java:717) at io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37) at io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:133) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) Caused by: io.grpc.StatusRuntimeException: INVALID_ARGUMENT: Output dir is in use by another batch translation job. output_uri_prefix: gs://java-docs-samples-testing/BATCH_TRANSLATION_OUTPUT_58ec83b6-dd3a-4214-badf-bd2aaec14128/ at io.grpc.Status.asRuntimeException(Status.java:535) ... 13 more </pre></details>
1.0
com.example.translate.BatchTranslateTextTests: testBatchTranslateText failed - Note: #709 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky. ---- commit: 8c1ae20536844353ffa329473c35e8a054a69d53 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/98a4e9bf-3fc6-49e1-afb6-db7c642449bc), [Sponge](http://sponge2/98a4e9bf-3fc6-49e1-afb6-db7c642449bc) status: failed <details><summary>Test output</summary><br><pre>java.util.concurrent.ExecutionException: com.google.api.gax.rpc.InvalidArgumentException: io.grpc.StatusRuntimeException: INVALID_ARGUMENT: Output dir is in use by another batch translation job. output_uri_prefix: gs://java-docs-samples-testing/BATCH_TRANSLATION_OUTPUT_58ec83b6-dd3a-4214-badf-bd2aaec14128/ at com.google.common.util.concurrent.AbstractFuture.getDoneValue(AbstractFuture.java:566) at com.google.common.util.concurrent.AbstractFuture.get(AbstractFuture.java:445) at com.google.common.util.concurrent.FluentFuture$TrustedFuture.get(FluentFuture.java:95) at com.google.common.util.concurrent.ForwardingFuture.get(ForwardingFuture.java:68) at com.google.api.gax.longrunning.OperationFutureImpl.get(OperationFutureImpl.java:133) at com.example.translate.BatchTranslateText.batchTranslateText(BatchTranslateText.java:92) at com.example.translate.BatchTranslateTextTests.testBatchTranslateText(BatchTranslateTextTests.java:118) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at com.google.cloud.testing.junit4.MultipleAttemptsRule$1.evaluate(MultipleAttemptsRule.java:94) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:364) at org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:272) at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:237) at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:158) at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:428) at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162) at org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:562) at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:548) Caused by: com.google.api.gax.rpc.InvalidArgumentException: io.grpc.StatusRuntimeException: INVALID_ARGUMENT: Output dir is in use by another batch translation job. output_uri_prefix: gs://java-docs-samples-testing/BATCH_TRANSLATION_OUTPUT_58ec83b6-dd3a-4214-badf-bd2aaec14128/ at com.google.api.gax.rpc.ApiExceptionFactory.createException(ApiExceptionFactory.java:49) at com.google.api.gax.grpc.GrpcApiExceptionFactory.create(GrpcApiExceptionFactory.java:72) at com.google.api.gax.grpc.GrpcApiExceptionFactory.create(GrpcApiExceptionFactory.java:60) at com.google.api.gax.grpc.GrpcExceptionCallable$ExceptionTransformingFuture.onFailure(GrpcExceptionCallable.java:97) at com.google.api.core.ApiFutures$1.onFailure(ApiFutures.java:68) at com.google.common.util.concurrent.Futures$CallbackListener.run(Futures.java:1074) at com.google.common.util.concurrent.DirectExecutor.execute(DirectExecutor.java:30) at com.google.common.util.concurrent.AbstractFuture.executeListener(AbstractFuture.java:1213) at com.google.common.util.concurrent.AbstractFuture.complete(AbstractFuture.java:983) at com.google.common.util.concurrent.AbstractFuture.setException(AbstractFuture.java:771) at io.grpc.stub.ClientCalls$GrpcFuture.setException(ClientCalls.java:563) at io.grpc.stub.ClientCalls$UnaryStreamToFuture.onClose(ClientCalls.java:533) at io.grpc.internal.DelayedClientCall$DelayedListener$3.run(DelayedClientCall.java:463) at io.grpc.internal.DelayedClientCall$DelayedListener.delayOrExecute(DelayedClientCall.java:427) at io.grpc.internal.DelayedClientCall$DelayedListener.onClose(DelayedClientCall.java:460) at io.grpc.internal.ClientCallImpl.closeObserver(ClientCallImpl.java:557) at io.grpc.internal.ClientCallImpl.access$300(ClientCallImpl.java:69) at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInternal(ClientCallImpl.java:738) at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInContext(ClientCallImpl.java:717) at io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37) at io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:133) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) Caused by: io.grpc.StatusRuntimeException: INVALID_ARGUMENT: Output dir is in use by another batch translation job. output_uri_prefix: gs://java-docs-samples-testing/BATCH_TRANSLATION_OUTPUT_58ec83b6-dd3a-4214-badf-bd2aaec14128/ at io.grpc.Status.asRuntimeException(Status.java:535) ... 13 more </pre></details>
process
com example translate batchtranslatetexttests testbatchtranslatetext failed note was also for this test but it was closed more than days ago so i didn t mark it flaky commit buildurl status failed test output java util concurrent executionexception com google api gax rpc invalidargumentexception io grpc statusruntimeexception invalid argument output dir is in use by another batch translation job output uri prefix gs java docs samples testing batch translation output badf at com google common util concurrent abstractfuture getdonevalue abstractfuture java at com google common util concurrent abstractfuture get abstractfuture java at com google common util concurrent fluentfuture trustedfuture get fluentfuture java at com google common util concurrent forwardingfuture get forwardingfuture java at com google api gax longrunning operationfutureimpl get operationfutureimpl java at com example translate batchtranslatetext batchtranslatetext batchtranslatetext java at com example translate batchtranslatetexttests testbatchtranslatetext batchtranslatetexttests java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at org junit internal runners statements runbefores evaluate runbefores java at org junit internal runners statements runafters evaluate runafters java at com google cloud testing multipleattemptsrule evaluate multipleattemptsrule java at org junit runners parentrunner evaluate parentrunner java at org junit runners evaluate java at org junit runners parentrunner runleaf parentrunner java at org junit runners runchild java at org junit runners runchild java at org junit runners parentrunner run parentrunner java at org junit runners parentrunner schedule parentrunner java at org junit runners parentrunner runchildren parentrunner java at org junit runners parentrunner access parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit internal runners statements runbefores evaluate runbefores java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner run parentrunner java at org apache maven surefire execute java at org apache maven surefire executewithrerun java at org apache maven surefire executetestset java at org apache maven surefire invoke java at org apache maven surefire booter forkedbooter runsuitesinprocess forkedbooter java at org apache maven surefire booter forkedbooter execute forkedbooter java at org apache maven surefire booter forkedbooter run forkedbooter java at org apache maven surefire booter forkedbooter main forkedbooter java caused by com google api gax rpc invalidargumentexception io grpc statusruntimeexception invalid argument output dir is in use by another batch translation job output uri prefix gs java docs samples testing batch translation output badf at com google api gax rpc apiexceptionfactory createexception apiexceptionfactory java at com google api gax grpc grpcapiexceptionfactory create grpcapiexceptionfactory java at com google api gax grpc grpcapiexceptionfactory create grpcapiexceptionfactory java at com google api gax grpc grpcexceptioncallable exceptiontransformingfuture onfailure grpcexceptioncallable java at com google api core apifutures onfailure apifutures java at com google common util concurrent futures callbacklistener run futures java at com google common util concurrent directexecutor execute directexecutor java at com google common util concurrent abstractfuture executelistener abstractfuture java at com google common util concurrent abstractfuture complete abstractfuture java at com google common util concurrent abstractfuture setexception abstractfuture java at io grpc stub clientcalls grpcfuture setexception clientcalls java at io grpc stub clientcalls unarystreamtofuture onclose clientcalls java at io grpc internal delayedclientcall delayedlistener run delayedclientcall java at io grpc internal delayedclientcall delayedlistener delayorexecute delayedclientcall java at io grpc internal delayedclientcall delayedlistener onclose delayedclientcall java at io grpc internal clientcallimpl closeobserver clientcallimpl java at io grpc internal clientcallimpl access clientcallimpl java at io grpc internal clientcallimpl clientstreamlistenerimpl runinternal clientcallimpl java at io grpc internal clientcallimpl clientstreamlistenerimpl runincontext clientcallimpl java at io grpc internal contextrunnable run contextrunnable java at io grpc internal serializingexecutor run serializingexecutor java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java caused by io grpc statusruntimeexception invalid argument output dir is in use by another batch translation job output uri prefix gs java docs samples testing batch translation output badf at io grpc status asruntimeexception status java more
1
75,423
9,854,021,970
IssuesEvent
2019-06-19 15:54:36
kubernetes-sigs/cluster-api
https://api.github.com/repos/kubernetes-sigs/cluster-api
closed
Add a new use case for cluster health check
kind/bug kind/documentation priority/backlog
/kind bug **What steps did you take and what happened:** [A clear and concise description of what the bug is.] **What did you expect to happen:** This is a follow up action for https://github.com/kubernetes-sigs/cluster-api/pull/903#discussion_r277693810 , I will create a PR for the use case of cluster health check. /assign FYI @vincepri @jichenjc @detiber
1.0
Add a new use case for cluster health check - /kind bug **What steps did you take and what happened:** [A clear and concise description of what the bug is.] **What did you expect to happen:** This is a follow up action for https://github.com/kubernetes-sigs/cluster-api/pull/903#discussion_r277693810 , I will create a PR for the use case of cluster health check. /assign FYI @vincepri @jichenjc @detiber
non_process
add a new use case for cluster health check kind bug what steps did you take and what happened what did you expect to happen this is a follow up action for i will create a pr for the use case of cluster health check assign fyi vincepri jichenjc detiber
0
19,663
26,026,579,981
IssuesEvent
2022-12-21 16:51:28
ConnorBaker/BSRT
https://api.github.com/repos/ConnorBaker/BSRT
closed
data_processing: Create separate module
enhancement module: data_processing
The `data-processing` module should be refactored into its own, standalone module outside the BSRT module. It offers a number of utility functions, like creating RAW bursts from a single image, which are very useful.
1.0
data_processing: Create separate module - The `data-processing` module should be refactored into its own, standalone module outside the BSRT module. It offers a number of utility functions, like creating RAW bursts from a single image, which are very useful.
process
data processing create separate module the data processing module should be refactored into its own standalone module outside the bsrt module it offers a number of utility functions like creating raw bursts from a single image which are very useful
1
181,644
30,719,826,838
IssuesEvent
2023-07-27 15:12:09
WordPress/gutenberg
https://api.github.com/repos/WordPress/gutenberg
opened
Allow large interactive areas
Needs Design Feedback
## What problem does this address? <!-- Please describe if this feature or enhancement is related to a current problem or pain point. For example, "I'm always frustrated when ..." or "It is currently difficult to ...". --> Sometimes we want a whole group of related visual items to have the same interaction (click, tap, keydown) result. For example in social media feeds sometimes the whole feed item links to the "detail" page. The feed item includes number of replies, the content, maybe other information too. The gist is to allow making HTML5 anchors which support just about any element set inside (including `div`) as long as none are interactive elements. ## What is your proposed solution? <!-- Please outline the feature or enhancement that you want and how it addresses any problem identified above. --> Make a group link block which does not allow buttons, videos etc as child? No idea really.
1.0
Allow large interactive areas - ## What problem does this address? <!-- Please describe if this feature or enhancement is related to a current problem or pain point. For example, "I'm always frustrated when ..." or "It is currently difficult to ...". --> Sometimes we want a whole group of related visual items to have the same interaction (click, tap, keydown) result. For example in social media feeds sometimes the whole feed item links to the "detail" page. The feed item includes number of replies, the content, maybe other information too. The gist is to allow making HTML5 anchors which support just about any element set inside (including `div`) as long as none are interactive elements. ## What is your proposed solution? <!-- Please outline the feature or enhancement that you want and how it addresses any problem identified above. --> Make a group link block which does not allow buttons, videos etc as child? No idea really.
non_process
allow large interactive areas what problem does this address please describe if this feature or enhancement is related to a current problem or pain point for example i m always frustrated when or it is currently difficult to sometimes we want a whole group of related visual items to have the same interaction click tap keydown result for example in social media feeds sometimes the whole feed item links to the detail page the feed item includes number of replies the content maybe other information too the gist is to allow making anchors which support just about any element set inside including div as long as none are interactive elements what is your proposed solution please outline the feature or enhancement that you want and how it addresses any problem identified above make a group link block which does not allow buttons videos etc as child no idea really
0
106,403
11,486,924,912
IssuesEvent
2020-02-11 10:53:29
fahlke/golibs
https://api.github.com/repos/fahlke/golibs
closed
Add documentation for Pearson hashing
documentation
- [x] adjust `README.md` - [x] add examples to `doc.go` - [x] add package description to `doc.go` - [x] document sources and tests
1.0
Add documentation for Pearson hashing - - [x] adjust `README.md` - [x] add examples to `doc.go` - [x] add package description to `doc.go` - [x] document sources and tests
non_process
add documentation for pearson hashing adjust readme md add examples to doc go add package description to doc go document sources and tests
0
16,549
21,568,599,087
IssuesEvent
2022-05-02 04:17:56
lynnandtonic/nestflix.fun
https://api.github.com/repos/lynnandtonic/nestflix.fun
closed
Add Today In Fashion
suggested title in process
Please add as much of the following info as you can: Title: Today In Fashion Type (film/tv show): tv reality Film or show in which it appears: Mars Attacks! Is the parent film/show streaming anywhere? HBO MAX About when in the parent film/show does it appear? First 20 mins Actual footage of the film/show can be seen (yes/no)? Yep
1.0
Add Today In Fashion - Please add as much of the following info as you can: Title: Today In Fashion Type (film/tv show): tv reality Film or show in which it appears: Mars Attacks! Is the parent film/show streaming anywhere? HBO MAX About when in the parent film/show does it appear? First 20 mins Actual footage of the film/show can be seen (yes/no)? Yep
process
add today in fashion please add as much of the following info as you can title today in fashion type film tv show tv reality film or show in which it appears mars attacks is the parent film show streaming anywhere hbo max about when in the parent film show does it appear first mins actual footage of the film show can be seen yes no yep
1
3,894
6,821,024,178
IssuesEvent
2017-11-07 15:37:39
syndesisio/syndesis-ui
https://api.github.com/repos/syndesisio/syndesis-ui
opened
Standardize the code linting flow for all contributors
dev process enhancement in progress Priority - Low
With more developers coming on board we should strive to keep consistency in our code in regards of format, style and syntax. Applying a linting phase bound to the transpilation/versioning phases of development seems a good start. The idea is to ensure that all developers get their code linted regardless the IDE in use and also get tips and hints on malformed code on dev-time. On a side note, a proper linting upon committing code will reduce the time required for peer-reviewing our code. **Requirements** - [ ] Update lint config manifest with industry standard conventions for Angular/TypeScript projects whereas required - [ ] Enable automated code linting regardless the IDE in use as a background process. - [ ] Ensure a linting pass is applied upon committing code to the forked repo - [ ] Update all install scripts necessary to make these rules available upon bootstrapping the frontend app
1.0
Standardize the code linting flow for all contributors - With more developers coming on board we should strive to keep consistency in our code in regards of format, style and syntax. Applying a linting phase bound to the transpilation/versioning phases of development seems a good start. The idea is to ensure that all developers get their code linted regardless the IDE in use and also get tips and hints on malformed code on dev-time. On a side note, a proper linting upon committing code will reduce the time required for peer-reviewing our code. **Requirements** - [ ] Update lint config manifest with industry standard conventions for Angular/TypeScript projects whereas required - [ ] Enable automated code linting regardless the IDE in use as a background process. - [ ] Ensure a linting pass is applied upon committing code to the forked repo - [ ] Update all install scripts necessary to make these rules available upon bootstrapping the frontend app
process
standardize the code linting flow for all contributors with more developers coming on board we should strive to keep consistency in our code in regards of format style and syntax applying a linting phase bound to the transpilation versioning phases of development seems a good start the idea is to ensure that all developers get their code linted regardless the ide in use and also get tips and hints on malformed code on dev time on a side note a proper linting upon committing code will reduce the time required for peer reviewing our code requirements update lint config manifest with industry standard conventions for angular typescript projects whereas required enable automated code linting regardless the ide in use as a background process ensure a linting pass is applied upon committing code to the forked repo update all install scripts necessary to make these rules available upon bootstrapping the frontend app
1
14,218
17,138,079,088
IssuesEvent
2021-07-13 06:16:44
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[QA] Push notifications not triggered from SB
Bug P0 Process: Fixed Unknown backend
**Steps:** 1. Send a manual app level/study level notification 2. Observe the status in 'Notifications' screen **Actual:** Push notifications not triggered from SB and status is 'Sending' **Expected:** Push notifications should be triggered Note: Issue tested in QA Issue also observed for automated push notifications ![Screenshot_1](https://user-images.githubusercontent.com/60386291/116519461-223ceb80-a8ef-11eb-81ea-5416854ddaa9.png)
1.0
[QA] Push notifications not triggered from SB - **Steps:** 1. Send a manual app level/study level notification 2. Observe the status in 'Notifications' screen **Actual:** Push notifications not triggered from SB and status is 'Sending' **Expected:** Push notifications should be triggered Note: Issue tested in QA Issue also observed for automated push notifications ![Screenshot_1](https://user-images.githubusercontent.com/60386291/116519461-223ceb80-a8ef-11eb-81ea-5416854ddaa9.png)
process
push notifications not triggered from sb steps send a manual app level study level notification observe the status in notifications screen actual push notifications not triggered from sb and status is sending expected push notifications should be triggered note issue tested in qa issue also observed for automated push notifications
1
18,455
24,548,599,278
IssuesEvent
2022-10-12 10:44:33
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[PM] [QA] Dashboard > Sites / Studies > UI Issue is observed in Participants details screen
Bug P2 Participant manager Process: Fixed Process: Tested dev
UI issue is observed in , Sites / Studies > Participants details screen ![1](https://user-images.githubusercontent.com/86007179/165103591-7fa78288-ce09-4c2e-aeac-94d530b853f5.png)
2.0
[PM] [QA] Dashboard > Sites / Studies > UI Issue is observed in Participants details screen - UI issue is observed in , Sites / Studies > Participants details screen ![1](https://user-images.githubusercontent.com/86007179/165103591-7fa78288-ce09-4c2e-aeac-94d530b853f5.png)
process
dashboard sites studies ui issue is observed in participants details screen ui issue is observed in sites studies participants details screen
1
8,344
11,498,511,672
IssuesEvent
2020-02-12 12:09:27
ESMValGroup/ESMValCore
https://api.github.com/repos/ESMValGroup/ESMValCore
closed
`extract_month` not functioning properly
bug preprocessor
**Describe the bug** The preprocessor `extract_month` does not work when the input cube does not yet have the auxiliary coordinate `month_number`. It needs to be added first if it is not there yet. Example recipe: ``` CMIP5_landcover: &CMIP5_landcover additional_datasets: - {dataset: MPI-ESM-LR, project: CMIP5, ensemble: r1i1p1} preprocessors: pp_cmip: custom_order: true regrid: target_grid: 2x2 scheme: linear extract_month: # This preprocessor returns None (bug) month: 7 climate_statistics: operator: mean diagnostics: albedolandcover_cmip5: description: test variables: treeFrac: &variable_settings_cmip5 preprocessor: pp_cmip mip: Lmon exp: historical start_year: 2000 end_year: 2004 <<: *CMIP5_landcover scripts: albedolandcover: script: landcover/jumpintoipython.py ```
1.0
`extract_month` not functioning properly - **Describe the bug** The preprocessor `extract_month` does not work when the input cube does not yet have the auxiliary coordinate `month_number`. It needs to be added first if it is not there yet. Example recipe: ``` CMIP5_landcover: &CMIP5_landcover additional_datasets: - {dataset: MPI-ESM-LR, project: CMIP5, ensemble: r1i1p1} preprocessors: pp_cmip: custom_order: true regrid: target_grid: 2x2 scheme: linear extract_month: # This preprocessor returns None (bug) month: 7 climate_statistics: operator: mean diagnostics: albedolandcover_cmip5: description: test variables: treeFrac: &variable_settings_cmip5 preprocessor: pp_cmip mip: Lmon exp: historical start_year: 2000 end_year: 2004 <<: *CMIP5_landcover scripts: albedolandcover: script: landcover/jumpintoipython.py ```
process
extract month not functioning properly describe the bug the preprocessor extract month does not work when the input cube does not yet have the auxiliary coordinate month number it needs to be added first if it is not there yet example recipe landcover landcover additional datasets dataset mpi esm lr project ensemble preprocessors pp cmip custom order true regrid target grid scheme linear extract month this preprocessor returns none bug month climate statistics operator mean diagnostics albedolandcover description test variables treefrac variable settings preprocessor pp cmip mip lmon exp historical start year end year landcover scripts albedolandcover script landcover jumpintoipython py
1
167,760
13,041,582,735
IssuesEvent
2020-07-28 20:40:05
apple/servicetalk
https://api.github.com/repos/apple/servicetalk
closed
ZipkinPublisherTest.testThriftRoundTrip test failure
flaky tests
https://ci.servicetalk.io/job/servicetalk-java8-prb/904/testReport/junit/io.servicetalk.opentracing.zipkin.publisher/ZipkinPublisherTest/testThriftRoundTrip/ ``` Error Message org.junit.runners.model.TestTimedOutException: test timed out after 90000 milliseconds Stacktrace org.junit.runners.model.TestTimedOutException: test timed out after 90000 milliseconds at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingDeque.takeFirst(LinkedBlockingDeque.java:492) at java.util.concurrent.LinkedBlockingDeque.take(LinkedBlockingDeque.java:680) at io.servicetalk.opentracing.zipkin.publisher.ZipkinPublisherTest.testRoundTrip(ZipkinPublisherTest.java:143) at io.servicetalk.opentracing.zipkin.publisher.ZipkinPublisherTest.testThriftRoundTrip(ZipkinPublisherTest.java:94) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at io.servicetalk.concurrent.internal.ServiceTalkTestTimeout$TimeoutStatement$CallableStatement.call(ServiceTalkTestTimeout.java:171) at io.servicetalk.concurrent.internal.ServiceTalkTestTimeout$TimeoutStatement$CallableStatement.call(ServiceTalkTestTimeout.java:163) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.lang.Thread.run(Thread.java:748) Standard Output "/127.0.0.1:38356 to /127.0.0.1:42084 workers Thread 2" #12 waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@5779cd9d java.lang.Thread.State: WAITING at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at org.gradle.internal.remote.internal.hub.queue.EndPointQueue.take(EndPointQueue.java:49) at org.gradle.internal.remote.internal.hub.MessageHub$ConnectionDispatch.run(MessageHub.java:320) at org.gradle.internal.concurrent.ExecutorPolicy$CatchAndRecordFailures.onExecute(ExecutorPolicy.java:64) at org.gradle.internal.concurrent.ManagedExecutorImpl$1.run(ManagedExecutorImpl.java:48) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at org.gradle.internal.concurrent.ThreadFactoryImpl$ManagedThreadRunnable.run(ThreadFactoryImpl.java:56) at java.lang.Thread.run(Thread.java:748) Number of locked synchronizers = 1 - java.util.concurrent.ThreadPoolExecutor$Worker@67205a84 "/127.0.0.1:38356 to /127.0.0.1:42084 workers Thread 3" #13 runnable (in native) java.lang.Thread.State: RUNNABLE at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) at sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) - locked sun.nio.ch.Util$3@61d52535 - locked java.util.Collections$UnmodifiableSet@4a2cd26b - locked sun.nio.ch.EPollSelectorImpl@28fe26fc at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:101) at org.gradle.internal.remote.internal.inet.SocketConnection$SocketInputStream.read(SocketConnection.java:182) at com.esotericsoftware.kryo.io.Input.fill(Input.java:146) at com.esotericsoftware.kryo.io.Input.require(Input.java:178) at com.esotericsoftware.kryo.io.Input.readByte(Input.java:295) at org.gradle.internal.serialize.kryo.KryoBackedDecoder.readByte(KryoBackedDecoder.java:81) at org.gradle.internal.remote.internal.hub.InterHubMessageSerializer$MessageReader.read(InterHubMessageSerializer.java:64) at org.gradle.internal.remote.internal.hub.InterHubMessageSerializer$MessageReader.read(InterHubMessageSerializer.java:52) at org.gradle.internal.remote.internal.inet.SocketConnection.receive(SocketConnection.java:81) at org.gradle.internal.remote.internal.hub.MessageHub$ConnectionReceive.run(MessageHub.java:268) at org.gradle.internal.concurrent.ExecutorPolicy$CatchAndRecordFailures.onExecute(ExecutorPolicy.java:64) at org.gradle.internal.concurrent.ManagedExecutorImpl$1.run(ManagedExecutorImpl.java:48) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at org.gradle.internal.concurrent.ThreadFactoryImpl$ManagedThreadRunnable.run(ThreadFactoryImpl.java:56) at java.lang.Thread.run(Thread.java:748) Number of locked synchronizers = 1 - java.util.concurrent.ThreadPoolExecutor$Worker@2a17b7b6 "Finalizer" #3 waiting on java.lang.ref.ReferenceQueue$Lock@1f4da6ca java.lang.Thread.State: WAITING at java.lang.Object.wait(Native Method) at java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:144) at java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:165) at java.lang.ref.Finalizer$FinalizerThread.run(Finalizer.java:216) "Reference Handler" #2 waiting on java.lang.ref.Reference$Lock@33de90c2 java.lang.Thread.State: WAITING at java.lang.Object.wait(Native Method) at java.lang.Object.wait(Object.java:502) at java.lang.ref.Reference.tryHandlePending(Reference.java:191) at java.lang.ref.Reference$ReferenceHandler.run(Reference.java:153) "Signal Dispatcher" #4 runnable java.lang.Thread.State: RUNNABLE "Test worker" #11 runnable java.lang.Thread.State: RUNNABLE at sun.management.ThreadImpl.dumpThreads0(Native Method) at sun.management.ThreadImpl.getThreadInfo(ThreadImpl.java:448) at io.servicetalk.concurrent.internal.ServiceTalkTestTimeout$TimeoutStatement.dumpAllStacks(ServiceTalkTestTimeout.java:187) at io.servicetalk.concurrent.internal.ServiceTalkTestTimeout$TimeoutStatement.getResult(ServiceTalkTestTimeout.java:147) at io.servicetalk.concurrent.internal.ServiceTalkTestTimeout$TimeoutStatement.evaluate(ServiceTalkTestTimeout.java:127) at org.junit.rules.RunRules.evaluate(RunRules.java:20) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57) at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288) at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268) at org.junit.runners.ParentRunner.run(ParentRunner.java:363) at org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecutor.runTestClass(JUnitTestClassExecutor.java:110) at org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecutor.execute(JUnitTestClassExecutor.java:58) at org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecutor.execute(JUnitTestClassExecutor.java:38) at org.gradle.api.internal.tasks.testing.junit.AbstractJUnitTestClassProcessor.processTestClass(AbstractJUnitTestClassProcessor.java:62) at org.gradle.api.internal.tasks.testing.SuiteTestClassProcessor.processTestClass(SuiteTestClassProcessor.java:51) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.gradle.internal.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:36) at org.gradle.internal.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:24) at org.gradle.internal.dispatch.ContextClassLoaderDispatch.dispatch(ContextClassLoaderDispatch.java:33) at org.gradle.internal.dispatch.ProxyDispatchAdapter$DispatchingInvocationHandler.invoke(ProxyDispatchAdapter.java:94) at com.sun.proxy.$Proxy2.processTestClass(Unknown Source) at org.gradle.api.internal.tasks.testing.worker.TestWorker.processTestClass(TestWorker.java:118) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.gradle.internal.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:36) at org.gradle.internal.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:24) at org.gradle.internal.remote.internal.hub.MessageHubBackedObjectConnection$DispatchWrapper.dispatch(MessageHubBackedObjectConnection.java:182) at org.gradle.internal.remote.internal.hub.MessageHubBackedObjectConnection$DispatchWrapper.dispatch(MessageHubBackedObjectConnection.java:164) at org.gradle.internal.remote.internal.hub.MessageHub$Handler.run(MessageHub.java:412) at org.gradle.internal.concurrent.ExecutorPolicy$CatchAndRecordFailures.onExecute(ExecutorPolicy.java:64) at org.gradle.internal.concurrent.ManagedExecutorImpl$1.run(ManagedExecutorImpl.java:48) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at org.gradle.internal.concurrent.ThreadFactoryImpl$ManagedThreadRunnable.run(ThreadFactoryImpl.java:56) at java.lang.Thread.run(Thread.java:748) Number of locked synchronizers = 1 - java.util.concurrent.ThreadPoolExecutor$Worker@6f1fba17 "Time-limited test" #29 waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@21cbc678 java.lang.Thread.State: WAITING at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingDeque.takeFirst(LinkedBlockingDeque.java:492) at java.util.concurrent.LinkedBlockingDeque.take(LinkedBlockingDeque.java:680) at io.servicetalk.opentracing.zipkin.publisher.ZipkinPublisherTest.testRoundTrip(ZipkinPublisherTest.java:143) at io.servicetalk.opentracing.zipkin.publisher.ZipkinPublisherTest.testThriftRoundTrip(ZipkinPublisherTest.java:94) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at io.servicetalk.concurrent.internal.ServiceTalkTestTimeout$TimeoutStatement$CallableStatement.call(ServiceTalkTestTimeout.java:171) at io.servicetalk.concurrent.internal.ServiceTalkTestTimeout$TimeoutStatement$CallableStatement.call(ServiceTalkTestTimeout.java:163) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.lang.Thread.run(Thread.java:748) "main" #1 waiting on java.util.concurrent.CountDownLatch$Sync@6a4e9983 java.lang.Thread.State: WAITING at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer.parkAndCheckInterrupt(AbstractQueuedSynchronizer.java:836) at java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(AbstractQueuedSynchronizer.java:997) at java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1304) at java.util.concurrent.CountDownLatch.await(CountDownLatch.java:231) at org.gradle.api.internal.tasks.testing.worker.TestWorker.execute(TestWorker.java:72) at org.gradle.api.internal.tasks.testing.worker.TestWorker.execute(TestWorker.java:46) at org.gradle.process.internal.worker.child.ActionExecutionWorker.execute(ActionExecutionWorker.java:91) at org.gradle.process.internal.worker.child.ActionExecutionWorker.execute(ActionExecutionWorker.java:34) at org.gradle.process.internal.worker.child.SystemApplicationClassLoaderWorker.call(SystemApplicationClassLoaderWorker.java:136) at org.gradle.process.internal.worker.child.SystemApplicationClassLoaderWorker.call(SystemApplicationClassLoaderWorker.java:69) at worker.org.gradle.process.internal.worker.GradleWorkerMain.run(GradleWorkerMain.java:68) at worker.org.gradle.process.internal.worker.GradleWorkerMain.main(GradleWorkerMain.java:73) "nioEventLoopGroup-8-1" #31 runnable (in native) java.lang.Thread.State: RUNNABLE at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) at sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) - locked io.netty.channel.nio.SelectedSelectionKeySet@46d0d1cd - locked java.util.Collections$UnmodifiableSet@69d8414d - locked sun.nio.ch.EPollSelectorImpl@2398351f at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) at io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:62) at io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:824) at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:457) at io.netty.util.concurrent.SingleThreadEventExecutor$6.run(SingleThreadEventExecutor.java:1050) at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) at java.lang.Thread.run(Thread.java:748) ```
1.0
ZipkinPublisherTest.testThriftRoundTrip test failure - https://ci.servicetalk.io/job/servicetalk-java8-prb/904/testReport/junit/io.servicetalk.opentracing.zipkin.publisher/ZipkinPublisherTest/testThriftRoundTrip/ ``` Error Message org.junit.runners.model.TestTimedOutException: test timed out after 90000 milliseconds Stacktrace org.junit.runners.model.TestTimedOutException: test timed out after 90000 milliseconds at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingDeque.takeFirst(LinkedBlockingDeque.java:492) at java.util.concurrent.LinkedBlockingDeque.take(LinkedBlockingDeque.java:680) at io.servicetalk.opentracing.zipkin.publisher.ZipkinPublisherTest.testRoundTrip(ZipkinPublisherTest.java:143) at io.servicetalk.opentracing.zipkin.publisher.ZipkinPublisherTest.testThriftRoundTrip(ZipkinPublisherTest.java:94) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at io.servicetalk.concurrent.internal.ServiceTalkTestTimeout$TimeoutStatement$CallableStatement.call(ServiceTalkTestTimeout.java:171) at io.servicetalk.concurrent.internal.ServiceTalkTestTimeout$TimeoutStatement$CallableStatement.call(ServiceTalkTestTimeout.java:163) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.lang.Thread.run(Thread.java:748) Standard Output "/127.0.0.1:38356 to /127.0.0.1:42084 workers Thread 2" #12 waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@5779cd9d java.lang.Thread.State: WAITING at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at org.gradle.internal.remote.internal.hub.queue.EndPointQueue.take(EndPointQueue.java:49) at org.gradle.internal.remote.internal.hub.MessageHub$ConnectionDispatch.run(MessageHub.java:320) at org.gradle.internal.concurrent.ExecutorPolicy$CatchAndRecordFailures.onExecute(ExecutorPolicy.java:64) at org.gradle.internal.concurrent.ManagedExecutorImpl$1.run(ManagedExecutorImpl.java:48) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at org.gradle.internal.concurrent.ThreadFactoryImpl$ManagedThreadRunnable.run(ThreadFactoryImpl.java:56) at java.lang.Thread.run(Thread.java:748) Number of locked synchronizers = 1 - java.util.concurrent.ThreadPoolExecutor$Worker@67205a84 "/127.0.0.1:38356 to /127.0.0.1:42084 workers Thread 3" #13 runnable (in native) java.lang.Thread.State: RUNNABLE at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) at sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) - locked sun.nio.ch.Util$3@61d52535 - locked java.util.Collections$UnmodifiableSet@4a2cd26b - locked sun.nio.ch.EPollSelectorImpl@28fe26fc at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:101) at org.gradle.internal.remote.internal.inet.SocketConnection$SocketInputStream.read(SocketConnection.java:182) at com.esotericsoftware.kryo.io.Input.fill(Input.java:146) at com.esotericsoftware.kryo.io.Input.require(Input.java:178) at com.esotericsoftware.kryo.io.Input.readByte(Input.java:295) at org.gradle.internal.serialize.kryo.KryoBackedDecoder.readByte(KryoBackedDecoder.java:81) at org.gradle.internal.remote.internal.hub.InterHubMessageSerializer$MessageReader.read(InterHubMessageSerializer.java:64) at org.gradle.internal.remote.internal.hub.InterHubMessageSerializer$MessageReader.read(InterHubMessageSerializer.java:52) at org.gradle.internal.remote.internal.inet.SocketConnection.receive(SocketConnection.java:81) at org.gradle.internal.remote.internal.hub.MessageHub$ConnectionReceive.run(MessageHub.java:268) at org.gradle.internal.concurrent.ExecutorPolicy$CatchAndRecordFailures.onExecute(ExecutorPolicy.java:64) at org.gradle.internal.concurrent.ManagedExecutorImpl$1.run(ManagedExecutorImpl.java:48) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at org.gradle.internal.concurrent.ThreadFactoryImpl$ManagedThreadRunnable.run(ThreadFactoryImpl.java:56) at java.lang.Thread.run(Thread.java:748) Number of locked synchronizers = 1 - java.util.concurrent.ThreadPoolExecutor$Worker@2a17b7b6 "Finalizer" #3 waiting on java.lang.ref.ReferenceQueue$Lock@1f4da6ca java.lang.Thread.State: WAITING at java.lang.Object.wait(Native Method) at java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:144) at java.lang.ref.ReferenceQueue.remove(ReferenceQueue.java:165) at java.lang.ref.Finalizer$FinalizerThread.run(Finalizer.java:216) "Reference Handler" #2 waiting on java.lang.ref.Reference$Lock@33de90c2 java.lang.Thread.State: WAITING at java.lang.Object.wait(Native Method) at java.lang.Object.wait(Object.java:502) at java.lang.ref.Reference.tryHandlePending(Reference.java:191) at java.lang.ref.Reference$ReferenceHandler.run(Reference.java:153) "Signal Dispatcher" #4 runnable java.lang.Thread.State: RUNNABLE "Test worker" #11 runnable java.lang.Thread.State: RUNNABLE at sun.management.ThreadImpl.dumpThreads0(Native Method) at sun.management.ThreadImpl.getThreadInfo(ThreadImpl.java:448) at io.servicetalk.concurrent.internal.ServiceTalkTestTimeout$TimeoutStatement.dumpAllStacks(ServiceTalkTestTimeout.java:187) at io.servicetalk.concurrent.internal.ServiceTalkTestTimeout$TimeoutStatement.getResult(ServiceTalkTestTimeout.java:147) at io.servicetalk.concurrent.internal.ServiceTalkTestTimeout$TimeoutStatement.evaluate(ServiceTalkTestTimeout.java:127) at org.junit.rules.RunRules.evaluate(RunRules.java:20) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57) at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288) at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268) at org.junit.runners.ParentRunner.run(ParentRunner.java:363) at org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecutor.runTestClass(JUnitTestClassExecutor.java:110) at org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecutor.execute(JUnitTestClassExecutor.java:58) at org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecutor.execute(JUnitTestClassExecutor.java:38) at org.gradle.api.internal.tasks.testing.junit.AbstractJUnitTestClassProcessor.processTestClass(AbstractJUnitTestClassProcessor.java:62) at org.gradle.api.internal.tasks.testing.SuiteTestClassProcessor.processTestClass(SuiteTestClassProcessor.java:51) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.gradle.internal.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:36) at org.gradle.internal.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:24) at org.gradle.internal.dispatch.ContextClassLoaderDispatch.dispatch(ContextClassLoaderDispatch.java:33) at org.gradle.internal.dispatch.ProxyDispatchAdapter$DispatchingInvocationHandler.invoke(ProxyDispatchAdapter.java:94) at com.sun.proxy.$Proxy2.processTestClass(Unknown Source) at org.gradle.api.internal.tasks.testing.worker.TestWorker.processTestClass(TestWorker.java:118) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.gradle.internal.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:36) at org.gradle.internal.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:24) at org.gradle.internal.remote.internal.hub.MessageHubBackedObjectConnection$DispatchWrapper.dispatch(MessageHubBackedObjectConnection.java:182) at org.gradle.internal.remote.internal.hub.MessageHubBackedObjectConnection$DispatchWrapper.dispatch(MessageHubBackedObjectConnection.java:164) at org.gradle.internal.remote.internal.hub.MessageHub$Handler.run(MessageHub.java:412) at org.gradle.internal.concurrent.ExecutorPolicy$CatchAndRecordFailures.onExecute(ExecutorPolicy.java:64) at org.gradle.internal.concurrent.ManagedExecutorImpl$1.run(ManagedExecutorImpl.java:48) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at org.gradle.internal.concurrent.ThreadFactoryImpl$ManagedThreadRunnable.run(ThreadFactoryImpl.java:56) at java.lang.Thread.run(Thread.java:748) Number of locked synchronizers = 1 - java.util.concurrent.ThreadPoolExecutor$Worker@6f1fba17 "Time-limited test" #29 waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@21cbc678 java.lang.Thread.State: WAITING at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039) at java.util.concurrent.LinkedBlockingDeque.takeFirst(LinkedBlockingDeque.java:492) at java.util.concurrent.LinkedBlockingDeque.take(LinkedBlockingDeque.java:680) at io.servicetalk.opentracing.zipkin.publisher.ZipkinPublisherTest.testRoundTrip(ZipkinPublisherTest.java:143) at io.servicetalk.opentracing.zipkin.publisher.ZipkinPublisherTest.testThriftRoundTrip(ZipkinPublisherTest.java:94) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at io.servicetalk.concurrent.internal.ServiceTalkTestTimeout$TimeoutStatement$CallableStatement.call(ServiceTalkTestTimeout.java:171) at io.servicetalk.concurrent.internal.ServiceTalkTestTimeout$TimeoutStatement$CallableStatement.call(ServiceTalkTestTimeout.java:163) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.lang.Thread.run(Thread.java:748) "main" #1 waiting on java.util.concurrent.CountDownLatch$Sync@6a4e9983 java.lang.Thread.State: WAITING at sun.misc.Unsafe.park(Native Method) at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175) at java.util.concurrent.locks.AbstractQueuedSynchronizer.parkAndCheckInterrupt(AbstractQueuedSynchronizer.java:836) at java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(AbstractQueuedSynchronizer.java:997) at java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1304) at java.util.concurrent.CountDownLatch.await(CountDownLatch.java:231) at org.gradle.api.internal.tasks.testing.worker.TestWorker.execute(TestWorker.java:72) at org.gradle.api.internal.tasks.testing.worker.TestWorker.execute(TestWorker.java:46) at org.gradle.process.internal.worker.child.ActionExecutionWorker.execute(ActionExecutionWorker.java:91) at org.gradle.process.internal.worker.child.ActionExecutionWorker.execute(ActionExecutionWorker.java:34) at org.gradle.process.internal.worker.child.SystemApplicationClassLoaderWorker.call(SystemApplicationClassLoaderWorker.java:136) at org.gradle.process.internal.worker.child.SystemApplicationClassLoaderWorker.call(SystemApplicationClassLoaderWorker.java:69) at worker.org.gradle.process.internal.worker.GradleWorkerMain.run(GradleWorkerMain.java:68) at worker.org.gradle.process.internal.worker.GradleWorkerMain.main(GradleWorkerMain.java:73) "nioEventLoopGroup-8-1" #31 runnable (in native) java.lang.Thread.State: RUNNABLE at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method) at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:269) at sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:93) at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:86) - locked io.netty.channel.nio.SelectedSelectionKeySet@46d0d1cd - locked java.util.Collections$UnmodifiableSet@69d8414d - locked sun.nio.ch.EPollSelectorImpl@2398351f at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:97) at io.netty.channel.nio.SelectedSelectionKeySetSelector.select(SelectedSelectionKeySetSelector.java:62) at io.netty.channel.nio.NioEventLoop.select(NioEventLoop.java:824) at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:457) at io.netty.util.concurrent.SingleThreadEventExecutor$6.run(SingleThreadEventExecutor.java:1050) at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) at java.lang.Thread.run(Thread.java:748) ```
non_process
zipkinpublishertest testthriftroundtrip test failure error message org junit runners model testtimedoutexception test timed out after milliseconds stacktrace org junit runners model testtimedoutexception test timed out after milliseconds at sun misc unsafe park native method at java util concurrent locks locksupport park locksupport java at java util concurrent locks abstractqueuedsynchronizer conditionobject await abstractqueuedsynchronizer java at java util concurrent linkedblockingdeque takefirst linkedblockingdeque java at java util concurrent linkedblockingdeque take linkedblockingdeque java at io servicetalk opentracing zipkin publisher zipkinpublishertest testroundtrip zipkinpublishertest java at io servicetalk opentracing zipkin publisher zipkinpublishertest testthriftroundtrip zipkinpublishertest java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at org junit internal runners statements runbefores evaluate runbefores java at org junit internal runners statements runafters evaluate runafters java at io servicetalk concurrent internal servicetalktesttimeout timeoutstatement callablestatement call servicetalktesttimeout java at io servicetalk concurrent internal servicetalktesttimeout timeoutstatement callablestatement call servicetalktesttimeout java at java util concurrent futuretask run futuretask java at java lang thread run thread java standard output to workers thread waiting on java util concurrent locks abstractqueuedsynchronizer conditionobject java lang thread state waiting at sun misc unsafe park native method at java util concurrent locks locksupport park locksupport java at java util concurrent locks abstractqueuedsynchronizer conditionobject await abstractqueuedsynchronizer java at org gradle internal remote internal hub queue endpointqueue take endpointqueue java at org gradle internal remote internal hub messagehub connectiondispatch run messagehub java at org gradle internal concurrent executorpolicy catchandrecordfailures onexecute executorpolicy java at org gradle internal concurrent managedexecutorimpl run managedexecutorimpl java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at org gradle internal concurrent threadfactoryimpl managedthreadrunnable run threadfactoryimpl java at java lang thread run thread java number of locked synchronizers java util concurrent threadpoolexecutor worker to workers thread runnable in native java lang thread state runnable at sun nio ch epollarraywrapper epollwait native method at sun nio ch epollarraywrapper poll epollarraywrapper java at sun nio ch epollselectorimpl doselect epollselectorimpl java at sun nio ch selectorimpl lockanddoselect selectorimpl java locked sun nio ch util locked java util collections unmodifiableset locked sun nio ch epollselectorimpl at sun nio ch selectorimpl select selectorimpl java at sun nio ch selectorimpl select selectorimpl java at org gradle internal remote internal inet socketconnection socketinputstream read socketconnection java at com esotericsoftware kryo io input fill input java at com esotericsoftware kryo io input require input java at com esotericsoftware kryo io input readbyte input java at org gradle internal serialize kryo kryobackeddecoder readbyte kryobackeddecoder java at org gradle internal remote internal hub interhubmessageserializer messagereader read interhubmessageserializer java at org gradle internal remote internal hub interhubmessageserializer messagereader read interhubmessageserializer java at org gradle internal remote internal inet socketconnection receive socketconnection java at org gradle internal remote internal hub messagehub connectionreceive run messagehub java at org gradle internal concurrent executorpolicy catchandrecordfailures onexecute executorpolicy java at org gradle internal concurrent managedexecutorimpl run managedexecutorimpl java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at org gradle internal concurrent threadfactoryimpl managedthreadrunnable run threadfactoryimpl java at java lang thread run thread java number of locked synchronizers java util concurrent threadpoolexecutor worker finalizer waiting on java lang ref referencequeue lock java lang thread state waiting at java lang object wait native method at java lang ref referencequeue remove referencequeue java at java lang ref referencequeue remove referencequeue java at java lang ref finalizer finalizerthread run finalizer java reference handler waiting on java lang ref reference lock java lang thread state waiting at java lang object wait native method at java lang object wait object java at java lang ref reference tryhandlepending reference java at java lang ref reference referencehandler run reference java signal dispatcher runnable java lang thread state runnable test worker runnable java lang thread state runnable at sun management threadimpl native method at sun management threadimpl getthreadinfo threadimpl java at io servicetalk concurrent internal servicetalktesttimeout timeoutstatement dumpallstacks servicetalktesttimeout java at io servicetalk concurrent internal servicetalktesttimeout timeoutstatement getresult servicetalktesttimeout java at io servicetalk concurrent internal servicetalktesttimeout timeoutstatement evaluate servicetalktesttimeout java at org junit rules runrules evaluate runrules java at org junit runners parentrunner runleaf parentrunner java at org junit runners runchild java at org junit runners runchild java at org junit runners parentrunner run parentrunner java at org junit runners parentrunner schedule parentrunner java at org junit runners parentrunner runchildren parentrunner java at org junit runners parentrunner access parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner run parentrunner java at org gradle api internal tasks testing junit junittestclassexecutor runtestclass junittestclassexecutor java at org gradle api internal tasks testing junit junittestclassexecutor execute junittestclassexecutor java at org gradle api internal tasks testing junit junittestclassexecutor execute junittestclassexecutor java at org gradle api internal tasks testing junit abstractjunittestclassprocessor processtestclass abstractjunittestclassprocessor java at org gradle api internal tasks testing suitetestclassprocessor processtestclass suitetestclassprocessor java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org gradle internal dispatch reflectiondispatch dispatch reflectiondispatch java at org gradle internal dispatch reflectiondispatch dispatch reflectiondispatch java at org gradle internal dispatch contextclassloaderdispatch dispatch contextclassloaderdispatch java at org gradle internal dispatch proxydispatchadapter dispatchinginvocationhandler invoke proxydispatchadapter java at com sun proxy processtestclass unknown source at org gradle api internal tasks testing worker testworker processtestclass testworker java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org gradle internal dispatch reflectiondispatch dispatch reflectiondispatch java at org gradle internal dispatch reflectiondispatch dispatch reflectiondispatch java at org gradle internal remote internal hub messagehubbackedobjectconnection dispatchwrapper dispatch messagehubbackedobjectconnection java at org gradle internal remote internal hub messagehubbackedobjectconnection dispatchwrapper dispatch messagehubbackedobjectconnection java at org gradle internal remote internal hub messagehub handler run messagehub java at org gradle internal concurrent executorpolicy catchandrecordfailures onexecute executorpolicy java at org gradle internal concurrent managedexecutorimpl run managedexecutorimpl java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at org gradle internal concurrent threadfactoryimpl managedthreadrunnable run threadfactoryimpl java at java lang thread run thread java number of locked synchronizers java util concurrent threadpoolexecutor worker time limited test waiting on java util concurrent locks abstractqueuedsynchronizer conditionobject java lang thread state waiting at sun misc unsafe park native method at java util concurrent locks locksupport park locksupport java at java util concurrent locks abstractqueuedsynchronizer conditionobject await abstractqueuedsynchronizer java at java util concurrent linkedblockingdeque takefirst linkedblockingdeque java at java util concurrent linkedblockingdeque take linkedblockingdeque java at io servicetalk opentracing zipkin publisher zipkinpublishertest testroundtrip zipkinpublishertest java at io servicetalk opentracing zipkin publisher zipkinpublishertest testthriftroundtrip zipkinpublishertest java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at org junit internal runners statements runbefores evaluate runbefores java at org junit internal runners statements runafters evaluate runafters java at io servicetalk concurrent internal servicetalktesttimeout timeoutstatement callablestatement call servicetalktesttimeout java at io servicetalk concurrent internal servicetalktesttimeout timeoutstatement callablestatement call servicetalktesttimeout java at java util concurrent futuretask run futuretask java at java lang thread run thread java main waiting on java util concurrent countdownlatch sync java lang thread state waiting at sun misc unsafe park native method at java util concurrent locks locksupport park locksupport java at java util concurrent locks abstractqueuedsynchronizer parkandcheckinterrupt abstractqueuedsynchronizer java at java util concurrent locks abstractqueuedsynchronizer doacquiresharedinterruptibly abstractqueuedsynchronizer java at java util concurrent locks abstractqueuedsynchronizer acquiresharedinterruptibly abstractqueuedsynchronizer java at java util concurrent countdownlatch await countdownlatch java at org gradle api internal tasks testing worker testworker execute testworker java at org gradle api internal tasks testing worker testworker execute testworker java at org gradle process internal worker child actionexecutionworker execute actionexecutionworker java at org gradle process internal worker child actionexecutionworker execute actionexecutionworker java at org gradle process internal worker child systemapplicationclassloaderworker call systemapplicationclassloaderworker java at org gradle process internal worker child systemapplicationclassloaderworker call systemapplicationclassloaderworker java at worker org gradle process internal worker gradleworkermain run gradleworkermain java at worker org gradle process internal worker gradleworkermain main gradleworkermain java nioeventloopgroup runnable in native java lang thread state runnable at sun nio ch epollarraywrapper epollwait native method at sun nio ch epollarraywrapper poll epollarraywrapper java at sun nio ch epollselectorimpl doselect epollselectorimpl java at sun nio ch selectorimpl lockanddoselect selectorimpl java locked io netty channel nio selectedselectionkeyset locked java util collections unmodifiableset locked sun nio ch epollselectorimpl at sun nio ch selectorimpl select selectorimpl java at io netty channel nio selectedselectionkeysetselector select selectedselectionkeysetselector java at io netty channel nio nioeventloop select nioeventloop java at io netty channel nio nioeventloop run nioeventloop java at io netty util concurrent singlethreadeventexecutor run singlethreadeventexecutor java at io netty util internal threadexecutormap run threadexecutormap java at io netty util concurrent fastthreadlocalrunnable run fastthreadlocalrunnable java at java lang thread run thread java
0
19,190
25,315,074,487
IssuesEvent
2022-11-17 20:50:54
carbon-design-system/ibm-cloud-cognitive
https://api.github.com/repos/carbon-design-system/ibm-cloud-cognitive
closed
Inline edit v2 release review
type: process improvement component: InlineEdit
## Review for release ### Readiness - [x] One or more scenarios for a design pattern have been identified as a useful unit of functionality to publish. - [x] A functioning component or components delivering those scenarios have been delivered and merged to main. - [x] Design maintainer has approved the implementation for those scenarios (via a comment on the relevant issue/epic). ### Engineering review - [x] Breaking changes have only been introduced with prior approval, discussion and documented in release notes. Ideally deprecation notices in the prior major version must have been added in a timely fashion. - [x] The implementation takes into account, and does not impair, remaining and anticipated design scenarios. - [x] Public component features (names, props, etc) are consistent with Carbon-defined conventions and are consistent internally. Where there isn't an existing convention to apply, ensure robust precedents are being established. - [x] The UI produced is accessible, responsive, translatable, cross-browser, and responds to the currently set Carbon theme. - [x] Components are functional components using hooks. - [x] Public components which produce DOM structures support className. - [x] Public components support a ref (via React.forwardRef). - [x] Public component supports a Devtools attribute - [x] All significant DOM elements have meaningful classes. - [x] Additional attributes that are not identified as props (such as title, aria-\*, etc) are passed through to an appropriate DOM node of the component as HTML attributes. - [x] No warnings, errors or log messages in the console. - [x] Each public component JS is exported in /src/components/index.js, each public component SCSS is included in /src/components/\_index.scss, and each public component has a flag in package-settings.js. - [x] Each public component SCSS lists all of the Carbon and C&CS components imported and used by the JavaScript code and explicitly imports the SCSS for each of these components. ### Standards - [x] No linter warnings or errors are produced. - [x] Carbon tokens (theme, layout, motion) are used unless the design specifies otherwise. - [x] All components utilizing motion must include reduced-motion queries for accessibility purposes - read more here. - [x] Code is formatted according to prettier rules (no use of //prettier-ignore). - [x] Code is clear, maintainable and follows standard React practices and the code guidelines. - [x] Copyright header in every source file (js, css, scss etc.) with the appropriate years. ### Testing - [x] There is a set of test cases for the components. - [x] The tests exercise all inputs (props, slots, etc) and verify appropriate outputs. - [x] The tests validate the behaviors and interactions defined in the design where practical. - [x] The tests achieve 100% coverage. Usage of istanbul ignore is appropriate and not extensive. - [x] No warnings, errors or log messages in the test output. ### Documentation - [x] Source code is satisfactorily commented and provides DocGen comments for all public components and their props. - [x] There is a story for each design scenario. The stories expose all relevant props and actions, and additional usage documentation and code samples are available on the 'Docs' tab along with the props table. - [ ] There is a sandbox (or multiple sandboxes if appropriate) on CodeSandbox for the components.
1.0
Inline edit v2 release review - ## Review for release ### Readiness - [x] One or more scenarios for a design pattern have been identified as a useful unit of functionality to publish. - [x] A functioning component or components delivering those scenarios have been delivered and merged to main. - [x] Design maintainer has approved the implementation for those scenarios (via a comment on the relevant issue/epic). ### Engineering review - [x] Breaking changes have only been introduced with prior approval, discussion and documented in release notes. Ideally deprecation notices in the prior major version must have been added in a timely fashion. - [x] The implementation takes into account, and does not impair, remaining and anticipated design scenarios. - [x] Public component features (names, props, etc) are consistent with Carbon-defined conventions and are consistent internally. Where there isn't an existing convention to apply, ensure robust precedents are being established. - [x] The UI produced is accessible, responsive, translatable, cross-browser, and responds to the currently set Carbon theme. - [x] Components are functional components using hooks. - [x] Public components which produce DOM structures support className. - [x] Public components support a ref (via React.forwardRef). - [x] Public component supports a Devtools attribute - [x] All significant DOM elements have meaningful classes. - [x] Additional attributes that are not identified as props (such as title, aria-\*, etc) are passed through to an appropriate DOM node of the component as HTML attributes. - [x] No warnings, errors or log messages in the console. - [x] Each public component JS is exported in /src/components/index.js, each public component SCSS is included in /src/components/\_index.scss, and each public component has a flag in package-settings.js. - [x] Each public component SCSS lists all of the Carbon and C&CS components imported and used by the JavaScript code and explicitly imports the SCSS for each of these components. ### Standards - [x] No linter warnings or errors are produced. - [x] Carbon tokens (theme, layout, motion) are used unless the design specifies otherwise. - [x] All components utilizing motion must include reduced-motion queries for accessibility purposes - read more here. - [x] Code is formatted according to prettier rules (no use of //prettier-ignore). - [x] Code is clear, maintainable and follows standard React practices and the code guidelines. - [x] Copyright header in every source file (js, css, scss etc.) with the appropriate years. ### Testing - [x] There is a set of test cases for the components. - [x] The tests exercise all inputs (props, slots, etc) and verify appropriate outputs. - [x] The tests validate the behaviors and interactions defined in the design where practical. - [x] The tests achieve 100% coverage. Usage of istanbul ignore is appropriate and not extensive. - [x] No warnings, errors or log messages in the test output. ### Documentation - [x] Source code is satisfactorily commented and provides DocGen comments for all public components and their props. - [x] There is a story for each design scenario. The stories expose all relevant props and actions, and additional usage documentation and code samples are available on the 'Docs' tab along with the props table. - [ ] There is a sandbox (or multiple sandboxes if appropriate) on CodeSandbox for the components.
process
inline edit release review review for release readiness one or more scenarios for a design pattern have been identified as a useful unit of functionality to publish a functioning component or components delivering those scenarios have been delivered and merged to main design maintainer has approved the implementation for those scenarios via a comment on the relevant issue epic engineering review breaking changes have only been introduced with prior approval discussion and documented in release notes ideally deprecation notices in the prior major version must have been added in a timely fashion the implementation takes into account and does not impair remaining and anticipated design scenarios public component features names props etc are consistent with carbon defined conventions and are consistent internally where there isn t an existing convention to apply ensure robust precedents are being established the ui produced is accessible responsive translatable cross browser and responds to the currently set carbon theme components are functional components using hooks public components which produce dom structures support classname public components support a ref via react forwardref public component supports a devtools attribute all significant dom elements have meaningful classes additional attributes that are not identified as props such as title aria etc are passed through to an appropriate dom node of the component as html attributes no warnings errors or log messages in the console each public component js is exported in src components index js each public component scss is included in src components index scss and each public component has a flag in package settings js each public component scss lists all of the carbon and c cs components imported and used by the javascript code and explicitly imports the scss for each of these components standards no linter warnings or errors are produced carbon tokens theme layout motion are used unless the design specifies otherwise all components utilizing motion must include reduced motion queries for accessibility purposes read more here code is formatted according to prettier rules no use of prettier ignore code is clear maintainable and follows standard react practices and the code guidelines copyright header in every source file js css scss etc with the appropriate years testing there is a set of test cases for the components the tests exercise all inputs props slots etc and verify appropriate outputs the tests validate the behaviors and interactions defined in the design where practical the tests achieve coverage usage of istanbul ignore is appropriate and not extensive no warnings errors or log messages in the test output documentation source code is satisfactorily commented and provides docgen comments for all public components and their props there is a story for each design scenario the stories expose all relevant props and actions and additional usage documentation and code samples are available on the docs tab along with the props table there is a sandbox or multiple sandboxes if appropriate on codesandbox for the components
1
7,048
10,208,530,512
IssuesEvent
2019-08-14 10:20:38
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Update Access Token
Pri2 assigned-to-author automation/svc process-automation/subsvc product-feedback triaged
There is no way to update the access token after it expires. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 83c90e64-b615-711f-a53d-fc76606e2ecd * Version Independent ID: 2d164036-6886-4440-50f7-369f99f41cea * Content: [Source Control integration in Azure Automation](https://docs.microsoft.com/en-us/azure/automation/source-control-integration#feedback) * Content Source: [articles/automation/source-control-integration.md](https://github.com/Microsoft/azure-docs/blob/master/articles/automation/source-control-integration.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @bobbytreed * Microsoft Alias: **robreed**
1.0
Update Access Token - There is no way to update the access token after it expires. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 83c90e64-b615-711f-a53d-fc76606e2ecd * Version Independent ID: 2d164036-6886-4440-50f7-369f99f41cea * Content: [Source Control integration in Azure Automation](https://docs.microsoft.com/en-us/azure/automation/source-control-integration#feedback) * Content Source: [articles/automation/source-control-integration.md](https://github.com/Microsoft/azure-docs/blob/master/articles/automation/source-control-integration.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @bobbytreed * Microsoft Alias: **robreed**
process
update access token there is no way to update the access token after it expires document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login bobbytreed microsoft alias robreed
1
221,723
17,366,826,698
IssuesEvent
2021-07-30 08:29:45
elastic/kibana
https://api.github.com/repos/elastic/kibana
closed
Failing test: X-Pack Accessibility Tests.x-pack/test/accessibility/apps/ml·ts - ml for user with full ML access with data loaded anomaly detection Anomaly Explorer page
:ml failed-test
A test failed on a tracked branch ``` Error: expected testSubject(mlPageAnomalyExplorer) to exist at TestSubjects.existOrFail (/dev/shm/workspace/parallel/17/kibana/test/functional/services/common/test_subjects.ts:45:13) at Object.openAnomalyExplorer (test/functional/services/ml/single_metric_viewer.ts:124:7) at Context.<anonymous> (test/accessibility/apps/ml.ts:223:11) at Object.apply (/dev/shm/workspace/parallel/17/kibana/node_modules/@kbn/test/target_node/functional_test_runner/lib/mocha/wrap_function.js:87:16) ``` First failure: [Jenkins Build](https://kibana-ci.elastic.co/job/elastic+kibana+7.x/15693/) <!-- kibanaCiData = {"failed-test":{"test.class":"X-Pack Accessibility Tests.x-pack/test/accessibility/apps/ml·ts","test.name":"ml for user with full ML access with data loaded anomaly detection Anomaly Explorer page","test.failCount":2}} -->
1.0
Failing test: X-Pack Accessibility Tests.x-pack/test/accessibility/apps/ml·ts - ml for user with full ML access with data loaded anomaly detection Anomaly Explorer page - A test failed on a tracked branch ``` Error: expected testSubject(mlPageAnomalyExplorer) to exist at TestSubjects.existOrFail (/dev/shm/workspace/parallel/17/kibana/test/functional/services/common/test_subjects.ts:45:13) at Object.openAnomalyExplorer (test/functional/services/ml/single_metric_viewer.ts:124:7) at Context.<anonymous> (test/accessibility/apps/ml.ts:223:11) at Object.apply (/dev/shm/workspace/parallel/17/kibana/node_modules/@kbn/test/target_node/functional_test_runner/lib/mocha/wrap_function.js:87:16) ``` First failure: [Jenkins Build](https://kibana-ci.elastic.co/job/elastic+kibana+7.x/15693/) <!-- kibanaCiData = {"failed-test":{"test.class":"X-Pack Accessibility Tests.x-pack/test/accessibility/apps/ml·ts","test.name":"ml for user with full ML access with data loaded anomaly detection Anomaly Explorer page","test.failCount":2}} -->
non_process
failing test x pack accessibility tests x pack test accessibility apps ml·ts ml for user with full ml access with data loaded anomaly detection anomaly explorer page a test failed on a tracked branch error expected testsubject mlpageanomalyexplorer to exist at testsubjects existorfail dev shm workspace parallel kibana test functional services common test subjects ts at object openanomalyexplorer test functional services ml single metric viewer ts at context test accessibility apps ml ts at object apply dev shm workspace parallel kibana node modules kbn test target node functional test runner lib mocha wrap function js first failure
0
94,580
15,987,536,852
IssuesEvent
2021-04-19 00:55:54
dotnet/aspnetcore
https://api.github.com/repos/dotnet/aspnetcore
closed
'Correlation failed.' on Google Authorization when using Docker - Cookie not found.
area-security
### Describe the bug I am getting a `Correlation failed.` exception when using the external Google authentication service in ASP.NET Core. I head to my Index page, and the Google account select box comes up. Then I select my account, and the exception happens. The The message also says something about a missing cookie (that is posted further down): ![image](https://user-images.githubusercontent.com/36459994/115150342-e9bb2900-a035-11eb-97c8-efe3652abda3.png) However, this exception ONLY happens when running the application with Docker: ![image](https://user-images.githubusercontent.com/36459994/115150367-00618000-a036-11eb-96db-76ffe7c134de.png) The authentication works totally fine when using Kestrel. Also, I have only seen the bug when using Firefox. It works totally fine in Chrome, even when using a Docker container. Even stranger, it will work in Firefox when I run the container on an external host (ie: my production scenario). The bug only happens for me when using a Docker container on `localhost` in Firefox. I have tried disabling all my extensions, clearing all cookies and cache for `localhost`, and using all the default security settings, but the crash will still happen in Firefox. ### To Reproduce The bug can be reproduced with a trivial application: [Github](https://github.com/jpbream/DockerBug) It contains only an `/Index` page, and an `AccountController` with the two authorization endpoints. On my machine, when I run this application with Docker, go to the homepage, and select my account, the exception will happen. However, if I don't use Docker (or I use Chrome), it proceeds to the `/Index` page with no issues. You will need to register an app in the `Google Developers Cloud` to test this however. I'm not going to share my Client ID and Secret. Doing so is very simple: 1. Head to the [Google Developers Console](https://console.developers.google.com/) 2. Create a new project 3. On the `OAuth Consent Screen` tab, configure the consent screen. Just use your email with all the default values. 4. On the `Credentials` tab, click `Create Credentials` -> `OAuth Client Id`. Choose `Web Application`. Be sure to leave `https://localhost:44380/signin-google` as an authorized redirect URI. This is the port I configured the Docker container to use. 5. The `Client ID` and `Client Secret` should pop up. Copy/paste these into the spots I left in the `appsettings.json` file in the test project. 6. Run the test project using Docker and see if you can reproduce the bug. You can delete the Google project when you are done with it. ### Exceptions (if any) ``` System.Exception: Correlation failed. System.Exception: An error was encountered while handling the remote login. ---> System.Exception: Correlation failed. --- End of inner exception stack trace --- at Microsoft.AspNetCore.Authentication.RemoteAuthenticationHandler`1.HandleRequestAsync() at Microsoft.AspNetCore.Authentication.AuthenticationMiddleware.Invoke(HttpContext context) at Microsoft.AspNetCore.Diagnostics.DeveloperExceptionPageMiddleware.Invoke(HttpContext context) ``` The container output also gives me this just before the exception happens: ``` warn: Microsoft.AspNetCore.Authentication.Google.GoogleHandler[15] '.AspNetCore.Correlation.gdbGbhLvCj-862NaogXrcQGhVS_mR1xlpIpbRtJ1pWA' cookie not found. Microsoft.AspNetCore.Authentication.Google.GoogleHandler: Warning: '.AspNetCore.Correlation.gdbGbhLvCj-862NaogXrcQGhVS_mR1xlpIpbRtJ1pWA' cookie not found. ``` The string of junk characters is random. Here is the console output from Firefox just as the exception happens: ![image](https://user-images.githubusercontent.com/36459994/115150722-a2ce3300-a037-11eb-8665-71096f772312.png) ### Further technical details - ASP.NET Core version: 5.0.5 - Output from `dotnet --info` ![image](https://user-images.githubusercontent.com/36459994/115150751-c98c6980-a037-11eb-9c8e-67f86a64a761.png) - Visual Studio v16.8.4 - My `WSL 2` Docker version: `20.10.5, build 55c4c88` - Firefox version: 87.0
True
'Correlation failed.' on Google Authorization when using Docker - Cookie not found. - ### Describe the bug I am getting a `Correlation failed.` exception when using the external Google authentication service in ASP.NET Core. I head to my Index page, and the Google account select box comes up. Then I select my account, and the exception happens. The The message also says something about a missing cookie (that is posted further down): ![image](https://user-images.githubusercontent.com/36459994/115150342-e9bb2900-a035-11eb-97c8-efe3652abda3.png) However, this exception ONLY happens when running the application with Docker: ![image](https://user-images.githubusercontent.com/36459994/115150367-00618000-a036-11eb-96db-76ffe7c134de.png) The authentication works totally fine when using Kestrel. Also, I have only seen the bug when using Firefox. It works totally fine in Chrome, even when using a Docker container. Even stranger, it will work in Firefox when I run the container on an external host (ie: my production scenario). The bug only happens for me when using a Docker container on `localhost` in Firefox. I have tried disabling all my extensions, clearing all cookies and cache for `localhost`, and using all the default security settings, but the crash will still happen in Firefox. ### To Reproduce The bug can be reproduced with a trivial application: [Github](https://github.com/jpbream/DockerBug) It contains only an `/Index` page, and an `AccountController` with the two authorization endpoints. On my machine, when I run this application with Docker, go to the homepage, and select my account, the exception will happen. However, if I don't use Docker (or I use Chrome), it proceeds to the `/Index` page with no issues. You will need to register an app in the `Google Developers Cloud` to test this however. I'm not going to share my Client ID and Secret. Doing so is very simple: 1. Head to the [Google Developers Console](https://console.developers.google.com/) 2. Create a new project 3. On the `OAuth Consent Screen` tab, configure the consent screen. Just use your email with all the default values. 4. On the `Credentials` tab, click `Create Credentials` -> `OAuth Client Id`. Choose `Web Application`. Be sure to leave `https://localhost:44380/signin-google` as an authorized redirect URI. This is the port I configured the Docker container to use. 5. The `Client ID` and `Client Secret` should pop up. Copy/paste these into the spots I left in the `appsettings.json` file in the test project. 6. Run the test project using Docker and see if you can reproduce the bug. You can delete the Google project when you are done with it. ### Exceptions (if any) ``` System.Exception: Correlation failed. System.Exception: An error was encountered while handling the remote login. ---> System.Exception: Correlation failed. --- End of inner exception stack trace --- at Microsoft.AspNetCore.Authentication.RemoteAuthenticationHandler`1.HandleRequestAsync() at Microsoft.AspNetCore.Authentication.AuthenticationMiddleware.Invoke(HttpContext context) at Microsoft.AspNetCore.Diagnostics.DeveloperExceptionPageMiddleware.Invoke(HttpContext context) ``` The container output also gives me this just before the exception happens: ``` warn: Microsoft.AspNetCore.Authentication.Google.GoogleHandler[15] '.AspNetCore.Correlation.gdbGbhLvCj-862NaogXrcQGhVS_mR1xlpIpbRtJ1pWA' cookie not found. Microsoft.AspNetCore.Authentication.Google.GoogleHandler: Warning: '.AspNetCore.Correlation.gdbGbhLvCj-862NaogXrcQGhVS_mR1xlpIpbRtJ1pWA' cookie not found. ``` The string of junk characters is random. Here is the console output from Firefox just as the exception happens: ![image](https://user-images.githubusercontent.com/36459994/115150722-a2ce3300-a037-11eb-8665-71096f772312.png) ### Further technical details - ASP.NET Core version: 5.0.5 - Output from `dotnet --info` ![image](https://user-images.githubusercontent.com/36459994/115150751-c98c6980-a037-11eb-9c8e-67f86a64a761.png) - Visual Studio v16.8.4 - My `WSL 2` Docker version: `20.10.5, build 55c4c88` - Firefox version: 87.0
non_process
correlation failed on google authorization when using docker cookie not found describe the bug i am getting a correlation failed exception when using the external google authentication service in asp net core i head to my index page and the google account select box comes up then i select my account and the exception happens the the message also says something about a missing cookie that is posted further down however this exception only happens when running the application with docker the authentication works totally fine when using kestrel also i have only seen the bug when using firefox it works totally fine in chrome even when using a docker container even stranger it will work in firefox when i run the container on an external host ie my production scenario the bug only happens for me when using a docker container on localhost in firefox i have tried disabling all my extensions clearing all cookies and cache for localhost and using all the default security settings but the crash will still happen in firefox to reproduce the bug can be reproduced with a trivial application it contains only an index page and an accountcontroller with the two authorization endpoints on my machine when i run this application with docker go to the homepage and select my account the exception will happen however if i don t use docker or i use chrome it proceeds to the index page with no issues you will need to register an app in the google developers cloud to test this however i m not going to share my client id and secret doing so is very simple head to the create a new project on the oauth consent screen tab configure the consent screen just use your email with all the default values on the credentials tab click create credentials oauth client id choose web application be sure to leave as an authorized redirect uri this is the port i configured the docker container to use the client id and client secret should pop up copy paste these into the spots i left in the appsettings json file in the test project run the test project using docker and see if you can reproduce the bug you can delete the google project when you are done with it exceptions if any system exception correlation failed system exception an error was encountered while handling the remote login system exception correlation failed end of inner exception stack trace at microsoft aspnetcore authentication remoteauthenticationhandler handlerequestasync at microsoft aspnetcore authentication authenticationmiddleware invoke httpcontext context at microsoft aspnetcore diagnostics developerexceptionpagemiddleware invoke httpcontext context the container output also gives me this just before the exception happens warn microsoft aspnetcore authentication google googlehandler aspnetcore correlation gdbgbhlvcj cookie not found microsoft aspnetcore authentication google googlehandler warning aspnetcore correlation gdbgbhlvcj cookie not found the string of junk characters is random here is the console output from firefox just as the exception happens further technical details asp net core version output from dotnet info visual studio my wsl docker version build firefox version
0
79,933
15,586,248,655
IssuesEvent
2021-03-18 01:30:41
hiucimon/ClamScanService
https://api.github.com/repos/hiucimon/ClamScanService
opened
CVE-2020-24616 (High) detected in jackson-databind-2.8.5.jar
security vulnerability
## CVE-2020-24616 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.5.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /ClamScanService/build.gradle</p> <p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.8.5/b3035f37e674c04dafe36a660c3815cc59f764e2/jackson-databind-2.8.5.jar,/root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.5/jackson-databind-2.8.5.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-thymeleaf-1.4.3.RELEASE.jar (Root Library) - spring-boot-starter-web-1.4.3.RELEASE.jar - :x: **jackson-databind-2.8.5.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.6 mishandles the interaction between serialization gadgets and typing, related to br.com.anteros.dbcp.AnterosDBCPDataSource (aka Anteros-DBCP). <p>Publish Date: 2020-08-25 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-24616>CVE-2020-24616</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-24616">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-24616</a></p> <p>Release Date: 2020-08-25</p> <p>Fix Resolution: 2.9.10.6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-24616 (High) detected in jackson-databind-2.8.5.jar - ## CVE-2020-24616 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.5.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /ClamScanService/build.gradle</p> <p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.8.5/b3035f37e674c04dafe36a660c3815cc59f764e2/jackson-databind-2.8.5.jar,/root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.5/jackson-databind-2.8.5.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-thymeleaf-1.4.3.RELEASE.jar (Root Library) - spring-boot-starter-web-1.4.3.RELEASE.jar - :x: **jackson-databind-2.8.5.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.6 mishandles the interaction between serialization gadgets and typing, related to br.com.anteros.dbcp.AnterosDBCPDataSource (aka Anteros-DBCP). <p>Publish Date: 2020-08-25 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-24616>CVE-2020-24616</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-24616">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-24616</a></p> <p>Release Date: 2020-08-25</p> <p>Fix Resolution: 2.9.10.6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file clamscanservice build gradle path to vulnerable library root gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar root repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring boot starter thymeleaf release jar root library spring boot starter web release jar x jackson databind jar vulnerable library vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to br com anteros dbcp anterosdbcpdatasource aka anteros dbcp publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
5,152
7,931,506,221
IssuesEvent
2018-07-07 00:57:27
ncbo/bioportal-project
https://api.github.com/repos/ncbo/bioportal-project
closed
ONTONEO: last 4 submissions failed to parse
ontology processing problem
The lastest 4 submissions for [ONTONEO](http://bioportal.bioontology.org/ontologies/ONTONEO) failed to parse. Error from parsing.log file: ``` E, [2017-08-03T19:13:05.004447 #2608] ERROR -- : ["Exception: Rapper cannot parse rdfxml file at /srv/ncbo/repository/ONTONEO/10/owlapi.xrdf: rapper: Parsing URI file:///srv/ncbo/repository/ONTONEO/10/owlapi.xrdf with parser rdfxml rapper: Serializing with serializer ntriples rapper: Warning - URI file:///srv/ncbo/repository/ONTONEO/10/owlapi.xrdf:21075 - rdf:comment is an unknown RDF namespaced element. rapper: Warning - URI file:///srv/ncbo/repository/ONTONEO/10/owlapi.xrdf:21095 - rdf:comment is an unknown RDF namespaced element. rapper: Warning - URI file:///srv/ncbo/repository/ONTONEO/10/owlapi.xrdf:21217 - rdf:comment is an unknown RDF namespaced element. rapper: Warning - URI file:///srv/ncbo/repository/ONTONEO/10/owlapi.xrdf:21371 - rdf:comment is an unknown RDF namespaced element. rapper: Warning - URI file:///srv/ncbo/repository/ONTONEO/10/owlapi.xrdf:21374 - rdf:comment is an unknown RDF namespaced element. rapper: Warning - URI file:///srv/ncbo/repository/ONTONEO/10/owlapi.xrdf:38750 - rdf:comment is an unknown RDF namespaced element. rapper: Warning - URI file:///srv/ncbo/repository/ONTONEO/10/owlapi.xrdf:38751 - rdf:comment is an unknown RDF namespaced element. rapper: Warning - URI file:///srv/ncbo/repository/ONTONEO/10/owlapi.xrdf:41892 - rdf:comment is an unknown RDF namespaced element. rapper: Warning - URI file:///srv/ncbo/repository/ONTONEO/10/owlapi.xrdf:42356 - rdf:comment is an unknown RDF namespaced element. rapper: Warning - URI file:///srv/ncbo/repository/ONTONEO/10/owlapi.xrdf:42374 - rdf:comment is an unknown RDF namespaced element. rapper: Warning - URI file:///srv/ncbo/repository/ONTONEO/10/owlapi.xrdf:42390 - rdf:comment is an unknown RDF namespaced element. rapper: Warning - URI file:///srv/ncbo/repository/ONTONEO/10/owlapi.xrdf:42446 - rdf:comment is an unknown RDF namespaced element. rapper: Warning - URI file:///srv/ncbo/repository/ONTONEO/10/owlapi.xrdf:42447 - rdf:comment is an unknown RDF namespaced element. rapper: Warning - URI file:///srv/ncbo/repository/ONTONEO/10/owlapi.xrdf:42489 - rdf:comment is an unknown RDF namespaced element. rapper: Warning - URI file:///srv/ncbo/repository/ONTONEO/10/owlapi.xrdf:42490 - rdf:comment is an unknown RDF namespaced element. rapper: Warning - URI file:///srv/ncbo/repository/ONTONEO/10/owlapi.xrdf:42518 - rdf:comment is an unknown RDF namespaced element. rapper: Warning - URI file:///srv/ncbo/repository/ONTONEO/10/owlapi.xrdf:42682 - rdf:comment is an unknown RDF namespaced element. rapper: Parsing returned 28845 triples /srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.3.0/bundler/gems/goo-dd1b78375bae/lib/goo/sparql/client.rb:59:in `bnodes_filter_file' /srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.3.0/bundler/gems/goo-dd1b78375bae/lib/goo/sparql/client.rb:80:in `append_triples_no_bnodes' /srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.3.0/bundler/gems/goo-dd1b78375bae/lib/goo/sparql/client.rb:128:in `put_triples' /srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.3.0/bundler/gems/ontologies_linked_data-5d17205300d9/lib/ontologies_linked_data/models/ontology_submission.rb:1415:in `delete_and_append' /srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.3.0/bundler/gems/ontologies_linked_data-5d17205300d9/lib/ontologies_linked_data/models/ontology_submission.rb:438:in `generate_rdf' /srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.3.0/bundler/gems/ontologies_linked_data-5d17205300d9/lib/ontologies_linked_data/models/ontology_submission.rb:892:in `process_submission' /srv/ncbo/ncbo_cron/lib/ncbo_cron/ontology_submission_parser.rb:177:in `process_submission' /srv/ncbo/ncbo_cron/lib/ncbo_cron/ontology_submission_parser.rb:47:in `block in process_queue_submissions' /srv/ncbo/ncbo_cron/lib/ncbo_cron/ontology_submission_parser.rb:41:in `each' /srv/ncbo/ncbo_cron/lib/ncbo_cron/ontology_submission_parser.rb:41:in `process_queue_submissions' /srv/ncbo/ncbo_cron/bin/ncbo_cron:228:in `block (3 levels) in <main>' /srv/ncbo/ncbo_cron/lib/ncbo_cron/scheduler.rb:65:in `block (3 levels) in scheduled_locking_job' /srv/ncbo/ncbo_cron/lib/ncbo_cron/scheduler.rb:51:in `fork' /srv/ncbo/ncbo_cron/lib/ncbo_cron/scheduler.rb:51:in `block (2 levels) in scheduled_locking_job' /srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.3.0/gems/mlanett-redis-lock-0.2.7/lib/redis-lock.rb:43:in `lock' /srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.3.0/gems/mlanett-redis-lock-0.2.7/lib/redis-lock.rb:234:in `lock' /srv/ncbo/ncbo_cron/lib/ncbo_cron/scheduler.rb:50:in `block in scheduled_locking_job' /srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.3.0/gems/rufus-scheduler-2.0.24/lib/rufus/sc/jobs.rb:230:in `trigger_block' /srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.3.0/gems/rufus-scheduler-2.0.24/lib/rufus/sc/jobs.rb:204:in `block in trigger' /srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.3.0/gems/rufus-scheduler-2.0.24/lib/rufus/sc/scheduler.rb:430:in `block in trigger_job'"] ```
1.0
ONTONEO: last 4 submissions failed to parse - The lastest 4 submissions for [ONTONEO](http://bioportal.bioontology.org/ontologies/ONTONEO) failed to parse. Error from parsing.log file: ``` E, [2017-08-03T19:13:05.004447 #2608] ERROR -- : ["Exception: Rapper cannot parse rdfxml file at /srv/ncbo/repository/ONTONEO/10/owlapi.xrdf: rapper: Parsing URI file:///srv/ncbo/repository/ONTONEO/10/owlapi.xrdf with parser rdfxml rapper: Serializing with serializer ntriples rapper: Warning - URI file:///srv/ncbo/repository/ONTONEO/10/owlapi.xrdf:21075 - rdf:comment is an unknown RDF namespaced element. rapper: Warning - URI file:///srv/ncbo/repository/ONTONEO/10/owlapi.xrdf:21095 - rdf:comment is an unknown RDF namespaced element. rapper: Warning - URI file:///srv/ncbo/repository/ONTONEO/10/owlapi.xrdf:21217 - rdf:comment is an unknown RDF namespaced element. rapper: Warning - URI file:///srv/ncbo/repository/ONTONEO/10/owlapi.xrdf:21371 - rdf:comment is an unknown RDF namespaced element. rapper: Warning - URI file:///srv/ncbo/repository/ONTONEO/10/owlapi.xrdf:21374 - rdf:comment is an unknown RDF namespaced element. rapper: Warning - URI file:///srv/ncbo/repository/ONTONEO/10/owlapi.xrdf:38750 - rdf:comment is an unknown RDF namespaced element. rapper: Warning - URI file:///srv/ncbo/repository/ONTONEO/10/owlapi.xrdf:38751 - rdf:comment is an unknown RDF namespaced element. rapper: Warning - URI file:///srv/ncbo/repository/ONTONEO/10/owlapi.xrdf:41892 - rdf:comment is an unknown RDF namespaced element. rapper: Warning - URI file:///srv/ncbo/repository/ONTONEO/10/owlapi.xrdf:42356 - rdf:comment is an unknown RDF namespaced element. rapper: Warning - URI file:///srv/ncbo/repository/ONTONEO/10/owlapi.xrdf:42374 - rdf:comment is an unknown RDF namespaced element. rapper: Warning - URI file:///srv/ncbo/repository/ONTONEO/10/owlapi.xrdf:42390 - rdf:comment is an unknown RDF namespaced element. rapper: Warning - URI file:///srv/ncbo/repository/ONTONEO/10/owlapi.xrdf:42446 - rdf:comment is an unknown RDF namespaced element. rapper: Warning - URI file:///srv/ncbo/repository/ONTONEO/10/owlapi.xrdf:42447 - rdf:comment is an unknown RDF namespaced element. rapper: Warning - URI file:///srv/ncbo/repository/ONTONEO/10/owlapi.xrdf:42489 - rdf:comment is an unknown RDF namespaced element. rapper: Warning - URI file:///srv/ncbo/repository/ONTONEO/10/owlapi.xrdf:42490 - rdf:comment is an unknown RDF namespaced element. rapper: Warning - URI file:///srv/ncbo/repository/ONTONEO/10/owlapi.xrdf:42518 - rdf:comment is an unknown RDF namespaced element. rapper: Warning - URI file:///srv/ncbo/repository/ONTONEO/10/owlapi.xrdf:42682 - rdf:comment is an unknown RDF namespaced element. rapper: Parsing returned 28845 triples /srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.3.0/bundler/gems/goo-dd1b78375bae/lib/goo/sparql/client.rb:59:in `bnodes_filter_file' /srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.3.0/bundler/gems/goo-dd1b78375bae/lib/goo/sparql/client.rb:80:in `append_triples_no_bnodes' /srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.3.0/bundler/gems/goo-dd1b78375bae/lib/goo/sparql/client.rb:128:in `put_triples' /srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.3.0/bundler/gems/ontologies_linked_data-5d17205300d9/lib/ontologies_linked_data/models/ontology_submission.rb:1415:in `delete_and_append' /srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.3.0/bundler/gems/ontologies_linked_data-5d17205300d9/lib/ontologies_linked_data/models/ontology_submission.rb:438:in `generate_rdf' /srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.3.0/bundler/gems/ontologies_linked_data-5d17205300d9/lib/ontologies_linked_data/models/ontology_submission.rb:892:in `process_submission' /srv/ncbo/ncbo_cron/lib/ncbo_cron/ontology_submission_parser.rb:177:in `process_submission' /srv/ncbo/ncbo_cron/lib/ncbo_cron/ontology_submission_parser.rb:47:in `block in process_queue_submissions' /srv/ncbo/ncbo_cron/lib/ncbo_cron/ontology_submission_parser.rb:41:in `each' /srv/ncbo/ncbo_cron/lib/ncbo_cron/ontology_submission_parser.rb:41:in `process_queue_submissions' /srv/ncbo/ncbo_cron/bin/ncbo_cron:228:in `block (3 levels) in <main>' /srv/ncbo/ncbo_cron/lib/ncbo_cron/scheduler.rb:65:in `block (3 levels) in scheduled_locking_job' /srv/ncbo/ncbo_cron/lib/ncbo_cron/scheduler.rb:51:in `fork' /srv/ncbo/ncbo_cron/lib/ncbo_cron/scheduler.rb:51:in `block (2 levels) in scheduled_locking_job' /srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.3.0/gems/mlanett-redis-lock-0.2.7/lib/redis-lock.rb:43:in `lock' /srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.3.0/gems/mlanett-redis-lock-0.2.7/lib/redis-lock.rb:234:in `lock' /srv/ncbo/ncbo_cron/lib/ncbo_cron/scheduler.rb:50:in `block in scheduled_locking_job' /srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.3.0/gems/rufus-scheduler-2.0.24/lib/rufus/sc/jobs.rb:230:in `trigger_block' /srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.3.0/gems/rufus-scheduler-2.0.24/lib/rufus/sc/jobs.rb:204:in `block in trigger' /srv/ncbo/ncbo_cron/vendor/bundle/ruby/2.3.0/gems/rufus-scheduler-2.0.24/lib/rufus/sc/scheduler.rb:430:in `block in trigger_job'"] ```
process
ontoneo last submissions failed to parse the lastest submissions for failed to parse error from parsing log file e error exception rapper cannot parse rdfxml file at srv ncbo repository ontoneo owlapi xrdf rapper parsing uri file srv ncbo repository ontoneo owlapi xrdf with parser rdfxml rapper serializing with serializer ntriples rapper warning uri file srv ncbo repository ontoneo owlapi xrdf rdf comment is an unknown rdf namespaced element rapper warning uri file srv ncbo repository ontoneo owlapi xrdf rdf comment is an unknown rdf namespaced element rapper warning uri file srv ncbo repository ontoneo owlapi xrdf rdf comment is an unknown rdf namespaced element rapper warning uri file srv ncbo repository ontoneo owlapi xrdf rdf comment is an unknown rdf namespaced element rapper warning uri file srv ncbo repository ontoneo owlapi xrdf rdf comment is an unknown rdf namespaced element rapper warning uri file srv ncbo repository ontoneo owlapi xrdf rdf comment is an unknown rdf namespaced element rapper warning uri file srv ncbo repository ontoneo owlapi xrdf rdf comment is an unknown rdf namespaced element rapper warning uri file srv ncbo repository ontoneo owlapi xrdf rdf comment is an unknown rdf namespaced element rapper warning uri file srv ncbo repository ontoneo owlapi xrdf rdf comment is an unknown rdf namespaced element rapper warning uri file srv ncbo repository ontoneo owlapi xrdf rdf comment is an unknown rdf namespaced element rapper warning uri file srv ncbo repository ontoneo owlapi xrdf rdf comment is an unknown rdf namespaced element rapper warning uri file srv ncbo repository ontoneo owlapi xrdf rdf comment is an unknown rdf namespaced element rapper warning uri file srv ncbo repository ontoneo owlapi xrdf rdf comment is an unknown rdf namespaced element rapper warning uri file srv ncbo repository ontoneo owlapi xrdf rdf comment is an unknown rdf namespaced element rapper warning uri file srv ncbo repository ontoneo owlapi xrdf rdf comment is an unknown rdf namespaced element rapper warning uri file srv ncbo repository ontoneo owlapi xrdf rdf comment is an unknown rdf namespaced element rapper warning uri file srv ncbo repository ontoneo owlapi xrdf rdf comment is an unknown rdf namespaced element rapper parsing returned triples srv ncbo ncbo cron vendor bundle ruby bundler gems goo lib goo sparql client rb in bnodes filter file srv ncbo ncbo cron vendor bundle ruby bundler gems goo lib goo sparql client rb in append triples no bnodes srv ncbo ncbo cron vendor bundle ruby bundler gems goo lib goo sparql client rb in put triples srv ncbo ncbo cron vendor bundle ruby bundler gems ontologies linked data lib ontologies linked data models ontology submission rb in delete and append srv ncbo ncbo cron vendor bundle ruby bundler gems ontologies linked data lib ontologies linked data models ontology submission rb in generate rdf srv ncbo ncbo cron vendor bundle ruby bundler gems ontologies linked data lib ontologies linked data models ontology submission rb in process submission srv ncbo ncbo cron lib ncbo cron ontology submission parser rb in process submission srv ncbo ncbo cron lib ncbo cron ontology submission parser rb in block in process queue submissions srv ncbo ncbo cron lib ncbo cron ontology submission parser rb in each srv ncbo ncbo cron lib ncbo cron ontology submission parser rb in process queue submissions srv ncbo ncbo cron bin ncbo cron in block levels in srv ncbo ncbo cron lib ncbo cron scheduler rb in block levels in scheduled locking job srv ncbo ncbo cron lib ncbo cron scheduler rb in fork srv ncbo ncbo cron lib ncbo cron scheduler rb in block levels in scheduled locking job srv ncbo ncbo cron vendor bundle ruby gems mlanett redis lock lib redis lock rb in lock srv ncbo ncbo cron vendor bundle ruby gems mlanett redis lock lib redis lock rb in lock srv ncbo ncbo cron lib ncbo cron scheduler rb in block in scheduled locking job srv ncbo ncbo cron vendor bundle ruby gems rufus scheduler lib rufus sc jobs rb in trigger block srv ncbo ncbo cron vendor bundle ruby gems rufus scheduler lib rufus sc jobs rb in block in trigger srv ncbo ncbo cron vendor bundle ruby gems rufus scheduler lib rufus sc scheduler rb in block in trigger job
1
20,052
6,808,668,294
IssuesEvent
2017-11-04 06:31:53
nasa/europa
https://api.github.com/repos/nasa/europa
closed
Symbol not found when launching europa.
Component-Build
#### Configuration: * OpenJDK 1.8 * ftjam 2.5.3 * GCC 7.2.0 * libantlr3c 3.5.2 I am trying to follow the quick start guide (using official binaries); When launching `ant` I have the following error: ``` Buildfile: /home/gandre/WIP/Aeroport/Planning/Light/build.xml init: compile: run: [echo] Running Light project [java] INFO: Custom library Light_g wasn't found and won't be loaded. [java] /usr/lib/jvm/java-8-openjdk/jre/bin/java: symbol lookup error: /home/gandre/WIP/Aeroport/Planning/europa//lib/libNDDL_g.so: undefined symbol: antlr3AsciiFileStreamNew BUILD FAILED /home/gandre/WIP/Aeroport/Planning/Light/build.xml:69: Java returned: 127 ``` I expect it to be because of libantrl3c version conflicts.
1.0
Symbol not found when launching europa. - #### Configuration: * OpenJDK 1.8 * ftjam 2.5.3 * GCC 7.2.0 * libantlr3c 3.5.2 I am trying to follow the quick start guide (using official binaries); When launching `ant` I have the following error: ``` Buildfile: /home/gandre/WIP/Aeroport/Planning/Light/build.xml init: compile: run: [echo] Running Light project [java] INFO: Custom library Light_g wasn't found and won't be loaded. [java] /usr/lib/jvm/java-8-openjdk/jre/bin/java: symbol lookup error: /home/gandre/WIP/Aeroport/Planning/europa//lib/libNDDL_g.so: undefined symbol: antlr3AsciiFileStreamNew BUILD FAILED /home/gandre/WIP/Aeroport/Planning/Light/build.xml:69: Java returned: 127 ``` I expect it to be because of libantrl3c version conflicts.
non_process
symbol not found when launching europa configuration openjdk ftjam gcc i am trying to follow the quick start guide using official binaries when launching ant i have the following error buildfile home gandre wip aeroport planning light build xml init compile run running light project info custom library light g wasn t found and won t be loaded usr lib jvm java openjdk jre bin java symbol lookup error home gandre wip aeroport planning europa lib libnddl g so undefined symbol build failed home gandre wip aeroport planning light build xml java returned i expect it to be because of version conflicts
0
314,486
23,524,953,295
IssuesEvent
2022-08-19 09:53:06
joshtom/josh-folio
https://api.github.com/repos/joshtom/josh-folio
opened
Working code for splitting Js in Next Js
documentation
https://github.com/shshaw/Splitting/issues/80#issuecomment-1220441652 import "splitting/dist/splitting.css"; import "splitting/dist/splitting-cells.css"; ```js const Component = () => { let target; setTimeout(() => { if ( window && document && target ) { const Splitting = require('Splitting'); Splitting({ by: "chars", target: target, }); } }); return ( <> <span ref={(el) => { target = el; }}>Lorem ipsum dolor sit amet</span> </> ) } export default Component; ```
1.0
Working code for splitting Js in Next Js - https://github.com/shshaw/Splitting/issues/80#issuecomment-1220441652 import "splitting/dist/splitting.css"; import "splitting/dist/splitting-cells.css"; ```js const Component = () => { let target; setTimeout(() => { if ( window && document && target ) { const Splitting = require('Splitting'); Splitting({ by: "chars", target: target, }); } }); return ( <> <span ref={(el) => { target = el; }}>Lorem ipsum dolor sit amet</span> </> ) } export default Component; ```
non_process
working code for splitting js in next js import splitting dist splitting css import splitting dist splitting cells css js const component let target settimeout if window document target const splitting require splitting splitting by chars target target return target el lorem ipsum dolor sit amet export default component
0
7,764
10,887,658,963
IssuesEvent
2019-11-18 14:56:09
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
opened
Merge GO:0060154 cellular process regulating host cell cycle in response to virus into GO:0060153 modulation by virus of host cell cycle
multi-species process
Merge GO:0060154 cellular process regulating host cell cycle in response to virus (Any cellular process that modulates the rate or extent of progression through the cell cycle in response to a virus.) into GO:0060153 modulation by virus of host cell cycle (Any viral process that modulates the rate or extent of progression through the cell cycle.)
1.0
Merge GO:0060154 cellular process regulating host cell cycle in response to virus into GO:0060153 modulation by virus of host cell cycle - Merge GO:0060154 cellular process regulating host cell cycle in response to virus (Any cellular process that modulates the rate or extent of progression through the cell cycle in response to a virus.) into GO:0060153 modulation by virus of host cell cycle (Any viral process that modulates the rate or extent of progression through the cell cycle.)
process
merge go cellular process regulating host cell cycle in response to virus into go modulation by virus of host cell cycle merge go cellular process regulating host cell cycle in response to virus any cellular process that modulates the rate or extent of progression through the cell cycle in response to a virus into go modulation by virus of host cell cycle any viral process that modulates the rate or extent of progression through the cell cycle
1
3,532
6,570,779,307
IssuesEvent
2017-09-10 04:24:44
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
closed
How to start the system default browser.
area-System.Diagnostics.Process bug up-for-grabs
The following line doesn't work: Process.Start("http://localhost:5000"); Microsoft said that that was the official way to start a new browser, a long time ago, at: https://support.microsoft.com/en-us/kb/305703 Is there any multiplatform way to do this now? Will it ever be?
1.0
How to start the system default browser. - The following line doesn't work: Process.Start("http://localhost:5000"); Microsoft said that that was the official way to start a new browser, a long time ago, at: https://support.microsoft.com/en-us/kb/305703 Is there any multiplatform way to do this now? Will it ever be?
process
how to start the system default browser the following line doesn t work process start microsoft said that that was the official way to start a new browser a long time ago at is there any multiplatform way to do this now will it ever be
1
76,128
3,481,800,847
IssuesEvent
2015-12-29 18:34:24
phetsims/tasks
https://api.github.com/repos/phetsims/tasks
opened
Test Java version of Isotopes and Atomic mass
Medium Priority QA
@bryo5363 can you test the Java version of Isotopes and Atomic Mass. We are porting this sim, and it would be good to identify any existing bugs.
1.0
Test Java version of Isotopes and Atomic mass - @bryo5363 can you test the Java version of Isotopes and Atomic Mass. We are porting this sim, and it would be good to identify any existing bugs.
non_process
test java version of isotopes and atomic mass can you test the java version of isotopes and atomic mass we are porting this sim and it would be good to identify any existing bugs
0
30,451
8,551,722,782
IssuesEvent
2018-11-07 18:53:33
servo/servo
https://api.github.com/repos/servo/servo
closed
Taskcluster builds fail in mozjs bindgen build step
A-build A-infrastructure
``` error: unknown argument: '-fno-sized-deallocation', err: true ``` Maybe we need to have a newer version of clang in our taskcluster setup?
1.0
Taskcluster builds fail in mozjs bindgen build step - ``` error: unknown argument: '-fno-sized-deallocation', err: true ``` Maybe we need to have a newer version of clang in our taskcluster setup?
non_process
taskcluster builds fail in mozjs bindgen build step error unknown argument fno sized deallocation err true maybe we need to have a newer version of clang in our taskcluster setup
0
324,062
9,883,342,285
IssuesEvent
2019-06-24 19:09:21
d2r2/go-dht
https://api.github.com/repos/d2r2/go-dht
closed
Decimal part of the temperature / humidity value for the DHT11
Priority: Medium Status: Pending Type: Maintenance
in the line 184, 185 of dht.go. why don't add the decimal part to the value? Based on the datasheet of DHT11, it supports the decimal part.
1.0
Decimal part of the temperature / humidity value for the DHT11 - in the line 184, 185 of dht.go. why don't add the decimal part to the value? Based on the datasheet of DHT11, it supports the decimal part.
non_process
decimal part of the temperature humidity value for the in the line of dht go why don t add the decimal part to the value based on the datasheet of it supports the decimal part
0
7,426
10,545,409,578
IssuesEvent
2019-10-02 19:04:20
openopps/openopps-platform
https://api.github.com/repos/openopps/openopps-platform
opened
DoS: Update text on next steps page
Apply Process State Dept.
We have already pulled USAJOBS data by now to remove that reference
1.0
DoS: Update text on next steps page - We have already pulled USAJOBS data by now to remove that reference
process
dos update text on next steps page we have already pulled usajobs data by now to remove that reference
1
7,696
10,780,639,921
IssuesEvent
2019-11-04 13:25:30
radis/radis
https://api.github.com/repos/radis/radis
closed
remove medium
interface post-process refactor
combination of `nm` + medium='air' or 'vacuum' is confusing I suggest discarding the `medium` parameter and using, as wavespace: `cm-1` `nm_air` `nm_vacuum` Question remains: what would the default `nm` points to?
1.0
remove medium - combination of `nm` + medium='air' or 'vacuum' is confusing I suggest discarding the `medium` parameter and using, as wavespace: `cm-1` `nm_air` `nm_vacuum` Question remains: what would the default `nm` points to?
process
remove medium combination of nm medium air or vacuum is confusing i suggest discarding the medium parameter and using as wavespace cm nm air nm vacuum question remains what would the default nm points to
1
16,266
20,862,996,560
IssuesEvent
2022-03-22 02:09:26
streamnative/flink
https://api.github.com/repos/streamnative/flink
closed
[Feature] Support Pulsar schema evolution in PulsarSink
compute/data-processing type/feature
Pulsar has a built-in schema and supports schema evolution out of the box. We should add this support in the Pulsar sink.
1.0
[Feature] Support Pulsar schema evolution in PulsarSink - Pulsar has a built-in schema and supports schema evolution out of the box. We should add this support in the Pulsar sink.
process
support pulsar schema evolution in pulsarsink pulsar has a built in schema and supports schema evolution out of the box we should add this support in the pulsar sink
1
208,358
15,887,115,901
IssuesEvent
2021-04-10 00:48:59
backend-br/vagas
https://api.github.com/repos/backend-br/vagas
closed
[REMOTO] Desenvolvedor(a) Python @ Instruct Soluções em Tecnologia
CLT Django Docker Git GraphQL Linux Python Remoto Rest SQL Stale Testes Unitários
## Descrição da vaga Aqui na Instruct nossos times são multidisciplinares e valorizamos a troca de experiência dos colaboradores entre as diversas áreas. Sabemos que os resultados dependem do trabalho em equipe! Prezamos por um ambiente amigável, inclusivo e seguro e estamos sempre abertos e dispostos a contribuir com a comunidade. Acreditamos na cultura blameless como forma de melhoria constante dos nossos processos, identificando rapidamente as falhas para corrigi-las, sem deixar de lado as responsabilidades inerentes de cada função. Compartilhamos as informações e o conhecimento diariamente e acima de tudo, valorizamos a transparência, respeito e honestidade nos relacionamentos. Mesmo de longe! Sim! Somos 100% remotos! Mas estamos juntos todos os dias. Venha provar um café de qualidade! Quer saber como? Então participe da nossa equipe! O que esperamos de você? Esperamos que você entenda problemas em sistemas que já tem familiaridade e possa ir além, propondo, criando, mudando, construindo, crescendo e evoluindo conosco. Que adquira visão da estratégia de produto e deseje colaborar com outros membros da equipe. Queremos que você aplique seus conhecimentos a cada nova tarefa e reconheça áreas que não tem entendimento para resolver um problema buscando ajuda do time. Queremos que você participe de planejamentos e estime o tempo de suas atividades, compartilhando responsabilidades, acertos e aprendizados. Remuneração e benefícios: Regime de contratação CLT. Salário inicial entre R$ 2.500,00 e R$ 4.000,00. Plano de saúde e odontológico Sul América Saúde para o titular e seus dependentes. Vale refeição de R$ 400,00. Vale alimentação de R$ 600,00. Seguro de vida corporativo custeado pela empresa. Verbas para investimento em treinamentos e educação. Acesso aos treinamentos da empresa sem custos. Prêmio de participação nos resultados. Prêmio quando se completa aniversário de empresa. A empresa fornecerá um laptop com as seguintes configurações: Dell Inspiron 14 5480 (Core i7-8565U, RAM 16GB, GeForce MX150, HDD 1TB, SSD 128GB, Full HD). Informações adicionais: O trabalho é 100% home office, dentro do horário comercial e seguindo o horário de Brasília, com viagens ocasionais custeadas pela empresa. Portanto, é fundamental ter disponibilidade para viajar. Horário comercial - das 9h às 18h (40h semanais). ## Local REMOTO - São Paulo ## Requisitos Conhecimentos e experiências fundamentais: De 1 a 3 anos de experiência em desenvolvimento de software. Linguagem Python 3 e bons conhecimentos sobre orientação a objeto. Fundamentos de banco de dados relacional/SQL. Fundamentos de sistemas operacionais e Linux. Comunicação via APIs REST. Uso de Git para versionamento de seu código. Conhecimentos diferenciais: Django. Docker. Testes unitários. GraphQL. ## Contratação CLT ## Como se candidatar Se inscreva [clicando aqui](https://www.pyjobs.com.br/job/2044) ## Labels
1.0
[REMOTO] Desenvolvedor(a) Python @ Instruct Soluções em Tecnologia - ## Descrição da vaga Aqui na Instruct nossos times são multidisciplinares e valorizamos a troca de experiência dos colaboradores entre as diversas áreas. Sabemos que os resultados dependem do trabalho em equipe! Prezamos por um ambiente amigável, inclusivo e seguro e estamos sempre abertos e dispostos a contribuir com a comunidade. Acreditamos na cultura blameless como forma de melhoria constante dos nossos processos, identificando rapidamente as falhas para corrigi-las, sem deixar de lado as responsabilidades inerentes de cada função. Compartilhamos as informações e o conhecimento diariamente e acima de tudo, valorizamos a transparência, respeito e honestidade nos relacionamentos. Mesmo de longe! Sim! Somos 100% remotos! Mas estamos juntos todos os dias. Venha provar um café de qualidade! Quer saber como? Então participe da nossa equipe! O que esperamos de você? Esperamos que você entenda problemas em sistemas que já tem familiaridade e possa ir além, propondo, criando, mudando, construindo, crescendo e evoluindo conosco. Que adquira visão da estratégia de produto e deseje colaborar com outros membros da equipe. Queremos que você aplique seus conhecimentos a cada nova tarefa e reconheça áreas que não tem entendimento para resolver um problema buscando ajuda do time. Queremos que você participe de planejamentos e estime o tempo de suas atividades, compartilhando responsabilidades, acertos e aprendizados. Remuneração e benefícios: Regime de contratação CLT. Salário inicial entre R$ 2.500,00 e R$ 4.000,00. Plano de saúde e odontológico Sul América Saúde para o titular e seus dependentes. Vale refeição de R$ 400,00. Vale alimentação de R$ 600,00. Seguro de vida corporativo custeado pela empresa. Verbas para investimento em treinamentos e educação. Acesso aos treinamentos da empresa sem custos. Prêmio de participação nos resultados. Prêmio quando se completa aniversário de empresa. A empresa fornecerá um laptop com as seguintes configurações: Dell Inspiron 14 5480 (Core i7-8565U, RAM 16GB, GeForce MX150, HDD 1TB, SSD 128GB, Full HD). Informações adicionais: O trabalho é 100% home office, dentro do horário comercial e seguindo o horário de Brasília, com viagens ocasionais custeadas pela empresa. Portanto, é fundamental ter disponibilidade para viajar. Horário comercial - das 9h às 18h (40h semanais). ## Local REMOTO - São Paulo ## Requisitos Conhecimentos e experiências fundamentais: De 1 a 3 anos de experiência em desenvolvimento de software. Linguagem Python 3 e bons conhecimentos sobre orientação a objeto. Fundamentos de banco de dados relacional/SQL. Fundamentos de sistemas operacionais e Linux. Comunicação via APIs REST. Uso de Git para versionamento de seu código. Conhecimentos diferenciais: Django. Docker. Testes unitários. GraphQL. ## Contratação CLT ## Como se candidatar Se inscreva [clicando aqui](https://www.pyjobs.com.br/job/2044) ## Labels
non_process
desenvolvedor a python instruct soluções em tecnologia descrição da vaga aqui na instruct nossos times são multidisciplinares e valorizamos a troca de experiência dos colaboradores entre as diversas áreas sabemos que os resultados dependem do trabalho em equipe prezamos por um ambiente amigável inclusivo e seguro e estamos sempre abertos e dispostos a contribuir com a comunidade acreditamos na cultura blameless como forma de melhoria constante dos nossos processos identificando rapidamente as falhas para corrigi las sem deixar de lado as responsabilidades inerentes de cada função compartilhamos as informações e o conhecimento diariamente e acima de tudo valorizamos a transparência respeito e honestidade nos relacionamentos mesmo de longe sim somos remotos mas estamos juntos todos os dias venha provar um café de qualidade quer saber como então participe da nossa equipe o que esperamos de você esperamos que você entenda problemas em sistemas que já tem familiaridade e possa ir além propondo criando mudando construindo crescendo e evoluindo conosco que adquira visão da estratégia de produto e deseje colaborar com outros membros da equipe queremos que você aplique seus conhecimentos a cada nova tarefa e reconheça áreas que não tem entendimento para resolver um problema buscando ajuda do time queremos que você participe de planejamentos e estime o tempo de suas atividades compartilhando responsabilidades acertos e aprendizados remuneração e benefícios regime de contratação clt salário inicial entre r e r plano de saúde e odontológico sul américa saúde para o titular e seus dependentes vale refeição de r vale alimentação de r seguro de vida corporativo custeado pela empresa verbas para investimento em treinamentos e educação acesso aos treinamentos da empresa sem custos prêmio de participação nos resultados prêmio quando se completa aniversário de empresa a empresa fornecerá um laptop com as seguintes configurações dell inspiron core ram geforce hdd ssd full hd informações adicionais o trabalho é home office dentro do horário comercial e seguindo o horário de brasília com viagens ocasionais custeadas pela empresa portanto é fundamental ter disponibilidade para viajar horário comercial das às semanais local remoto são paulo requisitos conhecimentos e experiências fundamentais de a anos de experiência em desenvolvimento de software linguagem python e bons conhecimentos sobre orientação a objeto fundamentos de banco de dados relacional sql fundamentos de sistemas operacionais e linux comunicação via apis rest uso de git para versionamento de seu código conhecimentos diferenciais django docker testes unitários graphql contratação clt como se candidatar se inscreva labels
0
17,361
23,185,583,565
IssuesEvent
2022-08-01 08:06:11
streamnative/flink
https://api.github.com/repos/streamnative/flink
closed
[SQL Connector] fix the TimeoutException of PulsarSourceITCase
compute/data-processing
PulsarSourceITCase>SourceTestSuiteBase.testSavepoint:236->SourceTestSuiteBase.restartFromSavepoint:388->SourceTestSuiteBase.checkResultWithSemantic:744 is failing, we can see if increase the timeout value would fix the problem. https://github.com/streamnative/streamnative-ci/runs/7378238879?check_suite_focus=true
1.0
[SQL Connector] fix the TimeoutException of PulsarSourceITCase - PulsarSourceITCase>SourceTestSuiteBase.testSavepoint:236->SourceTestSuiteBase.restartFromSavepoint:388->SourceTestSuiteBase.checkResultWithSemantic:744 is failing, we can see if increase the timeout value would fix the problem. https://github.com/streamnative/streamnative-ci/runs/7378238879?check_suite_focus=true
process
fix the timeoutexception of pulsarsourceitcase pulsarsourceitcase sourcetestsuitebase testsavepoint sourcetestsuitebase restartfromsavepoint sourcetestsuitebase checkresultwithsemantic is failing we can see if increase the timeout value would fix the problem
1
448,510
12,951,848,795
IssuesEvent
2020-07-19 18:18:27
rtc-focus-org/focus_issues
https://api.github.com/repos/rtc-focus-org/focus_issues
closed
Firefox needs dropdown to choose sharing type
enhancement priority screenshare
_From @rtc-focus on June 19, 2018 18:30_ Firefox needs to be told in advance of requesting screen sharing whether full screen or a window is to shared. So to support Firefox properly we need to present this choice after user clicks screen share button. _Copied from original issue: rtc-focus-org/focus#1_
1.0
Firefox needs dropdown to choose sharing type - _From @rtc-focus on June 19, 2018 18:30_ Firefox needs to be told in advance of requesting screen sharing whether full screen or a window is to shared. So to support Firefox properly we need to present this choice after user clicks screen share button. _Copied from original issue: rtc-focus-org/focus#1_
non_process
firefox needs dropdown to choose sharing type from rtc focus on june firefox needs to be told in advance of requesting screen sharing whether full screen or a window is to shared so to support firefox properly we need to present this choice after user clicks screen share button copied from original issue rtc focus org focus
0
72,528
7,300,672,537
IssuesEvent
2018-02-27 00:52:52
littlevgl/lvgl
https://api.github.com/repos/littlevgl/lvgl
closed
Input pointer drag on empty container causes invalidation/refresh
bug need test
Easiest way to see and replicate this issue is to run the theme demo. Go to Tab 3 - everything on this page fits on screen, so there is no vertical scrollbar. If you drag in any direction in the empty area to the right of the floating windows (in the area where there are no objects to interact with) the result is essentially a refresh of the entire screen. You can see this on Tab 2 as well, if you drag in the area between the graph/text/keyboard. The expected behavior would be no areas redrawn, since no objects were interacted with/nothing actually changed on screen.
1.0
Input pointer drag on empty container causes invalidation/refresh - Easiest way to see and replicate this issue is to run the theme demo. Go to Tab 3 - everything on this page fits on screen, so there is no vertical scrollbar. If you drag in any direction in the empty area to the right of the floating windows (in the area where there are no objects to interact with) the result is essentially a refresh of the entire screen. You can see this on Tab 2 as well, if you drag in the area between the graph/text/keyboard. The expected behavior would be no areas redrawn, since no objects were interacted with/nothing actually changed on screen.
non_process
input pointer drag on empty container causes invalidation refresh easiest way to see and replicate this issue is to run the theme demo go to tab everything on this page fits on screen so there is no vertical scrollbar if you drag in any direction in the empty area to the right of the floating windows in the area where there are no objects to interact with the result is essentially a refresh of the entire screen you can see this on tab as well if you drag in the area between the graph text keyboard the expected behavior would be no areas redrawn since no objects were interacted with nothing actually changed on screen
0
11,197
13,957,702,554
IssuesEvent
2020-10-24 08:13:38
alexanderkotsev/geoportal
https://api.github.com/repos/alexanderkotsev/geoportal
opened
MT - MITA: Harvest
Geoportal Harvesting process MT - Malta
Good Morning Angelo, Kindly can you perform a harvest on the Maltese CSW please as we did some changes and would like to see the result. Thanks, Rene
1.0
MT - MITA: Harvest - Good Morning Angelo, Kindly can you perform a harvest on the Maltese CSW please as we did some changes and would like to see the result. Thanks, Rene
process
mt mita harvest good morning angelo kindly can you perform a harvest on the maltese csw please as we did some changes and would like to see the result thanks rene
1
5,285
8,071,955,383
IssuesEvent
2018-08-06 14:38:44
dita-ot/dita-ot
https://api.github.com/repos/dita-ot/dita-ot
closed
NPE in TopicFragmentFilter module [DITA OT 2.x]
P2 bug preprocess preprocess/conref
DITA Map: ``` xml <!DOCTYPE map PUBLIC "-//OASIS//DTD DITA Map//EN" "map.dtd"> <map> <title>DITA Topic Map</title> <keydef href="zuzu.dita" keys="zuzu"/> <topicref href="test.dita"/> </map> ``` zuzu.dita: ``` xml <!DOCTYPE dita PUBLIC "-//OASIS//DTD DITA Composite//EN" "ditabase.dtd"> <dita> <topic id="a"> <title>A</title> </topic> <topic id="Z"> <title>A</title> </topic> </dita> ``` test.dita: ``` xml <!DOCTYPE topic PUBLIC "-//OASIS//DTD DITA Topic//EN" "topic.dtd"> <topic id="topic_fyk_zzc_x5"> <title>AAA</title> <body/> <topic id="id_kq2_11d_x5" conkeyref="zuzu" conrefend="default.dita#default/Z"> <title/> </topic> </topic> ``` NPE when publishing: ``` BUILD FAILED D:\projects\eXml\frameworks\dita\DITA-OT2.x\build.xml:41: The following error occurred while executing this line: D:\projects\eXml\frameworks\dita\DITA-OT2.x\plugins\org.dita.base\build_preprocess.xml:226: java.lang.NullPointerException at java.util.ArrayDeque.addFirst(Unknown Source) at org.dita.dost.writer.TopicFragmentFilter.startElement(TopicFragmentFilter.java:48) at org.apache.xerces.parsers.AbstractSAXParser.startElement(Unknown Source) at org.ditang.relaxng.defaults.RelaxNGDefaultsComponent.startElement(RelaxNGDefaultsComponent.java:200) at org.apache.xerces.impl.XMLNSDocumentScannerImpl.scanStartElement(XMLNSDocumentScannerImpl.java:268) at org.apache.xerces.impl.XMLDocumentFragmentScannerImpl$FragmentContentDispatcher.dispatch(XMLDocumentFragmentScannerImpl.java:1655) at org.apache.xerces.impl.XMLDocumentFragmentScannerImpl.scanDocument(XMLDocumentFragmentScannerImpl.java:325) ```
2.0
NPE in TopicFragmentFilter module [DITA OT 2.x] - DITA Map: ``` xml <!DOCTYPE map PUBLIC "-//OASIS//DTD DITA Map//EN" "map.dtd"> <map> <title>DITA Topic Map</title> <keydef href="zuzu.dita" keys="zuzu"/> <topicref href="test.dita"/> </map> ``` zuzu.dita: ``` xml <!DOCTYPE dita PUBLIC "-//OASIS//DTD DITA Composite//EN" "ditabase.dtd"> <dita> <topic id="a"> <title>A</title> </topic> <topic id="Z"> <title>A</title> </topic> </dita> ``` test.dita: ``` xml <!DOCTYPE topic PUBLIC "-//OASIS//DTD DITA Topic//EN" "topic.dtd"> <topic id="topic_fyk_zzc_x5"> <title>AAA</title> <body/> <topic id="id_kq2_11d_x5" conkeyref="zuzu" conrefend="default.dita#default/Z"> <title/> </topic> </topic> ``` NPE when publishing: ``` BUILD FAILED D:\projects\eXml\frameworks\dita\DITA-OT2.x\build.xml:41: The following error occurred while executing this line: D:\projects\eXml\frameworks\dita\DITA-OT2.x\plugins\org.dita.base\build_preprocess.xml:226: java.lang.NullPointerException at java.util.ArrayDeque.addFirst(Unknown Source) at org.dita.dost.writer.TopicFragmentFilter.startElement(TopicFragmentFilter.java:48) at org.apache.xerces.parsers.AbstractSAXParser.startElement(Unknown Source) at org.ditang.relaxng.defaults.RelaxNGDefaultsComponent.startElement(RelaxNGDefaultsComponent.java:200) at org.apache.xerces.impl.XMLNSDocumentScannerImpl.scanStartElement(XMLNSDocumentScannerImpl.java:268) at org.apache.xerces.impl.XMLDocumentFragmentScannerImpl$FragmentContentDispatcher.dispatch(XMLDocumentFragmentScannerImpl.java:1655) at org.apache.xerces.impl.XMLDocumentFragmentScannerImpl.scanDocument(XMLDocumentFragmentScannerImpl.java:325) ```
process
npe in topicfragmentfilter module dita map xml dita topic map zuzu dita xml a a test dita xml aaa npe when publishing build failed d projects exml frameworks dita dita x build xml the following error occurred while executing this line d projects exml frameworks dita dita x plugins org dita base build preprocess xml java lang nullpointerexception at java util arraydeque addfirst unknown source at org dita dost writer topicfragmentfilter startelement topicfragmentfilter java at org apache xerces parsers abstractsaxparser startelement unknown source at org ditang relaxng defaults relaxngdefaultscomponent startelement relaxngdefaultscomponent java at org apache xerces impl xmlnsdocumentscannerimpl scanstartelement xmlnsdocumentscannerimpl java at org apache xerces impl xmldocumentfragmentscannerimpl fragmentcontentdispatcher dispatch xmldocumentfragmentscannerimpl java at org apache xerces impl xmldocumentfragmentscannerimpl scandocument xmldocumentfragmentscannerimpl java
1
565
3,024,104,411
IssuesEvent
2015-08-02 07:59:13
HazyResearch/dd-genomics
https://api.github.com/repos/HazyResearch/dd-genomics
closed
Modify XML parser code so as to pick up years for titles
Preprocessing
Pull *all* sections of a document before outputting objects. (@Colossus- I could do a join as you suggested to compensate for the titles all having null year, but seems like this will be better long term solution to fix it here upstream / probably will take net similar amount of time. If not this should still be open as issue anyway...)
1.0
Modify XML parser code so as to pick up years for titles - Pull *all* sections of a document before outputting objects. (@Colossus- I could do a join as you suggested to compensate for the titles all having null year, but seems like this will be better long term solution to fix it here upstream / probably will take net similar amount of time. If not this should still be open as issue anyway...)
process
modify xml parser code so as to pick up years for titles pull all sections of a document before outputting objects colossus i could do a join as you suggested to compensate for the titles all having null year but seems like this will be better long term solution to fix it here upstream probably will take net similar amount of time if not this should still be open as issue anyway
1
61,778
14,640,709,879
IssuesEvent
2020-12-25 03:21:37
fu1771695yongxie/pm
https://api.github.com/repos/fu1771695yongxie/pm
opened
CVE-2018-14732 (High) detected in webpack-dev-server-1.16.5.tgz
security vulnerability
## CVE-2018-14732 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>webpack-dev-server-1.16.5.tgz</b></p></summary> <p>Serves a webpack app. Updates the browser on changes.</p> <p>Library home page: <a href="https://registry.npmjs.org/webpack-dev-server/-/webpack-dev-server-1.16.5.tgz">https://registry.npmjs.org/webpack-dev-server/-/webpack-dev-server-1.16.5.tgz</a></p> <p>Path to dependency file: pm/package.json</p> <p>Path to vulnerable library: pm/node_modules/webpack-dev-server/package.json</p> <p> Dependency Hierarchy: - :x: **webpack-dev-server-1.16.5.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/fu1771695yongxie/pm/commit/1c06cbe4c354bfe6922fec380958337d18de7e44">1c06cbe4c354bfe6922fec380958337d18de7e44</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in lib/Server.js in webpack-dev-server before 3.1.6. Attackers are able to steal developer's code because the origin of requests is not checked by the WebSocket server, which is used for HMR (Hot Module Replacement). Anyone can receive the HMR message sent by the WebSocket server via a ws://127.0.0.1:8080/ connection from any origin. <p>Publish Date: 2018-09-21 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14732>CVE-2018-14732</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-14732">https://nvd.nist.gov/vuln/detail/CVE-2018-14732</a></p> <p>Release Date: 2018-09-21</p> <p>Fix Resolution: 3.1.6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-14732 (High) detected in webpack-dev-server-1.16.5.tgz - ## CVE-2018-14732 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>webpack-dev-server-1.16.5.tgz</b></p></summary> <p>Serves a webpack app. Updates the browser on changes.</p> <p>Library home page: <a href="https://registry.npmjs.org/webpack-dev-server/-/webpack-dev-server-1.16.5.tgz">https://registry.npmjs.org/webpack-dev-server/-/webpack-dev-server-1.16.5.tgz</a></p> <p>Path to dependency file: pm/package.json</p> <p>Path to vulnerable library: pm/node_modules/webpack-dev-server/package.json</p> <p> Dependency Hierarchy: - :x: **webpack-dev-server-1.16.5.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/fu1771695yongxie/pm/commit/1c06cbe4c354bfe6922fec380958337d18de7e44">1c06cbe4c354bfe6922fec380958337d18de7e44</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in lib/Server.js in webpack-dev-server before 3.1.6. Attackers are able to steal developer's code because the origin of requests is not checked by the WebSocket server, which is used for HMR (Hot Module Replacement). Anyone can receive the HMR message sent by the WebSocket server via a ws://127.0.0.1:8080/ connection from any origin. <p>Publish Date: 2018-09-21 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14732>CVE-2018-14732</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-14732">https://nvd.nist.gov/vuln/detail/CVE-2018-14732</a></p> <p>Release Date: 2018-09-21</p> <p>Fix Resolution: 3.1.6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in webpack dev server tgz cve high severity vulnerability vulnerable library webpack dev server tgz serves a webpack app updates the browser on changes library home page a href path to dependency file pm package json path to vulnerable library pm node modules webpack dev server package json dependency hierarchy x webpack dev server tgz vulnerable library found in head commit a href found in base branch master vulnerability details an issue was discovered in lib server js in webpack dev server before attackers are able to steal developer s code because the origin of requests is not checked by the websocket server which is used for hmr hot module replacement anyone can receive the hmr message sent by the websocket server via a ws connection from any origin publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
16,701
21,802,251,838
IssuesEvent
2022-05-16 06:59:02
NationalSecurityAgency/ghidra
https://api.github.com/repos/NationalSecurityAgency/ghidra
closed
Win x64 calling convention: wrong XMM registers assignment
Feature: Processor/x86
Register assignment is incorrect for Windows x64 calling convention when a function accepts arguments of type `double` or `float`. Here is the minimal example showing the problem: ```cpp __declspec(noinline) __declspec(dllexport) int func3(int a, double b, int c, float d) { return a * b + (int)(b / d); } ``` Compile it with Visual C++ as `cl repro.cpp /LD`, then import resulting DLL into Ghidra with default settings, run default analysis. Windows x64 calling convention is described on [this documentation page](https://docs.microsoft.com/en-us/cpp/build/x64-calling-convention?view=vs-2019#parameter-passing). Pay special attention to [example 3](https://docs.microsoft.com/en-us/cpp/build/x64-calling-convention?view=vs-2019#example-of-argument-passing-3---mixed-ints-and-floats), which reads: func3(int a, double b, int c, float d); // a in RCX, b in XMM1, c in R8, d in XMM3 This is what Ghidra deduces for this function: ``` int <RETURN> EAX:4 1 int param_1 ECX:4 2 double param_2 XMM0_Qa:8 3 int param_3 EDX:4 4 float param_4 XMM1_Da:4 ``` ```cpp /* int __cdecl func3(int, double, int, float) int __cdecl func3(int,double,int,float) */ int func3(int param_1,double param_2,int param_3,float param_4) { undefined4 in_XMM1_Db; float in_XMM3_Da; /* 0x1000 1 ?func3@@YAHHNHM@Z */ return (int)((double)param_1 * (double)CONCAT44(in_XMM1_Db,param_4) + (double)(int)((double)CONCAT44(in_XMM1_Db,param_4) / (double)in_XMM3_Da)); } ``` **Environment:** - OS: Windows 7 x64 - Java Version: 11.0.5 - Ghidra Version: 9.0.4 2019-May-16 1545 EDT **P.S.** I have got used to switching all encountered functions with double arguments to "Use Custom Storage" and fixing register assignments. But now I'd like to make decompiler understand virtual function calls using the manual approach described [here](https://github.com/NationalSecurityAgency/ghidra/issues/516#issuecomment-487051576) (from #516), and it turns out that I cannot define custom register assignment for a function pointer type (reported as #1436).
1.0
Win x64 calling convention: wrong XMM registers assignment - Register assignment is incorrect for Windows x64 calling convention when a function accepts arguments of type `double` or `float`. Here is the minimal example showing the problem: ```cpp __declspec(noinline) __declspec(dllexport) int func3(int a, double b, int c, float d) { return a * b + (int)(b / d); } ``` Compile it with Visual C++ as `cl repro.cpp /LD`, then import resulting DLL into Ghidra with default settings, run default analysis. Windows x64 calling convention is described on [this documentation page](https://docs.microsoft.com/en-us/cpp/build/x64-calling-convention?view=vs-2019#parameter-passing). Pay special attention to [example 3](https://docs.microsoft.com/en-us/cpp/build/x64-calling-convention?view=vs-2019#example-of-argument-passing-3---mixed-ints-and-floats), which reads: func3(int a, double b, int c, float d); // a in RCX, b in XMM1, c in R8, d in XMM3 This is what Ghidra deduces for this function: ``` int <RETURN> EAX:4 1 int param_1 ECX:4 2 double param_2 XMM0_Qa:8 3 int param_3 EDX:4 4 float param_4 XMM1_Da:4 ``` ```cpp /* int __cdecl func3(int, double, int, float) int __cdecl func3(int,double,int,float) */ int func3(int param_1,double param_2,int param_3,float param_4) { undefined4 in_XMM1_Db; float in_XMM3_Da; /* 0x1000 1 ?func3@@YAHHNHM@Z */ return (int)((double)param_1 * (double)CONCAT44(in_XMM1_Db,param_4) + (double)(int)((double)CONCAT44(in_XMM1_Db,param_4) / (double)in_XMM3_Da)); } ``` **Environment:** - OS: Windows 7 x64 - Java Version: 11.0.5 - Ghidra Version: 9.0.4 2019-May-16 1545 EDT **P.S.** I have got used to switching all encountered functions with double arguments to "Use Custom Storage" and fixing register assignments. But now I'd like to make decompiler understand virtual function calls using the manual approach described [here](https://github.com/NationalSecurityAgency/ghidra/issues/516#issuecomment-487051576) (from #516), and it turns out that I cannot define custom register assignment for a function pointer type (reported as #1436).
process
win calling convention wrong xmm registers assignment register assignment is incorrect for windows calling convention when a function accepts arguments of type double or float here is the minimal example showing the problem cpp declspec noinline declspec dllexport int int a double b int c float d return a b int b d compile it with visual c as cl repro cpp ld then import resulting dll into ghidra with default settings run default analysis windows calling convention is described on pay special attention to which reads int a double b int c float d a in rcx b in c in d in this is what ghidra deduces for this function int eax int param ecx double param qa int param edx float param da cpp int cdecl int double int float int cdecl int double int float int int param double param int param float param in db float in da yahhnhm z return int double param double in db param double int double in db param double in da environment os windows java version ghidra version may edt p s i have got used to switching all encountered functions with double arguments to use custom storage and fixing register assignments but now i d like to make decompiler understand virtual function calls using the manual approach described from and it turns out that i cannot define custom register assignment for a function pointer type reported as
1
524,421
15,213,309,342
IssuesEvent
2021-02-17 11:36:45
nf-core/tools
https://api.github.com/repos/nf-core/tools
closed
Shields.io badge not recognised
linting low-priority
From https://shields.io/ > We support .svg and .json. The default is .svg, which can be omitted from the URL. So such a badge works: ```Markdown [![install with bioconda](https://img.shields.io/badge/install%20with-bioconda-brightgreen)](http://bioconda.github.io/) ``` [![install with bioconda](https://img.shields.io/badge/install%20with-bioconda-brightgreen)](http://bioconda.github.io/) But it's not recognised by linting. I'm guessing we're checking for the exact existence of this line: ```Markdown [![install with bioconda](https://img.shields.io/badge/install%20with-bioconda-brightgreen.svg)](http://bioconda.github.io/) ``` We could be slightly more lenient
1.0
Shields.io badge not recognised - From https://shields.io/ > We support .svg and .json. The default is .svg, which can be omitted from the URL. So such a badge works: ```Markdown [![install with bioconda](https://img.shields.io/badge/install%20with-bioconda-brightgreen)](http://bioconda.github.io/) ``` [![install with bioconda](https://img.shields.io/badge/install%20with-bioconda-brightgreen)](http://bioconda.github.io/) But it's not recognised by linting. I'm guessing we're checking for the exact existence of this line: ```Markdown [![install with bioconda](https://img.shields.io/badge/install%20with-bioconda-brightgreen.svg)](http://bioconda.github.io/) ``` We could be slightly more lenient
non_process
shields io badge not recognised from we support svg and json the default is svg which can be omitted from the url so such a badge works markdown but it s not recognised by linting i m guessing we re checking for the exact existence of this line markdown we could be slightly more lenient
0
18,045
24,055,926,521
IssuesEvent
2022-09-16 16:52:11
GregTechCEu/gt-ideas
https://api.github.com/repos/GregTechCEu/gt-ideas
opened
UDMH and Rocket Fuel Production
processing chain
## Details This processing chain is used for producing UDMH, Unsymmetrical Dimethylhydrazine (aka 1,1-dimethylhydrazine), which is used as rocket fuel. ## Products Main Product: UDMH Side Product(s): None ## Steps 2 Chlorine (g) + 6 Sodium Hydroxide (s) + Water (l) -> 1000L Salt Water (l) + 1000L Bleach (l) `2Cl (g) + 2NaOH (s) + H2O (l) -> NaCl (aq) + NaClO (aq) + 2H2O (l)` Ammonia (g) + Bleach (l) -> Monochloramine (g) + 3 Sodium Hydroxide (s) + Water (l) `NH3 (g) + NaClO (aq) + H2O (l) -> NH2Cl (g) + NaOH (s) + H2O (l)` This is the first step in the Olin-Raschig Process for production of Hydrazine. 2 Methanol (l) + Ammonia (g) -> Dimethylamine (g) + Water (l) `2CH3OH (l) + NH3 (g) -> (CH3)2NH (g) + 2H2O (l)` Monochloramine (g) + Dimethylamine (g) -> Unsymmetrical Dimethylhydrazine (l) + Hydrogen Chloride (g) `NH2Cl (g) + (CH3)2NH (g) -> (CH3)2NNH2 (l) + HCl (g)` UDMH uses NTO and IRFNA as an oxidizer in a bipropellant rockets. 2 Nitrogen Dioxide (g) -> Nitrogen Tetroxide (NTO) (l) `2NO2 (g) -> N2O4 (l)` This step is performed at below 262K. 850L Nitric Acid (aq) + 150L NTO (l) -> 1000L Red Fuming Nitric Acid (RFNA) (aq) (mixer) 990L RNFA (aq) + 10L Hydrofluoric Acid (aq) -> 1000L Inhibited Red Fuming Nitric Acid (IRFNA) (aq) (mixer) UDMH (l) + NTO (l) -> 2 Rocket Fuel (l) (mixer) UDMH (l) + IRFNA (aq) -> 3 Rocket Fuel (l) (mixer) ## Yield 1 Methanol + 0.5 Ammonia + 1 Nitrogen Dioxide -> 1 Rocket Fuel 2 Methanol + 1 Ammonia + ~2.15 Nitrogen Dioxide + ~0.85 Nitric Acid + 0.01 Hydrofluoric Acid -> 3 Rocket Fuel ## Sources https://en.wikipedia.org/wiki/Monochloramine https://en.wikipedia.org/wiki/Dimethylamine https://en.wikipedia.org/wiki/Unsymmetrical_dimethylhydrazine https://en.wikipedia.org/wiki/Dinitrogen_tetroxide https://en.wikipedia.org/wiki/Red_fuming_nitric_acid
1.0
UDMH and Rocket Fuel Production - ## Details This processing chain is used for producing UDMH, Unsymmetrical Dimethylhydrazine (aka 1,1-dimethylhydrazine), which is used as rocket fuel. ## Products Main Product: UDMH Side Product(s): None ## Steps 2 Chlorine (g) + 6 Sodium Hydroxide (s) + Water (l) -> 1000L Salt Water (l) + 1000L Bleach (l) `2Cl (g) + 2NaOH (s) + H2O (l) -> NaCl (aq) + NaClO (aq) + 2H2O (l)` Ammonia (g) + Bleach (l) -> Monochloramine (g) + 3 Sodium Hydroxide (s) + Water (l) `NH3 (g) + NaClO (aq) + H2O (l) -> NH2Cl (g) + NaOH (s) + H2O (l)` This is the first step in the Olin-Raschig Process for production of Hydrazine. 2 Methanol (l) + Ammonia (g) -> Dimethylamine (g) + Water (l) `2CH3OH (l) + NH3 (g) -> (CH3)2NH (g) + 2H2O (l)` Monochloramine (g) + Dimethylamine (g) -> Unsymmetrical Dimethylhydrazine (l) + Hydrogen Chloride (g) `NH2Cl (g) + (CH3)2NH (g) -> (CH3)2NNH2 (l) + HCl (g)` UDMH uses NTO and IRFNA as an oxidizer in a bipropellant rockets. 2 Nitrogen Dioxide (g) -> Nitrogen Tetroxide (NTO) (l) `2NO2 (g) -> N2O4 (l)` This step is performed at below 262K. 850L Nitric Acid (aq) + 150L NTO (l) -> 1000L Red Fuming Nitric Acid (RFNA) (aq) (mixer) 990L RNFA (aq) + 10L Hydrofluoric Acid (aq) -> 1000L Inhibited Red Fuming Nitric Acid (IRFNA) (aq) (mixer) UDMH (l) + NTO (l) -> 2 Rocket Fuel (l) (mixer) UDMH (l) + IRFNA (aq) -> 3 Rocket Fuel (l) (mixer) ## Yield 1 Methanol + 0.5 Ammonia + 1 Nitrogen Dioxide -> 1 Rocket Fuel 2 Methanol + 1 Ammonia + ~2.15 Nitrogen Dioxide + ~0.85 Nitric Acid + 0.01 Hydrofluoric Acid -> 3 Rocket Fuel ## Sources https://en.wikipedia.org/wiki/Monochloramine https://en.wikipedia.org/wiki/Dimethylamine https://en.wikipedia.org/wiki/Unsymmetrical_dimethylhydrazine https://en.wikipedia.org/wiki/Dinitrogen_tetroxide https://en.wikipedia.org/wiki/Red_fuming_nitric_acid
process
udmh and rocket fuel production details this processing chain is used for producing udmh unsymmetrical dimethylhydrazine aka dimethylhydrazine which is used as rocket fuel products main product udmh side product s none steps chlorine g sodium hydroxide s water l salt water l bleach l g s l nacl aq naclo aq l ammonia g bleach l monochloramine g sodium hydroxide s water l g naclo aq l g naoh s l this is the first step in the olin raschig process for production of hydrazine methanol l ammonia g dimethylamine g water l l g g l monochloramine g dimethylamine g unsymmetrical dimethylhydrazine l hydrogen chloride g g g l hcl g udmh uses nto and irfna as an oxidizer in a bipropellant rockets nitrogen dioxide g nitrogen tetroxide nto l g l this step is performed at below nitric acid aq nto l red fuming nitric acid rfna aq mixer rnfa aq hydrofluoric acid aq inhibited red fuming nitric acid irfna aq mixer udmh l nto l rocket fuel l mixer udmh l irfna aq rocket fuel l mixer yield methanol ammonia nitrogen dioxide rocket fuel methanol ammonia nitrogen dioxide nitric acid hydrofluoric acid rocket fuel sources
1
4,063
6,995,392,342
IssuesEvent
2017-12-15 19:06:15
syndesisio/syndesis
https://api.github.com/repos/syndesisio/syndesis
opened
Solidify Definitions and Expectations of Labels and Statuses
cat/discussion cat/process cat/retro
We discussed a few issues that are related to the agile methodology and how we follow it, what the expectations are, how we label items, etc. Topics of interest: 1. Large undertakings that are not necessarily features, such as solving critical architectural problems, but improve the sustainability of the project, are not usually Epics. Technical previews and sprints seem to be driven by feature requirements, which is a practice that creates a brittle architectural foundation and builds technical debt. - What should we expect of Epics if they are mostly for features, and how can we ensure that the development process accommodates scalability, growth, and maintainability for (hopefully) the possibility of multitenancy and increasing feature requests in the future? - Should epics spawn other epics that are not features, or that are branches of existing Epics that perhaps were originally too broad? 2. When it becomes clear that an Epic is too broad of a scope, we tend to create Issues. An example here is the [API Connector](https://github.com/syndesisio/syndesis/issues/183) and all of which it entails. As you can see, it becomes a bit unwieldy and difficult to see the progress of. It seems that Epics should sometimes spawn other Epics. I hate to say it, but maybe we do need a "Feature" label that issues pertain to, and those Features could be sprint or release specific (ZenHub had 'Scenarios' but this does not make any sense at all in terms of wording). An example would be the API Client Connector Wizard. Currently: - Epic: API Client Connector (targets a TP release) - All Issues Perhaps something like this would work? - Epic: API Client Connector (spans multiple sprints, possibly TPs) - Feature 1: API Client Connector Wizard (~1-3 sprints, who knows?)* - Issue: REST: Page-Specific Endpoints (e.g. Sprint 22) - Issue: UX: Page-Specific Designs (e.g. Spint 22) - Issue: UI: Step 1 (e.g. Sprint 21) - Issue: UI: Step 2 - Security (e.g. Sprint 22) - Issue: UI: [..] - Feature 2: API Client Connector List - Issue: REST: Page-Specific Endpoints (e.g. Sprint 23) - Feature 3: API Client Connector Detail - Epic: Technical Extensions [..] _* Anyone that pretends to know is likely guesstimating._ This is driven by the user requirements, not by anything else. But the idea is that the issues are what determine what can or cannot be accomplished in a single sprint, or even release. It would ensure to management that we have a new feature in each technical preview release, without over-promising that we will have everything API Client Connector (as an example) in a technical preview release before we've even had time to delve into the code and determine how feasible this really is.
1.0
Solidify Definitions and Expectations of Labels and Statuses - We discussed a few issues that are related to the agile methodology and how we follow it, what the expectations are, how we label items, etc. Topics of interest: 1. Large undertakings that are not necessarily features, such as solving critical architectural problems, but improve the sustainability of the project, are not usually Epics. Technical previews and sprints seem to be driven by feature requirements, which is a practice that creates a brittle architectural foundation and builds technical debt. - What should we expect of Epics if they are mostly for features, and how can we ensure that the development process accommodates scalability, growth, and maintainability for (hopefully) the possibility of multitenancy and increasing feature requests in the future? - Should epics spawn other epics that are not features, or that are branches of existing Epics that perhaps were originally too broad? 2. When it becomes clear that an Epic is too broad of a scope, we tend to create Issues. An example here is the [API Connector](https://github.com/syndesisio/syndesis/issues/183) and all of which it entails. As you can see, it becomes a bit unwieldy and difficult to see the progress of. It seems that Epics should sometimes spawn other Epics. I hate to say it, but maybe we do need a "Feature" label that issues pertain to, and those Features could be sprint or release specific (ZenHub had 'Scenarios' but this does not make any sense at all in terms of wording). An example would be the API Client Connector Wizard. Currently: - Epic: API Client Connector (targets a TP release) - All Issues Perhaps something like this would work? - Epic: API Client Connector (spans multiple sprints, possibly TPs) - Feature 1: API Client Connector Wizard (~1-3 sprints, who knows?)* - Issue: REST: Page-Specific Endpoints (e.g. Sprint 22) - Issue: UX: Page-Specific Designs (e.g. Spint 22) - Issue: UI: Step 1 (e.g. Sprint 21) - Issue: UI: Step 2 - Security (e.g. Sprint 22) - Issue: UI: [..] - Feature 2: API Client Connector List - Issue: REST: Page-Specific Endpoints (e.g. Sprint 23) - Feature 3: API Client Connector Detail - Epic: Technical Extensions [..] _* Anyone that pretends to know is likely guesstimating._ This is driven by the user requirements, not by anything else. But the idea is that the issues are what determine what can or cannot be accomplished in a single sprint, or even release. It would ensure to management that we have a new feature in each technical preview release, without over-promising that we will have everything API Client Connector (as an example) in a technical preview release before we've even had time to delve into the code and determine how feasible this really is.
process
solidify definitions and expectations of labels and statuses we discussed a few issues that are related to the agile methodology and how we follow it what the expectations are how we label items etc topics of interest large undertakings that are not necessarily features such as solving critical architectural problems but improve the sustainability of the project are not usually epics technical previews and sprints seem to be driven by feature requirements which is a practice that creates a brittle architectural foundation and builds technical debt what should we expect of epics if they are mostly for features and how can we ensure that the development process accommodates scalability growth and maintainability for hopefully the possibility of multitenancy and increasing feature requests in the future should epics spawn other epics that are not features or that are branches of existing epics that perhaps were originally too broad when it becomes clear that an epic is too broad of a scope we tend to create issues an example here is the and all of which it entails as you can see it becomes a bit unwieldy and difficult to see the progress of it seems that epics should sometimes spawn other epics i hate to say it but maybe we do need a feature label that issues pertain to and those features could be sprint or release specific zenhub had scenarios but this does not make any sense at all in terms of wording an example would be the api client connector wizard currently epic api client connector targets a tp release all issues perhaps something like this would work epic api client connector spans multiple sprints possibly tps feature api client connector wizard sprints who knows issue rest page specific endpoints e g sprint issue ux page specific designs e g spint issue ui step e g sprint issue ui step security e g sprint issue ui feature api client connector list issue rest page specific endpoints e g sprint feature api client connector detail epic technical extensions anyone that pretends to know is likely guesstimating this is driven by the user requirements not by anything else but the idea is that the issues are what determine what can or cannot be accomplished in a single sprint or even release it would ensure to management that we have a new feature in each technical preview release without over promising that we will have everything api client connector as an example in a technical preview release before we ve even had time to delve into the code and determine how feasible this really is
1
330,918
24,283,370,136
IssuesEvent
2022-09-28 19:31:21
chartjs/Chart.js
https://api.github.com/repos/chartjs/Chart.js
closed
Samples in Documentation are small and tabs are not working
type: bug type: documentation
### Expected behavior The charts in the documentation should be big enough that you can see what is going on in them. When the charts are in tabs, when you switch to another tab it should show the chart ### Current behavior The charts are extremely small so that the information and customization is almost not vissable. When switching tabs the chart does not load and no space gets reserved Tabs not working: https://www.chartjs.org/docs/master/charts/doughnut.html#doughnut https://www.chartjs.org/docs/master/configuration/canvas-background.html#color Some charts not showing information/customization because they are too small: https://www.chartjs.org/docs/master/samples/area/line-datasets.html https://www.chartjs.org/docs/master/samples/area/line-stacked.html https://www.chartjs.org/docs/master/samples/bar/stacked.html https://www.chartjs.org/docs/master/samples/scale-options/grid.html ### Reproducible sample https://www.chartjs.org/docs/master/ ### Optional extra steps/info to reproduce _No response_ ### Possible solution _No response_ ### Context _No response_ ### chart.js version 4.0.0dev current master ### Browser name and version _No response_ ### Link to your project _No response_
1.0
Samples in Documentation are small and tabs are not working - ### Expected behavior The charts in the documentation should be big enough that you can see what is going on in them. When the charts are in tabs, when you switch to another tab it should show the chart ### Current behavior The charts are extremely small so that the information and customization is almost not vissable. When switching tabs the chart does not load and no space gets reserved Tabs not working: https://www.chartjs.org/docs/master/charts/doughnut.html#doughnut https://www.chartjs.org/docs/master/configuration/canvas-background.html#color Some charts not showing information/customization because they are too small: https://www.chartjs.org/docs/master/samples/area/line-datasets.html https://www.chartjs.org/docs/master/samples/area/line-stacked.html https://www.chartjs.org/docs/master/samples/bar/stacked.html https://www.chartjs.org/docs/master/samples/scale-options/grid.html ### Reproducible sample https://www.chartjs.org/docs/master/ ### Optional extra steps/info to reproduce _No response_ ### Possible solution _No response_ ### Context _No response_ ### chart.js version 4.0.0dev current master ### Browser name and version _No response_ ### Link to your project _No response_
non_process
samples in documentation are small and tabs are not working expected behavior the charts in the documentation should be big enough that you can see what is going on in them when the charts are in tabs when you switch to another tab it should show the chart current behavior the charts are extremely small so that the information and customization is almost not vissable when switching tabs the chart does not load and no space gets reserved tabs not working some charts not showing information customization because they are too small reproducible sample optional extra steps info to reproduce no response possible solution no response context no response chart js version current master browser name and version no response link to your project no response
0
17,147
22,693,563,808
IssuesEvent
2022-07-05 01:43:39
Carlosmtp/DomuzSGI
https://api.github.com/repos/Carlosmtp/DomuzSGI
closed
Creación de nuevas columnas
Enhancement Medium Process Management
- [x] Crear una columna goal en los indicadores de procesos - [x] Crear la columna goal en el proceso - [x] Funcion para crear un registro en periodic_records - [x] Funcion para retornar los periodic_records en base a un año. Ex "get/periodic_records/",2021 (Nose si en la base de datos se puede tener una **fecha** que es la que llega desde el front como 3/07/2022 y un mes como **Julio** para facilitar el filtrado por mes)
1.0
Creación de nuevas columnas - - [x] Crear una columna goal en los indicadores de procesos - [x] Crear la columna goal en el proceso - [x] Funcion para crear un registro en periodic_records - [x] Funcion para retornar los periodic_records en base a un año. Ex "get/periodic_records/",2021 (Nose si en la base de datos se puede tener una **fecha** que es la que llega desde el front como 3/07/2022 y un mes como **Julio** para facilitar el filtrado por mes)
process
creación de nuevas columnas crear una columna goal en los indicadores de procesos crear la columna goal en el proceso funcion para crear un registro en periodic records funcion para retornar los periodic records en base a un año ex get periodic records nose si en la base de datos se puede tener una fecha que es la que llega desde el front como y un mes como julio para facilitar el filtrado por mes
1
22,308
30,861,130,129
IssuesEvent
2023-08-03 03:11:32
metabase/metabase
https://api.github.com/repos/metabase/metabase
closed
[MLv2] Tweaks to `join-lhs-display-name` method
.metabase-lib .Team/QueryProcessor :hammer_and_wrench:
Apparently, the `join-lhs-display-name` logic is a bit more comprehensive when adding a new join. Scenarios: ### 1. Adding a join when there're no joins When adding a new join and there're no other joins, we should use the source table/model/question name as the LHS table display name (even before the RHS table is selected and before there're any conditions added) <details> <summary>Screenshot</summary> <img width="383" alt="join-lhs-display-name-1" src="https://github.com/metabase/metabase/assets/17258145/c683ea1f-b054-4355-88d0-8fdbb19a2646"> </details> ### 2. Adding a join when there're other joins When adding a new join and there's at least one join on the same stage, we should display "Previous results" before at least an LHS column is selected. After the LHS column is selected, the LHS display name should be the LHS column's table name. I believe there're both FE and BE parts of the fix: 1. When the LHS column is selected, but the RHS one isn't yet selected, there's no join condition yet. This means `join-lhs-display-name` has nothing to use to figure out the LHS table display name. We probably want the FE to figure this out until there's at least one condition 2. Once there's at least one join condition, we probably want to completely rely on `join-lhs-display-name` (BE) here <details> <summary>Screenshot (no RHS table selected)</summary> <img width="664" alt="join-lhs-display-name-2" src="https://github.com/metabase/metabase/assets/17258145/b9fef436-2dab-4f6e-b3b4-f050e42d0294"> </details> <details> <summary>Screenshot (RHS table selected, no columns selected)</summary> <img width="934" alt="join-lhs-display-name-2 1" src="https://github.com/metabase/metabase/assets/17258145/c9c6b16e-f78b-49e7-8c78-abdb94860ded"> </details> <details> <summary>Screenshot (RHS table selected, LHS column selected)</summary> <img width="931" alt="join-lhs-display-name-2 2" src="https://github.com/metabase/metabase/assets/17258145/b80ca08b-1d65-4b20-86ad-fe98d292223a"> </details>
1.0
[MLv2] Tweaks to `join-lhs-display-name` method - Apparently, the `join-lhs-display-name` logic is a bit more comprehensive when adding a new join. Scenarios: ### 1. Adding a join when there're no joins When adding a new join and there're no other joins, we should use the source table/model/question name as the LHS table display name (even before the RHS table is selected and before there're any conditions added) <details> <summary>Screenshot</summary> <img width="383" alt="join-lhs-display-name-1" src="https://github.com/metabase/metabase/assets/17258145/c683ea1f-b054-4355-88d0-8fdbb19a2646"> </details> ### 2. Adding a join when there're other joins When adding a new join and there's at least one join on the same stage, we should display "Previous results" before at least an LHS column is selected. After the LHS column is selected, the LHS display name should be the LHS column's table name. I believe there're both FE and BE parts of the fix: 1. When the LHS column is selected, but the RHS one isn't yet selected, there's no join condition yet. This means `join-lhs-display-name` has nothing to use to figure out the LHS table display name. We probably want the FE to figure this out until there's at least one condition 2. Once there's at least one join condition, we probably want to completely rely on `join-lhs-display-name` (BE) here <details> <summary>Screenshot (no RHS table selected)</summary> <img width="664" alt="join-lhs-display-name-2" src="https://github.com/metabase/metabase/assets/17258145/b9fef436-2dab-4f6e-b3b4-f050e42d0294"> </details> <details> <summary>Screenshot (RHS table selected, no columns selected)</summary> <img width="934" alt="join-lhs-display-name-2 1" src="https://github.com/metabase/metabase/assets/17258145/c9c6b16e-f78b-49e7-8c78-abdb94860ded"> </details> <details> <summary>Screenshot (RHS table selected, LHS column selected)</summary> <img width="931" alt="join-lhs-display-name-2 2" src="https://github.com/metabase/metabase/assets/17258145/b80ca08b-1d65-4b20-86ad-fe98d292223a"> </details>
process
tweaks to join lhs display name method apparently the join lhs display name logic is a bit more comprehensive when adding a new join scenarios adding a join when there re no joins when adding a new join and there re no other joins we should use the source table model question name as the lhs table display name even before the rhs table is selected and before there re any conditions added screenshot img width alt join lhs display name src adding a join when there re other joins when adding a new join and there s at least one join on the same stage we should display previous results before at least an lhs column is selected after the lhs column is selected the lhs display name should be the lhs column s table name i believe there re both fe and be parts of the fix when the lhs column is selected but the rhs one isn t yet selected there s no join condition yet this means join lhs display name has nothing to use to figure out the lhs table display name we probably want the fe to figure this out until there s at least one condition once there s at least one join condition we probably want to completely rely on join lhs display name be here screenshot no rhs table selected img width alt join lhs display name src screenshot rhs table selected no columns selected img width alt join lhs display name src screenshot rhs table selected lhs column selected img width alt join lhs display name src
1
19,813
26,202,927,834
IssuesEvent
2023-01-03 19:20:02
zammad/zammad
https://api.github.com/repos/zammad/zammad
opened
mail processing not possible if reply-to header is faulty
bug verified mail processing
### Used Zammad Version 5.3 / stable ### Environment - Installation method: any - Operating system (if you're unsure: `cat /etc/os-release` ): any - Database + version: any - Elasticsearch version: any - Browser + version: any ### Actual behaviour Zammad cannot process emails that contain invalid email addresses. While this is working as designed, this also affects reply-to headers. There is a special kind of experts (some Outlook users are affected too) do set a reply-to header but without mail address like so: ``` Reply-To: "Jane Doe" Reply-To: John Doe ``` By default mails containing one of the above line will cause Zammad to not import mails but instead write them to local storage. I thus suggest to ignore faulty reply to headers if the "from" address does fit. This may also be relevant for contact form based mails were spammers decide to use RFC invalid mail addresses. This may be PM relevant however. ### Expected behaviour Zammad ignores invalid reply-to headers and imports the mail. ### Steps to reproduce the behaviour send a email to Zammad with set reply-to header which is incomplete (missing email address) ### Support Ticket _No response_ ### I'm sure this is a bug and no feature request or a general question. yes
1.0
mail processing not possible if reply-to header is faulty - ### Used Zammad Version 5.3 / stable ### Environment - Installation method: any - Operating system (if you're unsure: `cat /etc/os-release` ): any - Database + version: any - Elasticsearch version: any - Browser + version: any ### Actual behaviour Zammad cannot process emails that contain invalid email addresses. While this is working as designed, this also affects reply-to headers. There is a special kind of experts (some Outlook users are affected too) do set a reply-to header but without mail address like so: ``` Reply-To: "Jane Doe" Reply-To: John Doe ``` By default mails containing one of the above line will cause Zammad to not import mails but instead write them to local storage. I thus suggest to ignore faulty reply to headers if the "from" address does fit. This may also be relevant for contact form based mails were spammers decide to use RFC invalid mail addresses. This may be PM relevant however. ### Expected behaviour Zammad ignores invalid reply-to headers and imports the mail. ### Steps to reproduce the behaviour send a email to Zammad with set reply-to header which is incomplete (missing email address) ### Support Ticket _No response_ ### I'm sure this is a bug and no feature request or a general question. yes
process
mail processing not possible if reply to header is faulty used zammad version stable environment installation method any operating system if you re unsure cat etc os release any database version any elasticsearch version any browser version any actual behaviour zammad cannot process emails that contain invalid email addresses while this is working as designed this also affects reply to headers there is a special kind of experts some outlook users are affected too do set a reply to header but without mail address like so reply to jane doe reply to john doe by default mails containing one of the above line will cause zammad to not import mails but instead write them to local storage i thus suggest to ignore faulty reply to headers if the from address does fit this may also be relevant for contact form based mails were spammers decide to use rfc invalid mail addresses this may be pm relevant however expected behaviour zammad ignores invalid reply to headers and imports the mail steps to reproduce the behaviour send a email to zammad with set reply to header which is incomplete missing email address support ticket no response i m sure this is a bug and no feature request or a general question yes
1
2,789
5,721,829,510
IssuesEvent
2017-04-20 07:53:12
openvstorage/volumedriver
https://api.github.com/repos/openvstorage/volumedriver
closed
Integrate Alba with block cache
process_wontfix type_enhancement
API change: new `block_path` field in `struct alba::statistics::RoraCounter`.
1.0
Integrate Alba with block cache - API change: new `block_path` field in `struct alba::statistics::RoraCounter`.
process
integrate alba with block cache api change new block path field in struct alba statistics roracounter
1
795,806
28,086,938,096
IssuesEvent
2023-03-30 10:24:05
GoogleCloudPlatform/python-docs-samples
https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples
closed
asset.snippets.quickstart_getfeed_test: test_get_feed failed
priority: p1 type: bug api: cloudasset samples flakybot: issue
Note: #9021 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky. ---- commit: a0b5ecff8961110ffaa3fad9cf013c7268779e19 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/a7f4b84c-bd62-4b1f-bc1c-e447bc4457f9), [Sponge](http://sponge2/a7f4b84c-bd62-4b1f-bc1c-e447bc4457f9) status: failed <details><summary>Test output</summary><br><pre>Traceback (most recent call last): File "/workspace/asset/snippets/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/grpc_helpers.py", line 72, in error_remapped_callable return callable_(*args, **kwargs) File "/workspace/asset/snippets/.nox/py-3-7/lib/python3.7/site-packages/grpc/_channel.py", line 1030, in __call__ return _end_unary_response_blocking(state, call, False, None) File "/workspace/asset/snippets/.nox/py-3-7/lib/python3.7/site-packages/grpc/_channel.py", line 910, in _end_unary_response_blocking raise _InactiveRpcError(state) # pytype: disable=not-instantiable grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with: status = StatusCode.DEADLINE_EXCEEDED details = "Deadline Exceeded" debug_error_string = "UNKNOWN:Deadline Exceeded {grpc_status:4, created_time:"2023-03-29T20:44:06.93084194+00:00"}" > The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/workspace/asset/snippets/conftest.py", line 40, in test_topic topic = publisher.create_topic(request={"name": topic_path}) File "/workspace/asset/snippets/.nox/py-3-7/lib/python3.7/site-packages/google/pubsub_v1/services/publisher/client.py", line 590, in create_topic metadata=metadata, File "/workspace/asset/snippets/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/gapic_v1/method.py", line 113, in __call__ return wrapped_func(*args, **kwargs) File "/workspace/asset/snippets/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/retry.py", line 354, in retry_wrapped_func on_error=on_error, File "/workspace/asset/snippets/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/retry.py", line 191, in retry_target return target() File "/workspace/asset/snippets/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/timeout.py", line 120, in func_with_timeout return func(*args, **kwargs) File "/workspace/asset/snippets/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/grpc_helpers.py", line 74, in error_remapped_callable raise exceptions.from_grpc_error(exc) from exc google.api_core.exceptions.DeadlineExceeded: 504 Deadline Exceeded</pre></details>
1.0
asset.snippets.quickstart_getfeed_test: test_get_feed failed - Note: #9021 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky. ---- commit: a0b5ecff8961110ffaa3fad9cf013c7268779e19 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/a7f4b84c-bd62-4b1f-bc1c-e447bc4457f9), [Sponge](http://sponge2/a7f4b84c-bd62-4b1f-bc1c-e447bc4457f9) status: failed <details><summary>Test output</summary><br><pre>Traceback (most recent call last): File "/workspace/asset/snippets/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/grpc_helpers.py", line 72, in error_remapped_callable return callable_(*args, **kwargs) File "/workspace/asset/snippets/.nox/py-3-7/lib/python3.7/site-packages/grpc/_channel.py", line 1030, in __call__ return _end_unary_response_blocking(state, call, False, None) File "/workspace/asset/snippets/.nox/py-3-7/lib/python3.7/site-packages/grpc/_channel.py", line 910, in _end_unary_response_blocking raise _InactiveRpcError(state) # pytype: disable=not-instantiable grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with: status = StatusCode.DEADLINE_EXCEEDED details = "Deadline Exceeded" debug_error_string = "UNKNOWN:Deadline Exceeded {grpc_status:4, created_time:"2023-03-29T20:44:06.93084194+00:00"}" > The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/workspace/asset/snippets/conftest.py", line 40, in test_topic topic = publisher.create_topic(request={"name": topic_path}) File "/workspace/asset/snippets/.nox/py-3-7/lib/python3.7/site-packages/google/pubsub_v1/services/publisher/client.py", line 590, in create_topic metadata=metadata, File "/workspace/asset/snippets/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/gapic_v1/method.py", line 113, in __call__ return wrapped_func(*args, **kwargs) File "/workspace/asset/snippets/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/retry.py", line 354, in retry_wrapped_func on_error=on_error, File "/workspace/asset/snippets/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/retry.py", line 191, in retry_target return target() File "/workspace/asset/snippets/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/timeout.py", line 120, in func_with_timeout return func(*args, **kwargs) File "/workspace/asset/snippets/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/grpc_helpers.py", line 74, in error_remapped_callable raise exceptions.from_grpc_error(exc) from exc google.api_core.exceptions.DeadlineExceeded: 504 Deadline Exceeded</pre></details>
non_process
asset snippets quickstart getfeed test test get feed failed note was also for this test but it was closed more than days ago so i didn t mark it flaky commit buildurl status failed test output traceback most recent call last file workspace asset snippets nox py lib site packages google api core grpc helpers py line in error remapped callable return callable args kwargs file workspace asset snippets nox py lib site packages grpc channel py line in call return end unary response blocking state call false none file workspace asset snippets nox py lib site packages grpc channel py line in end unary response blocking raise inactiverpcerror state pytype disable not instantiable grpc channel inactiverpcerror inactiverpcerror of rpc that terminated with status statuscode deadline exceeded details deadline exceeded debug error string unknown deadline exceeded grpc status created time the above exception was the direct cause of the following exception traceback most recent call last file workspace asset snippets conftest py line in test topic topic publisher create topic request name topic path file workspace asset snippets nox py lib site packages google pubsub services publisher client py line in create topic metadata metadata file workspace asset snippets nox py lib site packages google api core gapic method py line in call return wrapped func args kwargs file workspace asset snippets nox py lib site packages google api core retry py line in retry wrapped func on error on error file workspace asset snippets nox py lib site packages google api core retry py line in retry target return target file workspace asset snippets nox py lib site packages google api core timeout py line in func with timeout return func args kwargs file workspace asset snippets nox py lib site packages google api core grpc helpers py line in error remapped callable raise exceptions from grpc error exc from exc google api core exceptions deadlineexceeded deadline exceeded
0
8,108
11,300,888,030
IssuesEvent
2020-01-17 14:32:44
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
opened
Problem with multiprocessing, custom __getstate__ with Tensors and forkserver
module: multiprocessing module: serialization
## 🐛 Bug **TL;DR**: multiprocessing forkserver / spawn + custom class with `__getstate__` returning tensors give ```python RuntimeError: unable to resize file <filename not specified> to the right size ``` ### Context Suppose the user created a custom class which holds a (large) list of torch tensors inside. In order to handle the large list of tensors as a single tensor (in order to overcome the limit on the number of file descriptors in the system), the user implemented a custom `__getstate__` and `__setstate__`. Pretty smart, this avoids the limitation on file descriptors without having to change the internal representation of the class! Everything works fine, until the multiprocessing method is changed from `fork` to either `forkserver` or `spawn`, in which case we get a the following error: ```python Traceback (most recent call last): File "<string>", line 1, in <module> File "/private/home/fmassa/.conda/envs/classyvision/lib/python3.7/multiprocessing/spawn.py", line 105, in spawn_main exitcode = _main(fd) File "/private/home/fmassa/.conda/envs/classyvision/lib/python3.7/multiprocessing/spawn.py", line 115, in _main self = reduction.pickle.load(from_parent) File "/private/home/fmassa/.conda/envs/classyvision/lib/python3.7/site-packages/torch/multiprocessing/reductions.py", line 299, in rebuild_storage_fd storage = cls._new_shared_fd(fd, size) RuntimeError: unable to resize file <filename not specified> to the right size ``` ## To Reproduce The following code reproduces the problem. I illustrate the issue in two cases (which are actually the same): one when using the `DataLoader`, and another one just when creating a new process with `forkserver`. ```python import torch class D: def __init__(self): # the size of the tensor doesn't matter here self.x = torch.rand(2).unbind(0) def __getitem__(self, idx): return self.x[idx] def __len__(self): return len(self.x) # let's handle our (?)large list of tensors def __getstate__(self): x = torch.stack(self.x) return x def __setstate__(self, d): self.x = d.unbind(0) def loop(ds): for i in range(2): print(ds[i]) if __name__ == "__main__": torch.multiprocessing.set_start_method('forkserver') d = D() # two cases if False: # users normally hit this error in the dataloader # but it is not specific to it dl = torch.utils.data.DataLoader(d, batch_size=4, num_workers=1) it = iter(dl) print(next(it)) else: # as we can see, it also happen when just creating # a process import multiprocessing w = multiprocessing.Process( target=loop, args=(d,)) w.start() ``` ## Expected behavior No error? :-) ## Environment ``` PyTorch version: 1.4.0 Is debug build: No CUDA used to build PyTorch: 9.2 OS: Ubuntu 18.04.3 LTS GCC version: (Ubuntu 7.4.0-1ubuntu1~18.04.1) 7.4.0 CMake version: version 3.10.2 Python version: 3.7 Is CUDA available: Yes CUDA runtime version: 9.2.88 GPU models and configuration: GPU 0: Quadro GP100 GPU 1: Quadro GP100 Nvidia driver version: 418.116.00 cuDNN version: Could not collect Versions of relevant libraries: [pip] numpy==1.17.2 [pip] torch==1.4.0 [pip] torchvision==0.5.0a0+07cbb46 [conda] blas 1.0 mkl [conda] mkl 2019.4 243 [conda] mkl-service 2.3.0 py37he904b0f_0 [conda] mkl_fft 1.0.14 py37ha843d7b_0 [conda] mkl_random 1.1.0 py37hd6b4f25_0 [conda] pytorch 1.4.0 py3.7_cuda9.2.148_cudnn7.6.3_0 pytorch-nightly [conda] torchvision 0.5.0a0+07cbb46 <pip> ``` ## Additional context This has been reported in the past, see https://github.com/pytorch/pytorch/issues/20409, but without a repro so it was hard to act on it
1.0
Problem with multiprocessing, custom __getstate__ with Tensors and forkserver - ## 🐛 Bug **TL;DR**: multiprocessing forkserver / spawn + custom class with `__getstate__` returning tensors give ```python RuntimeError: unable to resize file <filename not specified> to the right size ``` ### Context Suppose the user created a custom class which holds a (large) list of torch tensors inside. In order to handle the large list of tensors as a single tensor (in order to overcome the limit on the number of file descriptors in the system), the user implemented a custom `__getstate__` and `__setstate__`. Pretty smart, this avoids the limitation on file descriptors without having to change the internal representation of the class! Everything works fine, until the multiprocessing method is changed from `fork` to either `forkserver` or `spawn`, in which case we get a the following error: ```python Traceback (most recent call last): File "<string>", line 1, in <module> File "/private/home/fmassa/.conda/envs/classyvision/lib/python3.7/multiprocessing/spawn.py", line 105, in spawn_main exitcode = _main(fd) File "/private/home/fmassa/.conda/envs/classyvision/lib/python3.7/multiprocessing/spawn.py", line 115, in _main self = reduction.pickle.load(from_parent) File "/private/home/fmassa/.conda/envs/classyvision/lib/python3.7/site-packages/torch/multiprocessing/reductions.py", line 299, in rebuild_storage_fd storage = cls._new_shared_fd(fd, size) RuntimeError: unable to resize file <filename not specified> to the right size ``` ## To Reproduce The following code reproduces the problem. I illustrate the issue in two cases (which are actually the same): one when using the `DataLoader`, and another one just when creating a new process with `forkserver`. ```python import torch class D: def __init__(self): # the size of the tensor doesn't matter here self.x = torch.rand(2).unbind(0) def __getitem__(self, idx): return self.x[idx] def __len__(self): return len(self.x) # let's handle our (?)large list of tensors def __getstate__(self): x = torch.stack(self.x) return x def __setstate__(self, d): self.x = d.unbind(0) def loop(ds): for i in range(2): print(ds[i]) if __name__ == "__main__": torch.multiprocessing.set_start_method('forkserver') d = D() # two cases if False: # users normally hit this error in the dataloader # but it is not specific to it dl = torch.utils.data.DataLoader(d, batch_size=4, num_workers=1) it = iter(dl) print(next(it)) else: # as we can see, it also happen when just creating # a process import multiprocessing w = multiprocessing.Process( target=loop, args=(d,)) w.start() ``` ## Expected behavior No error? :-) ## Environment ``` PyTorch version: 1.4.0 Is debug build: No CUDA used to build PyTorch: 9.2 OS: Ubuntu 18.04.3 LTS GCC version: (Ubuntu 7.4.0-1ubuntu1~18.04.1) 7.4.0 CMake version: version 3.10.2 Python version: 3.7 Is CUDA available: Yes CUDA runtime version: 9.2.88 GPU models and configuration: GPU 0: Quadro GP100 GPU 1: Quadro GP100 Nvidia driver version: 418.116.00 cuDNN version: Could not collect Versions of relevant libraries: [pip] numpy==1.17.2 [pip] torch==1.4.0 [pip] torchvision==0.5.0a0+07cbb46 [conda] blas 1.0 mkl [conda] mkl 2019.4 243 [conda] mkl-service 2.3.0 py37he904b0f_0 [conda] mkl_fft 1.0.14 py37ha843d7b_0 [conda] mkl_random 1.1.0 py37hd6b4f25_0 [conda] pytorch 1.4.0 py3.7_cuda9.2.148_cudnn7.6.3_0 pytorch-nightly [conda] torchvision 0.5.0a0+07cbb46 <pip> ``` ## Additional context This has been reported in the past, see https://github.com/pytorch/pytorch/issues/20409, but without a repro so it was hard to act on it
process
problem with multiprocessing custom getstate with tensors and forkserver 🐛 bug tl dr multiprocessing forkserver spawn custom class with getstate returning tensors give python runtimeerror unable to resize file to the right size context suppose the user created a custom class which holds a large list of torch tensors inside in order to handle the large list of tensors as a single tensor in order to overcome the limit on the number of file descriptors in the system the user implemented a custom getstate and setstate pretty smart this avoids the limitation on file descriptors without having to change the internal representation of the class everything works fine until the multiprocessing method is changed from fork to either forkserver or spawn in which case we get a the following error python traceback most recent call last file line in file private home fmassa conda envs classyvision lib multiprocessing spawn py line in spawn main exitcode main fd file private home fmassa conda envs classyvision lib multiprocessing spawn py line in main self reduction pickle load from parent file private home fmassa conda envs classyvision lib site packages torch multiprocessing reductions py line in rebuild storage fd storage cls new shared fd fd size runtimeerror unable to resize file to the right size to reproduce the following code reproduces the problem i illustrate the issue in two cases which are actually the same one when using the dataloader and another one just when creating a new process with forkserver python import torch class d def init self the size of the tensor doesn t matter here self x torch rand unbind def getitem self idx return self x def len self return len self x let s handle our large list of tensors def getstate self x torch stack self x return x def setstate self d self x d unbind def loop ds for i in range print ds if name main torch multiprocessing set start method forkserver d d two cases if false users normally hit this error in the dataloader but it is not specific to it dl torch utils data dataloader d batch size num workers it iter dl print next it else as we can see it also happen when just creating a process import multiprocessing w multiprocessing process target loop args d w start expected behavior no error environment pytorch version is debug build no cuda used to build pytorch os ubuntu lts gcc version ubuntu cmake version version python version is cuda available yes cuda runtime version gpu models and configuration gpu quadro gpu quadro nvidia driver version cudnn version could not collect versions of relevant libraries numpy torch torchvision blas mkl mkl mkl service mkl fft mkl random pytorch pytorch nightly torchvision additional context this has been reported in the past see but without a repro so it was hard to act on it
1
17,320
23,141,162,271
IssuesEvent
2022-07-28 18:37:53
benthosdev/benthos
https://api.github.com/repos/benthosdev/benthos
closed
Processing Avro OCF
enhancement processors inputs
Hi! Am i right that currently it is not possible to process Avro OCF (Object Container Files, http://avro.apache.org/docs/current/spec.html#Object+Container+Files) with benthos? The main difference here is, that the schema is within the header an can be used to decode the message. Cheers Armin
1.0
Processing Avro OCF - Hi! Am i right that currently it is not possible to process Avro OCF (Object Container Files, http://avro.apache.org/docs/current/spec.html#Object+Container+Files) with benthos? The main difference here is, that the schema is within the header an can be used to decode the message. Cheers Armin
process
processing avro ocf hi am i right that currently it is not possible to process avro ocf object container files with benthos the main difference here is that the schema is within the header an can be used to decode the message cheers armin
1
597,588
18,167,053,718
IssuesEvent
2021-09-27 15:36:17
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
closed
The coap_server sample is missing the actual send in the retransmit routine
bug priority: medium area: Networking
**Describe the bug** If a device can't reach it's destination, it will get into `retransmit_request` for several times to decrement its retries, however no actual send is performed which makes the retransmit request useless. **To Reproduce** Steps to reproduce the behavior: 1. Build the coap_server sample 2. Observe the resource i.e. `coap get -o coap://xx.xx.xx.xx/obs` 3. Stop the coap client 4. Wait until the coap server gets into the retransmit_request **Expected behavior** I would expect the coap server to actual resend a message instead of just decrementing the retries. **Impact** It leaves the coap server retry mechanism useless. **Environment (please complete the following information):** - Zephyr v2.6.0
1.0
The coap_server sample is missing the actual send in the retransmit routine - **Describe the bug** If a device can't reach it's destination, it will get into `retransmit_request` for several times to decrement its retries, however no actual send is performed which makes the retransmit request useless. **To Reproduce** Steps to reproduce the behavior: 1. Build the coap_server sample 2. Observe the resource i.e. `coap get -o coap://xx.xx.xx.xx/obs` 3. Stop the coap client 4. Wait until the coap server gets into the retransmit_request **Expected behavior** I would expect the coap server to actual resend a message instead of just decrementing the retries. **Impact** It leaves the coap server retry mechanism useless. **Environment (please complete the following information):** - Zephyr v2.6.0
non_process
the coap server sample is missing the actual send in the retransmit routine describe the bug if a device can t reach it s destination it will get into retransmit request for several times to decrement its retries however no actual send is performed which makes the retransmit request useless to reproduce steps to reproduce the behavior build the coap server sample observe the resource i e coap get o coap xx xx xx xx obs stop the coap client wait until the coap server gets into the retransmit request expected behavior i would expect the coap server to actual resend a message instead of just decrementing the retries impact it leaves the coap server retry mechanism useless environment please complete the following information zephyr
0
4,604
7,451,846,824
IssuesEvent
2018-03-29 05:41:40
TEAMMATES/teammates
https://api.github.com/repos/TEAMMATES/teammates
closed
Instructor: edit session: multiple choice questions: support generating options for all other teams
a-Process d.Contributors e.4 f-Questions
Similar to #8412 but for 'all other teams' instead of 'all other students'
1.0
Instructor: edit session: multiple choice questions: support generating options for all other teams - Similar to #8412 but for 'all other teams' instead of 'all other students'
process
instructor edit session multiple choice questions support generating options for all other teams similar to but for all other teams instead of all other students
1
338,563
24,590,729,329
IssuesEvent
2022-10-14 01:46:23
nvaccess/nvda
https://api.github.com/repos/nvaccess/nvda
closed
The delete profile dialog should name the profile that is going to be deleted
component/documentation feature/configuration-profiles p4 good first issue triaged
### Steps to reproduce: 1. Open the NVDA menu. 2. Select `Configuration profiles`. 3. Choose a profile. 4. Press `delete`. ### Actual behavior: Receive a dialog including the following text: >This profile will be permanently deleted. This action cannot be undone. >OK Cancel ### Expected behavior: Receive a dialog containing text similar to this: >The "PROFILE_NAME" profile will be permanently deleted. This action cannot be undone. >OK Cancel ### System configuration #### NVDA installed/portable/running from source: All #### NVDA version: 2022.1 #### Windows version: Windows 10 #### Name and version of other software in use when reproducing the issue: N/A #### Other information about your system: ### Other questions #### Does the issue still occur after restarting your computer? Yes #### Have you tried any other versions of NVDA? If so, please report their behaviors. Yes, many, same. #### If NVDA add-ons are disabled, is your problem still occurring? Yes. #### Does the issue still occur after you run the COM Registration Fixing Tool in NVDA's tools menu? Yes.
1.0
The delete profile dialog should name the profile that is going to be deleted - ### Steps to reproduce: 1. Open the NVDA menu. 2. Select `Configuration profiles`. 3. Choose a profile. 4. Press `delete`. ### Actual behavior: Receive a dialog including the following text: >This profile will be permanently deleted. This action cannot be undone. >OK Cancel ### Expected behavior: Receive a dialog containing text similar to this: >The "PROFILE_NAME" profile will be permanently deleted. This action cannot be undone. >OK Cancel ### System configuration #### NVDA installed/portable/running from source: All #### NVDA version: 2022.1 #### Windows version: Windows 10 #### Name and version of other software in use when reproducing the issue: N/A #### Other information about your system: ### Other questions #### Does the issue still occur after restarting your computer? Yes #### Have you tried any other versions of NVDA? If so, please report their behaviors. Yes, many, same. #### If NVDA add-ons are disabled, is your problem still occurring? Yes. #### Does the issue still occur after you run the COM Registration Fixing Tool in NVDA's tools menu? Yes.
non_process
the delete profile dialog should name the profile that is going to be deleted steps to reproduce open the nvda menu select configuration profiles choose a profile press delete actual behavior receive a dialog including the following text this profile will be permanently deleted this action cannot be undone ok cancel expected behavior receive a dialog containing text similar to this the profile name profile will be permanently deleted this action cannot be undone ok cancel system configuration nvda installed portable running from source all nvda version windows version windows name and version of other software in use when reproducing the issue n a other information about your system other questions does the issue still occur after restarting your computer yes have you tried any other versions of nvda if so please report their behaviors yes many same if nvda add ons are disabled is your problem still occurring yes does the issue still occur after you run the com registration fixing tool in nvda s tools menu yes
0
17,410
23,225,296,537
IssuesEvent
2022-08-02 23:00:50
hashgraph/hedera-mirror-node
https://api.github.com/repos/hashgraph/hedera-mirror-node
closed
Release checklist 0.61
enhancement P1 process
### Problem We need a checklist to verify the release is rolled out successfully. ### Solution ## Preparation - [x] Milestone field populated on relevant [issues](https://github.com/hashgraph/hedera-mirror-node/issues?q=is%3Aclosed+no%3Amilestone+sort%3Aupdated-desc) - [x] Nothing open for [milestone](https://github.com/hashgraph/hedera-mirror-node/issues?q=is%3Aopen+sort%3Aupdated-desc+milestone%3A0.61.0) - [x] GitHub checks for branch are passing - [x] Automated Kubernetes deployment successful - [x] Tag release - [x] Upload release artifacts - [x] Publish release ## Integration - [x] Deploy to VM ## Performance - [x] Deploy to Kubernetes - [x] Deploy to VM - [x] gRPC API performance tests - [x] Importer performance tests - [x] REST API performance tests - [x] Migrations tested against mainnet clone ## Previewnet - [x] Deploy to VM ## Staging - [x] Deploy to Kubernetes EU - [x] Deploy to Kubernetes NA ## Testnet - [x] Deploy to VM ## Mainnet - [x] Deploy to Kubernetes EU - [x] Deploy to Kubernetes NA - [x] Deploy to VM - [x] Deploy to ETL ### Alternatives _No response_
1.0
Release checklist 0.61 - ### Problem We need a checklist to verify the release is rolled out successfully. ### Solution ## Preparation - [x] Milestone field populated on relevant [issues](https://github.com/hashgraph/hedera-mirror-node/issues?q=is%3Aclosed+no%3Amilestone+sort%3Aupdated-desc) - [x] Nothing open for [milestone](https://github.com/hashgraph/hedera-mirror-node/issues?q=is%3Aopen+sort%3Aupdated-desc+milestone%3A0.61.0) - [x] GitHub checks for branch are passing - [x] Automated Kubernetes deployment successful - [x] Tag release - [x] Upload release artifacts - [x] Publish release ## Integration - [x] Deploy to VM ## Performance - [x] Deploy to Kubernetes - [x] Deploy to VM - [x] gRPC API performance tests - [x] Importer performance tests - [x] REST API performance tests - [x] Migrations tested against mainnet clone ## Previewnet - [x] Deploy to VM ## Staging - [x] Deploy to Kubernetes EU - [x] Deploy to Kubernetes NA ## Testnet - [x] Deploy to VM ## Mainnet - [x] Deploy to Kubernetes EU - [x] Deploy to Kubernetes NA - [x] Deploy to VM - [x] Deploy to ETL ### Alternatives _No response_
process
release checklist problem we need a checklist to verify the release is rolled out successfully solution preparation milestone field populated on relevant nothing open for github checks for branch are passing automated kubernetes deployment successful tag release upload release artifacts publish release integration deploy to vm performance deploy to kubernetes deploy to vm grpc api performance tests importer performance tests rest api performance tests migrations tested against mainnet clone previewnet deploy to vm staging deploy to kubernetes eu deploy to kubernetes na testnet deploy to vm mainnet deploy to kubernetes eu deploy to kubernetes na deploy to vm deploy to etl alternatives no response
1
2,432
5,205,381,900
IssuesEvent
2017-01-24 17:47:44
Alfresco/alfresco-ng2-components
https://api.github.com/repos/Alfresco/alfresco-ng2-components
opened
Cannot get back to process instance after clicking task
comp: activiti-processList comp: activiti-taskList
<!-- PLEASE FILL OUT THE FOLLOWING INFORMATION, THIS WILL HELP US TO RESOLVE YOUR PROBLEM FASTER. REMEMBER FOR SUPPORT REQUESTS YOU CAN ALSO ASK ON OUR GITTER CHAT: Please ask before on our gitter channel https://gitter.im/Alfresco/alfresco-ng2-components --> **Type of issue:** (check with "[x]") ``` - [x] New feature request - [ ] Bug - [ ] Support request ``` **Current behavior:** <!-- Describe the current behavior. --> When the user clicks a task item inside a process instance in order to view the details or complete the task, they cannot get back to the process instance if they change their mind, and if they complete the task then the next running task is shown, which may not be connected with the process instance that they were previously looking at. **Expected behavior:** <!-- Describe the expected behavior. --> The user should be able to use the back button and some in-app link to get back to the parent process instance. After completing an active task, the user should be taken back to the process instance from where they can choose to action the next task. **Steps to reproduce the issue:** <!-- Describe the steps to reproduce the issue. --> 1. Click into an app from the Activiti component and select *Processes* 2. Select an existing process instance or create a new one 3. Click into any task on that process instance, either active, completed or start form 4. Try to get back to the previous process instance **Component name and version:** <!-- Example: ng2-alfresco-login. Check before if this issue is still present in the most recent version --> **Browser and version:** <!-- [all | Chrome XX | Firefox XX | IE XX | Safari XX | Mobile Chrome XX | Android X.X Web Browser | iOS XX Safari | iOS XX UIWebView | iOS XX WKWebView ] --> **Node version (for build issues):** <!-- To check the version: node --version --> **New feature request:** <!-- Describe the feature, motivation and the concrete use case (only in case of new feature request) -->
1.0
Cannot get back to process instance after clicking task - <!-- PLEASE FILL OUT THE FOLLOWING INFORMATION, THIS WILL HELP US TO RESOLVE YOUR PROBLEM FASTER. REMEMBER FOR SUPPORT REQUESTS YOU CAN ALSO ASK ON OUR GITTER CHAT: Please ask before on our gitter channel https://gitter.im/Alfresco/alfresco-ng2-components --> **Type of issue:** (check with "[x]") ``` - [x] New feature request - [ ] Bug - [ ] Support request ``` **Current behavior:** <!-- Describe the current behavior. --> When the user clicks a task item inside a process instance in order to view the details or complete the task, they cannot get back to the process instance if they change their mind, and if they complete the task then the next running task is shown, which may not be connected with the process instance that they were previously looking at. **Expected behavior:** <!-- Describe the expected behavior. --> The user should be able to use the back button and some in-app link to get back to the parent process instance. After completing an active task, the user should be taken back to the process instance from where they can choose to action the next task. **Steps to reproduce the issue:** <!-- Describe the steps to reproduce the issue. --> 1. Click into an app from the Activiti component and select *Processes* 2. Select an existing process instance or create a new one 3. Click into any task on that process instance, either active, completed or start form 4. Try to get back to the previous process instance **Component name and version:** <!-- Example: ng2-alfresco-login. Check before if this issue is still present in the most recent version --> **Browser and version:** <!-- [all | Chrome XX | Firefox XX | IE XX | Safari XX | Mobile Chrome XX | Android X.X Web Browser | iOS XX Safari | iOS XX UIWebView | iOS XX WKWebView ] --> **Node version (for build issues):** <!-- To check the version: node --version --> **New feature request:** <!-- Describe the feature, motivation and the concrete use case (only in case of new feature request) -->
process
cannot get back to process instance after clicking task please fill out the following information this will help us to resolve your problem faster remember for support requests you can also ask on our gitter chat please ask before on our gitter channel type of issue check with new feature request bug support request current behavior when the user clicks a task item inside a process instance in order to view the details or complete the task they cannot get back to the process instance if they change their mind and if they complete the task then the next running task is shown which may not be connected with the process instance that they were previously looking at expected behavior the user should be able to use the back button and some in app link to get back to the parent process instance after completing an active task the user should be taken back to the process instance from where they can choose to action the next task steps to reproduce the issue click into an app from the activiti component and select processes select an existing process instance or create a new one click into any task on that process instance either active completed or start form try to get back to the previous process instance component name and version browser and version node version for build issues new feature request
1
21,070
28,017,274,159
IssuesEvent
2023-03-28 00:25:22
kubernetes/minikube
https://api.github.com/repos/kubernetes/minikube
closed
Make ISO/Kicbase commit pushing more robust
priority/backlog lifecycle/rotten kind/process
We had an issue where user created a PR and `ok-to-build-iso` was ran in it (https://github.com/kubernetes/minikube/pull/15232#issuecomment-1295528339). The bot commented that the commit was pushed, but there was no commit. Under further investigation the user made the PR from a fork under the org, but the script seems to assume that the user is using a fork under their own user and the branch creation failed. If the branch creation or commit fails we should comment a user notifying them of the issue, but included the comments to make the changes in the comment so the user and manually update the ISO references themselves. We should also make sure we use the true fork and not just assume the fork is under the user.
1.0
Make ISO/Kicbase commit pushing more robust - We had an issue where user created a PR and `ok-to-build-iso` was ran in it (https://github.com/kubernetes/minikube/pull/15232#issuecomment-1295528339). The bot commented that the commit was pushed, but there was no commit. Under further investigation the user made the PR from a fork under the org, but the script seems to assume that the user is using a fork under their own user and the branch creation failed. If the branch creation or commit fails we should comment a user notifying them of the issue, but included the comments to make the changes in the comment so the user and manually update the ISO references themselves. We should also make sure we use the true fork and not just assume the fork is under the user.
process
make iso kicbase commit pushing more robust we had an issue where user created a pr and ok to build iso was ran in it the bot commented that the commit was pushed but there was no commit under further investigation the user made the pr from a fork under the org but the script seems to assume that the user is using a fork under their own user and the branch creation failed if the branch creation or commit fails we should comment a user notifying them of the issue but included the comments to make the changes in the comment so the user and manually update the iso references themselves we should also make sure we use the true fork and not just assume the fork is under the user
1
114,721
9,753,071,201
IssuesEvent
2019-06-04 08:16:39
Neufund/platform-frontend
https://api.github.com/repos/Neufund/platform-frontend
closed
[E2E] Test for external wallets (inject web3)
E2E test
I did something similar for ICBM tests. Take a look at https://github.com/Neufund/commit.neufund.org/tree/master/test-e2e
1.0
[E2E] Test for external wallets (inject web3) - I did something similar for ICBM tests. Take a look at https://github.com/Neufund/commit.neufund.org/tree/master/test-e2e
non_process
test for external wallets inject i did something similar for icbm tests take a look at
0
16,483
10,964,135,833
IssuesEvent
2019-11-27 21:37:46
daisy/pipeline-gui
https://api.github.com/repos/daisy/pipeline-gui
closed
Start menu shortcut missing after some Windows install?
aoft help wanted verifying usability
This issue has been reported by Mayu (from ATDO in Japan): > I'm trying Pipeline2 Windows version. > > Before, when I install it, I could get a shortcut and it appeared on start menu. > > However, recently, when I install, short cut doesn't appear and also it doesn't appear on the start menu. > > So the only way is to launch from the program folder.. > > It's OK for me but for some end user, it is confusing.
True
Start menu shortcut missing after some Windows install? - This issue has been reported by Mayu (from ATDO in Japan): > I'm trying Pipeline2 Windows version. > > Before, when I install it, I could get a shortcut and it appeared on start menu. > > However, recently, when I install, short cut doesn't appear and also it doesn't appear on the start menu. > > So the only way is to launch from the program folder.. > > It's OK for me but for some end user, it is confusing.
non_process
start menu shortcut missing after some windows install this issue has been reported by mayu from atdo in japan i m trying windows version before when i install it i could get a shortcut and it appeared on start menu however recently when i install short cut doesn t appear and also it doesn t appear on the start menu so the only way is to launch from the program folder it s ok for me but for some end user it is confusing
0
11,023
13,820,666,296
IssuesEvent
2020-10-13 00:10:16
cs-wiki/cs-wiki
https://api.github.com/repos/cs-wiki/cs-wiki
opened
혹시 프로세스 메모리에서 스택 영역은 실행시간 때 크기가 결정되나요? 아니면 컴파일 때 결정되나요??
Compile DB Memory Process Runtime
``` Date : 201012 (Mon) Nickname : 무민 ```
1.0
혹시 프로세스 메모리에서 스택 영역은 실행시간 때 크기가 결정되나요? 아니면 컴파일 때 결정되나요?? - ``` Date : 201012 (Mon) Nickname : 무민 ```
process
혹시 프로세스 메모리에서 스택 영역은 실행시간 때 크기가 결정되나요 아니면 컴파일 때 결정되나요 date mon nickname 무민
1
12,068
14,739,759,399
IssuesEvent
2021-01-07 07:52:26
kdjstudios/SABillingGitlab
https://api.github.com/repos/kdjstudios/SABillingGitlab
closed
Improving the delete draft invoice process
anc-process anp-2 ant-enhancement grt-billing has attachment
In GitLab by @kdjstudios on Sep 18, 2018, 11:17 Over the last few months I have received some feedback from both externals and internals about the Red X button on the draft invoice screen. most of them I was able to simply clarify its purpose for, but feel that maybe we should be updating this to be more user friendly from the start. Currently the draft invoice screen shows the screenshot below when you click the Red X in the top corner. ![image](/uploads/db55aa1be8f5d3073081b94f296f073c/image.png) I feel this may not be giving enough attention to what is actually occurring. My thoughts on improvement - Remove/replace the red "X" and have a "Delete" Button. (For me the X is associated with "close" not "delete" so I can understand the confusion most users see.) - Update Text to "Are you sure you wish to delete/remove the current draft invoice. There is no way to undo this. If you are processing a billing cycle invoice based upon usage, this will remove only this account's invoice and you may have to revert and redo the billing process to generate a replacement invoice with usage." - Since it is just a 'Chrome' dialog pop up that is being displayed it seems to be more of a confirmation, rather then a warning or alert. I think we should be displaying this with more emphasis if possible. IE. some sort of warning icon or bold text, etc?
1.0
Improving the delete draft invoice process - In GitLab by @kdjstudios on Sep 18, 2018, 11:17 Over the last few months I have received some feedback from both externals and internals about the Red X button on the draft invoice screen. most of them I was able to simply clarify its purpose for, but feel that maybe we should be updating this to be more user friendly from the start. Currently the draft invoice screen shows the screenshot below when you click the Red X in the top corner. ![image](/uploads/db55aa1be8f5d3073081b94f296f073c/image.png) I feel this may not be giving enough attention to what is actually occurring. My thoughts on improvement - Remove/replace the red "X" and have a "Delete" Button. (For me the X is associated with "close" not "delete" so I can understand the confusion most users see.) - Update Text to "Are you sure you wish to delete/remove the current draft invoice. There is no way to undo this. If you are processing a billing cycle invoice based upon usage, this will remove only this account's invoice and you may have to revert and redo the billing process to generate a replacement invoice with usage." - Since it is just a 'Chrome' dialog pop up that is being displayed it seems to be more of a confirmation, rather then a warning or alert. I think we should be displaying this with more emphasis if possible. IE. some sort of warning icon or bold text, etc?
process
improving the delete draft invoice process in gitlab by kdjstudios on sep over the last few months i have received some feedback from both externals and internals about the red x button on the draft invoice screen most of them i was able to simply clarify its purpose for but feel that maybe we should be updating this to be more user friendly from the start currently the draft invoice screen shows the screenshot below when you click the red x in the top corner uploads image png i feel this may not be giving enough attention to what is actually occurring my thoughts on improvement remove replace the red x and have a delete button for me the x is associated with close not delete so i can understand the confusion most users see update text to are you sure you wish to delete remove the current draft invoice there is no way to undo this if you are processing a billing cycle invoice based upon usage this will remove only this account s invoice and you may have to revert and redo the billing process to generate a replacement invoice with usage since it is just a chrome dialog pop up that is being displayed it seems to be more of a confirmation rather then a warning or alert i think we should be displaying this with more emphasis if possible ie some sort of warning icon or bold text etc
1
22,176
30,727,147,331
IssuesEvent
2023-07-27 20:42:04
h4sh5/npm-auto-scanner
https://api.github.com/repos/h4sh5/npm-auto-scanner
opened
mongodb-client-encryption 2.9.0 has 2 guarddog issues
npm-install-script npm-silent-process-execution
```{"npm-install-script":[{"code":" \"install\": \"prebuild-install --runtime napi --tag-prefix node-v || node-gyp rebuild\",","location":"package/package.json:19","message":"The package.json has a script automatically running when the package is installed"}],"npm-silent-process-execution":[{"code":" this._child = spawn(cmdName, this.spawnArgs, {\n stdio: 'ignore',\n detached: true\n });","location":"package/lib/mongocryptdManager.js:50","message":"This package is silently executing another executable"}]}```
1.0
mongodb-client-encryption 2.9.0 has 2 guarddog issues - ```{"npm-install-script":[{"code":" \"install\": \"prebuild-install --runtime napi --tag-prefix node-v || node-gyp rebuild\",","location":"package/package.json:19","message":"The package.json has a script automatically running when the package is installed"}],"npm-silent-process-execution":[{"code":" this._child = spawn(cmdName, this.spawnArgs, {\n stdio: 'ignore',\n detached: true\n });","location":"package/lib/mongocryptdManager.js:50","message":"This package is silently executing another executable"}]}```
process
mongodb client encryption has guarddog issues npm install script npm silent process execution
1
34,332
6,311,984,421
IssuesEvent
2017-07-24 00:29:45
mangapress/mangapress
https://api.github.com/repos/mangapress/mangapress
opened
Lightbox integration
documentation enhancement question
Add tutorial on integrating a lightbox with Manga+Press, to allow a comic image to be full-screen. Could this become a feature or separate plugin?
1.0
Lightbox integration - Add tutorial on integrating a lightbox with Manga+Press, to allow a comic image to be full-screen. Could this become a feature or separate plugin?
non_process
lightbox integration add tutorial on integrating a lightbox with manga press to allow a comic image to be full screen could this become a feature or separate plugin
0
6,742
9,872,954,640
IssuesEvent
2019-06-22 09:47:01
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Change default setting to 'Checked' for Processing Options "Use filename as layer name"
Feature Request Processing
Author Name: **Steve Lowman** (Steve Lowman) Original Redmine Issue: [17303](https://issues.qgis.org/issues/17303) Redmine category:processing/core Assignee: Victor Olaya --- The 2.14 ftools in the Vector menu always set the layer name to be the same as the file name, which seems more sensible than the other option. The Processing tools now in the Vector menu follow the Processing Option "Use filename as layer name", which is currently has a default setting of 'Unchecked'. Many users will not easily find the Processing Option to change this. (Currently, if they do find this possibility, they will then encounter the bug reported in #24338 but this is another issue) I suggest changing the default setting to 'Checked' for "Use filename as layer name" will be better for most users.
1.0
Change default setting to 'Checked' for Processing Options "Use filename as layer name" - Author Name: **Steve Lowman** (Steve Lowman) Original Redmine Issue: [17303](https://issues.qgis.org/issues/17303) Redmine category:processing/core Assignee: Victor Olaya --- The 2.14 ftools in the Vector menu always set the layer name to be the same as the file name, which seems more sensible than the other option. The Processing tools now in the Vector menu follow the Processing Option "Use filename as layer name", which is currently has a default setting of 'Unchecked'. Many users will not easily find the Processing Option to change this. (Currently, if they do find this possibility, they will then encounter the bug reported in #24338 but this is another issue) I suggest changing the default setting to 'Checked' for "Use filename as layer name" will be better for most users.
process
change default setting to checked for processing options use filename as layer name author name steve lowman steve lowman original redmine issue redmine category processing core assignee victor olaya the ftools in the vector menu always set the layer name to be the same as the file name which seems more sensible than the other option the processing tools now in the vector menu follow the processing option use filename as layer name which is currently has a default setting of unchecked many users will not easily find the processing option to change this currently if they do find this possibility they will then encounter the bug reported in but this is another issue i suggest changing the default setting to checked for use filename as layer name will be better for most users
1
165,522
12,856,603,441
IssuesEvent
2020-07-09 07:55:36
gyselroth/docxbox
https://api.github.com/repos/gyselroth/docxbox
closed
Feature: Add valgrind memory-leak test
Functional Test
* allow test.sh to receive optional argument: "valgrind" * when running in valgrind mode: add bats-preparation for executing test-suite w/ valgrind: copy functional-suites, replace doxbox binary-path within all copied suites prefixed w/ valgrind execution * test leak detection and reporting: provoke leakage, ensure correct error-message and return-signal
1.0
Feature: Add valgrind memory-leak test - * allow test.sh to receive optional argument: "valgrind" * when running in valgrind mode: add bats-preparation for executing test-suite w/ valgrind: copy functional-suites, replace doxbox binary-path within all copied suites prefixed w/ valgrind execution * test leak detection and reporting: provoke leakage, ensure correct error-message and return-signal
non_process
feature add valgrind memory leak test allow test sh to receive optional argument valgrind when running in valgrind mode add bats preparation for executing test suite w valgrind copy functional suites replace doxbox binary path within all copied suites prefixed w valgrind execution test leak detection and reporting provoke leakage ensure correct error message and return signal
0
4,427
7,307,017,152
IssuesEvent
2018-02-28 00:33:42
P2Poker/P2Poker
https://api.github.com/repos/P2Poker/P2Poker
opened
As a developer, I need a clear and concise directory structure for source code
c) dev origin d) release 0.1 e) dev tools f) priority 2 g) change request h) in process j) difficult workaround l) minor completion cost l) no ux impact n) no impact n) no users affected o) as a developer … p) triage completed
The project needs a directory structure for source code, which will make it easier to manage future source code files.
1.0
As a developer, I need a clear and concise directory structure for source code - The project needs a directory structure for source code, which will make it easier to manage future source code files.
process
as a developer i need a clear and concise directory structure for source code the project needs a directory structure for source code which will make it easier to manage future source code files
1
4,233
7,186,755,980
IssuesEvent
2018-02-02 00:54:33
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
reopened
Test: System.ServiceProcess.Tests.SafeServiceControllerTests/GetServices failed with "Xunit.Sdk.TrueException"
area-System.ServiceProcess disabled-test os-windows-nano test-run-core
Opened on behalf of @Jiayili1 The test `System.ServiceProcess.Tests.SafeServiceControllerTests/GetServices` has failed. KeyIso.CanStop\r Expected: True\r Actual: False Stack Trace: at System.ServiceProcess.Tests.SafeServiceControllerTests.GetServices() Build : Master - 20170502.01 (Core Tests) Failing configurations: - Windows.10.Nano.Amd64-x64 - Debug Detail: https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20170502.01/workItem/System.ServiceProcess.ServiceController.Tests/analysis/xunit/System.ServiceProcess.Tests.SafeServiceControllerTests~2FGetServices
1.0
Test: System.ServiceProcess.Tests.SafeServiceControllerTests/GetServices failed with "Xunit.Sdk.TrueException" - Opened on behalf of @Jiayili1 The test `System.ServiceProcess.Tests.SafeServiceControllerTests/GetServices` has failed. KeyIso.CanStop\r Expected: True\r Actual: False Stack Trace: at System.ServiceProcess.Tests.SafeServiceControllerTests.GetServices() Build : Master - 20170502.01 (Core Tests) Failing configurations: - Windows.10.Nano.Amd64-x64 - Debug Detail: https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20170502.01/workItem/System.ServiceProcess.ServiceController.Tests/analysis/xunit/System.ServiceProcess.Tests.SafeServiceControllerTests~2FGetServices
process
test system serviceprocess tests safeservicecontrollertests getservices failed with xunit sdk trueexception opened on behalf of the test system serviceprocess tests safeservicecontrollertests getservices has failed keyiso canstop r expected true r actual false stack trace at system serviceprocess tests safeservicecontrollertests getservices build master core tests failing configurations windows nano debug detail
1
1,513
4,104,453,020
IssuesEvent
2016-06-05 11:34:11
opentrials/opentrials
https://api.github.com/repos/opentrials/opentrials
closed
Implement processor for HRA data
Processors
# Description Once we have HARP Research Summaries in the warehouse #128, we need to process it to match data onto trials, adjust database schemas as required, and push data forward to the API database. # Tasks - [x] read `processors` overview - https://github.com/opentrials/processors/blob/master/docs/overview.md - [x] read contributing giude - https://github.com/opentrials/processors/blob/master/CONTRIBUTING.md - [x] setup dev environment as written in contributing guide - [x] clone and setup https://github.com/opentrials/api to run `npm run migrate` against development `database` (second database, not a `warehouse`) - [x] read a few `processors` examples like `nct`, `gsk` etc (when there are trial extractors) - [x] copy https://github.com/opentrials/processors/tree/master/processors/nct processor as `hra` processor - [x] update `processors.py` to work with `hra` table instead of `nct` - [x] update `extractors.py` to extract data from HRA source (consult with othter processors examples, `base.writers` and @roll about dict keys extractors should return, related to #180) - [x] run `make test` to lint and test code - [x] run `hra` collector as `make start hra` - [x] check database is filled correctly after `hra` processor run @vitorbaptista @roll I just don't know the tech implementation details to write a proper task list for this. Please provide so @BenGosub can proceed after he finishes #128
1.0
Implement processor for HRA data - # Description Once we have HARP Research Summaries in the warehouse #128, we need to process it to match data onto trials, adjust database schemas as required, and push data forward to the API database. # Tasks - [x] read `processors` overview - https://github.com/opentrials/processors/blob/master/docs/overview.md - [x] read contributing giude - https://github.com/opentrials/processors/blob/master/CONTRIBUTING.md - [x] setup dev environment as written in contributing guide - [x] clone and setup https://github.com/opentrials/api to run `npm run migrate` against development `database` (second database, not a `warehouse`) - [x] read a few `processors` examples like `nct`, `gsk` etc (when there are trial extractors) - [x] copy https://github.com/opentrials/processors/tree/master/processors/nct processor as `hra` processor - [x] update `processors.py` to work with `hra` table instead of `nct` - [x] update `extractors.py` to extract data from HRA source (consult with othter processors examples, `base.writers` and @roll about dict keys extractors should return, related to #180) - [x] run `make test` to lint and test code - [x] run `hra` collector as `make start hra` - [x] check database is filled correctly after `hra` processor run @vitorbaptista @roll I just don't know the tech implementation details to write a proper task list for this. Please provide so @BenGosub can proceed after he finishes #128
process
implement processor for hra data description once we have harp research summaries in the warehouse we need to process it to match data onto trials adjust database schemas as required and push data forward to the api database tasks read processors overview read contributing giude setup dev environment as written in contributing guide clone and setup to run npm run migrate against development database second database not a warehouse read a few processors examples like nct gsk etc when there are trial extractors copy processor as hra processor update processors py to work with hra table instead of nct update extractors py to extract data from hra source consult with othter processors examples base writers and roll about dict keys extractors should return related to run make test to lint and test code run hra collector as make start hra check database is filled correctly after hra processor run vitorbaptista roll i just don t know the tech implementation details to write a proper task list for this please provide so bengosub can proceed after he finishes
1
73,756
19,804,011,243
IssuesEvent
2022-01-19 03:10:48
microsoft/recommenders
https://api.github.com/repos/microsoft/recommenders
closed
[BUG] Build failed on Pyspark 3.2.0
bug build
### Description Our smoke and unit tests failed on the latest pyspark v3.2.0 in `staging` branch. ### In which platform does it happen? In our CI pipelines and in linux virtual machines. For details of the errors, please visit the following links: - [nightly](https://github.com/microsoft/recommenders/runs/3928935302?check_suite_focus=true) - [pr-gate](https://github.com/microsoft/recommenders/runs/3940426052?check_suite_focus=true) - [dsvm_notebook_linux_pyspark](https://dev.azure.com/best-practices/recommenders/_build/results?buildId=52781&view=results) (ADO pipeline) ### How do we replicate the issue? - `git clone https://github.com/microsoft/recommenders.git` - `git checkout staging` - `pip install -e .[spark,example,dev]` - `pytest tests/unit/examples/test_notebooks_pyspark.py` ### Expected behavior (i.e. solution) Our current dependency bounds for pyspark is `>=2.4.5, <4.0.0`. However, pyspark[ released 3.2.0 yesterday](https://pypi.org/project/pyspark/#history) (10/18/2021) that broke our build We should set the bounds for pyspark to `>=2.4.5, <3.2.0` as a hotfix.
1.0
[BUG] Build failed on Pyspark 3.2.0 - ### Description Our smoke and unit tests failed on the latest pyspark v3.2.0 in `staging` branch. ### In which platform does it happen? In our CI pipelines and in linux virtual machines. For details of the errors, please visit the following links: - [nightly](https://github.com/microsoft/recommenders/runs/3928935302?check_suite_focus=true) - [pr-gate](https://github.com/microsoft/recommenders/runs/3940426052?check_suite_focus=true) - [dsvm_notebook_linux_pyspark](https://dev.azure.com/best-practices/recommenders/_build/results?buildId=52781&view=results) (ADO pipeline) ### How do we replicate the issue? - `git clone https://github.com/microsoft/recommenders.git` - `git checkout staging` - `pip install -e .[spark,example,dev]` - `pytest tests/unit/examples/test_notebooks_pyspark.py` ### Expected behavior (i.e. solution) Our current dependency bounds for pyspark is `>=2.4.5, <4.0.0`. However, pyspark[ released 3.2.0 yesterday](https://pypi.org/project/pyspark/#history) (10/18/2021) that broke our build We should set the bounds for pyspark to `>=2.4.5, <3.2.0` as a hotfix.
non_process
build failed on pyspark description our smoke and unit tests failed on the latest pyspark in staging branch in which platform does it happen in our ci pipelines and in linux virtual machines for details of the errors please visit the following links ado pipeline how do we replicate the issue git clone git checkout staging pip install e pytest tests unit examples test notebooks pyspark py expected behavior i e solution our current dependency bounds for pyspark is however pyspark that broke our build we should set the bounds for pyspark to as a hotfix
0
16,425
21,263,713,327
IssuesEvent
2022-04-13 07:54:07
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
iOS > Create passcode screen is displayed twice after signup/login in below scenario
Bug P2 iOS Process: Fixed Process: Tested QA Process: Tested dev
**Steps:** 1. Login/Signup 2. Create passcode screen is displayed 3. Kill and relaunch the app 4. Create passcode screen is displayed 5. Enter passcode 6. Observe create passcode screen is again displayed **Actual:** Create passcode screen is displayed twice after signup/login in below scenario **Expected:** Create passcode screen should be displayed only once **Refer video:** https://user-images.githubusercontent.com/60386291/150069335-f171708a-16b6-479f-8242-c4001d2bd764.MOV
3.0
iOS > Create passcode screen is displayed twice after signup/login in below scenario - **Steps:** 1. Login/Signup 2. Create passcode screen is displayed 3. Kill and relaunch the app 4. Create passcode screen is displayed 5. Enter passcode 6. Observe create passcode screen is again displayed **Actual:** Create passcode screen is displayed twice after signup/login in below scenario **Expected:** Create passcode screen should be displayed only once **Refer video:** https://user-images.githubusercontent.com/60386291/150069335-f171708a-16b6-479f-8242-c4001d2bd764.MOV
process
ios create passcode screen is displayed twice after signup login in below scenario steps login signup create passcode screen is displayed kill and relaunch the app create passcode screen is displayed enter passcode observe create passcode screen is again displayed actual create passcode screen is displayed twice after signup login in below scenario expected create passcode screen should be displayed only once refer video
1
92,520
15,857,091,421
IssuesEvent
2021-04-08 03:56:50
heholek/better-onetab
https://api.github.com/repos/heholek/better-onetab
closed
CVE-2018-11695 (High) detected in opennms-opennms-source-24.1.2-1 - autoclosed
security vulnerability
## CVE-2018-11695 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>opennmsopennms-source-24.1.2-1</b></p></summary> <p> <p>A Java based fault and performance management system</p> <p>Library home page: <a href=https://sourceforge.net/projects/opennms/>https://sourceforge.net/projects/opennms/</a></p> <p>Found in HEAD commit: <a href="https://github.com/heholek/better-onetab/commit/e8c845acee9d25a919684414aabd7529ac3e9e4b">e8c845acee9d25a919684414aabd7529ac3e9e4b</a></p> </p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Library Source Files (12)</summary> <p></p> <p> * The source files were matched to this source library based on a best effort match. Source libraries are selected from a list of probable public libraries.</p> <p> - /better-onetab/node_modules/node-sass/src/libsass/src/util.hpp - /better-onetab/node_modules/node-sass/src/libsass/src/cssize.cpp - /better-onetab/node_modules/node-sass/src/sass_context_wrapper.cpp - /better-onetab/node_modules/node-sass/src/libsass/src/functions.cpp - /better-onetab/node_modules/node-sass/src/libsass/src/expand.cpp - /better-onetab/node_modules/node-sass/src/libsass/src/prelexer.cpp - /better-onetab/node_modules/node-sass/src/callback_bridge.h - /better-onetab/node_modules/node-sass/src/libsass/src/sass.cpp - /better-onetab/node_modules/node-sass/src/sass_context_wrapper.h - /better-onetab/node_modules/node-sass/src/libsass/src/parser.hpp - /better-onetab/node_modules/node-sass/src/libsass/src/eval.cpp - /better-onetab/node_modules/node-sass/src/libsass/src/debugger.hpp </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in LibSass through 3.5.2. A NULL pointer dereference was found in the function Sass::Expand::operator which could be leveraged by an attacker to cause a denial of service (application crash) or possibly have unspecified other impact. <p>Publish Date: 2018-06-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-11695>CVE-2018-11695</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-11695">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-11695</a></p> <p>Release Date: 2018-06-04</p> <p>Fix Resolution: LibSass - 3.6.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-11695 (High) detected in opennms-opennms-source-24.1.2-1 - autoclosed - ## CVE-2018-11695 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>opennmsopennms-source-24.1.2-1</b></p></summary> <p> <p>A Java based fault and performance management system</p> <p>Library home page: <a href=https://sourceforge.net/projects/opennms/>https://sourceforge.net/projects/opennms/</a></p> <p>Found in HEAD commit: <a href="https://github.com/heholek/better-onetab/commit/e8c845acee9d25a919684414aabd7529ac3e9e4b">e8c845acee9d25a919684414aabd7529ac3e9e4b</a></p> </p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Library Source Files (12)</summary> <p></p> <p> * The source files were matched to this source library based on a best effort match. Source libraries are selected from a list of probable public libraries.</p> <p> - /better-onetab/node_modules/node-sass/src/libsass/src/util.hpp - /better-onetab/node_modules/node-sass/src/libsass/src/cssize.cpp - /better-onetab/node_modules/node-sass/src/sass_context_wrapper.cpp - /better-onetab/node_modules/node-sass/src/libsass/src/functions.cpp - /better-onetab/node_modules/node-sass/src/libsass/src/expand.cpp - /better-onetab/node_modules/node-sass/src/libsass/src/prelexer.cpp - /better-onetab/node_modules/node-sass/src/callback_bridge.h - /better-onetab/node_modules/node-sass/src/libsass/src/sass.cpp - /better-onetab/node_modules/node-sass/src/sass_context_wrapper.h - /better-onetab/node_modules/node-sass/src/libsass/src/parser.hpp - /better-onetab/node_modules/node-sass/src/libsass/src/eval.cpp - /better-onetab/node_modules/node-sass/src/libsass/src/debugger.hpp </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in LibSass through 3.5.2. A NULL pointer dereference was found in the function Sass::Expand::operator which could be leveraged by an attacker to cause a denial of service (application crash) or possibly have unspecified other impact. <p>Publish Date: 2018-06-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-11695>CVE-2018-11695</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-11695">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-11695</a></p> <p>Release Date: 2018-06-04</p> <p>Fix Resolution: LibSass - 3.6.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in opennms opennms source autoclosed cve high severity vulnerability vulnerable library opennmsopennms source a java based fault and performance management system library home page a href found in head commit a href library source files the source files were matched to this source library based on a best effort match source libraries are selected from a list of probable public libraries better onetab node modules node sass src libsass src util hpp better onetab node modules node sass src libsass src cssize cpp better onetab node modules node sass src sass context wrapper cpp better onetab node modules node sass src libsass src functions cpp better onetab node modules node sass src libsass src expand cpp better onetab node modules node sass src libsass src prelexer cpp better onetab node modules node sass src callback bridge h better onetab node modules node sass src libsass src sass cpp better onetab node modules node sass src sass context wrapper h better onetab node modules node sass src libsass src parser hpp better onetab node modules node sass src libsass src eval cpp better onetab node modules node sass src libsass src debugger hpp vulnerability details an issue was discovered in libsass through a null pointer dereference was found in the function sass expand operator which could be leveraged by an attacker to cause a denial of service application crash or possibly have unspecified other impact publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution libsass step up your open source security game with whitesource
0
5,748
8,596,499,709
IssuesEvent
2018-11-15 16:07:51
kiwicom/orbit-components
https://api.github.com/repos/kiwicom/orbit-components
closed
RTL languages
Enhancement Processing
Make sure these boxes are ticked before submitting a new issue: * [x] I have checked that this is not a duplicate * [x] I have added a label categorizing this under BLKN | Kiwicom | Icons | Global I want to start a discussion on the subject of RTL languages and how to solve its support in our components. Maybe @tomashapl do have some experience?
1.0
RTL languages - Make sure these boxes are ticked before submitting a new issue: * [x] I have checked that this is not a duplicate * [x] I have added a label categorizing this under BLKN | Kiwicom | Icons | Global I want to start a discussion on the subject of RTL languages and how to solve its support in our components. Maybe @tomashapl do have some experience?
process
rtl languages make sure these boxes are ticked before submitting a new issue i have checked that this is not a duplicate i have added a label categorizing this under blkn kiwicom icons global i want to start a discussion on the subject of rtl languages and how to solve its support in our components maybe tomashapl do have some experience
1
400,663
27,294,513,126
IssuesEvent
2023-02-23 19:08:46
PepperDash/Essentials
https://api.github.com/repos/PepperDash/Essentials
closed
[FEATURE]-Add card cage to Wiki documentation
documentation enhancement planning
``` { "key": "cardCage1", "uid": 1, "name": "Internal Card Cage", "type": "internalcardcage", "group": "cardCage", "properties": { "cards": { "1": "c3com3", "2": "c3com3", "3": "" } } }, ```
1.0
[FEATURE]-Add card cage to Wiki documentation - ``` { "key": "cardCage1", "uid": 1, "name": "Internal Card Cage", "type": "internalcardcage", "group": "cardCage", "properties": { "cards": { "1": "c3com3", "2": "c3com3", "3": "" } } }, ```
non_process
add card cage to wiki documentation key uid name internal card cage type internalcardcage group cardcage properties cards
0
1,848
4,647,825,102
IssuesEvent
2016-10-01 18:25:41
opentrials/opentrials
https://api.github.com/repos/opentrials/opentrials
closed
Processors document writing procedures are out of sync with DB schema
4. Ready for Review bug Processors
After we extracted the `files` entity from the `documents`, we moved the `url` and `documentcloud_url` from `documents` to `files`. However, we didn't update the processors logic to cope with the new schema.
1.0
Processors document writing procedures are out of sync with DB schema - After we extracted the `files` entity from the `documents`, we moved the `url` and `documentcloud_url` from `documents` to `files`. However, we didn't update the processors logic to cope with the new schema.
process
processors document writing procedures are out of sync with db schema after we extracted the files entity from the documents we moved the url and documentcloud url from documents to files however we didn t update the processors logic to cope with the new schema
1
117,861
17,552,264,443
IssuesEvent
2021-08-13 00:18:54
turkdevops/kiwiirc
https://api.github.com/repos/turkdevops/kiwiirc
closed
CVE-2021-23343 (High) detected in path-parse-1.0.6.tgz - autoclosed
security vulnerability
## CVE-2021-23343 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>path-parse-1.0.6.tgz</b></p></summary> <p>Node.js path.parse() ponyfill</p> <p>Library home page: <a href="https://registry.npmjs.org/path-parse/-/path-parse-1.0.6.tgz">https://registry.npmjs.org/path-parse/-/path-parse-1.0.6.tgz</a></p> <p>Path to dependency file: kiwiirc/package.json</p> <p>Path to vulnerable library: kiwiirc/node_modules/path-parse</p> <p> Dependency Hierarchy: - cli-service-4.3.1.tgz (Root Library) - cli-shared-utils-4.3.1.tgz - read-pkg-5.2.0.tgz - normalize-package-data-2.5.0.tgz - resolve-1.15.1.tgz - :x: **path-parse-1.0.6.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/turkdevops/kiwiirc/commit/b1c301f509befb592ae72f3fc114b5c239f5eb04">b1c301f509befb592ae72f3fc114b5c239f5eb04</a></p> <p>Found in base branch: <b>captcha-on-demand</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> All versions of package path-parse are vulnerable to Regular Expression Denial of Service (ReDoS) via splitDeviceRe, splitTailRe, and splitPathRe regular expressions. ReDoS exhibits polynomial worst-case time complexity. <p>Publish Date: 2021-05-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23343>CVE-2021-23343</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/jbgutierrez/path-parse/issues/8">https://github.com/jbgutierrez/path-parse/issues/8</a></p> <p>Release Date: 2021-05-04</p> <p>Fix Resolution: path-parse - 1.0.7</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-23343 (High) detected in path-parse-1.0.6.tgz - autoclosed - ## CVE-2021-23343 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>path-parse-1.0.6.tgz</b></p></summary> <p>Node.js path.parse() ponyfill</p> <p>Library home page: <a href="https://registry.npmjs.org/path-parse/-/path-parse-1.0.6.tgz">https://registry.npmjs.org/path-parse/-/path-parse-1.0.6.tgz</a></p> <p>Path to dependency file: kiwiirc/package.json</p> <p>Path to vulnerable library: kiwiirc/node_modules/path-parse</p> <p> Dependency Hierarchy: - cli-service-4.3.1.tgz (Root Library) - cli-shared-utils-4.3.1.tgz - read-pkg-5.2.0.tgz - normalize-package-data-2.5.0.tgz - resolve-1.15.1.tgz - :x: **path-parse-1.0.6.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/turkdevops/kiwiirc/commit/b1c301f509befb592ae72f3fc114b5c239f5eb04">b1c301f509befb592ae72f3fc114b5c239f5eb04</a></p> <p>Found in base branch: <b>captcha-on-demand</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> All versions of package path-parse are vulnerable to Regular Expression Denial of Service (ReDoS) via splitDeviceRe, splitTailRe, and splitPathRe regular expressions. ReDoS exhibits polynomial worst-case time complexity. <p>Publish Date: 2021-05-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23343>CVE-2021-23343</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/jbgutierrez/path-parse/issues/8">https://github.com/jbgutierrez/path-parse/issues/8</a></p> <p>Release Date: 2021-05-04</p> <p>Fix Resolution: path-parse - 1.0.7</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in path parse tgz autoclosed cve high severity vulnerability vulnerable library path parse tgz node js path parse ponyfill library home page a href path to dependency file kiwiirc package json path to vulnerable library kiwiirc node modules path parse dependency hierarchy cli service tgz root library cli shared utils tgz read pkg tgz normalize package data tgz resolve tgz x path parse tgz vulnerable library found in head commit a href found in base branch captcha on demand vulnerability details all versions of package path parse are vulnerable to regular expression denial of service redos via splitdevicere splittailre and splitpathre regular expressions redos exhibits polynomial worst case time complexity publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution path parse step up your open source security game with whitesource
0
201,331
7,029,520,814
IssuesEvent
2017-12-26 01:51:41
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
youtube.com - video or audio doesn't play
browser-firefox priority-important
<!-- @browser: Firefox 59.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:59.0) Gecko/20100101 Firefox/59.0 --> <!-- @reported_with: desktop-reporter --> **URL**: https://youtube.com **Browser / Version**: Firefox 59.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes **Problem type**: Video or audio doesn't play **Description**: there is no audio **Steps to Reproduce**: update windows and update nightly [![Screenshot Description](https://webcompat.com/uploads/2017/12/69a011f4-6cc5-40bb-a829-185061093592-thumb.jpg)](https://webcompat.com/uploads/2017/12/69a011f4-6cc5-40bb-a829-185061093592.jpg) _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
youtube.com - video or audio doesn't play - <!-- @browser: Firefox 59.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:59.0) Gecko/20100101 Firefox/59.0 --> <!-- @reported_with: desktop-reporter --> **URL**: https://youtube.com **Browser / Version**: Firefox 59.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes **Problem type**: Video or audio doesn't play **Description**: there is no audio **Steps to Reproduce**: update windows and update nightly [![Screenshot Description](https://webcompat.com/uploads/2017/12/69a011f4-6cc5-40bb-a829-185061093592-thumb.jpg)](https://webcompat.com/uploads/2017/12/69a011f4-6cc5-40bb-a829-185061093592.jpg) _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
youtube com video or audio doesn t play url browser version firefox operating system windows tested another browser yes problem type video or audio doesn t play description there is no audio steps to reproduce update windows and update nightly from with ❤️
0
67,910
28,089,059,377
IssuesEvent
2023-03-30 11:52:02
odoo/odoo
https://api.github.com/repos/odoo/odoo
closed
Probability measuring in Recruitment: No values
Services 11.0 12.0
Impacted versions: 11, 12 **Steps to reproduce:** - To to Recruitment --> Reports --> Measure by Probability **Current behavior:** - No values for Probability despite the filter is there. - No option to enter Probability in Recruitment form. However, the field does exist in the model. `probability = fields.Float("Probability")` **Expected behavior:** - If we plan to use Probability in the pivot report, this field should be visible in the Recruitment form view - If we do not plan to have the Probability measuring feature, please remove this filter in Recruitment > Reports > Measure > Probability (and all related places in Recruitment). Video/Screenshot link (optional): ![probability](https://user-images.githubusercontent.com/10968690/54797778-2b771b00-4c88-11e9-98b3-6a067aa1adc9.png)
1.0
Probability measuring in Recruitment: No values - Impacted versions: 11, 12 **Steps to reproduce:** - To to Recruitment --> Reports --> Measure by Probability **Current behavior:** - No values for Probability despite the filter is there. - No option to enter Probability in Recruitment form. However, the field does exist in the model. `probability = fields.Float("Probability")` **Expected behavior:** - If we plan to use Probability in the pivot report, this field should be visible in the Recruitment form view - If we do not plan to have the Probability measuring feature, please remove this filter in Recruitment > Reports > Measure > Probability (and all related places in Recruitment). Video/Screenshot link (optional): ![probability](https://user-images.githubusercontent.com/10968690/54797778-2b771b00-4c88-11e9-98b3-6a067aa1adc9.png)
non_process
probability measuring in recruitment no values impacted versions steps to reproduce to to recruitment reports measure by probability current behavior no values for probability despite the filter is there no option to enter probability in recruitment form however the field does exist in the model probability fields float probability expected behavior if we plan to use probability in the pivot report this field should be visible in the recruitment form view if we do not plan to have the probability measuring feature please remove this filter in recruitment reports measure probability and all related places in recruitment video screenshot link optional
0
138,433
5,341,385,044
IssuesEvent
2017-02-17 02:38:19
mmisw/mmiorr
https://api.github.com/repos/mmisw/mmiorr
closed
user-customizable mappings in VINE, um, mapping tool
Addressed_in_ORR3 enhancement mmiorr Priority-High vine xdomes
The mapping tool is hard-coded with the 5 SKOS mappings. Some users need to create other mappings on a regular basis. This task is to create the ability for a user to manage their own custom mappings. The ideal scenario would be transferrable to other users, either within the tool or by exporting to a text file (JSON fine) then importing it. (This would also let a user just edit their mapping options offline, then re-import. I think the key fields are 'symbol', property definition (full IRI), short label (?), and description (which goes in the pop-up tip). (Would we need to include the originating property vocabulary and term ID for anything?) The 'symbol' could at first be a letter or special character. A significant improvement is to allow any Unicode, or any emoji, or an icon representation. We need this rather badly, because right now some users are creating 'false mappings' using our 5 SKOS mappings, then mapping those properties to their own (very different) concepts in their own tools. We need to get them to use our tool to do their mappings the right way, or all the other users will be very confused. Reference the last comment of #79, which alludes to this desire: > A general ability to adapt a mapping interface to any kind of properties (not just SKOS properties) would be pretty slick. But is a different topic then we've initially addressed. (Actually, I think this may have been driven originally by the desire to easily add mapping relations to the VINE interface. But this request deserves a separate issue.)
1.0
user-customizable mappings in VINE, um, mapping tool - The mapping tool is hard-coded with the 5 SKOS mappings. Some users need to create other mappings on a regular basis. This task is to create the ability for a user to manage their own custom mappings. The ideal scenario would be transferrable to other users, either within the tool or by exporting to a text file (JSON fine) then importing it. (This would also let a user just edit their mapping options offline, then re-import. I think the key fields are 'symbol', property definition (full IRI), short label (?), and description (which goes in the pop-up tip). (Would we need to include the originating property vocabulary and term ID for anything?) The 'symbol' could at first be a letter or special character. A significant improvement is to allow any Unicode, or any emoji, or an icon representation. We need this rather badly, because right now some users are creating 'false mappings' using our 5 SKOS mappings, then mapping those properties to their own (very different) concepts in their own tools. We need to get them to use our tool to do their mappings the right way, or all the other users will be very confused. Reference the last comment of #79, which alludes to this desire: > A general ability to adapt a mapping interface to any kind of properties (not just SKOS properties) would be pretty slick. But is a different topic then we've initially addressed. (Actually, I think this may have been driven originally by the desire to easily add mapping relations to the VINE interface. But this request deserves a separate issue.)
non_process
user customizable mappings in vine um mapping tool the mapping tool is hard coded with the skos mappings some users need to create other mappings on a regular basis this task is to create the ability for a user to manage their own custom mappings the ideal scenario would be transferrable to other users either within the tool or by exporting to a text file json fine then importing it this would also let a user just edit their mapping options offline then re import i think the key fields are symbol property definition full iri short label and description which goes in the pop up tip would we need to include the originating property vocabulary and term id for anything the symbol could at first be a letter or special character a significant improvement is to allow any unicode or any emoji or an icon representation we need this rather badly because right now some users are creating false mappings using our skos mappings then mapping those properties to their own very different concepts in their own tools we need to get them to use our tool to do their mappings the right way or all the other users will be very confused reference the last comment of which alludes to this desire a general ability to adapt a mapping interface to any kind of properties not just skos properties would be pretty slick but is a different topic then we ve initially addressed actually i think this may have been driven originally by the desire to easily add mapping relations to the vine interface but this request deserves a separate issue
0
7,378
10,514,634,638
IssuesEvent
2019-09-28 02:15:12
metabase/metabase
https://api.github.com/repos/metabase/metabase
closed
When converting Oracle questions to SQL timestamps are converted wrong
.Backend Database/Oracle Priority:P3 Query Processor Type:Bug
Someone whose name rhymes with fam accidentally used `yyyy-MM-dd hh:mm:ss.SSS ZZ` as his format string instead of `yyyy-MM-dd HH:mm:ss.SSS ZZ`. Say you had a "count of checkins" query where the date is on or after September 27th, 2019. The SQL generated would be: ```sql SELECT count(*) FROM checkins WHERE checkin_date >= timestamp '2019-09-27 12:00:00.000 ZZ' ``` Note the hour is `12:00` where we'd expect `00:00` instead. `hh` formats using 12-hour time as opposed to `HH` which uses 24 hour. This is a recent bug because splicing literals into queries when converting to SQL was added in either 0.31 or 0.32.
1.0
When converting Oracle questions to SQL timestamps are converted wrong - Someone whose name rhymes with fam accidentally used `yyyy-MM-dd hh:mm:ss.SSS ZZ` as his format string instead of `yyyy-MM-dd HH:mm:ss.SSS ZZ`. Say you had a "count of checkins" query where the date is on or after September 27th, 2019. The SQL generated would be: ```sql SELECT count(*) FROM checkins WHERE checkin_date >= timestamp '2019-09-27 12:00:00.000 ZZ' ``` Note the hour is `12:00` where we'd expect `00:00` instead. `hh` formats using 12-hour time as opposed to `HH` which uses 24 hour. This is a recent bug because splicing literals into queries when converting to SQL was added in either 0.31 or 0.32.
process
when converting oracle questions to sql timestamps are converted wrong someone whose name rhymes with fam accidentally used yyyy mm dd hh mm ss sss zz as his format string instead of yyyy mm dd hh mm ss sss zz say you had a count of checkins query where the date is on or after september the sql generated would be sql select count from checkins where checkin date timestamp zz note the hour is where we d expect instead hh formats using hour time as opposed to hh which uses hour this is a recent bug because splicing literals into queries when converting to sql was added in either or
1
21,643
30,056,766,255
IssuesEvent
2023-06-28 07:26:42
pydata/pydata-sphinx-theme
https://api.github.com/repos/pydata/pydata-sphinx-theme
opened
rationalize our examples
needs: discussion tag: team process
I already raised this concern earlier last year, we are not consistent in how we show examples and how we create documentation pages in our documentation. I think it is now leading to usage issues with our newest users (https://github.com/pydata/pydata-sphinx-theme/issues/1365). I know that @choldgraf is a very strong advocate of mystnb, I'm not. This issue was about the way we write the documentation and the conclusion was anyone with the will to write down a page should be able to do it either in .rst or .md as we use the mystnb extention. Now I would like to focus on the examples as people may not use mystnb in their doc. Could we show them by default in what sphinx is using e.g. .rst and optionally show how it looks in .md ?
1.0
rationalize our examples - I already raised this concern earlier last year, we are not consistent in how we show examples and how we create documentation pages in our documentation. I think it is now leading to usage issues with our newest users (https://github.com/pydata/pydata-sphinx-theme/issues/1365). I know that @choldgraf is a very strong advocate of mystnb, I'm not. This issue was about the way we write the documentation and the conclusion was anyone with the will to write down a page should be able to do it either in .rst or .md as we use the mystnb extention. Now I would like to focus on the examples as people may not use mystnb in their doc. Could we show them by default in what sphinx is using e.g. .rst and optionally show how it looks in .md ?
process
rationalize our examples i already raised this concern earlier last year we are not consistent in how we show examples and how we create documentation pages in our documentation i think it is now leading to usage issues with our newest users i know that choldgraf is a very strong advocate of mystnb i m not this issue was about the way we write the documentation and the conclusion was anyone with the will to write down a page should be able to do it either in rst or md as we use the mystnb extention now i would like to focus on the examples as people may not use mystnb in their doc could we show them by default in what sphinx is using e g rst and optionally show how it looks in md
1
267,784
28,509,220,438
IssuesEvent
2023-04-19 01:45:51
dpteam/RK3188_TABLET
https://api.github.com/repos/dpteam/RK3188_TABLET
closed
CVE-2014-3673 (High) detected in linuxv3.0 - autoclosed
Mend: dependency security vulnerability
## CVE-2014-3673 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv3.0</b></p></summary> <p> <p>Linux kernel source tree</p> <p>Library home page: <a href=https://github.com/verygreen/linux.git>https://github.com/verygreen/linux.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/dpteam/RK3188_TABLET/commit/0c501f5a0fd72c7b2ac82904235363bd44fd8f9e">0c501f5a0fd72c7b2ac82904235363bd44fd8f9e</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/include/net/sctp/sm.h</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/include/net/sctp/sm.h</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The SCTP implementation in the Linux kernel through 3.17.2 allows remote attackers to cause a denial of service (system crash) via a malformed ASCONF chunk, related to net/sctp/sm_make_chunk.c and net/sctp/sm_statefuns.c. <p>Publish Date: 2014-11-10 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2014-3673>CVE-2014-3673</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2014-3673">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2014-3673</a></p> <p>Release Date: 2014-11-10</p> <p>Fix Resolution: v3.18-rc1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2014-3673 (High) detected in linuxv3.0 - autoclosed - ## CVE-2014-3673 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv3.0</b></p></summary> <p> <p>Linux kernel source tree</p> <p>Library home page: <a href=https://github.com/verygreen/linux.git>https://github.com/verygreen/linux.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/dpteam/RK3188_TABLET/commit/0c501f5a0fd72c7b2ac82904235363bd44fd8f9e">0c501f5a0fd72c7b2ac82904235363bd44fd8f9e</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/include/net/sctp/sm.h</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/include/net/sctp/sm.h</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The SCTP implementation in the Linux kernel through 3.17.2 allows remote attackers to cause a denial of service (system crash) via a malformed ASCONF chunk, related to net/sctp/sm_make_chunk.c and net/sctp/sm_statefuns.c. <p>Publish Date: 2014-11-10 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2014-3673>CVE-2014-3673</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2014-3673">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2014-3673</a></p> <p>Release Date: 2014-11-10</p> <p>Fix Resolution: v3.18-rc1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in autoclosed cve high severity vulnerability vulnerable library linux kernel source tree library home page a href found in head commit a href found in base branch master vulnerable source files include net sctp sm h include net sctp sm h vulnerability details the sctp implementation in the linux kernel through allows remote attackers to cause a denial of service system crash via a malformed asconf chunk related to net sctp sm make chunk c and net sctp sm statefuns c publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
5,512
8,378,603,015
IssuesEvent
2018-10-06 15:58:47
pwittchen/InfiniteScroll
https://api.github.com/repos/pwittchen/InfiniteScroll
closed
Release 0.0.2
release process
**Release notes**: - applied `androidx.appcompat` thanks to @dotkebi - applied `com.google.android.material` thanks to @dotkebi - bumped version of `mockito-core` and `truth` - applied nexus-staging-plugin for gradle - re-organized gradle build config **Things to do**: - [x] bump library version - [x] upload Archives to Maven Central Repository - [x] close and release artifact on Nexus (done experimentally by nexus staging plugin this time) - [x] update gh-pages with JavaDoc - [x] update `CHANGELOG.md` after Maven Sync - [x] update Download section in `README.md` after Maven Sync - [x] create GitHub release
1.0
Release 0.0.2 - **Release notes**: - applied `androidx.appcompat` thanks to @dotkebi - applied `com.google.android.material` thanks to @dotkebi - bumped version of `mockito-core` and `truth` - applied nexus-staging-plugin for gradle - re-organized gradle build config **Things to do**: - [x] bump library version - [x] upload Archives to Maven Central Repository - [x] close and release artifact on Nexus (done experimentally by nexus staging plugin this time) - [x] update gh-pages with JavaDoc - [x] update `CHANGELOG.md` after Maven Sync - [x] update Download section in `README.md` after Maven Sync - [x] create GitHub release
process
release release notes applied androidx appcompat thanks to dotkebi applied com google android material thanks to dotkebi bumped version of mockito core and truth applied nexus staging plugin for gradle re organized gradle build config things to do bump library version upload archives to maven central repository close and release artifact on nexus done experimentally by nexus staging plugin this time update gh pages with javadoc update changelog md after maven sync update download section in readme md after maven sync create github release
1