Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
259,011
19,582,218,190
IssuesEvent
2022-01-04 23:15:10
ZauJulio/Taskiano
https://api.github.com/repos/ZauJulio/Taskiano
opened
Create file COBERTURA_TESTES.md
documentation
Descreva os requisitos necessários e como devemos executar para gerar relatórios de cobertura dos testes. O arquivo deverá estar salvo no diretório docs/. Lembre-se de colocar o link para o documento no README
1.0
Create file COBERTURA_TESTES.md - Descreva os requisitos necessários e como devemos executar para gerar relatórios de cobertura dos testes. O arquivo deverá estar salvo no diretório docs/. Lembre-se de colocar o link para o documento no README
non_process
create file cobertura testes md descreva os requisitos necessários e como devemos executar para gerar relatórios de cobertura dos testes o arquivo deverá estar salvo no diretório docs lembre se de colocar o link para o documento no readme
0
19,172
25,279,561,747
IssuesEvent
2022-11-16 14:54:26
googleapis/google-api-dotnet-client
https://api.github.com/repos/googleapis/google-api-dotnet-client
closed
Reconsider the discovery JSON patcher.
type: process priority: p3
- We are only patching Admin Directory now. - Maybe APIs could fix in origin what we are patching, taking into account whether the changes would be breaking or not, etc. - The patcher is too brittle (see #2243 as an example)
1.0
Reconsider the discovery JSON patcher. - - We are only patching Admin Directory now. - Maybe APIs could fix in origin what we are patching, taking into account whether the changes would be breaking or not, etc. - The patcher is too brittle (see #2243 as an example)
process
reconsider the discovery json patcher we are only patching admin directory now maybe apis could fix in origin what we are patching taking into account whether the changes would be breaking or not etc the patcher is too brittle see as an example
1
3,695
6,722,174,173
IssuesEvent
2017-10-16 14:20:56
fluent/fluent-bit
https://api.github.com/repos/fluent/fluent-bit
closed
Is FLB_TESTS_RUNTIME unnecessary?
work-in-process
FLB_TESTS_RUNTIME is disabled since July. Now cannot compile it. I think the runtime tests are useful for the regression tests of plugins. So I will make that PR if necessary. How you do you feel about it?
1.0
Is FLB_TESTS_RUNTIME unnecessary? - FLB_TESTS_RUNTIME is disabled since July. Now cannot compile it. I think the runtime tests are useful for the regression tests of plugins. So I will make that PR if necessary. How you do you feel about it?
process
is flb tests runtime unnecessary flb tests runtime is disabled since july now cannot compile it i think the runtime tests are useful for the regression tests of plugins so i will make that pr if necessary how you do you feel about it
1
59,428
6,651,031,288
IssuesEvent
2017-09-28 18:30:38
istio/test-infra
https://api.github.com/repos/istio/test-infra
closed
Prow should update pending and failed status to run_after_success jobs.
bug internal-infra-bug oncall test-infra
Mungegithub merges the pr based on second round e2e and first round presubmit presubmit: green e2e: green <--- mungegithub: /test all presubmit: pending e2e: green ---> presubmit succeeds presubmit: green (second round) e2e: green (first round) Racing ----> mungegithub merges the pr (two green while they may not bases on same code) | ---> run_after_succeed rule triggers e2e (second round) presubmit: green e2e: failed !! --- But pr is already merged
1.0
Prow should update pending and failed status to run_after_success jobs. - Mungegithub merges the pr based on second round e2e and first round presubmit presubmit: green e2e: green <--- mungegithub: /test all presubmit: pending e2e: green ---> presubmit succeeds presubmit: green (second round) e2e: green (first round) Racing ----> mungegithub merges the pr (two green while they may not bases on same code) | ---> run_after_succeed rule triggers e2e (second round) presubmit: green e2e: failed !! --- But pr is already merged
non_process
prow should update pending and failed status to run after success jobs mungegithub merges the pr based on second round and first round presubmit presubmit green green mungegithub test all presubmit pending green presubmit succeeds presubmit green second round green first round racing mungegithub merges the pr two green while they may not bases on same code run after succeed rule triggers second round presubmit green failed but pr is already merged
0
133,813
10,864,177,794
IssuesEvent
2019-11-14 16:24:06
dzhw/SLC-IntEr
https://api.github.com/repos/dzhw/SLC-IntEr
closed
Modul Ausbildung: udpate 2
Modul: Ausbildung testing
PV in Modul ausbildung überarbeitet (2 neue Seiten, tc und tr überarbeitet, Antwortoptionen in aus_01 und aus_04). https://github.com/dzhw/SLC-IntEr/commit/739f39363bbc95d6fe2eec58ef485fd579a8629e
1.0
Modul Ausbildung: udpate 2 - PV in Modul ausbildung überarbeitet (2 neue Seiten, tc und tr überarbeitet, Antwortoptionen in aus_01 und aus_04). https://github.com/dzhw/SLC-IntEr/commit/739f39363bbc95d6fe2eec58ef485fd579a8629e
non_process
modul ausbildung udpate pv in modul ausbildung überarbeitet neue seiten tc und tr überarbeitet antwortoptionen in aus und aus
0
273,476
29,820,316,604
IssuesEvent
2023-06-17 01:26:27
pazhanivel07/frameworks_base_2021-0970
https://api.github.com/repos/pazhanivel07/frameworks_base_2021-0970
closed
CVE-2020-0257 (High) detected in baseandroid-10.0.0_r44 - autoclosed
Mend: dependency security vulnerability
## CVE-2020-0257 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>baseandroid-10.0.0_r44</b></p></summary> <p> <p>Android framework classes and services</p> <p>Library home page: <a href=https://android.googlesource.com/platform/frameworks/base>https://android.googlesource.com/platform/frameworks/base</a></p> <p>Found in HEAD commit: <a href="https://github.com/pazhanivel07/frameworks_base_2021-0970/commit/ad3ed522c9ac4d72ed6d51ed523780ad73330cbe">ad3ed522c9ac4d72ed6d51ed523780ad73330cbe</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/core/jni/com_android_internal_os_Zygote.cpp</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> In SpecializeCommon of com_android_internal_os_Zygote.cpp, there is a permissions bypass due to an incomplete cleanup. This could lead to local escalation of privilege in isolated processes with no additional execution privileges needed. User interaction is not needed for exploitation.Product: AndroidVersions: Android-10Android ID: A-156741968 <p>Publish Date: 2020-08-11 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-0257>CVE-2020-0257</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://source.android.com/security/bulletin/2020-08-01">https://source.android.com/security/bulletin/2020-08-01</a></p> <p>Release Date: 2020-07-21</p> <p>Fix Resolution: android-10.0.0_r41</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-0257 (High) detected in baseandroid-10.0.0_r44 - autoclosed - ## CVE-2020-0257 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>baseandroid-10.0.0_r44</b></p></summary> <p> <p>Android framework classes and services</p> <p>Library home page: <a href=https://android.googlesource.com/platform/frameworks/base>https://android.googlesource.com/platform/frameworks/base</a></p> <p>Found in HEAD commit: <a href="https://github.com/pazhanivel07/frameworks_base_2021-0970/commit/ad3ed522c9ac4d72ed6d51ed523780ad73330cbe">ad3ed522c9ac4d72ed6d51ed523780ad73330cbe</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/core/jni/com_android_internal_os_Zygote.cpp</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> In SpecializeCommon of com_android_internal_os_Zygote.cpp, there is a permissions bypass due to an incomplete cleanup. This could lead to local escalation of privilege in isolated processes with no additional execution privileges needed. User interaction is not needed for exploitation.Product: AndroidVersions: Android-10Android ID: A-156741968 <p>Publish Date: 2020-08-11 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-0257>CVE-2020-0257</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://source.android.com/security/bulletin/2020-08-01">https://source.android.com/security/bulletin/2020-08-01</a></p> <p>Release Date: 2020-07-21</p> <p>Fix Resolution: android-10.0.0_r41</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in baseandroid autoclosed cve high severity vulnerability vulnerable library baseandroid android framework classes and services library home page a href found in head commit a href found in base branch master vulnerable source files core jni com android internal os zygote cpp vulnerability details in specializecommon of com android internal os zygote cpp there is a permissions bypass due to an incomplete cleanup this could lead to local escalation of privilege in isolated processes with no additional execution privileges needed user interaction is not needed for exploitation product androidversions android id a publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution android step up your open source security game with mend
0
10,617
13,439,054,753
IssuesEvent
2020-09-07 19:53:18
timberio/vector
https://api.github.com/repos/timberio/vector
closed
Fingerprint/Hashing transform
domain: processing have: should type: enhancement
### Current Vector Version <!--- Run `vector --version` to show the version, and paste the result between the ``` marks below. This will record which version was current at the time of your feature request, to help manage the request backlog. If you're not using the latest version, please check to see if something related to your request has already been implemented in a later version. --> ``` ❯ vector --version vector 0.10.0 (g0f0311a x86_64-unknown-linux-gnu 2020-07-22) ``` ### Use-cases <!--- In order to properly evaluate a feature request, it is necessary to understand the use-cases for it. Please describe below the _end goal_ you are trying to achieve that has led you to request this feature. Please keep this section focused on the problem and not on the suggested solution. We'll get to that in a moment, below! --> I'm evaluating vector as a replacement for logstash/fluentd at my company and we currently use [the fingerprint plugin](https://www.elastic.co/guide/en/logstash/current/plugins-filters-fingerprint.html) that logstash provides (with fluentd offering similar [here](https://github.com/mtakemi/fluent-plugin-genhashvalue). This helps us avoid duplicated events in elasticsearch if we ever need to replay messages. ### Attempted Solutions <!--- If you've already tried to solve the problem within Vector's existing features and found a limitation that prevented you from succeeding, please describe it below in as much detail as possible. Ideally, this would include real configuration snippets that you tried and what results you got in each case. Please remove any sensitive information such as passwords before sharing configuration snippets and command lines. ---> This could likely be written in lua or done with the wasm transform but it could be useful for users migrating to vector (also possibly suggested already here #1129) ### Proposal <!--- If you have an idea for a way to address the problem via a change to Vector features, please describe it below. In this section, it's helpful to include specific examples of how what you are suggesting might look in configuration files, or on the command line, since that allows us to understand the full picture of what you are proposing. If you're not sure of some details, don't worry! When we evaluate the feature request we may suggest modifications as necessary to work within the design constraints of Vector. --> Provide a built-in transform that allows users to hash with common hashing functions. ### References <!-- Are there any other GitHub issues, whether open or closed, that are related to the problem you've described above or to the suggested solution? If so, please create a list below that mentions each of them. For example: - #7023 -->
1.0
Fingerprint/Hashing transform - ### Current Vector Version <!--- Run `vector --version` to show the version, and paste the result between the ``` marks below. This will record which version was current at the time of your feature request, to help manage the request backlog. If you're not using the latest version, please check to see if something related to your request has already been implemented in a later version. --> ``` ❯ vector --version vector 0.10.0 (g0f0311a x86_64-unknown-linux-gnu 2020-07-22) ``` ### Use-cases <!--- In order to properly evaluate a feature request, it is necessary to understand the use-cases for it. Please describe below the _end goal_ you are trying to achieve that has led you to request this feature. Please keep this section focused on the problem and not on the suggested solution. We'll get to that in a moment, below! --> I'm evaluating vector as a replacement for logstash/fluentd at my company and we currently use [the fingerprint plugin](https://www.elastic.co/guide/en/logstash/current/plugins-filters-fingerprint.html) that logstash provides (with fluentd offering similar [here](https://github.com/mtakemi/fluent-plugin-genhashvalue). This helps us avoid duplicated events in elasticsearch if we ever need to replay messages. ### Attempted Solutions <!--- If you've already tried to solve the problem within Vector's existing features and found a limitation that prevented you from succeeding, please describe it below in as much detail as possible. Ideally, this would include real configuration snippets that you tried and what results you got in each case. Please remove any sensitive information such as passwords before sharing configuration snippets and command lines. ---> This could likely be written in lua or done with the wasm transform but it could be useful for users migrating to vector (also possibly suggested already here #1129) ### Proposal <!--- If you have an idea for a way to address the problem via a change to Vector features, please describe it below. In this section, it's helpful to include specific examples of how what you are suggesting might look in configuration files, or on the command line, since that allows us to understand the full picture of what you are proposing. If you're not sure of some details, don't worry! When we evaluate the feature request we may suggest modifications as necessary to work within the design constraints of Vector. --> Provide a built-in transform that allows users to hash with common hashing functions. ### References <!-- Are there any other GitHub issues, whether open or closed, that are related to the problem you've described above or to the suggested solution? If so, please create a list below that mentions each of them. For example: - #7023 -->
process
fingerprint hashing transform current vector version run vector version to show the version and paste the result between the marks below this will record which version was current at the time of your feature request to help manage the request backlog if you re not using the latest version please check to see if something related to your request has already been implemented in a later version ❯ vector version vector unknown linux gnu use cases in order to properly evaluate a feature request it is necessary to understand the use cases for it please describe below the end goal you are trying to achieve that has led you to request this feature please keep this section focused on the problem and not on the suggested solution we ll get to that in a moment below i m evaluating vector as a replacement for logstash fluentd at my company and we currently use that logstash provides with fluentd offering similar this helps us avoid duplicated events in elasticsearch if we ever need to replay messages attempted solutions if you ve already tried to solve the problem within vector s existing features and found a limitation that prevented you from succeeding please describe it below in as much detail as possible ideally this would include real configuration snippets that you tried and what results you got in each case please remove any sensitive information such as passwords before sharing configuration snippets and command lines this could likely be written in lua or done with the wasm transform but it could be useful for users migrating to vector also possibly suggested already here proposal if you have an idea for a way to address the problem via a change to vector features please describe it below in this section it s helpful to include specific examples of how what you are suggesting might look in configuration files or on the command line since that allows us to understand the full picture of what you are proposing if you re not sure of some details don t worry when we evaluate the feature request we may suggest modifications as necessary to work within the design constraints of vector provide a built in transform that allows users to hash with common hashing functions references are there any other github issues whether open or closed that are related to the problem you ve described above or to the suggested solution if so please create a list below that mentions each of them for example
1
9,175
6,152,928,938
IssuesEvent
2017-06-28 08:41:23
raiden-network/raiden
https://api.github.com/repos/raiden-network/raiden
closed
Raiden RESTful API
usability
*Expose the API of raiden to external clients.* ## General API Spec: - **Setting up** the initial connection to the raiden network should be possible **manually** (by opening channels and depositing into them from single API methods) - Additionally we want to provide a **(semi-)automatic setup** to the network from given configuration-parameters *(e.g. based on available assets, so that the node takes care of an optimal connectivity to the raiden network by opening and depositing into channels without any additional considerations from the user)* - API calls and responses should include a specification of the **API-version** (up from v1.0), so that the API is **future-proof**. The app has to include all versions of the API in the future. - **MvP**: this version of the API needs to provide all neccessary tools to make it possible to build a **Wallet style UI** (issue #285) ## Functionality add following methods to `raiden-service.RaidenAPI`: - `query_endpoint(raiden_address)` - for light clients - `register_endpoint(endpoint, signature)` - `get_channel_info()` - retrieve the timeout, deposit, etc let the caller decide if the called method should be executed asynchronous or waiting (== `gevent.Greenlet.wait()`) - `asset_swap(..., wait=<bool>)` - `expect_asset_swap(..., wait=<bool>)` - `transfer(..., wait=<bool>)` The asynchronous method should be the default option, but when calling via REST the waiting method should be default. That's because in e.g. JavaScript one can work easily with Promises. ## Protocol: - Endpoints of all methods in the `raiden_service.RaidenAPI` are accessible through http-requests and are structured complying to a **RESTful**-pattern. Resources are kept minimal: - (e.g. on `/api/channel` -`GET` - get channel info, -`POST` - open channel, -`DELETE` - close channel, -`UPDATE` - deposit to channel ) - All http-requests (*-> client-initiated*) are *'blocking'* and at least include information about the **success of the call** in their response message. That way we don't have to work with callback-ids on the client side (see #280 ) but rather can work with promises. - For **non-client initiated** actions (e.g. incoming transfers, opened channels, channel closing etc.) there will be the **emission of events** from the node to the client - **Filtering** of the events will be done **on the client side**, so all events will be put in a FIFO-queue and sequentially forwarded to the client - We assume that **only 1 client per node** is connected - Events and responses will be **serialized to JSON**, 'difficult' data encoding will be reused from the Ethereum protocol (pyethapp/pyethereum) where possible (e.g. hex encoding for addresses) ## Transport: - Method-calls: **TCP** (http-requests) - Events: the client can choose between 1) TCP (**polling** http-requests) 2) TCP ( **long-polling**; *'fake' push notifications*) - The **interface** for the API has to be **uniform**, independent of the chosen transport-layer ## Tasklist - [x] [Expose blockchain events in the REST api](https://github.com/raiden-network/raiden/issues/467) - [x] [Add http endpoint for exchange task](https://github.com/raiden-network/raiden/issues/465) - [x] [Expose Raiden Rest API at Raiden start](https://github.com/raiden-network/raiden/issues/464) - [ ] [Expose Raiden internal events from the REST API](https://github.com/raiden-network/raiden/issues/466) --> #594 #466 - [x] [Add rpc endpoints for `connect` and `leave` of token networks](https://github.com/raiden-network/raiden/issues/508)
True
Raiden RESTful API - *Expose the API of raiden to external clients.* ## General API Spec: - **Setting up** the initial connection to the raiden network should be possible **manually** (by opening channels and depositing into them from single API methods) - Additionally we want to provide a **(semi-)automatic setup** to the network from given configuration-parameters *(e.g. based on available assets, so that the node takes care of an optimal connectivity to the raiden network by opening and depositing into channels without any additional considerations from the user)* - API calls and responses should include a specification of the **API-version** (up from v1.0), so that the API is **future-proof**. The app has to include all versions of the API in the future. - **MvP**: this version of the API needs to provide all neccessary tools to make it possible to build a **Wallet style UI** (issue #285) ## Functionality add following methods to `raiden-service.RaidenAPI`: - `query_endpoint(raiden_address)` - for light clients - `register_endpoint(endpoint, signature)` - `get_channel_info()` - retrieve the timeout, deposit, etc let the caller decide if the called method should be executed asynchronous or waiting (== `gevent.Greenlet.wait()`) - `asset_swap(..., wait=<bool>)` - `expect_asset_swap(..., wait=<bool>)` - `transfer(..., wait=<bool>)` The asynchronous method should be the default option, but when calling via REST the waiting method should be default. That's because in e.g. JavaScript one can work easily with Promises. ## Protocol: - Endpoints of all methods in the `raiden_service.RaidenAPI` are accessible through http-requests and are structured complying to a **RESTful**-pattern. Resources are kept minimal: - (e.g. on `/api/channel` -`GET` - get channel info, -`POST` - open channel, -`DELETE` - close channel, -`UPDATE` - deposit to channel ) - All http-requests (*-> client-initiated*) are *'blocking'* and at least include information about the **success of the call** in their response message. That way we don't have to work with callback-ids on the client side (see #280 ) but rather can work with promises. - For **non-client initiated** actions (e.g. incoming transfers, opened channels, channel closing etc.) there will be the **emission of events** from the node to the client - **Filtering** of the events will be done **on the client side**, so all events will be put in a FIFO-queue and sequentially forwarded to the client - We assume that **only 1 client per node** is connected - Events and responses will be **serialized to JSON**, 'difficult' data encoding will be reused from the Ethereum protocol (pyethapp/pyethereum) where possible (e.g. hex encoding for addresses) ## Transport: - Method-calls: **TCP** (http-requests) - Events: the client can choose between 1) TCP (**polling** http-requests) 2) TCP ( **long-polling**; *'fake' push notifications*) - The **interface** for the API has to be **uniform**, independent of the chosen transport-layer ## Tasklist - [x] [Expose blockchain events in the REST api](https://github.com/raiden-network/raiden/issues/467) - [x] [Add http endpoint for exchange task](https://github.com/raiden-network/raiden/issues/465) - [x] [Expose Raiden Rest API at Raiden start](https://github.com/raiden-network/raiden/issues/464) - [ ] [Expose Raiden internal events from the REST API](https://github.com/raiden-network/raiden/issues/466) --> #594 #466 - [x] [Add rpc endpoints for `connect` and `leave` of token networks](https://github.com/raiden-network/raiden/issues/508)
non_process
raiden restful api expose the api of raiden to external clients general api spec setting up the initial connection to the raiden network should be possible manually by opening channels and depositing into them from single api methods additionally we want to provide a semi automatic setup to the network from given configuration parameters e g based on available assets so that the node takes care of an optimal connectivity to the raiden network by opening and depositing into channels without any additional considerations from the user api calls and responses should include a specification of the api version up from so that the api is future proof the app has to include all versions of the api in the future mvp this version of the api needs to provide all neccessary tools to make it possible to build a wallet style ui issue functionality add following methods to raiden service raidenapi query endpoint raiden address for light clients register endpoint endpoint signature get channel info retrieve the timeout deposit etc let the caller decide if the called method should be executed asynchronous or waiting gevent greenlet wait asset swap wait expect asset swap wait transfer wait the asynchronous method should be the default option but when calling via rest the waiting method should be default that s because in e g javascript one can work easily with promises protocol endpoints of all methods in the raiden service raidenapi are accessible through http requests and are structured complying to a restful pattern resources are kept minimal e g on api channel get get channel info post open channel delete close channel update deposit to channel all http requests client initiated are blocking and at least include information about the success of the call in their response message that way we don t have to work with callback ids on the client side see but rather can work with promises for non client initiated actions e g incoming transfers opened channels channel closing etc there will be the emission of events from the node to the client filtering of the events will be done on the client side so all events will be put in a fifo queue and sequentially forwarded to the client we assume that only client per node is connected events and responses will be serialized to json difficult data encoding will be reused from the ethereum protocol pyethapp pyethereum where possible e g hex encoding for addresses transport method calls tcp http requests events the client can choose between tcp polling http requests tcp long polling fake push notifications the interface for the api has to be uniform independent of the chosen transport layer tasklist
0
42,760
9,306,950,909
IssuesEvent
2019-03-25 10:58:36
gbif/portal16
https://api.github.com/repos/gbif/portal16
closed
Delete menu.json from translations
code maintenance
it appears though that it is used as headlines a few places. For example `menu.nameParser` probably belongs better with the nameparser tool translations. Either way, those translation keys that aren't used should be deleted.
1.0
Delete menu.json from translations - it appears though that it is used as headlines a few places. For example `menu.nameParser` probably belongs better with the nameparser tool translations. Either way, those translation keys that aren't used should be deleted.
non_process
delete menu json from translations it appears though that it is used as headlines a few places for example menu nameparser probably belongs better with the nameparser tool translations either way those translation keys that aren t used should be deleted
0
22,411
31,142,293,235
IssuesEvent
2023-08-16 01:44:51
cypress-io/cypress
https://api.github.com/repos/cypress-io/cypress
closed
Flaky test: Error: done() called multiple times III
OS: linux process: flaky test topic: flake ❄️ stage: flake "topic: done()" stale
### Link to dashboard or CircleCI failure https://app.circleci.com/pipelines/github/cypress-io/cypress/41414/workflows/d24eebfe-e9ce-4304-8063-85d8a9b8b296/jobs/1715026 ### Link to failing test in GitHub https://github.com/cypress-io/cypress/blob/develop/packages/driver/cypress/e2e/commands/net_stubbing.cy.ts#L1946 ### Analysis done() called multiple times ### Cypress Version 10.4.0 ### Other Search for this issue number in the codebase to find the test(s) skipped until this issue is fixed
1.0
Flaky test: Error: done() called multiple times III - ### Link to dashboard or CircleCI failure https://app.circleci.com/pipelines/github/cypress-io/cypress/41414/workflows/d24eebfe-e9ce-4304-8063-85d8a9b8b296/jobs/1715026 ### Link to failing test in GitHub https://github.com/cypress-io/cypress/blob/develop/packages/driver/cypress/e2e/commands/net_stubbing.cy.ts#L1946 ### Analysis done() called multiple times ### Cypress Version 10.4.0 ### Other Search for this issue number in the codebase to find the test(s) skipped until this issue is fixed
process
flaky test error done called multiple times iii link to dashboard or circleci failure link to failing test in github analysis done called multiple times cypress version other search for this issue number in the codebase to find the test s skipped until this issue is fixed
1
197,828
22,606,294,771
IssuesEvent
2022-06-29 13:32:02
elikkatzgit/TestingPOM
https://api.github.com/repos/elikkatzgit/TestingPOM
closed
CVE-2018-11307 (High) detected in jackson-databind-2.7.2.jar - autoclosed
security vulnerability
## CVE-2018-11307 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.7.2.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.7.2.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/elikkatzgit/TestingPOM/commit/349fffeed7cf25f2cf5b8b6a05b5e4367130406e">349fffeed7cf25f2cf5b8b6a05b5e4367130406e</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in FasterXML jackson-databind 2.0.0 through 2.9.5. Use of Jackson default typing along with a gadget class from iBatis allows exfiltration of content. Fixed in 2.7.9.4, 2.8.11.2, and 2.9.6. <p>Publish Date: 2019-07-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-11307>CVE-2018-11307</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2032">https://github.com/FasterXML/jackson-databind/issues/2032</a></p> <p>Release Date: 2019-07-09</p> <p>Fix Resolution: 2.7.9.4</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.7.2","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.7.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.7.9.4","isBinary":true}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2018-11307","vulnerabilityDetails":"An issue was discovered in FasterXML jackson-databind 2.0.0 through 2.9.5. Use of Jackson default typing along with a gadget class from iBatis allows exfiltration of content. Fixed in 2.7.9.4, 2.8.11.2, and 2.9.6.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-11307","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2018-11307 (High) detected in jackson-databind-2.7.2.jar - autoclosed - ## CVE-2018-11307 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.7.2.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.7.2.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/elikkatzgit/TestingPOM/commit/349fffeed7cf25f2cf5b8b6a05b5e4367130406e">349fffeed7cf25f2cf5b8b6a05b5e4367130406e</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in FasterXML jackson-databind 2.0.0 through 2.9.5. Use of Jackson default typing along with a gadget class from iBatis allows exfiltration of content. Fixed in 2.7.9.4, 2.8.11.2, and 2.9.6. <p>Publish Date: 2019-07-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-11307>CVE-2018-11307</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2032">https://github.com/FasterXML/jackson-databind/issues/2032</a></p> <p>Release Date: 2019-07-09</p> <p>Fix Resolution: 2.7.9.4</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.7.2","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.7.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.7.9.4","isBinary":true}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2018-11307","vulnerabilityDetails":"An issue was discovered in FasterXML jackson-databind 2.0.0 through 2.9.5. Use of Jackson default typing along with a gadget class from iBatis allows exfiltration of content. Fixed in 2.7.9.4, 2.8.11.2, and 2.9.6.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-11307","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_process
cve high detected in jackson databind jar autoclosed cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href dependency hierarchy x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details an issue was discovered in fasterxml jackson databind through use of jackson default typing along with a gadget class from ibatis allows exfiltration of content fixed in and publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion isbinary true basebranches vulnerabilityidentifier cve vulnerabilitydetails an issue was discovered in fasterxml jackson databind through use of jackson default typing along with a gadget class from ibatis allows exfiltration of content fixed in and vulnerabilityurl
0
11,386
14,223,132,755
IssuesEvent
2020-11-17 17:47:34
JuliaParallel/Dagger.jl
https://api.github.com/repos/JuliaParallel/Dagger.jl
opened
Remove scheduler plugin machinery, make Sch programmable
processors scheduler
The scheduler plugin system is (to my knowledge) unused by any modern users of Dagger. However, the presence of the potential for multiple external schedulers makes changing the scheduler API technically a breaking change, which would slow Dagger's development were that actually a concern. Of course, we still want to support different kinds of scheduling algorithms and optimizations that best suit a user's use case and DAG structure, so we should make the Sch scheduler user-programmable by making the current worker pressure algorithm optional, and having `schedule!` call a user-defined callable. The option to change the internal scheduler algorithm is intended to only be used by adventurous users who understand that deadlocks/livelocks/hangs/etc. are all possible when changing the default scheduler, although we should expose "safe" semi-internal APIs that can perform common tasks correctly (which the default scheduler should also use whenever possible). Hopefully this change can spur developers and adventurous users to experiment with (and contribute) new scheduling algorithms which are better than the default scheduler for certain classes of workloads, with Dagger itself becoming the foundation for user-defined scheduling of distributed Julia code.
1.0
Remove scheduler plugin machinery, make Sch programmable - The scheduler plugin system is (to my knowledge) unused by any modern users of Dagger. However, the presence of the potential for multiple external schedulers makes changing the scheduler API technically a breaking change, which would slow Dagger's development were that actually a concern. Of course, we still want to support different kinds of scheduling algorithms and optimizations that best suit a user's use case and DAG structure, so we should make the Sch scheduler user-programmable by making the current worker pressure algorithm optional, and having `schedule!` call a user-defined callable. The option to change the internal scheduler algorithm is intended to only be used by adventurous users who understand that deadlocks/livelocks/hangs/etc. are all possible when changing the default scheduler, although we should expose "safe" semi-internal APIs that can perform common tasks correctly (which the default scheduler should also use whenever possible). Hopefully this change can spur developers and adventurous users to experiment with (and contribute) new scheduling algorithms which are better than the default scheduler for certain classes of workloads, with Dagger itself becoming the foundation for user-defined scheduling of distributed Julia code.
process
remove scheduler plugin machinery make sch programmable the scheduler plugin system is to my knowledge unused by any modern users of dagger however the presence of the potential for multiple external schedulers makes changing the scheduler api technically a breaking change which would slow dagger s development were that actually a concern of course we still want to support different kinds of scheduling algorithms and optimizations that best suit a user s use case and dag structure so we should make the sch scheduler user programmable by making the current worker pressure algorithm optional and having schedule call a user defined callable the option to change the internal scheduler algorithm is intended to only be used by adventurous users who understand that deadlocks livelocks hangs etc are all possible when changing the default scheduler although we should expose safe semi internal apis that can perform common tasks correctly which the default scheduler should also use whenever possible hopefully this change can spur developers and adventurous users to experiment with and contribute new scheduling algorithms which are better than the default scheduler for certain classes of workloads with dagger itself becoming the foundation for user defined scheduling of distributed julia code
1
4,566
7,393,778,888
IssuesEvent
2018-03-17 01:41:23
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Java "Support"
app-service cxp doc-bug in-process triaged
According to the following more recently updated document for Functions, Java is "not supported" for WebJobs. https://docs.microsoft.com/en-us/azure/azure-functions/functions-compare-logic-apps-ms-flow-webjobs#functions-vs-webjobs However, this document lists .jar (using Java) as a supported script type. This seems to be contradictory information, which one is correct? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 2af36172-e046-c87e-2171-cb55dfc8a8c5 * Version Independent ID: a6cdf0ce-973f-713e-984a-e5f5a08afa57 * Content: [Run Background tasks with WebJobs in Azure App Service | Microsoft Docs](https://docs.microsoft.com/en-us/azure/app-service/web-sites-create-web-jobs) * Content Source: [articles/app-service/web-sites-create-web-jobs.md](https://github.com/Microsoft/azure-docs/blob/master/articles/app-service/web-sites-create-web-jobs.md) * Service: **app-service** * GitHub Login: @ggailey777 * Microsoft Alias: **glenga;david.ebbo;suwatch;pbatum;naren.soni**
1.0
Java "Support" - According to the following more recently updated document for Functions, Java is "not supported" for WebJobs. https://docs.microsoft.com/en-us/azure/azure-functions/functions-compare-logic-apps-ms-flow-webjobs#functions-vs-webjobs However, this document lists .jar (using Java) as a supported script type. This seems to be contradictory information, which one is correct? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 2af36172-e046-c87e-2171-cb55dfc8a8c5 * Version Independent ID: a6cdf0ce-973f-713e-984a-e5f5a08afa57 * Content: [Run Background tasks with WebJobs in Azure App Service | Microsoft Docs](https://docs.microsoft.com/en-us/azure/app-service/web-sites-create-web-jobs) * Content Source: [articles/app-service/web-sites-create-web-jobs.md](https://github.com/Microsoft/azure-docs/blob/master/articles/app-service/web-sites-create-web-jobs.md) * Service: **app-service** * GitHub Login: @ggailey777 * Microsoft Alias: **glenga;david.ebbo;suwatch;pbatum;naren.soni**
process
java support according to the following more recently updated document for functions java is not supported for webjobs however this document lists jar using java as a supported script type this seems to be contradictory information which one is correct document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service app service github login microsoft alias glenga david ebbo suwatch pbatum naren soni
1
28,548
8,178,099,851
IssuesEvent
2018-08-28 12:54:56
nbs-system/snuffleupagus
https://api.github.com/repos/nbs-system/snuffleupagus
closed
Warnings when building the Debian package
build system good first issue help wanted
``` make[2]: Leaving directory '/home/jvoisin/Dev/snuffleupagus' dh_installdocs dh_installchangelogs dh_installexamples dh_perl dh_link dh_strip_nondeterminism dh_compress dh_fixperms dh_missing dh_strip dh_makeshlibs dh_shlibdeps dpkg-shlibdeps: warning: debian/snuffleupagus/usr/lib/php/20160303/snuffleupagus.so contains an unresolvable reference to symbol zend_register_ini_entries: it's probably a plugin dpkg-shlibdeps: warning: 67 other similar warnings have been skipped (use -v to see them all) dh_installdeb dh_gencontrol dpkg-gencontrol: warning: package snuffleupagus: unused substitution variable ${shlibs:Depends} dpkg-gencontrol: warning: package snuffleupagus: unused substitution variable ${shlibs:Depends} dh_md5sums dh_builddeb dpkg-deb: building package 'snuffleupagus' in '../snuffleupagus_0.2.1_amd64.deb'. dpkg-deb: building package 'snuffleupagus-dbgsym' in 'debian/.debhelper/scratch-space/build-snu ```
1.0
Warnings when building the Debian package - ``` make[2]: Leaving directory '/home/jvoisin/Dev/snuffleupagus' dh_installdocs dh_installchangelogs dh_installexamples dh_perl dh_link dh_strip_nondeterminism dh_compress dh_fixperms dh_missing dh_strip dh_makeshlibs dh_shlibdeps dpkg-shlibdeps: warning: debian/snuffleupagus/usr/lib/php/20160303/snuffleupagus.so contains an unresolvable reference to symbol zend_register_ini_entries: it's probably a plugin dpkg-shlibdeps: warning: 67 other similar warnings have been skipped (use -v to see them all) dh_installdeb dh_gencontrol dpkg-gencontrol: warning: package snuffleupagus: unused substitution variable ${shlibs:Depends} dpkg-gencontrol: warning: package snuffleupagus: unused substitution variable ${shlibs:Depends} dh_md5sums dh_builddeb dpkg-deb: building package 'snuffleupagus' in '../snuffleupagus_0.2.1_amd64.deb'. dpkg-deb: building package 'snuffleupagus-dbgsym' in 'debian/.debhelper/scratch-space/build-snu ```
non_process
warnings when building the debian package make leaving directory home jvoisin dev snuffleupagus dh installdocs dh installchangelogs dh installexamples dh perl dh link dh strip nondeterminism dh compress dh fixperms dh missing dh strip dh makeshlibs dh shlibdeps dpkg shlibdeps warning debian snuffleupagus usr lib php snuffleupagus so contains an unresolvable reference to symbol zend register ini entries it s probably a plugin dpkg shlibdeps warning other similar warnings have been skipped use v to see them all dh installdeb dh gencontrol dpkg gencontrol warning package snuffleupagus unused substitution variable shlibs depends dpkg gencontrol warning package snuffleupagus unused substitution variable shlibs depends dh dh builddeb dpkg deb building package snuffleupagus in snuffleupagus deb dpkg deb building package snuffleupagus dbgsym in debian debhelper scratch space build snu
0
578,575
17,148,283,837
IssuesEvent
2021-07-13 17:01:46
googleapis/repo-automation-bots
https://api.github.com/repos/googleapis/repo-automation-bots
opened
owl-bot frontend dies with "Error: 14 UNAVAILABLE: read ECONNRESET"
bot: owl-bot priority: p2
Starring as P2 because it's very rare. The error log: ``` { "insertId": "60ed5b52000534eabd06fd2b", "jsonPayload": { "details": "read ECONNRESET", "message": "14 UNAVAILABLE: read ECONNRESET", "metadata": { "options": {}, "internalRepr": {} }, "code": 14, "type": "Error", "trigger": { "trigger_source_repo": { "owner_type": "Organization", "repo_name": "google-resumable-media-python", "url": "https://github.com/googleapis/google-resumable-media-python", "owner": "googleapis" }, "trigger_sender": "gcf-merge-on-green[bot]", "github_event_type": "pull_request.unlabeled", "github_delivery_guid": "d602a860-e3bb-11eb-9f5f-01f9e172f02b", "trigger_type": "GitHub Webhook", "payload_hash": "c4d522aac65d2bb87ffdecd9f7c49869" }, "level": 50, "stack": "Error: 14 UNAVAILABLE: read ECONNRESET\n at Object.callErrorFromStatus (/usr/src/app/node_modules/@grpc/grpc-js/build/src/call.js:31:26)\n at Object.onReceiveStatus (/usr/src/app/node_modules/@grpc/grpc-js/build/src/client.js:179:52)\n at Object.onReceiveStatus (/usr/src/app/node_modules/@grpc/grpc-js/build/src/client-interceptors.js:336:141)\n at Object.onReceiveStatus (/usr/src/app/node_modules/@grpc/grpc-js/build/src/client-interceptors.js:299:181)\n at /usr/src/app/node_modules/@grpc/grpc-js/build/src/call-stream.js:145:78\n at processTicksAndRejections (internal/process/task_queues.js:77:11)" }, "resource": { "type": "cloud_run_revision", "labels": { "project_id": "repo-automation-bots", "revision_name": "owl-bot-00007-zag", "location": "us-central1", "configuration_name": "owl-bot", "service_name": "owl-bot" } }, "timestamp": "2021-07-13T09:22:26.341226Z", "severity": "ERROR", "labels": { "instanceId": "00bf4bf02d480c07e7ac82c2a94eedcb21772fc9b421079001923328aadc9c17206d6cbdd09c42e99eaf03bbfd7759770d2a1010e3d0dd86a84eb0f6ed" }, "logName": "projects/repo-automation-bots/logs/run.googleapis.com%2Fstdout", "receiveTimestamp": "2021-07-13T09:22:26.409211893Z" } ```
1.0
owl-bot frontend dies with "Error: 14 UNAVAILABLE: read ECONNRESET" - Starring as P2 because it's very rare. The error log: ``` { "insertId": "60ed5b52000534eabd06fd2b", "jsonPayload": { "details": "read ECONNRESET", "message": "14 UNAVAILABLE: read ECONNRESET", "metadata": { "options": {}, "internalRepr": {} }, "code": 14, "type": "Error", "trigger": { "trigger_source_repo": { "owner_type": "Organization", "repo_name": "google-resumable-media-python", "url": "https://github.com/googleapis/google-resumable-media-python", "owner": "googleapis" }, "trigger_sender": "gcf-merge-on-green[bot]", "github_event_type": "pull_request.unlabeled", "github_delivery_guid": "d602a860-e3bb-11eb-9f5f-01f9e172f02b", "trigger_type": "GitHub Webhook", "payload_hash": "c4d522aac65d2bb87ffdecd9f7c49869" }, "level": 50, "stack": "Error: 14 UNAVAILABLE: read ECONNRESET\n at Object.callErrorFromStatus (/usr/src/app/node_modules/@grpc/grpc-js/build/src/call.js:31:26)\n at Object.onReceiveStatus (/usr/src/app/node_modules/@grpc/grpc-js/build/src/client.js:179:52)\n at Object.onReceiveStatus (/usr/src/app/node_modules/@grpc/grpc-js/build/src/client-interceptors.js:336:141)\n at Object.onReceiveStatus (/usr/src/app/node_modules/@grpc/grpc-js/build/src/client-interceptors.js:299:181)\n at /usr/src/app/node_modules/@grpc/grpc-js/build/src/call-stream.js:145:78\n at processTicksAndRejections (internal/process/task_queues.js:77:11)" }, "resource": { "type": "cloud_run_revision", "labels": { "project_id": "repo-automation-bots", "revision_name": "owl-bot-00007-zag", "location": "us-central1", "configuration_name": "owl-bot", "service_name": "owl-bot" } }, "timestamp": "2021-07-13T09:22:26.341226Z", "severity": "ERROR", "labels": { "instanceId": "00bf4bf02d480c07e7ac82c2a94eedcb21772fc9b421079001923328aadc9c17206d6cbdd09c42e99eaf03bbfd7759770d2a1010e3d0dd86a84eb0f6ed" }, "logName": "projects/repo-automation-bots/logs/run.googleapis.com%2Fstdout", "receiveTimestamp": "2021-07-13T09:22:26.409211893Z" } ```
non_process
owl bot frontend dies with error unavailable read econnreset starring as because it s very rare the error log insertid jsonpayload details read econnreset message unavailable read econnreset metadata options internalrepr code type error trigger trigger source repo owner type organization repo name google resumable media python url owner googleapis trigger sender gcf merge on green github event type pull request unlabeled github delivery guid trigger type github webhook payload hash level stack error unavailable read econnreset n at object callerrorfromstatus usr src app node modules grpc grpc js build src call js n at object onreceivestatus usr src app node modules grpc grpc js build src client js n at object onreceivestatus usr src app node modules grpc grpc js build src client interceptors js n at object onreceivestatus usr src app node modules grpc grpc js build src client interceptors js n at usr src app node modules grpc grpc js build src call stream js n at processticksandrejections internal process task queues js resource type cloud run revision labels project id repo automation bots revision name owl bot zag location us configuration name owl bot service name owl bot timestamp severity error labels instanceid logname projects repo automation bots logs run googleapis com receivetimestamp
0
20,325
26,964,908,109
IssuesEvent
2023-02-08 21:24:53
tokio-rs/tokio
https://api.github.com/repos/tokio-rs/tokio
closed
15 minute hang when reading from subprocess stderr
C-question A-tokio M-process I-hang
<!-- Thank you for reporting an issue. Please fill in as much of the template below as you're able. --> ## Version <!-- List the versions of all `tokio` crates you are using. The easiest way to get this information is using `cargo-tree`. `cargo install cargo-tree` (see install here: https://github.com/sfackler/cargo-tree) Then: `cargo tree | grep tokio` --> Tokio v0.2.13 and v0.2.14 <details><summary>click for full tree</summary> ``` ├── tokio v0.2.14 │ └── tokio-macros v0.2.5 │ │ ├── tokio v0.2.14 (*) │ │ └── tokio-util v0.3.1 │ │ └── tokio v0.2.14 (*) │ ├── tokio v0.2.14 (*) ├── tokio v0.2.14 (*) ├── tokio-rustls v0.13.0 │ ├── tokio v0.2.14 (*) ├── tokio-util v0.3.1 (*) │ │ ├── tokio v0.2.14 (*) │ │ ├── tokio v0.2.14 (*) │ │ │ ├── tokio v0.2.14 (*) │ │ ├── tokio v0.2.14 (*) │ │ ├── tokio v0.2.14 (*) │ ├── tokio v0.2.14 (*) │ │ ├── tokio v0.2.14 (*) │ │ ├── tokio v0.2.14 (*) ``` </details> ## Platform <!--- Output of `uname -a` (UNIX), or version and 32 or 64-bit (Windows) --> > Linux code13sim 3.10.0-957.el7.x86_64 #1 SMP Thu Oct 4 20:48:51 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux > Red Hat Enterprise Linux Server release 7.6 (Maipo) ## Description This scenario is a little hard to describe in detail. At a high level, I am trying to spawn a subprocess, read stdout and stderr, wait for the subprocess to end, and stream the data from stdout/stderr into a MPSC channel (which is attached to a tonic service handler. Important details about this below) The core code looks basically like this: ```rust let (mut tx, rx) = mpsc::channel(4); let mut cmd = tokio::process::Command::new("sh"); cmd.arg("run.sh"); cmd.stdout(Stdio::piped()).stderr(Stdio::piped()).stdin(Stdio::null()); let mut child: tokio::process::Child = cmd.spawn().unwrap(); let mut stdout = child.stdout.take().expect("No stdout handle"); let mut stderr = child.stderr.take().expect("No stderr handle"); // capture stderr let mut tx1 = tx.clone(); let stderr_task = tokio::spawn(async move { let mut buf = [0u8; 1024]; while let Ok(bytes) = stderr.read(&mut buf).await { println!("stderr Line: {} bytes @ {:?}", bytes, std::time::SystemTime::now().duration_since(UNIX_EPOCH)); if bytes == 0 { break; } let res = tx1.send(Ok(ExecuteReply::from_stderr(buf[..bytes].to_vec()))).await; println!("Done sending stderr line to tx1: {:?}", res) } println!("Stderr task is done"); }); // capture stdout let mut tx1 = tx.clone(); let stdout_task = tokio::spawn(async move { let mut buf = [0u8; 1024]; while let Ok(bytes) = stdout.read(&mut buf).await { println!("stdout Line: {} bytes @ {:?}", bytes, std::time::SystemTime::now().duration_since(UNIX_EPOCH)); if bytes == 0 { break; } let res = tx1.send(Ok(ExecuteReply::from_stdout(buf[..bytes].to_vec()))).await; println!("Done sending stdout line to tx1: {:?}", res); } println!("stdout task is done"); }); // wait for the child process to exit // let mut tx = tx.clone(); tokio::spawn(async move { // wait for child and stdout and stderr tasks all at once let (stdout_r, stderr_r, status) = tokio::join!(stdout_task, stderr_task, child); let status: std::process::ExitStatus = status.expect("child process encountered an error"); println!("child status was: {}", status); stdout_r.context("Waiting for stdout task").unwrap(); stderr_r.context("Waiting for stderr task").unwrap(); println!("stdout and stderr tasks are done"); tx.send(Ok(ExecuteReply::from_exit_status(status))) .await .unwrap(); }); ``` The problem that I'm seeing is that the call to `stderr.read(&mut buf).await` is blocking for about 15 minutes before returning `Ok(0)`. This delay can be seen in the timestamps of this output: ``` stdout Line: 57 bytes @ Ok(1585802719.11239218s) Done sending stdout line to tx1: Ok(()) stderr Line: 120 bytes @ Ok(1585802719.172319169s) Done sending stderr line to tx1: Ok(()) stdout Line: 163 bytes @ Ok(1585802719.193320336s) Done sending stdout line to tx1: Ok(()) stdout Line: 90 bytes @ Ok(1585802719.708493195s) Done sending stdout line to tx1: Ok(()) stdout Line: 0 bytes @ Ok(1585802719.708636111s) stdout task is done stderr Line: 0 bytes @ Ok(1585803649.608731897s) Stderr task is done child status was: exit code: 0 stdout and stderr tasks are done ``` Observe the 930 second delay between these two lines: > stdout Line: 0 bytes @ Ok(1585802719.708636111s) > stderr Line: 0 bytes @ Ok(1585803649.608731897s) Here's where it gets weird: * There is a particular `run.sh` script that causes this problem. Tweaking it relatively trivial ways makes the problem disappear. By "problem disappear" I mean there is no 15 minute delay when awaiting stderr. A complete copy of run.sh is included below. * I only ever see this problem when waiting for the stderr task (never the stdout task) * While stuck in the 15 minute nap, a process listing confirms that the subprocess has exited cleanly. It normally should take about 500ms to complete. * As mentioned above, all of this code is running in the context of a tonic gRPC server (so the code I've written above is running in a tokio task itself [and not directly in a `main()` function]). In the above code, I don't show what's consuming the reading end of the MPSC channel, but it's being passed to tonic as the read end of a streaming service. Trying to extract this code into a standalone executable doesn't reproduce this problem (though I'm still trying) * The problem is very "bistable". I can generally reproduce the problem reliably... until I can't. When the problem reproduces, it'll reproduce reliably 30 out of 30 times. But then it'll stop reproducing for a while. * I *think* the problem doesn't happen on SLES machines (but does happen on RHEL machines), but I'm not totally sure on that. Still chasing this down. Here's the full copy of the problematic run.sh. It looks a little weird since I started as a generated script from a gitlab runner that I've tried to minify <details><summary>click to show</summary> ```sh #!/usr/bin/env bash set -eo pipefail set +o noclobber mkdir -p "builds/0/lxd-gitlab-custom-executor/FtpgesaK/0/lxd/gitlab-custom-executor.tmp" echo $'\x1b[32;1mFetching changes with git depth set to 50...\x1b[0;m' $'mkdir' "-p" "builds/0/lxd-gitlab-custom-executor/FtpgesaK/0/lxd/gitlab-custom-executor.tmp/git-template" $'git' "config" "-f" "builds/0/lxd-gitlab-custom-executor/FtpgesaK/0/lxd/gitlab-custom-executor.tmp/git-template/config" "fetch.recurseSubmodules" "false" $'git' "init" "builds/0/lxd-gitlab-custom-executor/FtpgesaK/0/lxd/gitlab-custom-executor" "--template" "builds/0/lxd-gitlab-custom-executor/FtpgesaK/0/lxd/gitlab-custom-executor.tmp/git-template" 'cd' "builds/0/lxd-gitlab-custom-executor/FtpgesaK/0/lxd/gitlab-custom-executor" $'rm' "-f" ".git/index.lock" $'rm' "-f" ".git/shallow.lock" $'rm' "-f" ".git/HEAD.lock" $'rm' "-f" ".git/hooks/post-checkout" $'git' "fetch" "origin" "--prune" "+refs/pipelines/*:refs/pipelines/*" "+refs/heads/testing:refs/remotes/origin/testing" "--depth" "50" echo $'\x1b[32;1mChecking out 56fb23e1 as testing...\x1b[0;m' echo $'\x1b[32;1mSkipping Git submodules setup\x1b[0;m' # exit 0 ``` </details> ## Summary As mentioned above, I see this problem reliably in my tonic gRPC server, but not in a standalone program. I don't think tonic is at fault, but I do think that perhaps all of the other tasks in the executor are having an impact on this. Given all my problems trying to reproduce this, I'm fairly certain that you won't be able to. So what I'm asking for is hints on how to debug this: * Have I done anything obviously wrong in the above code? * Are there flags/envvars that I can enable to get a better idea from tokio about what's happening? (Stracing the process while it's taking its 15 minute nap don't show anything helpful just blocking on a futex call) * If I wanted to instrument tokio itself to help track this down, where would I start? Thank you!
1.0
15 minute hang when reading from subprocess stderr - <!-- Thank you for reporting an issue. Please fill in as much of the template below as you're able. --> ## Version <!-- List the versions of all `tokio` crates you are using. The easiest way to get this information is using `cargo-tree`. `cargo install cargo-tree` (see install here: https://github.com/sfackler/cargo-tree) Then: `cargo tree | grep tokio` --> Tokio v0.2.13 and v0.2.14 <details><summary>click for full tree</summary> ``` ├── tokio v0.2.14 │ └── tokio-macros v0.2.5 │ │ ├── tokio v0.2.14 (*) │ │ └── tokio-util v0.3.1 │ │ └── tokio v0.2.14 (*) │ ├── tokio v0.2.14 (*) ├── tokio v0.2.14 (*) ├── tokio-rustls v0.13.0 │ ├── tokio v0.2.14 (*) ├── tokio-util v0.3.1 (*) │ │ ├── tokio v0.2.14 (*) │ │ ├── tokio v0.2.14 (*) │ │ │ ├── tokio v0.2.14 (*) │ │ ├── tokio v0.2.14 (*) │ │ ├── tokio v0.2.14 (*) │ ├── tokio v0.2.14 (*) │ │ ├── tokio v0.2.14 (*) │ │ ├── tokio v0.2.14 (*) ``` </details> ## Platform <!--- Output of `uname -a` (UNIX), or version and 32 or 64-bit (Windows) --> > Linux code13sim 3.10.0-957.el7.x86_64 #1 SMP Thu Oct 4 20:48:51 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux > Red Hat Enterprise Linux Server release 7.6 (Maipo) ## Description This scenario is a little hard to describe in detail. At a high level, I am trying to spawn a subprocess, read stdout and stderr, wait for the subprocess to end, and stream the data from stdout/stderr into a MPSC channel (which is attached to a tonic service handler. Important details about this below) The core code looks basically like this: ```rust let (mut tx, rx) = mpsc::channel(4); let mut cmd = tokio::process::Command::new("sh"); cmd.arg("run.sh"); cmd.stdout(Stdio::piped()).stderr(Stdio::piped()).stdin(Stdio::null()); let mut child: tokio::process::Child = cmd.spawn().unwrap(); let mut stdout = child.stdout.take().expect("No stdout handle"); let mut stderr = child.stderr.take().expect("No stderr handle"); // capture stderr let mut tx1 = tx.clone(); let stderr_task = tokio::spawn(async move { let mut buf = [0u8; 1024]; while let Ok(bytes) = stderr.read(&mut buf).await { println!("stderr Line: {} bytes @ {:?}", bytes, std::time::SystemTime::now().duration_since(UNIX_EPOCH)); if bytes == 0 { break; } let res = tx1.send(Ok(ExecuteReply::from_stderr(buf[..bytes].to_vec()))).await; println!("Done sending stderr line to tx1: {:?}", res) } println!("Stderr task is done"); }); // capture stdout let mut tx1 = tx.clone(); let stdout_task = tokio::spawn(async move { let mut buf = [0u8; 1024]; while let Ok(bytes) = stdout.read(&mut buf).await { println!("stdout Line: {} bytes @ {:?}", bytes, std::time::SystemTime::now().duration_since(UNIX_EPOCH)); if bytes == 0 { break; } let res = tx1.send(Ok(ExecuteReply::from_stdout(buf[..bytes].to_vec()))).await; println!("Done sending stdout line to tx1: {:?}", res); } println!("stdout task is done"); }); // wait for the child process to exit // let mut tx = tx.clone(); tokio::spawn(async move { // wait for child and stdout and stderr tasks all at once let (stdout_r, stderr_r, status) = tokio::join!(stdout_task, stderr_task, child); let status: std::process::ExitStatus = status.expect("child process encountered an error"); println!("child status was: {}", status); stdout_r.context("Waiting for stdout task").unwrap(); stderr_r.context("Waiting for stderr task").unwrap(); println!("stdout and stderr tasks are done"); tx.send(Ok(ExecuteReply::from_exit_status(status))) .await .unwrap(); }); ``` The problem that I'm seeing is that the call to `stderr.read(&mut buf).await` is blocking for about 15 minutes before returning `Ok(0)`. This delay can be seen in the timestamps of this output: ``` stdout Line: 57 bytes @ Ok(1585802719.11239218s) Done sending stdout line to tx1: Ok(()) stderr Line: 120 bytes @ Ok(1585802719.172319169s) Done sending stderr line to tx1: Ok(()) stdout Line: 163 bytes @ Ok(1585802719.193320336s) Done sending stdout line to tx1: Ok(()) stdout Line: 90 bytes @ Ok(1585802719.708493195s) Done sending stdout line to tx1: Ok(()) stdout Line: 0 bytes @ Ok(1585802719.708636111s) stdout task is done stderr Line: 0 bytes @ Ok(1585803649.608731897s) Stderr task is done child status was: exit code: 0 stdout and stderr tasks are done ``` Observe the 930 second delay between these two lines: > stdout Line: 0 bytes @ Ok(1585802719.708636111s) > stderr Line: 0 bytes @ Ok(1585803649.608731897s) Here's where it gets weird: * There is a particular `run.sh` script that causes this problem. Tweaking it relatively trivial ways makes the problem disappear. By "problem disappear" I mean there is no 15 minute delay when awaiting stderr. A complete copy of run.sh is included below. * I only ever see this problem when waiting for the stderr task (never the stdout task) * While stuck in the 15 minute nap, a process listing confirms that the subprocess has exited cleanly. It normally should take about 500ms to complete. * As mentioned above, all of this code is running in the context of a tonic gRPC server (so the code I've written above is running in a tokio task itself [and not directly in a `main()` function]). In the above code, I don't show what's consuming the reading end of the MPSC channel, but it's being passed to tonic as the read end of a streaming service. Trying to extract this code into a standalone executable doesn't reproduce this problem (though I'm still trying) * The problem is very "bistable". I can generally reproduce the problem reliably... until I can't. When the problem reproduces, it'll reproduce reliably 30 out of 30 times. But then it'll stop reproducing for a while. * I *think* the problem doesn't happen on SLES machines (but does happen on RHEL machines), but I'm not totally sure on that. Still chasing this down. Here's the full copy of the problematic run.sh. It looks a little weird since I started as a generated script from a gitlab runner that I've tried to minify <details><summary>click to show</summary> ```sh #!/usr/bin/env bash set -eo pipefail set +o noclobber mkdir -p "builds/0/lxd-gitlab-custom-executor/FtpgesaK/0/lxd/gitlab-custom-executor.tmp" echo $'\x1b[32;1mFetching changes with git depth set to 50...\x1b[0;m' $'mkdir' "-p" "builds/0/lxd-gitlab-custom-executor/FtpgesaK/0/lxd/gitlab-custom-executor.tmp/git-template" $'git' "config" "-f" "builds/0/lxd-gitlab-custom-executor/FtpgesaK/0/lxd/gitlab-custom-executor.tmp/git-template/config" "fetch.recurseSubmodules" "false" $'git' "init" "builds/0/lxd-gitlab-custom-executor/FtpgesaK/0/lxd/gitlab-custom-executor" "--template" "builds/0/lxd-gitlab-custom-executor/FtpgesaK/0/lxd/gitlab-custom-executor.tmp/git-template" 'cd' "builds/0/lxd-gitlab-custom-executor/FtpgesaK/0/lxd/gitlab-custom-executor" $'rm' "-f" ".git/index.lock" $'rm' "-f" ".git/shallow.lock" $'rm' "-f" ".git/HEAD.lock" $'rm' "-f" ".git/hooks/post-checkout" $'git' "fetch" "origin" "--prune" "+refs/pipelines/*:refs/pipelines/*" "+refs/heads/testing:refs/remotes/origin/testing" "--depth" "50" echo $'\x1b[32;1mChecking out 56fb23e1 as testing...\x1b[0;m' echo $'\x1b[32;1mSkipping Git submodules setup\x1b[0;m' # exit 0 ``` </details> ## Summary As mentioned above, I see this problem reliably in my tonic gRPC server, but not in a standalone program. I don't think tonic is at fault, but I do think that perhaps all of the other tasks in the executor are having an impact on this. Given all my problems trying to reproduce this, I'm fairly certain that you won't be able to. So what I'm asking for is hints on how to debug this: * Have I done anything obviously wrong in the above code? * Are there flags/envvars that I can enable to get a better idea from tokio about what's happening? (Stracing the process while it's taking its 15 minute nap don't show anything helpful just blocking on a futex call) * If I wanted to instrument tokio itself to help track this down, where would I start? Thank you!
process
minute hang when reading from subprocess stderr thank you for reporting an issue please fill in as much of the template below as you re able version list the versions of all tokio crates you are using the easiest way to get this information is using cargo tree cargo install cargo tree see install here then cargo tree grep tokio tokio and click for full tree ├── tokio │ └── tokio macros │ │ ├── tokio │ │ └── tokio util │ │ └── tokio │ ├── tokio ├── tokio ├── tokio rustls │ ├── tokio ├── tokio util │ │ ├── tokio │ │ ├── tokio │ │ │ ├── tokio │ │ ├── tokio │ │ ├── tokio │ ├── tokio │ │ ├── tokio │ │ ├── tokio platform output of uname a unix or version and or bit windows linux smp thu oct utc gnu linux red hat enterprise linux server release maipo description this scenario is a little hard to describe in detail at a high level i am trying to spawn a subprocess read stdout and stderr wait for the subprocess to end and stream the data from stdout stderr into a mpsc channel which is attached to a tonic service handler important details about this below the core code looks basically like this rust let mut tx rx mpsc channel let mut cmd tokio process command new sh cmd arg run sh cmd stdout stdio piped stderr stdio piped stdin stdio null let mut child tokio process child cmd spawn unwrap let mut stdout child stdout take expect no stdout handle let mut stderr child stderr take expect no stderr handle capture stderr let mut tx clone let stderr task tokio spawn async move let mut buf while let ok bytes stderr read mut buf await println stderr line bytes bytes std time systemtime now duration since unix epoch if bytes break let res send ok executereply from stderr buf to vec await println done sending stderr line to res println stderr task is done capture stdout let mut tx clone let stdout task tokio spawn async move let mut buf while let ok bytes stdout read mut buf await println stdout line bytes bytes std time systemtime now duration since unix epoch if bytes break let res send ok executereply from stdout buf to vec await println done sending stdout line to res println stdout task is done wait for the child process to exit let mut tx tx clone tokio spawn async move wait for child and stdout and stderr tasks all at once let stdout r stderr r status tokio join stdout task stderr task child let status std process exitstatus status expect child process encountered an error println child status was status stdout r context waiting for stdout task unwrap stderr r context waiting for stderr task unwrap println stdout and stderr tasks are done tx send ok executereply from exit status status await unwrap the problem that i m seeing is that the call to stderr read mut buf await is blocking for about minutes before returning ok this delay can be seen in the timestamps of this output stdout line bytes ok done sending stdout line to ok stderr line bytes ok done sending stderr line to ok stdout line bytes ok done sending stdout line to ok stdout line bytes ok done sending stdout line to ok stdout line bytes ok stdout task is done stderr line bytes ok stderr task is done child status was exit code stdout and stderr tasks are done observe the second delay between these two lines stdout line bytes ok stderr line bytes ok here s where it gets weird there is a particular run sh script that causes this problem tweaking it relatively trivial ways makes the problem disappear by problem disappear i mean there is no minute delay when awaiting stderr a complete copy of run sh is included below i only ever see this problem when waiting for the stderr task never the stdout task while stuck in the minute nap a process listing confirms that the subprocess has exited cleanly it normally should take about to complete as mentioned above all of this code is running in the context of a tonic grpc server so the code i ve written above is running in a tokio task itself in the above code i don t show what s consuming the reading end of the mpsc channel but it s being passed to tonic as the read end of a streaming service trying to extract this code into a standalone executable doesn t reproduce this problem though i m still trying the problem is very bistable i can generally reproduce the problem reliably until i can t when the problem reproduces it ll reproduce reliably out of times but then it ll stop reproducing for a while i think the problem doesn t happen on sles machines but does happen on rhel machines but i m not totally sure on that still chasing this down here s the full copy of the problematic run sh it looks a little weird since i started as a generated script from a gitlab runner that i ve tried to minify click to show sh usr bin env bash set eo pipefail set o noclobber mkdir p builds lxd gitlab custom executor ftpgesak lxd gitlab custom executor tmp echo changes with git depth set to m mkdir p builds lxd gitlab custom executor ftpgesak lxd gitlab custom executor tmp git template git config f builds lxd gitlab custom executor ftpgesak lxd gitlab custom executor tmp git template config fetch recursesubmodules false git init builds lxd gitlab custom executor ftpgesak lxd gitlab custom executor template builds lxd gitlab custom executor ftpgesak lxd gitlab custom executor tmp git template cd builds lxd gitlab custom executor ftpgesak lxd gitlab custom executor rm f git index lock rm f git shallow lock rm f git head lock rm f git hooks post checkout git fetch origin prune refs pipelines refs pipelines refs heads testing refs remotes origin testing depth echo out as testing m echo git submodules setup m exit summary as mentioned above i see this problem reliably in my tonic grpc server but not in a standalone program i don t think tonic is at fault but i do think that perhaps all of the other tasks in the executor are having an impact on this given all my problems trying to reproduce this i m fairly certain that you won t be able to so what i m asking for is hints on how to debug this have i done anything obviously wrong in the above code are there flags envvars that i can enable to get a better idea from tokio about what s happening stracing the process while it s taking its minute nap don t show anything helpful just blocking on a futex call if i wanted to instrument tokio itself to help track this down where would i start thank you
1
6,625
9,725,439,403
IssuesEvent
2019-05-30 08:42:50
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Error with Proccesing tools and GeoPackage (QGIS v2.99)
Bug Feedback Processing
Author Name: **Samuel Fernando Mesa Giraldo** (Samuel Fernando Mesa Giraldo) Original Redmine Issue: [17614](https://issues.qgis.org/issues/17614) Affected QGIS version: master Redmine category:processing/core Assignee: Victor Olaya --- I'm using the default GeoPackage format for databases, but when I try to use a processing tool of GDAL, GRASS GIS or SAGA GIS, it does not recognize the format or it just closes unexpectedly. I have used _Buffer Vector_ from GDAL, _v.buffer.distance_ from GRASS GIS and F_ixed distance buffer_ from SAGA GIS and none with results. In the case of GDAL, it always converts the layer to SHAPEFILE and consequently does not recognize the shortened fields. In the case of GRASS GIS, it always appears that the GPKG format is not recognized. In the case of SAGA GIS, returns the following error, closing the QGIS program: ``` (Make sure 'QTextCursor' is registered using qRegisterMetaType().) src/core/qgsmessagelog.cpp: 27: (logMessage) [1ms] [thread:0x7f66d8008f30] 2017-12-03T21:31:03 Processing[0] SAGA execution commands shapes_tools "Shapes Buffer" -SHAPES "/tmp/processing_ce203365cf1d4a0dba6dc6c42a7923fb/88f882df27f04f47ba0de04def37c1b7/SHAPES.shp" -DIST_FIELD_DEFAULT 100.0 -NZONES 1 -DARC 5.0 -DISSOLVE true -POLY_INNER false -BUFFER "/home/samtux/tmp/salida.gpkg" *** Error in `./qgis': double free or corruption (!prev): 0x0000564e272961e0 *** Abortado (`core' generado) ``` --- - [test_points.gpkg](https://issues.qgis.org/attachments/download/11802/test_points.gpkg) (Samuel Fernando Mesa Giraldo)
1.0
Error with Proccesing tools and GeoPackage (QGIS v2.99) - Author Name: **Samuel Fernando Mesa Giraldo** (Samuel Fernando Mesa Giraldo) Original Redmine Issue: [17614](https://issues.qgis.org/issues/17614) Affected QGIS version: master Redmine category:processing/core Assignee: Victor Olaya --- I'm using the default GeoPackage format for databases, but when I try to use a processing tool of GDAL, GRASS GIS or SAGA GIS, it does not recognize the format or it just closes unexpectedly. I have used _Buffer Vector_ from GDAL, _v.buffer.distance_ from GRASS GIS and F_ixed distance buffer_ from SAGA GIS and none with results. In the case of GDAL, it always converts the layer to SHAPEFILE and consequently does not recognize the shortened fields. In the case of GRASS GIS, it always appears that the GPKG format is not recognized. In the case of SAGA GIS, returns the following error, closing the QGIS program: ``` (Make sure 'QTextCursor' is registered using qRegisterMetaType().) src/core/qgsmessagelog.cpp: 27: (logMessage) [1ms] [thread:0x7f66d8008f30] 2017-12-03T21:31:03 Processing[0] SAGA execution commands shapes_tools "Shapes Buffer" -SHAPES "/tmp/processing_ce203365cf1d4a0dba6dc6c42a7923fb/88f882df27f04f47ba0de04def37c1b7/SHAPES.shp" -DIST_FIELD_DEFAULT 100.0 -NZONES 1 -DARC 5.0 -DISSOLVE true -POLY_INNER false -BUFFER "/home/samtux/tmp/salida.gpkg" *** Error in `./qgis': double free or corruption (!prev): 0x0000564e272961e0 *** Abortado (`core' generado) ``` --- - [test_points.gpkg](https://issues.qgis.org/attachments/download/11802/test_points.gpkg) (Samuel Fernando Mesa Giraldo)
process
error with proccesing tools and geopackage qgis author name samuel fernando mesa giraldo samuel fernando mesa giraldo original redmine issue affected qgis version master redmine category processing core assignee victor olaya i m using the default geopackage format for databases but when i try to use a processing tool of gdal grass gis or saga gis it does not recognize the format or it just closes unexpectedly i have used buffer vector from gdal v buffer distance from grass gis and f ixed distance buffer from saga gis and none with results in the case of gdal it always converts the layer to shapefile and consequently does not recognize the shortened fields in the case of grass gis it always appears that the gpkg format is not recognized in the case of saga gis returns the following error closing the qgis program make sure qtextcursor is registered using qregistermetatype src core qgsmessagelog cpp logmessage processing saga execution commands shapes tools shapes buffer shapes tmp processing shapes shp dist field default nzones darc dissolve true poly inner false buffer home samtux tmp salida gpkg error in qgis double free or corruption prev abortado core generado samuel fernando mesa giraldo
1
17,844
23,783,952,004
IssuesEvent
2022-09-02 08:19:58
emacs-ess/ESS
https://api.github.com/repos/emacs-ess/ESS
closed
How can I salvage a bad start? -- i.e, when no "ESSR" in `search()`
process lang:r
This currently happens quite regularly to me, notably when R is started *after* I work with an `*.R` file in emacs (and ESS is loaded of course). It also happens when I run R in the (C-level) debugger, i.e., after `C-u M-x R` (because in the `*R*` buffer, there's the `gdb> ` prompt, and I first have to type `r` (for `run`) there to start R and all this takes little while. Anyway back to the question: Notably when R runs from `R -d gdb` it's a pain when "ESSR" is missing because practically everything interactive fails (such as TAB completion for R function names). What M-x .... (or other way) can I use to "restart" that second part of `M-x R` ? {{yes, I think I could find out myself; but it's something we should be aware of that users eventually want to be able to do}}
1.0
How can I salvage a bad start? -- i.e, when no "ESSR" in `search()` - This currently happens quite regularly to me, notably when R is started *after* I work with an `*.R` file in emacs (and ESS is loaded of course). It also happens when I run R in the (C-level) debugger, i.e., after `C-u M-x R` (because in the `*R*` buffer, there's the `gdb> ` prompt, and I first have to type `r` (for `run`) there to start R and all this takes little while. Anyway back to the question: Notably when R runs from `R -d gdb` it's a pain when "ESSR" is missing because practically everything interactive fails (such as TAB completion for R function names). What M-x .... (or other way) can I use to "restart" that second part of `M-x R` ? {{yes, I think I could find out myself; but it's something we should be aware of that users eventually want to be able to do}}
process
how can i salvage a bad start i e when no essr in search this currently happens quite regularly to me notably when r is started after i work with an r file in emacs and ess is loaded of course it also happens when i run r in the c level debugger i e after c u m x r because in the r buffer there s the gdb prompt and i first have to type r for run there to start r and all this takes little while anyway back to the question notably when r runs from r d gdb it s a pain when essr is missing because practically everything interactive fails such as tab completion for r function names what m x or other way can i use to restart that second part of m x r yes i think i could find out myself but it s something we should be aware of that users eventually want to be able to do
1
631
3,092,112,250
IssuesEvent
2015-08-26 16:12:24
e-government-ua/iBP
https://api.github.com/repos/e-government-ua/iBP
opened
Кузнецовск - Надання дозволу на проведення публічних богослужінь
in process of creating
Алгоритм: https://drive.google.com/file/d/0B1NO5RoNKst7Sy12SERLa2puRFNrcjlBYjVrNDNSSDVCaTlj/view?usp=sharing Заява: https://drive.google.com/file/d/0B1NO5RoNKst7aDRFMUhMcFAxSWZEeVFiUUtCNnVmV0dnWGU4/view?usp=sharing Інформаційна карта послуги: https://drive.google.com/file/d/0B1NO5RoNKst7YVBmZXMtRGE2a2psYlpwN0xENmU0MUlhZ21F/view?usp=sharing
1.0
Кузнецовск - Надання дозволу на проведення публічних богослужінь - Алгоритм: https://drive.google.com/file/d/0B1NO5RoNKst7Sy12SERLa2puRFNrcjlBYjVrNDNSSDVCaTlj/view?usp=sharing Заява: https://drive.google.com/file/d/0B1NO5RoNKst7aDRFMUhMcFAxSWZEeVFiUUtCNnVmV0dnWGU4/view?usp=sharing Інформаційна карта послуги: https://drive.google.com/file/d/0B1NO5RoNKst7YVBmZXMtRGE2a2psYlpwN0xENmU0MUlhZ21F/view?usp=sharing
process
кузнецовск надання дозволу на проведення публічних богослужінь алгоритм заява інформаційна карта послуги
1
62,653
12,229,117,608
IssuesEvent
2020-05-03 22:31:48
Regalis11/Barotrauma
https://api.github.com/repos/Regalis11/Barotrauma
closed
Red Cells do not die after defeating Brain
Code Design
- [ ] I have searched the issue tracker to check if the issue has already been reported. **Description** Not sure if it's intended or not. These cells tend to get stuck in places in the sub that players have a hard time getting to as well. Killing the brain doesn't stop them from tearing the sub apart though. **Version** ~0.9.9.1
1.0
Red Cells do not die after defeating Brain - - [ ] I have searched the issue tracker to check if the issue has already been reported. **Description** Not sure if it's intended or not. These cells tend to get stuck in places in the sub that players have a hard time getting to as well. Killing the brain doesn't stop them from tearing the sub apart though. **Version** ~0.9.9.1
non_process
red cells do not die after defeating brain i have searched the issue tracker to check if the issue has already been reported description not sure if it s intended or not these cells tend to get stuck in places in the sub that players have a hard time getting to as well killing the brain doesn t stop them from tearing the sub apart though version
0
4,400
7,296,251,739
IssuesEvent
2018-02-26 10:08:31
DevExpress/testcafe-hammerhead
https://api.github.com/repos/DevExpress/testcafe-hammerhead
closed
Wrong processing of 'for...in' expression
AREA: server SYSTEM: resource processing TYPE: bug health-monitor
Original script: ```js for (utag._i in utag.loader.GV(utag_cfg_ovrd)) utag.cfg[utag._i] = utag_cfg_ovrd[utag._i] ``` Processed script: ```js for ( var __set$temp in utag.loader.GV(utag_cfg_ovrd)) __set$(utag.cfg,utag._i,__get$(utag_cfg_ovrd,utag._i))tag.cfg[utag._i] = utag_cfg_ovrd[utag._i] utag._i=__set$temp;tag.cfg[utag._i] = utag_cfg_ovrd[utag._i] ``` Site: https://www.wellsfargo.com/ Error text: `Uncaught SyntaxError: Unexpected identifier`
1.0
Wrong processing of 'for...in' expression - Original script: ```js for (utag._i in utag.loader.GV(utag_cfg_ovrd)) utag.cfg[utag._i] = utag_cfg_ovrd[utag._i] ``` Processed script: ```js for ( var __set$temp in utag.loader.GV(utag_cfg_ovrd)) __set$(utag.cfg,utag._i,__get$(utag_cfg_ovrd,utag._i))tag.cfg[utag._i] = utag_cfg_ovrd[utag._i] utag._i=__set$temp;tag.cfg[utag._i] = utag_cfg_ovrd[utag._i] ``` Site: https://www.wellsfargo.com/ Error text: `Uncaught SyntaxError: Unexpected identifier`
process
wrong processing of for in expression original script js for utag i in utag loader gv utag cfg ovrd utag cfg utag cfg ovrd processed script js for var set temp in utag loader gv utag cfg ovrd set utag cfg utag i get utag cfg ovrd utag i tag cfg utag cfg ovrd utag i set temp tag cfg utag cfg ovrd site error text uncaught syntaxerror unexpected identifier
1
3,290
6,384,773,029
IssuesEvent
2017-08-03 06:31:23
rubberduck-vba/Rubberduck
https://api.github.com/repos/rubberduck-vba/Rubberduck
closed
Resolve: implicit variable creation
parse-tree-processing
Does the resolver automatically create declarations for "implicitly created variables" if Option Explicit is not specified? Should we do that?
1.0
Resolve: implicit variable creation - Does the resolver automatically create declarations for "implicitly created variables" if Option Explicit is not specified? Should we do that?
process
resolve implicit variable creation does the resolver automatically create declarations for implicitly created variables if option explicit is not specified should we do that
1
190,203
15,222,723,272
IssuesEvent
2021-02-18 00:57:33
tl-its-umich-edu/my-learning-analytics
https://api.github.com/repos/tl-its-umich-edu/my-learning-analytics
closed
update LTI config docs with generic Canvas URLs
📖 documentation
#1223 changed several configuration URLs from domain-specific ones to generic Canvas ones. Check whether the LTI configuration documentation in the wiki needs to be updated and act accordingly.
1.0
update LTI config docs with generic Canvas URLs - #1223 changed several configuration URLs from domain-specific ones to generic Canvas ones. Check whether the LTI configuration documentation in the wiki needs to be updated and act accordingly.
non_process
update lti config docs with generic canvas urls changed several configuration urls from domain specific ones to generic canvas ones check whether the lti configuration documentation in the wiki needs to be updated and act accordingly
0
474,899
13,684,071,450
IssuesEvent
2020-09-30 03:47:35
wright-group/PyCMDS
https://api.github.com/repos/wright-group/PyCMDS
closed
Poynting tune load zero curve
Priority: Low discussion
Gives a good starting location after aligning from scratch Not as useful for iterative alignment, though we may decide to make it part of the general procedure
1.0
Poynting tune load zero curve - Gives a good starting location after aligning from scratch Not as useful for iterative alignment, though we may decide to make it part of the general procedure
non_process
poynting tune load zero curve gives a good starting location after aligning from scratch not as useful for iterative alignment though we may decide to make it part of the general procedure
0
953
3,418,199,248
IssuesEvent
2015-12-08 00:30:24
martensonbj/traffic-spy-skeleton
https://api.github.com/repos/martensonbj/traffic-spy-skeleton
opened
processing_requests_sad_path_403
processing requests user story
As a user When I send a POST request to 'http://yourapplication:port/sources/IDENTIFIER/data' And I have no yet registered Then I get a response of 'Application Not Registered - 403 Forbidden'
1.0
processing_requests_sad_path_403 - As a user When I send a POST request to 'http://yourapplication:port/sources/IDENTIFIER/data' And I have no yet registered Then I get a response of 'Application Not Registered - 403 Forbidden'
process
processing requests sad path as a user when i send a post request to and i have no yet registered then i get a response of application not registered forbidden
1
238,084
7,770,129,651
IssuesEvent
2018-06-04 07:46:24
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
discordapp.com - see bug description
browser-firefox priority-important status-needsinfo
<!-- @browser: Firefox 61.0 --> <!-- @ua_header: Mozilla/5.0 (Macintosh; Intel Mac OS X 10.13; rv:61.0) Gecko/20100101 Firefox/61.0 --> <!-- @reported_with: desktop-reporter --> **URL**: https://discordapp.com/partners **Browser / Version**: Firefox 61.0 **Operating System**: Mac OS X 10.13 **Tested Another Browser**: Unknown **Problem type**: Something else **Description**: CSS animation stutters **Steps to Reproduce**: No steps required. [![Screenshot Description](https://webcompat.com/uploads/2018/5/d146b494-403b-4028-a09a-2fb094e90683-thumb.jpg)](https://webcompat.com/uploads/2018/5/d146b494-403b-4028-a09a-2fb094e90683.jpg) <details> <summary>Browser Configuration</summary> <ul> <li>buildID: 20180510160705</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.all: false</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>channel: aurora</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
discordapp.com - see bug description - <!-- @browser: Firefox 61.0 --> <!-- @ua_header: Mozilla/5.0 (Macintosh; Intel Mac OS X 10.13; rv:61.0) Gecko/20100101 Firefox/61.0 --> <!-- @reported_with: desktop-reporter --> **URL**: https://discordapp.com/partners **Browser / Version**: Firefox 61.0 **Operating System**: Mac OS X 10.13 **Tested Another Browser**: Unknown **Problem type**: Something else **Description**: CSS animation stutters **Steps to Reproduce**: No steps required. [![Screenshot Description](https://webcompat.com/uploads/2018/5/d146b494-403b-4028-a09a-2fb094e90683-thumb.jpg)](https://webcompat.com/uploads/2018/5/d146b494-403b-4028-a09a-2fb094e90683.jpg) <details> <summary>Browser Configuration</summary> <ul> <li>buildID: 20180510160705</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.all: false</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>channel: aurora</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
discordapp com see bug description url browser version firefox operating system mac os x tested another browser unknown problem type something else description css animation stutters steps to reproduce no steps required browser configuration buildid gfx webrender blob images true gfx webrender all false gfx webrender enabled false image mem shared true channel aurora from with ❤️
0
787,051
27,703,310,799
IssuesEvent
2023-03-14 09:33:54
ballerina-platform/ballerina-lang
https://api.github.com/repos/ballerina-platform/ballerina-lang
closed
Class cast exception for nested query action
Type/Bug Priority/High Team/CompilerFE Crash Lang/Actions/Query
**Description:** ```ballerina public type Point record {|anydata...;|}; public type Feature record {|Point location;|}; Feature[] FEATURES = []; client class Service { remote function RecordRoute(stream<Point, error?> pointStream) returns error? { check from var point in pointStream do { check from Feature feature in FEATURES do { Point streamPoint = point; }; }; } } ``` stacktrace ``` Running executable [2022-08-08 14:15:39,579] SEVERE {b7a.log.crash} - class io.ballerina.runtime.internal.values.StreamValue cannot be cast to class io.ballerina.runtime.internal.values.MapValue (io.ballerina.runtime.internal.values.StreamValue and io.ballerina.runtime.internal.values.MapValue are in unnamed module of loader 'app') java.lang.ClassCastException: class io.ballerina.runtime.internal.values.StreamValue cannot be cast to class io.ballerina.runtime.internal.values.MapValue (io.ballerina.runtime.internal.values.StreamValue and io.ballerina.runtime.internal.values.MapValue are in unnamed module of loader 'app') at $value$Service.call(test.bal) at test.main(test.bal:58) at $_init.$lambda$main$(.) at io.ballerina.runtime.internal.scheduling.SchedulerItem.execute(Scheduler.java:610) at io.ballerina.runtime.internal.scheduling.Scheduler.run(Scheduler.java:316) at io.ballerina.runtime.internal.scheduling.Scheduler.runSafely(Scheduler.java:283) at java.base/java.lang.Thread.run(Thread.java:834) ``` **Steps to reproduce:** **Affected Versions:** 2201.1.1 **OS, DB, other environment details and versions:** **Related Issues (optional):** <!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. --> **Suggested Labels (optional):** <!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels--> **Suggested Assignees (optional):** <!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
1.0
Class cast exception for nested query action - **Description:** ```ballerina public type Point record {|anydata...;|}; public type Feature record {|Point location;|}; Feature[] FEATURES = []; client class Service { remote function RecordRoute(stream<Point, error?> pointStream) returns error? { check from var point in pointStream do { check from Feature feature in FEATURES do { Point streamPoint = point; }; }; } } ``` stacktrace ``` Running executable [2022-08-08 14:15:39,579] SEVERE {b7a.log.crash} - class io.ballerina.runtime.internal.values.StreamValue cannot be cast to class io.ballerina.runtime.internal.values.MapValue (io.ballerina.runtime.internal.values.StreamValue and io.ballerina.runtime.internal.values.MapValue are in unnamed module of loader 'app') java.lang.ClassCastException: class io.ballerina.runtime.internal.values.StreamValue cannot be cast to class io.ballerina.runtime.internal.values.MapValue (io.ballerina.runtime.internal.values.StreamValue and io.ballerina.runtime.internal.values.MapValue are in unnamed module of loader 'app') at $value$Service.call(test.bal) at test.main(test.bal:58) at $_init.$lambda$main$(.) at io.ballerina.runtime.internal.scheduling.SchedulerItem.execute(Scheduler.java:610) at io.ballerina.runtime.internal.scheduling.Scheduler.run(Scheduler.java:316) at io.ballerina.runtime.internal.scheduling.Scheduler.runSafely(Scheduler.java:283) at java.base/java.lang.Thread.run(Thread.java:834) ``` **Steps to reproduce:** **Affected Versions:** 2201.1.1 **OS, DB, other environment details and versions:** **Related Issues (optional):** <!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. --> **Suggested Labels (optional):** <!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels--> **Suggested Assignees (optional):** <!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
non_process
class cast exception for nested query action description ballerina public type point record anydata public type feature record point location feature features client class service remote function recordroute stream pointstream returns error check from var point in pointstream do check from feature feature in features do point streampoint point stacktrace running executable severe log crash class io ballerina runtime internal values streamvalue cannot be cast to class io ballerina runtime internal values mapvalue io ballerina runtime internal values streamvalue and io ballerina runtime internal values mapvalue are in unnamed module of loader app java lang classcastexception class io ballerina runtime internal values streamvalue cannot be cast to class io ballerina runtime internal values mapvalue io ballerina runtime internal values streamvalue and io ballerina runtime internal values mapvalue are in unnamed module of loader app at value service call test bal at test main test bal at init lambda main at io ballerina runtime internal scheduling scheduleritem execute scheduler java at io ballerina runtime internal scheduling scheduler run scheduler java at io ballerina runtime internal scheduling scheduler runsafely scheduler java at java base java lang thread run thread java steps to reproduce affected versions os db other environment details and versions related issues optional suggested labels optional suggested assignees optional
0
22,723
32,043,217,248
IssuesEvent
2023-09-22 21:27:14
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Split Vector Layer fails if attributes contain a slash
Processing Bug
### What is the bug or the crash? I'm running Vector > Data Management Tools > Split Vector Layer... on a layer showing city council districts for the state of New Mexico. I'm trying to split by city, but one district spans two cities and its "city" field is "Rio Rancho/Corrales". The relevant error in the dialog that pops up is: ``` Creating layer: /home/akkana/Data/votingdistricts/2023/citycouncil/City_Rio Rancho/Corrales.shp Could not create layer /home/akkana/Data/votingdistricts/2023/citycouncil/City_Rio Rancho/Corrales.shp: Creation of layer failed (OGR error: Failed to create file /home/akkana/Data/votingdistricts/2023/citycouncil/City_Rio Rancho/Corrales.shp: No such file or directory) ``` Of course, / is an illegal character in a filename (at least on Linux/Unix). It would be great if qgis detected this and omitted the slash, or replaced it with some legal character. I tried fixing this by right-clicking on the layer, opening the attribute table and editing the city to be "Rio Rancho Corrales" for each of the districts in that city. After editing, if I click on them with Identify, it shows no slash, but Split Vector Layers is still using the slash so it still fails. I had to save, quit and restart, after which things worked. (Except that it found a bunch of ring self-intersections that I had to fix, but at least no more slash in filename problems.) ### Steps to reproduce the issue 1. Load a dataset where some of the layer attributes have slashes. (I'll attach a zip of the shapefiles.) 2. Vector > Data Management Tools > Split Vector Layer... 3. Split on the "city" field, choose an output directory, and try to run the splitter. ### Versions QGIS version 3.28.8-Firenze QGIS code branch Release 3.28 Qt version 5.15.10 Python version 3.11.4 Compiled against GDAL/OGR 3.7.0 Running against GDAL/OGR 3.7.1 PROJ version 9.2.1 EPSG Registry database version v10.088 (2023-05-13) GEOS version 3.12.0-CAPI-1.18.0 SQLite version 3.42.0 PostgreSQL client version unknown SpatiaLite version 5.0.1 QWT version 6.1.4 QScintilla2 version 2.13.3 OS version Debian GNU/Linux trixie/sid Active Python plugins SRTM-Downloader 3.1.17 quick_map_services 0.19.33 processing 2.12.99 db_manager 0.1.20 grassprovider 2.12.99 sagaprovider 2.12.99 MetaSearch 0.3.6 ### Supported QGIS version - [X] I'm running a supported QGIS version according to [the roadmap](https://www.qgis.org/en/site/getinvolved/development/roadmap.html#release-schedule). ### New profile - [ ] I tried with a new [QGIS profile](https://docs.qgis.org/latest/en/docs/user_manual/introduction/qgis_configuration.html#working-with-user-profiles) ### Additional context _No response_
1.0
Split Vector Layer fails if attributes contain a slash - ### What is the bug or the crash? I'm running Vector > Data Management Tools > Split Vector Layer... on a layer showing city council districts for the state of New Mexico. I'm trying to split by city, but one district spans two cities and its "city" field is "Rio Rancho/Corrales". The relevant error in the dialog that pops up is: ``` Creating layer: /home/akkana/Data/votingdistricts/2023/citycouncil/City_Rio Rancho/Corrales.shp Could not create layer /home/akkana/Data/votingdistricts/2023/citycouncil/City_Rio Rancho/Corrales.shp: Creation of layer failed (OGR error: Failed to create file /home/akkana/Data/votingdistricts/2023/citycouncil/City_Rio Rancho/Corrales.shp: No such file or directory) ``` Of course, / is an illegal character in a filename (at least on Linux/Unix). It would be great if qgis detected this and omitted the slash, or replaced it with some legal character. I tried fixing this by right-clicking on the layer, opening the attribute table and editing the city to be "Rio Rancho Corrales" for each of the districts in that city. After editing, if I click on them with Identify, it shows no slash, but Split Vector Layers is still using the slash so it still fails. I had to save, quit and restart, after which things worked. (Except that it found a bunch of ring self-intersections that I had to fix, but at least no more slash in filename problems.) ### Steps to reproduce the issue 1. Load a dataset where some of the layer attributes have slashes. (I'll attach a zip of the shapefiles.) 2. Vector > Data Management Tools > Split Vector Layer... 3. Split on the "city" field, choose an output directory, and try to run the splitter. ### Versions QGIS version 3.28.8-Firenze QGIS code branch Release 3.28 Qt version 5.15.10 Python version 3.11.4 Compiled against GDAL/OGR 3.7.0 Running against GDAL/OGR 3.7.1 PROJ version 9.2.1 EPSG Registry database version v10.088 (2023-05-13) GEOS version 3.12.0-CAPI-1.18.0 SQLite version 3.42.0 PostgreSQL client version unknown SpatiaLite version 5.0.1 QWT version 6.1.4 QScintilla2 version 2.13.3 OS version Debian GNU/Linux trixie/sid Active Python plugins SRTM-Downloader 3.1.17 quick_map_services 0.19.33 processing 2.12.99 db_manager 0.1.20 grassprovider 2.12.99 sagaprovider 2.12.99 MetaSearch 0.3.6 ### Supported QGIS version - [X] I'm running a supported QGIS version according to [the roadmap](https://www.qgis.org/en/site/getinvolved/development/roadmap.html#release-schedule). ### New profile - [ ] I tried with a new [QGIS profile](https://docs.qgis.org/latest/en/docs/user_manual/introduction/qgis_configuration.html#working-with-user-profiles) ### Additional context _No response_
process
split vector layer fails if attributes contain a slash what is the bug or the crash i m running vector data management tools split vector layer on a layer showing city council districts for the state of new mexico i m trying to split by city but one district spans two cities and its city field is rio rancho corrales the relevant error in the dialog that pops up is creating layer home akkana data votingdistricts citycouncil city rio rancho corrales shp could not create layer home akkana data votingdistricts citycouncil city rio rancho corrales shp creation of layer failed ogr error failed to create file home akkana data votingdistricts citycouncil city rio rancho corrales shp no such file or directory of course is an illegal character in a filename at least on linux unix it would be great if qgis detected this and omitted the slash or replaced it with some legal character i tried fixing this by right clicking on the layer opening the attribute table and editing the city to be rio rancho corrales for each of the districts in that city after editing if i click on them with identify it shows no slash but split vector layers is still using the slash so it still fails i had to save quit and restart after which things worked except that it found a bunch of ring self intersections that i had to fix but at least no more slash in filename problems steps to reproduce the issue load a dataset where some of the layer attributes have slashes i ll attach a zip of the shapefiles vector data management tools split vector layer split on the city field choose an output directory and try to run the splitter versions qgis version firenze qgis code branch release qt version python version compiled against gdal ogr running against gdal ogr proj version epsg registry database version geos version capi sqlite version postgresql client version unknown spatialite version qwt version version os version debian gnu linux trixie sid active python plugins srtm downloader quick map services processing db manager grassprovider sagaprovider metasearch supported qgis version i m running a supported qgis version according to new profile i tried with a new additional context no response
1
7,804
10,960,331,134
IssuesEvent
2019-11-27 13:22:48
deeplearningunb/pneumonia-diagnose
https://api.github.com/repos/deeplearningunb/pneumonia-diagnose
closed
Criar função para ler e processar imagens
preprocessing tensorflow
- [ ] Ler imagens 16x16 - [ ] Adicionar etapas de data augmentation
1.0
Criar função para ler e processar imagens - - [ ] Ler imagens 16x16 - [ ] Adicionar etapas de data augmentation
process
criar função para ler e processar imagens ler imagens adicionar etapas de data augmentation
1
558,455
16,533,950,888
IssuesEvent
2021-05-27 09:33:17
threefoldtech/digitaltwin
https://api.github.com/repos/threefoldtech/digitaltwin
closed
Integrate file manager - W12
priority_major type_feature
File manager should have full integration in the digital twin. - [x] Create folders - [x] Upload files - [x] Delete files - [x] Browse folders / folder up - [x] Open files in OnlyOffice - [ ] Error handling (support TXT; DOCX ; DOC ; XLSX ; ... for editing, others only view in list / download) - [x] Copy / Pasting files - [x] Renaming - [x] Readable file sizes - [x] Readable last accessed - [x] Sorting Always write to real file system
1.0
Integrate file manager - W12 - File manager should have full integration in the digital twin. - [x] Create folders - [x] Upload files - [x] Delete files - [x] Browse folders / folder up - [x] Open files in OnlyOffice - [ ] Error handling (support TXT; DOCX ; DOC ; XLSX ; ... for editing, others only view in list / download) - [x] Copy / Pasting files - [x] Renaming - [x] Readable file sizes - [x] Readable last accessed - [x] Sorting Always write to real file system
non_process
integrate file manager file manager should have full integration in the digital twin create folders upload files delete files browse folders folder up open files in onlyoffice error handling support txt docx doc xlsx for editing others only view in list download copy pasting files renaming readable file sizes readable last accessed sorting always write to real file system
0
13,449
15,894,986,642
IssuesEvent
2021-04-11 12:22:39
nodejs/node
https://api.github.com/repos/nodejs/node
reopened
Throw within queueMicrotask callbacks should not crash Node
process
<!-- Thank you for reporting an issue. This issue tracker is for bugs and issues found within Node.js core. If you require more general support please file an issue on our help repo. https://github.com/nodejs/help Please fill in as much of the template below as you're able. Version: output of `node -v` Platform: output of `uname -a` (UNIX), or output of `"$([Environment]::OSVersion | ForEach-Object VersionString) $(if ([Environment]::Is64BitOperatingSystem) { "x64" } else { "x86" })"` in PowerShell console (Windows) Subsystem: if known, please specify affected core module name --> * **Version**:v15.9.0 * **Platform**: macOS x86 * **Subsystem**: task_queues ### What steps will reproduce the bug? ``` λ node --trace-uncaught Welcome to Node.js v15.9.0. Type ".help" for more information. > queueMicrotask(_ => { throw 1 }); undefined > node:internal/process/task_queues:94 runMicrotasks(); ^ 1 Thrown at: at processTicksAndRejections (node:internal/process/task_queues:94:5) ``` ### How often does it reproduce? Is there a required condition? Always. ### What is the expected behavior? If I understood the [spec](https://html.spec.whatwg.org/multipage/timers-and-user-prompts.html#microtask-queuing) correct: > The queueMicrotask(callback) method must queue a microtask to invoke callback, and if callback throws an exception, report the exception. It should behave similarly as Timer tasks. ``` > setTimeout(_ => {throw 1}); Timeout { ... } > Uncaught 1 ``` ### What do you see instead? Node crashed ### Additional information I did some preliminary triages that hopefully helps a bit: - `task_queues.js` invoke `runMicrotask` - which is a JS binding from `node_task_queue.cc` which invokes V8's `PerformCheckpoint` under the hood. - V8's `PerformCheckpoint` invoked V8's `RunMicrotasks` to actually evaluate those JS callbacks. But since V8's `PerformCheckpoint` returns `void`. I'm not sure how can Node be aware of any exceptions threw during the evaluation of those JS callbacks. Chromium seems to handle this fine but I haven't got time looking at its source.
1.0
Throw within queueMicrotask callbacks should not crash Node - <!-- Thank you for reporting an issue. This issue tracker is for bugs and issues found within Node.js core. If you require more general support please file an issue on our help repo. https://github.com/nodejs/help Please fill in as much of the template below as you're able. Version: output of `node -v` Platform: output of `uname -a` (UNIX), or output of `"$([Environment]::OSVersion | ForEach-Object VersionString) $(if ([Environment]::Is64BitOperatingSystem) { "x64" } else { "x86" })"` in PowerShell console (Windows) Subsystem: if known, please specify affected core module name --> * **Version**:v15.9.0 * **Platform**: macOS x86 * **Subsystem**: task_queues ### What steps will reproduce the bug? ``` λ node --trace-uncaught Welcome to Node.js v15.9.0. Type ".help" for more information. > queueMicrotask(_ => { throw 1 }); undefined > node:internal/process/task_queues:94 runMicrotasks(); ^ 1 Thrown at: at processTicksAndRejections (node:internal/process/task_queues:94:5) ``` ### How often does it reproduce? Is there a required condition? Always. ### What is the expected behavior? If I understood the [spec](https://html.spec.whatwg.org/multipage/timers-and-user-prompts.html#microtask-queuing) correct: > The queueMicrotask(callback) method must queue a microtask to invoke callback, and if callback throws an exception, report the exception. It should behave similarly as Timer tasks. ``` > setTimeout(_ => {throw 1}); Timeout { ... } > Uncaught 1 ``` ### What do you see instead? Node crashed ### Additional information I did some preliminary triages that hopefully helps a bit: - `task_queues.js` invoke `runMicrotask` - which is a JS binding from `node_task_queue.cc` which invokes V8's `PerformCheckpoint` under the hood. - V8's `PerformCheckpoint` invoked V8's `RunMicrotasks` to actually evaluate those JS callbacks. But since V8's `PerformCheckpoint` returns `void`. I'm not sure how can Node be aware of any exceptions threw during the evaluation of those JS callbacks. Chromium seems to handle this fine but I haven't got time looking at its source.
process
throw within queuemicrotask callbacks should not crash node thank you for reporting an issue this issue tracker is for bugs and issues found within node js core if you require more general support please file an issue on our help repo please fill in as much of the template below as you re able version output of node v platform output of uname a unix or output of osversion foreach object versionstring if else in powershell console windows subsystem if known please specify affected core module name version platform macos subsystem task queues what steps will reproduce the bug λ node trace uncaught welcome to node js type help for more information queuemicrotask throw undefined node internal process task queues runmicrotasks thrown at at processticksandrejections node internal process task queues how often does it reproduce is there a required condition always what is the expected behavior if i understood the correct the queuemicrotask callback method must queue a microtask to invoke callback and if callback throws an exception report the exception it should behave similarly as timer tasks settimeout throw timeout uncaught what do you see instead node crashed additional information i did some preliminary triages that hopefully helps a bit task queues js invoke runmicrotask which is a js binding from node task queue cc which invokes s performcheckpoint under the hood s performcheckpoint invoked s runmicrotasks to actually evaluate those js callbacks but since s performcheckpoint returns void i m not sure how can node be aware of any exceptions threw during the evaluation of those js callbacks chromium seems to handle this fine but i haven t got time looking at its source
1
3,073
5,971,430,005
IssuesEvent
2017-05-31 02:29:34
facebook/hhvm
https://api.github.com/repos/facebook/hhvm
closed
UTF-8: Chinese (probably multybyte)
php5 incompatibility
On travis (sadly private) i have following failing test on hhvm ``` 1) LispBaseTest\UtilsTest::testMbStrReplace Failed asserting that two strings are equal. --- Expected +++ Actual @@ @@ -'漢-漢 asdf' +'漢語漢 asdf' ``` With following code ``` php <?php namespace LispBaseTest; use LispBase\Utils; /** * @covers \LispBase\Utils */ class UtilsTest extends \PHPUnit_Framework_TestCase { public function testMbStrReplace() { $this->assertEquals('blu-.png', Utils::mb_str_replace('bb', '-', 'blubb.png')); $this->assertEquals('漢-漢 asdf', Utils::mb_str_replace('語', '-', '漢語漢 asdf')); $this->assertEquals('blu-..png', Utils::mb_str_replace([ 'bb', '.', ], [ '-', '..', ], 'blubb.png')); $this->assertEquals('漢-漢 語漢', Utils::mb_str_replace([ '語', 'asdf', ], [ '-', '語漢', ], '漢語漢 asdf')); } } <?php namespace LispBase; class Utils { /** * PHP DOES NOT HAVE A mb_str_replace -.- * * @param mixed $search * @param mixed $replace * @param mixed $subject * * @return string */ public static function mb_str_replace($search, $replace, $subject, &$count = 0) { // Normalize $search and $replace so they are both arrays of the same length $searches = is_array($search) ? array_values($search) : [ $search, ]; $replacements = is_array($replace) ? array_values($replace) : [ $replace, ]; $replacements = array_pad($replacements, count($searches), ''); foreach ($searches as $key => $search) { $parts = mb_split(preg_quote($search), $subject); $count += count($parts) - 1; $subject = implode($replacements[$key], $parts); } return $subject; } } ```
True
UTF-8: Chinese (probably multybyte) - On travis (sadly private) i have following failing test on hhvm ``` 1) LispBaseTest\UtilsTest::testMbStrReplace Failed asserting that two strings are equal. --- Expected +++ Actual @@ @@ -'漢-漢 asdf' +'漢語漢 asdf' ``` With following code ``` php <?php namespace LispBaseTest; use LispBase\Utils; /** * @covers \LispBase\Utils */ class UtilsTest extends \PHPUnit_Framework_TestCase { public function testMbStrReplace() { $this->assertEquals('blu-.png', Utils::mb_str_replace('bb', '-', 'blubb.png')); $this->assertEquals('漢-漢 asdf', Utils::mb_str_replace('語', '-', '漢語漢 asdf')); $this->assertEquals('blu-..png', Utils::mb_str_replace([ 'bb', '.', ], [ '-', '..', ], 'blubb.png')); $this->assertEquals('漢-漢 語漢', Utils::mb_str_replace([ '語', 'asdf', ], [ '-', '語漢', ], '漢語漢 asdf')); } } <?php namespace LispBase; class Utils { /** * PHP DOES NOT HAVE A mb_str_replace -.- * * @param mixed $search * @param mixed $replace * @param mixed $subject * * @return string */ public static function mb_str_replace($search, $replace, $subject, &$count = 0) { // Normalize $search and $replace so they are both arrays of the same length $searches = is_array($search) ? array_values($search) : [ $search, ]; $replacements = is_array($replace) ? array_values($replace) : [ $replace, ]; $replacements = array_pad($replacements, count($searches), ''); foreach ($searches as $key => $search) { $parts = mb_split(preg_quote($search), $subject); $count += count($parts) - 1; $subject = implode($replacements[$key], $parts); } return $subject; } } ```
non_process
utf chinese probably multybyte on travis sadly private i have following failing test on hhvm lispbasetest utilstest testmbstrreplace failed asserting that two strings are equal expected actual 漢 漢 asdf 漢語漢 asdf with following code php php namespace lispbasetest use lispbase utils covers lispbase utils class utilstest extends phpunit framework testcase public function testmbstrreplace this assertequals blu png utils mb str replace bb blubb png this assertequals 漢 漢 asdf utils mb str replace 語 漢語漢 asdf this assertequals blu png utils mb str replace bb blubb png this assertequals 漢 漢 語漢 utils mb str replace 語 asdf 語漢 漢語漢 asdf php namespace lispbase class utils php does not have a mb str replace param mixed search param mixed replace param mixed subject return string public static function mb str replace search replace subject count normalize search and replace so they are both arrays of the same length searches is array search array values search search replacements is array replace array values replace replace replacements array pad replacements count searches foreach searches as key search parts mb split preg quote search subject count count parts subject implode replacements parts return subject
0
578,781
17,154,347,805
IssuesEvent
2021-07-14 03:37:52
openmsupply/mobile
https://api.github.com/repos/openmsupply/mobile
closed
Support for list of checkboxes in JSON schema form
Feature Priority: normal
## Is your feature request related to a problem? Please describe. Would like to be able to use a list of checkboxes in a JSON form. For example, adding this to the schema: ``` "causes": { "title": "Possible causes", "description": "Select which vaccinations you think may have caused this reaction", "type": "array", "items": { "type": "string", "enum": ["Option One", "Option Two", "Other"] }, "uniqueItems": true }, ``` and this to the UI ``` "causes": { "ui:widget": "checkboxes" }, ``` should display a list of checkboxes.
1.0
Support for list of checkboxes in JSON schema form - ## Is your feature request related to a problem? Please describe. Would like to be able to use a list of checkboxes in a JSON form. For example, adding this to the schema: ``` "causes": { "title": "Possible causes", "description": "Select which vaccinations you think may have caused this reaction", "type": "array", "items": { "type": "string", "enum": ["Option One", "Option Two", "Other"] }, "uniqueItems": true }, ``` and this to the UI ``` "causes": { "ui:widget": "checkboxes" }, ``` should display a list of checkboxes.
non_process
support for list of checkboxes in json schema form is your feature request related to a problem please describe would like to be able to use a list of checkboxes in a json form for example adding this to the schema causes title possible causes description select which vaccinations you think may have caused this reaction type array items type string enum uniqueitems true and this to the ui causes ui widget checkboxes should display a list of checkboxes
0
19,622
25,976,619,201
IssuesEvent
2022-12-19 15:20:15
prisma/prisma
https://api.github.com/repos/prisma/prisma
closed
Cross schema reference with foreign key not working in 4.7.0.
bug/2-confirmed kind/bug process/candidate tech/engines/migration engine team/schema topic: prisma migrate dev topic: multiSchema
### Bug description Running `prisma migrate dev` gives me following error: ``` Error: P4002 The schema of the introspected database was inconsistent: Illegal cross schema reference from `public.Users` to `auth.users` in constraint `Users_authId_fkey`. Foreign keys between database schemas are not supported in Prisma. Please follow the GitHub ticket: https://github.com/prisma/prisma/issues/1175 ``` I am using postgres + supabase auth and cross reference is used from public schema to auth schema. ### How to reproduce 1. Use two separate postgres schemas 2. Reference with foreign key from one table to another 3. Run prisma migrate dev 4. See error ``` Error: P4002 The schema of the introspected database was inconsistent: Illegal cross schema reference from `public.Users` to `auth.users` in constraint `Users_authId_fkey`. Foreign keys between database schemas are not supported in Prisma. Please follow the GitHub ticket: https://github.com/prisma/prisma/issues/1175 ``` ### Expected behavior _No response_ ### Prisma information Prisma & prisma-client 4.7.0 ### Environment & setup - OS: <!--[e.g. macOS, Windows, Debian, CentOS, ...]--> MacOS - Database: <!--[PostgreSQL, MySQL, PlanetScale, MariaDB, SQL Server, SQLite, MongoDB or CockroachDB]--> Postgres 14.1 (supabase) - Node.js version: <!--[Run `node -v` to see your Node.js version]--> ### Prisma Version ``` generator client { provider = "prisma-client-js" previewFeatures = ["multiSchema"] binaryTargets = ["native", "rhel-openssl-1.0.x"] } datasource db { provider = "postgresql" url = env("DATABASE_URL") schemas = ["auth", "public"] } model User { id Int @id @default(autoincrement()) authId String @unique @db.Uuid createdAt DateTime @default(now()) users users @relation(fields: [authId], references: [id], onDelete: Cascade, onUpdate: NoAction) @@map("Users") @@schema("public") } model users { instance_id String? @db.Uuid id String @id @db.Uuid aud String? @db.VarChar(255) role String? @db.VarChar(255) email String? @unique @db.VarChar(255) encrypted_password String? @db.VarChar(255) email_confirmed_at DateTime? @db.Timestamptz(6) invited_at DateTime? @db.Timestamptz(6) confirmation_token String? @db.VarChar(255) confirmation_sent_at DateTime? @db.Timestamptz(6) recovery_token String? @db.VarChar(255) recovery_sent_at DateTime? @db.Timestamptz(6) email_change_token_new String? @db.VarChar(255) email_change String? @db.VarChar(255) email_change_sent_at DateTime? @db.Timestamptz(6) last_sign_in_at DateTime? @db.Timestamptz(6) raw_app_meta_data Json? raw_user_meta_data Json? is_super_admin Boolean? created_at DateTime? @db.Timestamptz(6) updated_at DateTime? @db.Timestamptz(6) phone String? @unique @db.VarChar(15) phone_confirmed_at DateTime? @db.Timestamptz(6) phone_change String? @default("") @db.VarChar(15) phone_change_token String? @default("") @db.VarChar(255) phone_change_sent_at DateTime? @db.Timestamptz(6) confirmed_at DateTime? @default(dbgenerated("LEAST(email_confirmed_at, phone_confirmed_at)")) @db.Timestamptz(6) email_change_token_current String? @default("") @db.VarChar(255) email_change_confirm_status Int? @default(0) @db.SmallInt banned_until DateTime? @db.Timestamptz(6) reauthentication_token String? @default("") @db.VarChar(255) reauthentication_sent_at DateTime? @db.Timestamptz(6) identities identities[] mfa_factors mfa_factors[] sessions sessions[] User User? @@index([instance_id]) @@schema("auth") } ```
1.0
Cross schema reference with foreign key not working in 4.7.0. - ### Bug description Running `prisma migrate dev` gives me following error: ``` Error: P4002 The schema of the introspected database was inconsistent: Illegal cross schema reference from `public.Users` to `auth.users` in constraint `Users_authId_fkey`. Foreign keys between database schemas are not supported in Prisma. Please follow the GitHub ticket: https://github.com/prisma/prisma/issues/1175 ``` I am using postgres + supabase auth and cross reference is used from public schema to auth schema. ### How to reproduce 1. Use two separate postgres schemas 2. Reference with foreign key from one table to another 3. Run prisma migrate dev 4. See error ``` Error: P4002 The schema of the introspected database was inconsistent: Illegal cross schema reference from `public.Users` to `auth.users` in constraint `Users_authId_fkey`. Foreign keys between database schemas are not supported in Prisma. Please follow the GitHub ticket: https://github.com/prisma/prisma/issues/1175 ``` ### Expected behavior _No response_ ### Prisma information Prisma & prisma-client 4.7.0 ### Environment & setup - OS: <!--[e.g. macOS, Windows, Debian, CentOS, ...]--> MacOS - Database: <!--[PostgreSQL, MySQL, PlanetScale, MariaDB, SQL Server, SQLite, MongoDB or CockroachDB]--> Postgres 14.1 (supabase) - Node.js version: <!--[Run `node -v` to see your Node.js version]--> ### Prisma Version ``` generator client { provider = "prisma-client-js" previewFeatures = ["multiSchema"] binaryTargets = ["native", "rhel-openssl-1.0.x"] } datasource db { provider = "postgresql" url = env("DATABASE_URL") schemas = ["auth", "public"] } model User { id Int @id @default(autoincrement()) authId String @unique @db.Uuid createdAt DateTime @default(now()) users users @relation(fields: [authId], references: [id], onDelete: Cascade, onUpdate: NoAction) @@map("Users") @@schema("public") } model users { instance_id String? @db.Uuid id String @id @db.Uuid aud String? @db.VarChar(255) role String? @db.VarChar(255) email String? @unique @db.VarChar(255) encrypted_password String? @db.VarChar(255) email_confirmed_at DateTime? @db.Timestamptz(6) invited_at DateTime? @db.Timestamptz(6) confirmation_token String? @db.VarChar(255) confirmation_sent_at DateTime? @db.Timestamptz(6) recovery_token String? @db.VarChar(255) recovery_sent_at DateTime? @db.Timestamptz(6) email_change_token_new String? @db.VarChar(255) email_change String? @db.VarChar(255) email_change_sent_at DateTime? @db.Timestamptz(6) last_sign_in_at DateTime? @db.Timestamptz(6) raw_app_meta_data Json? raw_user_meta_data Json? is_super_admin Boolean? created_at DateTime? @db.Timestamptz(6) updated_at DateTime? @db.Timestamptz(6) phone String? @unique @db.VarChar(15) phone_confirmed_at DateTime? @db.Timestamptz(6) phone_change String? @default("") @db.VarChar(15) phone_change_token String? @default("") @db.VarChar(255) phone_change_sent_at DateTime? @db.Timestamptz(6) confirmed_at DateTime? @default(dbgenerated("LEAST(email_confirmed_at, phone_confirmed_at)")) @db.Timestamptz(6) email_change_token_current String? @default("") @db.VarChar(255) email_change_confirm_status Int? @default(0) @db.SmallInt banned_until DateTime? @db.Timestamptz(6) reauthentication_token String? @default("") @db.VarChar(255) reauthentication_sent_at DateTime? @db.Timestamptz(6) identities identities[] mfa_factors mfa_factors[] sessions sessions[] User User? @@index([instance_id]) @@schema("auth") } ```
process
cross schema reference with foreign key not working in bug description running prisma migrate dev gives me following error error the schema of the introspected database was inconsistent illegal cross schema reference from public users to auth users in constraint users authid fkey foreign keys between database schemas are not supported in prisma please follow the github ticket i am using postgres supabase auth and cross reference is used from public schema to auth schema how to reproduce use two separate postgres schemas reference with foreign key from one table to another run prisma migrate dev see error error the schema of the introspected database was inconsistent illegal cross schema reference from public users to auth users in constraint users authid fkey foreign keys between database schemas are not supported in prisma please follow the github ticket expected behavior no response prisma information prisma prisma client environment setup os macos database postgres supabase node js version prisma version generator client provider prisma client js previewfeatures binarytargets datasource db provider postgresql url env database url schemas model user id int id default autoincrement authid string unique db uuid createdat datetime default now users users relation fields references ondelete cascade onupdate noaction map users schema public model users instance id string db uuid id string id db uuid aud string db varchar role string db varchar email string unique db varchar encrypted password string db varchar email confirmed at datetime db timestamptz invited at datetime db timestamptz confirmation token string db varchar confirmation sent at datetime db timestamptz recovery token string db varchar recovery sent at datetime db timestamptz email change token new string db varchar email change string db varchar email change sent at datetime db timestamptz last sign in at datetime db timestamptz raw app meta data json raw user meta data json is super admin boolean created at datetime db timestamptz updated at datetime db timestamptz phone string unique db varchar phone confirmed at datetime db timestamptz phone change string default db varchar phone change token string default db varchar phone change sent at datetime db timestamptz confirmed at datetime default dbgenerated least email confirmed at phone confirmed at db timestamptz email change token current string default db varchar email change confirm status int default db smallint banned until datetime db timestamptz reauthentication token string default db varchar reauthentication sent at datetime db timestamptz identities identities mfa factors mfa factors sessions sessions user user index schema auth
1
8,848
11,951,366,714
IssuesEvent
2020-04-03 16:45:03
MicrosoftDocs/vsts-docs
https://api.github.com/repos/MicrosoftDocs/vsts-docs
closed
Runtime expression syntax for variables doesn't work for env parameter of the powershell task
Pri1 devops-cicd-process/tech devops/prod
``` variables: myVar: 'My Var' steps: - powershell: | Write-Host 'My environment variables:' Get-ChildItem env:my* | Sort-Object Name | % { Write-Host ($_.Name + ': ' + $_.Value) } env: myEnvVarRunTime1: $(myVar) myEnvVarRunTime2: $[variables.myVar] myEnvVarCompileTime: ${{ variables.myVar }} myEnvVarUndefRunTime1: $(myUndefVar) myEnvVarUndefRunTime2: $[variables.myUndefVar] myEnvVarUndefCompileTime: ${{ variables.myUndefVar }} ``` Produces this output: ``` My environment variables: myEnvVarCompileTime: My Var myEnvVarRunTime1: My Var myEnvVarRunTime2: $[variables.myVar] myEnvVarUndefCompileTime: myEnvVarUndefRunTime1: $(myUndefVar) myEnvVarUndefRunTime2: $[variables.myUndefVar] MYVAR: My Var ``` Why doesn't this match the runtime expression variable behavior as described in the `Understand variable syntax` section? 1. The variable isn't being expanded when it is defined 2. The variable should be silently coalesced to the empty string when it isn't found --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 77c58a78-a567-e99a-9eb7-62dddd1b90b6 * Version Independent ID: 680a79bc-11de-39fc-43e3-e07dc762db18 * Content: [Expressions - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/expressions?view=azure-devops#feedback) * Content Source: [docs/pipelines/process/expressions.md](https://github.com/MicrosoftDocs/vsts-docs/blob/master/docs/pipelines/process/expressions.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
1.0
Runtime expression syntax for variables doesn't work for env parameter of the powershell task - ``` variables: myVar: 'My Var' steps: - powershell: | Write-Host 'My environment variables:' Get-ChildItem env:my* | Sort-Object Name | % { Write-Host ($_.Name + ': ' + $_.Value) } env: myEnvVarRunTime1: $(myVar) myEnvVarRunTime2: $[variables.myVar] myEnvVarCompileTime: ${{ variables.myVar }} myEnvVarUndefRunTime1: $(myUndefVar) myEnvVarUndefRunTime2: $[variables.myUndefVar] myEnvVarUndefCompileTime: ${{ variables.myUndefVar }} ``` Produces this output: ``` My environment variables: myEnvVarCompileTime: My Var myEnvVarRunTime1: My Var myEnvVarRunTime2: $[variables.myVar] myEnvVarUndefCompileTime: myEnvVarUndefRunTime1: $(myUndefVar) myEnvVarUndefRunTime2: $[variables.myUndefVar] MYVAR: My Var ``` Why doesn't this match the runtime expression variable behavior as described in the `Understand variable syntax` section? 1. The variable isn't being expanded when it is defined 2. The variable should be silently coalesced to the empty string when it isn't found --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 77c58a78-a567-e99a-9eb7-62dddd1b90b6 * Version Independent ID: 680a79bc-11de-39fc-43e3-e07dc762db18 * Content: [Expressions - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/expressions?view=azure-devops#feedback) * Content Source: [docs/pipelines/process/expressions.md](https://github.com/MicrosoftDocs/vsts-docs/blob/master/docs/pipelines/process/expressions.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
process
runtime expression syntax for variables doesn t work for env parameter of the powershell task variables myvar my var steps powershell write host my environment variables get childitem env my sort object name write host name value env myvar myenvvarcompiletime variables myvar myundefvar myenvvarundefcompiletime variables myundefvar produces this output my environment variables myenvvarcompiletime my var my var myenvvarundefcompiletime myundefvar myvar my var why doesn t this match the runtime expression variable behavior as described in the understand variable syntax section the variable isn t being expanded when it is defined the variable should be silently coalesced to the empty string when it isn t found document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
1
15,924
20,142,201,896
IssuesEvent
2022-02-09 01:10:59
JustBru00/RenamePlugin
https://api.github.com/repos/JustBru00/RenamePlugin
opened
Harden Grindstone Exploit Protection
TODO High Priority Processing CRITICAL
Requested by @ralphie12 on spigotmc.org. Details will be filled out here after a fix has been released.
1.0
Harden Grindstone Exploit Protection - Requested by @ralphie12 on spigotmc.org. Details will be filled out here after a fix has been released.
process
harden grindstone exploit protection requested by on spigotmc org details will be filled out here after a fix has been released
1
278,142
30,702,211,796
IssuesEvent
2023-07-27 01:11:50
snykiotcubedev/arangodb-3.7.6
https://api.github.com/repos/snykiotcubedev/arangodb-3.7.6
reopened
CVE-2021-33623 (High) detected in trim-newlines-1.0.0.tgz
Mend: dependency security vulnerability
## CVE-2021-33623 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>trim-newlines-1.0.0.tgz</b></p></summary> <p>Trim newlines from the start and/or end of a string</p> <p>Library home page: <a href="https://registry.npmjs.org/trim-newlines/-/trim-newlines-1.0.0.tgz">https://registry.npmjs.org/trim-newlines/-/trim-newlines-1.0.0.tgz</a></p> <p> Dependency Hierarchy: - node-sass-4.14.1.tgz (Root Library) - meow-3.7.0.tgz - :x: **trim-newlines-1.0.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/snykiotcubedev/arangodb-3.7.6/commit/fce8f85f1c2f070c8e6a8e76d17210a2117d3833">fce8f85f1c2f070c8e6a8e76d17210a2117d3833</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> The trim-newlines package before 3.0.1 and 4.x before 4.0.1 for Node.js has an issue related to regular expression denial-of-service (ReDoS) for the .end() method. <p>Publish Date: 2021-05-28 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-33623>CVE-2021-33623</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-33623">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-33623</a></p> <p>Release Date: 2021-05-28</p> <p>Fix Resolution (trim-newlines): 3.0.1</p> <p>Direct dependency fix Resolution (node-sass): 6.0.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-33623 (High) detected in trim-newlines-1.0.0.tgz - ## CVE-2021-33623 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>trim-newlines-1.0.0.tgz</b></p></summary> <p>Trim newlines from the start and/or end of a string</p> <p>Library home page: <a href="https://registry.npmjs.org/trim-newlines/-/trim-newlines-1.0.0.tgz">https://registry.npmjs.org/trim-newlines/-/trim-newlines-1.0.0.tgz</a></p> <p> Dependency Hierarchy: - node-sass-4.14.1.tgz (Root Library) - meow-3.7.0.tgz - :x: **trim-newlines-1.0.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/snykiotcubedev/arangodb-3.7.6/commit/fce8f85f1c2f070c8e6a8e76d17210a2117d3833">fce8f85f1c2f070c8e6a8e76d17210a2117d3833</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> The trim-newlines package before 3.0.1 and 4.x before 4.0.1 for Node.js has an issue related to regular expression denial-of-service (ReDoS) for the .end() method. <p>Publish Date: 2021-05-28 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-33623>CVE-2021-33623</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-33623">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-33623</a></p> <p>Release Date: 2021-05-28</p> <p>Fix Resolution (trim-newlines): 3.0.1</p> <p>Direct dependency fix Resolution (node-sass): 6.0.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in trim newlines tgz cve high severity vulnerability vulnerable library trim newlines tgz trim newlines from the start and or end of a string library home page a href dependency hierarchy node sass tgz root library meow tgz x trim newlines tgz vulnerable library found in head commit a href found in base branch main vulnerability details the trim newlines package before and x before for node js has an issue related to regular expression denial of service redos for the end method publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution trim newlines direct dependency fix resolution node sass step up your open source security game with mend
0
14,258
17,192,781,556
IssuesEvent
2021-07-16 13:23:27
qgis/QGIS-Documentation
https://api.github.com/repos/qgis/QGIS-Documentation
closed
[processing][needs-docs] Add friendlier API for running algorithms as sub-steps of main algorithm
3.4 Automatic new feature Processing enhancement
Original commit: https://github.com/qgis/QGIS/commit/7f7c7a97899637f59ceb76f30abbec6ba924bec6 by nyalldawson Using code like: buffered_layer = processing.run(..., context, feedback)['OUTPUT'] ... return {'OUTPUT': buffered_layer} can cause issues if done as a sub-step of a larger processing algorithm. This is because ownership of the generated layer is transferred to the caller (Python) by processing.run. When the algorithm returns, Processing attempts to move ownership of the layer from the context to the caller, resulting in a crash. (This is by design, because processing.run has been optimised for the most common use case, which is one-off execution of algorithms as part of a script, not as part of another processing algorithm. Accordingly by design it returns layers and ownership to the caller, making things easier for callers as they do not then have to resolve the layer reference from the context object and handle ownership themselves) This commit adds a new "is_child_algorithm" argument to processing.run. For algorithms which are executed as sub-steps of a larger algorithm is_child_algorithm should be set to True to avoid any ownership issues with layers. E.g. buffered_layer = processing.run(..., context, feedback, is_child_algorithm=True)['OUTPUT'] ... return {'OUTPUT': buffered_layer}
1.0
[processing][needs-docs] Add friendlier API for running algorithms as sub-steps of main algorithm - Original commit: https://github.com/qgis/QGIS/commit/7f7c7a97899637f59ceb76f30abbec6ba924bec6 by nyalldawson Using code like: buffered_layer = processing.run(..., context, feedback)['OUTPUT'] ... return {'OUTPUT': buffered_layer} can cause issues if done as a sub-step of a larger processing algorithm. This is because ownership of the generated layer is transferred to the caller (Python) by processing.run. When the algorithm returns, Processing attempts to move ownership of the layer from the context to the caller, resulting in a crash. (This is by design, because processing.run has been optimised for the most common use case, which is one-off execution of algorithms as part of a script, not as part of another processing algorithm. Accordingly by design it returns layers and ownership to the caller, making things easier for callers as they do not then have to resolve the layer reference from the context object and handle ownership themselves) This commit adds a new "is_child_algorithm" argument to processing.run. For algorithms which are executed as sub-steps of a larger algorithm is_child_algorithm should be set to True to avoid any ownership issues with layers. E.g. buffered_layer = processing.run(..., context, feedback, is_child_algorithm=True)['OUTPUT'] ... return {'OUTPUT': buffered_layer}
process
add friendlier api for running algorithms as sub steps of main algorithm original commit by nyalldawson using code like buffered layer processing run context feedback return output buffered layer can cause issues if done as a sub step of a larger processing algorithm this is because ownership of the generated layer is transferred to the caller python by processing run when the algorithm returns processing attempts to move ownership of the layer from the context to the caller resulting in a crash this is by design because processing run has been optimised for the most common use case which is one off execution of algorithms as part of a script not as part of another processing algorithm accordingly by design it returns layers and ownership to the caller making things easier for callers as they do not then have to resolve the layer reference from the context object and handle ownership themselves this commit adds a new is child algorithm argument to processing run for algorithms which are executed as sub steps of a larger algorithm is child algorithm should be set to true to avoid any ownership issues with layers e g buffered layer processing run context feedback is child algorithm true return output buffered layer
1
25,872
12,307,848,315
IssuesEvent
2020-05-12 05:51:22
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Unable to view virtual node after enabling virtual node addon with advanced networking
Pri2 container-service/svc cxp product-question triaged
I created an AKS cluster via Azure CLI with advanced networking (CNI). During the creation of the cluster, I set the **--max-pods** to 250 and after creating the cluster I enabled an a**ddon for the virtual node**, however, the command `$kubectl get nodes` does not show the created virtual node. Please help! --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 916ac90f-8a94-5f3d-eb99-fe06c39eb6cf * Version Independent ID: 2aad2e47-7fde-8d24-f824-236a26e5d887 * Content: [Create virtual nodes using Azure CLI - Azure Kubernetes Service](https://docs.microsoft.com/en-us/azure/aks/virtual-nodes-cli) * Content Source: [articles/aks/virtual-nodes-cli.md](https://github.com/Microsoft/azure-docs/blob/master/articles/aks/virtual-nodes-cli.md) * Service: **container-service** * GitHub Login: @mlearned * Microsoft Alias: **mlearned**
1.0
Unable to view virtual node after enabling virtual node addon with advanced networking - I created an AKS cluster via Azure CLI with advanced networking (CNI). During the creation of the cluster, I set the **--max-pods** to 250 and after creating the cluster I enabled an a**ddon for the virtual node**, however, the command `$kubectl get nodes` does not show the created virtual node. Please help! --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 916ac90f-8a94-5f3d-eb99-fe06c39eb6cf * Version Independent ID: 2aad2e47-7fde-8d24-f824-236a26e5d887 * Content: [Create virtual nodes using Azure CLI - Azure Kubernetes Service](https://docs.microsoft.com/en-us/azure/aks/virtual-nodes-cli) * Content Source: [articles/aks/virtual-nodes-cli.md](https://github.com/Microsoft/azure-docs/blob/master/articles/aks/virtual-nodes-cli.md) * Service: **container-service** * GitHub Login: @mlearned * Microsoft Alias: **mlearned**
non_process
unable to view virtual node after enabling virtual node addon with advanced networking i created an aks cluster via azure cli with advanced networking cni during the creation of the cluster i set the max pods to and after creating the cluster i enabled an a ddon for the virtual node however the command kubectl get nodes does not show the created virtual node please help document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service container service github login mlearned microsoft alias mlearned
0
75,225
20,732,569,513
IssuesEvent
2022-03-14 10:46:47
ElektraInitiative/libelektra
https://api.github.com/repos/ElektraInitiative/libelektra
closed
process: dump plugin excluded
build lang/cmake
process plugin has wrong inclusion logic and a broken error message: ```sh dump plugin not found (excluded by category EXPERIMENTAL), excluding the process plugin ... -- Include plugin dump ```
1.0
process: dump plugin excluded - process plugin has wrong inclusion logic and a broken error message: ```sh dump plugin not found (excluded by category EXPERIMENTAL), excluding the process plugin ... -- Include plugin dump ```
non_process
process dump plugin excluded process plugin has wrong inclusion logic and a broken error message sh dump plugin not found excluded by category experimental excluding the process plugin include plugin dump
0
9,167
12,220,498,670
IssuesEvent
2020-05-02 01:56:44
BlesseNtumble/GalaxySpace
https://api.github.com/repos/BlesseNtumble/GalaxySpace
closed
Can't load world
1.12.2 already fixed in next version in the process of correcting
Versions: ------------------------------------------------------------------------ 1. Minecraft version: 1.12.2 2. Galacticraft version: 4.0.2.254 3. GalaxySpace version: 2.0.12 4. AsmodeusCore version: 0.0.13 5. Side : Single player ------------------------------------------------------------------------ Description of the issue: When loading a world it crashes without debug mode on. but will always crash when trying to reload the world. ------------------------------------------------------------------------ Screenshot/Video: None ------------------------------------------------------------------------ Attached log file (or url on pastebin.com): [latest.log](https://github.com/BlesseNtumble/GalaxySpace/files/4503506/latest.log) ------------------------------------------------------------------------
1.0
Can't load world - Versions: ------------------------------------------------------------------------ 1. Minecraft version: 1.12.2 2. Galacticraft version: 4.0.2.254 3. GalaxySpace version: 2.0.12 4. AsmodeusCore version: 0.0.13 5. Side : Single player ------------------------------------------------------------------------ Description of the issue: When loading a world it crashes without debug mode on. but will always crash when trying to reload the world. ------------------------------------------------------------------------ Screenshot/Video: None ------------------------------------------------------------------------ Attached log file (or url on pastebin.com): [latest.log](https://github.com/BlesseNtumble/GalaxySpace/files/4503506/latest.log) ------------------------------------------------------------------------
process
can t load world versions minecraft version galacticraft version galaxyspace version asmodeuscore version side single player description of the issue when loading a world it crashes without debug mode on but will always crash when trying to reload the world screenshot video none attached log file or url on pastebin com
1
561,446
16,617,593,810
IssuesEvent
2021-06-02 18:52:07
openforcefield/openff-toolkit
https://api.github.com/repos/openforcefield/openff-toolkit
closed
VdW Cutoff Ignored when Electrostatic Handler Present
bug effort:low priority:high
**Describe the bug** The non-bonded cutoff is hardcoded in several places (the [`vdWHandler`](https://github.com/openforcefield/openforcefield/blob/7091de5daf03abfa487e7c3619eafcfcbff91d1d/openforcefield/typing/engines/smirnoff/parameters.py#L3226) and the [`ElectrostaticsHandler`](https://github.com/openforcefield/openforcefield/blob/7091de5daf03abfa487e7c3619eafcfcbff91d1d/openforcefield/typing/engines/smirnoff/parameters.py#L3524)). This hardcoded value overwrites the value set by both the `vdWHandler` and the `ElectrostaticsHandler`. **To Reproduce** ``` # Create a dummy topology containing only argon and give it a set of # box vectors. topology = Molecule.from_smiles("[Ar]").to_topology() topology.box_vectors = numpy.eye(3) * unit.nanometers # Create the FF force_field = ForceField() # Add only a VdW handler. vdw_handler = vdWHandler(version=0.3) vdw_handler.method = "cutoff" vdw_handler.cutoff = 6.0 * unit.angstrom vdw_handler.scale14 = 1.0 vdw_handler.add_parameter( { "smirks": "[#18:1]", "epsilon": 1.0 * unit.kilojoules_per_mole, "sigma": 1.0 * unit.angstrom } ) force_field.register_parameter_handler(vdw_handler) # Map the FF to the dummy topology system = force_field.create_openmm_system(topology) assert numpy.isclose( system.getForce(0).getCutoffDistance().value_in_unit(unit.angstrom), 6.0 ) # Add an electrostatics handler. electrostatics_handler = ElectrostaticsHandler(version=0.3) electrostatics_handler.cutoff = 6.0 * unit.angstrom electrostatics_handler.method = "PME" force_field.register_parameter_handler(electrostatics_handler) library_charges = LibraryChargeHandler(version=0.3) library_charges.add_parameter( { "smirks": "[#18:1]", "charge1": 0.0 * unit.elementary_charge, } ) force_field.register_parameter_handler(library_charges) system = force_field.create_openmm_system(topology) assert numpy.isclose( system.getForce(0).getCutoffDistance().value_in_unit(unit.angstrom), 6.0 ) ``` **Output** ```AssertionError``` on the final assert. **Computing environment (please complete the following information):** - `openforcefield = 0.7.2`
1.0
VdW Cutoff Ignored when Electrostatic Handler Present - **Describe the bug** The non-bonded cutoff is hardcoded in several places (the [`vdWHandler`](https://github.com/openforcefield/openforcefield/blob/7091de5daf03abfa487e7c3619eafcfcbff91d1d/openforcefield/typing/engines/smirnoff/parameters.py#L3226) and the [`ElectrostaticsHandler`](https://github.com/openforcefield/openforcefield/blob/7091de5daf03abfa487e7c3619eafcfcbff91d1d/openforcefield/typing/engines/smirnoff/parameters.py#L3524)). This hardcoded value overwrites the value set by both the `vdWHandler` and the `ElectrostaticsHandler`. **To Reproduce** ``` # Create a dummy topology containing only argon and give it a set of # box vectors. topology = Molecule.from_smiles("[Ar]").to_topology() topology.box_vectors = numpy.eye(3) * unit.nanometers # Create the FF force_field = ForceField() # Add only a VdW handler. vdw_handler = vdWHandler(version=0.3) vdw_handler.method = "cutoff" vdw_handler.cutoff = 6.0 * unit.angstrom vdw_handler.scale14 = 1.0 vdw_handler.add_parameter( { "smirks": "[#18:1]", "epsilon": 1.0 * unit.kilojoules_per_mole, "sigma": 1.0 * unit.angstrom } ) force_field.register_parameter_handler(vdw_handler) # Map the FF to the dummy topology system = force_field.create_openmm_system(topology) assert numpy.isclose( system.getForce(0).getCutoffDistance().value_in_unit(unit.angstrom), 6.0 ) # Add an electrostatics handler. electrostatics_handler = ElectrostaticsHandler(version=0.3) electrostatics_handler.cutoff = 6.0 * unit.angstrom electrostatics_handler.method = "PME" force_field.register_parameter_handler(electrostatics_handler) library_charges = LibraryChargeHandler(version=0.3) library_charges.add_parameter( { "smirks": "[#18:1]", "charge1": 0.0 * unit.elementary_charge, } ) force_field.register_parameter_handler(library_charges) system = force_field.create_openmm_system(topology) assert numpy.isclose( system.getForce(0).getCutoffDistance().value_in_unit(unit.angstrom), 6.0 ) ``` **Output** ```AssertionError``` on the final assert. **Computing environment (please complete the following information):** - `openforcefield = 0.7.2`
non_process
vdw cutoff ignored when electrostatic handler present describe the bug the non bonded cutoff is hardcoded in several places the and the this hardcoded value overwrites the value set by both the vdwhandler and the electrostaticshandler to reproduce create a dummy topology containing only argon and give it a set of box vectors topology molecule from smiles to topology topology box vectors numpy eye unit nanometers create the ff force field forcefield add only a vdw handler vdw handler vdwhandler version vdw handler method cutoff vdw handler cutoff unit angstrom vdw handler vdw handler add parameter smirks epsilon unit kilojoules per mole sigma unit angstrom force field register parameter handler vdw handler map the ff to the dummy topology system force field create openmm system topology assert numpy isclose system getforce getcutoffdistance value in unit unit angstrom add an electrostatics handler electrostatics handler electrostaticshandler version electrostatics handler cutoff unit angstrom electrostatics handler method pme force field register parameter handler electrostatics handler library charges librarychargehandler version library charges add parameter smirks unit elementary charge force field register parameter handler library charges system force field create openmm system topology assert numpy isclose system getforce getcutoffdistance value in unit unit angstrom output assertionerror on the final assert computing environment please complete the following information openforcefield
0
6,128
8,996,635,764
IssuesEvent
2019-02-02 03:04:17
bow-simulation/virtualbow
https://api.github.com/repos/bow-simulation/virtualbow
opened
Write tests for all numerical utilities (splines, root-finding, geometry)
area: fem/numerics area: software process prio: normal type: improvement
In GitLab by **spfeifer** on Dec 7, 2018, 12:14
1.0
Write tests for all numerical utilities (splines, root-finding, geometry) - In GitLab by **spfeifer** on Dec 7, 2018, 12:14
process
write tests for all numerical utilities splines root finding geometry in gitlab by spfeifer on dec
1
34,929
30,593,627,697
IssuesEvent
2023-07-21 19:28:38
google/site-kit-wp
https://api.github.com/repos/google/site-kit-wp
opened
Reorganize ABR components for consistency
P1 Module: AdSense Type: Infrastructure
## Feature Description There are a few inconsistencies that have been introduced with the components for ad blocking recovery regarding component location in the repo and naming. --------------- _Do not alter or remove anything below. The following sections will be managed by moderators only._ ## Acceptance criteria * The `AdBlockingRecoveryWidget` component should be **renamed** to `AdBlockingRecoverySetupCTAWidget` * The `AdBlockingRecoveryApp` component should be **moved** under the `components/setup` directory * The `AdBlockingRecoveryToggle` component should be **moved** under the `components/settings` directory * The `AdBlockingRecoveryCTA` component should be **moved** under the `components/settings` directory and **renamed** to `AdBlockingRecoverySetupCTANotice` * The `AdBlockingRecoveryNotification` component should be renamed to `AdBlockingRecoverySetupSuccessBannerNotification` * All relocated components should include their complementary co-located stories and tests * Stories and related VRT scenarios should be renamed accordingly ## Implementation Brief * <!-- One or more bullet points for how to technically implement the feature. Make sure to include changes to Storybook and visual regression tests where relevant. --> ### Test Coverage * <!-- One or more bullet points for how to implement automated tests to verify the feature works. --> ## QA Brief * <!-- One or more bullet points for how to test that the feature works as expected. --> ## Changelog entry * <!-- One sentence summarizing the PR, to be used in the changelog. -->
1.0
Reorganize ABR components for consistency - ## Feature Description There are a few inconsistencies that have been introduced with the components for ad blocking recovery regarding component location in the repo and naming. --------------- _Do not alter or remove anything below. The following sections will be managed by moderators only._ ## Acceptance criteria * The `AdBlockingRecoveryWidget` component should be **renamed** to `AdBlockingRecoverySetupCTAWidget` * The `AdBlockingRecoveryApp` component should be **moved** under the `components/setup` directory * The `AdBlockingRecoveryToggle` component should be **moved** under the `components/settings` directory * The `AdBlockingRecoveryCTA` component should be **moved** under the `components/settings` directory and **renamed** to `AdBlockingRecoverySetupCTANotice` * The `AdBlockingRecoveryNotification` component should be renamed to `AdBlockingRecoverySetupSuccessBannerNotification` * All relocated components should include their complementary co-located stories and tests * Stories and related VRT scenarios should be renamed accordingly ## Implementation Brief * <!-- One or more bullet points for how to technically implement the feature. Make sure to include changes to Storybook and visual regression tests where relevant. --> ### Test Coverage * <!-- One or more bullet points for how to implement automated tests to verify the feature works. --> ## QA Brief * <!-- One or more bullet points for how to test that the feature works as expected. --> ## Changelog entry * <!-- One sentence summarizing the PR, to be used in the changelog. -->
non_process
reorganize abr components for consistency feature description there are a few inconsistencies that have been introduced with the components for ad blocking recovery regarding component location in the repo and naming do not alter or remove anything below the following sections will be managed by moderators only acceptance criteria the adblockingrecoverywidget component should be renamed to adblockingrecoverysetupctawidget the adblockingrecoveryapp component should be moved under the components setup directory the adblockingrecoverytoggle component should be moved under the components settings directory the adblockingrecoverycta component should be moved under the components settings directory and renamed to adblockingrecoverysetupctanotice the adblockingrecoverynotification component should be renamed to adblockingrecoverysetupsuccessbannernotification all relocated components should include their complementary co located stories and tests stories and related vrt scenarios should be renamed accordingly implementation brief test coverage qa brief changelog entry
0
5,391
3,213,258,454
IssuesEvent
2015-10-06 19:03:35
PancakeBot/PancakePainter
https://api.github.com/repos/PancakeBot/PancakePainter
closed
Color Transitions in GCODE
gcode render question
We may have color selections from #12, but it doesn't tell the entire story. When I originally discussed with Miguel, he mentioned there would be "some special GCODE command" that would either allow for 30(?) seconds of time to pass (to let the first layer cook darker), _or_ would allow an undetermined amount of time to allow for the user to change the bottles (and thus, change the batter color), only continuing after they've pressed a button on the machine. This is perfectly reasonable, though there's two clear issues: 0. I've not yet seen any such special GCODE command, I suppose we could just force a 30 second dwell. 0. It is not clear if this is meant to happen every time regardless of which method is intended. If a drawing is done all in "dark", would it draw everything, then simply wait for three "cycles" worth of time? Would user interaction be required for all of those? 0. Do we need a GCODE command for both this magic wait, AND a config as to whether it's a TIME wait, or a user interaction wait? For now I'm moving on with a made up GCODE (`M142`), and I'll train the simulator to fake batter color offset.
1.0
Color Transitions in GCODE - We may have color selections from #12, but it doesn't tell the entire story. When I originally discussed with Miguel, he mentioned there would be "some special GCODE command" that would either allow for 30(?) seconds of time to pass (to let the first layer cook darker), _or_ would allow an undetermined amount of time to allow for the user to change the bottles (and thus, change the batter color), only continuing after they've pressed a button on the machine. This is perfectly reasonable, though there's two clear issues: 0. I've not yet seen any such special GCODE command, I suppose we could just force a 30 second dwell. 0. It is not clear if this is meant to happen every time regardless of which method is intended. If a drawing is done all in "dark", would it draw everything, then simply wait for three "cycles" worth of time? Would user interaction be required for all of those? 0. Do we need a GCODE command for both this magic wait, AND a config as to whether it's a TIME wait, or a user interaction wait? For now I'm moving on with a made up GCODE (`M142`), and I'll train the simulator to fake batter color offset.
non_process
color transitions in gcode we may have color selections from but it doesn t tell the entire story when i originally discussed with miguel he mentioned there would be some special gcode command that would either allow for seconds of time to pass to let the first layer cook darker or would allow an undetermined amount of time to allow for the user to change the bottles and thus change the batter color only continuing after they ve pressed a button on the machine this is perfectly reasonable though there s two clear issues i ve not yet seen any such special gcode command i suppose we could just force a second dwell it is not clear if this is meant to happen every time regardless of which method is intended if a drawing is done all in dark would it draw everything then simply wait for three cycles worth of time would user interaction be required for all of those do we need a gcode command for both this magic wait and a config as to whether it s a time wait or a user interaction wait for now i m moving on with a made up gcode and i ll train the simulator to fake batter color offset
0
6,993
10,142,815,047
IssuesEvent
2019-08-04 05:41:29
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
TIN interpolation crash
Bug Crash/Data Corruption Processing
TIN interpolation crash with this contour line Crash ID: 185de4a9a11eb29f05672d8b176e731c0d36ee2f Stack Trace MathUtils::lineIntersection : DualEdgeTriangulation::insertForcedSegment : DualEdgeTriangulation::addLine : QgsTinInterpolator::insertData : QgsTinInterpolator::initialize : QgsTinInterpolator::interpolatePoint : PyInit__analysis : QgsGridFileWriter::writeFile : PyInit__analysis : PyMethodDef_RawFastCallKeywords : PyMethodDef_RawFastCallKeywords : PyEval_EvalFrameDefault : PyFunction_FastCallDict : PyMethodDef_RawFastCallDict : PyObject_Call : PyInit_sip : CPLStringList::empty : PyInit__core : QgsProcessingAlgorithm::runPrepared : QgsProcessingAlgRunnerTask::run : PyInit__core : QgsTask::start : QThreadPoolPrivate::reset : QThread::start : BaseThreadInitThunk : RtlUserThreadStart : [Isolinee_1953.zip](https://github.com/qgis/QGIS/files/3452397/Isolinee_1953.zip)
1.0
TIN interpolation crash - TIN interpolation crash with this contour line Crash ID: 185de4a9a11eb29f05672d8b176e731c0d36ee2f Stack Trace MathUtils::lineIntersection : DualEdgeTriangulation::insertForcedSegment : DualEdgeTriangulation::addLine : QgsTinInterpolator::insertData : QgsTinInterpolator::initialize : QgsTinInterpolator::interpolatePoint : PyInit__analysis : QgsGridFileWriter::writeFile : PyInit__analysis : PyMethodDef_RawFastCallKeywords : PyMethodDef_RawFastCallKeywords : PyEval_EvalFrameDefault : PyFunction_FastCallDict : PyMethodDef_RawFastCallDict : PyObject_Call : PyInit_sip : CPLStringList::empty : PyInit__core : QgsProcessingAlgorithm::runPrepared : QgsProcessingAlgRunnerTask::run : PyInit__core : QgsTask::start : QThreadPoolPrivate::reset : QThread::start : BaseThreadInitThunk : RtlUserThreadStart : [Isolinee_1953.zip](https://github.com/qgis/QGIS/files/3452397/Isolinee_1953.zip)
process
tin interpolation crash tin interpolation crash with this contour line crash id stack trace mathutils lineintersection dualedgetriangulation insertforcedsegment dualedgetriangulation addline qgstininterpolator insertdata qgstininterpolator initialize qgstininterpolator interpolatepoint pyinit analysis qgsgridfilewriter writefile pyinit analysis pymethoddef rawfastcallkeywords pymethoddef rawfastcallkeywords pyeval evalframedefault pyfunction fastcalldict pymethoddef rawfastcalldict pyobject call pyinit sip cplstringlist empty pyinit core qgsprocessingalgorithm runprepared qgsprocessingalgrunnertask run pyinit core qgstask start qthreadpoolprivate reset qthread start basethreadinitthunk rtluserthreadstart
1
19,087
25,134,692,101
IssuesEvent
2022-11-09 17:37:11
dart-lang/linter
https://api.github.com/repos/dart-lang/linter
opened
☂️ [tracking] rename lints
P3 long-term meta process
A host of lints have prescriptive language in the names (e.g., prefixes like `prefer` and `avoid`). Historically, these reflected language in the style guide and seemed to make sense at the time. In retrospect this choice has some significant downsides and there have been a number of conversations about a big rename. This issue is a **tracking issue** for such an effort. The current suggestion for naming is more neutral: > [Lints should be given a] short name using [Dart package naming conventions](https://dart.dev/tools/pub/pubspec#name). Naming is hard but strive to be concise and consistent. Prefer to use the problem as the name, as in the existing lints control_flow_in_finally and empty_catches. Do not start a lint's name with "always", "avoid", or "prefer". Where possible, use existing rules for inspiration and observe the rules of [parallel construction](https://en.wikipedia.org/wiki/Parallelism_(grammar)). (from [writing-lints.md](https://github.com/dart-lang/linter/blob/main/doc/writing-lints.md)). The goal is to update all lints to this new guidance. Concretely, I propose we - [ ] identify lints that should be renamed - [ ] settle on an _aliasing approach_ so we can safely **_deprecate_** unfavored names - [ ] ensure that `dart fix` can perform the renames I've got some thoughts on all 3 and will fill in some more details here but any early (and ongoing feedback) welcome!
1.0
☂️ [tracking] rename lints - A host of lints have prescriptive language in the names (e.g., prefixes like `prefer` and `avoid`). Historically, these reflected language in the style guide and seemed to make sense at the time. In retrospect this choice has some significant downsides and there have been a number of conversations about a big rename. This issue is a **tracking issue** for such an effort. The current suggestion for naming is more neutral: > [Lints should be given a] short name using [Dart package naming conventions](https://dart.dev/tools/pub/pubspec#name). Naming is hard but strive to be concise and consistent. Prefer to use the problem as the name, as in the existing lints control_flow_in_finally and empty_catches. Do not start a lint's name with "always", "avoid", or "prefer". Where possible, use existing rules for inspiration and observe the rules of [parallel construction](https://en.wikipedia.org/wiki/Parallelism_(grammar)). (from [writing-lints.md](https://github.com/dart-lang/linter/blob/main/doc/writing-lints.md)). The goal is to update all lints to this new guidance. Concretely, I propose we - [ ] identify lints that should be renamed - [ ] settle on an _aliasing approach_ so we can safely **_deprecate_** unfavored names - [ ] ensure that `dart fix` can perform the renames I've got some thoughts on all 3 and will fill in some more details here but any early (and ongoing feedback) welcome!
process
☂️ rename lints a host of lints have prescriptive language in the names e g prefixes like prefer and avoid historically these reflected language in the style guide and seemed to make sense at the time in retrospect this choice has some significant downsides and there have been a number of conversations about a big rename this issue is a tracking issue for such an effort the current suggestion for naming is more neutral short name using naming is hard but strive to be concise and consistent prefer to use the problem as the name as in the existing lints control flow in finally and empty catches do not start a lint s name with always avoid or prefer where possible use existing rules for inspiration and observe the rules of from the goal is to update all lints to this new guidance concretely i propose we identify lints that should be renamed settle on an aliasing approach so we can safely deprecate unfavored names ensure that dart fix can perform the renames i ve got some thoughts on all and will fill in some more details here but any early and ongoing feedback welcome
1
10,708
13,504,120,121
IssuesEvent
2020-09-13 16:32:58
openslide/openslide
https://api.github.com/repos/openslide/openslide
opened
Add code of conduct
development-process enhancement
Codes of conduct are becoming the norm for open source projects. We should select one and apply it.
1.0
Add code of conduct - Codes of conduct are becoming the norm for open source projects. We should select one and apply it.
process
add code of conduct codes of conduct are becoming the norm for open source projects we should select one and apply it
1
6,902
10,054,231,294
IssuesEvent
2019-07-21 23:50:07
ION28/BLUESPAWN
https://api.github.com/repos/ION28/BLUESPAWN
closed
Create 1 good example hunt with new class structure
basic easy enhancement good first issue high priority processes
then Jake can model others off this format and follow best practices
1.0
Create 1 good example hunt with new class structure - then Jake can model others off this format and follow best practices
process
create good example hunt with new class structure then jake can model others off this format and follow best practices
1
337,338
24,535,923,088
IssuesEvent
2022-10-11 20:44:33
yugabyte/yb-voyager
https://api.github.com/repos/yugabyte/yb-voyager
opened
[docs] CLI reference page modifications
documentation
Description Review all commands' args list. Slight changes in default and others. https://docs.yugabyte.com/preview/migrate/yb-voyager-cli/#commands
1.0
[docs] CLI reference page modifications - Description Review all commands' args list. Slight changes in default and others. https://docs.yugabyte.com/preview/migrate/yb-voyager-cli/#commands
non_process
cli reference page modifications description review all commands args list slight changes in default and others
0
104,944
13,146,698,025
IssuesEvent
2020-08-08 11:28:19
mapbox/mapbox-navigation-ios
https://api.github.com/repos/mapbox/mapbox-navigation-ios
closed
Destination pin should match style
archived topic: design
We’re currently using the map SDK’s default annotation icon for the destination pin. The red has good contrast with the default navigation style, but maybe we could be using a different color and different shape that more closely matches the style. Down the line, we might also want to make it easier for the developer to customize the annotation to match the annotation they’re using to represent the destination on a preview map or some other map. But matching the style would be a good start. /cc @aparlato @pveugen @bsudekum
1.0
Destination pin should match style - We’re currently using the map SDK’s default annotation icon for the destination pin. The red has good contrast with the default navigation style, but maybe we could be using a different color and different shape that more closely matches the style. Down the line, we might also want to make it easier for the developer to customize the annotation to match the annotation they’re using to represent the destination on a preview map or some other map. But matching the style would be a good start. /cc @aparlato @pveugen @bsudekum
non_process
destination pin should match style we’re currently using the map sdk’s default annotation icon for the destination pin the red has good contrast with the default navigation style but maybe we could be using a different color and different shape that more closely matches the style down the line we might also want to make it easier for the developer to customize the annotation to match the annotation they’re using to represent the destination on a preview map or some other map but matching the style would be a good start cc aparlato pveugen bsudekum
0
279,203
30,702,467,284
IssuesEvent
2023-07-27 01:32:38
nidhi7598/linux-3.0.35_CVE-2018-13405
https://api.github.com/repos/nidhi7598/linux-3.0.35_CVE-2018-13405
closed
CVE-2018-8822 (High) detected in linux-stable-rtv3.8.6 - autoclosed
Mend: dependency security vulnerability
## CVE-2018-8822 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv3.8.6</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-3.0.35_CVE-2018-13405/commit/662fbf6e1ed61fd353add2f52e2dd27e990364c7">662fbf6e1ed61fd353add2f52e2dd27e990364c7</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/ncpfs/ncplib_kernel.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/ncpfs/ncplib_kernel.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/ncpfs/ncplib_kernel.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> Incorrect buffer length handling in the ncp_read_kernel function in fs/ncpfs/ncplib_kernel.c in the Linux kernel through 4.15.11, and in drivers/staging/ncpfs/ncplib_kernel.c in the Linux kernel 4.16-rc through 4.16-rc6, could be exploited by malicious NCPFS servers to crash the kernel or execute code. <p>Publish Date: 2018-03-20 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-8822>CVE-2018-8822</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-8822">https://nvd.nist.gov/vuln/detail/CVE-2018-8822</a></p> <p>Release Date: 2018-03-20</p> <p>Fix Resolution: v4.16-rc6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-8822 (High) detected in linux-stable-rtv3.8.6 - autoclosed - ## CVE-2018-8822 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv3.8.6</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-3.0.35_CVE-2018-13405/commit/662fbf6e1ed61fd353add2f52e2dd27e990364c7">662fbf6e1ed61fd353add2f52e2dd27e990364c7</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/ncpfs/ncplib_kernel.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/ncpfs/ncplib_kernel.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/ncpfs/ncplib_kernel.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> Incorrect buffer length handling in the ncp_read_kernel function in fs/ncpfs/ncplib_kernel.c in the Linux kernel through 4.15.11, and in drivers/staging/ncpfs/ncplib_kernel.c in the Linux kernel 4.16-rc through 4.16-rc6, could be exploited by malicious NCPFS servers to crash the kernel or execute code. <p>Publish Date: 2018-03-20 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-8822>CVE-2018-8822</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-8822">https://nvd.nist.gov/vuln/detail/CVE-2018-8822</a></p> <p>Release Date: 2018-03-20</p> <p>Fix Resolution: v4.16-rc6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in linux stable autoclosed cve high severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files fs ncpfs ncplib kernel c fs ncpfs ncplib kernel c fs ncpfs ncplib kernel c vulnerability details incorrect buffer length handling in the ncp read kernel function in fs ncpfs ncplib kernel c in the linux kernel through and in drivers staging ncpfs ncplib kernel c in the linux kernel rc through could be exploited by malicious ncpfs servers to crash the kernel or execute code publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
50,091
6,060,468,930
IssuesEvent
2017-06-14 01:54:33
ampproject/amphtml
https://api.github.com/repos/ampproject/amphtml
opened
Flaky integration test: extension load order
Related to: Flaky Tests
https://travis-ci.org/ampproject/amphtml/jobs/242641942 Chrome 45.0.2454 (Linux 0.0.0) test extensions loading in multiple orders two extensions, one of extension scripts and v0 in header FAILED Error: Timeout waiting for elements to start loading. at /home/travis/build/ampproject/amphtml/testing/iframe.js:152:15 <- /tmp/1cdfdf8c28006aa7b83f2c72cb85513d.browserify:55167:16
1.0
Flaky integration test: extension load order - https://travis-ci.org/ampproject/amphtml/jobs/242641942 Chrome 45.0.2454 (Linux 0.0.0) test extensions loading in multiple orders two extensions, one of extension scripts and v0 in header FAILED Error: Timeout waiting for elements to start loading. at /home/travis/build/ampproject/amphtml/testing/iframe.js:152:15 <- /tmp/1cdfdf8c28006aa7b83f2c72cb85513d.browserify:55167:16
non_process
flaky integration test extension load order chrome linux test extensions loading in multiple orders two extensions one of extension scripts and in header failed error timeout waiting for elements to start loading at home travis build ampproject amphtml testing iframe js tmp browserify
0
813
3,287,761,805
IssuesEvent
2015-10-29 12:10:31
nodejs/node
https://api.github.com/repos/nodejs/node
closed
child_process: fork() from -e does not exit
child_process
From [here](https://github.com/nodejs/node/pull/3516#discussion_r43278600): ``` $ out/Release/node -e \ 'require("child_process").fork("test/fixtures/empty.js").on("exit", console.log)' # doesn't print, doesn't exit ``` `parallel/test-child-process-fork3` does the same thing from a file and works fine. Dropping the 'exit' listener doesn't make a difference. @Trott points out that this goes back to at least v0.8.28.
1.0
child_process: fork() from -e does not exit - From [here](https://github.com/nodejs/node/pull/3516#discussion_r43278600): ``` $ out/Release/node -e \ 'require("child_process").fork("test/fixtures/empty.js").on("exit", console.log)' # doesn't print, doesn't exit ``` `parallel/test-child-process-fork3` does the same thing from a file and works fine. Dropping the 'exit' listener doesn't make a difference. @Trott points out that this goes back to at least v0.8.28.
process
child process fork from e does not exit from out release node e require child process fork test fixtures empty js on exit console log doesn t print doesn t exit parallel test child process does the same thing from a file and works fine dropping the exit listener doesn t make a difference trott points out that this goes back to at least
1
875
3,332,583,476
IssuesEvent
2015-11-11 20:48:31
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
closed
possible NTR: catalytic spliceosome grouping term
auto-migrated RNA processes SGD
Hi, I was wondering about adding a grouping term for a subset of the spliceosomal complex terms. The Ohi et al. 2002 paper I'm reading \(citation and PMID below\) describes isolation of complexes complexed with either TAP-tagged versions of Sp Cdc5 or Sc CEF1, PRP19, SNT309, PRP46/CWC1, and CWC2. They get large complexes containing U2, U5, and U6 snRNAs. But they don't do anything that allows me to distinguish between catalytic step 1 or catalytic step 2 spliceosomes, but it's clearly more specific than just "spliceosomal complex", which is the term I would have to use currently. So, I was thinking that it might be nice to have a grouping term for "catalytic spliceosome". The def could be something like this: catalytic spliceosome: A spliceosomal complex that contains three snRNPs, including U5, as well as numerous proteins in addition to those found in the associated snRNPs, bound to a substrate RNA and then if that's OK, we should also add the specific terms for the U2-type and the U12-type Does this sound like a reasonable idea? -Karen Ohi MD, Link AJ, Ren L, Jennings JL, McDonald WH, Gould KL \(2002\) Proteomics analysis reveals stable multiprotein complexes in both fission and budding yeasts containing Myb-related Cdc5p/Cef1p, novel pre-mRNA splicing factors, and snRNAs. Mol Cell Biol 22\(7\):2011-24 PMID:11884590 Reported by: krchristie Original Ticket: [geneontology/ontology-requests/6610](https://sourceforge.net/p/geneontology/ontology-requests/6610)
1.0
possible NTR: catalytic spliceosome grouping term - Hi, I was wondering about adding a grouping term for a subset of the spliceosomal complex terms. The Ohi et al. 2002 paper I'm reading \(citation and PMID below\) describes isolation of complexes complexed with either TAP-tagged versions of Sp Cdc5 or Sc CEF1, PRP19, SNT309, PRP46/CWC1, and CWC2. They get large complexes containing U2, U5, and U6 snRNAs. But they don't do anything that allows me to distinguish between catalytic step 1 or catalytic step 2 spliceosomes, but it's clearly more specific than just "spliceosomal complex", which is the term I would have to use currently. So, I was thinking that it might be nice to have a grouping term for "catalytic spliceosome". The def could be something like this: catalytic spliceosome: A spliceosomal complex that contains three snRNPs, including U5, as well as numerous proteins in addition to those found in the associated snRNPs, bound to a substrate RNA and then if that's OK, we should also add the specific terms for the U2-type and the U12-type Does this sound like a reasonable idea? -Karen Ohi MD, Link AJ, Ren L, Jennings JL, McDonald WH, Gould KL \(2002\) Proteomics analysis reveals stable multiprotein complexes in both fission and budding yeasts containing Myb-related Cdc5p/Cef1p, novel pre-mRNA splicing factors, and snRNAs. Mol Cell Biol 22\(7\):2011-24 PMID:11884590 Reported by: krchristie Original Ticket: [geneontology/ontology-requests/6610](https://sourceforge.net/p/geneontology/ontology-requests/6610)
process
possible ntr catalytic spliceosome grouping term hi i was wondering about adding a grouping term for a subset of the spliceosomal complex terms the ohi et al paper i m reading citation and pmid below describes isolation of complexes complexed with either tap tagged versions of sp or sc and they get large complexes containing and snrnas but they don t do anything that allows me to distinguish between catalytic step or catalytic step spliceosomes but it s clearly more specific than just spliceosomal complex which is the term i would have to use currently so i was thinking that it might be nice to have a grouping term for catalytic spliceosome the def could be something like this catalytic spliceosome a spliceosomal complex that contains three snrnps including as well as numerous proteins in addition to those found in the associated snrnps bound to a substrate rna and then if that s ok we should also add the specific terms for the type and the type does this sound like a reasonable idea karen ohi md link aj ren l jennings jl mcdonald wh gould kl proteomics analysis reveals stable multiprotein complexes in both fission and budding yeasts containing myb related novel pre mrna splicing factors and snrnas mol cell biol pmid reported by krchristie original ticket
1
117,155
17,439,301,515
IssuesEvent
2021-08-05 01:00:31
lukebroganws/Umbraco-CMS
https://api.github.com/repos/lukebroganws/Umbraco-CMS
opened
CVE-2021-32804 (High) detected in tar-4.4.13.tgz, tar-4.4.8.tgz
security vulnerability
## CVE-2021-32804 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>tar-4.4.13.tgz</b>, <b>tar-4.4.8.tgz</b></p></summary> <p> <details><summary><b>tar-4.4.13.tgz</b></p></summary> <p>tar for node</p> <p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-4.4.13.tgz">https://registry.npmjs.org/tar/-/tar-4.4.13.tgz</a></p> <p>Path to dependency file: Umbraco-CMS/src/Umbraco.Web.UI.Client/package.json</p> <p>Path to vulnerable library: Umbraco-CMS/src/Umbraco.Web.UI.Client/node_modules/npm/node_modules/tar/package.json</p> <p> Dependency Hierarchy: - npm-6.14.9.tgz (Root Library) - :x: **tar-4.4.13.tgz** (Vulnerable Library) </details> <details><summary><b>tar-4.4.8.tgz</b></p></summary> <p>tar for node</p> <p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-4.4.8.tgz">https://registry.npmjs.org/tar/-/tar-4.4.8.tgz</a></p> <p> Dependency Hierarchy: - gulp-watch-5.0.1.tgz (Root Library) - chokidar-2.1.8.tgz - fsevents-1.2.9.tgz - node-pre-gyp-0.12.0.tgz - :x: **tar-4.4.8.tgz** (Vulnerable Library) </details> <p>Found in base branch: <b>v8/contrib</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The npm package "tar" (aka node-tar) before versions 6.1.1, 5.0.6, 4.4.14, and 3.3.2 has a arbitrary File Creation/Overwrite vulnerability due to insufficient absolute path sanitization. node-tar aims to prevent extraction of absolute file paths by turning absolute paths into relative paths when the `preservePaths` flag is not set to `true`. This is achieved by stripping the absolute path root from any absolute file paths contained in a tar file. For example `/home/user/.bashrc` would turn into `home/user/.bashrc`. This logic was insufficient when file paths contained repeated path roots such as `////home/user/.bashrc`. `node-tar` would only strip a single path root from such paths. When given an absolute file path with repeating path roots, the resulting path (e.g. `///home/user/.bashrc`) would still resolve to an absolute path, thus allowing arbitrary file creation and overwrite. This issue was addressed in releases 3.2.2, 4.4.14, 5.0.6 and 6.1.1. Users may work around this vulnerability without upgrading by creating a custom `onentry` method which sanitizes the `entry.path` or a `filter` method which removes entries with absolute paths. See referenced GitHub Advisory for details. Be aware of CVE-2021-32803 which fixes a similar bug in later versions of tar. <p>Publish Date: 2021-08-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32804>CVE-2021-32804</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.2</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: N/A - Attack Complexity: N/A - Privileges Required: N/A - User Interaction: N/A - Scope: N/A - Impact Metrics: - Confidentiality Impact: N/A - Integrity Impact: N/A - Availability Impact: N/A </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/npm/node-tar/security/advisories/GHSA-3jfq-g458-7qm9">https://github.com/npm/node-tar/security/advisories/GHSA-3jfq-g458-7qm9</a></p> <p>Release Date: 2021-08-03</p> <p>Fix Resolution: tar - 3.2.2, 4.4.14, 5.0.6, 6.1.1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"tar","packageVersion":"4.4.13","packageFilePaths":["/src/Umbraco.Web.UI.Client/package.json"],"isTransitiveDependency":true,"dependencyTree":"npm:6.14.9;tar:4.4.13","isMinimumFixVersionAvailable":true,"minimumFixVersion":"tar - 3.2.2, 4.4.14, 5.0.6, 6.1.1"},{"packageType":"javascript/Node.js","packageName":"tar","packageVersion":"4.4.8","packageFilePaths":[],"isTransitiveDependency":true,"dependencyTree":"gulp-watch:5.0.1;chokidar:2.1.8;fsevents:1.2.9;node-pre-gyp:0.12.0;tar:4.4.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"tar - 3.2.2, 4.4.14, 5.0.6, 6.1.1"}],"baseBranches":["v8/contrib"],"vulnerabilityIdentifier":"CVE-2021-32804","vulnerabilityDetails":"The npm package \"tar\" (aka node-tar) before versions 6.1.1, 5.0.6, 4.4.14, and 3.3.2 has a arbitrary File Creation/Overwrite vulnerability due to insufficient absolute path sanitization. node-tar aims to prevent extraction of absolute file paths by turning absolute paths into relative paths when the `preservePaths` flag is not set to `true`. This is achieved by stripping the absolute path root from any absolute file paths contained in a tar file. For example `/home/user/.bashrc` would turn into `home/user/.bashrc`. This logic was insufficient when file paths contained repeated path roots such as `////home/user/.bashrc`. `node-tar` would only strip a single path root from such paths. When given an absolute file path with repeating path roots, the resulting path (e.g. `///home/user/.bashrc`) would still resolve to an absolute path, thus allowing arbitrary file creation and overwrite. This issue was addressed in releases 3.2.2, 4.4.14, 5.0.6 and 6.1.1. Users may work around this vulnerability without upgrading by creating a custom `onentry` method which sanitizes the `entry.path` or a `filter` method which removes entries with absolute paths. See referenced GitHub Advisory for details. Be aware of CVE-2021-32803 which fixes a similar bug in later versions of tar.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32804","cvss3Severity":"high","cvss3Score":"8.2","cvss3Metrics":{"A":"N/A","AC":"N/A","PR":"N/A","S":"N/A","C":"N/A","UI":"N/A","AV":"N/A","I":"N/A"},"extraData":{}}</REMEDIATE> -->
True
CVE-2021-32804 (High) detected in tar-4.4.13.tgz, tar-4.4.8.tgz - ## CVE-2021-32804 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>tar-4.4.13.tgz</b>, <b>tar-4.4.8.tgz</b></p></summary> <p> <details><summary><b>tar-4.4.13.tgz</b></p></summary> <p>tar for node</p> <p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-4.4.13.tgz">https://registry.npmjs.org/tar/-/tar-4.4.13.tgz</a></p> <p>Path to dependency file: Umbraco-CMS/src/Umbraco.Web.UI.Client/package.json</p> <p>Path to vulnerable library: Umbraco-CMS/src/Umbraco.Web.UI.Client/node_modules/npm/node_modules/tar/package.json</p> <p> Dependency Hierarchy: - npm-6.14.9.tgz (Root Library) - :x: **tar-4.4.13.tgz** (Vulnerable Library) </details> <details><summary><b>tar-4.4.8.tgz</b></p></summary> <p>tar for node</p> <p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-4.4.8.tgz">https://registry.npmjs.org/tar/-/tar-4.4.8.tgz</a></p> <p> Dependency Hierarchy: - gulp-watch-5.0.1.tgz (Root Library) - chokidar-2.1.8.tgz - fsevents-1.2.9.tgz - node-pre-gyp-0.12.0.tgz - :x: **tar-4.4.8.tgz** (Vulnerable Library) </details> <p>Found in base branch: <b>v8/contrib</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The npm package "tar" (aka node-tar) before versions 6.1.1, 5.0.6, 4.4.14, and 3.3.2 has a arbitrary File Creation/Overwrite vulnerability due to insufficient absolute path sanitization. node-tar aims to prevent extraction of absolute file paths by turning absolute paths into relative paths when the `preservePaths` flag is not set to `true`. This is achieved by stripping the absolute path root from any absolute file paths contained in a tar file. For example `/home/user/.bashrc` would turn into `home/user/.bashrc`. This logic was insufficient when file paths contained repeated path roots such as `////home/user/.bashrc`. `node-tar` would only strip a single path root from such paths. When given an absolute file path with repeating path roots, the resulting path (e.g. `///home/user/.bashrc`) would still resolve to an absolute path, thus allowing arbitrary file creation and overwrite. This issue was addressed in releases 3.2.2, 4.4.14, 5.0.6 and 6.1.1. Users may work around this vulnerability without upgrading by creating a custom `onentry` method which sanitizes the `entry.path` or a `filter` method which removes entries with absolute paths. See referenced GitHub Advisory for details. Be aware of CVE-2021-32803 which fixes a similar bug in later versions of tar. <p>Publish Date: 2021-08-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32804>CVE-2021-32804</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.2</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: N/A - Attack Complexity: N/A - Privileges Required: N/A - User Interaction: N/A - Scope: N/A - Impact Metrics: - Confidentiality Impact: N/A - Integrity Impact: N/A - Availability Impact: N/A </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/npm/node-tar/security/advisories/GHSA-3jfq-g458-7qm9">https://github.com/npm/node-tar/security/advisories/GHSA-3jfq-g458-7qm9</a></p> <p>Release Date: 2021-08-03</p> <p>Fix Resolution: tar - 3.2.2, 4.4.14, 5.0.6, 6.1.1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"tar","packageVersion":"4.4.13","packageFilePaths":["/src/Umbraco.Web.UI.Client/package.json"],"isTransitiveDependency":true,"dependencyTree":"npm:6.14.9;tar:4.4.13","isMinimumFixVersionAvailable":true,"minimumFixVersion":"tar - 3.2.2, 4.4.14, 5.0.6, 6.1.1"},{"packageType":"javascript/Node.js","packageName":"tar","packageVersion":"4.4.8","packageFilePaths":[],"isTransitiveDependency":true,"dependencyTree":"gulp-watch:5.0.1;chokidar:2.1.8;fsevents:1.2.9;node-pre-gyp:0.12.0;tar:4.4.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"tar - 3.2.2, 4.4.14, 5.0.6, 6.1.1"}],"baseBranches":["v8/contrib"],"vulnerabilityIdentifier":"CVE-2021-32804","vulnerabilityDetails":"The npm package \"tar\" (aka node-tar) before versions 6.1.1, 5.0.6, 4.4.14, and 3.3.2 has a arbitrary File Creation/Overwrite vulnerability due to insufficient absolute path sanitization. node-tar aims to prevent extraction of absolute file paths by turning absolute paths into relative paths when the `preservePaths` flag is not set to `true`. This is achieved by stripping the absolute path root from any absolute file paths contained in a tar file. For example `/home/user/.bashrc` would turn into `home/user/.bashrc`. This logic was insufficient when file paths contained repeated path roots such as `////home/user/.bashrc`. `node-tar` would only strip a single path root from such paths. When given an absolute file path with repeating path roots, the resulting path (e.g. `///home/user/.bashrc`) would still resolve to an absolute path, thus allowing arbitrary file creation and overwrite. This issue was addressed in releases 3.2.2, 4.4.14, 5.0.6 and 6.1.1. Users may work around this vulnerability without upgrading by creating a custom `onentry` method which sanitizes the `entry.path` or a `filter` method which removes entries with absolute paths. See referenced GitHub Advisory for details. Be aware of CVE-2021-32803 which fixes a similar bug in later versions of tar.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32804","cvss3Severity":"high","cvss3Score":"8.2","cvss3Metrics":{"A":"N/A","AC":"N/A","PR":"N/A","S":"N/A","C":"N/A","UI":"N/A","AV":"N/A","I":"N/A"},"extraData":{}}</REMEDIATE> -->
non_process
cve high detected in tar tgz tar tgz cve high severity vulnerability vulnerable libraries tar tgz tar tgz tar tgz tar for node library home page a href path to dependency file umbraco cms src umbraco web ui client package json path to vulnerable library umbraco cms src umbraco web ui client node modules npm node modules tar package json dependency hierarchy npm tgz root library x tar tgz vulnerable library tar tgz tar for node library home page a href dependency hierarchy gulp watch tgz root library chokidar tgz fsevents tgz node pre gyp tgz x tar tgz vulnerable library found in base branch contrib vulnerability details the npm package tar aka node tar before versions and has a arbitrary file creation overwrite vulnerability due to insufficient absolute path sanitization node tar aims to prevent extraction of absolute file paths by turning absolute paths into relative paths when the preservepaths flag is not set to true this is achieved by stripping the absolute path root from any absolute file paths contained in a tar file for example home user bashrc would turn into home user bashrc this logic was insufficient when file paths contained repeated path roots such as home user bashrc node tar would only strip a single path root from such paths when given an absolute file path with repeating path roots the resulting path e g home user bashrc would still resolve to an absolute path thus allowing arbitrary file creation and overwrite this issue was addressed in releases and users may work around this vulnerability without upgrading by creating a custom onentry method which sanitizes the entry path or a filter method which removes entries with absolute paths see referenced github advisory for details be aware of cve which fixes a similar bug in later versions of tar publish date url a href cvss score details base score metrics exploitability metrics attack vector n a attack complexity n a privileges required n a user interaction n a scope n a impact metrics confidentiality impact n a integrity impact n a availability impact n a for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tar isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree npm tar isminimumfixversionavailable true minimumfixversion tar packagetype javascript node js packagename tar packageversion packagefilepaths istransitivedependency true dependencytree gulp watch chokidar fsevents node pre gyp tar isminimumfixversionavailable true minimumfixversion tar basebranches vulnerabilityidentifier cve vulnerabilitydetails the npm package tar aka node tar before versions and has a arbitrary file creation overwrite vulnerability due to insufficient absolute path sanitization node tar aims to prevent extraction of absolute file paths by turning absolute paths into relative paths when the preservepaths flag is not set to true this is achieved by stripping the absolute path root from any absolute file paths contained in a tar file for example home user bashrc would turn into home user bashrc this logic was insufficient when file paths contained repeated path roots such as home user bashrc node tar would only strip a single path root from such paths when given an absolute file path with repeating path roots the resulting path e g home user bashrc would still resolve to an absolute path thus allowing arbitrary file creation and overwrite this issue was addressed in releases and users may work around this vulnerability without upgrading by creating a custom onentry method which sanitizes the entry path or a filter method which removes entries with absolute paths see referenced github advisory for details be aware of cve which fixes a similar bug in later versions of tar vulnerabilityurl
0
366,067
10,817,032,089
IssuesEvent
2019-11-08 08:48:31
goharbor/harbor
https://api.github.com/repos/goharbor/harbor
closed
Error message is not correctly displayed in the UI page
area/ui candidate/1.10 kind/bug priority/medium
Trigger a scan all action, `412` HTTP code returned. The error message is displayed as `{{error}}` <img width="1122" alt="Screen Shot 2019-11-05 at 10 23 32" src="https://user-images.githubusercontent.com/5753287/68173745-e73a5000-ffb6-11e9-9445-aa70ce683d6f.png">
1.0
Error message is not correctly displayed in the UI page - Trigger a scan all action, `412` HTTP code returned. The error message is displayed as `{{error}}` <img width="1122" alt="Screen Shot 2019-11-05 at 10 23 32" src="https://user-images.githubusercontent.com/5753287/68173745-e73a5000-ffb6-11e9-9445-aa70ce683d6f.png">
non_process
error message is not correctly displayed in the ui page trigger a scan all action http code returned the error message is displayed as error img width alt screen shot at src
0
4,556
7,388,551,334
IssuesEvent
2018-03-16 03:23:12
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Sample does not work
assigned-to-author doc-bug functions in-process triaged
Hi, I tried to create the GitHub Webhook sample in C# in the portal and in a second project in VS2017. None of both projects work. Even when I create the project in the portal and try the TEST folder on the right hand side. I get that error response. { "Message": "Could not find a valid configuration for WebHook receiver 'github' and instance 'githubwebhookcsharp1,_master'. The setting must be set to a value between 16 and 128 characters long." } --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 9255b6c5-d435-f89b-d813-cf85c5326041 * Version Independent ID: cbb10e07-56dd-99ba-d48f-97e48386dc52 * [Content](https://docs.microsoft.com/en-us/azure/azure-functions/functions-create-github-webhook-triggered-function) * [Content Source](https://github.com/Microsoft/azure-docs/blob/master/articles/azure-functions/functions-create-github-webhook-triggered-function.md) * Service: functions
1.0
Sample does not work - Hi, I tried to create the GitHub Webhook sample in C# in the portal and in a second project in VS2017. None of both projects work. Even when I create the project in the portal and try the TEST folder on the right hand side. I get that error response. { "Message": "Could not find a valid configuration for WebHook receiver 'github' and instance 'githubwebhookcsharp1,_master'. The setting must be set to a value between 16 and 128 characters long." } --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 9255b6c5-d435-f89b-d813-cf85c5326041 * Version Independent ID: cbb10e07-56dd-99ba-d48f-97e48386dc52 * [Content](https://docs.microsoft.com/en-us/azure/azure-functions/functions-create-github-webhook-triggered-function) * [Content Source](https://github.com/Microsoft/azure-docs/blob/master/articles/azure-functions/functions-create-github-webhook-triggered-function.md) * Service: functions
process
sample does not work hi i tried to create the github webhook sample in c in the portal and in a second project in none of both projects work even when i create the project in the portal and try the test folder on the right hand side i get that error response message could not find a valid configuration for webhook receiver github and instance master the setting must be set to a value between and characters long document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id service functions
1
3,943
6,885,908,551
IssuesEvent
2017-11-21 17:32:37
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
closed
new terms for mRNA 3' cleavage factors? need advice
auto-migrated PRO RNA processes SGD
Hi, The Cellular Component ontology may need fleshing out in the area of mRNA cleavage and polyadenylation factors, but I'm not sure how granular we want to get. The existing terms are: # GO:0005849 mRNA cleavage factor complex [133 gene products] # [is_a relation] GO:0005847 mRNA cleavage and polyadenylation specificity factor complex [104 gene products] # [is_a relation] GO:0005848 mRNA cleavage stimulating factor complex [5 gene products] PMID: 18158581 is a recent review that describes these complexes in both yeast and mammals. There are many more! See Table 1 of the review. Do we want to add terms for all of these individual complexes? I'm happy to work on the yeast ones but just wanted to check whether this is our strategy, or whether we want to annotate most things to the more general term GO:0005849. Thanks, Maria Reported by: mariacostanzo Original Ticket: [geneontology/ontology-requests/8217](https://sourceforge.net/p/geneontology/ontology-requests/8217)
1.0
new terms for mRNA 3' cleavage factors? need advice - Hi, The Cellular Component ontology may need fleshing out in the area of mRNA cleavage and polyadenylation factors, but I'm not sure how granular we want to get. The existing terms are: # GO:0005849 mRNA cleavage factor complex [133 gene products] # [is_a relation] GO:0005847 mRNA cleavage and polyadenylation specificity factor complex [104 gene products] # [is_a relation] GO:0005848 mRNA cleavage stimulating factor complex [5 gene products] PMID: 18158581 is a recent review that describes these complexes in both yeast and mammals. There are many more! See Table 1 of the review. Do we want to add terms for all of these individual complexes? I'm happy to work on the yeast ones but just wanted to check whether this is our strategy, or whether we want to annotate most things to the more general term GO:0005849. Thanks, Maria Reported by: mariacostanzo Original Ticket: [geneontology/ontology-requests/8217](https://sourceforge.net/p/geneontology/ontology-requests/8217)
process
new terms for mrna cleavage factors need advice hi the cellular component ontology may need fleshing out in the area of mrna cleavage and polyadenylation factors but i m not sure how granular we want to get the existing terms are go mrna cleavage factor complex go mrna cleavage and polyadenylation specificity factor complex go mrna cleavage stimulating factor complex pmid is a recent review that describes these complexes in both yeast and mammals there are many more see table of the review do we want to add terms for all of these individual complexes i m happy to work on the yeast ones but just wanted to check whether this is our strategy or whether we want to annotate most things to the more general term go thanks maria reported by mariacostanzo original ticket
1
649
3,114,813,052
IssuesEvent
2015-09-03 11:08:04
e-government-ua/iBP
https://api.github.com/repos/e-government-ua/iBP
closed
Днепропетровск ЦНАП - Копії рішення міської ради про надання дозволу на розроблення проекту відведення земельної ділянки
In process of testing in work test
Тех.карті - https://drive.google.com/file/d/0BydenhvN5xzeQ2lZdGRNYWJRUnk4VXUtTGlsUUpuVU92bElZ/view?usp=sharing Заявление - https://docs.google.com/document/d/16ll4Wer7KnbCdUrdywbZrharMZs4w03mMapTIZ0tYRg/edit
1.0
Днепропетровск ЦНАП - Копії рішення міської ради про надання дозволу на розроблення проекту відведення земельної ділянки - Тех.карті - https://drive.google.com/file/d/0BydenhvN5xzeQ2lZdGRNYWJRUnk4VXUtTGlsUUpuVU92bElZ/view?usp=sharing Заявление - https://docs.google.com/document/d/16ll4Wer7KnbCdUrdywbZrharMZs4w03mMapTIZ0tYRg/edit
process
днепропетровск цнап копії рішення міської ради про надання дозволу на розроблення проекту відведення земельної ділянки тех карті заявление
1
38,042
8,639,913,268
IssuesEvent
2018-11-23 22:39:15
bridgedotnet/Bridge
https://api.github.com/repos/bridgedotnet/Bridge
closed
Comparer<string>.Default not using System.String.compare
defect in-progress
Comparer<string>.Default.Compare(str1, str2) gives a different result the str1CompareTo(str2). The first case calls Bridge.compare, and seems to perform char sorting. The second case calls System.String.compare and seems to correctly sort text. https://deck.net/42fe359ec3524e6024107106ec9c9386 I would expect that Comparer<string>.Default would call System.String.compare as System.String implements IComparer<System.String>.
1.0
Comparer<string>.Default not using System.String.compare - Comparer<string>.Default.Compare(str1, str2) gives a different result the str1CompareTo(str2). The first case calls Bridge.compare, and seems to perform char sorting. The second case calls System.String.compare and seems to correctly sort text. https://deck.net/42fe359ec3524e6024107106ec9c9386 I would expect that Comparer<string>.Default would call System.String.compare as System.String implements IComparer<System.String>.
non_process
comparer default not using system string compare comparer default compare gives a different result the the first case calls bridge compare and seems to perform char sorting the second case calls system string compare and seems to correctly sort text i would expect that comparer default would call system string compare as system string implements icomparer
0
10,776
13,595,912,103
IssuesEvent
2020-09-22 04:36:18
Ghost-chu/QuickShop-Reremake
https://api.github.com/repos/Ghost-chu/QuickShop-Reremake
closed
[BUG] Incompatible with Spartan Anti-Cheat
Bug In Process Priority:Major
**Describe the bug** When trying to setup quick shops Spartan plugin denies my permissions to create a shop. After trying enough, I get kicked for using nuker hacks. **To Reproduce** Steps to reproduce the behavior: Install Spartan Anti Cheat and Quick Shop. Try to setup a shop as a non administrative user. **Expected behavior** Spartan and QuickShops should be compatible. **Screenshots** If applicable, add screenshots to help explain your problem. **Paste link:** Execute command /qs paste, you will get a link contains your server information, paste it under this text. You must create a paste, except plugin completely won't work. If you create failed, you should find a paste file under the plugin/QuickShop folder. - 【https://paste.enginehub.org/wMyEKRTAp】 **Additional context** Please work with Evan over at https://discord.me/spartan to see if he can help you sort this incompatibility.
1.0
[BUG] Incompatible with Spartan Anti-Cheat - **Describe the bug** When trying to setup quick shops Spartan plugin denies my permissions to create a shop. After trying enough, I get kicked for using nuker hacks. **To Reproduce** Steps to reproduce the behavior: Install Spartan Anti Cheat and Quick Shop. Try to setup a shop as a non administrative user. **Expected behavior** Spartan and QuickShops should be compatible. **Screenshots** If applicable, add screenshots to help explain your problem. **Paste link:** Execute command /qs paste, you will get a link contains your server information, paste it under this text. You must create a paste, except plugin completely won't work. If you create failed, you should find a paste file under the plugin/QuickShop folder. - 【https://paste.enginehub.org/wMyEKRTAp】 **Additional context** Please work with Evan over at https://discord.me/spartan to see if he can help you sort this incompatibility.
process
incompatible with spartan anti cheat describe the bug when trying to setup quick shops spartan plugin denies my permissions to create a shop after trying enough i get kicked for using nuker hacks to reproduce steps to reproduce the behavior install spartan anti cheat and quick shop try to setup a shop as a non administrative user expected behavior spartan and quickshops should be compatible screenshots if applicable add screenshots to help explain your problem paste link execute command qs paste you will get a link contains your server information paste it under this text you must create a paste except plugin completely won t work if you create failed you should find a paste file under the plugin quickshop folder 【 additional context please work with evan over at to see if he can help you sort this incompatibility
1
152,645
13,463,071,026
IssuesEvent
2020-09-09 17:00:25
cyberark/conjur-puppet
https://api.github.com/repos/cyberark/conjur-puppet
opened
Update diagrams
component/puppet kind/documentation
## Summary With the code change for v3. we can now update our diagrams that are outdated: https://github.com/cyberark/conjur-puppet/tree/master/diagrams ## AC - [ ] Diagrams [here](https://github.com/cyberark/conjur-puppet/tree/master/diagrams) are updated to reflect the new architesture. ## Additional Information Add any other context about the problem here.
1.0
Update diagrams - ## Summary With the code change for v3. we can now update our diagrams that are outdated: https://github.com/cyberark/conjur-puppet/tree/master/diagrams ## AC - [ ] Diagrams [here](https://github.com/cyberark/conjur-puppet/tree/master/diagrams) are updated to reflect the new architesture. ## Additional Information Add any other context about the problem here.
non_process
update diagrams summary with the code change for we can now update our diagrams that are outdated ac diagrams are updated to reflect the new architesture additional information add any other context about the problem here
0
20,527
27,188,590,185
IssuesEvent
2023-02-19 14:23:22
CATcher-org/WATcher
https://api.github.com/repos/CATcher-org/WATcher
closed
Set up Github Actions for Setup builds & tests
aspect-Process priority.Medium
It will be good to set up Github actions for setup builds & simple tests so that it is easier and safer to maintain out WATcher repository. We can refer to CATcher's Github Actions for this
1.0
Set up Github Actions for Setup builds & tests - It will be good to set up Github actions for setup builds & simple tests so that it is easier and safer to maintain out WATcher repository. We can refer to CATcher's Github Actions for this
process
set up github actions for setup builds tests it will be good to set up github actions for setup builds simple tests so that it is easier and safer to maintain out watcher repository we can refer to catcher s github actions for this
1
80,920
15,604,068,413
IssuesEvent
2021-03-19 03:08:27
intellij-rust/intellij-rust
https://api.github.com/repos/intellij-rust/intellij-rust
closed
Can't detect "use of undeclared crate or module" error, the same code in vscode can be detected.
feature subsystem::code insight
<!-- Hello and thank you for the issue! If you would like to report a bug, we have added some points below that you can fill out. Consider using "Help | Create New Issue" action that fills out all necessary information automatically. Feel free to remove all the irrelevant text to request a new feature. --> ## Environment * **IntelliJ Rust plugin version:0.3.142.3705-203** * **Rust toolchain version:*1.5* * **IDE name and version:Clion 2020.3.2** * **Operating system:MacOS 11.2.2** ## Problem description Can't detect "use of undeclared crate or module" error, the same code in vscode can be detected. ![image](https://user-images.githubusercontent.com/32605477/110341501-5b7e7b00-8065-11eb-824e-7912d9f27222.png) ![image](https://user-images.githubusercontent.com/32605477/110341631-82d54800-8065-11eb-9436-f3f8adfee8c7.png) ## Steps to reproduce 1. git clone https://github.com/SummerVibes/lite-raft.git 2. open the project and open lite-rpc/src/client/client.rs <!-- Please include as much of your codebase as needed to reproduce the error. If the relevant files are large, please provide a link to a public repository or a [Gist](https://gist.github.com/). -->
1.0
Can't detect "use of undeclared crate or module" error, the same code in vscode can be detected. - <!-- Hello and thank you for the issue! If you would like to report a bug, we have added some points below that you can fill out. Consider using "Help | Create New Issue" action that fills out all necessary information automatically. Feel free to remove all the irrelevant text to request a new feature. --> ## Environment * **IntelliJ Rust plugin version:0.3.142.3705-203** * **Rust toolchain version:*1.5* * **IDE name and version:Clion 2020.3.2** * **Operating system:MacOS 11.2.2** ## Problem description Can't detect "use of undeclared crate or module" error, the same code in vscode can be detected. ![image](https://user-images.githubusercontent.com/32605477/110341501-5b7e7b00-8065-11eb-824e-7912d9f27222.png) ![image](https://user-images.githubusercontent.com/32605477/110341631-82d54800-8065-11eb-9436-f3f8adfee8c7.png) ## Steps to reproduce 1. git clone https://github.com/SummerVibes/lite-raft.git 2. open the project and open lite-rpc/src/client/client.rs <!-- Please include as much of your codebase as needed to reproduce the error. If the relevant files are large, please provide a link to a public repository or a [Gist](https://gist.github.com/). -->
non_process
can t detect use of undeclared crate or module error the same code in vscode can be detected hello and thank you for the issue if you would like to report a bug we have added some points below that you can fill out consider using help create new issue action that fills out all necessary information automatically feel free to remove all the irrelevant text to request a new feature environment intellij rust plugin version rust toolchain version ide name and version clion operating system macos problem description can t detect use of undeclared crate or module error the same code in vscode can be detected steps to reproduce git clone open the project and open lite rpc src client client rs please include as much of your codebase as needed to reproduce the error if the relevant files are large please provide a link to a public repository or a
0
73,734
9,694,819,552
IssuesEvent
2019-05-24 20:09:50
OfficeDev/office-js-docs-reference
https://api.github.com/repos/OfficeDev/office-js-docs-reference
closed
Is "complete" defined somewhere?
area: excel bug: documentation
In just reading this page, it's not clear to what qualifies as a "complete" or "partial" match. e.g., Does "complete" mean that the search string must match the entire cell contents? Or is it more akin to a "whole word" requirement, in which case it would be helpful to know which characters constitute word components vs. word separators. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 61b7d244-ee86-561c-853a-b289b8184c01 * Version Independent ID: 208ba922-39f7-21ac-9481-646216edec6e * Content: [Excel.WorksheetSearchCriteria interface - Office Add-ins](https://docs.microsoft.com/en-us/javascript/api/excel/excel.worksheetsearchcriteria?view=office-js#feedback) * Content Source: [docs/docs-ref-autogen/excel/excel.worksheetsearchcriteria.yml](https://github.com/OfficeDev/office-js-docs-reference/blob/live/docs/docs-ref-autogen/excel/excel.worksheetsearchcriteria.yml) * Product: **excel** * Technology: **add-ins** * GitHub Login: @o365devx * Microsoft Alias: **o365devx**
1.0
Is "complete" defined somewhere? - In just reading this page, it's not clear to what qualifies as a "complete" or "partial" match. e.g., Does "complete" mean that the search string must match the entire cell contents? Or is it more akin to a "whole word" requirement, in which case it would be helpful to know which characters constitute word components vs. word separators. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 61b7d244-ee86-561c-853a-b289b8184c01 * Version Independent ID: 208ba922-39f7-21ac-9481-646216edec6e * Content: [Excel.WorksheetSearchCriteria interface - Office Add-ins](https://docs.microsoft.com/en-us/javascript/api/excel/excel.worksheetsearchcriteria?view=office-js#feedback) * Content Source: [docs/docs-ref-autogen/excel/excel.worksheetsearchcriteria.yml](https://github.com/OfficeDev/office-js-docs-reference/blob/live/docs/docs-ref-autogen/excel/excel.worksheetsearchcriteria.yml) * Product: **excel** * Technology: **add-ins** * GitHub Login: @o365devx * Microsoft Alias: **o365devx**
non_process
is complete defined somewhere in just reading this page it s not clear to what qualifies as a complete or partial match e g does complete mean that the search string must match the entire cell contents or is it more akin to a whole word requirement in which case it would be helpful to know which characters constitute word components vs word separators document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product excel technology add ins github login microsoft alias
0
241,584
18,468,301,410
IssuesEvent
2021-10-17 09:30:34
Leofeng10/alpha
https://api.github.com/repos/Leofeng10/alpha
opened
UserGuide improvement
severity.Low type.DocumentationBug
![Screen Shot 2021-10-17 at 5.27.12 PM.png](https://raw.githubusercontent.com/Leofeng10/alpha/main/files/446fa0b3-0929-43cb-b963-a7e735304c62.png) Could put more images into the UserGuide to help readers confirm that they are following the guide correctly <!--session: 1634462458282-2e133861-111a-43a6-95dc-4c347a8ed8e6--> <!--Version: Web v3.4.1-->
1.0
UserGuide improvement - ![Screen Shot 2021-10-17 at 5.27.12 PM.png](https://raw.githubusercontent.com/Leofeng10/alpha/main/files/446fa0b3-0929-43cb-b963-a7e735304c62.png) Could put more images into the UserGuide to help readers confirm that they are following the guide correctly <!--session: 1634462458282-2e133861-111a-43a6-95dc-4c347a8ed8e6--> <!--Version: Web v3.4.1-->
non_process
userguide improvement could put more images into the userguide to help readers confirm that they are following the guide correctly
0
493,176
14,227,322,528
IssuesEvent
2020-11-18 01:00:50
Sonarr/Sonarr
https://api.github.com/repos/Sonarr/Sonarr
closed
Synology Indexer no longer syncs with Video Station
bug priority:high
Some time after September 15th the previously working connection with synoindex seems to have broken. There was a minor DSM update that week but so far I can't find anything that seems related in any change logs. I will spend some time this weekend trying to reproduce this issue on the command line and keep this ticket up to date with any info I find.
1.0
Synology Indexer no longer syncs with Video Station - Some time after September 15th the previously working connection with synoindex seems to have broken. There was a minor DSM update that week but so far I can't find anything that seems related in any change logs. I will spend some time this weekend trying to reproduce this issue on the command line and keep this ticket up to date with any info I find.
non_process
synology indexer no longer syncs with video station some time after september the previously working connection with synoindex seems to have broken there was a minor dsm update that week but so far i can t find anything that seems related in any change logs i will spend some time this weekend trying to reproduce this issue on the command line and keep this ticket up to date with any info i find
0
61,887
6,761,101,070
IssuesEvent
2017-10-24 23:33:44
gsfraley/dotrust
https://api.github.com/repos/gsfraley/dotrust
closed
Create .NET artifact for testing and include in Rust build phase
testing wontfix
Needed for #1, #3 This ticket covers creating some sort of .NET artifact or included project that can be used for testing. Ideally, we'll include this in the project itself, as well as source, via the Cargo build phase.
1.0
Create .NET artifact for testing and include in Rust build phase - Needed for #1, #3 This ticket covers creating some sort of .NET artifact or included project that can be used for testing. Ideally, we'll include this in the project itself, as well as source, via the Cargo build phase.
non_process
create net artifact for testing and include in rust build phase needed for this ticket covers creating some sort of net artifact or included project that can be used for testing ideally we ll include this in the project itself as well as source via the cargo build phase
0
746,964
26,052,160,827
IssuesEvent
2022-12-22 19:57:04
google/ground-platform
https://api.github.com/repos/google/ground-platform
closed
[Code health] Set up prettier for web and functions
type: cleanup priority: p2
Currently a prettier config exists in web, but the prettier plugin is not installed. This guide may help: https://khalilstemmler.com/blogs/tooling/prettier/
1.0
[Code health] Set up prettier for web and functions - Currently a prettier config exists in web, but the prettier plugin is not installed. This guide may help: https://khalilstemmler.com/blogs/tooling/prettier/
non_process
set up prettier for web and functions currently a prettier config exists in web but the prettier plugin is not installed this guide may help
0
856
3,316,616,446
IssuesEvent
2015-11-06 17:41:49
pwittchen/kirai
https://api.github.com/repos/pwittchen/kirai
closed
Release 1.1.0
release process
**Initial release notes**: - removed `formatter(...)` method from `Kirai` class - added `format(...)` method accepting implementation of `Formatter` interface to `Kirai` class - added `Syntax` interface and `HtmlSyntax` class implementing this interface - added `put(String key, Object value, Syntax syntax)` method to `Piece` class - set `HtmlSyntax` as default `Syntax` implementation in `Piece` class - removed dependencies to Android SDK - updated project dependencies - applied `Square` code style - updated tests, sample app and code snippets in `README.md` - added gh-pages with JavaDoc **Things to do**: - [x] bump library version - [x] upload archives to Maven Central - [x] close and release artifact on Maven Central - [x] update `CHANGELOG.md` - [x] bump library version in `README.md` - [x] create new GitHub release
1.0
Release 1.1.0 - **Initial release notes**: - removed `formatter(...)` method from `Kirai` class - added `format(...)` method accepting implementation of `Formatter` interface to `Kirai` class - added `Syntax` interface and `HtmlSyntax` class implementing this interface - added `put(String key, Object value, Syntax syntax)` method to `Piece` class - set `HtmlSyntax` as default `Syntax` implementation in `Piece` class - removed dependencies to Android SDK - updated project dependencies - applied `Square` code style - updated tests, sample app and code snippets in `README.md` - added gh-pages with JavaDoc **Things to do**: - [x] bump library version - [x] upload archives to Maven Central - [x] close and release artifact on Maven Central - [x] update `CHANGELOG.md` - [x] bump library version in `README.md` - [x] create new GitHub release
process
release initial release notes removed formatter method from kirai class added format method accepting implementation of formatter interface to kirai class added syntax interface and htmlsyntax class implementing this interface added put string key object value syntax syntax method to piece class set htmlsyntax as default syntax implementation in piece class removed dependencies to android sdk updated project dependencies applied square code style updated tests sample app and code snippets in readme md added gh pages with javadoc things to do bump library version upload archives to maven central close and release artifact on maven central update changelog md bump library version in readme md create new github release
1
25,013
5,117,642,395
IssuesEvent
2017-01-07 19:03:35
golang/go
https://api.github.com/repos/golang/go
closed
website: exclude line numbers when copying from source files
Documentation
by **Dmitry.Gudkov**: <pre>This bug is related to <a href="http://golang.org">http://golang.org</a> website Code samples (e.g. <a href="http://golang.org/doc/articles/wiki/final.go)">http://golang.org/doc/articles/wiki/final.go)</a> contain line numbers which makes copy-pasting of the entire code or some part of it littered with line numbers so additional effort is required to clean up code before compiling. Ubuntu 12.10, Chrome 23.0.1271.97</pre>
1.0
website: exclude line numbers when copying from source files - by **Dmitry.Gudkov**: <pre>This bug is related to <a href="http://golang.org">http://golang.org</a> website Code samples (e.g. <a href="http://golang.org/doc/articles/wiki/final.go)">http://golang.org/doc/articles/wiki/final.go)</a> contain line numbers which makes copy-pasting of the entire code or some part of it littered with line numbers so additional effort is required to clean up code before compiling. Ubuntu 12.10, Chrome 23.0.1271.97</pre>
non_process
website exclude line numbers when copying from source files by dmitry gudkov this bug is related to a href website code samples e g a href contain line numbers which makes copy pasting of the entire code or some part of it littered with line numbers so additional effort is required to clean up code before compiling ubuntu chrome
0
21,404
29,326,705,633
IssuesEvent
2023-05-26 00:17:28
ethereum/EIPs
https://api.github.com/repos/ethereum/EIPs
closed
Errata'd EIPs
w-stale enhancement r-ci r-process r-eips e-consensus
### Proposed Change I propose that with the approval of an original non-errata'd Final EIP's author, a new EIP can be created with the EIP "number" `<original-eip-number>-e<number-erratas>`, where `<original-eip-number>` is the original Final EIP's number, and `<number-erratas>` is a number that starts at `1` and increments by 1 every time an errata EIP for the given original EIP number reaches final, withdrawn, or stagnant. So, instead of modifying EIP-721, a new EIP would be created with the number `721-e1`, referenced as "EIP-721-e1".
1.0
Errata'd EIPs - ### Proposed Change I propose that with the approval of an original non-errata'd Final EIP's author, a new EIP can be created with the EIP "number" `<original-eip-number>-e<number-erratas>`, where `<original-eip-number>` is the original Final EIP's number, and `<number-erratas>` is a number that starts at `1` and increments by 1 every time an errata EIP for the given original EIP number reaches final, withdrawn, or stagnant. So, instead of modifying EIP-721, a new EIP would be created with the number `721-e1`, referenced as "EIP-721-e1".
process
errata d eips proposed change i propose that with the approval of an original non errata d final eip s author a new eip can be created with the eip number e where is the original final eip s number and is a number that starts at and increments by every time an errata eip for the given original eip number reaches final withdrawn or stagnant so instead of modifying eip a new eip would be created with the number referenced as eip
1
3,692
6,718,062,579
IssuesEvent
2017-10-15 07:28:53
nicolashainaux/mathmaker
https://api.github.com/repos/nicolashainaux/mathmaker
closed
Ensure database is changed at each install
database install/upgrade process
So far, `mathmaker` creates a copy of `mathmaker.db-dist` (as `mathmaker.db`) if `mathmaker.db-dist` is newer than `mathmaker.db`. This is not enough, in some cases it does not ensure the replacement of the old version by the newer one. Possible solutions: - Include the version number in a dedicated field in the database replace the time condition to trigger the copy by a check of this version number the script `toolbox/build_db.py` create the copy at the end of its process inclusion of the version number in the database should be made by `setup.py` - Looks better: include the version number in the name of the database, like `mathmaker-0.7.2dev1.db` the script `toolbox/build_db.py` create the copy including version number in the name, at the end of its process `mathmaker` will only check if there's a file with the matching version in the name (AND STILL check the time condition), if not it will create the copy (and delete older versions). - Even better: use a post-install script to create the copy right after install. Check it works also for upgrades.
1.0
Ensure database is changed at each install - So far, `mathmaker` creates a copy of `mathmaker.db-dist` (as `mathmaker.db`) if `mathmaker.db-dist` is newer than `mathmaker.db`. This is not enough, in some cases it does not ensure the replacement of the old version by the newer one. Possible solutions: - Include the version number in a dedicated field in the database replace the time condition to trigger the copy by a check of this version number the script `toolbox/build_db.py` create the copy at the end of its process inclusion of the version number in the database should be made by `setup.py` - Looks better: include the version number in the name of the database, like `mathmaker-0.7.2dev1.db` the script `toolbox/build_db.py` create the copy including version number in the name, at the end of its process `mathmaker` will only check if there's a file with the matching version in the name (AND STILL check the time condition), if not it will create the copy (and delete older versions). - Even better: use a post-install script to create the copy right after install. Check it works also for upgrades.
process
ensure database is changed at each install so far mathmaker creates a copy of mathmaker db dist as mathmaker db if mathmaker db dist is newer than mathmaker db this is not enough in some cases it does not ensure the replacement of the old version by the newer one possible solutions include the version number in a dedicated field in the database replace the time condition to trigger the copy by a check of this version number the script toolbox build db py create the copy at the end of its process inclusion of the version number in the database should be made by setup py looks better include the version number in the name of the database like mathmaker db the script toolbox build db py create the copy including version number in the name at the end of its process mathmaker will only check if there s a file with the matching version in the name and still check the time condition if not it will create the copy and delete older versions even better use a post install script to create the copy right after install check it works also for upgrades
1
322,585
23,915,315,505
IssuesEvent
2022-09-09 12:10:02
DLR-RM/stable-baselines3
https://api.github.com/repos/DLR-RM/stable-baselines3
closed
set_parameters() listed as load_parameters() in documentation.
documentation
### 📚 Documentation The [documentation](https://stable-baselines.readthedocs.io/en/master/modules/td3.html) defines the set_parameters() method for TD3 as load_parameters(). Following the documentation results in ``` AttributeError: 'TD3' object has no attribute 'load_parameters' ``` ``` python import gym from stable_baselines3 import TD3 env = gym.make('Pendulum-v0') parameters = model.get_parameters() model.load_parameters(parameters) ``` Where as the following works properly: ``` python import gym from stable_baselines3 import TD3 env = gym.make('Pendulum-v0') parameters = model.get_parameters() model.set_parameters(parameters) ``` This is using stable-baselines3 version 1.1.0.
1.0
set_parameters() listed as load_parameters() in documentation. - ### 📚 Documentation The [documentation](https://stable-baselines.readthedocs.io/en/master/modules/td3.html) defines the set_parameters() method for TD3 as load_parameters(). Following the documentation results in ``` AttributeError: 'TD3' object has no attribute 'load_parameters' ``` ``` python import gym from stable_baselines3 import TD3 env = gym.make('Pendulum-v0') parameters = model.get_parameters() model.load_parameters(parameters) ``` Where as the following works properly: ``` python import gym from stable_baselines3 import TD3 env = gym.make('Pendulum-v0') parameters = model.get_parameters() model.set_parameters(parameters) ``` This is using stable-baselines3 version 1.1.0.
non_process
set parameters listed as load parameters in documentation 📚 documentation the defines the set parameters method for as load parameters following the documentation results in attributeerror object has no attribute load parameters python import gym from stable import env gym make pendulum parameters model get parameters model load parameters parameters where as the following works properly python import gym from stable import env gym make pendulum parameters model get parameters model set parameters parameters this is using stable version
0
478,037
13,771,769,794
IssuesEvent
2020-10-07 22:45:12
department-of-veterans-affairs/caseflow
https://api.github.com/repos/department-of-veterans-affairs/caseflow
closed
Remove diagnostic code validation
Feature: caseflow-decisions Priority: Medium Product: caseflow-queue Source: Sentry Alert Team: Echo 🐬 Type: Tech-Improvement
## Description The sentry error alert brought to light in #10948 was determined to be caused by the fact that we were receiving diagnostic codes that were not stored in our code. The overall effort to correct this problem is being addressed in several different issues: #14870, #14871, #14872 & #14877. However, since we have determined that this error will prevent judges/attorneys from completing reviews we have decided to remove the validation to remove the blockage. ## Acceptance criteria - [ ] Remove [the validation](https://github.com/department-of-veterans-affairs/caseflow/blob/788ab1dd58a43cdd1055d86250e3892ffdbbb111/app/models/decision_issue.rb#L12-L13) that causes [the error](https://github.com/department-of-veterans-affairs/caseflow/blob/993fb3c00ab2ce06bfbdfef1cdb5a3ae4b4fe7b4/app/models/concerns/issue_updater.rb#L15)
1.0
Remove diagnostic code validation - ## Description The sentry error alert brought to light in #10948 was determined to be caused by the fact that we were receiving diagnostic codes that were not stored in our code. The overall effort to correct this problem is being addressed in several different issues: #14870, #14871, #14872 & #14877. However, since we have determined that this error will prevent judges/attorneys from completing reviews we have decided to remove the validation to remove the blockage. ## Acceptance criteria - [ ] Remove [the validation](https://github.com/department-of-veterans-affairs/caseflow/blob/788ab1dd58a43cdd1055d86250e3892ffdbbb111/app/models/decision_issue.rb#L12-L13) that causes [the error](https://github.com/department-of-veterans-affairs/caseflow/blob/993fb3c00ab2ce06bfbdfef1cdb5a3ae4b4fe7b4/app/models/concerns/issue_updater.rb#L15)
non_process
remove diagnostic code validation description the sentry error alert brought to light in was determined to be caused by the fact that we were receiving diagnostic codes that were not stored in our code the overall effort to correct this problem is being addressed in several different issues however since we have determined that this error will prevent judges attorneys from completing reviews we have decided to remove the validation to remove the blockage acceptance criteria remove that causes
0
14,468
17,573,604,534
IssuesEvent
2021-08-15 06:58:21
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Export to PostgreSQL available connection fails to run when schema name contains a space
Processing Bug
# Problem 1. Create a PostgreSQL schema with spaces ie `845 fix` 2. Navigate to the processing tool and select `Export to PostgreSQL available connection 3. Specify the schema with space and run the tool 4. Error occurs ![error-import](https://user-images.githubusercontent.com/2510900/117858192-ebbb8500-b28d-11eb-8170-7455d3b2c28c.png) # Solution This is an easy fix. The column name should be passed to the algorithm as a single-quoted string like the username and password Tested with 3.18.2-Zürich
1.0
Export to PostgreSQL available connection fails to run when schema name contains a space - # Problem 1. Create a PostgreSQL schema with spaces ie `845 fix` 2. Navigate to the processing tool and select `Export to PostgreSQL available connection 3. Specify the schema with space and run the tool 4. Error occurs ![error-import](https://user-images.githubusercontent.com/2510900/117858192-ebbb8500-b28d-11eb-8170-7455d3b2c28c.png) # Solution This is an easy fix. The column name should be passed to the algorithm as a single-quoted string like the username and password Tested with 3.18.2-Zürich
process
export to postgresql available connection fails to run when schema name contains a space problem create a postgresql schema with spaces ie fix navigate to the processing tool and select export to postgresql available connection specify the schema with space and run the tool error occurs solution this is an easy fix the column name should be passed to the algorithm as a single quoted string like the username and password tested with zürich
1
1,169
3,663,742,602
IssuesEvent
2016-02-19 08:13:15
ViDA-NYU/genotet
https://api.github.com/repos/ViDA-NYU/genotet
closed
Change the way of reading TFA files from binary file (.bin) to text file.
data processing
expression.readTFAmat_ reads TFA binary files now. It should be changed to read TFA text file.
1.0
Change the way of reading TFA files from binary file (.bin) to text file. - expression.readTFAmat_ reads TFA binary files now. It should be changed to read TFA text file.
process
change the way of reading tfa files from binary file bin to text file expression readtfamat reads tfa binary files now it should be changed to read tfa text file
1
9,572
12,522,259,754
IssuesEvent
2020-06-03 18:51:05
pwittchen/ReactiveNetwork
https://api.github.com/repos/pwittchen/ReactiveNetwork
opened
Release 4.0.0-rx3
RxJava3.x release process
Release notes: - **migrated library to RxJava3.x** on RxJava3.x branch and released it as `reactivenetwork-rx3` artifact - the `onError` method in `NetworkObservingStrategy` now requires a `Throwable` instead of an `Exception`. - updated project dependencies - update gradle version Things to do: TBD.
1.0
Release 4.0.0-rx3 - Release notes: - **migrated library to RxJava3.x** on RxJava3.x branch and released it as `reactivenetwork-rx3` artifact - the `onError` method in `NetworkObservingStrategy` now requires a `Throwable` instead of an `Exception`. - updated project dependencies - update gradle version Things to do: TBD.
process
release release notes migrated library to x on x branch and released it as reactivenetwork artifact the onerror method in networkobservingstrategy now requires a throwable instead of an exception updated project dependencies update gradle version things to do tbd
1
10,407
13,204,246,574
IssuesEvent
2020-08-14 15:33:27
usgs-makerspace/makerspace-sandbox
https://api.github.com/repos/usgs-makerspace/makerspace-sandbox
opened
Transition Items
Makerspace Processes
I (Aaron) have started to compile a list of tasks/tutorials/demonstrations that might be useful to complete before I move to Internet of Water. Please feel free to add/modify as desired. Also add your name behind the tutorials if you wish to be included. - [ ] Browserstack Tutorial (Ellen, Colleen) - [ ] Makerspace Project Setup (part 1) using the base template (Ellen, Colleen) - [ ] Makerspace Project Setup (part 2) build configuration, Jenkins etc. (Ellen, Colleen) - [ ] AWS S3 hosting (Ellen, Colleen) - [ ] Google Analytics and Google Tag Manager (Ellen, Colleen) - [ ] NPM (package.json) and dependency management/updating (Ellen, Colleen) - [ ] Vue /Webpack Code Splitting/Lazy loading/load ordering (Ellen, Colleen)
1.0
Transition Items - I (Aaron) have started to compile a list of tasks/tutorials/demonstrations that might be useful to complete before I move to Internet of Water. Please feel free to add/modify as desired. Also add your name behind the tutorials if you wish to be included. - [ ] Browserstack Tutorial (Ellen, Colleen) - [ ] Makerspace Project Setup (part 1) using the base template (Ellen, Colleen) - [ ] Makerspace Project Setup (part 2) build configuration, Jenkins etc. (Ellen, Colleen) - [ ] AWS S3 hosting (Ellen, Colleen) - [ ] Google Analytics and Google Tag Manager (Ellen, Colleen) - [ ] NPM (package.json) and dependency management/updating (Ellen, Colleen) - [ ] Vue /Webpack Code Splitting/Lazy loading/load ordering (Ellen, Colleen)
process
transition items i aaron have started to compile a list of tasks tutorials demonstrations that might be useful to complete before i move to internet of water please feel free to add modify as desired also add your name behind the tutorials if you wish to be included browserstack tutorial ellen colleen makerspace project setup part using the base template ellen colleen makerspace project setup part build configuration jenkins etc ellen colleen aws hosting ellen colleen google analytics and google tag manager ellen colleen npm package json and dependency management updating ellen colleen vue webpack code splitting lazy loading load ordering ellen colleen
1
20,038
26,521,822,393
IssuesEvent
2023-01-19 03:55:07
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
[Android] System.Diagnostics.Process LongProcessNamesAreSupported Fails
area-System.Diagnostics.Process untriaged
Failures in Runfo in last 30 days as of 9/23: Day | Run | OS -- | -- | -- 9/22 | Rolling run [26785](https://runfo.azurewebsites.net/view/build/?number=26785) | net7.0-Linux-Release-x64-CoreCLR_checked-(Alpine.314.Amd64.Open)Ubuntu.1804.Amd64.Open net7.0-Linux-Release-x64-CoreCLR_checked-(Alpine.314.Amd64.Open)Ubuntu.1804.Amd64.Open - [Console log](https://helixre107v0xdeko0k025g8.blob.core.windows.net/dotnet-runtime-refs-heads-main-2f9c0a9a463943b991/System.Diagnostics.Process.Tests/1/console.c446c4aa.log?%3Fhelixlogtype%3Dresult) - 9/22: ``` System.Diagnostics.Tests.ProcessTests.LongProcessNamesAreSupported [FAIL] Assert.Contains() Failure Not found: (filter expression) In value: Process[] [System.Diagnostics.Process (helix_docker_work.sh), System.Diagnostics.Process (execute.sh), System.Diagnostics.Process (bash), System.Diagnostics.Process (dotnet), System.Diagnostics.Process (123456789012345)] Stack Trace: /_/src/libraries/System.Diagnostics.Process/tests/ProcessTests.cs(2221,0): at System.Diagnostics.Tests.ProcessTests.LongProcessNamesAreSupported() at System.RuntimeMethodHandle.InvokeMethod(Object target, Void** arguments, Signature sig, Boolean isConstructor) /_/src/libraries/System.Private.CoreLib/src/System/Reflection/MethodInvoker.cs(69,0): at System.Reflection.MethodInvoker.Invoke(Object obj, IntPtr* args, BindingFlags invokeAttr) ``` Android Arm and Arm64 ``` System.Diagnostics.Tests.ProcessTests.LongProcessNamesAreSupported Assert.Contains() Failure Not found: (filter expression) In value: Process[] [System.Diagnostics.Process (s.Process.Tests), System.Diagnostics.Process (sleep) at System.Diagnostics.Tests.ProcessTests.LongProcessNamesAreSupported() in /_/src/libraries/System.Diagnostics.Process/tests/ProcessTests.cs:line 2128 at System.Reflection.RuntimeMethodInfo.Invoke(Object obj, BindingFlags invokeAttr, Binder binder, Object[] parameters, CultureInfo culture) in System.Private.CoreLib.dll:token 0x6004a27+0x6a
1.0
[Android] System.Diagnostics.Process LongProcessNamesAreSupported Fails - Failures in Runfo in last 30 days as of 9/23: Day | Run | OS -- | -- | -- 9/22 | Rolling run [26785](https://runfo.azurewebsites.net/view/build/?number=26785) | net7.0-Linux-Release-x64-CoreCLR_checked-(Alpine.314.Amd64.Open)Ubuntu.1804.Amd64.Open net7.0-Linux-Release-x64-CoreCLR_checked-(Alpine.314.Amd64.Open)Ubuntu.1804.Amd64.Open - [Console log](https://helixre107v0xdeko0k025g8.blob.core.windows.net/dotnet-runtime-refs-heads-main-2f9c0a9a463943b991/System.Diagnostics.Process.Tests/1/console.c446c4aa.log?%3Fhelixlogtype%3Dresult) - 9/22: ``` System.Diagnostics.Tests.ProcessTests.LongProcessNamesAreSupported [FAIL] Assert.Contains() Failure Not found: (filter expression) In value: Process[] [System.Diagnostics.Process (helix_docker_work.sh), System.Diagnostics.Process (execute.sh), System.Diagnostics.Process (bash), System.Diagnostics.Process (dotnet), System.Diagnostics.Process (123456789012345)] Stack Trace: /_/src/libraries/System.Diagnostics.Process/tests/ProcessTests.cs(2221,0): at System.Diagnostics.Tests.ProcessTests.LongProcessNamesAreSupported() at System.RuntimeMethodHandle.InvokeMethod(Object target, Void** arguments, Signature sig, Boolean isConstructor) /_/src/libraries/System.Private.CoreLib/src/System/Reflection/MethodInvoker.cs(69,0): at System.Reflection.MethodInvoker.Invoke(Object obj, IntPtr* args, BindingFlags invokeAttr) ``` Android Arm and Arm64 ``` System.Diagnostics.Tests.ProcessTests.LongProcessNamesAreSupported Assert.Contains() Failure Not found: (filter expression) In value: Process[] [System.Diagnostics.Process (s.Process.Tests), System.Diagnostics.Process (sleep) at System.Diagnostics.Tests.ProcessTests.LongProcessNamesAreSupported() in /_/src/libraries/System.Diagnostics.Process/tests/ProcessTests.cs:line 2128 at System.Reflection.RuntimeMethodInfo.Invoke(Object obj, BindingFlags invokeAttr, Binder binder, Object[] parameters, CultureInfo culture) in System.Private.CoreLib.dll:token 0x6004a27+0x6a
process
system diagnostics process longprocessnamesaresupported fails failures in runfo in last days as of day run os rolling run linux release coreclr checked alpine open ubuntu open linux release coreclr checked alpine open ubuntu open system diagnostics tests processtests longprocessnamesaresupported assert contains failure not found filter expression in value process stack trace src libraries system diagnostics process tests processtests cs at system diagnostics tests processtests longprocessnamesaresupported at system runtimemethodhandle invokemethod object target void arguments signature sig boolean isconstructor src libraries system private corelib src system reflection methodinvoker cs at system reflection methodinvoker invoke object obj intptr args bindingflags invokeattr android arm and system diagnostics tests processtests longprocessnamesaresupported assert contains failure not found filter expression in value process system diagnostics process s process tests system diagnostics process sleep at system diagnostics tests processtests longprocessnamesaresupported in src libraries system diagnostics process tests processtests cs line at system reflection runtimemethodinfo invoke object obj bindingflags invokeattr binder binder object parameters cultureinfo culture in system private corelib dll token
1
28,228
2,700,615,707
IssuesEvent
2015-04-04 10:57:45
palkopatel/pryced
https://api.github.com/repos/palkopatel/pryced
closed
Остановка слежения
auto-migrated Component-Logic Maintainability Milestone-Release1.0 Priority-Medium Type-Task Usability
``` Сделать возможность отключать слежение за книгами и за отдельными магазинами ``` Original issue reported on code.google.com by `palkopa...@gmail.com` on 26 Oct 2011 at 11:24
1.0
Остановка слежения - ``` Сделать возможность отключать слежение за книгами и за отдельными магазинами ``` Original issue reported on code.google.com by `palkopa...@gmail.com` on 26 Oct 2011 at 11:24
non_process
остановка слежения сделать возможность отключать слежение за книгами и за отдельными магазинами original issue reported on code google com by palkopa gmail com on oct at
0
17,606
23,427,753,066
IssuesEvent
2022-08-14 16:45:54
vortexntnu/Vortex-CV
https://api.github.com/repos/vortexntnu/Vortex-CV
closed
Depth Frequency Analysis (Object Detection)
feature moderate priority Pointcloud Processing
**Time estimate:** 15-20 hours **Description of task:** Create a ROS node that maps depth points versus number of pixels (frequency) that a specific depth has in the current view frame - a histogram. This can be used to see if a specific depth has a 'peak' in frequency, which would indicate an object. Find a way to get the indices of the pixels in the areas of interest of the histogram to be able to see the 'peak' in an image format.
1.0
Depth Frequency Analysis (Object Detection) - **Time estimate:** 15-20 hours **Description of task:** Create a ROS node that maps depth points versus number of pixels (frequency) that a specific depth has in the current view frame - a histogram. This can be used to see if a specific depth has a 'peak' in frequency, which would indicate an object. Find a way to get the indices of the pixels in the areas of interest of the histogram to be able to see the 'peak' in an image format.
process
depth frequency analysis object detection time estimate hours description of task create a ros node that maps depth points versus number of pixels frequency that a specific depth has in the current view frame a histogram this can be used to see if a specific depth has a peak in frequency which would indicate an object find a way to get the indices of the pixels in the areas of interest of the histogram to be able to see the peak in an image format
1
18,902
24,840,021,073
IssuesEvent
2022-10-26 12:00:55
TUM-Dev/NavigaTUM
https://api.github.com/repos/TUM-Dev/NavigaTUM
opened
[Entry] [5414.EG.001]: Edit coordinate
entry webform delete-after-processing
Hello, I would like to add this coordinate to the roomfinder: ``` "5414.EG.001": {coords: {lat: 48.26748730805014, lon: 11.668340402841157}},```
1.0
[Entry] [5414.EG.001]: Edit coordinate - Hello, I would like to add this coordinate to the roomfinder: ``` "5414.EG.001": {coords: {lat: 48.26748730805014, lon: 11.668340402841157}},```
process
edit coordinate hello i would like to add this coordinate to the roomfinder eg coords lat lon
1
20,139
26,687,105,911
IssuesEvent
2023-01-26 23:09:31
GoogleCloudPlatform/spring-cloud-gcp
https://api.github.com/repos/GoogleCloudPlatform/spring-cloud-gcp
opened
Investigate flaky unit test in kotlin sample
priority: p3 type: process
I have been observing this flaky failure in `spring-cloud-gcp-kotlin-app-sample` in a number of recent pull requests, opening an issue for it to track and investigate. (Example error log - [unitTests(19)](https://github.com/GoogleCloudPlatform/spring-cloud-gcp/actions/runs/4018331943/jobs/6903823230)) ``` Error: Failed to execute goal org.jetbrains.kotlin:kotlin-maven-plugin:1.7.21:compile (compile) on project spring-cloud-gcp-kotlin-app-sample: Compilation failure Error: java.lang.ExceptionInInitializerError Error: at com.intellij.openapi.util.BuildNumber.currentVersion(BuildNumber.java:297) Error: at com.intellij.ide.plugins.PluginManagerCore.getBuildNumber(PluginManagerCore.java:876) Error: at com.intellij.ide.plugins.PluginManagerCore.lambda$createLoadingResult$16(PluginManagerCore.java:822) Error: at com.intellij.ide.plugins.DescriptorListLoadingContext.getDefaultVersion(DescriptorListLoadingContext.java:145) Error: at com.intellij.ide.plugins.IdeaPluginDescriptorImpl.readExternal(IdeaPluginDescriptorImpl.java:166) Error: at com.intellij.ide.plugins.PluginDescriptorLoader.loadDescriptorFromJar(PluginDescriptorLoader.java:94) Error: at com.intellij.ide.plugins.PluginManagerCore.registerExtensionPointAndExtensions(PluginManagerCore.java:1325) Error: at com.intellij.core.CoreApplicationEnvironment.registerExtensionPointAndExtensions(CoreApplicationEnvironment.java:287) ```
1.0
Investigate flaky unit test in kotlin sample - I have been observing this flaky failure in `spring-cloud-gcp-kotlin-app-sample` in a number of recent pull requests, opening an issue for it to track and investigate. (Example error log - [unitTests(19)](https://github.com/GoogleCloudPlatform/spring-cloud-gcp/actions/runs/4018331943/jobs/6903823230)) ``` Error: Failed to execute goal org.jetbrains.kotlin:kotlin-maven-plugin:1.7.21:compile (compile) on project spring-cloud-gcp-kotlin-app-sample: Compilation failure Error: java.lang.ExceptionInInitializerError Error: at com.intellij.openapi.util.BuildNumber.currentVersion(BuildNumber.java:297) Error: at com.intellij.ide.plugins.PluginManagerCore.getBuildNumber(PluginManagerCore.java:876) Error: at com.intellij.ide.plugins.PluginManagerCore.lambda$createLoadingResult$16(PluginManagerCore.java:822) Error: at com.intellij.ide.plugins.DescriptorListLoadingContext.getDefaultVersion(DescriptorListLoadingContext.java:145) Error: at com.intellij.ide.plugins.IdeaPluginDescriptorImpl.readExternal(IdeaPluginDescriptorImpl.java:166) Error: at com.intellij.ide.plugins.PluginDescriptorLoader.loadDescriptorFromJar(PluginDescriptorLoader.java:94) Error: at com.intellij.ide.plugins.PluginManagerCore.registerExtensionPointAndExtensions(PluginManagerCore.java:1325) Error: at com.intellij.core.CoreApplicationEnvironment.registerExtensionPointAndExtensions(CoreApplicationEnvironment.java:287) ```
process
investigate flaky unit test in kotlin sample i have been observing this flaky failure in spring cloud gcp kotlin app sample in a number of recent pull requests opening an issue for it to track and investigate example error log error failed to execute goal org jetbrains kotlin kotlin maven plugin compile compile on project spring cloud gcp kotlin app sample compilation failure error java lang exceptionininitializererror error at com intellij openapi util buildnumber currentversion buildnumber java error at com intellij ide plugins pluginmanagercore getbuildnumber pluginmanagercore java error at com intellij ide plugins pluginmanagercore lambda createloadingresult pluginmanagercore java error at com intellij ide plugins descriptorlistloadingcontext getdefaultversion descriptorlistloadingcontext java error at com intellij ide plugins ideaplugindescriptorimpl readexternal ideaplugindescriptorimpl java error at com intellij ide plugins plugindescriptorloader loaddescriptorfromjar plugindescriptorloader java error at com intellij ide plugins pluginmanagercore registerextensionpointandextensions pluginmanagercore java error at com intellij core coreapplicationenvironment registerextensionpointandextensions coreapplicationenvironment java
1
17,099
22,614,264,837
IssuesEvent
2022-06-29 20:10:38
keras-team/keras-cv
https://api.github.com/repos/keras-team/keras-cv
closed
Implement RandomResizedCrop layer
contribution-welcome preprocessing
Randomly resized cropping is pretty much a standard augmentation transformation that is used to train vision models. Recently, a team at Facebook also showed why it's often necessary to include this transformation for better generalization [1]. An implementation of this layer in PyTorch is available as [`RandomResizedCrop`](https://pytorch.org/vision/main/generated/torchvision.transforms.RandomResizedCrop.html). **References**: [1] https://arxiv.org/abs/2106.05121
1.0
Implement RandomResizedCrop layer - Randomly resized cropping is pretty much a standard augmentation transformation that is used to train vision models. Recently, a team at Facebook also showed why it's often necessary to include this transformation for better generalization [1]. An implementation of this layer in PyTorch is available as [`RandomResizedCrop`](https://pytorch.org/vision/main/generated/torchvision.transforms.RandomResizedCrop.html). **References**: [1] https://arxiv.org/abs/2106.05121
process
implement randomresizedcrop layer randomly resized cropping is pretty much a standard augmentation transformation that is used to train vision models recently a team at facebook also showed why it s often necessary to include this transformation for better generalization an implementation of this layer in pytorch is available as references
1
4,901
7,782,161,716
IssuesEvent
2018-06-06 04:55:27
neuropoly/spinalcordtoolbox
https://api.github.com/repos/neuropoly/spinalcordtoolbox
closed
UnboundLocalError: local variable 'z_centerline_voxel' referenced before assignment
bug priority:HIGH sct_process_segmentation
I think this issue is related to the centerline not being defined if the flag `-no-angle` is set to 1. Needs further investigations. ~~~ Spinal Cord Toolbox (master/fc72f34d4251de09d3e92071c66b40ac7fe4c92a) Running /Users/julien/code/sct/scripts/sct_process_segmentation.py -i t2s_gmseg_manual.nii.gz -p csa -no-angle 1 -vert 4 -vertfile t1_seg_labeled_reg.nii.gz -ofolder csa_gm Folder csa_gm has been created. Check parameters: .. segmentation file: t2s_gmseg_manual.nii.gz Create temporary folder (/var/folders/6f/wy6ljmx9453cgth2qwv5l1s80000gn/T/sct-180530091520-LylRGO)... Copying input data to tmp folder and convert to nii... sct_convert -i /Users/julien/data/spine_generic/20180509_julien-skyra/t2s/t2s_gmseg_manual.nii.gz -o /var/folders/6f/wy6ljmx9453cgth2qwv5l1s80000gn/T/sct-180530091520-LylRGO/segmentation.nii.gz # in /Users/julien/data/spine_generic/20180509_julien-skyra/t2s Change orientation to RPI... sct_image -i segmentation.nii.gz -setorient RPI -o segmentation_RPI.nii.gz # in /private/var/folders/6f/wy6ljmx9453cgth2qwv5l1s80000gn/T/sct-180530091520-LylRGO Open segmentation volume... Get data dimensions... 448 x 448 x 15 Compute CSA... Smooth CSA across slices... .. No smoothing! Create volume of CSA values... Create volume of angle values... sct_image -i csa_volume_RPI.nii.gz -setorient RPI -o csa_volume_in_initial_orientation.nii.gz # in /private/var/folders/6f/wy6ljmx9453cgth2qwv5l1s80000gn/T/sct-180530091520-LylRGO sct_image -i angle_volume_RPI.nii.gz -setorient RPI -o angle_volume_in_initial_orientation.nii.gz # in /private/var/folders/6f/wy6ljmx9453cgth2qwv5l1s80000gn/T/sct-180530091520-LylRGO Generate output files... WARNING: File csa_gm/csa_image.nii.gz already exists. Deleting it... File created: csa_gm/csa_image.nii.gz WARNING: File csa_gm/angle_image.nii.gz already exists. Deleting it... File created: csa_gm/angle_image.nii.gz Display CSA per slice: z = 0, CSA = 16.000000 mm^2, Angle = 0.000000 deg z = 1, CSA = 17.000000 mm^2, Angle = 0.000000 deg z = 2, CSA = 18.000000 mm^2, Angle = 0.000000 deg z = 3, CSA = 16.750000 mm^2, Angle = 0.000000 deg z = 4, CSA = 15.250000 mm^2, Angle = 0.000000 deg z = 5, CSA = 15.750000 mm^2, Angle = 0.000000 deg z = 6, CSA = 15.500000 mm^2, Angle = 0.000000 deg z = 7, CSA = 16.250000 mm^2, Angle = 0.000000 deg z = 8, CSA = 16.000000 mm^2, Angle = 0.000000 deg z = 9, CSA = 13.750000 mm^2, Angle = 0.000000 deg z = 10, CSA = 14.500000 mm^2, Angle = 0.000000 deg z = 11, CSA = 13.500000 mm^2, Angle = 0.000000 deg z = 12, CSA = 13.000000 mm^2, Angle = 0.000000 deg z = 13, CSA = 14.000000 mm^2, Angle = 0.000000 deg z = 14, CSA = 13.500000 mm^2, Angle = 0.000000 deg Save results in: csa_gm/csa_per_slice.txt Save results in: csa_gm/csa_per_slice.pickle Selected vertebral levels... 4 OK: t1_seg_labeled_reg.nii.gz Traceback (most recent call last): File "/Users/julien/code/sct/scripts/sct_process_segmentation.py", line 1230, in <module> main(sys.argv[1:]) File "/Users/julien/code/sct/scripts/sct_process_segmentation.py", line 242, in main compute_csa(fname_segmentation, output_folder, overwrite, verbose, remove_temp_files, step, smoothing_param, slices, vert_lev, fname_vertebral_labeling, algo_fitting=param.algo_fitting, type_window=param.type_window, window_length=param.window_length, angle_correction=angle_correction, use_phys_coord=use_phys_coord) File "/Users/julien/code/sct/scripts/sct_process_segmentation.py", line 827, in compute_csa slices, vert_levels_list, warning = get_slices_matching_with_vertebral_levels_based_centerline(vert_levels, im_vertebral_labeling.data, z_centerline_voxel) UnboundLocalError: local variable 'z_centerline_voxel' referenced before assignment ~~~
1.0
UnboundLocalError: local variable 'z_centerline_voxel' referenced before assignment - I think this issue is related to the centerline not being defined if the flag `-no-angle` is set to 1. Needs further investigations. ~~~ Spinal Cord Toolbox (master/fc72f34d4251de09d3e92071c66b40ac7fe4c92a) Running /Users/julien/code/sct/scripts/sct_process_segmentation.py -i t2s_gmseg_manual.nii.gz -p csa -no-angle 1 -vert 4 -vertfile t1_seg_labeled_reg.nii.gz -ofolder csa_gm Folder csa_gm has been created. Check parameters: .. segmentation file: t2s_gmseg_manual.nii.gz Create temporary folder (/var/folders/6f/wy6ljmx9453cgth2qwv5l1s80000gn/T/sct-180530091520-LylRGO)... Copying input data to tmp folder and convert to nii... sct_convert -i /Users/julien/data/spine_generic/20180509_julien-skyra/t2s/t2s_gmseg_manual.nii.gz -o /var/folders/6f/wy6ljmx9453cgth2qwv5l1s80000gn/T/sct-180530091520-LylRGO/segmentation.nii.gz # in /Users/julien/data/spine_generic/20180509_julien-skyra/t2s Change orientation to RPI... sct_image -i segmentation.nii.gz -setorient RPI -o segmentation_RPI.nii.gz # in /private/var/folders/6f/wy6ljmx9453cgth2qwv5l1s80000gn/T/sct-180530091520-LylRGO Open segmentation volume... Get data dimensions... 448 x 448 x 15 Compute CSA... Smooth CSA across slices... .. No smoothing! Create volume of CSA values... Create volume of angle values... sct_image -i csa_volume_RPI.nii.gz -setorient RPI -o csa_volume_in_initial_orientation.nii.gz # in /private/var/folders/6f/wy6ljmx9453cgth2qwv5l1s80000gn/T/sct-180530091520-LylRGO sct_image -i angle_volume_RPI.nii.gz -setorient RPI -o angle_volume_in_initial_orientation.nii.gz # in /private/var/folders/6f/wy6ljmx9453cgth2qwv5l1s80000gn/T/sct-180530091520-LylRGO Generate output files... WARNING: File csa_gm/csa_image.nii.gz already exists. Deleting it... File created: csa_gm/csa_image.nii.gz WARNING: File csa_gm/angle_image.nii.gz already exists. Deleting it... File created: csa_gm/angle_image.nii.gz Display CSA per slice: z = 0, CSA = 16.000000 mm^2, Angle = 0.000000 deg z = 1, CSA = 17.000000 mm^2, Angle = 0.000000 deg z = 2, CSA = 18.000000 mm^2, Angle = 0.000000 deg z = 3, CSA = 16.750000 mm^2, Angle = 0.000000 deg z = 4, CSA = 15.250000 mm^2, Angle = 0.000000 deg z = 5, CSA = 15.750000 mm^2, Angle = 0.000000 deg z = 6, CSA = 15.500000 mm^2, Angle = 0.000000 deg z = 7, CSA = 16.250000 mm^2, Angle = 0.000000 deg z = 8, CSA = 16.000000 mm^2, Angle = 0.000000 deg z = 9, CSA = 13.750000 mm^2, Angle = 0.000000 deg z = 10, CSA = 14.500000 mm^2, Angle = 0.000000 deg z = 11, CSA = 13.500000 mm^2, Angle = 0.000000 deg z = 12, CSA = 13.000000 mm^2, Angle = 0.000000 deg z = 13, CSA = 14.000000 mm^2, Angle = 0.000000 deg z = 14, CSA = 13.500000 mm^2, Angle = 0.000000 deg Save results in: csa_gm/csa_per_slice.txt Save results in: csa_gm/csa_per_slice.pickle Selected vertebral levels... 4 OK: t1_seg_labeled_reg.nii.gz Traceback (most recent call last): File "/Users/julien/code/sct/scripts/sct_process_segmentation.py", line 1230, in <module> main(sys.argv[1:]) File "/Users/julien/code/sct/scripts/sct_process_segmentation.py", line 242, in main compute_csa(fname_segmentation, output_folder, overwrite, verbose, remove_temp_files, step, smoothing_param, slices, vert_lev, fname_vertebral_labeling, algo_fitting=param.algo_fitting, type_window=param.type_window, window_length=param.window_length, angle_correction=angle_correction, use_phys_coord=use_phys_coord) File "/Users/julien/code/sct/scripts/sct_process_segmentation.py", line 827, in compute_csa slices, vert_levels_list, warning = get_slices_matching_with_vertebral_levels_based_centerline(vert_levels, im_vertebral_labeling.data, z_centerline_voxel) UnboundLocalError: local variable 'z_centerline_voxel' referenced before assignment ~~~
process
unboundlocalerror local variable z centerline voxel referenced before assignment i think this issue is related to the centerline not being defined if the flag no angle is set to needs further investigations spinal cord toolbox master running users julien code sct scripts sct process segmentation py i gmseg manual nii gz p csa no angle vert vertfile seg labeled reg nii gz ofolder csa gm folder csa gm has been created check parameters segmentation file gmseg manual nii gz create temporary folder var folders t sct lylrgo copying input data to tmp folder and convert to nii sct convert i users julien data spine generic julien skyra gmseg manual nii gz o var folders t sct lylrgo segmentation nii gz in users julien data spine generic julien skyra change orientation to rpi sct image i segmentation nii gz setorient rpi o segmentation rpi nii gz in private var folders t sct lylrgo open segmentation volume get data dimensions x x compute csa smooth csa across slices no smoothing create volume of csa values create volume of angle values sct image i csa volume rpi nii gz setorient rpi o csa volume in initial orientation nii gz in private var folders t sct lylrgo sct image i angle volume rpi nii gz setorient rpi o angle volume in initial orientation nii gz in private var folders t sct lylrgo generate output files warning file csa gm csa image nii gz already exists deleting it file created csa gm csa image nii gz warning file csa gm angle image nii gz already exists deleting it file created csa gm angle image nii gz display csa per slice z csa mm angle deg z csa mm angle deg z csa mm angle deg z csa mm angle deg z csa mm angle deg z csa mm angle deg z csa mm angle deg z csa mm angle deg z csa mm angle deg z csa mm angle deg z csa mm angle deg z csa mm angle deg z csa mm angle deg z csa mm angle deg z csa mm angle deg save results in csa gm csa per slice txt save results in csa gm csa per slice pickle selected vertebral levels ok seg labeled reg nii gz traceback most recent call last file users julien code sct scripts sct process segmentation py line in main sys argv file users julien code sct scripts sct process segmentation py line in main compute csa fname segmentation output folder overwrite verbose remove temp files step smoothing param slices vert lev fname vertebral labeling algo fitting param algo fitting type window param type window window length param window length angle correction angle correction use phys coord use phys coord file users julien code sct scripts sct process segmentation py line in compute csa slices vert levels list warning get slices matching with vertebral levels based centerline vert levels im vertebral labeling data z centerline voxel unboundlocalerror local variable z centerline voxel referenced before assignment
1
429,293
30,034,220,411
IssuesEvent
2023-06-27 11:41:29
intersystems-community/vscode-objectscript
https://api.github.com/repos/intersystems-community/vscode-objectscript
closed
Document VS Code's default not to offer completions automatically when user types into snippet fields
documentation
See https://github.com/intersystems/language-server/issues/65#issuecomment-722532927 and the pointer within it.
1.0
Document VS Code's default not to offer completions automatically when user types into snippet fields - See https://github.com/intersystems/language-server/issues/65#issuecomment-722532927 and the pointer within it.
non_process
document vs code s default not to offer completions automatically when user types into snippet fields see and the pointer within it
0
111,714
9,539,702,596
IssuesEvent
2019-04-30 17:38:01
RocketChat/Rocket.Chat.Electron
https://api.github.com/repos/RocketChat/Rocket.Chat.Electron
closed
Desktop notifications wont work even with new install
feat: notifications stat: needs testing type: bug
<!-- Thanks for opening an issue! A few things to keep in mind: - Before reporting a bug, please try reproducing your issue with the latest version. - Please verify that the bug is related to the Rocket.Chat Desktop app, and NOT the main web app by testing in Chrome/Firefox. - If the issue occurs in the browser, report to github.com/RocketChat/Rocket.Chat instead --> ## My Setup - Operating System: Win 10 - App Version: 2.15.1 and tried all version past 6months - Installation type: <!-- exe, dmg, package manager etc. --> exe <!-- Answer questions by putting x in box, e.g. [x] --> - [x] I have tested with the latest version - [x] I can simulate the issue easily ## Description <!-- If relevant, attach screenshots, and how to reproduce --> I'm a member of one hobby group where timely desktop alerts are very important. Desktop notifications stopped working 6months ago, settings doesn't help, updates doesn't help, different chat channels doesn't make any difference. I don't get @here or @all notifications. Task bar turns to orange but no sound or popup. I have checked windows settings and allowed all notifications. My SSD died and I MADE FRESH WIN 10 INSTALL with latest app versions and updates but still no luck. What can I do? ### Current Behavior <!-- What actually happens? --> No sound or pop-up alert, task bar turns to orange. From user settings, test notification button works and gives me pop-up. Earlier with old versions, this didn't work (if it gives any clue for solution). ### Expected Behavior <!-- What do you think should happen? --> Pop-up alert with sound when @here is mentioned in my channel. Attached is image of settings bar: ![image](https://user-images.githubusercontent.com/49494915/55904417-71186b00-5bd8-11e9-801d-bdeed3c4bd00.png)
1.0
Desktop notifications wont work even with new install - <!-- Thanks for opening an issue! A few things to keep in mind: - Before reporting a bug, please try reproducing your issue with the latest version. - Please verify that the bug is related to the Rocket.Chat Desktop app, and NOT the main web app by testing in Chrome/Firefox. - If the issue occurs in the browser, report to github.com/RocketChat/Rocket.Chat instead --> ## My Setup - Operating System: Win 10 - App Version: 2.15.1 and tried all version past 6months - Installation type: <!-- exe, dmg, package manager etc. --> exe <!-- Answer questions by putting x in box, e.g. [x] --> - [x] I have tested with the latest version - [x] I can simulate the issue easily ## Description <!-- If relevant, attach screenshots, and how to reproduce --> I'm a member of one hobby group where timely desktop alerts are very important. Desktop notifications stopped working 6months ago, settings doesn't help, updates doesn't help, different chat channels doesn't make any difference. I don't get @here or @all notifications. Task bar turns to orange but no sound or popup. I have checked windows settings and allowed all notifications. My SSD died and I MADE FRESH WIN 10 INSTALL with latest app versions and updates but still no luck. What can I do? ### Current Behavior <!-- What actually happens? --> No sound or pop-up alert, task bar turns to orange. From user settings, test notification button works and gives me pop-up. Earlier with old versions, this didn't work (if it gives any clue for solution). ### Expected Behavior <!-- What do you think should happen? --> Pop-up alert with sound when @here is mentioned in my channel. Attached is image of settings bar: ![image](https://user-images.githubusercontent.com/49494915/55904417-71186b00-5bd8-11e9-801d-bdeed3c4bd00.png)
non_process
desktop notifications wont work even with new install thanks for opening an issue a few things to keep in mind before reporting a bug please try reproducing your issue with the latest version please verify that the bug is related to the rocket chat desktop app and not the main web app by testing in chrome firefox if the issue occurs in the browser report to github com rocketchat rocket chat instead my setup operating system win app version and tried all version past installation type exe i have tested with the latest version i can simulate the issue easily description i m a member of one hobby group where timely desktop alerts are very important desktop notifications stopped working ago settings doesn t help updates doesn t help different chat channels doesn t make any difference i don t get here or all notifications task bar turns to orange but no sound or popup i have checked windows settings and allowed all notifications my ssd died and i made fresh win install with latest app versions and updates but still no luck what can i do current behavior no sound or pop up alert task bar turns to orange from user settings test notification button works and gives me pop up earlier with old versions this didn t work if it gives any clue for solution expected behavior pop up alert with sound when here is mentioned in my channel attached is image of settings bar
0
67,630
14,885,266,367
IssuesEvent
2021-01-20 15:31:58
quarkusio/quarkus
https://api.github.com/repos/quarkusio/quarkus
closed
More config options for FormAuthenticationMechanism
area/security kind/enhancement
**Description** The `FormAuthenticationMechanism` can be great to use to perform form based auth, however it lacks some personalisation features, the post location for example is hard coded (/j_security_check), the username/password too ```java public class FormAuthenticationMechanism implements HttpAuthenticationMechanism { private static final Logger log = Logger.getLogger(FormAuthenticationMechanism.class); public static final String DEFAULT_POST_LOCATION = "/j_security_check"; private final String loginPage; private final String errorPage; private final String postLocation = DEFAULT_POST_LOCATION; private final String locationCookie = "quarkus-redirect-location"; private final String landingPage; private final boolean redirectAfterLogin; private final PersistentLoginManager loginManager; public FormAuthenticationMechanism(String loginPage, String errorPage, String landingPage, boolean redirectAfterLogin, PersistentLoginManager loginManager) { this.loginPage = loginPage; this.errorPage = errorPage; this.landingPage = landingPage; this.redirectAfterLogin = redirectAfterLogin; this.loginManager = loginManager; } public Uni<SecurityIdentity> runFormAuth(final RoutingContext exchange, final IdentityProviderManager securityContext) { exchange.request().setExpectMultipart(true); return Uni.createFrom().emitter(new Consumer<UniEmitter<? super SecurityIdentity>>() { @Override public void accept(UniEmitter<? super SecurityIdentity> uniEmitter) { exchange.request().endHandler(new Handler<Void>() { @Override public void handle(Void event) { try { MultiMap res = exchange.request().formAttributes(); final String jUsername = res.get("j_username"); final String jPassword = res.get("j_password"); ``` **Implementation ideas** We can simply add 4 options to the configuration, to allow the developer specify the following : - username/password field names - post location. - locationCookie
True
More config options for FormAuthenticationMechanism - **Description** The `FormAuthenticationMechanism` can be great to use to perform form based auth, however it lacks some personalisation features, the post location for example is hard coded (/j_security_check), the username/password too ```java public class FormAuthenticationMechanism implements HttpAuthenticationMechanism { private static final Logger log = Logger.getLogger(FormAuthenticationMechanism.class); public static final String DEFAULT_POST_LOCATION = "/j_security_check"; private final String loginPage; private final String errorPage; private final String postLocation = DEFAULT_POST_LOCATION; private final String locationCookie = "quarkus-redirect-location"; private final String landingPage; private final boolean redirectAfterLogin; private final PersistentLoginManager loginManager; public FormAuthenticationMechanism(String loginPage, String errorPage, String landingPage, boolean redirectAfterLogin, PersistentLoginManager loginManager) { this.loginPage = loginPage; this.errorPage = errorPage; this.landingPage = landingPage; this.redirectAfterLogin = redirectAfterLogin; this.loginManager = loginManager; } public Uni<SecurityIdentity> runFormAuth(final RoutingContext exchange, final IdentityProviderManager securityContext) { exchange.request().setExpectMultipart(true); return Uni.createFrom().emitter(new Consumer<UniEmitter<? super SecurityIdentity>>() { @Override public void accept(UniEmitter<? super SecurityIdentity> uniEmitter) { exchange.request().endHandler(new Handler<Void>() { @Override public void handle(Void event) { try { MultiMap res = exchange.request().formAttributes(); final String jUsername = res.get("j_username"); final String jPassword = res.get("j_password"); ``` **Implementation ideas** We can simply add 4 options to the configuration, to allow the developer specify the following : - username/password field names - post location. - locationCookie
non_process
more config options for formauthenticationmechanism description the formauthenticationmechanism can be great to use to perform form based auth however it lacks some personalisation features the post location for example is hard coded j security check the username password too java public class formauthenticationmechanism implements httpauthenticationmechanism private static final logger log logger getlogger formauthenticationmechanism class public static final string default post location j security check private final string loginpage private final string errorpage private final string postlocation default post location private final string locationcookie quarkus redirect location private final string landingpage private final boolean redirectafterlogin private final persistentloginmanager loginmanager public formauthenticationmechanism string loginpage string errorpage string landingpage boolean redirectafterlogin persistentloginmanager loginmanager this loginpage loginpage this errorpage errorpage this landingpage landingpage this redirectafterlogin redirectafterlogin this loginmanager loginmanager public uni runformauth final routingcontext exchange final identityprovidermanager securitycontext exchange request setexpectmultipart true return uni createfrom emitter new consumer override public void accept uniemitter uniemitter exchange request endhandler new handler override public void handle void event try multimap res exchange request formattributes final string jusername res get j username final string jpassword res get j password implementation ideas we can simply add options to the configuration to allow the developer specify the following username password field names post location locationcookie
0
239,194
19,828,369,384
IssuesEvent
2022-01-20 09:24:01
bic-org/Facility-Code
https://api.github.com/repos/bic-org/Facility-Code
closed
Find Nearby Facilities by Address or Lat/Long
enhancement help wanted BETA Testing
To assist with finding facilities using the API a new endpoint is being considered to satisfy the following search queries: 1. Finding container facilities where the depot name is not known, but the address or part of the address is known 2. Finding container facilities near a location (street or lat/long) to check if one exists before necessarily requesting a new container facility code We are considering the requirements around this and have the following thoughts for a new endpoint: `/facilities/byAddress` for which the following query strings would be required to enable efficient searching (similar to if you were looking for holiday accommodation for example) * Country Code (ISO Code) * Street With optional query strings of: * City * Zip/Post Code * State The search would perform weighting based on the data provided in each query string value and if a lat/long is provided the distance from that lat/long to nearby facilities could also be returned to assist. The other option being considered is by a point on a map to return nearby facilities, for example : `facilities/byLatLong/{latitude}/{longitude}` with optional query string * searchTerm This will return all facilities near the lat/long in distance order that have the optional search term included. Similar to how you may enter on a phone "find coffee shops near me" Welcome any thoughts on your user cases around this topic.
1.0
Find Nearby Facilities by Address or Lat/Long - To assist with finding facilities using the API a new endpoint is being considered to satisfy the following search queries: 1. Finding container facilities where the depot name is not known, but the address or part of the address is known 2. Finding container facilities near a location (street or lat/long) to check if one exists before necessarily requesting a new container facility code We are considering the requirements around this and have the following thoughts for a new endpoint: `/facilities/byAddress` for which the following query strings would be required to enable efficient searching (similar to if you were looking for holiday accommodation for example) * Country Code (ISO Code) * Street With optional query strings of: * City * Zip/Post Code * State The search would perform weighting based on the data provided in each query string value and if a lat/long is provided the distance from that lat/long to nearby facilities could also be returned to assist. The other option being considered is by a point on a map to return nearby facilities, for example : `facilities/byLatLong/{latitude}/{longitude}` with optional query string * searchTerm This will return all facilities near the lat/long in distance order that have the optional search term included. Similar to how you may enter on a phone "find coffee shops near me" Welcome any thoughts on your user cases around this topic.
non_process
find nearby facilities by address or lat long to assist with finding facilities using the api a new endpoint is being considered to satisfy the following search queries finding container facilities where the depot name is not known but the address or part of the address is known finding container facilities near a location street or lat long to check if one exists before necessarily requesting a new container facility code we are considering the requirements around this and have the following thoughts for a new endpoint facilities byaddress for which the following query strings would be required to enable efficient searching similar to if you were looking for holiday accommodation for example country code iso code street with optional query strings of city zip post code state the search would perform weighting based on the data provided in each query string value and if a lat long is provided the distance from that lat long to nearby facilities could also be returned to assist the other option being considered is by a point on a map to return nearby facilities for example facilities bylatlong latitude longitude with optional query string searchterm this will return all facilities near the lat long in distance order that have the optional search term included similar to how you may enter on a phone find coffee shops near me welcome any thoughts on your user cases around this topic
0
35,115
14,620,231,794
IssuesEvent
2020-12-22 19:20:00
layer5io/meshery.io
https://api.github.com/repos/layer5io/meshery.io
closed
[Traefik Mesh] Adapter and Repo name change
kind/enhancement service-mesh/maesh
**Current State:** The Maesh project underwent a name change and is now referred to as "Traefik Mesh". **Desired State:** The Meshery site needs updated to reflect this project name change. The GitHub repository for Meshery Adapter for Traefik Mesh has also recently been updated. All links to https://github.com/layer5io/meshery-maesh need to be updated to https://github.com/layer5io/meshery-traefik-mesh --- **Contributor Resources** - See the [Contributing to Meshery.io Website](https://github.com/layer5io/meshery.io#contributing-to-the-mesheryio-website) section of the readme.md.
1.0
[Traefik Mesh] Adapter and Repo name change - **Current State:** The Maesh project underwent a name change and is now referred to as "Traefik Mesh". **Desired State:** The Meshery site needs updated to reflect this project name change. The GitHub repository for Meshery Adapter for Traefik Mesh has also recently been updated. All links to https://github.com/layer5io/meshery-maesh need to be updated to https://github.com/layer5io/meshery-traefik-mesh --- **Contributor Resources** - See the [Contributing to Meshery.io Website](https://github.com/layer5io/meshery.io#contributing-to-the-mesheryio-website) section of the readme.md.
non_process
adapter and repo name change current state the maesh project underwent a name change and is now referred to as traefik mesh desired state the meshery site needs updated to reflect this project name change the github repository for meshery adapter for traefik mesh has also recently been updated all links to need to be updated to contributor resources see the section of the readme md
0
4,060
6,993,425,173
IssuesEvent
2017-12-15 11:16:45
hbinderup94/E17-PROJ3-Gruppe-7
https://api.github.com/repos/hbinderup94/E17-PROJ3-Gruppe-7
closed
Ansigtsgenkendelses feedback ift. placering af kamera
In process SPRINT 5
Skal kunne placere kameratet på COREYXY, så kameraet har ansigtet indenfor 10% af centrum på billedet.
1.0
Ansigtsgenkendelses feedback ift. placering af kamera - Skal kunne placere kameratet på COREYXY, så kameraet har ansigtet indenfor 10% af centrum på billedet.
process
ansigtsgenkendelses feedback ift placering af kamera skal kunne placere kameratet på coreyxy så kameraet har ansigtet indenfor af centrum på billedet
1
8,189
11,386,988,012
IssuesEvent
2020-01-29 14:17:13
devssa/onde-codar-em-salvador
https://api.github.com/repos/devssa/onde-codar-em-salvador
opened
ANALISTA DE SISTEMAS na [AVANSYS]
ARCGIS GEOPROCESSAMENTO HELP WANTED JOOMLA SALVADOR WORDPRESS
<!-- ================================================== POR FAVOR, SÓ POSTE SE A VAGA FOR PARA SALVADOR E CIDADES VIZINHAS! Use: "Desenvolvedor Front-end" ao invés de "Front-End Developer" \o/ Exemplo: `[JAVASCRIPT] [MYSQL] [NODE.JS] Desenvolvedor Front-End na [NOME DA EMPRESA]` ================================================== --> ## Local - Salvador ## Requisitos **Obrigatórios:** - Superior completo - Experiência em manutenção, atualização de conteúdos de sistemas e gestão de conteúdos em Geoprocessamentos, ARCGIS, JOOMLA e Wordpress em ambientes similares ## Contratação - a combinar ## Nossa empresa - A Avansys Tecnologia presta efetivamente serviços especializados de desenvolvimento e manutenção de softwares utilizando sua Fabrica de Software CMMI3, Consultoria para todos os tipos de organização, serviço de Service Desk e Outsourcing para suprir todas as necessidades de sua organização. ## Como se candidatar - Por favor envie um email para curriculo@avansys.com.br com seu CV anexado - enviar no assunto: curriculo@avansys.com.br - 2020
1.0
ANALISTA DE SISTEMAS na [AVANSYS] - <!-- ================================================== POR FAVOR, SÓ POSTE SE A VAGA FOR PARA SALVADOR E CIDADES VIZINHAS! Use: "Desenvolvedor Front-end" ao invés de "Front-End Developer" \o/ Exemplo: `[JAVASCRIPT] [MYSQL] [NODE.JS] Desenvolvedor Front-End na [NOME DA EMPRESA]` ================================================== --> ## Local - Salvador ## Requisitos **Obrigatórios:** - Superior completo - Experiência em manutenção, atualização de conteúdos de sistemas e gestão de conteúdos em Geoprocessamentos, ARCGIS, JOOMLA e Wordpress em ambientes similares ## Contratação - a combinar ## Nossa empresa - A Avansys Tecnologia presta efetivamente serviços especializados de desenvolvimento e manutenção de softwares utilizando sua Fabrica de Software CMMI3, Consultoria para todos os tipos de organização, serviço de Service Desk e Outsourcing para suprir todas as necessidades de sua organização. ## Como se candidatar - Por favor envie um email para curriculo@avansys.com.br com seu CV anexado - enviar no assunto: curriculo@avansys.com.br - 2020
process
analista de sistemas na por favor só poste se a vaga for para salvador e cidades vizinhas use desenvolvedor front end ao invés de front end developer o exemplo desenvolvedor front end na local salvador requisitos obrigatórios superior completo experiência em manutenção atualização de conteúdos de sistemas e gestão de conteúdos em geoprocessamentos arcgis joomla e wordpress em ambientes similares contratação a combinar nossa empresa a avansys tecnologia presta efetivamente serviços especializados de desenvolvimento e manutenção de softwares utilizando sua fabrica de software consultoria para todos os tipos de organização serviço de service desk e outsourcing para suprir todas as necessidades de sua organização como se candidatar por favor envie um email para curriculo avansys com br com seu cv anexado enviar no assunto curriculo avansys com br
1
11,960
14,727,645,160
IssuesEvent
2021-01-06 08:51:01
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[PM] Signin > Remove 'Forgot Password?' hyperlink in blank spaces
Bug P2 Participant manager Process: Tested QA UI
Steps: 1. Navigate to login page 2. Mouse over the blank space present next to signin button 3. Observe Refer attached screenshot: ![Signin](https://user-images.githubusercontent.com/60386291/103064232-9bdb6a80-45d9-11eb-8407-c66814b288bb.png)
1.0
[PM] Signin > Remove 'Forgot Password?' hyperlink in blank spaces - Steps: 1. Navigate to login page 2. Mouse over the blank space present next to signin button 3. Observe Refer attached screenshot: ![Signin](https://user-images.githubusercontent.com/60386291/103064232-9bdb6a80-45d9-11eb-8407-c66814b288bb.png)
process
signin remove forgot password hyperlink in blank spaces steps navigate to login page mouse over the blank space present next to signin button observe refer attached screenshot
1
18,170
24,207,798,477
IssuesEvent
2022-09-25 13:34:14
sebastianbergmann/phpunit
https://api.github.com/repos/sebastianbergmann/phpunit
closed
Annotation @runTestsInSeparateProcesses makes false positives
type/bug feature/process-isolation
| Q | A | --------------------| --------------- | PHPUnit version | 7.4.3 | PHP version | 7.2.9 | Installation Method | PHAR https://stackoverflow.com/questions/52965511/phpunit-false-positives-when-running-tests-in-separate-processes ``` /** * @runTestsInSeparateProcesses */ class ProfileTest extends TestCase { public function testFalsePositive() { $this->assertFalse(true); } } ``` In short: thiscode returns all tests passed, when there is annotation @runTestsInSeparateProcesses. Without the annotation, test correctly fails. Checked in PHPUnit v.5, with or without annotation, tests fail.
1.0
Annotation @runTestsInSeparateProcesses makes false positives - | Q | A | --------------------| --------------- | PHPUnit version | 7.4.3 | PHP version | 7.2.9 | Installation Method | PHAR https://stackoverflow.com/questions/52965511/phpunit-false-positives-when-running-tests-in-separate-processes ``` /** * @runTestsInSeparateProcesses */ class ProfileTest extends TestCase { public function testFalsePositive() { $this->assertFalse(true); } } ``` In short: thiscode returns all tests passed, when there is annotation @runTestsInSeparateProcesses. Without the annotation, test correctly fails. Checked in PHPUnit v.5, with or without annotation, tests fail.
process
annotation runtestsinseparateprocesses makes false positives q a phpunit version php version installation method phar runtestsinseparateprocesses class profiletest extends testcase public function testfalsepositive this assertfalse true in short thiscode returns all tests passed when there is annotation runtestsinseparateprocesses without the annotation test correctly fails checked in phpunit v with or without annotation tests fail
1
55,026
3,071,831,261
IssuesEvent
2015-08-19 14:13:56
RobotiumTech/robotium
https://api.github.com/repos/RobotiumTech/robotium
closed
Code coverage for Robotium tests with Emma
bug imported invalid Priority-Medium
_From [feu...@gmail.com](https://code.google.com/u/103174306080771765636/) on March 21, 2011 04:36:02_ Hello, What steps will reproduce the problem? I am trying to integrate Emma on my build system for Robotium test coverage. I allays have this error when : export EMMA_INSTRUMENT=true What is the expected output? What do you see instead? UNEXPECTED TOP-LEVEL EXCEPTION: com.android.dx.cf.code.SimException: local variable type mismatch: attempt to set or access a value of type java.lang.Object using a local variable of type int. This is symptomatic of .class transformation tools that ignore local variable information. at com.android.dx.cf.code.BaseMachine.throwLocalMismatch(BaseMachine.java:537) at com.android.dx.cf.code.Simulator$SimVisitor.visitLocal(Simulator.java:523) at com.android.dx.cf.code.BytecodeArray.parseInstruction(BytecodeArray.java:472) at com.android.dx.cf.code.Simulator.simulate(Simulator.java:99) at com.android.dx.cf.code.Ropper.processBlock(Ropper.java:678) at com.android.dx.cf.code.Ropper.doit(Ropper.java:633) at com.android.dx.cf.code.Ropper.convert(Ropper.java:250) at com.android.dx.dex.cf.CfTranslator.processMethods(CfTranslator.java:252) at com.android.dx.dex.cf.CfTranslator.translate0(CfTranslator.java:131) at com.android.dx.dex.cf.CfTranslator.translate(CfTranslator.java:85) at com.android.dx.command.dexer.Main.processClass(Main.java:299) at com.android.dx.command.dexer.Main.processFileBytes(Main.java:278) at com.android.dx.command.dexer.Main.access$100(Main.java:56) at com.android.dx.command.dexer.Main$1.processFileBytes(Main.java:229) at com.android.dx.cf.direct.ClassPathOpener.processArchive(ClassPathOpener.java:244) at com.android.dx.cf.direct.ClassPathOpener.processOne(ClassPathOpener.java:130) at com.android.dx.cf.direct.ClassPathOpener.process(ClassPathOpener.java:108) at com.android.dx.command.dexer.Main.processOne(Main.java:247) at com.android.dx.command.dexer.Main.processAllFiles(Main.java:183) at com.android.dx.command.dexer.Main.run(Main.java:139) at com.android.dx.command.dexer.Main.main(Main.java:120) at com.android.dx.command.Main.main(Main.java:89) ...at bytecode offset 00000071 locals[0000]: Lcom/jayway/android/robotium/solo/Presser; locals[0001]: I locals[0002]: I locals[0003]: [I locals[0004]: I locals[0005]: [Z stack[top0]: type{java.lang.SecurityException} ...while working on block 0071 ...while working on method pressMenuItem:(II)V ...while processing pressMenuItem (II)V ...while processing com/jayway/android/robotium/solo/Presser.class 1 error; aborting make: *** [out/target/common/obj/APPS/PhoneCallTests_intermediates/classes.dex] Fehler 1 What version of the product are you using? On what operating system? robotium-solo-2.2 Operating system : Ubuntu 10.04 LTS Please provide any additional information below. Please can you help me out of this problem? Thank You. Chindji _Original issue: http://code.google.com/p/robotium/issues/detail?id=89_
1.0
Code coverage for Robotium tests with Emma - _From [feu...@gmail.com](https://code.google.com/u/103174306080771765636/) on March 21, 2011 04:36:02_ Hello, What steps will reproduce the problem? I am trying to integrate Emma on my build system for Robotium test coverage. I allays have this error when : export EMMA_INSTRUMENT=true What is the expected output? What do you see instead? UNEXPECTED TOP-LEVEL EXCEPTION: com.android.dx.cf.code.SimException: local variable type mismatch: attempt to set or access a value of type java.lang.Object using a local variable of type int. This is symptomatic of .class transformation tools that ignore local variable information. at com.android.dx.cf.code.BaseMachine.throwLocalMismatch(BaseMachine.java:537) at com.android.dx.cf.code.Simulator$SimVisitor.visitLocal(Simulator.java:523) at com.android.dx.cf.code.BytecodeArray.parseInstruction(BytecodeArray.java:472) at com.android.dx.cf.code.Simulator.simulate(Simulator.java:99) at com.android.dx.cf.code.Ropper.processBlock(Ropper.java:678) at com.android.dx.cf.code.Ropper.doit(Ropper.java:633) at com.android.dx.cf.code.Ropper.convert(Ropper.java:250) at com.android.dx.dex.cf.CfTranslator.processMethods(CfTranslator.java:252) at com.android.dx.dex.cf.CfTranslator.translate0(CfTranslator.java:131) at com.android.dx.dex.cf.CfTranslator.translate(CfTranslator.java:85) at com.android.dx.command.dexer.Main.processClass(Main.java:299) at com.android.dx.command.dexer.Main.processFileBytes(Main.java:278) at com.android.dx.command.dexer.Main.access$100(Main.java:56) at com.android.dx.command.dexer.Main$1.processFileBytes(Main.java:229) at com.android.dx.cf.direct.ClassPathOpener.processArchive(ClassPathOpener.java:244) at com.android.dx.cf.direct.ClassPathOpener.processOne(ClassPathOpener.java:130) at com.android.dx.cf.direct.ClassPathOpener.process(ClassPathOpener.java:108) at com.android.dx.command.dexer.Main.processOne(Main.java:247) at com.android.dx.command.dexer.Main.processAllFiles(Main.java:183) at com.android.dx.command.dexer.Main.run(Main.java:139) at com.android.dx.command.dexer.Main.main(Main.java:120) at com.android.dx.command.Main.main(Main.java:89) ...at bytecode offset 00000071 locals[0000]: Lcom/jayway/android/robotium/solo/Presser; locals[0001]: I locals[0002]: I locals[0003]: [I locals[0004]: I locals[0005]: [Z stack[top0]: type{java.lang.SecurityException} ...while working on block 0071 ...while working on method pressMenuItem:(II)V ...while processing pressMenuItem (II)V ...while processing com/jayway/android/robotium/solo/Presser.class 1 error; aborting make: *** [out/target/common/obj/APPS/PhoneCallTests_intermediates/classes.dex] Fehler 1 What version of the product are you using? On what operating system? robotium-solo-2.2 Operating system : Ubuntu 10.04 LTS Please provide any additional information below. Please can you help me out of this problem? Thank You. Chindji _Original issue: http://code.google.com/p/robotium/issues/detail?id=89_
non_process
code coverage for robotium tests with emma from on march hello what steps will reproduce the problem i am trying to integrate emma on my build system for robotium test coverage i allays have this error when export emma instrument true what is the expected output what do you see instead unexpected top level exception com android dx cf code simexception local variable type mismatch attempt to set or access a value of type java lang object using a local variable of type int this is symptomatic of class transformation tools that ignore local variable information at com android dx cf code basemachine throwlocalmismatch basemachine java at com android dx cf code simulator simvisitor visitlocal simulator java at com android dx cf code bytecodearray parseinstruction bytecodearray java at com android dx cf code simulator simulate simulator java at com android dx cf code ropper processblock ropper java at com android dx cf code ropper doit ropper java at com android dx cf code ropper convert ropper java at com android dx dex cf cftranslator processmethods cftranslator java at com android dx dex cf cftranslator cftranslator java at com android dx dex cf cftranslator translate cftranslator java at com android dx command dexer main processclass main java at com android dx command dexer main processfilebytes main java at com android dx command dexer main access main java at com android dx command dexer main processfilebytes main java at com android dx cf direct classpathopener processarchive classpathopener java at com android dx cf direct classpathopener processone classpathopener java at com android dx cf direct classpathopener process classpathopener java at com android dx command dexer main processone main java at com android dx command dexer main processallfiles main java at com android dx command dexer main run main java at com android dx command dexer main main main java at com android dx command main main main java at bytecode offset locals lcom jayway android robotium solo presser locals i locals i locals i locals i locals z stack type java lang securityexception while working on block while working on method pressmenuitem ii v while processing pressmenuitem ii v while processing com jayway android robotium solo presser class error aborting make fehler what version of the product are you using on what operating system robotium solo operating system ubuntu lts please provide any additional information below please can you help me out of this problem thank you chindji original issue
0
346,882
31,031,158,739
IssuesEvent
2023-08-10 12:34:07
frozen-signal/Apple_IIe_MMU_IOU
https://api.github.com/repos/frozen-signal/Apple_IIe_MMU_IOU
closed
Fix the test IOU_TB_SCANNER_MUX
IOU tests
the MUX RA has changed. Need to update the test. The test is disabled in the meantime.
1.0
Fix the test IOU_TB_SCANNER_MUX - the MUX RA has changed. Need to update the test. The test is disabled in the meantime.
non_process
fix the test iou tb scanner mux the mux ra has changed need to update the test the test is disabled in the meantime
0
11,306
14,107,663,198
IssuesEvent
2020-11-06 16:37:15
retaildevcrews/ngsa
https://api.github.com/repos/retaildevcrews/ngsa
closed
GamePlan Documentation
GPR Process
Next Steps: - [x] Draft Game Plan document - [x] Setup Game Plan Review - [x] Upload to Artifact Hub
1.0
GamePlan Documentation - Next Steps: - [x] Draft Game Plan document - [x] Setup Game Plan Review - [x] Upload to Artifact Hub
process
gameplan documentation next steps draft game plan document setup game plan review upload to artifact hub
1
19,215
25,350,266,085
IssuesEvent
2022-11-19 17:34:17
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
closed
membrane curvature/membrane bending follow up
low priority PomBase cellular processes MF_in_BP
https://github.com/geneontology/go-ontology/issues/12387 There still seem to be some connections missing: ![membrane bending terms](https://cloud.githubusercontent.com/assets/7359272/23076407/3b6b93be-f538-11e6-8a57-87376da20aba.jpg) - [ ] spore membrane bending pathway -> rename spore membrane bending? - [ ] spore membrane bending pathwa, add membrane bending parent? - [ ] lipid tube assembly appears to be describing some sort of "membrane tubulation"? - [ ] endoplasmic reticulum tubular network assembly appears to be describing some sort of "membrane tubulation"?
1.0
membrane curvature/membrane bending follow up - https://github.com/geneontology/go-ontology/issues/12387 There still seem to be some connections missing: ![membrane bending terms](https://cloud.githubusercontent.com/assets/7359272/23076407/3b6b93be-f538-11e6-8a57-87376da20aba.jpg) - [ ] spore membrane bending pathway -> rename spore membrane bending? - [ ] spore membrane bending pathwa, add membrane bending parent? - [ ] lipid tube assembly appears to be describing some sort of "membrane tubulation"? - [ ] endoplasmic reticulum tubular network assembly appears to be describing some sort of "membrane tubulation"?
process
membrane curvature membrane bending follow up there still seem to be some connections missing spore membrane bending pathway rename spore membrane bending spore membrane bending pathwa add membrane bending parent lipid tube assembly appears to be describing some sort of membrane tubulation endoplasmic reticulum tubular network assembly appears to be describing some sort of membrane tubulation
1