Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 7 112 | repo_url stringlengths 36 141 | action stringclasses 3 values | title stringlengths 1 744 | labels stringlengths 4 574 | body stringlengths 9 211k | index stringclasses 10 values | text_combine stringlengths 96 211k | label stringclasses 2 values | text stringlengths 96 188k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
712,523 | 24,497,979,547 | IssuesEvent | 2022-10-10 10:19:22 | open-telemetry/opentelemetry-js-contrib | https://api.github.com/repos/open-telemetry/opentelemetry-js-contrib | closed | fs.exists with util.promisify causes to unhandled rejection | bug priority:p1 pkg:instrumentation-fs | <!--
Please answer these questions before submitting a bug report.
-->
### What version of OpenTelemetry are you using?
```
"name": "@opentelemetry/api",
"version": "1.2.0"
```
**Latest installed globally.**
Installed dependencies:
npm i --unsafe-perm -g @splunk/otel @opentelemetry/api @opentelemetry/instrumentation-http @opentelemetry/instrumentation-express @opentelemetry/instrumentation-pino @opentelemetry/instrumentation-winston @opentelemetry/instrumentation-nestjs-core @opentelemetry/instrumentation-ioredis @opentelemetry/instrumentation-fs opentelemetry-instrumentation-mongoose @opentelemetry/instrumentation-dns @opentelemetry/instrumentation-amqplib
Note: Also tried to install everything except fs but same behaviour.
### What version of Node are you using?
v16.17.0
### What did you do?
Reproducing is easy:
test.mjs
```mjs
import * as fs from 'fs';
import * as util from 'util';
const result = await util.promisify(fs.exists)('./index.js');
console.log(result);
```
### What did you expect to see?
fs.exists should return 'true' as a resolved value and not as a rejected value.
### What did you see instead?
Executing node with `node -r @splunk/otel/instrument test.mjs` causes to this error:
node:internal/process/esm_loader:97
internalBinding('errors').triggerUncaughtException(
^
true
(Use `node --trace-uncaught ...` to show where the exception was thrown)
Without -r @splunk/otel/instrument it works as expected.
### Additional context
It looks like the promisified fs.exists is handled incorrectly (CB parameters) by opentelemetry and Promise.reject is called with (true) instead of resolve.
| 1.0 | fs.exists with util.promisify causes to unhandled rejection - <!--
Please answer these questions before submitting a bug report.
-->
### What version of OpenTelemetry are you using?
```
"name": "@opentelemetry/api",
"version": "1.2.0"
```
**Latest installed globally.**
Installed dependencies:
npm i --unsafe-perm -g @splunk/otel @opentelemetry/api @opentelemetry/instrumentation-http @opentelemetry/instrumentation-express @opentelemetry/instrumentation-pino @opentelemetry/instrumentation-winston @opentelemetry/instrumentation-nestjs-core @opentelemetry/instrumentation-ioredis @opentelemetry/instrumentation-fs opentelemetry-instrumentation-mongoose @opentelemetry/instrumentation-dns @opentelemetry/instrumentation-amqplib
Note: Also tried to install everything except fs but same behaviour.
### What version of Node are you using?
v16.17.0
### What did you do?
Reproducing is easy:
test.mjs
```mjs
import * as fs from 'fs';
import * as util from 'util';
const result = await util.promisify(fs.exists)('./index.js');
console.log(result);
```
### What did you expect to see?
fs.exists should return 'true' as a resolved value and not as a rejected value.
### What did you see instead?
Executing node with `node -r @splunk/otel/instrument test.mjs` causes to this error:
node:internal/process/esm_loader:97
internalBinding('errors').triggerUncaughtException(
^
true
(Use `node --trace-uncaught ...` to show where the exception was thrown)
Without -r @splunk/otel/instrument it works as expected.
### Additional context
It looks like the promisified fs.exists is handled incorrectly (CB parameters) by opentelemetry and Promise.reject is called with (true) instead of resolve.
| non_process | fs exists with util promisify causes to unhandled rejection please answer these questions before submitting a bug report what version of opentelemetry are you using name opentelemetry api version latest installed globally installed dependencies npm i unsafe perm g splunk otel opentelemetry api opentelemetry instrumentation http opentelemetry instrumentation express opentelemetry instrumentation pino opentelemetry instrumentation winston opentelemetry instrumentation nestjs core opentelemetry instrumentation ioredis opentelemetry instrumentation fs opentelemetry instrumentation mongoose opentelemetry instrumentation dns opentelemetry instrumentation amqplib note also tried to install everything except fs but same behaviour what version of node are you using what did you do reproducing is easy test mjs mjs import as fs from fs import as util from util const result await util promisify fs exists index js console log result what did you expect to see fs exists should return true as a resolved value and not as a rejected value what did you see instead executing node with node r splunk otel instrument test mjs causes to this error node internal process esm loader internalbinding errors triggeruncaughtexception true use node trace uncaught to show where the exception was thrown without r splunk otel instrument it works as expected additional context it looks like the promisified fs exists is handled incorrectly cb parameters by opentelemetry and promise reject is called with true instead of resolve | 0 |
32,176 | 12,093,666,320 | IssuesEvent | 2020-04-19 20:34:12 | AOSC-Dev/aosc-os-abbs | https://api.github.com/repos/AOSC-Dev/aosc-os-abbs | closed | libvirt: multiple security updates | security to-stable | **CVE IDs:** CVE-2019-10132, CVE-2019-10161, CVE-2019-10166, CVE-2019-10167, CVE-2019-10168
**Other security advisory IDs:** RHSA-2019:1264-01
**Descriptions:**
* Restrict user access to virt-admin, virtlogd and virtlockd
* The intended users for these facilities are the root user and the libvirtd service respectively, but these restrictions were not enforced correctly. CVE-2019-10132.
* api: Prevent access to several APIs over read-only connections
* Certain APIs give root-equivalent access to the host, and as such should be limited to privileged users. CVE-2019-10161, CVE-2019-10166, CVE-2019-10167, CVE-2019-10168.
**Patches:** https://security.libvirt.org/2019/0003.html, https://security.libvirt.org/2019/0004.html, https://security.libvirt.org/2019/0005.html, https://security.libvirt.org/2019/0006.html, https://security.libvirt.org/2019/0007.html
**PoC(s):** N/A
**Architectural progress:**
- [x] AMD64 `amd64`
| True | libvirt: multiple security updates - **CVE IDs:** CVE-2019-10132, CVE-2019-10161, CVE-2019-10166, CVE-2019-10167, CVE-2019-10168
**Other security advisory IDs:** RHSA-2019:1264-01
**Descriptions:**
* Restrict user access to virt-admin, virtlogd and virtlockd
* The intended users for these facilities are the root user and the libvirtd service respectively, but these restrictions were not enforced correctly. CVE-2019-10132.
* api: Prevent access to several APIs over read-only connections
* Certain APIs give root-equivalent access to the host, and as such should be limited to privileged users. CVE-2019-10161, CVE-2019-10166, CVE-2019-10167, CVE-2019-10168.
**Patches:** https://security.libvirt.org/2019/0003.html, https://security.libvirt.org/2019/0004.html, https://security.libvirt.org/2019/0005.html, https://security.libvirt.org/2019/0006.html, https://security.libvirt.org/2019/0007.html
**PoC(s):** N/A
**Architectural progress:**
- [x] AMD64 `amd64`
| non_process | libvirt multiple security updates cve ids cve cve cve cve cve other security advisory ids rhsa descriptions restrict user access to virt admin virtlogd and virtlockd the intended users for these facilities are the root user and the libvirtd service respectively but these restrictions were not enforced correctly cve api prevent access to several apis over read only connections certain apis give root equivalent access to the host and as such should be limited to privileged users cve cve cve cve patches poc s n a architectural progress | 0 |
221,502 | 17,354,136,141 | IssuesEvent | 2021-07-29 12:32:24 | elastic/kibana | https://api.github.com/repos/elastic/kibana | closed | [Security Solution] For Trusted Apps, we can see the summary is exactly the same on the tab and flyout too. But it is not consistent Event Filters | QA:Ready for Testing Team: SecuritySolution Team:Onboarding and Lifecycle Mgt bug impact:medium v7.14.0 | **Describe the feature**
For Trusted Apps, we can see the summary is exactly the same on the tab and flyout too. But it is not consistent Event Filters
**Build Details:**
```
VERSION: 7.14.0-BC3
BUILD: 42545
COMMIT: c314921a9893e0b46d9a3958f5520e3d6b1ce7d5
ARTIFACT: https://staging.elastic.co/7.14.0-682a8012/summary-7.14.0.html
```
**Preconditions**
1. Elastic 7.14.0 environment should be deployed.
**Steps to Reproduce**
1.
**Test data**
N/A
**Impacted Test case(s)**
N/A
**Actual Result**
For Trusted Apps, we can see the summary is exactly the same on the tab and flyout too. But it is not consistent Event Filters
**Expected Result**
For Event Filters, the summary should be exactly the same on the tab and flyout too.
**What's Working**
N/A
**What's Not Working**
N/A
**Screenshots**
- Trusted Applications Tab

- Trusted Applications Flyout

- Filter Events Tab

- Filter Events Flyout

**Logs**:
N/A | 1.0 | [Security Solution] For Trusted Apps, we can see the summary is exactly the same on the tab and flyout too. But it is not consistent Event Filters - **Describe the feature**
For Trusted Apps, we can see the summary is exactly the same on the tab and flyout too. But it is not consistent Event Filters
**Build Details:**
```
VERSION: 7.14.0-BC3
BUILD: 42545
COMMIT: c314921a9893e0b46d9a3958f5520e3d6b1ce7d5
ARTIFACT: https://staging.elastic.co/7.14.0-682a8012/summary-7.14.0.html
```
**Preconditions**
1. Elastic 7.14.0 environment should be deployed.
**Steps to Reproduce**
1.
**Test data**
N/A
**Impacted Test case(s)**
N/A
**Actual Result**
For Trusted Apps, we can see the summary is exactly the same on the tab and flyout too. But it is not consistent Event Filters
**Expected Result**
For Event Filters, the summary should be exactly the same on the tab and flyout too.
**What's Working**
N/A
**What's Not Working**
N/A
**Screenshots**
- Trusted Applications Tab

- Trusted Applications Flyout

- Filter Events Tab

- Filter Events Flyout

**Logs**:
N/A | non_process | for trusted apps we can see the summary is exactly the same on the tab and flyout too but it is not consistent event filters describe the feature for trusted apps we can see the summary is exactly the same on the tab and flyout too but it is not consistent event filters build details version build commit artifact preconditions elastic environment should be deployed steps to reproduce test data n a impacted test case s n a actual result for trusted apps we can see the summary is exactly the same on the tab and flyout too but it is not consistent event filters expected result for event filters the summary should be exactly the same on the tab and flyout too what s working n a what s not working n a screenshots trusted applications tab trusted applications flyout filter events tab filter events flyout logs n a | 0 |
10,234 | 12,227,815,820 | IssuesEvent | 2020-05-03 16:46:45 | RKrahl/photo-tools | https://api.github.com/repos/RKrahl/photo-tools | opened | Move away from gexiv2 | compatibility enhancement | We currently use `gexiv2` to read the EXIF metadata from the images. The disadvantage is that `gexiv2` is part of GNOME and depends on the `Exiv2` library which makes it more difficult to install.
Consider alternative packages such as [exif](https://pypi.org/project/exif/). | True | Move away from gexiv2 - We currently use `gexiv2` to read the EXIF metadata from the images. The disadvantage is that `gexiv2` is part of GNOME and depends on the `Exiv2` library which makes it more difficult to install.
Consider alternative packages such as [exif](https://pypi.org/project/exif/). | non_process | move away from we currently use to read the exif metadata from the images the disadvantage is that is part of gnome and depends on the library which makes it more difficult to install consider alternative packages such as | 0 |
30,944 | 8,630,450,038 | IssuesEvent | 2018-11-22 01:58:06 | apache/incubator-mxnet | https://api.github.com/repos/apache/incubator-mxnet | closed | Check failed: err == CUDNN_STATUS_SUCCESS (6 vs. 0) CUDNN_STATUS_ARCH_MISMATCH | Build Pending Requester Info | Note: Providing complete information in the most concise form is the best way to get help. This issue template serves as the checklist for essential information to most of the technical issues and bug reports. For non-technical issues and feature requests, feel free to present the information in what you believe is the best form.
For Q & A and discussion, please start a discussion thread at https://discuss.mxnet.io
## GTX680, mxnet-cu90==1.0.0, Check failed: err == CUDNN_STATUS_SUCCESS (6 vs. 0) CUDNN_STATUS_ARCH_MISMATCH
##
1.Install CUDA 9.0
2.Install cuDNN 7.0.5.15 for CUDA 9.0
3.export LD_LIBRARY_PATH=/usr/local/cuda/lib64/:$LD_LIBRARY_PATH
4.pip install mxnet-cu90==1.0.0
5. run 'python'
yong@yong-desktop:~$ python
Python 2.7.14 |Anaconda, Inc.| (default, Oct 16 2017, 17:29:19)
[GCC 7.2.0] on linux2
Type "help", "copyright", "credits" or "license" for more information.
>>> import mxnet as mx
/home/yong/anaconda2/lib/python2.7/site-packages/urllib3/contrib/pyopenssl.py:46: DeprecationWarning: OpenSSL.rand is deprecated - you should use os.urandom instead
import OpenSSL.SSL
>>> a = mx.nd.ones((2, 3), mx.gpu())
[17:11:45] /home/travis/build/dmlc/mxnet-distro/mxnet-build/dmlc-core/include/dmlc/logging.h:308: [17:11:45] /home/travis/build/dmlc/mxnet-distro/mxnet-build/mshadow/mshadow/./stream_gpu-inl.h:171: Check failed: err == CUDNN_STATUS_SUCCESS (6 vs. 0) CUDNN_STATUS_ARCH_MISMATCH
Stack trace returned 9 entries:
[bt] (0) /home/yong/anaconda2/lib/python2.7/site-packages/mxnet/libmxnet.so(+0x28980c) [0x7ff7be0e080c]
[bt] (1) /home/yong/anaconda2/lib/python2.7/site-packages/mxnet/libmxnet.so(+0x24188dd) [0x7ff7c026f8dd]
[bt] (2) /home/yong/anaconda2/lib/python2.7/site-packages/mxnet/libmxnet.so(+0x24191e0) [0x7ff7c02701e0]
[bt] (3) /home/yong/anaconda2/lib/python2.7/site-packages/mxnet/libmxnet.so(+0x24308b7) [0x7ff7c02878b7]
[bt] (4) /home/yong/anaconda2/lib/python2.7/site-packages/mxnet/libmxnet.so(+0x2430a46) [0x7ff7c0287a46]
[bt] (5) /home/yong/anaconda2/lib/python2.7/site-packages/mxnet/libmxnet.so(+0x242d26b) [0x7ff7c028426b]
[bt] (6) /home/yong/anaconda2/bin/../lib/libstdc++.so.6(+0xafc5c) [0x7ff79f058c5c]
[bt] (7) /lib/x86_64-linux-gnu/libpthread.so.0(+0x76ba) [0x7ff7df2156ba]
[bt] (8) /lib/x86_64-linux-gnu/libc.so.6(clone+0x6d) [0x7ff7de83b41d]
[17:11:45] /home/travis/build/dmlc/mxnet-distro/mxnet-build/dmlc-core/include/dmlc/logging.h:308: [17:11:45] /home/travis/build/dmlc/mxnet-distro/mxnet-build/mshadow/mshadow/./stream_gpu-inl.h:171: Check failed: err == CUDNN_STATUS_SUCCESS (6 vs. 0) CUDNN_STATUS_ARCH_MISMATCH
Stack trace returned 9 entries:
[bt] (0) /home/yong/anaconda2/lib/python2.7/site-packages/mxnet/libmxnet.so(+0x28980c) [0x7ff7be0e080c]
[bt] (1) /home/yong/anaconda2/lib/python2.7/site-packages/mxnet/libmxnet.so(+0x24188dd) [0x7ff7c026f8dd]
[bt] (2) /home/yong/anaconda2/lib/python2.7/site-packages/mxnet/libmxnet.so(+0x24191e0) [0x7ff7c02701e0]
[bt] (3) /home/yong/anaconda2/lib/python2.7/site-packages/mxnet/libmxnet.so(+0x24308b7) [0x7ff7c02878b7]
[bt] (4) /home/yong/anaconda2/lib/python2.7/site-packages/mxnet/libmxnet.so(+0x2430a46) [0x7ff7c0287a46]
[bt] (5) /home/yong/anaconda2/lib/python2.7/site-packages/mxnet/libmxnet.so(+0x242d26b) [0x7ff7c028426b]
[bt] (6) /home/yong/anaconda2/bin/../lib/libstdc++.so.6(+0xafc5c) [0x7ff79f058c5c]
[bt] (7) /lib/x86_64-linux-gnu/libpthread.so.0(+0x76ba) [0x7ff7df2156ba]
[bt] (8) /lib/x86_64-linux-gnu/libc.so.6(clone+0x6d) [0x7ff7de83b41d]
terminate called after throwing an instance of 'dmlc::Error'
what(): [17:11:45] /home/travis/build/dmlc/mxnet-distro/mxnet-build/mshadow/mshadow/./stream_gpu-inl.h:171: Check failed: err == CUDNN_STATUS_SUCCESS (6 vs. 0) CUDNN_STATUS_ARCH_MISMATCH
Stack trace returned 9 entries:
[bt] (0) /home/yong/anaconda2/lib/python2.7/site-packages/mxnet/libmxnet.so(+0x28980c) [0x7ff7be0e080c]
[bt] (1) /home/yong/anaconda2/lib/python2.7/site-packages/mxnet/libmxnet.so(+0x24188dd) [0x7ff7c026f8dd]
[bt] (2) /home/yong/anaconda2/lib/python2.7/site-packages/mxnet/libmxnet.so(+0x24191e0) [0x7ff7c02701e0]
[bt] (3) /home/yong/anaconda2/lib/python2.7/site-packages/mxnet/libmxnet.so(+0x24308b7) [0x7ff7c02878b7]
[bt] (4) /home/yong/anaconda2/lib/python2.7/site-packages/mxnet/libmxnet.so(+0x2430a46) [0x7ff7c0287a46]
[bt] (5) /home/yong/anaconda2/lib/python2.7/site-packages/mxnet/libmxnet.so(+0x242d26b) [0x7ff7c028426b]
[bt] (6) /home/yong/anaconda2/bin/../lib/libstdc++.so.6(+0xafc5c) [0x7ff79f058c5c]
[bt] (7) /lib/x86_64-linux-gnu/libpthread.so.0(+0x76ba) [0x7ff7df2156ba]
[bt] (8) /lib/x86_64-linux-gnu/libc.so.6(clone+0x6d) [0x7ff7de83b41d]
(I'm using Python 2.7.14 |Anaconda.)
| 1.0 | Check failed: err == CUDNN_STATUS_SUCCESS (6 vs. 0) CUDNN_STATUS_ARCH_MISMATCH - Note: Providing complete information in the most concise form is the best way to get help. This issue template serves as the checklist for essential information to most of the technical issues and bug reports. For non-technical issues and feature requests, feel free to present the information in what you believe is the best form.
For Q & A and discussion, please start a discussion thread at https://discuss.mxnet.io
## GTX680, mxnet-cu90==1.0.0, Check failed: err == CUDNN_STATUS_SUCCESS (6 vs. 0) CUDNN_STATUS_ARCH_MISMATCH
##
1.Install CUDA 9.0
2.Install cuDNN 7.0.5.15 for CUDA 9.0
3.export LD_LIBRARY_PATH=/usr/local/cuda/lib64/:$LD_LIBRARY_PATH
4.pip install mxnet-cu90==1.0.0
5. run 'python'
yong@yong-desktop:~$ python
Python 2.7.14 |Anaconda, Inc.| (default, Oct 16 2017, 17:29:19)
[GCC 7.2.0] on linux2
Type "help", "copyright", "credits" or "license" for more information.
>>> import mxnet as mx
/home/yong/anaconda2/lib/python2.7/site-packages/urllib3/contrib/pyopenssl.py:46: DeprecationWarning: OpenSSL.rand is deprecated - you should use os.urandom instead
import OpenSSL.SSL
>>> a = mx.nd.ones((2, 3), mx.gpu())
[17:11:45] /home/travis/build/dmlc/mxnet-distro/mxnet-build/dmlc-core/include/dmlc/logging.h:308: [17:11:45] /home/travis/build/dmlc/mxnet-distro/mxnet-build/mshadow/mshadow/./stream_gpu-inl.h:171: Check failed: err == CUDNN_STATUS_SUCCESS (6 vs. 0) CUDNN_STATUS_ARCH_MISMATCH
Stack trace returned 9 entries:
[bt] (0) /home/yong/anaconda2/lib/python2.7/site-packages/mxnet/libmxnet.so(+0x28980c) [0x7ff7be0e080c]
[bt] (1) /home/yong/anaconda2/lib/python2.7/site-packages/mxnet/libmxnet.so(+0x24188dd) [0x7ff7c026f8dd]
[bt] (2) /home/yong/anaconda2/lib/python2.7/site-packages/mxnet/libmxnet.so(+0x24191e0) [0x7ff7c02701e0]
[bt] (3) /home/yong/anaconda2/lib/python2.7/site-packages/mxnet/libmxnet.so(+0x24308b7) [0x7ff7c02878b7]
[bt] (4) /home/yong/anaconda2/lib/python2.7/site-packages/mxnet/libmxnet.so(+0x2430a46) [0x7ff7c0287a46]
[bt] (5) /home/yong/anaconda2/lib/python2.7/site-packages/mxnet/libmxnet.so(+0x242d26b) [0x7ff7c028426b]
[bt] (6) /home/yong/anaconda2/bin/../lib/libstdc++.so.6(+0xafc5c) [0x7ff79f058c5c]
[bt] (7) /lib/x86_64-linux-gnu/libpthread.so.0(+0x76ba) [0x7ff7df2156ba]
[bt] (8) /lib/x86_64-linux-gnu/libc.so.6(clone+0x6d) [0x7ff7de83b41d]
[17:11:45] /home/travis/build/dmlc/mxnet-distro/mxnet-build/dmlc-core/include/dmlc/logging.h:308: [17:11:45] /home/travis/build/dmlc/mxnet-distro/mxnet-build/mshadow/mshadow/./stream_gpu-inl.h:171: Check failed: err == CUDNN_STATUS_SUCCESS (6 vs. 0) CUDNN_STATUS_ARCH_MISMATCH
Stack trace returned 9 entries:
[bt] (0) /home/yong/anaconda2/lib/python2.7/site-packages/mxnet/libmxnet.so(+0x28980c) [0x7ff7be0e080c]
[bt] (1) /home/yong/anaconda2/lib/python2.7/site-packages/mxnet/libmxnet.so(+0x24188dd) [0x7ff7c026f8dd]
[bt] (2) /home/yong/anaconda2/lib/python2.7/site-packages/mxnet/libmxnet.so(+0x24191e0) [0x7ff7c02701e0]
[bt] (3) /home/yong/anaconda2/lib/python2.7/site-packages/mxnet/libmxnet.so(+0x24308b7) [0x7ff7c02878b7]
[bt] (4) /home/yong/anaconda2/lib/python2.7/site-packages/mxnet/libmxnet.so(+0x2430a46) [0x7ff7c0287a46]
[bt] (5) /home/yong/anaconda2/lib/python2.7/site-packages/mxnet/libmxnet.so(+0x242d26b) [0x7ff7c028426b]
[bt] (6) /home/yong/anaconda2/bin/../lib/libstdc++.so.6(+0xafc5c) [0x7ff79f058c5c]
[bt] (7) /lib/x86_64-linux-gnu/libpthread.so.0(+0x76ba) [0x7ff7df2156ba]
[bt] (8) /lib/x86_64-linux-gnu/libc.so.6(clone+0x6d) [0x7ff7de83b41d]
terminate called after throwing an instance of 'dmlc::Error'
what(): [17:11:45] /home/travis/build/dmlc/mxnet-distro/mxnet-build/mshadow/mshadow/./stream_gpu-inl.h:171: Check failed: err == CUDNN_STATUS_SUCCESS (6 vs. 0) CUDNN_STATUS_ARCH_MISMATCH
Stack trace returned 9 entries:
[bt] (0) /home/yong/anaconda2/lib/python2.7/site-packages/mxnet/libmxnet.so(+0x28980c) [0x7ff7be0e080c]
[bt] (1) /home/yong/anaconda2/lib/python2.7/site-packages/mxnet/libmxnet.so(+0x24188dd) [0x7ff7c026f8dd]
[bt] (2) /home/yong/anaconda2/lib/python2.7/site-packages/mxnet/libmxnet.so(+0x24191e0) [0x7ff7c02701e0]
[bt] (3) /home/yong/anaconda2/lib/python2.7/site-packages/mxnet/libmxnet.so(+0x24308b7) [0x7ff7c02878b7]
[bt] (4) /home/yong/anaconda2/lib/python2.7/site-packages/mxnet/libmxnet.so(+0x2430a46) [0x7ff7c0287a46]
[bt] (5) /home/yong/anaconda2/lib/python2.7/site-packages/mxnet/libmxnet.so(+0x242d26b) [0x7ff7c028426b]
[bt] (6) /home/yong/anaconda2/bin/../lib/libstdc++.so.6(+0xafc5c) [0x7ff79f058c5c]
[bt] (7) /lib/x86_64-linux-gnu/libpthread.so.0(+0x76ba) [0x7ff7df2156ba]
[bt] (8) /lib/x86_64-linux-gnu/libc.so.6(clone+0x6d) [0x7ff7de83b41d]
(I'm using Python 2.7.14 |Anaconda.)
| non_process | check failed err cudnn status success vs cudnn status arch mismatch note providing complete information in the most concise form is the best way to get help this issue template serves as the checklist for essential information to most of the technical issues and bug reports for non technical issues and feature requests feel free to present the information in what you believe is the best form for q a and discussion please start a discussion thread at mxnet check failed err cudnn status success vs cudnn status arch mismatch install cuda install cudnn for cuda export ld library path usr local cuda ld library path pip install mxnet run python yong yong desktop python python anaconda inc default oct on type help copyright credits or license for more information import mxnet as mx home yong lib site packages contrib pyopenssl py deprecationwarning openssl rand is deprecated you should use os urandom instead import openssl ssl a mx nd ones mx gpu home travis build dmlc mxnet distro mxnet build dmlc core include dmlc logging h home travis build dmlc mxnet distro mxnet build mshadow mshadow stream gpu inl h check failed err cudnn status success vs cudnn status arch mismatch stack trace returned entries home yong lib site packages mxnet libmxnet so home yong lib site packages mxnet libmxnet so home yong lib site packages mxnet libmxnet so home yong lib site packages mxnet libmxnet so home yong lib site packages mxnet libmxnet so home yong lib site packages mxnet libmxnet so home yong bin lib libstdc so lib linux gnu libpthread so lib linux gnu libc so clone home travis build dmlc mxnet distro mxnet build dmlc core include dmlc logging h home travis build dmlc mxnet distro mxnet build mshadow mshadow stream gpu inl h check failed err cudnn status success vs cudnn status arch mismatch stack trace returned entries home yong lib site packages mxnet libmxnet so home yong lib site packages mxnet libmxnet so home yong lib site packages mxnet libmxnet so home yong lib site packages mxnet libmxnet so home yong lib site packages mxnet libmxnet so home yong lib site packages mxnet libmxnet so home yong bin lib libstdc so lib linux gnu libpthread so lib linux gnu libc so clone terminate called after throwing an instance of dmlc error what home travis build dmlc mxnet distro mxnet build mshadow mshadow stream gpu inl h check failed err cudnn status success vs cudnn status arch mismatch stack trace returned entries home yong lib site packages mxnet libmxnet so home yong lib site packages mxnet libmxnet so home yong lib site packages mxnet libmxnet so home yong lib site packages mxnet libmxnet so home yong lib site packages mxnet libmxnet so home yong lib site packages mxnet libmxnet so home yong bin lib libstdc so lib linux gnu libpthread so lib linux gnu libc so clone i m using python anaconda | 0 |
26,907 | 5,293,828,862 | IssuesEvent | 2017-02-09 08:58:57 | architecture-building-systems/CEAforArcGIS | https://api.github.com/repos/architecture-building-systems/CEAforArcGIS | closed | Installation guide-CEA documentation | documentation Priority 1 | Make a list of the steps involved in installation of CEA along with screenshots or the add the text output provided by the command prompt | 1.0 | Installation guide-CEA documentation - Make a list of the steps involved in installation of CEA along with screenshots or the add the text output provided by the command prompt | non_process | installation guide cea documentation make a list of the steps involved in installation of cea along with screenshots or the add the text output provided by the command prompt | 0 |
206,824 | 16,058,123,232 | IssuesEvent | 2021-04-23 08:39:05 | jOOQ/jOOQ | https://api.github.com/repos/jOOQ/jOOQ | closed | Add a manual section about LATERAL and APPLY | C: Documentation E: All Editions P: Medium T: Enhancement | `LATERAL` and `APPLY` are documented only as a part of the `JOIN` section, when they should have their own section:
https://www.jooq.org/doc/3.13/manual/sql-building/sql-statements/select-statement/join-clause/ | 1.0 | Add a manual section about LATERAL and APPLY - `LATERAL` and `APPLY` are documented only as a part of the `JOIN` section, when they should have their own section:
https://www.jooq.org/doc/3.13/manual/sql-building/sql-statements/select-statement/join-clause/ | non_process | add a manual section about lateral and apply lateral and apply are documented only as a part of the join section when they should have their own section | 0 |
265,811 | 23,200,423,413 | IssuesEvent | 2022-08-01 20:50:10 | foundry-rs/foundry | https://api.github.com/repos/foundry-rs/foundry | closed | expectRevert fails when returning a struct | T-bug Cmd-forge-test C-forge | ### Component
Forge
### Have you ensured that all of these are up to date?
- [X] Foundry
- [X] Foundryup
### What version of Foundry are you on?
33dbaa5 2022-07-23T00:08:24.676044Z
### What command(s) is the bug in?
forge test
### Operating System
macOS (Apple Silicon)
### Describe the bug
My team and I are having an issue where tests expect a revert with the onlyOwner modifier reason and the logs show the call reverting with the expected reason, but the test fails. If you remove the reason in the expectRevert the test fails, but with onlyOwner modifier reason. We debugged the test and found that the issue occurs when returning the metadata struct. If we just return the address of the deployed vault from the deployVaultV1 function, the test passes as it is written with the expected revert reason. When we also try to return the metadata in the function, the test fails and reverts without reason.

1
function being tested from the factory contract
``` solidity
function deployVaultV1(
VaultParams memory _vaultParams,
uint256 _type,
bool _enabled,
address _stakingAddress,
address _submitter,
string memory _metadataCID,
address[8] memory _swapTokenAddresses,
address _swapAddress,
uint256 _exchange
) external onlyOwner returns (VaultMetadata memory, address) {
Vault vault = new Vault(
_vaultParams.token,
_vaultParams.yearnRegistry,
_vaultParams.contractRegistry,
_vaultParams.staking,
_vaultParams.feeStructure
);
VaultMetadata memory metadata = VaultMetadata({
vaultAddress: address(vault),
vaultType: _type,
enabled: _enabled,
stakingAddress: _stakingAddress,
submitter: _submitter,
metadataCID: _metadataCID,
swapTokenAddresses: _swapTokenAddresses,
swapAddress: _swapAddress,
exchange: _exchange
});
emit VaultV1Deployment(address(vault));
return (metadata, address(vault));
}
```
expectRevert onlyOwner test (passes if deploy function modified to only return vault address, fails, and reverts without reason when the function is as written and returning the vault address and metadata struct)
``` solidity
function test__deployVaultV1NotOwnerReverts() public {
vm.expectRevert("Only the contract owner may perform this action");
(VaultMetadata memory metadata, address vault) = vaultsV1Factory.deployVaultV1(
vaultParams,
1,
true,
address(0x1111),
vaultsV1ControllerOwner,
"someCID",
swapTokenAddresses,
address(0x2222),
1
);
assertEq(vault, address(0), "vault deployment failed");
assertEq(metadata.vaultAddress, address(0), "metadata not constructed");
}
```
[Factory contract](https://github.com/popcorndao/sweet-caramel/blob/feature/vault-registry/packages/hardhat/contracts/core/defi/vault/VaultsV1Factory.sol)
[Factory test contract](https://github.com/popcorndao/sweet-caramel/blob/feature/vault-registry/packages/hardhat/test/forge/fork/VaultsV1Factory.t.sol) | 1.0 | expectRevert fails when returning a struct - ### Component
Forge
### Have you ensured that all of these are up to date?
- [X] Foundry
- [X] Foundryup
### What version of Foundry are you on?
33dbaa5 2022-07-23T00:08:24.676044Z
### What command(s) is the bug in?
forge test
### Operating System
macOS (Apple Silicon)
### Describe the bug
My team and I are having an issue where tests expect a revert with the onlyOwner modifier reason and the logs show the call reverting with the expected reason, but the test fails. If you remove the reason in the expectRevert the test fails, but with onlyOwner modifier reason. We debugged the test and found that the issue occurs when returning the metadata struct. If we just return the address of the deployed vault from the deployVaultV1 function, the test passes as it is written with the expected revert reason. When we also try to return the metadata in the function, the test fails and reverts without reason.

1
function being tested from the factory contract
``` solidity
function deployVaultV1(
VaultParams memory _vaultParams,
uint256 _type,
bool _enabled,
address _stakingAddress,
address _submitter,
string memory _metadataCID,
address[8] memory _swapTokenAddresses,
address _swapAddress,
uint256 _exchange
) external onlyOwner returns (VaultMetadata memory, address) {
Vault vault = new Vault(
_vaultParams.token,
_vaultParams.yearnRegistry,
_vaultParams.contractRegistry,
_vaultParams.staking,
_vaultParams.feeStructure
);
VaultMetadata memory metadata = VaultMetadata({
vaultAddress: address(vault),
vaultType: _type,
enabled: _enabled,
stakingAddress: _stakingAddress,
submitter: _submitter,
metadataCID: _metadataCID,
swapTokenAddresses: _swapTokenAddresses,
swapAddress: _swapAddress,
exchange: _exchange
});
emit VaultV1Deployment(address(vault));
return (metadata, address(vault));
}
```
expectRevert onlyOwner test (passes if deploy function modified to only return vault address, fails, and reverts without reason when the function is as written and returning the vault address and metadata struct)
``` solidity
function test__deployVaultV1NotOwnerReverts() public {
vm.expectRevert("Only the contract owner may perform this action");
(VaultMetadata memory metadata, address vault) = vaultsV1Factory.deployVaultV1(
vaultParams,
1,
true,
address(0x1111),
vaultsV1ControllerOwner,
"someCID",
swapTokenAddresses,
address(0x2222),
1
);
assertEq(vault, address(0), "vault deployment failed");
assertEq(metadata.vaultAddress, address(0), "metadata not constructed");
}
```
[Factory contract](https://github.com/popcorndao/sweet-caramel/blob/feature/vault-registry/packages/hardhat/contracts/core/defi/vault/VaultsV1Factory.sol)
[Factory test contract](https://github.com/popcorndao/sweet-caramel/blob/feature/vault-registry/packages/hardhat/test/forge/fork/VaultsV1Factory.t.sol) | non_process | expectrevert fails when returning a struct component forge have you ensured that all of these are up to date foundry foundryup what version of foundry are you on what command s is the bug in forge test operating system macos apple silicon describe the bug my team and i are having an issue where tests expect a revert with the onlyowner modifier reason and the logs show the call reverting with the expected reason but the test fails if you remove the reason in the expectrevert the test fails but with onlyowner modifier reason we debugged the test and found that the issue occurs when returning the metadata struct if we just return the address of the deployed vault from the function the test passes as it is written with the expected revert reason when we also try to return the metadata in the function the test fails and reverts without reason function being tested from the factory contract solidity function vaultparams memory vaultparams type bool enabled address stakingaddress address submitter string memory metadatacid address memory swaptokenaddresses address swapaddress exchange external onlyowner returns vaultmetadata memory address vault vault new vault vaultparams token vaultparams yearnregistry vaultparams contractregistry vaultparams staking vaultparams feestructure vaultmetadata memory metadata vaultmetadata vaultaddress address vault vaulttype type enabled enabled stakingaddress stakingaddress submitter submitter metadatacid metadatacid swaptokenaddresses swaptokenaddresses swapaddress swapaddress exchange exchange emit address vault return metadata address vault expectrevert onlyowner test passes if deploy function modified to only return vault address fails and reverts without reason when the function is as written and returning the vault address and metadata struct solidity function test public vm expectrevert only the contract owner may perform this action vaultmetadata memory metadata address vault vaultparams true address somecid swaptokenaddresses address asserteq vault address vault deployment failed asserteq metadata vaultaddress address metadata not constructed | 0 |
321,715 | 9,807,910,834 | IssuesEvent | 2019-06-12 14:37:41 | lbryio/lbry | https://api.github.com/repos/lbryio/lbry | closed | Support intra-protocol thumbnail publishing | good first issue level: 1 priority: low | It is silly that we encourage thumbnails to be published/accessed over HTTP when we have a protocol that is designed to distribute data.
Trivially, thumbnails could also be published as claims, but this is expensive and creates blockchain bloat. Ideally, there would be a solution that allows the `thumbnail` metadata property to support fetching from the LBRY data network without an additional claim.
One idea for a solution is that the `thumbnail` property can simply be a blob hash. When performing a `resolve` or `get`, if the SDK sees a `thumbnail` that contains a valid blob hash, it would automatically fetch them. It would also verify that the blob is a valid image type and return a signature that includes the path to the thumbnail on local disk.
We should consider it a requirement that the SDK does not require any additional application-level calls to fetch thumbnails. Whatever the actual solution, all thumbnail fetching should happen behind the scenes. | 1.0 | Support intra-protocol thumbnail publishing - It is silly that we encourage thumbnails to be published/accessed over HTTP when we have a protocol that is designed to distribute data.
Trivially, thumbnails could also be published as claims, but this is expensive and creates blockchain bloat. Ideally, there would be a solution that allows the `thumbnail` metadata property to support fetching from the LBRY data network without an additional claim.
One idea for a solution is that the `thumbnail` property can simply be a blob hash. When performing a `resolve` or `get`, if the SDK sees a `thumbnail` that contains a valid blob hash, it would automatically fetch them. It would also verify that the blob is a valid image type and return a signature that includes the path to the thumbnail on local disk.
We should consider it a requirement that the SDK does not require any additional application-level calls to fetch thumbnails. Whatever the actual solution, all thumbnail fetching should happen behind the scenes. | non_process | support intra protocol thumbnail publishing it is silly that we encourage thumbnails to be published accessed over http when we have a protocol that is designed to distribute data trivially thumbnails could also be published as claims but this is expensive and creates blockchain bloat ideally there would be a solution that allows the thumbnail metadata property to support fetching from the lbry data network without an additional claim one idea for a solution is that the thumbnail property can simply be a blob hash when performing a resolve or get if the sdk sees a thumbnail that contains a valid blob hash it would automatically fetch them it would also verify that the blob is a valid image type and return a signature that includes the path to the thumbnail on local disk we should consider it a requirement that the sdk does not require any additional application level calls to fetch thumbnails whatever the actual solution all thumbnail fetching should happen behind the scenes | 0 |
4,330 | 7,241,281,529 | IssuesEvent | 2018-02-14 00:20:55 | hashicorp/packer | https://api.github.com/repos/hashicorp/packer | closed | [RFC] Deprecate Shell-Local postprocessor | post-processor/shell-local | The shell-local post-processor will never be as seamless or full-featured as running shell scripts locally after the build is done. The main reason for having this post-processor is being able to access packer variables in the post-processing script, but this is easily handled by using the manifest post-processor (https://www.packer.io/docs/post-processors/manifest.html). Rather than spending energy on a redundant feature, let's remove it and make sure that all use cases that the shell-local post-processor handled have workarounds.
TODO:
- [x] Add deprecation warning to v 1.1.0
- [ ] Write blog post demonstrating how to migrate off of it
- [ ] better document manifest pp [#5413]
- [ ] Remove processor (or leave it, but document that it is unsupported?)
| 1.0 | [RFC] Deprecate Shell-Local postprocessor - The shell-local post-processor will never be as seamless or full-featured as running shell scripts locally after the build is done. The main reason for having this post-processor is being able to access packer variables in the post-processing script, but this is easily handled by using the manifest post-processor (https://www.packer.io/docs/post-processors/manifest.html). Rather than spending energy on a redundant feature, let's remove it and make sure that all use cases that the shell-local post-processor handled have workarounds.
TODO:
- [x] Add deprecation warning to v 1.1.0
- [ ] Write blog post demonstrating how to migrate off of it
- [ ] better document manifest pp [#5413]
- [ ] Remove processor (or leave it, but document that it is unsupported?)
| process | deprecate shell local postprocessor the shell local post processor will never be as seamless or full featured as running shell scripts locally after the build is done the main reason for having this post processor is being able to access packer variables in the post processing script but this is easily handled by using the manifest post processor rather than spending energy on a redundant feature let s remove it and make sure that all use cases that the shell local post processor handled have workarounds todo add deprecation warning to v write blog post demonstrating how to migrate off of it better document manifest pp remove processor or leave it but document that it is unsupported | 1 |
99,760 | 4,064,497,281 | IssuesEvent | 2016-05-26 07:10:42 | kubernetes/kubernetes | https://api.github.com/repos/kubernetes/kubernetes | opened | cluster-autoscaler: incompatible variable names with GKE | area/autoscaling priority/P1 team/control-plane | In kube-up we use var `ENABLE_NODE_AUTOSCALER` while in GKE, documentation, issues etc I think we call it cluster autoscaler. We should make it consistent. | 1.0 | cluster-autoscaler: incompatible variable names with GKE - In kube-up we use var `ENABLE_NODE_AUTOSCALER` while in GKE, documentation, issues etc I think we call it cluster autoscaler. We should make it consistent. | non_process | cluster autoscaler incompatible variable names with gke in kube up we use var enable node autoscaler while in gke documentation issues etc i think we call it cluster autoscaler we should make it consistent | 0 |
48,822 | 10,281,712,290 | IssuesEvent | 2019-08-26 09:11:49 | mozilla/fxa | https://api.github.com/repos/mozilla/fxa | closed | Simplify fxa-js-client release process | cat:code quality help wanted | Running `grunt release` in the js client directory still tries to build and deploy the docs, which fails and is redundant anyway because we have the docs being built and deployed automagically in Circle now. We should take that stuff out. | 1.0 | Simplify fxa-js-client release process - Running `grunt release` in the js client directory still tries to build and deploy the docs, which fails and is redundant anyway because we have the docs being built and deployed automagically in Circle now. We should take that stuff out. | non_process | simplify fxa js client release process running grunt release in the js client directory still tries to build and deploy the docs which fails and is redundant anyway because we have the docs being built and deployed automagically in circle now we should take that stuff out | 0 |
11,039 | 8,271,641,648 | IssuesEvent | 2018-09-16 11:38:25 | ioswinews/hadoop | https://api.github.com/repos/ioswinews/hadoop | closed | CVE-2017-5637 high severity vulnerability detected by WhiteSource | security vulnerability | ## CVE-2017-5637 - High Vulnerability
<details><summary>:warning: Vulnerable Library - <b>zookeeper-3.4.9.jar</b></p></summary>
<p>path: 2/repository/org/apache/zookeeper/zookeeper/3.4.9/zookeeper-3.4.9.jar,/root/.m2/repository/org/apache/zookeeper/zookeeper/3.4.9/zookeeper-3.4.9.jar</p>
<p>
Dependency Hierarchy:
- :x: **zookeeper-3.4.9.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary>:red_circle: Vulnerability Details</summary>
<p>
Two four letter word commands "wchp/wchc" are CPU intensive and could cause spike of CPU utilization on Apache ZooKeeper server if abused, which leads to the server unable to serve legitimate client requests. Apache ZooKeeper thru version 3.4.9 and 3.5.2 suffer from this issue, fixed in 3.4.10, 3.5.3, and later.
<p>Publish Date: 2017-10-10
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-5637>CVE-2017-5637</a></p>
</p>
</details>
<p></p>
<details><summary>:dart: CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary>:white_check_mark: Suggested Fix</summary>
<p>
<p>Type: Change files</p>
<p>Origin: <a href="https://github.com/apache/zookeeper/commit/5fe68506f217246c7ebd96803f9c78e13ec2f11a">https://github.com/apache/zookeeper/commit/5fe68506f217246c7ebd96803f9c78e13ec2f11a</a></p>
<p>Release Date: 2017-03-04</p>
<p>Fix Resolution: Replace or update the following files: NIOServerCnxn.java, ZKTestCase.java, zookeeperAdmin.xml, FourLetterCommands.java, NopCommand.java, NettyServerCnxn.java, FourLetterWordsWhiteListTest.java</p>
</p>
</details>
<p></p>
***
Want to learn more about the open source vulnerabilities in your products? Click [here](https://www.whitesourcesoftware.com) | True | CVE-2017-5637 high severity vulnerability detected by WhiteSource - ## CVE-2017-5637 - High Vulnerability
<details><summary>:warning: Vulnerable Library - <b>zookeeper-3.4.9.jar</b></p></summary>
<p>path: 2/repository/org/apache/zookeeper/zookeeper/3.4.9/zookeeper-3.4.9.jar,/root/.m2/repository/org/apache/zookeeper/zookeeper/3.4.9/zookeeper-3.4.9.jar</p>
<p>
Dependency Hierarchy:
- :x: **zookeeper-3.4.9.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary>:red_circle: Vulnerability Details</summary>
<p>
Two four letter word commands "wchp/wchc" are CPU intensive and could cause spike of CPU utilization on Apache ZooKeeper server if abused, which leads to the server unable to serve legitimate client requests. Apache ZooKeeper thru version 3.4.9 and 3.5.2 suffer from this issue, fixed in 3.4.10, 3.5.3, and later.
<p>Publish Date: 2017-10-10
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-5637>CVE-2017-5637</a></p>
</p>
</details>
<p></p>
<details><summary>:dart: CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary>:white_check_mark: Suggested Fix</summary>
<p>
<p>Type: Change files</p>
<p>Origin: <a href="https://github.com/apache/zookeeper/commit/5fe68506f217246c7ebd96803f9c78e13ec2f11a">https://github.com/apache/zookeeper/commit/5fe68506f217246c7ebd96803f9c78e13ec2f11a</a></p>
<p>Release Date: 2017-03-04</p>
<p>Fix Resolution: Replace or update the following files: NIOServerCnxn.java, ZKTestCase.java, zookeeperAdmin.xml, FourLetterCommands.java, NopCommand.java, NettyServerCnxn.java, FourLetterWordsWhiteListTest.java</p>
</p>
</details>
<p></p>
***
Want to learn more about the open source vulnerabilities in your products? Click [here](https://www.whitesourcesoftware.com) | non_process | cve high severity vulnerability detected by whitesource cve high vulnerability warning vulnerable library zookeeper jar path repository org apache zookeeper zookeeper zookeeper jar root repository org apache zookeeper zookeeper zookeeper jar dependency hierarchy x zookeeper jar vulnerable library red circle vulnerability details two four letter word commands wchp wchc are cpu intensive and could cause spike of cpu utilization on apache zookeeper server if abused which leads to the server unable to serve legitimate client requests apache zookeeper thru version and suffer from this issue fixed in and later publish date url a href dart cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href white check mark suggested fix type change files origin a href release date fix resolution replace or update the following files nioservercnxn java zktestcase java zookeeperadmin xml fourlettercommands java nopcommand java nettyservercnxn java fourletterwordswhitelisttest java want to learn more about the open source vulnerabilities in your products click | 0 |
126,379 | 4,989,959,510 | IssuesEvent | 2016-12-08 13:43:00 | fossasia/open-event-orga-server | https://api.github.com/repos/fossasia/open-event-orga-server | closed | Permanent delete of events not working | Admin bug Priority: High | **Location:** `/admin/events/` - `All Deleted Events` Tab
**Possible cause:** Foreign key constraints
**Fix:** Set proper `ON DELETE` directives to all foreign keys | 1.0 | Permanent delete of events not working - **Location:** `/admin/events/` - `All Deleted Events` Tab
**Possible cause:** Foreign key constraints
**Fix:** Set proper `ON DELETE` directives to all foreign keys | non_process | permanent delete of events not working location admin events all deleted events tab possible cause foreign key constraints fix set proper on delete directives to all foreign keys | 0 |
632,444 | 20,197,054,518 | IssuesEvent | 2022-02-11 11:38:54 | nens/threedi-api-qgis-client | https://api.github.com/repos/nens/threedi-api-qgis-client | closed | When new upload starts, check if loaded model (3Di Toolbox) is same as loaded schematisation (3Di API Client) | ⏰ Priority: 3. Medium | If this isn't the case, warn the user about this
Perhaps we can include this in the overview page of the upload wizard? | 1.0 | When new upload starts, check if loaded model (3Di Toolbox) is same as loaded schematisation (3Di API Client) - If this isn't the case, warn the user about this
Perhaps we can include this in the overview page of the upload wizard? | non_process | when new upload starts check if loaded model toolbox is same as loaded schematisation api client if this isn t the case warn the user about this perhaps we can include this in the overview page of the upload wizard | 0 |
16,786 | 22,013,499,602 | IssuesEvent | 2022-05-27 01:23:13 | allinurl/goaccess | https://api.github.com/repos/allinurl/goaccess | closed | Increase max-items for minute specificity | change log-processing command-line options | # Problem
When using `--date-spec=min` the visitor's graph does only show the past 6 hours. This is probably fine for real-time reporting, but not suitable to generate daily graphs. Turns out, this behavior is related to `--max-items` defaulting to `366`.
# Proposal
When `--date-spec=min` is set, the default for `--max-items` could be increased to `1440` to display the data for the last 24 hours.
**Alternative:** If the described use case (daily graphs) is not strong enough, a note in the [man pages](https://goaccess.io/man#:~:text=%2D%2Ddate%2Dspec%3D%3Cdate%7Chr%7Cmin%3E) could be added, indicating, that number of displayed data point might be increased via `--max-items` to render a longer time range.
# Side note
While searching for an answer to my problem, I found two ticket that are probably fixed and could be closed:
1. [Need ability to break down VISITORS data by minute (#2277)](https://github.com/allinurl/goaccess/issues/2277#)
2. [Add option to use actual minutes, instead of every 'tenth minute' (#1251)](https://github.com/allinurl/goaccess/issues/1251#) | 1.0 | Increase max-items for minute specificity - # Problem
When using `--date-spec=min` the visitor's graph does only show the past 6 hours. This is probably fine for real-time reporting, but not suitable to generate daily graphs. Turns out, this behavior is related to `--max-items` defaulting to `366`.
# Proposal
When `--date-spec=min` is set, the default for `--max-items` could be increased to `1440` to display the data for the last 24 hours.
**Alternative:** If the described use case (daily graphs) is not strong enough, a note in the [man pages](https://goaccess.io/man#:~:text=%2D%2Ddate%2Dspec%3D%3Cdate%7Chr%7Cmin%3E) could be added, indicating, that number of displayed data point might be increased via `--max-items` to render a longer time range.
# Side note
While searching for an answer to my problem, I found two ticket that are probably fixed and could be closed:
1. [Need ability to break down VISITORS data by minute (#2277)](https://github.com/allinurl/goaccess/issues/2277#)
2. [Add option to use actual minutes, instead of every 'tenth minute' (#1251)](https://github.com/allinurl/goaccess/issues/1251#) | process | increase max items for minute specificity problem when using date spec min the visitor s graph does only show the past hours this is probably fine for real time reporting but not suitable to generate daily graphs turns out this behavior is related to max items defaulting to proposal when date spec min is set the default for max items could be increased to to display the data for the last hours alternative if the described use case daily graphs is not strong enough a note in the could be added indicating that number of displayed data point might be increased via max items to render a longer time range side note while searching for an answer to my problem i found two ticket that are probably fixed and could be closed | 1 |
21,623 | 30,022,547,097 | IssuesEvent | 2023-06-27 01:36:51 | bazelbuild/bazel | https://api.github.com/repos/bazelbuild/bazel | closed | bazel coverage do not generate coverage.dat | P3 type: support / not a bug (process) coverage team-Rules-Server stale |
### Description of the problem / feature request:
I have lots of repos which is build by bazel, I using another repo called `main` to build all this repos and get the `coverage.dat`.
but the `coverage.dat` is empty in the `main` repo. it works for each bazel repo.
How can I deal with this? is there any example for me to reference.
### What operating system are you running Bazel on?
> Ubuntu 18.04
### What's the output of `bazel info release`?
> release 4.2.0
| 1.0 | bazel coverage do not generate coverage.dat -
### Description of the problem / feature request:
I have lots of repos which is build by bazel, I using another repo called `main` to build all this repos and get the `coverage.dat`.
but the `coverage.dat` is empty in the `main` repo. it works for each bazel repo.
How can I deal with this? is there any example for me to reference.
### What operating system are you running Bazel on?
> Ubuntu 18.04
### What's the output of `bazel info release`?
> release 4.2.0
| process | bazel coverage do not generate coverage dat description of the problem feature request i have lots of repos which is build by bazel i using another repo called main to build all this repos and get the coverage dat but the coverage dat is empty in the main repo it works for each bazel repo how can i deal with this is there any example for me to reference what operating system are you running bazel on ubuntu what s the output of bazel info release release | 1 |
368,109 | 25,776,961,608 | IssuesEvent | 2022-12-09 12:50:14 | Qiskit/qiskit | https://api.github.com/repos/Qiskit/qiskit | opened | Remove Aer and IBMQ provider from the sidebar | documentation | Currently, the side bar includes *Qiskit Aer* and *Qiskit IBM Quantum Provider* as part of the sidebar.

As the metapackage deprecated the inclusion of these modules (and will be removed mid-2023) I think is time to remove them from the sidebar.
## Does existing documentation need to be updated with new content?
- Which files need to be updated?
`docs/index.rst`
- What content needs to be updated or replaced? (You can copy and paste)
```udiff
API Reference <apidoc/terra>
circuit_library
- aer
- ibmq-provider
release_notes
configuration
```
- What new content needs to be added?
None.
| 1.0 | Remove Aer and IBMQ provider from the sidebar - Currently, the side bar includes *Qiskit Aer* and *Qiskit IBM Quantum Provider* as part of the sidebar.

As the metapackage deprecated the inclusion of these modules (and will be removed mid-2023) I think is time to remove them from the sidebar.
## Does existing documentation need to be updated with new content?
- Which files need to be updated?
`docs/index.rst`
- What content needs to be updated or replaced? (You can copy and paste)
```udiff
API Reference <apidoc/terra>
circuit_library
- aer
- ibmq-provider
release_notes
configuration
```
- What new content needs to be added?
None.
| non_process | remove aer and ibmq provider from the sidebar currently the side bar includes qiskit aer and qiskit ibm quantum provider as part of the sidebar as the metapackage deprecated the inclusion of these modules and will be removed mid i think is time to remove them from the sidebar does existing documentation need to be updated with new content which files need to be updated docs index rst what content needs to be updated or replaced you can copy and paste udiff api reference circuit library aer ibmq provider release notes configuration what new content needs to be added none | 0 |
440,894 | 12,705,832,391 | IssuesEvent | 2020-06-23 05:48:10 | wso2/product-apim | https://api.github.com/repos/wso2/product-apim | closed | Resources do not load in Dev Portal when not signed in | 3.2.0 Priority/Highest Type/Bug Type/UX | ### Description:
The resources of an API that is Published or Deployed as a Prototype do not show (keeps loading) if not signed into the Dev Portal.
### Steps to reproduce:
- Create a new API
- Publish or Deploy as Prototype
- View in Dev Portal
- Go to Try Out

| 1.0 | Resources do not load in Dev Portal when not signed in - ### Description:
The resources of an API that is Published or Deployed as a Prototype do not show (keeps loading) if not signed into the Dev Portal.
### Steps to reproduce:
- Create a new API
- Publish or Deploy as Prototype
- View in Dev Portal
- Go to Try Out

| non_process | resources do not load in dev portal when not signed in description the resources of an api that is published or deployed as a prototype do not show keeps loading if not signed into the dev portal steps to reproduce create a new api publish or deploy as prototype view in dev portal go to try out | 0 |
436,207 | 30,542,136,191 | IssuesEvent | 2023-07-19 22:36:15 | Nadzey/Peshkariki | https://api.github.com/repos/Nadzey/Peshkariki | opened | Test Suite: Create new order. Test Case #4: Find Created Order | documentation | Description: Verify that the created order can be found using the order number.
Test Steps:
Navigate to the order list page.
Find the order with the previously obtained order number | 1.0 | Test Suite: Create new order. Test Case #4: Find Created Order - Description: Verify that the created order can be found using the order number.
Test Steps:
Navigate to the order list page.
Find the order with the previously obtained order number | non_process | test suite create new order test case find created order description verify that the created order can be found using the order number test steps navigate to the order list page find the order with the previously obtained order number | 0 |
6,842 | 2,610,297,182 | IssuesEvent | 2015-02-26 19:35:32 | chrsmith/hedgewars | https://api.github.com/repos/chrsmith/hedgewars | closed | Unlimited Attacks mode: Irregular delays between weapons using ability | auto-migrated Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
1. Run a fight with game scheme with Unlimited Attack mode.
2. Use bazooka or any instant use weapon (I used dynamites).
3. Try to use it again.
What is the expected output? What do you see instead?
The delays between weapons using aren't constant. Sometimes you may be able to
use the weapon again intermediately after firing, but most of the time you'll
have to wait for an indefinite amount of time before you can use it again.
What version of the product are you using? On what operating system?
0.9.15 on Windows XP SP2
Please provide any additional information below.
```
-----
Original issue reported on code.google.com by `adibiaz...@gmail.com` on 21 Jan 2011 at 4:38
* Merged into: #107 | 1.0 | Unlimited Attacks mode: Irregular delays between weapons using ability - ```
What steps will reproduce the problem?
1. Run a fight with game scheme with Unlimited Attack mode.
2. Use bazooka or any instant use weapon (I used dynamites).
3. Try to use it again.
What is the expected output? What do you see instead?
The delays between weapons using aren't constant. Sometimes you may be able to
use the weapon again intermediately after firing, but most of the time you'll
have to wait for an indefinite amount of time before you can use it again.
What version of the product are you using? On what operating system?
0.9.15 on Windows XP SP2
Please provide any additional information below.
```
-----
Original issue reported on code.google.com by `adibiaz...@gmail.com` on 21 Jan 2011 at 4:38
* Merged into: #107 | non_process | unlimited attacks mode irregular delays between weapons using ability what steps will reproduce the problem run a fight with game scheme with unlimited attack mode use bazooka or any instant use weapon i used dynamites try to use it again what is the expected output what do you see instead the delays between weapons using aren t constant sometimes you may be able to use the weapon again intermediately after firing but most of the time you ll have to wait for an indefinite amount of time before you can use it again what version of the product are you using on what operating system on windows xp please provide any additional information below original issue reported on code google com by adibiaz gmail com on jan at merged into | 0 |
73,113 | 15,252,740,378 | IssuesEvent | 2021-02-20 04:19:39 | 1Blackdiamondsc/BlackDiamond-SC | https://api.github.com/repos/1Blackdiamondsc/BlackDiamond-SC | closed | WS-2019-0026 (Medium) detected in marked-0.3.19.js | security vulnerability | ## WS-2019-0026 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>marked-0.3.19.js</b></p></summary>
<p>A markdown parser built for speed</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/marked/0.3.19/marked.js">https://cdnjs.cloudflare.com/ajax/libs/marked/0.3.19/marked.js</a></p>
<p>Path to dependency file: BlackDiamond-SC/node_modules/marked/www/demo.html</p>
<p>Path to vulnerable library: BlackDiamond-SC/node_modules/marked/www/../lib/marked.js</p>
<p>
Dependency Hierarchy:
- :x: **marked-0.3.19.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/1Blackdiamondsc/BlackDiamond-SC/commit/ed44029275ee23e85087fe55659efce3b9315325">ed44029275ee23e85087fe55659efce3b9315325</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Versions 0.3.7 and earlier of marked unescape only lowercase while owsers support both lowercase and uppercase x in hexadecimal form of HTML character entity
<p>Publish Date: 2017-12-23
<p>URL: <a href=https://github.com/markedjs/marked/commit/6d1901ff71abb83aa32ca9a5ce47471382ea42a9>WS-2019-0026</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/markedjs/marked/commit/6d1901ff71abb83aa32ca9a5ce47471382ea42a9">https://github.com/markedjs/marked/commit/6d1901ff71abb83aa32ca9a5ce47471382ea42a9</a></p>
<p>Release Date: 2019-03-17</p>
<p>Fix Resolution: 0.3.9</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | WS-2019-0026 (Medium) detected in marked-0.3.19.js - ## WS-2019-0026 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>marked-0.3.19.js</b></p></summary>
<p>A markdown parser built for speed</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/marked/0.3.19/marked.js">https://cdnjs.cloudflare.com/ajax/libs/marked/0.3.19/marked.js</a></p>
<p>Path to dependency file: BlackDiamond-SC/node_modules/marked/www/demo.html</p>
<p>Path to vulnerable library: BlackDiamond-SC/node_modules/marked/www/../lib/marked.js</p>
<p>
Dependency Hierarchy:
- :x: **marked-0.3.19.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/1Blackdiamondsc/BlackDiamond-SC/commit/ed44029275ee23e85087fe55659efce3b9315325">ed44029275ee23e85087fe55659efce3b9315325</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Versions 0.3.7 and earlier of marked unescape only lowercase while owsers support both lowercase and uppercase x in hexadecimal form of HTML character entity
<p>Publish Date: 2017-12-23
<p>URL: <a href=https://github.com/markedjs/marked/commit/6d1901ff71abb83aa32ca9a5ce47471382ea42a9>WS-2019-0026</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/markedjs/marked/commit/6d1901ff71abb83aa32ca9a5ce47471382ea42a9">https://github.com/markedjs/marked/commit/6d1901ff71abb83aa32ca9a5ce47471382ea42a9</a></p>
<p>Release Date: 2019-03-17</p>
<p>Fix Resolution: 0.3.9</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_process | ws medium detected in marked js ws medium severity vulnerability vulnerable library marked js a markdown parser built for speed library home page a href path to dependency file blackdiamond sc node modules marked www demo html path to vulnerable library blackdiamond sc node modules marked www lib marked js dependency hierarchy x marked js vulnerable library found in head commit a href found in base branch main vulnerability details versions and earlier of marked unescape only lowercase while owsers support both lowercase and uppercase x in hexadecimal form of html character entity publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
406,001 | 11,885,911,277 | IssuesEvent | 2020-03-27 20:39:16 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | closed | website feedback | Priority: Medium | 
- [x] Tooltips for team members should display their title and bio (perhaps with a 'more...' button if the bio is long). Tooltip should be something like this:
Eco Dev: John Krajewski
Currently working on 'Biomes' <web link to feature ideally>
Eco Designer, Strange Loop CEO
<bio here>
- [x] Can you make the venture beat review link go directly to the section on Eco on that page?

- [x] Lists are a too-small font
- [x] Change color of notification bubbles to yellow | 1.0 | website feedback - 
- [x] Tooltips for team members should display their title and bio (perhaps with a 'more...' button if the bio is long). Tooltip should be something like this:
Eco Dev: John Krajewski
Currently working on 'Biomes' <web link to feature ideally>
Eco Designer, Strange Loop CEO
<bio here>
- [x] Can you make the venture beat review link go directly to the section on Eco on that page?

- [x] Lists are a too-small font
- [x] Change color of notification bubbles to yellow | non_process | website feedback tooltips for team members should display their title and bio perhaps with a more button if the bio is long tooltip should be something like this eco dev john krajewski currently working on biomes eco designer strange loop ceo can you make the venture beat review link go directly to the section on eco on that page lists are a too small font change color of notification bubbles to yellow | 0 |
9,805 | 12,819,684,203 | IssuesEvent | 2020-07-06 03:05:37 | JustBru00/RenamePlugin | https://api.github.com/repos/JustBru00/RenamePlugin | closed | i need help | Need More Information Processing Question | Hi, I want to get help from you on something ı want to use your plugin in my server but ı having trouble blacklist,rename character limit and xp
does not work even though I set. If you could help with this situation, I would greatly appreciate it I look forward to hearing from you have a nice day | 1.0 | i need help - Hi, I want to get help from you on something ı want to use your plugin in my server but ı having trouble blacklist,rename character limit and xp
does not work even though I set. If you could help with this situation, I would greatly appreciate it I look forward to hearing from you have a nice day | process | i need help hi i want to get help from you on something ı want to use your plugin in my server but ı having trouble blacklist rename character limit and xp does not work even though i set if you could help with this situation i would greatly appreciate it i look forward to hearing from you have a nice day | 1 |
7,216 | 10,347,026,422 | IssuesEvent | 2019-09-04 16:26:01 | geneontology/go-ontology | https://api.github.com/repos/geneontology/go-ontology | closed | obsolete or merge GO:0061487 | PomBase cell cycle and DNA processes obsoletion regulation | This should have been done as part of #12533, but seems to have been overlooked:
DNA replication initiation from late origin ! GO:0061487 should either be made obsolete or merged into its superclass, GO:0006270. I have no preference which is done, but there are no biological grounds for GO:0061487 to be retained as a separate term.
Rationale, from the older ticket:
> ... initiation is initiation is initiation. The mechanism and involved gene products are the same at early origins and at late origins, so there is no differentium to distinguish "initiation at late origin" from "initiation". | 1.0 | obsolete or merge GO:0061487 - This should have been done as part of #12533, but seems to have been overlooked:
DNA replication initiation from late origin ! GO:0061487 should either be made obsolete or merged into its superclass, GO:0006270. I have no preference which is done, but there are no biological grounds for GO:0061487 to be retained as a separate term.
Rationale, from the older ticket:
> ... initiation is initiation is initiation. The mechanism and involved gene products are the same at early origins and at late origins, so there is no differentium to distinguish "initiation at late origin" from "initiation". | process | obsolete or merge go this should have been done as part of but seems to have been overlooked dna replication initiation from late origin go should either be made obsolete or merged into its superclass go i have no preference which is done but there are no biological grounds for go to be retained as a separate term rationale from the older ticket initiation is initiation is initiation the mechanism and involved gene products are the same at early origins and at late origins so there is no differentium to distinguish initiation at late origin from initiation | 1 |
138,192 | 30,828,392,187 | IssuesEvent | 2023-08-01 22:18:25 | backdrop-contrib/examples | https://api.github.com/repos/backdrop-contrib/examples | closed | Remove the hook_uninstall() implementations that are not necessary | needs code review task | `hook_uninstall()` implementations that remove a table defined in `hook_schema()` or that delete the configuration values associated with the module should be removed. Backdrop already does that in [`backdrop_uninstall_modules()`](https://docs.backdropcms.org/api/backdrop/core%21includes%21install.inc/function/backdrop_uninstall_modules/1). | 1.0 | Remove the hook_uninstall() implementations that are not necessary - `hook_uninstall()` implementations that remove a table defined in `hook_schema()` or that delete the configuration values associated with the module should be removed. Backdrop already does that in [`backdrop_uninstall_modules()`](https://docs.backdropcms.org/api/backdrop/core%21includes%21install.inc/function/backdrop_uninstall_modules/1). | non_process | remove the hook uninstall implementations that are not necessary hook uninstall implementations that remove a table defined in hook schema or that delete the configuration values associated with the module should be removed backdrop already does that in | 0 |
14,873 | 2,831,390,046 | IssuesEvent | 2015-05-24 15:54:55 | nobodyguy/dslrdashboard | https://api.github.com/repos/nobodyguy/dslrdashboard | closed | Feature request: Initiate timelapse from LRTimelapse | auto-migrated Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
1. When initiating timelapse from dsldashboard am unable to move to LRTimelapse
section
2.
3.
What is the expected output? What do you see instead?
Not having a separate intervalometer it would be great to be able to control
shutter and make LRTimelapse holy grail changes from within dslrdashboard
What version of the product are you using? On what operating system?
0.30.27 on Nexus7 4.3. Canon 40D
Please provide any additional information below.
Fantastic tool that would be even better if I could make shutter/iso changes
with histograms while driving a timelapse sequence.
App is well worth paying for so I've made a small donation. Thanks!
```
Original issue reported on code.google.com by `philipjh...@gmail.com` on 28 Aug 2013 at 10:28 | 1.0 | Feature request: Initiate timelapse from LRTimelapse - ```
What steps will reproduce the problem?
1. When initiating timelapse from dsldashboard am unable to move to LRTimelapse
section
2.
3.
What is the expected output? What do you see instead?
Not having a separate intervalometer it would be great to be able to control
shutter and make LRTimelapse holy grail changes from within dslrdashboard
What version of the product are you using? On what operating system?
0.30.27 on Nexus7 4.3. Canon 40D
Please provide any additional information below.
Fantastic tool that would be even better if I could make shutter/iso changes
with histograms while driving a timelapse sequence.
App is well worth paying for so I've made a small donation. Thanks!
```
Original issue reported on code.google.com by `philipjh...@gmail.com` on 28 Aug 2013 at 10:28 | non_process | feature request initiate timelapse from lrtimelapse what steps will reproduce the problem when initiating timelapse from dsldashboard am unable to move to lrtimelapse section what is the expected output what do you see instead not having a separate intervalometer it would be great to be able to control shutter and make lrtimelapse holy grail changes from within dslrdashboard what version of the product are you using on what operating system on canon please provide any additional information below fantastic tool that would be even better if i could make shutter iso changes with histograms while driving a timelapse sequence app is well worth paying for so i ve made a small donation thanks original issue reported on code google com by philipjh gmail com on aug at | 0 |
323,895 | 23,972,597,231 | IssuesEvent | 2022-09-13 08:59:07 | mantidproject/mantid | https://api.github.com/repos/mantidproject/mantid | closed | Update EngDiff UI dev testing docs | Documentation Maintenance ISIS Team: Diffraction | **Describe the outcome that is desired.**
Points arising from manual testing for 6.3 release - #33191
- update image for the calibration tab
- update image for focus tab
- under test 5 need to add instructions on how to load multiple runs
| 1.0 | Update EngDiff UI dev testing docs - **Describe the outcome that is desired.**
Points arising from manual testing for 6.3 release - #33191
- update image for the calibration tab
- update image for focus tab
- under test 5 need to add instructions on how to load multiple runs
| non_process | update engdiff ui dev testing docs describe the outcome that is desired points arising from manual testing for release update image for the calibration tab update image for focus tab under test need to add instructions on how to load multiple runs | 0 |
5,896 | 8,711,706,826 | IssuesEvent | 2018-12-06 19:59:37 | hashicorp/packer | https://api.github.com/repos/hashicorp/packer | closed | Progress bar for vagrant-cloud upload | enhancement post-processor/vagrant-cloud | When running the vagrant-cloud post processor it appears to stall during the upload. There is not feedback letting me know that the upload is still happening or how long it will take. Packer even ignores `ctrl+c` which at first led me to believe that it had indeed stalled or become stuck.
Once I realised it was just my upload speeds (which are legitimately terrible, <50kbps 😢) then I decided to leave the upload running overnight, it eventually completed in > 3hrs.
I've managed to find a work around for the moment, by running `nethogs` I can see the total MB that packer has uploaded which I can use to take a guess, but it would be really great if there was a progress bar, or even just some text saying `xMB / yMB uploaded` that updated every 10 seconds or so.
I noticed this was a feature request for the atlas post processor (#4364) but it was closed because you were focused on releasing 1.0. Would it be worth taking another look at it now? | 1.0 | Progress bar for vagrant-cloud upload - When running the vagrant-cloud post processor it appears to stall during the upload. There is not feedback letting me know that the upload is still happening or how long it will take. Packer even ignores `ctrl+c` which at first led me to believe that it had indeed stalled or become stuck.
Once I realised it was just my upload speeds (which are legitimately terrible, <50kbps 😢) then I decided to leave the upload running overnight, it eventually completed in > 3hrs.
I've managed to find a work around for the moment, by running `nethogs` I can see the total MB that packer has uploaded which I can use to take a guess, but it would be really great if there was a progress bar, or even just some text saying `xMB / yMB uploaded` that updated every 10 seconds or so.
I noticed this was a feature request for the atlas post processor (#4364) but it was closed because you were focused on releasing 1.0. Would it be worth taking another look at it now? | process | progress bar for vagrant cloud upload when running the vagrant cloud post processor it appears to stall during the upload there is not feedback letting me know that the upload is still happening or how long it will take packer even ignores ctrl c which at first led me to believe that it had indeed stalled or become stuck once i realised it was just my upload speeds which are legitimately terrible i ve managed to find a work around for the moment by running nethogs i can see the total mb that packer has uploaded which i can use to take a guess but it would be really great if there was a progress bar or even just some text saying xmb ymb uploaded that updated every seconds or so i noticed this was a feature request for the atlas post processor but it was closed because you were focused on releasing would it be worth taking another look at it now | 1 |
138,675 | 11,211,251,332 | IssuesEvent | 2020-01-06 15:03:39 | CGCookie/retopoflow | https://api.github.com/repos/CGCookie/retopoflow | closed | Some Contour transforms do not work in orthographic mode | Ready for Testing bug | If you attempt to grab or rotate a symeterized loop (loop cutting across the symmetry line) in Orthographic view, from Side view, then the transform does nothing from. No errors thrown.
Rotating works from front/back view, but not side. If you rotate far enough then the loop eventually snaps to a distorted angle.
Grab does not work at all in ortho on symmetry loops.
Everything works as expected in perspective.
Video:
https://share.getcloudapp.com/jkunL8WQ | 1.0 | Some Contour transforms do not work in orthographic mode - If you attempt to grab or rotate a symeterized loop (loop cutting across the symmetry line) in Orthographic view, from Side view, then the transform does nothing from. No errors thrown.
Rotating works from front/back view, but not side. If you rotate far enough then the loop eventually snaps to a distorted angle.
Grab does not work at all in ortho on symmetry loops.
Everything works as expected in perspective.
Video:
https://share.getcloudapp.com/jkunL8WQ | non_process | some contour transforms do not work in orthographic mode if you attempt to grab or rotate a symeterized loop loop cutting across the symmetry line in orthographic view from side view then the transform does nothing from no errors thrown rotating works from front back view but not side if you rotate far enough then the loop eventually snaps to a distorted angle grab does not work at all in ortho on symmetry loops everything works as expected in perspective video | 0 |
1,912 | 4,750,541,188 | IssuesEvent | 2016-10-22 11:51:08 | kerubistan/kerub | https://api.github.com/repos/kerubistan/kerub | opened | virtual network expectations | component:data processing component:scheduling enhancement priority: normal | some expectations to restrict the use of virtual networks, ideas:
* location requirements (e.g. not in given countries)
* technology requirements
* site features: only on equipment in locations with armed guards, etc... | 1.0 | virtual network expectations - some expectations to restrict the use of virtual networks, ideas:
* location requirements (e.g. not in given countries)
* technology requirements
* site features: only on equipment in locations with armed guards, etc... | process | virtual network expectations some expectations to restrict the use of virtual networks ideas location requirements e g not in given countries technology requirements site features only on equipment in locations with armed guards etc | 1 |
82,610 | 3,617,296,983 | IssuesEvent | 2016-02-08 01:47:30 | pufexi/multiorder | https://api.github.com/repos/pufexi/multiorder | opened | Vygenerovat XML faktury | high priority | \app\templates\Homepage\invexport.latte jsem opravil, predim se generovali XML faktury spatne, bylo prazdne date, dal jsem tam accountingdate, takze jsem vsechny XML faktury ve slozce smazal, ale nyni nevim jak je hromadne vygenerovat, hodne to hori | 1.0 | Vygenerovat XML faktury - \app\templates\Homepage\invexport.latte jsem opravil, predim se generovali XML faktury spatne, bylo prazdne date, dal jsem tam accountingdate, takze jsem vsechny XML faktury ve slozce smazal, ale nyni nevim jak je hromadne vygenerovat, hodne to hori | non_process | vygenerovat xml faktury app templates homepage invexport latte jsem opravil predim se generovali xml faktury spatne bylo prazdne date dal jsem tam accountingdate takze jsem vsechny xml faktury ve slozce smazal ale nyni nevim jak je hromadne vygenerovat hodne to hori | 0 |
38,724 | 6,690,885,502 | IssuesEvent | 2017-10-09 10:56:05 | goncalotomas/FMKe | https://api.github.com/repos/goncalotomas/FMKe | opened | Users don't know they have to install rebar3 to use FMKe | documentation | We stopped bundling the rebar3 binary with the repository since it was constantly needing updates, but now there is no mention in the README that FMKe requires Erlang >20 and rebar3 | 1.0 | Users don't know they have to install rebar3 to use FMKe - We stopped bundling the rebar3 binary with the repository since it was constantly needing updates, but now there is no mention in the README that FMKe requires Erlang >20 and rebar3 | non_process | users don t know they have to install to use fmke we stopped bundling the binary with the repository since it was constantly needing updates but now there is no mention in the readme that fmke requires erlang and | 0 |
121,540 | 12,128,655,480 | IssuesEvent | 2020-04-22 20:54:33 | SteveMacenski/ros2_ouster_drivers | https://api.github.com/repos/SteveMacenski/ros2_ouster_drivers | closed | Document components use / test | documentation | e.g.
component manager for dynamic loading (also may be done through launch)
$ros2 run rclcpp_components component_container
load this component
$ros2 component load /ComponentManager ros2_ouster ros2_ouster::OS1Driver
transition to configuring
ros2 lifecycle set OusterDriver 1
transition to active
ros2 lifecycle set OusterDriver 3
** parameters | 1.0 | Document components use / test - e.g.
component manager for dynamic loading (also may be done through launch)
$ros2 run rclcpp_components component_container
load this component
$ros2 component load /ComponentManager ros2_ouster ros2_ouster::OS1Driver
transition to configuring
ros2 lifecycle set OusterDriver 1
transition to active
ros2 lifecycle set OusterDriver 3
** parameters | non_process | document components use test e g component manager for dynamic loading also may be done through launch run rclcpp components component container load this component component load componentmanager ouster ouster transition to configuring lifecycle set ousterdriver transition to active lifecycle set ousterdriver parameters | 0 |
2,234 | 5,084,903,721 | IssuesEvent | 2016-12-30 08:24:55 | f2etw/f2e-notes | https://api.github.com/repos/f2etw/f2e-notes | opened | http://lebab.io/ ,Lebab 將 ES5 轉為 ES6 語法 | js npm online tool preprocessor | Lebab
http://lebab.io/
> Turn your ES5 code into readable ES6 (sugar-syntax). It does the opposite of what Babel does.
> github repo: https://github.com/lebab/lebab
將舊的 ES5 語法自動轉換為 ES6 的語法
可輔助學習新語法結構~
可以線上玩玩看
http://lebab.io/try-it | 1.0 | http://lebab.io/ ,Lebab 將 ES5 轉為 ES6 語法 - Lebab
http://lebab.io/
> Turn your ES5 code into readable ES6 (sugar-syntax). It does the opposite of what Babel does.
> github repo: https://github.com/lebab/lebab
將舊的 ES5 語法自動轉換為 ES6 的語法
可輔助學習新語法結構~
可以線上玩玩看
http://lebab.io/try-it | process | ,lebab 將 轉為 語法 lebab turn your code into readable sugar syntax it does the opposite of what babel does github repo 將舊的 語法自動轉換為 的語法 可輔助學習新語法結構~ 可以線上玩玩看 | 1 |
12,630 | 15,016,217,617 | IssuesEvent | 2021-02-01 09:18:33 | micado-eu/ngo_application | https://api.github.com/repos/micado-eu/ngo_application | closed | Provide annotation possibility for NGO users | guided processes wontfix | Rethink and redesign the comment function:
NGO users should be able to provide concrete annotations in the steps themselves (not only basic comments), as this is an USP for the app | 1.0 | Provide annotation possibility for NGO users - Rethink and redesign the comment function:
NGO users should be able to provide concrete annotations in the steps themselves (not only basic comments), as this is an USP for the app | process | provide annotation possibility for ngo users rethink and redesign the comment function ngo users should be able to provide concrete annotations in the steps themselves not only basic comments as this is an usp for the app | 1 |
1,617 | 4,231,157,168 | IssuesEvent | 2016-07-04 14:51:42 | nodejs/node | https://api.github.com/repos/nodejs/node | closed | process.argv[0] contains full node.exe path on Windows, docs say it must be just "node" | doc docs-requested process | * **Version**: 4.4.6.0
* **Platform**: Windows 7 Enterprise SP1
* **Subsystem**: no idea
Documentation https://nodejs.org/api/process.html#process_process_argv says `process.argv[0]` must be `node` but I run this code on Windows
console.log('argv[0]: ' + process.argv[0]);
and it outputs this
argv[0]: C:\Program Files (x86)\nodejs\node.exe
which kind of isn't the same as `node`. It was `node` on some earlier versions - perhaps a couple years ago.
Here's how I run my code: I started `cmd.exe`, `cd`d into `C:\Program Files (x86)\nodejs`, so it's now the current path and I run
node fullPathToJsFile
| 1.0 | process.argv[0] contains full node.exe path on Windows, docs say it must be just "node" - * **Version**: 4.4.6.0
* **Platform**: Windows 7 Enterprise SP1
* **Subsystem**: no idea
Documentation https://nodejs.org/api/process.html#process_process_argv says `process.argv[0]` must be `node` but I run this code on Windows
console.log('argv[0]: ' + process.argv[0]);
and it outputs this
argv[0]: C:\Program Files (x86)\nodejs\node.exe
which kind of isn't the same as `node`. It was `node` on some earlier versions - perhaps a couple years ago.
Here's how I run my code: I started `cmd.exe`, `cd`d into `C:\Program Files (x86)\nodejs`, so it's now the current path and I run
node fullPathToJsFile
| process | process argv contains full node exe path on windows docs say it must be just node version platform windows enterprise subsystem no idea documentation says process argv must be node but i run this code on windows console log argv process argv and it outputs this argv c program files nodejs node exe which kind of isn t the same as node it was node on some earlier versions perhaps a couple years ago here s how i run my code i started cmd exe cd d into c program files nodejs so it s now the current path and i run node fullpathtojsfile | 1 |
654 | 3,125,609,434 | IssuesEvent | 2015-09-08 01:40:49 | nodejs/node | https://api.github.com/repos/nodejs/node | closed | Yet another segmentation fault in node-v4.0.0-rc.1 | child_process confirmed-bug v8 | ```
var cp = require('child_process');
cp.spawn('/bin/ls', [], {
stdio: [
process.stdin,
process.stdout,
process.stderr,
]
});
```

| 1.0 | Yet another segmentation fault in node-v4.0.0-rc.1 - ```
var cp = require('child_process');
cp.spawn('/bin/ls', [], {
stdio: [
process.stdin,
process.stdout,
process.stderr,
]
});
```

| process | yet another segmentation fault in node rc var cp require child process cp spawn bin ls stdio process stdin process stdout process stderr | 1 |
64,714 | 26,856,483,095 | IssuesEvent | 2023-02-03 14:59:25 | coq/coq | https://api.github.com/repos/coq/coq | closed | .vo files vary from build order | Coq Consortium support services kind: bug | #### Description of the problem
While working on reproducible builds for openSUSE, I found that
In addition to #11227 , there are variations in .vo files
that go away when I build on a filesystem that has deterministic `readdir` order **and** build with `make` instead of `make -j4` - so that the builds happen in deterministic order.
```diff
/usr/lib64/coq/theories/FSets/FSetList.vo differs at offset '9059' (data)
--- /tmp/tmp.KtRUkpybQT/old.yPM 2019-12-02 10:39:36.148042421 +0000
+++ /tmp/tmp.KtRUkpybQT/new.vUR 2019-12-02 10:39:36.148042421 +0000
@@ -1,11 +1,11 @@
00002340 8b e6 08 d0 e1 81 57 ec a0 a0 28 4d 53 65 74 4c |......W...(MSetL|
00002350 69 73 74 a0 25 4d 53 65 74 73 a0 23 43 6f 71 40 |ist.%MSets.#Coq@|
-00002360 90 30 35 20 c8 a4 e3 1b b6 aa ca 98 d2 da c7 7f |.05 ............|
-00002370 f0 c1 a0 a0 29 4f 72 64 65 72 73 41 6c 74 a0 2a |....)OrdersAlt.*|
+00002360 90 30 b5 b2 78 25 9a ee e4 62 59 2b 46 34 9c 6c |.0..x%...bY+F4.l|
+00002370 48 e4 a0 a0 29 4f 72 64 65 72 73 41 6c 74 a0 2a |H...)OrdersAlt.*|
00002380 53 74 72 75 63 74 75 72 65 73 a0 23 43 6f 71 40 |Structures.#Coq@|
00002390 90 30 5c 91 d4 af b1 94 2f 7a 6f 13 17 a2 75 69 |.0\...../zo...ui|
-000023a0 ed 79 b1 cf 42 6c ca 2c 09 6f 38 78 f9 cc b9 cc |.y..Bl.,.o8x....|
-000023b0 d4 7a 00 02 af 1c 84 95 a6 be 00 02 8b 52 00 00 |.z...........R..|
+000023a0 ed 79 6e 87 f6 25 58 27 a4 20 92 36 b7 1c 3b 0e |.yn..%X'. .6..;.|
+000023b0 f3 06 00 02 af 1c 84 95 a6 be 00 02 8b 52 00 00 |.............R..|
```
https://github.com/bmwiedemann/openSUSE/blob/master/packages/c/coq/coq.spec#L77 has the details of how we build.
#### Coq Version
8.9.1 | 1.0 | .vo files vary from build order - #### Description of the problem
While working on reproducible builds for openSUSE, I found that
In addition to #11227 , there are variations in .vo files
that go away when I build on a filesystem that has deterministic `readdir` order **and** build with `make` instead of `make -j4` - so that the builds happen in deterministic order.
```diff
/usr/lib64/coq/theories/FSets/FSetList.vo differs at offset '9059' (data)
--- /tmp/tmp.KtRUkpybQT/old.yPM 2019-12-02 10:39:36.148042421 +0000
+++ /tmp/tmp.KtRUkpybQT/new.vUR 2019-12-02 10:39:36.148042421 +0000
@@ -1,11 +1,11 @@
00002340 8b e6 08 d0 e1 81 57 ec a0 a0 28 4d 53 65 74 4c |......W...(MSetL|
00002350 69 73 74 a0 25 4d 53 65 74 73 a0 23 43 6f 71 40 |ist.%MSets.#Coq@|
-00002360 90 30 35 20 c8 a4 e3 1b b6 aa ca 98 d2 da c7 7f |.05 ............|
-00002370 f0 c1 a0 a0 29 4f 72 64 65 72 73 41 6c 74 a0 2a |....)OrdersAlt.*|
+00002360 90 30 b5 b2 78 25 9a ee e4 62 59 2b 46 34 9c 6c |.0..x%...bY+F4.l|
+00002370 48 e4 a0 a0 29 4f 72 64 65 72 73 41 6c 74 a0 2a |H...)OrdersAlt.*|
00002380 53 74 72 75 63 74 75 72 65 73 a0 23 43 6f 71 40 |Structures.#Coq@|
00002390 90 30 5c 91 d4 af b1 94 2f 7a 6f 13 17 a2 75 69 |.0\...../zo...ui|
-000023a0 ed 79 b1 cf 42 6c ca 2c 09 6f 38 78 f9 cc b9 cc |.y..Bl.,.o8x....|
-000023b0 d4 7a 00 02 af 1c 84 95 a6 be 00 02 8b 52 00 00 |.z...........R..|
+000023a0 ed 79 6e 87 f6 25 58 27 a4 20 92 36 b7 1c 3b 0e |.yn..%X'. .6..;.|
+000023b0 f3 06 00 02 af 1c 84 95 a6 be 00 02 8b 52 00 00 |.............R..|
```
https://github.com/bmwiedemann/openSUSE/blob/master/packages/c/coq/coq.spec#L77 has the details of how we build.
#### Coq Version
8.9.1 | non_process | vo files vary from build order description of the problem while working on reproducible builds for opensuse i found that in addition to there are variations in vo files that go away when i build on a filesystem that has deterministic readdir order and build with make instead of make so that the builds happen in deterministic order diff usr coq theories fsets fsetlist vo differs at offset data tmp tmp ktrukpybqt old ypm tmp tmp ktrukpybqt new vur ec w msetl ist msets coq aa ca da ordersalt ee x by l h ordersalt structures coq af zo ui ed cf ca cc cc y bl af be z r ed yn x af be r has the details of how we build coq version | 0 |
13,332 | 15,790,807,138 | IssuesEvent | 2021-04-02 02:34:34 | googlefonts/noto-fonts | https://api.github.com/repos/googlefonts/noto-fonts | closed | Complete set of script info and testing web pages | Noto-Process-Issue | https://motaitalic.github.io/devanagari-documentation/languages/dogri/dogri.html is an example of excellent background information on the Dogri language written with Devanagari.
I'd love to see such script info and testing web pages available for all languages supported by Noto. | 1.0 | Complete set of script info and testing web pages - https://motaitalic.github.io/devanagari-documentation/languages/dogri/dogri.html is an example of excellent background information on the Dogri language written with Devanagari.
I'd love to see such script info and testing web pages available for all languages supported by Noto. | process | complete set of script info and testing web pages is an example of excellent background information on the dogri language written with devanagari i d love to see such script info and testing web pages available for all languages supported by noto | 1 |
183,415 | 6,688,547,408 | IssuesEvent | 2017-10-08 15:59:57 | OperationCode/operationcode_bot | https://api.github.com/repos/OperationCode/operationcode_bot | closed | Change tests to be less reliant on copy | hacktoberfest Needs: Engineer Priority: High | # Feature
## Why is this feature being added?
Any copy changes completely break the tests on operationcode_bot. It should check to see that messages are sent, not that the specific wording is exactly the same.
## What should your feature do?
* `test_it_sends_a_message_to_the_user_if_an_env_var_is_set` just checks to see if a message is sent to the user.
* `test_it_welcomes_the_user_on_new_user_join` just checks to see if a message is sent to the user. | 1.0 | Change tests to be less reliant on copy - # Feature
## Why is this feature being added?
Any copy changes completely break the tests on operationcode_bot. It should check to see that messages are sent, not that the specific wording is exactly the same.
## What should your feature do?
* `test_it_sends_a_message_to_the_user_if_an_env_var_is_set` just checks to see if a message is sent to the user.
* `test_it_welcomes_the_user_on_new_user_join` just checks to see if a message is sent to the user. | non_process | change tests to be less reliant on copy feature why is this feature being added any copy changes completely break the tests on operationcode bot it should check to see that messages are sent not that the specific wording is exactly the same what should your feature do test it sends a message to the user if an env var is set just checks to see if a message is sent to the user test it welcomes the user on new user join just checks to see if a message is sent to the user | 0 |
267,589 | 23,307,680,182 | IssuesEvent | 2022-08-08 04:08:40 | vgstation-coders/vgstation13 | https://api.github.com/repos/vgstation-coders/vgstation13 | closed | Supply Ordering Console Not Loading | Needs Moar Testing Needs Pics / Needs Moar Info | (WEB REPORT BY: germansanta REMOTE: 206.221.180.138:7777)
# Supply Ordering Console Screen always shows up blank. Despite using Fix NanoUI, Clicking again, or even restarting byond and computer.
May be a resource bug. Going to reinstall for fix. | 1.0 | Supply Ordering Console Not Loading - (WEB REPORT BY: germansanta REMOTE: 206.221.180.138:7777)
# Supply Ordering Console Screen always shows up blank. Despite using Fix NanoUI, Clicking again, or even restarting byond and computer.
May be a resource bug. Going to reinstall for fix. | non_process | supply ordering console not loading web report by germansanta remote supply ordering console screen always shows up blank despite using fix nanoui clicking again or even restarting byond and computer may be a resource bug going to reinstall for fix | 0 |
20,705 | 27,392,655,431 | IssuesEvent | 2023-02-28 17:16:41 | MicrosoftDocs/azure-docs | https://api.github.com/repos/MicrosoftDocs/azure-docs | closed | Is GitHub Enterprise Cloud a supported Source Control type for Automation feature Source Control? | automation/svc triaged cxp product-feedback process-automation/subsvc Pri2 | Hello,
Is source control integration with a repository from GitHub Enterprise Cloud supported in the Source Control feature of Automation or is only GitHub explicitly supported, not GitHub Enterprise Cloud?
I do see that we once did support it but now the link to the doc is dead: https://docs.microsoft.com/en-us/azure/automation/automation-scenario-source-control-integration-with-github-ent. I have a Cx that is trying to integrate with the SaaS (GHEC) product but cannot save the connection. Was wondering if it's because we don't support it.
---
#### Document Details
⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.*
* ID: 83c90e64-b615-711f-a53d-fc76606e2ecd
* Version Independent ID: 2d164036-6886-4440-50f7-369f99f41cea
* Content: [Use source control integration in Azure Automation](https://learn.microsoft.com/en-us/azure/automation/source-control-integration#prerequisites)
* Content Source: [articles/automation/source-control-integration.md](https://github.com/MicrosoftDocs/azure-docs/blob/main/articles/automation/source-control-integration.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @SnehaSudhirG
* Microsoft Alias: **sudhirsneha** | 1.0 | Is GitHub Enterprise Cloud a supported Source Control type for Automation feature Source Control? - Hello,
Is source control integration with a repository from GitHub Enterprise Cloud supported in the Source Control feature of Automation or is only GitHub explicitly supported, not GitHub Enterprise Cloud?
I do see that we once did support it but now the link to the doc is dead: https://docs.microsoft.com/en-us/azure/automation/automation-scenario-source-control-integration-with-github-ent. I have a Cx that is trying to integrate with the SaaS (GHEC) product but cannot save the connection. Was wondering if it's because we don't support it.
---
#### Document Details
⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.*
* ID: 83c90e64-b615-711f-a53d-fc76606e2ecd
* Version Independent ID: 2d164036-6886-4440-50f7-369f99f41cea
* Content: [Use source control integration in Azure Automation](https://learn.microsoft.com/en-us/azure/automation/source-control-integration#prerequisites)
* Content Source: [articles/automation/source-control-integration.md](https://github.com/MicrosoftDocs/azure-docs/blob/main/articles/automation/source-control-integration.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @SnehaSudhirG
* Microsoft Alias: **sudhirsneha** | process | is github enterprise cloud a supported source control type for automation feature source control hello is source control integration with a repository from github enterprise cloud supported in the source control feature of automation or is only github explicitly supported not github enterprise cloud i do see that we once did support it but now the link to the doc is dead i have a cx that is trying to integrate with the saas ghec product but cannot save the connection was wondering if it s because we don t support it document details ⚠ do not edit this section it is required for learn microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login snehasudhirg microsoft alias sudhirsneha | 1 |
7,544 | 10,659,983,680 | IssuesEvent | 2019-10-18 08:57:23 | prisma/photonjs | https://api.github.com/repos/prisma/photonjs | closed | `count` queries do not invoke `beforeRequest` hook | bug/0-needs-info kind/bug process/candidate | The `beforeRequest` hook that Studio configures for requests is not called for a query like so: `photon.users.count()` | 1.0 | `count` queries do not invoke `beforeRequest` hook - The `beforeRequest` hook that Studio configures for requests is not called for a query like so: `photon.users.count()` | process | count queries do not invoke beforerequest hook the beforerequest hook that studio configures for requests is not called for a query like so photon users count | 1 |
746,874 | 26,049,175,066 | IssuesEvent | 2022-12-22 16:56:22 | yugabyte/yugabyte-db | https://api.github.com/repos/yugabyte/yugabyte-db | reopened | [YSQL] PANIC: ERRORDATA_STACK_SIZE exceeded happens commonly | kind/bug area/ysql priority/high | Jira Link: [DB-2701](https://yugabyte.atlassian.net/browse/DB-2701)
### Description
Happens with [LST](https://github.com/yugabyte/yb-long-system-test/) on c28ce4c7f36d577fd18a091eb366076c42f1c548 (Release build), server run with `bin/yb-ctl --replication_factor 3 create --tserver_flags=ysql_enable_packed_row=true,ysql_packed_row_size_limit=1700 --master_flags=ysql_enable_packed_row=true,ysql_packed_row_size_limit=1700`:
```
2022-06-21 10:15:55,098 MainThread INFO
2022-06-21 10:15:55,098 MainThread INFO --------------------------------------------------------------------------------
2022-06-21 10:15:55,098 MainThread INFO Running Long System Test 0.1
2022-06-21 10:15:55,098 MainThread INFO --------------------------------------------------------------------------------
2022-06-21 10:15:55,098 MainThread INFO
2022-06-21 10:15:55,105 MainThread INFO Reproduce with: git checkout 6fdda644 && ./long_system_test.py --nodes=127.0.0.1:5433,127.0.0.2:5433,127.0.0.3:5433 --threads=10 --runtime=0 --complexity=full --max-columns=10 --seed=032216
2022-06-21 10:15:55,413 MainThread INFO Database version: PostgreSQL 11.2-YB-2.15.1.0-b0 on x86_64-pc-linux-gnu, compiled by clang version 12.0.1 (https://github.com/yugabyte/llvm-project.git bdb147e675d8c87cee72cc1f87c4b82855977d94), 64-bit
2022-06-21 10:15:55,416 MainThread INFO Creating tables for database db_lst_032216
2022-06-21 10:16:15,948 MainThread INFO Starting worker_0: RandomSelectAction, SetConfigAction
2022-06-21 10:16:15,950 MainThread INFO Starting worker_1: CreateIndexAction, DropIndexAction, SetConfigAction, AddColumnAction
2022-06-21 10:16:15,950 MainThread INFO Starting worker_2: CreateIndexAction, DropIndexAction, SetConfigAction, AddColumnAction
2022-06-21 10:16:15,952 MainThread INFO Starting worker_3: CreateIndexAction, DropIndexAction, SetConfigAction, AddColumnAction
2022-06-21 10:16:15,952 MainThread INFO Starting worker_4: RandomSelectAction, SetConfigAction
2022-06-21 10:16:15,953 MainThread INFO Starting worker_5: SingleInsertAction, SingleUpdateAction, SingleDeleteAction, BulkInsertAction, BulkUpdateAction, SetConfigAction
2022-06-21 10:16:15,953 MainThread INFO Starting worker_6: RandomSelectAction, SetConfigAction
2022-06-21 10:16:15,954 MainThread INFO Starting worker_7: CreateIndexAction, DropIndexAction, SetConfigAction, AddColumnAction
2022-06-21 10:16:15,955 MainThread INFO Starting worker_8: SingleInsertAction, SingleUpdateAction, SingleDeleteAction, BulkInsertAction, BulkUpdateAction, SetConfigAction
2022-06-21 10:16:15,955 MainThread INFO Starting worker_9: RandomSelectAction, SetConfigAction
2022-06-21 10:16:25,967 MainThread INFO Worker queries/s: [010.6][001.4][000.9][001.2][012.6][014.2][012.2][000.9][013.6][016.1]
2022-06-21 10:16:29,233 worker_5 ERROR Unexpected query failure: OperationalError
Query: EXPLAIN (ANALYZE TRUE, VERBOSE TRUE, COSTS TRUE, BUFFERS FALSE, TIMING FALSE, SUMMARY TRUE, FORMAT YAML)
DELETE FROM tg0_4 WHERE TRUE;
values: None
runtime: 2022-06-21 10:16:28.109 - 2022-06-21 10:16:29.233
supports explain: False
supports rollback: False
affected rows: None
Action: SingleDeleteAction
Error class: OperationalError
Error code: None
Error message: ERROR: cache lookup failed for attribute -8 of relation 16464
ERROR: Operation expired: Transaction c562d5b7-7d23-4ab3-bf42-99422eaf599f expired or aborted by a conflict: 40001
ERROR: Operation expired: Transaction c562d5b7-7d23-4ab3-bf42-99422eaf599f expired or aborted by a conflict: 40001
ERROR: Operation expired: Transaction c562d5b7-7d23-4ab3-bf42-99422eaf599f expired or aborted by a conflict: 40001
ERROR: Operation expired: Transaction c562d5b7-7d23-4ab3-bf42-99422eaf599f expired or aborted by a conflict: 40001
PANIC: ERRORDATA_STACK_SIZE exceeded
server closed the connection unexpectedly
This probably means the server terminated abnormally
before or while processing the request.
Transaction isolation level: default
DB Node: host: 127.0.0.3, port: 5433
DB Backend PID: 162903
```
Failures like this are happening after a few seconds now on each run. There is always a `PANIC: ERRORDATA_STACK_SIZE exceeded` failure. I'll upload 4 runs with this:
[lst_12972.zip](https://github.com/yugabyte/yugabyte-db/files/8949917/lst_12972.zip)
There are multiple similar bugs open like https://github.com/yugabyte/yugabyte-db/issues/11487, but I haven't seen this in older code states (b0dcb81db7f0793ab7e6278ec1f2e34d344d3a7c was still fine). | 1.0 | [YSQL] PANIC: ERRORDATA_STACK_SIZE exceeded happens commonly - Jira Link: [DB-2701](https://yugabyte.atlassian.net/browse/DB-2701)
### Description
Happens with [LST](https://github.com/yugabyte/yb-long-system-test/) on c28ce4c7f36d577fd18a091eb366076c42f1c548 (Release build), server run with `bin/yb-ctl --replication_factor 3 create --tserver_flags=ysql_enable_packed_row=true,ysql_packed_row_size_limit=1700 --master_flags=ysql_enable_packed_row=true,ysql_packed_row_size_limit=1700`:
```
2022-06-21 10:15:55,098 MainThread INFO
2022-06-21 10:15:55,098 MainThread INFO --------------------------------------------------------------------------------
2022-06-21 10:15:55,098 MainThread INFO Running Long System Test 0.1
2022-06-21 10:15:55,098 MainThread INFO --------------------------------------------------------------------------------
2022-06-21 10:15:55,098 MainThread INFO
2022-06-21 10:15:55,105 MainThread INFO Reproduce with: git checkout 6fdda644 && ./long_system_test.py --nodes=127.0.0.1:5433,127.0.0.2:5433,127.0.0.3:5433 --threads=10 --runtime=0 --complexity=full --max-columns=10 --seed=032216
2022-06-21 10:15:55,413 MainThread INFO Database version: PostgreSQL 11.2-YB-2.15.1.0-b0 on x86_64-pc-linux-gnu, compiled by clang version 12.0.1 (https://github.com/yugabyte/llvm-project.git bdb147e675d8c87cee72cc1f87c4b82855977d94), 64-bit
2022-06-21 10:15:55,416 MainThread INFO Creating tables for database db_lst_032216
2022-06-21 10:16:15,948 MainThread INFO Starting worker_0: RandomSelectAction, SetConfigAction
2022-06-21 10:16:15,950 MainThread INFO Starting worker_1: CreateIndexAction, DropIndexAction, SetConfigAction, AddColumnAction
2022-06-21 10:16:15,950 MainThread INFO Starting worker_2: CreateIndexAction, DropIndexAction, SetConfigAction, AddColumnAction
2022-06-21 10:16:15,952 MainThread INFO Starting worker_3: CreateIndexAction, DropIndexAction, SetConfigAction, AddColumnAction
2022-06-21 10:16:15,952 MainThread INFO Starting worker_4: RandomSelectAction, SetConfigAction
2022-06-21 10:16:15,953 MainThread INFO Starting worker_5: SingleInsertAction, SingleUpdateAction, SingleDeleteAction, BulkInsertAction, BulkUpdateAction, SetConfigAction
2022-06-21 10:16:15,953 MainThread INFO Starting worker_6: RandomSelectAction, SetConfigAction
2022-06-21 10:16:15,954 MainThread INFO Starting worker_7: CreateIndexAction, DropIndexAction, SetConfigAction, AddColumnAction
2022-06-21 10:16:15,955 MainThread INFO Starting worker_8: SingleInsertAction, SingleUpdateAction, SingleDeleteAction, BulkInsertAction, BulkUpdateAction, SetConfigAction
2022-06-21 10:16:15,955 MainThread INFO Starting worker_9: RandomSelectAction, SetConfigAction
2022-06-21 10:16:25,967 MainThread INFO Worker queries/s: [010.6][001.4][000.9][001.2][012.6][014.2][012.2][000.9][013.6][016.1]
2022-06-21 10:16:29,233 worker_5 ERROR Unexpected query failure: OperationalError
Query: EXPLAIN (ANALYZE TRUE, VERBOSE TRUE, COSTS TRUE, BUFFERS FALSE, TIMING FALSE, SUMMARY TRUE, FORMAT YAML)
DELETE FROM tg0_4 WHERE TRUE;
values: None
runtime: 2022-06-21 10:16:28.109 - 2022-06-21 10:16:29.233
supports explain: False
supports rollback: False
affected rows: None
Action: SingleDeleteAction
Error class: OperationalError
Error code: None
Error message: ERROR: cache lookup failed for attribute -8 of relation 16464
ERROR: Operation expired: Transaction c562d5b7-7d23-4ab3-bf42-99422eaf599f expired or aborted by a conflict: 40001
ERROR: Operation expired: Transaction c562d5b7-7d23-4ab3-bf42-99422eaf599f expired or aborted by a conflict: 40001
ERROR: Operation expired: Transaction c562d5b7-7d23-4ab3-bf42-99422eaf599f expired or aborted by a conflict: 40001
ERROR: Operation expired: Transaction c562d5b7-7d23-4ab3-bf42-99422eaf599f expired or aborted by a conflict: 40001
PANIC: ERRORDATA_STACK_SIZE exceeded
server closed the connection unexpectedly
This probably means the server terminated abnormally
before or while processing the request.
Transaction isolation level: default
DB Node: host: 127.0.0.3, port: 5433
DB Backend PID: 162903
```
Failures like this are happening after a few seconds now on each run. There is always a `PANIC: ERRORDATA_STACK_SIZE exceeded` failure. I'll upload 4 runs with this:
[lst_12972.zip](https://github.com/yugabyte/yugabyte-db/files/8949917/lst_12972.zip)
There are multiple similar bugs open like https://github.com/yugabyte/yugabyte-db/issues/11487, but I haven't seen this in older code states (b0dcb81db7f0793ab7e6278ec1f2e34d344d3a7c was still fine). | non_process | panic errordata stack size exceeded happens commonly jira link description happens with on release build server run with bin yb ctl replication factor create tserver flags ysql enable packed row true ysql packed row size limit master flags ysql enable packed row true ysql packed row size limit mainthread info mainthread info mainthread info running long system test mainthread info mainthread info mainthread info reproduce with git checkout long system test py nodes threads runtime complexity full max columns seed mainthread info database version postgresql yb on pc linux gnu compiled by clang version bit mainthread info creating tables for database db lst mainthread info starting worker randomselectaction setconfigaction mainthread info starting worker createindexaction dropindexaction setconfigaction addcolumnaction mainthread info starting worker createindexaction dropindexaction setconfigaction addcolumnaction mainthread info starting worker createindexaction dropindexaction setconfigaction addcolumnaction mainthread info starting worker randomselectaction setconfigaction mainthread info starting worker singleinsertaction singleupdateaction singledeleteaction bulkinsertaction bulkupdateaction setconfigaction mainthread info starting worker randomselectaction setconfigaction mainthread info starting worker createindexaction dropindexaction setconfigaction addcolumnaction mainthread info starting worker singleinsertaction singleupdateaction singledeleteaction bulkinsertaction bulkupdateaction setconfigaction mainthread info starting worker randomselectaction setconfigaction mainthread info worker queries s worker error unexpected query failure operationalerror query explain analyze true verbose true costs true buffers false timing false summary true format yaml delete from where true values none runtime supports explain false supports rollback false affected rows none action singledeleteaction error class operationalerror error code none error message error cache lookup failed for attribute of relation error operation expired transaction expired or aborted by a conflict error operation expired transaction expired or aborted by a conflict error operation expired transaction expired or aborted by a conflict error operation expired transaction expired or aborted by a conflict panic errordata stack size exceeded server closed the connection unexpectedly this probably means the server terminated abnormally before or while processing the request transaction isolation level default db node host port db backend pid failures like this are happening after a few seconds now on each run there is always a panic errordata stack size exceeded failure i ll upload runs with this there are multiple similar bugs open like but i haven t seen this in older code states was still fine | 0 |
6,840 | 9,982,971,144 | IssuesEvent | 2019-07-10 11:12:12 | qgis/QGIS-Documentation | https://api.github.com/repos/qgis/QGIS-Documentation | closed | Move part of comments out of the Extending QgsProcessingalgorithm code bolck | Processing User Manual | ## Description
Imho, there's a lot of comment and explanation in the code block in [extending qgsprocessingalgorithm](https://docs.qgis.org/3.4/en/docs/user_manual/processing/scripts.html#extending-qgsprocessingalgorithm) section. If we could move part of this as simple text out of the code frame:
* the code will look shorter, focus on the steps, with readability improved
* explanation will be translatable, hence accessible to more people
Page URL: https://docs.qgis.org/3.4/en/docs/user_manual/processing/scripts.html#extending-qgsprocessingalgorithm
| 1.0 | Move part of comments out of the Extending QgsProcessingalgorithm code bolck - ## Description
Imho, there's a lot of comment and explanation in the code block in [extending qgsprocessingalgorithm](https://docs.qgis.org/3.4/en/docs/user_manual/processing/scripts.html#extending-qgsprocessingalgorithm) section. If we could move part of this as simple text out of the code frame:
* the code will look shorter, focus on the steps, with readability improved
* explanation will be translatable, hence accessible to more people
Page URL: https://docs.qgis.org/3.4/en/docs/user_manual/processing/scripts.html#extending-qgsprocessingalgorithm
| process | move part of comments out of the extending qgsprocessingalgorithm code bolck description imho there s a lot of comment and explanation in the code block in section if we could move part of this as simple text out of the code frame the code will look shorter focus on the steps with readability improved explanation will be translatable hence accessible to more people page url | 1 |
791,958 | 27,882,699,188 | IssuesEvent | 2023-03-21 20:45:31 | GoogleChrome/lighthouse | https://api.github.com/repos/GoogleChrome/lighthouse | closed | uses-responsive-images fails in snapshot but passes in navigation | needs-priority | found by @sofiayem
repro
1. https://resplendent-unicorn-332514.netlify.app/ in canary
2. open devtools > lighthouse panel
1. capture navigation. see that "Properly Size Images" is passing
2. capture snapshot. see that "Properly Size Images" is failing
On this page, the images are wayy too big so the audit should definitely be failing.
Not sure whats going on in navigation mode. I do see we have `uses-responsive-images` and `uses-responsive-images-snapshot`, so that explains part of whats going on. :)
#### navigation

#### snapshot

| 1.0 | uses-responsive-images fails in snapshot but passes in navigation - found by @sofiayem
repro
1. https://resplendent-unicorn-332514.netlify.app/ in canary
2. open devtools > lighthouse panel
1. capture navigation. see that "Properly Size Images" is passing
2. capture snapshot. see that "Properly Size Images" is failing
On this page, the images are wayy too big so the audit should definitely be failing.
Not sure whats going on in navigation mode. I do see we have `uses-responsive-images` and `uses-responsive-images-snapshot`, so that explains part of whats going on. :)
#### navigation

#### snapshot

| non_process | uses responsive images fails in snapshot but passes in navigation found by sofiayem repro in canary open devtools lighthouse panel capture navigation see that properly size images is passing capture snapshot see that properly size images is failing on this page the images are wayy too big so the audit should definitely be failing not sure whats going on in navigation mode i do see we have uses responsive images and uses responsive images snapshot so that explains part of whats going on navigation snapshot | 0 |
4,429 | 7,307,263,416 | IssuesEvent | 2018-02-28 01:51:06 | Great-Hill-Corporation/quickBlocks | https://api.github.com/repos/Great-Hill-Corporation/quickBlocks | closed | ethName option -s should hit on source where the same search with source won't | status-inprocess tools-ethName type-bug | This test:
ethName -s GHC
should hit, whereas this one should not
ethName GHC
The -s option searches the additional field 'source' which is not searched by default. Please add a test (or more). I modified the _source_ test cases to reflect this. Please confirm that they work as you would expect. | 1.0 | ethName option -s should hit on source where the same search with source won't - This test:
ethName -s GHC
should hit, whereas this one should not
ethName GHC
The -s option searches the additional field 'source' which is not searched by default. Please add a test (or more). I modified the _source_ test cases to reflect this. Please confirm that they work as you would expect. | process | ethname option s should hit on source where the same search with source won t this test ethname s ghc should hit whereas this one should not ethname ghc the s option searches the additional field source which is not searched by default please add a test or more i modified the source test cases to reflect this please confirm that they work as you would expect | 1 |
12,136 | 14,741,023,101 | IssuesEvent | 2021-01-07 09:59:03 | prisma/prisma | https://api.github.com/repos/prisma/prisma | closed | Add --non-interactive to @prisma/migrate | kind/feature process/candidate team/migrations topic: cli-migrate | ## Problem
I have a script that seeds the database with fake data. But before it does, it first checks that I'm not running this in production env/DB (😅) and after that, It tries to run `prisma migrate reset --preview-feature` like so:
```ts
const prismaBinary = join(__dirname, "..", "node_modules", ".bin", "prisma");
execSync(`${prismaBinary} migrate reset --preview-feature`, {
env: {
...process.env,
DATABASE_URL: config.DATABASE_URL,
},
});
```
The thing is, when it tries to run it, I get the following error:
```
Error: We detected that your environment is non-interactive. Running this command is not supported in this context.
Error: Command failed: /home/eliya/projects/screendow/packages/server/node_modules/.bin/prisma migrate reset --preview-feature
```
## Suggested solution
Add a flag `--non-interactive` or `--force`
## Alternatives
run the reset manually and only then run the seed script
## Additional context
<!-- Add any other context or screenshots about the feature request here. -->
None | 1.0 | Add --non-interactive to @prisma/migrate - ## Problem
I have a script that seeds the database with fake data. But before it does, it first checks that I'm not running this in production env/DB (😅) and after that, It tries to run `prisma migrate reset --preview-feature` like so:
```ts
const prismaBinary = join(__dirname, "..", "node_modules", ".bin", "prisma");
execSync(`${prismaBinary} migrate reset --preview-feature`, {
env: {
...process.env,
DATABASE_URL: config.DATABASE_URL,
},
});
```
The thing is, when it tries to run it, I get the following error:
```
Error: We detected that your environment is non-interactive. Running this command is not supported in this context.
Error: Command failed: /home/eliya/projects/screendow/packages/server/node_modules/.bin/prisma migrate reset --preview-feature
```
## Suggested solution
Add a flag `--non-interactive` or `--force`
## Alternatives
run the reset manually and only then run the seed script
## Additional context
<!-- Add any other context or screenshots about the feature request here. -->
None | process | add non interactive to prisma migrate problem i have a script that seeds the database with fake data but before it does it first checks that i m not running this in production env db 😅 and after that it tries to run prisma migrate reset preview feature like so ts const prismabinary join dirname node modules bin prisma execsync prismabinary migrate reset preview feature env process env database url config database url the thing is when it tries to run it i get the following error error we detected that your environment is non interactive running this command is not supported in this context error command failed home eliya projects screendow packages server node modules bin prisma migrate reset preview feature suggested solution add a flag non interactive or force alternatives run the reset manually and only then run the seed script additional context none | 1 |
4,134 | 2,711,378,885 | IssuesEvent | 2015-04-09 05:30:33 | jvalanen/diomber | https://api.github.com/repos/jvalanen/diomber | closed | Fix navbar with long diory names | 3 - Under testing / done UI / UX |
<!---
@huboard:{"order":44.0,"milestone_order":20,"custom_state":""}
-->
| 1.0 | Fix navbar with long diory names -
<!---
@huboard:{"order":44.0,"milestone_order":20,"custom_state":""}
-->
| non_process | fix navbar with long diory names huboard order milestone order custom state | 0 |
56,472 | 14,078,437,024 | IssuesEvent | 2020-11-04 13:34:20 | themagicalmammal/android_kernel_samsung_a5xelte | https://api.github.com/repos/themagicalmammal/android_kernel_samsung_a5xelte | opened | CVE-2016-9576 (High) detected in linuxv3.10 | security vulnerability | ## CVE-2016-9576 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv3.10</b></p></summary>
<p>
<p>Linux kernel source tree</p>
<p>Library home page: <a href=https://github.com/torvalds/linux.git>https://github.com/torvalds/linux.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/themagicalmammal/android_kernel_samsung_a5xelte/commit/738375813823cb33918102af385bdd5d82225e17">738375813823cb33918102af385bdd5d82225e17</a></p>
<p>Found in base branch: <b>cosmic-1.6-experimental</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>android_kernel_samsung_a5xelte/block/blk-map.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The blk_rq_map_user_iov function in block/blk-map.c in the Linux kernel before 4.8.14 does not properly restrict the type of iterator, which allows local users to read or write to arbitrary kernel memory locations or cause a denial of service (use-after-free) by leveraging access to a /dev/sg device.
<p>Publish Date: 2016-12-28
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-9576>CVE-2016-9576</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2016-9576">https://nvd.nist.gov/vuln/detail/CVE-2016-9576</a></p>
<p>Release Date: 2016-12-28</p>
<p>Fix Resolution: 4.8.14</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2016-9576 (High) detected in linuxv3.10 - ## CVE-2016-9576 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv3.10</b></p></summary>
<p>
<p>Linux kernel source tree</p>
<p>Library home page: <a href=https://github.com/torvalds/linux.git>https://github.com/torvalds/linux.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/themagicalmammal/android_kernel_samsung_a5xelte/commit/738375813823cb33918102af385bdd5d82225e17">738375813823cb33918102af385bdd5d82225e17</a></p>
<p>Found in base branch: <b>cosmic-1.6-experimental</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>android_kernel_samsung_a5xelte/block/blk-map.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The blk_rq_map_user_iov function in block/blk-map.c in the Linux kernel before 4.8.14 does not properly restrict the type of iterator, which allows local users to read or write to arbitrary kernel memory locations or cause a denial of service (use-after-free) by leveraging access to a /dev/sg device.
<p>Publish Date: 2016-12-28
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-9576>CVE-2016-9576</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2016-9576">https://nvd.nist.gov/vuln/detail/CVE-2016-9576</a></p>
<p>Release Date: 2016-12-28</p>
<p>Fix Resolution: 4.8.14</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_process | cve high detected in cve high severity vulnerability vulnerable library linux kernel source tree library home page a href found in head commit a href found in base branch cosmic experimental vulnerable source files android kernel samsung block blk map c vulnerability details the blk rq map user iov function in block blk map c in the linux kernel before does not properly restrict the type of iterator which allows local users to read or write to arbitrary kernel memory locations or cause a denial of service use after free by leveraging access to a dev sg device publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
87,034 | 3,736,325,020 | IssuesEvent | 2016-03-08 15:36:55 | OCHA-DAP/hdx-ckan | https://api.github.com/repos/OCHA-DAP/hdx-ckan | closed | New contribute: change default text for the resource note field | New Contribute Flow Priority-Low | To make the distinction between the description of the dataset and the description of the resource, I suggest we change the default text from
`Add description`
to
`Add a note about this resource`
This also will align better with the metadata guidance page.

| 1.0 | New contribute: change default text for the resource note field - To make the distinction between the description of the dataset and the description of the resource, I suggest we change the default text from
`Add description`
to
`Add a note about this resource`
This also will align better with the metadata guidance page.

| non_process | new contribute change default text for the resource note field to make the distinction between the description of the dataset and the description of the resource i suggest we change the default text from add description to add a note about this resource this also will align better with the metadata guidance page | 0 |
20,785 | 27,523,699,403 | IssuesEvent | 2023-03-06 16:34:43 | cse442-at-ub/project_s23-cinco | https://api.github.com/repos/cse442-at-ub/project_s23-cinco | closed | Update a readme on GitHub that explains organization of Repo | Processing Task Sprint 2 | **Task Tests**
*Test 1*
1) Go to url: https://github.com/cse442-at-ub/project_s23-cinco/tree/dev
2) Verify that you can see a description of the project below the files
3) See if the Readme that's displayed talks about the src, style, and image folder | 1.0 | Update a readme on GitHub that explains organization of Repo - **Task Tests**
*Test 1*
1) Go to url: https://github.com/cse442-at-ub/project_s23-cinco/tree/dev
2) Verify that you can see a description of the project below the files
3) See if the Readme that's displayed talks about the src, style, and image folder | process | update a readme on github that explains organization of repo task tests test go to url verify that you can see a description of the project below the files see if the readme that s displayed talks about the src style and image folder | 1 |
19,897 | 26,347,840,093 | IssuesEvent | 2023-01-11 00:25:51 | qgis/QGIS | https://api.github.com/repos/qgis/QGIS | closed | Distance to nearest hub - No destination hubs -> IndexError | Processing Bug | ### What is the bug or the crash?
When running the algorithms `Distance to nearest hub (points)` or `Distance to nearest hub (line to hub)` with the parameter `Destination hubs layer` being an layer without any features the algorithm crashes with the following error:
```
Traceback (most recent call last):
File "my/path/QGIS32~1.2/apps/qgis/./python/plugins\processing\algs\qgis\HubDistanceLines.py", line 139, in processAlgorithm
ft = next(hub_source.getFeatures(QgsFeatureRequest().setFilterFid(neighbors[0]).setSubsetOfAttributes([fieldName], hub_source.fields()).setDestinationCrs(point_source.sourceCrs(), context.transformContext())))
IndexError: list index out of range
```
I don't know what the expected output should be in such a case but an option would be e.g. to display an error message which indicates that the algorithm cannot be run with a `Destination hubs layer` without any features.
I can write a PR to fix the problem when it is decided how the algoritm(s) should behave in this case.
### Steps to reproduce the issue
Run the processing algorithms `Distance to nearest hub (points)` or `Distance to nearest hub (line to hub)` with the parameter `Destination hubs layer` being an layer without any features.
### Versions
<!--StartFragment--><!DOCTYPE HTML PUBLIC "-//W3C//DTD HTML 4.0//EN" "http://www.w3.org/TR/REC-html40/strict.dtd">
<html><head><meta http-equiv="Content-Type" content="text/html; charset=utf-8" /></head><body>
QGIS version | 3.24.2-Tisler | QGIS code revision | 13c1a028
-- | -- | -- | --
Qt version | 5.15.2
Python version | 3.9.5
GDAL/OGR version | 3.4.2
PROJ version | 9.0.0
EPSG Registry database version | v10.054 (2022-02-13)
GEOS version | 3.10.2-CAPI-1.16.0
SQLite version | 3.38.1
PDAL version | 2.3.0
PostgreSQL client version | unknown
SpatiaLite version | 5.0.1
QWT version | 6.1.3
QScintilla2 version | 2.11.5
OS version | Windows 10 Version 2009
| | |
Active Python plugins
FreehandRasterGeoreferencer | 0.8.3
osminfo | 0.7.1
qgis-maptiler-plugin | 2.0.0
QuickOSM | 2.0.0
quick_map_services | 0.19.26
db_manager | 0.1.20
grassprovider | 2.12.99
MetaSearch | 0.3.6
processing | 2.12.99
sagaprovider | 2.12.99
</body></html><!--EndFragment-->
### Supported QGIS version
- [X] I'm running a supported QGIS version according to the roadmap.
### New profile
- [ ] I tried with a new QGIS profile
### Additional context
_No response_ | 1.0 | Distance to nearest hub - No destination hubs -> IndexError - ### What is the bug or the crash?
When running the algorithms `Distance to nearest hub (points)` or `Distance to nearest hub (line to hub)` with the parameter `Destination hubs layer` being an layer without any features the algorithm crashes with the following error:
```
Traceback (most recent call last):
File "my/path/QGIS32~1.2/apps/qgis/./python/plugins\processing\algs\qgis\HubDistanceLines.py", line 139, in processAlgorithm
ft = next(hub_source.getFeatures(QgsFeatureRequest().setFilterFid(neighbors[0]).setSubsetOfAttributes([fieldName], hub_source.fields()).setDestinationCrs(point_source.sourceCrs(), context.transformContext())))
IndexError: list index out of range
```
I don't know what the expected output should be in such a case but an option would be e.g. to display an error message which indicates that the algorithm cannot be run with a `Destination hubs layer` without any features.
I can write a PR to fix the problem when it is decided how the algoritm(s) should behave in this case.
### Steps to reproduce the issue
Run the processing algorithms `Distance to nearest hub (points)` or `Distance to nearest hub (line to hub)` with the parameter `Destination hubs layer` being an layer without any features.
### Versions
<!--StartFragment--><!DOCTYPE HTML PUBLIC "-//W3C//DTD HTML 4.0//EN" "http://www.w3.org/TR/REC-html40/strict.dtd">
<html><head><meta http-equiv="Content-Type" content="text/html; charset=utf-8" /></head><body>
QGIS version | 3.24.2-Tisler | QGIS code revision | 13c1a028
-- | -- | -- | --
Qt version | 5.15.2
Python version | 3.9.5
GDAL/OGR version | 3.4.2
PROJ version | 9.0.0
EPSG Registry database version | v10.054 (2022-02-13)
GEOS version | 3.10.2-CAPI-1.16.0
SQLite version | 3.38.1
PDAL version | 2.3.0
PostgreSQL client version | unknown
SpatiaLite version | 5.0.1
QWT version | 6.1.3
QScintilla2 version | 2.11.5
OS version | Windows 10 Version 2009
| | |
Active Python plugins
FreehandRasterGeoreferencer | 0.8.3
osminfo | 0.7.1
qgis-maptiler-plugin | 2.0.0
QuickOSM | 2.0.0
quick_map_services | 0.19.26
db_manager | 0.1.20
grassprovider | 2.12.99
MetaSearch | 0.3.6
processing | 2.12.99
sagaprovider | 2.12.99
</body></html><!--EndFragment-->
### Supported QGIS version
- [X] I'm running a supported QGIS version according to the roadmap.
### New profile
- [ ] I tried with a new QGIS profile
### Additional context
_No response_ | process | distance to nearest hub no destination hubs indexerror what is the bug or the crash when running the algorithms distance to nearest hub points or distance to nearest hub line to hub with the parameter destination hubs layer being an layer without any features the algorithm crashes with the following error traceback most recent call last file my path apps qgis python plugins processing algs qgis hubdistancelines py line in processalgorithm ft next hub source getfeatures qgsfeaturerequest setfilterfid neighbors setsubsetofattributes hub source fields setdestinationcrs point source sourcecrs context transformcontext indexerror list index out of range i don t know what the expected output should be in such a case but an option would be e g to display an error message which indicates that the algorithm cannot be run with a destination hubs layer without any features i can write a pr to fix the problem when it is decided how the algoritm s should behave in this case steps to reproduce the issue run the processing algorithms distance to nearest hub points or distance to nearest hub line to hub with the parameter destination hubs layer being an layer without any features versions doctype html public dtd html en qgis version tisler qgis code revision qt version python version gdal ogr version proj version epsg registry database version geos version capi sqlite version pdal version postgresql client version unknown spatialite version qwt version version os version windows version active python plugins freehandrastergeoreferencer osminfo qgis maptiler plugin quickosm quick map services db manager grassprovider metasearch processing sagaprovider supported qgis version i m running a supported qgis version according to the roadmap new profile i tried with a new qgis profile additional context no response | 1 |
155,410 | 13,623,918,244 | IssuesEvent | 2020-09-24 07:13:22 | joadeveloped/dataStreaming | https://api.github.com/repos/joadeveloped/dataStreaming | closed | Definicion de Kpis | Epic documentation | Realizar una entrevista a supervisores con el siguiente formato para definir los kpis a mostrar.
Formato: [E-001](https://app.zenhub.com/files/298093456/eeee6282-6825-48c8-90a9-c3324581671a/download) | 1.0 | Definicion de Kpis - Realizar una entrevista a supervisores con el siguiente formato para definir los kpis a mostrar.
Formato: [E-001](https://app.zenhub.com/files/298093456/eeee6282-6825-48c8-90a9-c3324581671a/download) | non_process | definicion de kpis realizar una entrevista a supervisores con el siguiente formato para definir los kpis a mostrar formato | 0 |
7,490 | 10,579,775,320 | IssuesEvent | 2019-10-08 04:05:12 | fluent/fluent-bit | https://api.github.com/repos/fluent/fluent-bit | closed | Any plans for Docker container monitoring in plugins? | enhancement work-in-process | Hi,
I wonder if you guys have any plans to implement Docker container monitoring with Fluent Bit.
You already have node monitoring in plugins (cpu, mem, disk). It is also possible to use Fluent Bit instead of Fluentd as a Docker logging driver.
If you also had some In plugins for monitoring Docker containers (e.g. something similar to cAdvisor, Dockbeat, Docker stats) it would be possible to have most of the needed monitoring and log collection for Docker microservices/containers using Fluent Bit, Elasticsearch and Kibana.
Without the container monitoring, it looks like it's needed to also run other tools like cAdvisor to get a good setup for Docker/Docker Swarm deployments.
It looks like cAdvisor (with its web server and other goodies) will be much less efficient than Fluent Bit. Plus Fluent Bit is already somewhat in the same territory (cpu, mem and disk monitoring, Docker logging drivers).
It would be great to have only one extra tool per node (Fluent Bit, of course) to collect and forward all Docker-related info into ES.
Thus, I wonder if:
1. You have any plans to add Docker container monitoring as some kind of in plugins.
2. You might have some working current workarounds that allow full Docker container monitoring using just Fluent Bit.
Thank you. | 1.0 | Any plans for Docker container monitoring in plugins? - Hi,
I wonder if you guys have any plans to implement Docker container monitoring with Fluent Bit.
You already have node monitoring in plugins (cpu, mem, disk). It is also possible to use Fluent Bit instead of Fluentd as a Docker logging driver.
If you also had some In plugins for monitoring Docker containers (e.g. something similar to cAdvisor, Dockbeat, Docker stats) it would be possible to have most of the needed monitoring and log collection for Docker microservices/containers using Fluent Bit, Elasticsearch and Kibana.
Without the container monitoring, it looks like it's needed to also run other tools like cAdvisor to get a good setup for Docker/Docker Swarm deployments.
It looks like cAdvisor (with its web server and other goodies) will be much less efficient than Fluent Bit. Plus Fluent Bit is already somewhat in the same territory (cpu, mem and disk monitoring, Docker logging drivers).
It would be great to have only one extra tool per node (Fluent Bit, of course) to collect and forward all Docker-related info into ES.
Thus, I wonder if:
1. You have any plans to add Docker container monitoring as some kind of in plugins.
2. You might have some working current workarounds that allow full Docker container monitoring using just Fluent Bit.
Thank you. | process | any plans for docker container monitoring in plugins hi i wonder if you guys have any plans to implement docker container monitoring with fluent bit you already have node monitoring in plugins cpu mem disk it is also possible to use fluent bit instead of fluentd as a docker logging driver if you also had some in plugins for monitoring docker containers e g something similar to cadvisor dockbeat docker stats it would be possible to have most of the needed monitoring and log collection for docker microservices containers using fluent bit elasticsearch and kibana without the container monitoring it looks like it s needed to also run other tools like cadvisor to get a good setup for docker docker swarm deployments it looks like cadvisor with its web server and other goodies will be much less efficient than fluent bit plus fluent bit is already somewhat in the same territory cpu mem and disk monitoring docker logging drivers it would be great to have only one extra tool per node fluent bit of course to collect and forward all docker related info into es thus i wonder if you have any plans to add docker container monitoring as some kind of in plugins you might have some working current workarounds that allow full docker container monitoring using just fluent bit thank you | 1 |
265,866 | 23,206,584,210 | IssuesEvent | 2022-08-02 06:12:26 | longhorn/longhorn | https://api.github.com/repos/longhorn/longhorn | closed | [TEST] Add version information on each regression test build | kind/test | ## What's the test to develop? Please describe
Currently there's no way to know the version (commit id) of longhorn-manager and longhorn-engine used in each regression test build.
These information need to be revealed so we are able to know this build includes a commit or not.
| 1.0 | [TEST] Add version information on each regression test build - ## What's the test to develop? Please describe
Currently there's no way to know the version (commit id) of longhorn-manager and longhorn-engine used in each regression test build.
These information need to be revealed so we are able to know this build includes a commit or not.
| non_process | add version information on each regression test build what s the test to develop please describe currently there s no way to know the version commit id of longhorn manager and longhorn engine used in each regression test build these information need to be revealed so we are able to know this build includes a commit or not | 0 |
7,689 | 10,774,983,834 | IssuesEvent | 2019-11-03 11:02:55 | bisq-network/bisq | https://api.github.com/repos/bisq-network/bisq | closed | Show "Funds needed" with all the calculations at the "Create Offer" window | in:gui in:trade-process was:dropped | Based on the feedback from [this reddit post](https://www.reddit.com/r/bisq/comments/ba0xxy/feedback_show_funds_needed_with_all_the/), can we have "Funds needed" value showing up at the "Create offer" stage? | 1.0 | Show "Funds needed" with all the calculations at the "Create Offer" window - Based on the feedback from [this reddit post](https://www.reddit.com/r/bisq/comments/ba0xxy/feedback_show_funds_needed_with_all_the/), can we have "Funds needed" value showing up at the "Create offer" stage? | process | show funds needed with all the calculations at the create offer window based on the feedback from can we have funds needed value showing up at the create offer stage | 1 |
316,769 | 23,647,489,893 | IssuesEvent | 2022-08-26 00:51:23 | aws/aws-cdk | https://api.github.com/repos/aws/aws-cdk | closed | aws-cdk-lib: SplunkLogDriverProps does not accept token value | p2 effort/small documentation aws-cdk-lib | ### Describe the bug
When trying to provide a token to the SplunkLogDriver, I would like to provide the token as a string value as I did with CDKv1. CDK 2.38.1 internal errors say "Please provide either token or secretToken" and that if "secretToken" is not provided, that "token" will be used. But if token is provided, it says," 'token' does not exist in type 'SplunkLogDriverProps' ".
### Expected Behavior
Should be able to provide "token" or "secretToken" in SplunkLogDriver props.
### Current Behavior
If "token" is provided, get: Object literal may only specify known properties, and 'token' does not exist in type 'SplunkLogDriverProps'.
### Reproduction Steps
Provide a token to SplunkLogDriver, and do not provide secretToken.
### Possible Solution
Allow specifying "token" as with CDK v1.
### Additional Information/Context
_No response_
### CDK CLI Version
2.38.1
### Framework Version
_No response_
### Node.js Version
16.16.0
### OS
macOS 12.5.1
### Language
Typescript
### Language Version
TypeScript 4.7.4
### Other information
_No response_ | 1.0 | aws-cdk-lib: SplunkLogDriverProps does not accept token value - ### Describe the bug
When trying to provide a token to the SplunkLogDriver, I would like to provide the token as a string value as I did with CDKv1. CDK 2.38.1 internal errors say "Please provide either token or secretToken" and that if "secretToken" is not provided, that "token" will be used. But if token is provided, it says," 'token' does not exist in type 'SplunkLogDriverProps' ".
### Expected Behavior
Should be able to provide "token" or "secretToken" in SplunkLogDriver props.
### Current Behavior
If "token" is provided, get: Object literal may only specify known properties, and 'token' does not exist in type 'SplunkLogDriverProps'.
### Reproduction Steps
Provide a token to SplunkLogDriver, and do not provide secretToken.
### Possible Solution
Allow specifying "token" as with CDK v1.
### Additional Information/Context
_No response_
### CDK CLI Version
2.38.1
### Framework Version
_No response_
### Node.js Version
16.16.0
### OS
macOS 12.5.1
### Language
Typescript
### Language Version
TypeScript 4.7.4
### Other information
_No response_ | non_process | aws cdk lib splunklogdriverprops does not accept token value describe the bug when trying to provide a token to the splunklogdriver i would like to provide the token as a string value as i did with cdk internal errors say please provide either token or secrettoken and that if secrettoken is not provided that token will be used but if token is provided it says token does not exist in type splunklogdriverprops expected behavior should be able to provide token or secrettoken in splunklogdriver props current behavior if token is provided get object literal may only specify known properties and token does not exist in type splunklogdriverprops reproduction steps provide a token to splunklogdriver and do not provide secrettoken possible solution allow specifying token as with cdk additional information context no response cdk cli version framework version no response node js version os macos language typescript language version typescript other information no response | 0 |
88,140 | 15,800,744,369 | IssuesEvent | 2021-04-03 01:05:38 | rammatzkvosky/11 | https://api.github.com/repos/rammatzkvosky/11 | opened | CVE-2020-36184 (High) detected in jackson-databind-2.8.8.jar | security vulnerability | ## CVE-2020-36184 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.8.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: 11/pom.xml</p>
<p>Path to vulnerable library: canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.8/jackson-databind-2.8.8.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.8.8.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.tomcat.dbcp.dbcp2.datasources.PerUserPoolDataSource.
<p>Publish Date: 2021-01-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36184>CVE-2020-36184</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2998">https://github.com/FasterXML/jackson-databind/issues/2998</a></p>
<p>Release Date: 2021-01-06</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.8","packageFilePaths":["/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.8.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"}],"baseBranches":[],"vulnerabilityIdentifier":"CVE-2020-36184","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.tomcat.dbcp.dbcp2.datasources.PerUserPoolDataSource.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36184","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2020-36184 (High) detected in jackson-databind-2.8.8.jar - ## CVE-2020-36184 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.8.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: 11/pom.xml</p>
<p>Path to vulnerable library: canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.8/jackson-databind-2.8.8.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.8.8.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.tomcat.dbcp.dbcp2.datasources.PerUserPoolDataSource.
<p>Publish Date: 2021-01-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36184>CVE-2020-36184</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2998">https://github.com/FasterXML/jackson-databind/issues/2998</a></p>
<p>Release Date: 2021-01-06</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.8","packageFilePaths":["/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.8.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8"}],"baseBranches":[],"vulnerabilityIdentifier":"CVE-2020-36184","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.tomcat.dbcp.dbcp2.datasources.PerUserPoolDataSource.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36184","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_process | cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file pom xml path to vulnerable library canner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache tomcat dbcp datasources peruserpooldatasource publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind basebranches vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache tomcat dbcp datasources peruserpooldatasource vulnerabilityurl | 0 |
50,568 | 10,525,600,369 | IssuesEvent | 2019-09-30 15:22:37 | mozilla/addons-server | https://api.github.com/repos/mozilla/addons-server | closed | Remove ".jar" support | component: code quality component: devhub priority: p3 | This came up during an IRC discussion. We probably don't need to support `.jar` files anymore. We should remove this extension from `VALID_ADDON_FILE_EXTENSIONS` and remove potentially related code. | 1.0 | Remove ".jar" support - This came up during an IRC discussion. We probably don't need to support `.jar` files anymore. We should remove this extension from `VALID_ADDON_FILE_EXTENSIONS` and remove potentially related code. | non_process | remove jar support this came up during an irc discussion we probably don t need to support jar files anymore we should remove this extension from valid addon file extensions and remove potentially related code | 0 |
51,599 | 27,164,083,257 | IssuesEvent | 2023-02-17 14:10:33 | liquibase/liquibase | https://api.github.com/repos/liquibase/liquibase | closed | Improve snapshot performance for DBAAS databases | performance sprint2022-40 | ## Description
The snapshot process is never as fast as it could be, generally from making too many small queries when it could/should make less queries which each return more rows and/or more complete information.
For databases on a LAN that's not a big problem, but for cloud-based databases the network overhead is so much more it makes a very noticeable difference. For example, snapshotting the databasechangelog table takes about 30 seconds on bigquery.
There are work arounds in the code on when to do bulk queries and when to consider a snapshotted object "complete" but those are not extendable beyond what can be put directly into the code in liquibase-core and therefore cannot be leveraged by cloud-database extensions.
## Steps To Reproduce
Against a bigquery database:
- Run `liquibase update` with a changelog that doesn't have to execute anything
- Run `liquibase snapshot` against schemas of various sizes
## Expected/Desired Behavior
"Runs fast" is always a not-useful "expected behavior". As part of the investigation we need to figure out what can and should be fixed and likely come up with phases of improvement
## Additional Context
Columns are often times the biggest culprit because there are the most of those. From an initial poking around: do we keep re-snapshotting them because LIQUIBASE_COMPLETE is only true if it's not nullable (line 203)? If so, why is that setting based on nullable?
At some point, the whole system was designed more for single-object snapshotting and has gotten overly complex and hard to reason against due to trying to make it even vaguely performant on full-database snapshots. When does it make sense to do a larger overhaul that is more focused on performance and simplicity? | True | Improve snapshot performance for DBAAS databases - ## Description
The snapshot process is never as fast as it could be, generally from making too many small queries when it could/should make less queries which each return more rows and/or more complete information.
For databases on a LAN that's not a big problem, but for cloud-based databases the network overhead is so much more it makes a very noticeable difference. For example, snapshotting the databasechangelog table takes about 30 seconds on bigquery.
There are work arounds in the code on when to do bulk queries and when to consider a snapshotted object "complete" but those are not extendable beyond what can be put directly into the code in liquibase-core and therefore cannot be leveraged by cloud-database extensions.
## Steps To Reproduce
Against a bigquery database:
- Run `liquibase update` with a changelog that doesn't have to execute anything
- Run `liquibase snapshot` against schemas of various sizes
## Expected/Desired Behavior
"Runs fast" is always a not-useful "expected behavior". As part of the investigation we need to figure out what can and should be fixed and likely come up with phases of improvement
## Additional Context
Columns are often times the biggest culprit because there are the most of those. From an initial poking around: do we keep re-snapshotting them because LIQUIBASE_COMPLETE is only true if it's not nullable (line 203)? If so, why is that setting based on nullable?
At some point, the whole system was designed more for single-object snapshotting and has gotten overly complex and hard to reason against due to trying to make it even vaguely performant on full-database snapshots. When does it make sense to do a larger overhaul that is more focused on performance and simplicity? | non_process | improve snapshot performance for dbaas databases description the snapshot process is never as fast as it could be generally from making too many small queries when it could should make less queries which each return more rows and or more complete information for databases on a lan that s not a big problem but for cloud based databases the network overhead is so much more it makes a very noticeable difference for example snapshotting the databasechangelog table takes about seconds on bigquery there are work arounds in the code on when to do bulk queries and when to consider a snapshotted object complete but those are not extendable beyond what can be put directly into the code in liquibase core and therefore cannot be leveraged by cloud database extensions steps to reproduce against a bigquery database run liquibase update with a changelog that doesn t have to execute anything run liquibase snapshot against schemas of various sizes expected desired behavior runs fast is always a not useful expected behavior as part of the investigation we need to figure out what can and should be fixed and likely come up with phases of improvement additional context columns are often times the biggest culprit because there are the most of those from an initial poking around do we keep re snapshotting them because liquibase complete is only true if it s not nullable line if so why is that setting based on nullable at some point the whole system was designed more for single object snapshotting and has gotten overly complex and hard to reason against due to trying to make it even vaguely performant on full database snapshots when does it make sense to do a larger overhaul that is more focused on performance and simplicity | 0 |
179,212 | 21,531,997,736 | IssuesEvent | 2022-04-29 02:22:18 | drakeg/udemy_django_vue | https://api.github.com/repos/drakeg/udemy_django_vue | closed | CVE-2021-33503 (High) detected in urllib3-1.25.9-py2.py3-none-any.whl - autoclosed | security vulnerability | ## CVE-2021-33503 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>urllib3-1.25.9-py2.py3-none-any.whl</b></p></summary>
<p>HTTP library with thread-safe connection pooling, file post, and more.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/e1/e5/df302e8017440f111c11cc41a6b432838672f5a70aa29227bf58149dc72f/urllib3-1.25.9-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/e1/e5/df302e8017440f111c11cc41a6b432838672f5a70aa29227bf58149dc72f/urllib3-1.25.9-py2.py3-none-any.whl</a></p>
<p>Path to dependency file: /requirements.txt</p>
<p>Path to vulnerable library: /requirements.txt</p>
<p>
Dependency Hierarchy:
- requests_oauthlib-1.3.0-py2.py3-none-any.whl (Root Library)
- requests-2.26.0-py2.py3-none-any.whl
- :x: **urllib3-1.25.9-py2.py3-none-any.whl** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/drakeg/udemy_django_vue/commit/576ef05699bb9bea9bd18cb142d008bcee57ef8d">576ef05699bb9bea9bd18cb142d008bcee57ef8d</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in urllib3 before 1.26.5. When provided with a URL containing many @ characters in the authority component, the authority regular expression exhibits catastrophic backtracking, causing a denial of service if a URL were passed as a parameter or redirected to via an HTTP redirect.
<p>Publish Date: 2021-06-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33503>CVE-2021-33503</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/urllib3/urllib3/security/advisories/GHSA-q2q7-5pp4-w6pg">https://github.com/urllib3/urllib3/security/advisories/GHSA-q2q7-5pp4-w6pg</a></p>
<p>Release Date: 2021-06-29</p>
<p>Fix Resolution: urllib3 - 1.26.5</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-33503 (High) detected in urllib3-1.25.9-py2.py3-none-any.whl - autoclosed - ## CVE-2021-33503 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>urllib3-1.25.9-py2.py3-none-any.whl</b></p></summary>
<p>HTTP library with thread-safe connection pooling, file post, and more.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/e1/e5/df302e8017440f111c11cc41a6b432838672f5a70aa29227bf58149dc72f/urllib3-1.25.9-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/e1/e5/df302e8017440f111c11cc41a6b432838672f5a70aa29227bf58149dc72f/urllib3-1.25.9-py2.py3-none-any.whl</a></p>
<p>Path to dependency file: /requirements.txt</p>
<p>Path to vulnerable library: /requirements.txt</p>
<p>
Dependency Hierarchy:
- requests_oauthlib-1.3.0-py2.py3-none-any.whl (Root Library)
- requests-2.26.0-py2.py3-none-any.whl
- :x: **urllib3-1.25.9-py2.py3-none-any.whl** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/drakeg/udemy_django_vue/commit/576ef05699bb9bea9bd18cb142d008bcee57ef8d">576ef05699bb9bea9bd18cb142d008bcee57ef8d</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in urllib3 before 1.26.5. When provided with a URL containing many @ characters in the authority component, the authority regular expression exhibits catastrophic backtracking, causing a denial of service if a URL were passed as a parameter or redirected to via an HTTP redirect.
<p>Publish Date: 2021-06-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33503>CVE-2021-33503</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/urllib3/urllib3/security/advisories/GHSA-q2q7-5pp4-w6pg">https://github.com/urllib3/urllib3/security/advisories/GHSA-q2q7-5pp4-w6pg</a></p>
<p>Release Date: 2021-06-29</p>
<p>Fix Resolution: urllib3 - 1.26.5</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_process | cve high detected in none any whl autoclosed cve high severity vulnerability vulnerable library none any whl http library with thread safe connection pooling file post and more library home page a href path to dependency file requirements txt path to vulnerable library requirements txt dependency hierarchy requests oauthlib none any whl root library requests none any whl x none any whl vulnerable library found in head commit a href vulnerability details an issue was discovered in before when provided with a url containing many characters in the authority component the authority regular expression exhibits catastrophic backtracking causing a denial of service if a url were passed as a parameter or redirected to via an http redirect publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
113,496 | 17,145,337,845 | IssuesEvent | 2021-07-13 14:05:08 | OpenLiberty/open-liberty | https://api.github.com/repos/OpenLiberty/open-liberty | opened | SAML SSO shows error page before redirecting to resource TS005787316 | release bug team:Security SSO | - I am using samlWeb-2.0 with Liberty to configure SP-Initiated SSO.
- When I make a request to the application, the error screen of the application is displayed for a moment,
- After that, the screen of the expected application is displayed.
- When I checked the trace of the web container in Liberty, it surely returned status 401 and it seems to be returning the error screen of the application.
For an initial resource access, the SP should redirect to the IdP without showing the error page configured for the 401 error code. | True | SAML SSO shows error page before redirecting to resource TS005787316 - - I am using samlWeb-2.0 with Liberty to configure SP-Initiated SSO.
- When I make a request to the application, the error screen of the application is displayed for a moment,
- After that, the screen of the expected application is displayed.
- When I checked the trace of the web container in Liberty, it surely returned status 401 and it seems to be returning the error screen of the application.
For an initial resource access, the SP should redirect to the IdP without showing the error page configured for the 401 error code. | non_process | saml sso shows error page before redirecting to resource i am using samlweb with liberty to configure sp initiated sso when i make a request to the application the error screen of the application is displayed for a moment after that the screen of the expected application is displayed when i checked the trace of the web container in liberty it surely returned status and it seems to be returning the error screen of the application for an initial resource access the sp should redirect to the idp without showing the error page configured for the error code | 0 |
154,238 | 19,711,853,686 | IssuesEvent | 2022-01-13 06:42:14 | Shai-Demo-Org/JS-Demo | https://api.github.com/repos/Shai-Demo-Org/JS-Demo | opened | CVE-2019-10746 (High) detected in mixin-deep-1.3.1.tgz | security vulnerability | ## CVE-2019-10746 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mixin-deep-1.3.1.tgz</b></p></summary>
<p>Deeply mix the properties of objects into the first object. Like merge-deep, but doesn't clone.</p>
<p>Library home page: <a href="https://registry.npmjs.org/mixin-deep/-/mixin-deep-1.3.1.tgz">https://registry.npmjs.org/mixin-deep/-/mixin-deep-1.3.1.tgz</a></p>
<p>
Dependency Hierarchy:
- forever-2.0.0.tgz (Root Library)
- forever-monitor-2.0.0.tgz
- chokidar-2.1.8.tgz
- braces-2.3.2.tgz
- snapdragon-0.8.2.tgz
- base-0.11.2.tgz
- :x: **mixin-deep-1.3.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Shai-Demo-Org/JS-Demo/commit/5aaf68c266a0270ee6d4a0ac684351efb0b24dbf">5aaf68c266a0270ee6d4a0ac684351efb0b24dbf</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
mixin-deep is vulnerable to Prototype Pollution in versions before 1.3.2 and version 2.0.0. The function mixin-deep could be tricked into adding or modifying properties of Object.prototype using a constructor payload.
<p>Publish Date: 2019-08-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10746>CVE-2019-10746</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/jonschlinkert/mixin-deep/commit/8f464c8ce9761a8c9c2b3457eaeee9d404fa7af9">https://github.com/jonschlinkert/mixin-deep/commit/8f464c8ce9761a8c9c2b3457eaeee9d404fa7af9</a></p>
<p>Release Date: 2019-08-23</p>
<p>Fix Resolution: 1.3.2,2.0.1</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"mixin-deep","packageVersion":"1.3.1","packageFilePaths":[null],"isTransitiveDependency":true,"dependencyTree":"forever:2.0.0;forever-monitor:2.0.0;chokidar:2.1.8;braces:2.3.2;snapdragon:0.8.2;base:0.11.2;mixin-deep:1.3.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.3.2,2.0.1","isBinary":false}],"baseBranches":[],"vulnerabilityIdentifier":"CVE-2019-10746","vulnerabilityDetails":"mixin-deep is vulnerable to Prototype Pollution in versions before 1.3.2 and version 2.0.0. The function mixin-deep could be tricked into adding or modifying properties of Object.prototype using a constructor payload.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10746","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2019-10746 (High) detected in mixin-deep-1.3.1.tgz - ## CVE-2019-10746 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mixin-deep-1.3.1.tgz</b></p></summary>
<p>Deeply mix the properties of objects into the first object. Like merge-deep, but doesn't clone.</p>
<p>Library home page: <a href="https://registry.npmjs.org/mixin-deep/-/mixin-deep-1.3.1.tgz">https://registry.npmjs.org/mixin-deep/-/mixin-deep-1.3.1.tgz</a></p>
<p>
Dependency Hierarchy:
- forever-2.0.0.tgz (Root Library)
- forever-monitor-2.0.0.tgz
- chokidar-2.1.8.tgz
- braces-2.3.2.tgz
- snapdragon-0.8.2.tgz
- base-0.11.2.tgz
- :x: **mixin-deep-1.3.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Shai-Demo-Org/JS-Demo/commit/5aaf68c266a0270ee6d4a0ac684351efb0b24dbf">5aaf68c266a0270ee6d4a0ac684351efb0b24dbf</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
mixin-deep is vulnerable to Prototype Pollution in versions before 1.3.2 and version 2.0.0. The function mixin-deep could be tricked into adding or modifying properties of Object.prototype using a constructor payload.
<p>Publish Date: 2019-08-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10746>CVE-2019-10746</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/jonschlinkert/mixin-deep/commit/8f464c8ce9761a8c9c2b3457eaeee9d404fa7af9">https://github.com/jonschlinkert/mixin-deep/commit/8f464c8ce9761a8c9c2b3457eaeee9d404fa7af9</a></p>
<p>Release Date: 2019-08-23</p>
<p>Fix Resolution: 1.3.2,2.0.1</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"mixin-deep","packageVersion":"1.3.1","packageFilePaths":[null],"isTransitiveDependency":true,"dependencyTree":"forever:2.0.0;forever-monitor:2.0.0;chokidar:2.1.8;braces:2.3.2;snapdragon:0.8.2;base:0.11.2;mixin-deep:1.3.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.3.2,2.0.1","isBinary":false}],"baseBranches":[],"vulnerabilityIdentifier":"CVE-2019-10746","vulnerabilityDetails":"mixin-deep is vulnerable to Prototype Pollution in versions before 1.3.2 and version 2.0.0. The function mixin-deep could be tricked into adding or modifying properties of Object.prototype using a constructor payload.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10746","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_process | cve high detected in mixin deep tgz cve high severity vulnerability vulnerable library mixin deep tgz deeply mix the properties of objects into the first object like merge deep but doesn t clone library home page a href dependency hierarchy forever tgz root library forever monitor tgz chokidar tgz braces tgz snapdragon tgz base tgz x mixin deep tgz vulnerable library found in head commit a href vulnerability details mixin deep is vulnerable to prototype pollution in versions before and version the function mixin deep could be tricked into adding or modifying properties of object prototype using a constructor payload publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree forever forever monitor chokidar braces snapdragon base mixin deep isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails mixin deep is vulnerable to prototype pollution in versions before and version the function mixin deep could be tricked into adding or modifying properties of object prototype using a constructor payload vulnerabilityurl | 0 |
215 | 2,644,220,275 | IssuesEvent | 2015-03-12 15:50:38 | sci-visus/visus-issues | https://api.github.com/repos/sci-visus/visus-issues | opened | add support for new Visus Array operations | Feature Request Processing | Binarization and Bitwise operations
Automatic Normalization
Thresholding
This is not a complete list, but I wanted to start collecting ideas.
Essentially many of these are motivated by trying to write scripts that loop over the arrays to perform some operation. As this isn't possible, collective versions of these operations are necessary.
| 1.0 | add support for new Visus Array operations - Binarization and Bitwise operations
Automatic Normalization
Thresholding
This is not a complete list, but I wanted to start collecting ideas.
Essentially many of these are motivated by trying to write scripts that loop over the arrays to perform some operation. As this isn't possible, collective versions of these operations are necessary.
| process | add support for new visus array operations binarization and bitwise operations automatic normalization thresholding this is not a complete list but i wanted to start collecting ideas essentially many of these are motivated by trying to write scripts that loop over the arrays to perform some operation as this isn t possible collective versions of these operations are necessary | 1 |
19,126 | 25,177,739,776 | IssuesEvent | 2022-11-11 10:47:31 | googleapis/nodejs-eventarc-publishing | https://api.github.com/repos/googleapis/nodejs-eventarc-publishing | closed | Your .repo-metadata.json file has a problem 🤒 | type: process repo-metadata: lint | You have a problem with your .repo-metadata.json file:
Result of scan 📈:
* api_shortname 'publishing' invalid in .repo-metadata.json
☝️ Once you address these problems, you can close this issue.
### Need help?
* [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field.
* [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**.
* Reach out to **go/github-automation** if you have any questions. | 1.0 | Your .repo-metadata.json file has a problem 🤒 - You have a problem with your .repo-metadata.json file:
Result of scan 📈:
* api_shortname 'publishing' invalid in .repo-metadata.json
☝️ Once you address these problems, you can close this issue.
### Need help?
* [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field.
* [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**.
* Reach out to **go/github-automation** if you have any questions. | process | your repo metadata json file has a problem 🤒 you have a problem with your repo metadata json file result of scan 📈 api shortname publishing invalid in repo metadata json ☝️ once you address these problems you can close this issue need help lists valid options for each field for grpc libraries api shortname should match the subdomain of an api s hostname reach out to go github automation if you have any questions | 1 |
8,576 | 11,742,841,472 | IssuesEvent | 2020-03-12 02:18:01 | googleapis/google-cloud-cpp | https://api.github.com/repos/googleapis/google-cloud-cpp | opened | opensuse-tumbleweed build is failing continuously since 9 march | type: process | example: https://source.cloud.google.com/results/invocations/0de0824b-99d6-46a6-b7d2-0438502e4d64/targets
notes ftom @coryan
You are the operator this week right? This is a different failure from #3437, so I am not nothing to investigate any further than what you will find below. It is really hard to see what is going on because it is "the case of the dog that did not bark", the end of the log reads:
```
================================================================
Run Bigtable test programs against installed libraries opensuse-tumbleweed.
Created table tbl-1583871465-31780
bulk mutation successful
row-key-0: fam:c0 = v0
row-key-0: fam:c1 = v1
row-key-1: fam:c0 = v2
row-key-1: fam:c1 = v3
Deleted table tbl-1583871465-31780
Run Storage test programs against installed libraries opensuse-tumbleweed.
[ID: 4601247] Build finished after 1015 secs, exit value: 139
```
Okay, so a failure status, but what failed? That last message reads like this on success (it does succeed when running locally):
```
Run Storage test programs against installed libraries opensuse-tumbleweed.
Successfully created object, generation=1583871862928949
================================================================
```
So the test program for storage:
https://github.com/googleapis/google-cloud-cpp/blob/master/ci/test-install/storage/storage_install_test.cc
failed, but printed nothing. That is all I know at this point. | 1.0 | opensuse-tumbleweed build is failing continuously since 9 march - example: https://source.cloud.google.com/results/invocations/0de0824b-99d6-46a6-b7d2-0438502e4d64/targets
notes ftom @coryan
You are the operator this week right? This is a different failure from #3437, so I am not nothing to investigate any further than what you will find below. It is really hard to see what is going on because it is "the case of the dog that did not bark", the end of the log reads:
```
================================================================
Run Bigtable test programs against installed libraries opensuse-tumbleweed.
Created table tbl-1583871465-31780
bulk mutation successful
row-key-0: fam:c0 = v0
row-key-0: fam:c1 = v1
row-key-1: fam:c0 = v2
row-key-1: fam:c1 = v3
Deleted table tbl-1583871465-31780
Run Storage test programs against installed libraries opensuse-tumbleweed.
[ID: 4601247] Build finished after 1015 secs, exit value: 139
```
Okay, so a failure status, but what failed? That last message reads like this on success (it does succeed when running locally):
```
Run Storage test programs against installed libraries opensuse-tumbleweed.
Successfully created object, generation=1583871862928949
================================================================
```
So the test program for storage:
https://github.com/googleapis/google-cloud-cpp/blob/master/ci/test-install/storage/storage_install_test.cc
failed, but printed nothing. That is all I know at this point. | process | opensuse tumbleweed build is failing continuously since march example notes ftom coryan you are the operator this week right this is a different failure from so i am not nothing to investigate any further than what you will find below it is really hard to see what is going on because it is the case of the dog that did not bark the end of the log reads run bigtable test programs against installed libraries opensuse tumbleweed created table tbl bulk mutation successful row key fam row key fam row key fam row key fam deleted table tbl run storage test programs against installed libraries opensuse tumbleweed build finished after secs exit value okay so a failure status but what failed that last message reads like this on success it does succeed when running locally run storage test programs against installed libraries opensuse tumbleweed successfully created object generation so the test program for storage failed but printed nothing that is all i know at this point | 1 |
150,008 | 13,307,505,018 | IssuesEvent | 2020-08-25 22:20:31 | corretto/corretto-8 | https://api.github.com/repos/corretto/corretto-8 | closed | Update FAQ for JFR on JDK 1.8 | documentation | In https://aws.amazon.com/corretto/faqs/ it says:
> A: Corretto is designed as a drop-in replacement for all Java SE distributions unless you are using features (e.g., Java Flight Recorder) not available in OpenJDK. Once Corretto binaries are installed on a host and correctly invoked to run your Java applications (e.g., using the alternatives command on Linux), existing command-line options, tuning parameters, monitoring, and anything else in place will continue to work as before.
Since the backport of JFR in 8.262.10.1 with JDK-8223147, Java Flight Recorder is enabled and so this is no longer accurate :-)
https://github.com/corretto/corretto-8/blob/release-8.262.10.1/CHANGELOG.md#corretto-version-8262101
| 1.0 | Update FAQ for JFR on JDK 1.8 - In https://aws.amazon.com/corretto/faqs/ it says:
> A: Corretto is designed as a drop-in replacement for all Java SE distributions unless you are using features (e.g., Java Flight Recorder) not available in OpenJDK. Once Corretto binaries are installed on a host and correctly invoked to run your Java applications (e.g., using the alternatives command on Linux), existing command-line options, tuning parameters, monitoring, and anything else in place will continue to work as before.
Since the backport of JFR in 8.262.10.1 with JDK-8223147, Java Flight Recorder is enabled and so this is no longer accurate :-)
https://github.com/corretto/corretto-8/blob/release-8.262.10.1/CHANGELOG.md#corretto-version-8262101
| non_process | update faq for jfr on jdk in it says a corretto is designed as a drop in replacement for all java se distributions unless you are using features e g java flight recorder not available in openjdk once corretto binaries are installed on a host and correctly invoked to run your java applications e g using the alternatives command on linux existing command line options tuning parameters monitoring and anything else in place will continue to work as before since the backport of jfr in with jdk java flight recorder is enabled and so this is no longer accurate | 0 |
7,969 | 8,128,980,961 | IssuesEvent | 2018-08-17 13:45:06 | Microsoft/vscode-cpptools | https://api.github.com/repos/Microsoft/vscode-cpptools | closed | Mac Framework dependencies are not being discovered. | Language Service bug fixed (release pending) | I cannot make IntelliSense recognize the Mac OS frameworks...
My application makes use of the ApplicationServices framework and IntelliSense cannot find `AE/AE.h` header for example. The only way I found to make it work was creating a symlink to the Headers directory from the framework. I'm using VS Code 1.22.2 and the C++ extension 0.16.1
Here's my properties:
{
"name": "Mac",
"includePath": [
"${workspaceFolder}",
"/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/include/c++/v1",
"/usr/local/include",
"/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/lib/clang/9.1.0/include",
"/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/include",
"/usr/include",
"${workspaceFolder}/import/include"
],
"defines": [],
"intelliSenseMode": "clang-x64",
"browse": {
"path": [
"${workspaceFolder}",
"/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/include/c++/v1",
"/usr/local/include",
"/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/lib/clang/9.1.0/include",
"/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/include",
"/usr/include"
],
"limitSymbolsToIncludedHeaders": true,
"databaseFilename": ""
},
"macFrameworkPath": [
"/System/Library/Frameworks",
"/Library/Frameworks"
],
"compilerPath": "/usr/bin/clang++",
"cStandard": "c11",
"cppStandard": "c++17"
}
| 1.0 | Mac Framework dependencies are not being discovered. - I cannot make IntelliSense recognize the Mac OS frameworks...
My application makes use of the ApplicationServices framework and IntelliSense cannot find `AE/AE.h` header for example. The only way I found to make it work was creating a symlink to the Headers directory from the framework. I'm using VS Code 1.22.2 and the C++ extension 0.16.1
Here's my properties:
{
"name": "Mac",
"includePath": [
"${workspaceFolder}",
"/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/include/c++/v1",
"/usr/local/include",
"/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/lib/clang/9.1.0/include",
"/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/include",
"/usr/include",
"${workspaceFolder}/import/include"
],
"defines": [],
"intelliSenseMode": "clang-x64",
"browse": {
"path": [
"${workspaceFolder}",
"/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/include/c++/v1",
"/usr/local/include",
"/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/lib/clang/9.1.0/include",
"/Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/include",
"/usr/include"
],
"limitSymbolsToIncludedHeaders": true,
"databaseFilename": ""
},
"macFrameworkPath": [
"/System/Library/Frameworks",
"/Library/Frameworks"
],
"compilerPath": "/usr/bin/clang++",
"cStandard": "c11",
"cppStandard": "c++17"
}
| non_process | mac framework dependencies are not being discovered i cannot make intellisense recognize the mac os frameworks my application makes use of the applicationservices framework and intellisense cannot find ae ae h header for example the only way i found to make it work was creating a symlink to the headers directory from the framework i m using vs code and the c extension here s my properties name mac includepath workspacefolder applications xcode app contents developer toolchains xcodedefault xctoolchain usr include c usr local include applications xcode app contents developer toolchains xcodedefault xctoolchain usr lib clang include applications xcode app contents developer toolchains xcodedefault xctoolchain usr include usr include workspacefolder import include defines intellisensemode clang browse path workspacefolder applications xcode app contents developer toolchains xcodedefault xctoolchain usr include c usr local include applications xcode app contents developer toolchains xcodedefault xctoolchain usr lib clang include applications xcode app contents developer toolchains xcodedefault xctoolchain usr include usr include limitsymbolstoincludedheaders true databasefilename macframeworkpath system library frameworks library frameworks compilerpath usr bin clang cstandard cppstandard c | 0 |
59,626 | 14,620,876,396 | IssuesEvent | 2020-12-22 20:32:41 | MarlinFirmware/Marlin | https://api.github.com/repos/MarlinFirmware/Marlin | closed | [BUG] Build fails using FYSETC_F6_14 due to missing fysetc_f6_14 board | Bug: Confirmed ! C: Boards/Pins C: Build / Toolchain Fix Included | <!--
Please follow the instructions below. Failure to do so may result in your issue being closed.
### Before Reporting a Bug
1. Test with the `bugfix-2.0.x` branch to see whether the issue still exists.
2. Get troubleshooting help from the Marlin community to confirm it's a bug and not just a configuration error. Links at https://github.com/MarlinFirmware/Marlin/issues/new/choose
### Instructions
1. Fill out every section of the template below.
2. Always attach configuration files, regardless of whether you think they are involved.
3. Read and understand Marlin's Code of Conduct. By filing an Issue, you are expected to comply with it, including treating everyone with respect: https://github.com/MarlinFirmware/Marlin/blob/master/.github/code_of_conduct.md
-->
### Bug Description
When cloning Marlin, and trying to set up Marlin with PlatformIO through CLion, it always fails with the next error:
```
/home/me/.local/bin/platformio init --ide clion
The current working directory /home/me/Documents/3D_printing/firmware_compilation/SKR-1.1-FFS will be used for the project.
The next files/directories have been created in /home/me/Documents/3D_printing/firmware_compilation/SKR-1.1-FFS
include - Put project header files here
lib - Put here project specific (private) libraries
src - Put project source files here
platformio.ini - Project Configuration File
Error: Unknown board ID 'fysetc_f6_14'
Process finished with exit code 1
```
### Configuration Files
N/A for this bug
### Steps to Reproduce
1. Clone Marlin, optionally check out to a tag
2. Open the project in CLion and let it initialize (using the `platformio init --ide clion`), or alternatively initialize it yourself through a terminal
3. Note that it errors with the error reported above
**Expected behavior:**
PlatformIO initializes correctly
**Actual behavior:**
PlatformIO errors out with "Unknown board ID 'fysetc_f6_14'"
#### Additional Information
If I comment out the section of the fysetc board, it works as intended, and then I can use CLion in combination with PlatformIO.
| 1.0 | [BUG] Build fails using FYSETC_F6_14 due to missing fysetc_f6_14 board - <!--
Please follow the instructions below. Failure to do so may result in your issue being closed.
### Before Reporting a Bug
1. Test with the `bugfix-2.0.x` branch to see whether the issue still exists.
2. Get troubleshooting help from the Marlin community to confirm it's a bug and not just a configuration error. Links at https://github.com/MarlinFirmware/Marlin/issues/new/choose
### Instructions
1. Fill out every section of the template below.
2. Always attach configuration files, regardless of whether you think they are involved.
3. Read and understand Marlin's Code of Conduct. By filing an Issue, you are expected to comply with it, including treating everyone with respect: https://github.com/MarlinFirmware/Marlin/blob/master/.github/code_of_conduct.md
-->
### Bug Description
When cloning Marlin, and trying to set up Marlin with PlatformIO through CLion, it always fails with the next error:
```
/home/me/.local/bin/platformio init --ide clion
The current working directory /home/me/Documents/3D_printing/firmware_compilation/SKR-1.1-FFS will be used for the project.
The next files/directories have been created in /home/me/Documents/3D_printing/firmware_compilation/SKR-1.1-FFS
include - Put project header files here
lib - Put here project specific (private) libraries
src - Put project source files here
platformio.ini - Project Configuration File
Error: Unknown board ID 'fysetc_f6_14'
Process finished with exit code 1
```
### Configuration Files
N/A for this bug
### Steps to Reproduce
1. Clone Marlin, optionally check out to a tag
2. Open the project in CLion and let it initialize (using the `platformio init --ide clion`), or alternatively initialize it yourself through a terminal
3. Note that it errors with the error reported above
**Expected behavior:**
PlatformIO initializes correctly
**Actual behavior:**
PlatformIO errors out with "Unknown board ID 'fysetc_f6_14'"
#### Additional Information
If I comment out the section of the fysetc board, it works as intended, and then I can use CLion in combination with PlatformIO.
| non_process | build fails using fysetc due to missing fysetc board please follow the instructions below failure to do so may result in your issue being closed before reporting a bug test with the bugfix x branch to see whether the issue still exists get troubleshooting help from the marlin community to confirm it s a bug and not just a configuration error links at instructions fill out every section of the template below always attach configuration files regardless of whether you think they are involved read and understand marlin s code of conduct by filing an issue you are expected to comply with it including treating everyone with respect bug description when cloning marlin and trying to set up marlin with platformio through clion it always fails with the next error home me local bin platformio init ide clion the current working directory home me documents printing firmware compilation skr ffs will be used for the project the next files directories have been created in home me documents printing firmware compilation skr ffs include put project header files here lib put here project specific private libraries src put project source files here platformio ini project configuration file error unknown board id fysetc process finished with exit code configuration files n a for this bug steps to reproduce clone marlin optionally check out to a tag open the project in clion and let it initialize using the platformio init ide clion or alternatively initialize it yourself through a terminal note that it errors with the error reported above expected behavior platformio initializes correctly actual behavior platformio errors out with unknown board id fysetc additional information if i comment out the section of the fysetc board it works as intended and then i can use clion in combination with platformio | 0 |
3,855 | 6,808,620,159 | IssuesEvent | 2017-11-04 05:41:20 | Great-Hill-Corporation/quickBlocks | https://api.github.com/repos/Great-Hill-Corporation/quickBlocks | reopened | miniBlocks.bin and fullBlockIndex.bin require full scan of blockchain | libs-etherlib status-inprocess type-enhancement | I need a way to quickly scan the full blockchain or download the last fullBlockIndex / miniBlocks.bin files. Otherwise, some tools (whenBlock and cacheMan?) don't work.
You can figure out what doesn't work by clearing the cache and re-running all the tests. | 1.0 | miniBlocks.bin and fullBlockIndex.bin require full scan of blockchain - I need a way to quickly scan the full blockchain or download the last fullBlockIndex / miniBlocks.bin files. Otherwise, some tools (whenBlock and cacheMan?) don't work.
You can figure out what doesn't work by clearing the cache and re-running all the tests. | process | miniblocks bin and fullblockindex bin require full scan of blockchain i need a way to quickly scan the full blockchain or download the last fullblockindex miniblocks bin files otherwise some tools whenblock and cacheman don t work you can figure out what doesn t work by clearing the cache and re running all the tests | 1 |
242,381 | 7,841,553,536 | IssuesEvent | 2018-06-18 20:01:07 | google/google-api-php-client | https://api.github.com/repos/google/google-api-php-client | closed | Batch operations always returning 404: Not Found response | :rotating_light: priority: p2 type: bug | Using the Google_Service_PeopleService API provided in the [google-api-php-client-services](https://github.com/google/google-api-php-client-services) library, I'm attempting to create contacts via the Google_Http_Batch class.
I've tested executing the request on its own prior to adding it to the batch and it works correctly, however when attempting to use batching it always returns a 404 Not Found Exception as the Response.
I searched around for similar issues and it seems like this can be caused by methods not allowing batching but I don't see anywhere in the People API documentation that specifies which methods can and can't be batched.
Here is a variable dump (using Laravel's `dump()` function) for the Google_Http_Batch object:
```
Google_Http_Batch {#733 ▼
-boundary: 350902964
-requests: array:1 [▼
1 => Request {#764 ▼
-method: "POST"
-requestTarget: null
-uri: Uri {#766 ▼
-scheme: "https"
-userInfo: ""
-host: "people.googleapis.com"
-port: null
-path: "/v1/people:createContact"
-query: ""
-fragment: ""
}
-headers: array:3 [▶]
-headerNames: array:3 [▶]
-protocol: "1.1"
-stream: Stream {#765 ▶}
}
]
-client: Google_Client {#732 ▶}
-rootUrl: "https://www.googleapis.com"
-batchPath: "batch"
}
```
And here is both the request and response bodies for the attempted operation:
```
// Request Body
--350902964\n
Content-Type: application/http
Content-Transfer-Encoding: binary
MIME-Version: 1.0
Content-ID: 1
POST /v1/people:createContact HTTP/1.1
Host:people.googleapis.com
content-type:application/json
X-Php-Expected-Class:Google_Service_PeopleService_Person
{"names":[{"familyName":"Last","givenName":"First"}],"birthdays":[{"text":"1975-01-01"}],"emailAddresses":[{"type":"other","value":"firstlast@gmail.com"}],"phoneNumbers":[{"type":"other","value":"+18005555555"}]}\n
\n
--350902964--
// Response Body
--batch_pb1ntA148M4_AAcIp_VroDQ
Content-Type: application/http
Content-ID: response-1
\r\n
HTTP/1.1 404 Not Found
Content-Type: text/html; charset=UTF-8
Date: Thu, 19 Oct 2017 20:44:28 GMT
Expires: Thu, 19 Oct 2017 20:44:28 GMT
Cache-Control: private, max-age=0
Content-Length: 9
Not Found
--batch_pb1ntA148M4_AAcIp_VroDQ--
```
If this appears to be an issue with the `Google_Service_PeopleService` class instead of the `Google_Http_Batch` class, I'll be more than happy to create an issue on the [google-api-php-client-services](https://github.com/google/google-api-php-client-services) and close this one.
Any clarification on whether I'm doing anything incorrectly or if I'm even able to batch this method at all would be greatly appreciated!
| 1.0 | Batch operations always returning 404: Not Found response - Using the Google_Service_PeopleService API provided in the [google-api-php-client-services](https://github.com/google/google-api-php-client-services) library, I'm attempting to create contacts via the Google_Http_Batch class.
I've tested executing the request on its own prior to adding it to the batch and it works correctly, however when attempting to use batching it always returns a 404 Not Found Exception as the Response.
I searched around for similar issues and it seems like this can be caused by methods not allowing batching but I don't see anywhere in the People API documentation that specifies which methods can and can't be batched.
Here is a variable dump (using Laravel's `dump()` function) for the Google_Http_Batch object:
```
Google_Http_Batch {#733 ▼
-boundary: 350902964
-requests: array:1 [▼
1 => Request {#764 ▼
-method: "POST"
-requestTarget: null
-uri: Uri {#766 ▼
-scheme: "https"
-userInfo: ""
-host: "people.googleapis.com"
-port: null
-path: "/v1/people:createContact"
-query: ""
-fragment: ""
}
-headers: array:3 [▶]
-headerNames: array:3 [▶]
-protocol: "1.1"
-stream: Stream {#765 ▶}
}
]
-client: Google_Client {#732 ▶}
-rootUrl: "https://www.googleapis.com"
-batchPath: "batch"
}
```
And here is both the request and response bodies for the attempted operation:
```
// Request Body
--350902964\n
Content-Type: application/http
Content-Transfer-Encoding: binary
MIME-Version: 1.0
Content-ID: 1
POST /v1/people:createContact HTTP/1.1
Host:people.googleapis.com
content-type:application/json
X-Php-Expected-Class:Google_Service_PeopleService_Person
{"names":[{"familyName":"Last","givenName":"First"}],"birthdays":[{"text":"1975-01-01"}],"emailAddresses":[{"type":"other","value":"firstlast@gmail.com"}],"phoneNumbers":[{"type":"other","value":"+18005555555"}]}\n
\n
--350902964--
// Response Body
--batch_pb1ntA148M4_AAcIp_VroDQ
Content-Type: application/http
Content-ID: response-1
\r\n
HTTP/1.1 404 Not Found
Content-Type: text/html; charset=UTF-8
Date: Thu, 19 Oct 2017 20:44:28 GMT
Expires: Thu, 19 Oct 2017 20:44:28 GMT
Cache-Control: private, max-age=0
Content-Length: 9
Not Found
--batch_pb1ntA148M4_AAcIp_VroDQ--
```
If this appears to be an issue with the `Google_Service_PeopleService` class instead of the `Google_Http_Batch` class, I'll be more than happy to create an issue on the [google-api-php-client-services](https://github.com/google/google-api-php-client-services) and close this one.
Any clarification on whether I'm doing anything incorrectly or if I'm even able to batch this method at all would be greatly appreciated!
| non_process | batch operations always returning not found response using the google service peopleservice api provided in the library i m attempting to create contacts via the google http batch class i ve tested executing the request on its own prior to adding it to the batch and it works correctly however when attempting to use batching it always returns a not found exception as the response i searched around for similar issues and it seems like this can be caused by methods not allowing batching but i don t see anywhere in the people api documentation that specifies which methods can and can t be batched here is a variable dump using laravel s dump function for the google http batch object google http batch ▼ boundary requests array ▼ request ▼ method post requesttarget null uri uri ▼ scheme https userinfo host people googleapis com port null path people createcontact query fragment headers array headernames array protocol stream stream ▶ client google client ▶ rooturl batchpath batch and here is both the request and response bodies for the attempted operation request body n content type application http content transfer encoding binary mime version content id post people createcontact http host people googleapis com content type application json x php expected class google service peopleservice person names birthdays emailaddresses phonenumbers n n response body batch aacip vrodq content type application http content id response r n http not found content type text html charset utf date thu oct gmt expires thu oct gmt cache control private max age content length not found batch aacip vrodq if this appears to be an issue with the google service peopleservice class instead of the google http batch class i ll be more than happy to create an issue on the and close this one any clarification on whether i m doing anything incorrectly or if i m even able to batch this method at all would be greatly appreciated | 0 |
8,223 | 11,410,593,143 | IssuesEvent | 2020-02-01 00:03:21 | parcel-bundler/parcel | https://api.github.com/repos/parcel-bundler/parcel | closed | cannot resolve sass dependencies that have a "sass" entry at their package.json | :bug: Bug CSS Preprocessing Stale | # 🐛 bug report
When trying to import an installed sass module, in this case bootstrap (for testing), import fails.
## 🎛 Configuration (.babelrc, package.json, cli command)
```js
// not related but .babelrc
module.exports = api => {
api.cache(true);
const presets = ["@babel/preset-env"];
return {
presets
}
}
```
```json
{
"name": "parcel-sanity-check",
"version": "1.0.0",
"main": "index.js",
"repository": "git@github.com:renatodeleao/parcel-sanity-check.git",
"author": "Renato de Leão <renatodeleao@gmail.com>",
"license": "MIT",
"scripts": {
"dev:site": "parcel site/index.html --out-dir public --no-cache",
"build": "parcel build site/index.html --out-dir public"
},
"devDependencies": {
"autoprefixer": "^9.4.7",
"parcel-bundler": "^1.11.0",
"posthtml": "^0.11.3",
"posthtml-expressions": "^1.1.0",
"posthtml-extend": "^0.3.0",
"posthtml-modules": "^0.4.2",
"sass": "^1.17.0"
},
"dependencies": {
"@babel/core": "^7.2.2",
"@babel/preset-env": "^7.3.1",
"@babel/register": "^7.0.0",
"@whitesmith/qnorr-styles": "^0.1.1",
"bootstrap": "4.3.0"
}
}
```
## 🤔 Expected Behavior
It should import correctly an external sass modules dependency
## 😯 Current Behavior
It's throwing the following error because parcel expects `sass` from `package.json` to be an object, but bootstrap ships with `sass: "scss/bootstrap.scss"`
```
🚨 /localPath/node_modules/bootstrap/scss/bootstrap.scss: Cannot create property 'includePaths' on string 'scss/bootstrap.scss'
at SASSAsset.parse (/localPath/parcel-sanity-check/node_modules/parcel-bundler/src/assets/SASSAsset.js:27:23)
```
## 💁 Possible Solution
Don't have a solution, and non of the suggested at https://github.com/parcel-bundler/parcel/issues/645 which blame postCSS work for me ( i don't think postCSS has anything to do with my case)
I may have a suggestion:
What if entries in `package.json` that parcel uses for configuration were scoped under `parcel` namespace. Otherwise, packages like `bootstrap` that uses (the non-standard) `sass` will break. It's a non standard prop but, you know, they are kind of big and famous.
Note: This would be a breaking change, so i do not expect it to be shipped as fix for this.
```js
"parcel": {
sass: "",
posthtml: ""
...
}
```
## 🔦 Context
I accidentally discovered this because i am developing my companiy's own sass framework. I was looking for references on best practices for bundling and bootstrap immediately popped up. Then I found this two lines i've never seen in a `package.json` before `style` and a `sass` entries
https://github.com/twbs/bootstrap/blob/45ced606f479ecad1a8e5858367dbb3fcf1a9ee7/package.json#L88-L90
I try to dig on the webs
https://jaketrent.com/post/package-json-style-attribute/
https://stackoverflow.com/questions/32037150/style-field-in-package-json
so it looks like both are non standard, but `style` is used by `postcss-import`.
In my innocence i said:
> well they can't do no harm
It did, i was testing a simple parcel website with my new framework and realised that i couldn't import my library. I deleted the line `sass` from `package.json`, *et voilá*
## 💻 Code Sample
https://github.com/renatodeleao/parcel-sanity-check
## 🌍 Your Environment
| Software | Version(s) |
| ---------------- | ---------- |
| Parcel | 1.11.0
| Node | v11.8.0
| npm/Yarn | 1.13.0
| Operating System | macOs 10.14 | 1.0 | cannot resolve sass dependencies that have a "sass" entry at their package.json - # 🐛 bug report
When trying to import an installed sass module, in this case bootstrap (for testing), import fails.
## 🎛 Configuration (.babelrc, package.json, cli command)
```js
// not related but .babelrc
module.exports = api => {
api.cache(true);
const presets = ["@babel/preset-env"];
return {
presets
}
}
```
```json
{
"name": "parcel-sanity-check",
"version": "1.0.0",
"main": "index.js",
"repository": "git@github.com:renatodeleao/parcel-sanity-check.git",
"author": "Renato de Leão <renatodeleao@gmail.com>",
"license": "MIT",
"scripts": {
"dev:site": "parcel site/index.html --out-dir public --no-cache",
"build": "parcel build site/index.html --out-dir public"
},
"devDependencies": {
"autoprefixer": "^9.4.7",
"parcel-bundler": "^1.11.0",
"posthtml": "^0.11.3",
"posthtml-expressions": "^1.1.0",
"posthtml-extend": "^0.3.0",
"posthtml-modules": "^0.4.2",
"sass": "^1.17.0"
},
"dependencies": {
"@babel/core": "^7.2.2",
"@babel/preset-env": "^7.3.1",
"@babel/register": "^7.0.0",
"@whitesmith/qnorr-styles": "^0.1.1",
"bootstrap": "4.3.0"
}
}
```
## 🤔 Expected Behavior
It should import correctly an external sass modules dependency
## 😯 Current Behavior
It's throwing the following error because parcel expects `sass` from `package.json` to be an object, but bootstrap ships with `sass: "scss/bootstrap.scss"`
```
🚨 /localPath/node_modules/bootstrap/scss/bootstrap.scss: Cannot create property 'includePaths' on string 'scss/bootstrap.scss'
at SASSAsset.parse (/localPath/parcel-sanity-check/node_modules/parcel-bundler/src/assets/SASSAsset.js:27:23)
```
## 💁 Possible Solution
Don't have a solution, and non of the suggested at https://github.com/parcel-bundler/parcel/issues/645 which blame postCSS work for me ( i don't think postCSS has anything to do with my case)
I may have a suggestion:
What if entries in `package.json` that parcel uses for configuration were scoped under `parcel` namespace. Otherwise, packages like `bootstrap` that uses (the non-standard) `sass` will break. It's a non standard prop but, you know, they are kind of big and famous.
Note: This would be a breaking change, so i do not expect it to be shipped as fix for this.
```js
"parcel": {
sass: "",
posthtml: ""
...
}
```
## 🔦 Context
I accidentally discovered this because i am developing my companiy's own sass framework. I was looking for references on best practices for bundling and bootstrap immediately popped up. Then I found this two lines i've never seen in a `package.json` before `style` and a `sass` entries
https://github.com/twbs/bootstrap/blob/45ced606f479ecad1a8e5858367dbb3fcf1a9ee7/package.json#L88-L90
I try to dig on the webs
https://jaketrent.com/post/package-json-style-attribute/
https://stackoverflow.com/questions/32037150/style-field-in-package-json
so it looks like both are non standard, but `style` is used by `postcss-import`.
In my innocence i said:
> well they can't do no harm
It did, i was testing a simple parcel website with my new framework and realised that i couldn't import my library. I deleted the line `sass` from `package.json`, *et voilá*
## 💻 Code Sample
https://github.com/renatodeleao/parcel-sanity-check
## 🌍 Your Environment
| Software | Version(s) |
| ---------------- | ---------- |
| Parcel | 1.11.0
| Node | v11.8.0
| npm/Yarn | 1.13.0
| Operating System | macOs 10.14 | process | cannot resolve sass dependencies that have a sass entry at their package json 🐛 bug report when trying to import an installed sass module in this case bootstrap for testing import fails 🎛 configuration babelrc package json cli command js not related but babelrc module exports api api cache true const presets return presets json name parcel sanity check version main index js repository git github com renatodeleao parcel sanity check git author renato de leão license mit scripts dev site parcel site index html out dir public no cache build parcel build site index html out dir public devdependencies autoprefixer parcel bundler posthtml posthtml expressions posthtml extend posthtml modules sass dependencies babel core babel preset env babel register whitesmith qnorr styles bootstrap 🤔 expected behavior it should import correctly an external sass modules dependency 😯 current behavior it s throwing the following error because parcel expects sass from package json to be an object but bootstrap ships with sass scss bootstrap scss 🚨 localpath node modules bootstrap scss bootstrap scss cannot create property includepaths on string scss bootstrap scss at sassasset parse localpath parcel sanity check node modules parcel bundler src assets sassasset js 💁 possible solution don t have a solution and non of the suggested at which blame postcss work for me i don t think postcss has anything to do with my case i may have a suggestion what if entries in package json that parcel uses for configuration were scoped under parcel namespace otherwise packages like bootstrap that uses the non standard sass will break it s a non standard prop but you know they are kind of big and famous note this would be a breaking change so i do not expect it to be shipped as fix for this js parcel sass posthtml 🔦 context i accidentally discovered this because i am developing my companiy s own sass framework i was looking for references on best practices for bundling and bootstrap immediately popped up then i found this two lines i ve never seen in a package json before style and a sass entries i try to dig on the webs so it looks like both are non standard but style is used by postcss import in my innocence i said well they can t do no harm it did i was testing a simple parcel website with my new framework and realised that i couldn t import my library i deleted the line sass from package json et voilá 💻 code sample 🌍 your environment software version s parcel node npm yarn operating system macos | 1 |
945 | 3,410,649,937 | IssuesEvent | 2015-12-04 21:12:02 | MaretEngineering/MROV | https://api.github.com/repos/MaretEngineering/MROV | closed | Put variable list into array | enhancement Processing | The variables where we currently store our joystick positions are
```Java
int joy1x = 0;
int joy1y = 0;
int joy2x = 0;
int joy2y = 0;
int rTrig = 0;
int lTrig = 0;
```
These should be in arrays for better organization. Maybe
```Java
int[] joy1 = {0, 0}; // x,y
int[] joy2 = ...
``` | 1.0 | Put variable list into array - The variables where we currently store our joystick positions are
```Java
int joy1x = 0;
int joy1y = 0;
int joy2x = 0;
int joy2y = 0;
int rTrig = 0;
int lTrig = 0;
```
These should be in arrays for better organization. Maybe
```Java
int[] joy1 = {0, 0}; // x,y
int[] joy2 = ...
``` | process | put variable list into array the variables where we currently store our joystick positions are java int int int int int rtrig int ltrig these should be in arrays for better organization maybe java int x y int | 1 |
239,765 | 7,799,998,288 | IssuesEvent | 2018-06-09 03:14:49 | tine20/Tine-2.0-Open-Source-Groupware-and-CRM | https://api.github.com/repos/tine20/Tine-2.0-Open-Source-Groupware-and-CRM | closed | 0006546:
saving user with duplicate imap/smtp user entry fails | Admin Bug Mantis high priority | **Reported by pschuele on 31 May 2012 15:10**
**Version:** Milan (2012-03-3)
saving user with duplicate imap/smtp user entry fails
- js error: Uncaught TypeError: Cannot read property 'length' of null
- perhaps we should update this row if user id is not known in tine
- check if it throws correct exception
| 1.0 | 0006546:
saving user with duplicate imap/smtp user entry fails - **Reported by pschuele on 31 May 2012 15:10**
**Version:** Milan (2012-03-3)
saving user with duplicate imap/smtp user entry fails
- js error: Uncaught TypeError: Cannot read property 'length' of null
- perhaps we should update this row if user id is not known in tine
- check if it throws correct exception
| non_process | saving user with duplicate imap smtp user entry fails reported by pschuele on may version milan saving user with duplicate imap smtp user entry fails js error uncaught typeerror cannot read property length of null perhaps we should update this row if user id is not known in tine check if it throws correct exception | 0 |
738,544 | 25,565,606,944 | IssuesEvent | 2022-11-30 14:06:04 | kubermatic/dashboard | https://api.github.com/repos/kubermatic/dashboard | opened | Initial Node step is broken for editing Cluster Templates | kind/bug priority/high sig/ui | ### What happened
https://github.com/kubermatic/dashboard/pull/5258 broke the edit cluster template workflow. There are errors in the `Initial Node` step in the wizard.
### Expected behavior
### How to reproduce
### Environment
- UI Version:
- API Version:
- Domain:
- Others:
### Current workaround
### Affected user persona
### Business goal to be improved
### Metric to be improved
| 1.0 | Initial Node step is broken for editing Cluster Templates - ### What happened
https://github.com/kubermatic/dashboard/pull/5258 broke the edit cluster template workflow. There are errors in the `Initial Node` step in the wizard.
### Expected behavior
### How to reproduce
### Environment
- UI Version:
- API Version:
- Domain:
- Others:
### Current workaround
### Affected user persona
### Business goal to be improved
### Metric to be improved
| non_process | initial node step is broken for editing cluster templates what happened broke the edit cluster template workflow there are errors in the initial node step in the wizard expected behavior how to reproduce environment ui version api version domain others current workaround affected user persona business goal to be improved metric to be improved | 0 |
130,703 | 18,170,256,818 | IssuesEvent | 2021-09-27 19:06:55 | snowdensb/job-dsl-plugin | https://api.github.com/repos/snowdensb/job-dsl-plugin | opened | CVE-2021-21610 (Medium) detected in jenkins-core-2.176.jar | security vulnerability | ## CVE-2021-21610 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jenkins-core-2.176.jar</b></p></summary>
<p>Jenkins core code and view files to render HTML.</p>
<p>Library home page: <a href="https://jenkins.io/jenkins-parent/jenkins-core/">https://jenkins.io/jenkins-parent/jenkins-core/</a></p>
<p>Path to dependency file: job-dsl-plugin/job-dsl-plugin/build.gradle</p>
<p>Path to vulnerable library: caches/modules-2/files-2.1/org.jenkins-ci.main/jenkins-core/2.176/95ee06bed42207774c63dc8ead89f79f2a9daee9/jenkins-core-2.176.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.jenkins-ci.main/jenkins-core/2.176/95ee06bed42207774c63dc8ead89f79f2a9daee9/jenkins-core-2.176.jar</p>
<p>
Dependency Hierarchy:
- :x: **jenkins-core-2.176.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/snowdensb/job-dsl-plugin/commit/2e8946fc25b6060b5637631b8813fe86b6da687a">2e8946fc25b6060b5637631b8813fe86b6da687a</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Jenkins 2.274 and earlier, LTS 2.263.1 and earlier does not implement any restrictions for the URL rendering a formatted preview of markup passed as a query parameter, resulting in a reflected cross-site scripting (XSS) vulnerability if the configured markup formatter does not prohibit unsafe elements (JavaScript) in markup.
<p>Publish Date: 2021-01-13
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21610>CVE-2021-21610</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.jenkins.io/security/advisory/2021-01-13/">https://www.jenkins.io/security/advisory/2021-01-13/</a></p>
<p>Release Date: 2021-01-13</p>
<p>Fix Resolution: org.jenkins-ci.main:jenkins-core:2.275, org.jenkins-ci.main:jenkins-core:LTS 2.263.2</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.jenkins-ci.main","packageName":"jenkins-core","packageVersion":"2.176","packageFilePaths":["/job-dsl-plugin/build.gradle"],"isTransitiveDependency":false,"dependencyTree":"org.jenkins-ci.main:jenkins-core:2.176","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.jenkins-ci.main:jenkins-core:2.275, org.jenkins-ci.main:jenkins-core:LTS 2.263.2"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-21610","vulnerabilityDetails":"Jenkins 2.274 and earlier, LTS 2.263.1 and earlier does not implement any restrictions for the URL rendering a formatted preview of markup passed as a query parameter, resulting in a reflected cross-site scripting (XSS) vulnerability if the configured markup formatter does not prohibit unsafe elements (JavaScript) in markup.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21610","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> --> | True | CVE-2021-21610 (Medium) detected in jenkins-core-2.176.jar - ## CVE-2021-21610 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jenkins-core-2.176.jar</b></p></summary>
<p>Jenkins core code and view files to render HTML.</p>
<p>Library home page: <a href="https://jenkins.io/jenkins-parent/jenkins-core/">https://jenkins.io/jenkins-parent/jenkins-core/</a></p>
<p>Path to dependency file: job-dsl-plugin/job-dsl-plugin/build.gradle</p>
<p>Path to vulnerable library: caches/modules-2/files-2.1/org.jenkins-ci.main/jenkins-core/2.176/95ee06bed42207774c63dc8ead89f79f2a9daee9/jenkins-core-2.176.jar,/home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.jenkins-ci.main/jenkins-core/2.176/95ee06bed42207774c63dc8ead89f79f2a9daee9/jenkins-core-2.176.jar</p>
<p>
Dependency Hierarchy:
- :x: **jenkins-core-2.176.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/snowdensb/job-dsl-plugin/commit/2e8946fc25b6060b5637631b8813fe86b6da687a">2e8946fc25b6060b5637631b8813fe86b6da687a</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Jenkins 2.274 and earlier, LTS 2.263.1 and earlier does not implement any restrictions for the URL rendering a formatted preview of markup passed as a query parameter, resulting in a reflected cross-site scripting (XSS) vulnerability if the configured markup formatter does not prohibit unsafe elements (JavaScript) in markup.
<p>Publish Date: 2021-01-13
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21610>CVE-2021-21610</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.jenkins.io/security/advisory/2021-01-13/">https://www.jenkins.io/security/advisory/2021-01-13/</a></p>
<p>Release Date: 2021-01-13</p>
<p>Fix Resolution: org.jenkins-ci.main:jenkins-core:2.275, org.jenkins-ci.main:jenkins-core:LTS 2.263.2</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.jenkins-ci.main","packageName":"jenkins-core","packageVersion":"2.176","packageFilePaths":["/job-dsl-plugin/build.gradle"],"isTransitiveDependency":false,"dependencyTree":"org.jenkins-ci.main:jenkins-core:2.176","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.jenkins-ci.main:jenkins-core:2.275, org.jenkins-ci.main:jenkins-core:LTS 2.263.2"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-21610","vulnerabilityDetails":"Jenkins 2.274 and earlier, LTS 2.263.1 and earlier does not implement any restrictions for the URL rendering a formatted preview of markup passed as a query parameter, resulting in a reflected cross-site scripting (XSS) vulnerability if the configured markup formatter does not prohibit unsafe elements (JavaScript) in markup.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21610","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> --> | non_process | cve medium detected in jenkins core jar cve medium severity vulnerability vulnerable library jenkins core jar jenkins core code and view files to render html library home page a href path to dependency file job dsl plugin job dsl plugin build gradle path to vulnerable library caches modules files org jenkins ci main jenkins core jenkins core jar home wss scanner gradle caches modules files org jenkins ci main jenkins core jenkins core jar dependency hierarchy x jenkins core jar vulnerable library found in head commit a href found in base branch master vulnerability details jenkins and earlier lts and earlier does not implement any restrictions for the url rendering a formatted preview of markup passed as a query parameter resulting in a reflected cross site scripting xss vulnerability if the configured markup formatter does not prohibit unsafe elements javascript in markup publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org jenkins ci main jenkins core org jenkins ci main jenkins core lts rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree org jenkins ci main jenkins core isminimumfixversionavailable true minimumfixversion org jenkins ci main jenkins core org jenkins ci main jenkins core lts basebranches vulnerabilityidentifier cve vulnerabilitydetails jenkins and earlier lts and earlier does not implement any restrictions for the url rendering a formatted preview of markup passed as a query parameter resulting in a reflected cross site scripting xss vulnerability if the configured markup formatter does not prohibit unsafe elements javascript in markup vulnerabilityurl | 0 |
271,801 | 29,659,401,007 | IssuesEvent | 2023-06-10 01:27:09 | pazhanivel07/linux-4.19.72 | https://api.github.com/repos/pazhanivel07/linux-4.19.72 | closed | CVE-2019-19071 (High) detected in linuxlinux-4.19.83 - autoclosed | Mend: dependency security vulnerability | ## CVE-2019-19071 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.83</b></p></summary>
<p>
<p>Apache Software Foundation (ASF)</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/pazhanivel07/linux-4.19.72/commit/ce28e4f7a922d93d9b737061ae46827305c8c30a">ce28e4f7a922d93d9b737061ae46827305c8c30a</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/wireless/rsi/rsi_91x_mgmt.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/wireless/rsi/rsi_91x_mgmt.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
A memory leak in the rsi_send_beacon() function in drivers/net/wireless/rsi/rsi_91x_mgmt.c in the Linux kernel through 5.3.11 allows attackers to cause a denial of service (memory consumption) by triggering rsi_prepare_beacon() failures, aka CID-d563131ef23c.
<p>Publish Date: 2019-11-18
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-19071>CVE-2019-19071</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2019-19071">https://www.linuxkernelcves.com/cves/CVE-2019-19071</a></p>
<p>Release Date: 2020-08-24</p>
<p>Fix Resolution: v5.5-rc1,v4.14.159,v4.19.89,v5.3.16,v5.4.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2019-19071 (High) detected in linuxlinux-4.19.83 - autoclosed - ## CVE-2019-19071 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.83</b></p></summary>
<p>
<p>Apache Software Foundation (ASF)</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/pazhanivel07/linux-4.19.72/commit/ce28e4f7a922d93d9b737061ae46827305c8c30a">ce28e4f7a922d93d9b737061ae46827305c8c30a</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/wireless/rsi/rsi_91x_mgmt.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/wireless/rsi/rsi_91x_mgmt.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
A memory leak in the rsi_send_beacon() function in drivers/net/wireless/rsi/rsi_91x_mgmt.c in the Linux kernel through 5.3.11 allows attackers to cause a denial of service (memory consumption) by triggering rsi_prepare_beacon() failures, aka CID-d563131ef23c.
<p>Publish Date: 2019-11-18
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-19071>CVE-2019-19071</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2019-19071">https://www.linuxkernelcves.com/cves/CVE-2019-19071</a></p>
<p>Release Date: 2020-08-24</p>
<p>Fix Resolution: v5.5-rc1,v4.14.159,v4.19.89,v5.3.16,v5.4.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_process | cve high detected in linuxlinux autoclosed cve high severity vulnerability vulnerable library linuxlinux apache software foundation asf library home page a href found in head commit a href found in base branch master vulnerable source files drivers net wireless rsi rsi mgmt c drivers net wireless rsi rsi mgmt c vulnerability details a memory leak in the rsi send beacon function in drivers net wireless rsi rsi mgmt c in the linux kernel through allows attackers to cause a denial of service memory consumption by triggering rsi prepare beacon failures aka cid publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend | 0 |
8,220 | 11,410,592,214 | IssuesEvent | 2020-02-01 00:03:09 | parcel-bundler/parcel | https://api.github.com/repos/parcel-bundler/parcel | closed | css keyframes are missing after building | :bug: Bug CSS Preprocessing Stale | <!---
Thanks for filing an issue 😄 ! Before you submit, please read the following:
Search open/closed issues before submitting since someone might have asked the same thing before!
-->
**Choose one:** is this a 🐛 bug report or 🙋 feature request?
🐛 bug report
<!--- Provide a general summary of the issue in the title above -->
### 🎛 Configuration (.babelrc, package.json, cli command)
<!--- If describing a bug, tell us what your babel configuration looks like -->
```js
parcel build index.html -d build --public-url ./
```
### 🤔 Expected Behavior
No bug occurred
<!--- If you're describing a bug, tell us what should happen -->
<!--- If you're suggesting a change/improvement, tell us how it should work -->
### 😯 Current Behavior
My page cannot show loading animation
<!--- If describing a bug, tell us what happens instead of the expected behavior -->
<!--- If you are seeing an error, please include the full error message and stack trace -->
<!--- If suggesting a change/improvement, explain the difference from current behavior -->
### 💁 Possible Solution
set option below:
https://github.com/ben-eb/cssnano/issues/351#issuecomment-277406846
<!--- Not obligatory, but suggest a fix/reason for the bug, -->
<!--- or ideas how to implement the addition or change -->
### 🔦 Context
After finishing building my project, I found that all of the loading animations cannot show, then I tried to check @keyframes in my stylesheet but found nothing. But it surely exists in my source code.
<!--- How has this issue affected you? What are you trying to accomplish? -->
<!--- Providing context helps us come up with a solution that is most useful in the real world -->
### 💻 Code Sample
`import './assets/semantic/dist/semantic.css'`
<!-- If you are seeing an error, please provide a code repository, gist or sample files to reproduce the issue -->
### 🌍 Your Environment
<!--- Include as many relevant details about the environment you experienced the bug in -->
| Software | Version(s) |
| ---------------- | ---------- |
| Parcel |1.5.1
| Node |9.6.0
| npm/Yarn |5.6.0
| Operating System |windows 10
<!-- Love parcel? Please consider supporting our collective:
👉 https://opencollective.com/parcel/donate -->
| 1.0 | css keyframes are missing after building - <!---
Thanks for filing an issue 😄 ! Before you submit, please read the following:
Search open/closed issues before submitting since someone might have asked the same thing before!
-->
**Choose one:** is this a 🐛 bug report or 🙋 feature request?
🐛 bug report
<!--- Provide a general summary of the issue in the title above -->
### 🎛 Configuration (.babelrc, package.json, cli command)
<!--- If describing a bug, tell us what your babel configuration looks like -->
```js
parcel build index.html -d build --public-url ./
```
### 🤔 Expected Behavior
No bug occurred
<!--- If you're describing a bug, tell us what should happen -->
<!--- If you're suggesting a change/improvement, tell us how it should work -->
### 😯 Current Behavior
My page cannot show loading animation
<!--- If describing a bug, tell us what happens instead of the expected behavior -->
<!--- If you are seeing an error, please include the full error message and stack trace -->
<!--- If suggesting a change/improvement, explain the difference from current behavior -->
### 💁 Possible Solution
set option below:
https://github.com/ben-eb/cssnano/issues/351#issuecomment-277406846
<!--- Not obligatory, but suggest a fix/reason for the bug, -->
<!--- or ideas how to implement the addition or change -->
### 🔦 Context
After finishing building my project, I found that all of the loading animations cannot show, then I tried to check @keyframes in my stylesheet but found nothing. But it surely exists in my source code.
<!--- How has this issue affected you? What are you trying to accomplish? -->
<!--- Providing context helps us come up with a solution that is most useful in the real world -->
### 💻 Code Sample
`import './assets/semantic/dist/semantic.css'`
<!-- If you are seeing an error, please provide a code repository, gist or sample files to reproduce the issue -->
### 🌍 Your Environment
<!--- Include as many relevant details about the environment you experienced the bug in -->
| Software | Version(s) |
| ---------------- | ---------- |
| Parcel |1.5.1
| Node |9.6.0
| npm/Yarn |5.6.0
| Operating System |windows 10
<!-- Love parcel? Please consider supporting our collective:
👉 https://opencollective.com/parcel/donate -->
| process | css keyframes are missing after building thanks for filing an issue 😄 before you submit please read the following search open closed issues before submitting since someone might have asked the same thing before choose one is this a 🐛 bug report or 🙋 feature request 🐛 bug report 🎛 configuration babelrc package json cli command js parcel build index html d build public url 🤔 expected behavior no bug occurred 😯 current behavior my page cannot show loading animation 💁 possible solution set option below 🔦 context after finishing building my project i found that all of the loading animations cannot show then i tried to check keyframes in my stylesheet but found nothing but it surely exists in my source code 💻 code sample import assets semantic dist semantic css 🌍 your environment software version s parcel node npm yarn operating system windows love parcel please consider supporting our collective 👉 | 1 |
9,707 | 12,703,521,034 | IssuesEvent | 2020-06-22 22:32:18 | qgis/QGIS | https://api.github.com/repos/qgis/QGIS | closed | Using some functions in raster calulator crashes QGIS 3.10/3.12. | Bug Crash/Data Corruption Processing | Hey,
I am on Linux Mint 19.2 Cinnamon using QGIS 3.10 LTR apt package. Every time I use the native raster calculator with an expression, such as the following:
`("slope@1" > (6.94*0.85+((max("slope@1")-min("slope@1")))/4))*1`
it crashes the whole QGIS without any message.
The expression is valid as well as the raster layer I'm using. I suspect there is a problem with min/max functions, however, not sure about it, and not sure which other functions might be involved.
The expression: `("slope@1" > (6.94*0.85))*1` returned a valid result.
I reproduced the same issue on the 3.12 version.
I launched QGIS 3.12 through the console and captured this log on the crash:
```
Warning: QStandardPaths: XDG_RUNTIME_DIR not set, defaulting to '/tmp/runtime-root'
QGIS died on signal 11[New LWP 5663]
[New LWP 5664]
[New LWP 5665]
[New LWP 5666]
[New LWP 5673]
[New LWP 5674]
[New LWP 5675]
[New LWP 5676]
[New LWP 5679]
[New LWP 5690]
[New LWP 5691]
[New LWP 5692]
[Thread debugging using libthread_db enabled]
Using host libthread_db library "/lib/x86_64-linux-gnu/libthread_db.so.1".
0x00007fda6e2036c2 in __GI___waitpid (pid=5695, stat_loc=0x7ffeb33be6ec, options=0) at ../sysdeps/unix/sysv/linux/waitpid.c:30
30 ../sysdeps/unix/sysv/linux/waitpid.c: No such file or directory.
[Current thread is 1 (Thread 0x7fda732fe3c0 (LWP 5660))]
#0 0x00007fda6e2036c2 in __GI___waitpid (pid=5695, stat_loc=0x7ffeb33be6ec, options=0) at ../sysdeps/unix/sysv/linux/waitpid.c:30
resultvar = 18446744073709551104
sc_cancel_oldtype = 0
#1 0x000055ec465132be in ()
#2 0x00007fda6e15df20 in <signal handler called> () at /lib/x86_64-linux-gnu/libc.so.6
#3 0x00007fda6c855981 in QgsRasterMatrix::twoArgumentOperation(QgsRasterMatrix::TwoArgOperator, QgsRasterMatrix const&) () at /usr/lib/libqgis_analysis.so.3.12.1
#4 0x00007fda6c846aff in QgsRasterCalcNode::calculate(QMap<QString, QgsRasterBlock*>&, QgsRasterMatrix&, int) const () at /usr/lib/libqgis_analysis.so.3.12.1
#5 0x00007fda6c846a7a in QgsRasterCalcNode::calculate(QMap<QString, QgsRasterBlock*>&, QgsRasterMatrix&, int) const () at /usr/lib/libqgis_analysis.so.3.12.1
#6 0x00007fda6c846a7a in QgsRasterCalcNode::calculate(QMap<QString, QgsRasterBlock*>&, QgsRasterMatrix&, int) const () at /usr/lib/libqgis_analysis.so.3.12.1
#7 0x00007fda6c846ab1 in QgsRasterCalcNode::calculate(QMap<QString, QgsRasterBlock*>&, QgsRasterMatrix&, int) const () at /usr/lib/libqgis_analysis.so.3.12.1
#8 0x00007fda6c846ab1 in QgsRasterCalcNode::calculate(QMap<QString, QgsRasterBlock*>&, QgsRasterMatrix&, int) const () at /usr/lib/libqgis_analysis.so.3.12.1
#9 0x00007fda6c846a7a in QgsRasterCalcNode::calculate(QMap<QString, QgsRasterBlock*>&, QgsRasterMatrix&, int) const () at /usr/lib/libqgis_analysis.so.3.12.1
#10 0x00007fda6c850980 in QgsRasterCalculator::processCalculation(QgsFeedback*) () at /usr/lib/libqgis_analysis.so.3.12.1
#11 0x00007fda72773b4c in QgisApp::showRasterCalculator() () at /usr/lib/libqgis_app.so.3.12.1
#12 0x00007fda6ed6a66f in QMetaObject::activate(QObject*, int, int, void**) () at /usr/lib/x86_64-linux-gnu/libQt5Core.so.5
#13 0x00007fda6fab3122 in QAction::triggered(bool) () at /usr/lib/x86_64-linux-gnu/libQt5Widgets.so.5
#14 0x00007fda6fab580c in QAction::activate(QAction::ActionEvent) () at /usr/lib/x86_64-linux-gnu/libQt5Widgets.so.5
#15 0x00007fda6fc320fc in () at /usr/lib/x86_64-linux-gnu/libQt5Widgets.so.5
#16 0x00007fda6fc395ab in () at /usr/lib/x86_64-linux-gnu/libQt5Widgets.so.5
#17 0x00007fda6fc3a3d3 in QMenu::mouseReleaseEvent(QMouseEvent*) () at /usr/lib/x86_64-linux-gnu/libQt5Widgets.so.5
#18 0x00007fda6faf8048 in QWidget::event(QEvent*) () at /usr/lib/x86_64-linux-gnu/libQt5Widgets.so.5
#19 0x00007fda6fc3c6fb in QMenu::event(QEvent*) () at /usr/lib/x86_64-linux-gnu/libQt5Widgets.so.5
#20 0x00007fda6fab983c in QApplicationPrivate::notify_helper(QObject*, QEvent*) () at /usr/lib/x86_64-linux-gnu/libQt5Widgets.so.5
#21 0x00007fda6fac165f in QApplication::notify(QObject*, QEvent*) () at /usr/lib/x86_64-linux-gnu/libQt5Widgets.so.5
#22 0x00007fda709b434b in QgsApplication::notify(QObject*, QEvent*) () at /usr/lib/libqgis_core.so.3.12.1
#23 0x00007fda6ed3b8d8 in QCoreApplication::notifyInternal2(QObject*, QEvent*) () at /usr/lib/x86_64-linux-gnu/libQt5Core.so.5
#24 0x00007fda6fac0632 in QApplicationPrivate::sendMouseEvent(QWidget*, QMouseEvent*, QWidget*, QWidget*, QWidget**, QPointer<QWidget>&, bool) () at /usr/lib/x86_64-linux-gnu/libQt5Widgets.so.5
#25 0x00007fda6fb12ea5 in () at /usr/lib/x86_64-linux-gnu/libQt5Widgets.so.5
#26 0x00007fda6fb157da in () at /usr/lib/x86_64-linux-gnu/libQt5Widgets.so.5
#27 0x00007fda6fab983c in QApplicationPrivate::notify_helper(QObject*, QEvent*) () at /usr/lib/x86_64-linux-gnu/libQt5Widgets.so.5
#28 0x00007fda6fac1104 in QApplication::notify(QObject*, QEvent*) () at /usr/lib/x86_64-linux-gnu/libQt5Widgets.so.5
#29 0x00007fda709b434b in QgsApplication::notify(QObject*, QEvent*) () at /usr/lib/libqgis_core.so.3.12.1
#30 0x00007fda6ed3b8d8 in QCoreApplication::notifyInternal2(QObject*, QEvent*) () at /usr/lib/x86_64-linux-gnu/libQt5Core.so.5
#31 0x00007fda6f2fd583 in QGuiApplicationPrivate::processMouseEvent(QWindowSystemInterfacePrivate::MouseEvent*) () at /usr/lib/x86_64-linux-gnu/libQt5Gui.so.5
#32 0x00007fda6f2ff055 in QGuiApplicationPrivate::processWindowSystemEvent(QWindowSystemInterfacePrivate::WindowSystemEvent*) () at /usr/lib/x86_64-linux-gnu/libQt5Gui.so.5
#33 0x00007fda6f2d62eb in QWindowSystemInterface::sendWindowSystemEvents(QFlags<QEventLoop::ProcessEventsFlag>) () at /usr/lib/x86_64-linux-gnu/libQt5Gui.so.5
#34 0x00007fda4a187260 in () at /usr/lib/x86_64-linux-gnu/libQt5XcbQpa.so.5
#35 0x00007fda62d00417 in g_main_context_dispatch () at /usr/lib/x86_64-linux-gnu/libglib-2.0.so.0
#36 0x00007fda62d00650 in () at /usr/lib/x86_64-linux-gnu/libglib-2.0.so.0
#37 0x00007fda62d006dc in g_main_context_iteration () at /usr/lib/x86_64-linux-gnu/libglib-2.0.so.0
#38 0x00007fda6ed9488f in QEventDispatcherGlib::processEvents(QFlags<QEventLoop::ProcessEventsFlag>) () at /usr/lib/x86_64-linux-gnu/libQt5Core.so.5
#39 0x00007fda6ed3990a in QEventLoop::exec(QFlags<QEventLoop::ProcessEventsFlag>) () at /usr/lib/x86_64-linux-gnu/libQt5Core.so.5
#40 0x00007fda6ed429b4 in QCoreApplication::exec() () at /usr/lib/x86_64-linux-gnu/libQt5Core.so.5
#41 0x000055ec4650f56e in ()
#42 0x00007fda6e140b97 in __libc_start_main (main=0x55ec4650cd10, argc=1, argv=0x7ffeb33c15a8, init=<optimised out>, fini=<optimised out>, rtld_fini=<optimised out>, stack_end=0x7ffeb33c1598) at ../csu/libc-start.c:310
result = <optimised out>
unwind_buf = {cancel_jmp_buf = {{jmp_buf = {0, -3021662798714391275, 94473280366544, 140731905480096, 0, 0, -9064321723719109355, -9044154686778395371}, mask_was_saved = 0}}, priv = {pad = {0x0, 0x0, 0x7fda73167733 <_dl_init+259>, 0x7fda731231f8}, data = {prev = 0x0, cleanup = 0x0, canceltype = 1930852147}}}
not_first_call = <optimised out>
#43 0x000055ec465127fa in _start ()
gdb returned 0
Aborted
``` | 1.0 | Using some functions in raster calulator crashes QGIS 3.10/3.12. - Hey,
I am on Linux Mint 19.2 Cinnamon using QGIS 3.10 LTR apt package. Every time I use the native raster calculator with an expression, such as the following:
`("slope@1" > (6.94*0.85+((max("slope@1")-min("slope@1")))/4))*1`
it crashes the whole QGIS without any message.
The expression is valid as well as the raster layer I'm using. I suspect there is a problem with min/max functions, however, not sure about it, and not sure which other functions might be involved.
The expression: `("slope@1" > (6.94*0.85))*1` returned a valid result.
I reproduced the same issue on the 3.12 version.
I launched QGIS 3.12 through the console and captured this log on the crash:
```
Warning: QStandardPaths: XDG_RUNTIME_DIR not set, defaulting to '/tmp/runtime-root'
QGIS died on signal 11[New LWP 5663]
[New LWP 5664]
[New LWP 5665]
[New LWP 5666]
[New LWP 5673]
[New LWP 5674]
[New LWP 5675]
[New LWP 5676]
[New LWP 5679]
[New LWP 5690]
[New LWP 5691]
[New LWP 5692]
[Thread debugging using libthread_db enabled]
Using host libthread_db library "/lib/x86_64-linux-gnu/libthread_db.so.1".
0x00007fda6e2036c2 in __GI___waitpid (pid=5695, stat_loc=0x7ffeb33be6ec, options=0) at ../sysdeps/unix/sysv/linux/waitpid.c:30
30 ../sysdeps/unix/sysv/linux/waitpid.c: No such file or directory.
[Current thread is 1 (Thread 0x7fda732fe3c0 (LWP 5660))]
#0 0x00007fda6e2036c2 in __GI___waitpid (pid=5695, stat_loc=0x7ffeb33be6ec, options=0) at ../sysdeps/unix/sysv/linux/waitpid.c:30
resultvar = 18446744073709551104
sc_cancel_oldtype = 0
#1 0x000055ec465132be in ()
#2 0x00007fda6e15df20 in <signal handler called> () at /lib/x86_64-linux-gnu/libc.so.6
#3 0x00007fda6c855981 in QgsRasterMatrix::twoArgumentOperation(QgsRasterMatrix::TwoArgOperator, QgsRasterMatrix const&) () at /usr/lib/libqgis_analysis.so.3.12.1
#4 0x00007fda6c846aff in QgsRasterCalcNode::calculate(QMap<QString, QgsRasterBlock*>&, QgsRasterMatrix&, int) const () at /usr/lib/libqgis_analysis.so.3.12.1
#5 0x00007fda6c846a7a in QgsRasterCalcNode::calculate(QMap<QString, QgsRasterBlock*>&, QgsRasterMatrix&, int) const () at /usr/lib/libqgis_analysis.so.3.12.1
#6 0x00007fda6c846a7a in QgsRasterCalcNode::calculate(QMap<QString, QgsRasterBlock*>&, QgsRasterMatrix&, int) const () at /usr/lib/libqgis_analysis.so.3.12.1
#7 0x00007fda6c846ab1 in QgsRasterCalcNode::calculate(QMap<QString, QgsRasterBlock*>&, QgsRasterMatrix&, int) const () at /usr/lib/libqgis_analysis.so.3.12.1
#8 0x00007fda6c846ab1 in QgsRasterCalcNode::calculate(QMap<QString, QgsRasterBlock*>&, QgsRasterMatrix&, int) const () at /usr/lib/libqgis_analysis.so.3.12.1
#9 0x00007fda6c846a7a in QgsRasterCalcNode::calculate(QMap<QString, QgsRasterBlock*>&, QgsRasterMatrix&, int) const () at /usr/lib/libqgis_analysis.so.3.12.1
#10 0x00007fda6c850980 in QgsRasterCalculator::processCalculation(QgsFeedback*) () at /usr/lib/libqgis_analysis.so.3.12.1
#11 0x00007fda72773b4c in QgisApp::showRasterCalculator() () at /usr/lib/libqgis_app.so.3.12.1
#12 0x00007fda6ed6a66f in QMetaObject::activate(QObject*, int, int, void**) () at /usr/lib/x86_64-linux-gnu/libQt5Core.so.5
#13 0x00007fda6fab3122 in QAction::triggered(bool) () at /usr/lib/x86_64-linux-gnu/libQt5Widgets.so.5
#14 0x00007fda6fab580c in QAction::activate(QAction::ActionEvent) () at /usr/lib/x86_64-linux-gnu/libQt5Widgets.so.5
#15 0x00007fda6fc320fc in () at /usr/lib/x86_64-linux-gnu/libQt5Widgets.so.5
#16 0x00007fda6fc395ab in () at /usr/lib/x86_64-linux-gnu/libQt5Widgets.so.5
#17 0x00007fda6fc3a3d3 in QMenu::mouseReleaseEvent(QMouseEvent*) () at /usr/lib/x86_64-linux-gnu/libQt5Widgets.so.5
#18 0x00007fda6faf8048 in QWidget::event(QEvent*) () at /usr/lib/x86_64-linux-gnu/libQt5Widgets.so.5
#19 0x00007fda6fc3c6fb in QMenu::event(QEvent*) () at /usr/lib/x86_64-linux-gnu/libQt5Widgets.so.5
#20 0x00007fda6fab983c in QApplicationPrivate::notify_helper(QObject*, QEvent*) () at /usr/lib/x86_64-linux-gnu/libQt5Widgets.so.5
#21 0x00007fda6fac165f in QApplication::notify(QObject*, QEvent*) () at /usr/lib/x86_64-linux-gnu/libQt5Widgets.so.5
#22 0x00007fda709b434b in QgsApplication::notify(QObject*, QEvent*) () at /usr/lib/libqgis_core.so.3.12.1
#23 0x00007fda6ed3b8d8 in QCoreApplication::notifyInternal2(QObject*, QEvent*) () at /usr/lib/x86_64-linux-gnu/libQt5Core.so.5
#24 0x00007fda6fac0632 in QApplicationPrivate::sendMouseEvent(QWidget*, QMouseEvent*, QWidget*, QWidget*, QWidget**, QPointer<QWidget>&, bool) () at /usr/lib/x86_64-linux-gnu/libQt5Widgets.so.5
#25 0x00007fda6fb12ea5 in () at /usr/lib/x86_64-linux-gnu/libQt5Widgets.so.5
#26 0x00007fda6fb157da in () at /usr/lib/x86_64-linux-gnu/libQt5Widgets.so.5
#27 0x00007fda6fab983c in QApplicationPrivate::notify_helper(QObject*, QEvent*) () at /usr/lib/x86_64-linux-gnu/libQt5Widgets.so.5
#28 0x00007fda6fac1104 in QApplication::notify(QObject*, QEvent*) () at /usr/lib/x86_64-linux-gnu/libQt5Widgets.so.5
#29 0x00007fda709b434b in QgsApplication::notify(QObject*, QEvent*) () at /usr/lib/libqgis_core.so.3.12.1
#30 0x00007fda6ed3b8d8 in QCoreApplication::notifyInternal2(QObject*, QEvent*) () at /usr/lib/x86_64-linux-gnu/libQt5Core.so.5
#31 0x00007fda6f2fd583 in QGuiApplicationPrivate::processMouseEvent(QWindowSystemInterfacePrivate::MouseEvent*) () at /usr/lib/x86_64-linux-gnu/libQt5Gui.so.5
#32 0x00007fda6f2ff055 in QGuiApplicationPrivate::processWindowSystemEvent(QWindowSystemInterfacePrivate::WindowSystemEvent*) () at /usr/lib/x86_64-linux-gnu/libQt5Gui.so.5
#33 0x00007fda6f2d62eb in QWindowSystemInterface::sendWindowSystemEvents(QFlags<QEventLoop::ProcessEventsFlag>) () at /usr/lib/x86_64-linux-gnu/libQt5Gui.so.5
#34 0x00007fda4a187260 in () at /usr/lib/x86_64-linux-gnu/libQt5XcbQpa.so.5
#35 0x00007fda62d00417 in g_main_context_dispatch () at /usr/lib/x86_64-linux-gnu/libglib-2.0.so.0
#36 0x00007fda62d00650 in () at /usr/lib/x86_64-linux-gnu/libglib-2.0.so.0
#37 0x00007fda62d006dc in g_main_context_iteration () at /usr/lib/x86_64-linux-gnu/libglib-2.0.so.0
#38 0x00007fda6ed9488f in QEventDispatcherGlib::processEvents(QFlags<QEventLoop::ProcessEventsFlag>) () at /usr/lib/x86_64-linux-gnu/libQt5Core.so.5
#39 0x00007fda6ed3990a in QEventLoop::exec(QFlags<QEventLoop::ProcessEventsFlag>) () at /usr/lib/x86_64-linux-gnu/libQt5Core.so.5
#40 0x00007fda6ed429b4 in QCoreApplication::exec() () at /usr/lib/x86_64-linux-gnu/libQt5Core.so.5
#41 0x000055ec4650f56e in ()
#42 0x00007fda6e140b97 in __libc_start_main (main=0x55ec4650cd10, argc=1, argv=0x7ffeb33c15a8, init=<optimised out>, fini=<optimised out>, rtld_fini=<optimised out>, stack_end=0x7ffeb33c1598) at ../csu/libc-start.c:310
result = <optimised out>
unwind_buf = {cancel_jmp_buf = {{jmp_buf = {0, -3021662798714391275, 94473280366544, 140731905480096, 0, 0, -9064321723719109355, -9044154686778395371}, mask_was_saved = 0}}, priv = {pad = {0x0, 0x0, 0x7fda73167733 <_dl_init+259>, 0x7fda731231f8}, data = {prev = 0x0, cleanup = 0x0, canceltype = 1930852147}}}
not_first_call = <optimised out>
#43 0x000055ec465127fa in _start ()
gdb returned 0
Aborted
``` | process | using some functions in raster calulator crashes qgis hey i am on linux mint cinnamon using qgis ltr apt package every time i use the native raster calculator with an expression such as the following slope max slope min slope it crashes the whole qgis without any message the expression is valid as well as the raster layer i m using i suspect there is a problem with min max functions however not sure about it and not sure which other functions might be involved the expression slope returned a valid result i reproduced the same issue on the version i launched qgis through the console and captured this log on the crash warning qstandardpaths xdg runtime dir not set defaulting to tmp runtime root qgis died on signal using host libthread db library lib linux gnu libthread db so in gi waitpid pid stat loc options at sysdeps unix sysv linux waitpid c sysdeps unix sysv linux waitpid c no such file or directory in gi waitpid pid stat loc options at sysdeps unix sysv linux waitpid c resultvar sc cancel oldtype in in at lib linux gnu libc so in qgsrastermatrix twoargumentoperation qgsrastermatrix twoargoperator qgsrastermatrix const at usr lib libqgis analysis so in qgsrastercalcnode calculate qmap qgsrastermatrix int const at usr lib libqgis analysis so in qgsrastercalcnode calculate qmap qgsrastermatrix int const at usr lib libqgis analysis so in qgsrastercalcnode calculate qmap qgsrastermatrix int const at usr lib libqgis analysis so in qgsrastercalcnode calculate qmap qgsrastermatrix int const at usr lib libqgis analysis so in qgsrastercalcnode calculate qmap qgsrastermatrix int const at usr lib libqgis analysis so in qgsrastercalcnode calculate qmap qgsrastermatrix int const at usr lib libqgis analysis so in qgsrastercalculator processcalculation qgsfeedback at usr lib libqgis analysis so in qgisapp showrastercalculator at usr lib libqgis app so in qmetaobject activate qobject int int void at usr lib linux gnu so in qaction triggered bool at usr lib linux gnu so in qaction activate qaction actionevent at usr lib linux gnu so in at usr lib linux gnu so in at usr lib linux gnu so in qmenu mousereleaseevent qmouseevent at usr lib linux gnu so in qwidget event qevent at usr lib linux gnu so in qmenu event qevent at usr lib linux gnu so in qapplicationprivate notify helper qobject qevent at usr lib linux gnu so in qapplication notify qobject qevent at usr lib linux gnu so in qgsapplication notify qobject qevent at usr lib libqgis core so in qcoreapplication qobject qevent at usr lib linux gnu so in qapplicationprivate sendmouseevent qwidget qmouseevent qwidget qwidget qwidget qpointer bool at usr lib linux gnu so in at usr lib linux gnu so in at usr lib linux gnu so in qapplicationprivate notify helper qobject qevent at usr lib linux gnu so in qapplication notify qobject qevent at usr lib linux gnu so in qgsapplication notify qobject qevent at usr lib libqgis core so in qcoreapplication qobject qevent at usr lib linux gnu so in qguiapplicationprivate processmouseevent qwindowsysteminterfaceprivate mouseevent at usr lib linux gnu so in qguiapplicationprivate processwindowsystemevent qwindowsysteminterfaceprivate windowsystemevent at usr lib linux gnu so in qwindowsysteminterface sendwindowsystemevents qflags at usr lib linux gnu so in at usr lib linux gnu so in g main context dispatch at usr lib linux gnu libglib so in at usr lib linux gnu libglib so in g main context iteration at usr lib linux gnu libglib so in qeventdispatcherglib processevents qflags at usr lib linux gnu so in qeventloop exec qflags at usr lib linux gnu so in qcoreapplication exec at usr lib linux gnu so in in libc start main main argc argv init fini rtld fini stack end at csu libc start c result unwind buf cancel jmp buf jmp buf mask was saved priv pad data prev cleanup canceltype not first call in start gdb returned aborted | 1 |
22,194 | 30,750,569,512 | IssuesEvent | 2023-07-28 18:49:31 | MicrosoftDocs/azure-docs | https://api.github.com/repos/MicrosoftDocs/azure-docs | closed | Replacement for RunAs in Azure Automation for on-prem scripts | automation/svc triaged assigned-to-author doc-enhancement process-automation/subsvc Pri1 | Currently it is possible to execute PowerShell scripts against on-premise Active Directory infrastructure from Azure using Azure Automation with a hybrid worker and RunAs using Active Directory credentials for authentication (see https://practical365.com/how-to-manage-on-premises-infrastructure-using-azure-automation-hybrid-worker/).
Since Managed Identities cannot be used to authenticate to Active Directory, what is the recommended migration path for this use case as RunAs is being deprecated?
---
#### Document Details
⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.*
* ID: 3eedb810-487f-bc9c-89f5-d5fdbcc5d796
* Version Independent ID: 329e9ec7-d9ea-518b-625b-d39880365e80
* Content: [Migrate from a Run As account to a managed identity](https://learn.microsoft.com/en-us/azure/automation/migrate-run-as-accounts-managed-identity?tabs=run-as-account)
* Content Source: [articles/automation/migrate-run-as-accounts-managed-identity.md](https://github.com/MicrosoftDocs/azure-docs/blob/main/articles/automation/migrate-run-as-accounts-managed-identity.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @SnehaSudhirG
* Microsoft Alias: **sudhirsneha** | 1.0 | Replacement for RunAs in Azure Automation for on-prem scripts - Currently it is possible to execute PowerShell scripts against on-premise Active Directory infrastructure from Azure using Azure Automation with a hybrid worker and RunAs using Active Directory credentials for authentication (see https://practical365.com/how-to-manage-on-premises-infrastructure-using-azure-automation-hybrid-worker/).
Since Managed Identities cannot be used to authenticate to Active Directory, what is the recommended migration path for this use case as RunAs is being deprecated?
---
#### Document Details
⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.*
* ID: 3eedb810-487f-bc9c-89f5-d5fdbcc5d796
* Version Independent ID: 329e9ec7-d9ea-518b-625b-d39880365e80
* Content: [Migrate from a Run As account to a managed identity](https://learn.microsoft.com/en-us/azure/automation/migrate-run-as-accounts-managed-identity?tabs=run-as-account)
* Content Source: [articles/automation/migrate-run-as-accounts-managed-identity.md](https://github.com/MicrosoftDocs/azure-docs/blob/main/articles/automation/migrate-run-as-accounts-managed-identity.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @SnehaSudhirG
* Microsoft Alias: **sudhirsneha** | process | replacement for runas in azure automation for on prem scripts currently it is possible to execute powershell scripts against on premise active directory infrastructure from azure using azure automation with a hybrid worker and runas using active directory credentials for authentication see since managed identities cannot be used to authenticate to active directory what is the recommended migration path for this use case as runas is being deprecated document details ⚠ do not edit this section it is required for learn microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login snehasudhirg microsoft alias sudhirsneha | 1 |
390,667 | 26,868,425,700 | IssuesEvent | 2023-02-04 06:03:17 | liliana-krp/testing | https://api.github.com/repos/liliana-krp/testing | closed | ad some more details to this repo | documentation enhancement good first issue | I need to come back here later to add some more information to this repo.
Options include:
* [x] some folders
* [x] some R code
* [x] some data | 1.0 | ad some more details to this repo - I need to come back here later to add some more information to this repo.
Options include:
* [x] some folders
* [x] some R code
* [x] some data | non_process | ad some more details to this repo i need to come back here later to add some more information to this repo options include some folders some r code some data | 0 |
13,802 | 16,554,514,967 | IssuesEvent | 2021-05-28 12:30:25 | encode/uvicorn | https://api.github.com/repos/encode/uvicorn | closed | uvicorn.run with debug=True results in multiprocessing RuntimeError | multiprocessing need confirmation | When trying to call `uvicorn.run` with `debug=True`, I get a runtime error.
```
File "/usr/lib/python3.8/multiprocessing/spawn.py", line 134, in _check_not_importing_main
raise RuntimeError('''
RuntimeError:
An attempt has been made to start a new process before the
current process has finished its bootstrapping phase.
This probably means that you are not using fork to start your
child processes and you have forgotten to use the proper idiom
in the main module:
if __name__ == '__main__':
freeze_support()
...
The "freeze_support()" line can be omitted if the program
is not going to be frozen to produce an executable.
INFO: Stopping reloader process [12137]
Process finished with exit code 0
```
Code used to run:
```python
uvicorn.run(
"backend.app:app",
host="127.0.0.1",
port=8000,
debug=True,
)
```
When `debug=True` is not passed, the app runs normally.
| 1.0 | uvicorn.run with debug=True results in multiprocessing RuntimeError - When trying to call `uvicorn.run` with `debug=True`, I get a runtime error.
```
File "/usr/lib/python3.8/multiprocessing/spawn.py", line 134, in _check_not_importing_main
raise RuntimeError('''
RuntimeError:
An attempt has been made to start a new process before the
current process has finished its bootstrapping phase.
This probably means that you are not using fork to start your
child processes and you have forgotten to use the proper idiom
in the main module:
if __name__ == '__main__':
freeze_support()
...
The "freeze_support()" line can be omitted if the program
is not going to be frozen to produce an executable.
INFO: Stopping reloader process [12137]
Process finished with exit code 0
```
Code used to run:
```python
uvicorn.run(
"backend.app:app",
host="127.0.0.1",
port=8000,
debug=True,
)
```
When `debug=True` is not passed, the app runs normally.
| process | uvicorn run with debug true results in multiprocessing runtimeerror when trying to call uvicorn run with debug true i get a runtime error file usr lib multiprocessing spawn py line in check not importing main raise runtimeerror runtimeerror an attempt has been made to start a new process before the current process has finished its bootstrapping phase this probably means that you are not using fork to start your child processes and you have forgotten to use the proper idiom in the main module if name main freeze support the freeze support line can be omitted if the program is not going to be frozen to produce an executable info stopping reloader process process finished with exit code code used to run python uvicorn run backend app app host port debug true when debug true is not passed the app runs normally | 1 |
138,800 | 11,217,656,444 | IssuesEvent | 2020-01-07 09:44:40 | FreeRDP/FreeRDP | https://api.github.com/repos/FreeRDP/FreeRDP | closed | xfreerdp usb devices can not redirect auto | channel:usb fixed-waiting-test | when I via pid and vid to redirect a usb device , successful . but when use /usb:auto I can not redirect usb devices.
who can help me ?thank you very much
| 1.0 | xfreerdp usb devices can not redirect auto - when I via pid and vid to redirect a usb device , successful . but when use /usb:auto I can not redirect usb devices.
who can help me ?thank you very much
| non_process | xfreerdp usb devices can not redirect auto when i via pid and vid to redirect a usb device successful but when use usb auto i can not redirect usb devices who can help me thank you very much | 0 |
14,696 | 17,859,570,135 | IssuesEvent | 2021-09-05 18:01:09 | bazelbuild/bazel | https://api.github.com/repos/bazelbuild/bazel | closed | bazel can not generate enumeration related function xxx_Name. | type: support / not a bug (process) |
### Description of the problem / feature request:
Hi, all
* I want to use bazel to build protobuf for Enumerations to generate related functions.
follow the documents, the function *xxx_Name* should be generated, but I can not find the function in my generated cpp header file.
* But I can got the related *xxx_MIN*, *xxx_MAX*
* This is across-compile in x86 machine for aarch64.

* About my environment:
1. Ubuntu 18.04 x86_64 across-compile for aarch64
2. bazel 0.13.0
3. protobuf 3.10.1
* I directly use the protobuf software, I can got the function *xxx_Name*. When the protobuf version is used in my bazel project , the some errors are got.
eg:
1. rules_python/python/private/reexports.bzl:40:19: name 'PyInfo' is not defined (did you mean 'ApkInfo'?)
2. Extension file not found. Unable to load file '@bazel_tools//tools/python:toolchain.bzl': file doesn't exist or isn't a file.
* I want to know if the reason are right. would you give me some guide?
### Feature requests: what underlying problem are you trying to solve with this feature?
* about import protobuf package information in *WORKSPACE*
```
http_archive(
name = "com_google_protobuf",
sha256 = "d7a221b3d4fb4f05b7473795ccea9e05dab3b8721f6286a95fffbffc2d926f8b",
strip_prefix = "protobuf-3.6.1",
urls = [
"https://cnbj1.fds.api.xiaomi.com/mace/third-party/protobuf/protobuf-3.6.1.zip",
"https://github.com/google/protobuf/archive/v3.6.1.zip",
],
)
```
* about proto cpp configurate in the realted directory *BUILD* file
```
cc_proto_library(
name = "bench_proto_cc",
srcs = glob(["base.proto",
"aibench.proto"]),
# default_runtime = "@com_google_protobuf//:protobuf_lite",
default_runtime = "@com_google_protobuf//:protobuf",
protoc = "@com_google_protobuf//:protoc",
)
```
### Bugs: what's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible.
> Replace this line with your answer.
### What operating system are you running Bazel on?
> Replace this line with your answer.
### What's the output of `bazel info release`?
> Replace this line with your answer.
### If `bazel info release` returns "development version" or "(@non-git)", tell us how you built Bazel.
> Replace this line with your answer.
### What's the output of `git remote get-url origin ; git rev-parse master ; git rev-parse HEAD` ?
> Replace this line with your answer.
### Have you found anything relevant by searching the web?
> Replace these lines with your answer.
>
> Places to look:
> - StackOverflow: http://stackoverflow.com/questions/tagged/bazel
> - GitHub issues: https://github.com/bazelbuild/bazel/issues
> - email threads on https://groups.google.com/forum/#!forum/bazel-discuss
### Any other information, logs, or outputs that you want to share?
> Replace these lines with your answer.
>
> If the files are large, upload as attachment or provide link.
| 1.0 | bazel can not generate enumeration related function xxx_Name. -
### Description of the problem / feature request:
Hi, all
* I want to use bazel to build protobuf for Enumerations to generate related functions.
follow the documents, the function *xxx_Name* should be generated, but I can not find the function in my generated cpp header file.
* But I can got the related *xxx_MIN*, *xxx_MAX*
* This is across-compile in x86 machine for aarch64.

* About my environment:
1. Ubuntu 18.04 x86_64 across-compile for aarch64
2. bazel 0.13.0
3. protobuf 3.10.1
* I directly use the protobuf software, I can got the function *xxx_Name*. When the protobuf version is used in my bazel project , the some errors are got.
eg:
1. rules_python/python/private/reexports.bzl:40:19: name 'PyInfo' is not defined (did you mean 'ApkInfo'?)
2. Extension file not found. Unable to load file '@bazel_tools//tools/python:toolchain.bzl': file doesn't exist or isn't a file.
* I want to know if the reason are right. would you give me some guide?
### Feature requests: what underlying problem are you trying to solve with this feature?
* about import protobuf package information in *WORKSPACE*
```
http_archive(
name = "com_google_protobuf",
sha256 = "d7a221b3d4fb4f05b7473795ccea9e05dab3b8721f6286a95fffbffc2d926f8b",
strip_prefix = "protobuf-3.6.1",
urls = [
"https://cnbj1.fds.api.xiaomi.com/mace/third-party/protobuf/protobuf-3.6.1.zip",
"https://github.com/google/protobuf/archive/v3.6.1.zip",
],
)
```
* about proto cpp configurate in the realted directory *BUILD* file
```
cc_proto_library(
name = "bench_proto_cc",
srcs = glob(["base.proto",
"aibench.proto"]),
# default_runtime = "@com_google_protobuf//:protobuf_lite",
default_runtime = "@com_google_protobuf//:protobuf",
protoc = "@com_google_protobuf//:protoc",
)
```
### Bugs: what's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible.
> Replace this line with your answer.
### What operating system are you running Bazel on?
> Replace this line with your answer.
### What's the output of `bazel info release`?
> Replace this line with your answer.
### If `bazel info release` returns "development version" or "(@non-git)", tell us how you built Bazel.
> Replace this line with your answer.
### What's the output of `git remote get-url origin ; git rev-parse master ; git rev-parse HEAD` ?
> Replace this line with your answer.
### Have you found anything relevant by searching the web?
> Replace these lines with your answer.
>
> Places to look:
> - StackOverflow: http://stackoverflow.com/questions/tagged/bazel
> - GitHub issues: https://github.com/bazelbuild/bazel/issues
> - email threads on https://groups.google.com/forum/#!forum/bazel-discuss
### Any other information, logs, or outputs that you want to share?
> Replace these lines with your answer.
>
> If the files are large, upload as attachment or provide link.
| process | bazel can not generate enumeration related function xxx name description of the problem feature request hi all i want to use bazel to build protobuf for enumerations to generate related functions follow the documents the function xxx name should be generated but i can not find the function in my generated cpp header file but i can got the related xxx min xxx max this is across compile in machine for about my environment ubuntu across compile for bazel protobuf i directly use the protobuf software i can got the function xxx name when the protobuf version is used in my bazel project the some errors are got eg rules python python private reexports bzl name pyinfo is not defined did you mean apkinfo extension file not found unable to load file bazel tools tools python toolchain bzl file doesn t exist or isn t a file i want to know if the reason are right would you give me some guide feature requests what underlying problem are you trying to solve with this feature about import protobuf package information in workspace http archive name com google protobuf strip prefix protobuf urls about proto cpp configurate in the realted directory build file cc proto library name bench proto cc srcs glob base proto aibench proto default runtime com google protobuf protobuf lite default runtime com google protobuf protobuf protoc com google protobuf protoc bugs what s the simplest easiest way to reproduce this bug please provide a minimal example if possible replace this line with your answer what operating system are you running bazel on replace this line with your answer what s the output of bazel info release replace this line with your answer if bazel info release returns development version or non git tell us how you built bazel replace this line with your answer what s the output of git remote get url origin git rev parse master git rev parse head replace this line with your answer have you found anything relevant by searching the web replace these lines with your answer places to look stackoverflow github issues email threads on any other information logs or outputs that you want to share replace these lines with your answer if the files are large upload as attachment or provide link | 1 |
3,273 | 6,348,865,804 | IssuesEvent | 2017-07-28 11:13:09 | openvstorage/integrationtests | https://api.github.com/repos/openvstorage/integrationtests | closed | Failed test: add_remove_alba_vpool_test | process_duplicate | This night's builds started throwing these for this test on both of the unstable envs:
EE-PHY-3N-187.31-unstable - http://testrail.openvstorage.com/index.php?/plans/view/38378
EE-VIR-3N-199.181-unstable - http://testrail.openvstorage.com/index.php?/plans/view/38374
`Removal of preset 'ciaddremovevpool' for backend 'mybackend' failed with 'Albamgr exception(Albamgr_protocol.Protocol.Error.Unknown,Protocol_common.XException(4, "/alba/preset/propagation/ciaddremovevpool"))'`
Most likely related to https://github.com/openvstorage/integrationtests/issues/556 | 1.0 | Failed test: add_remove_alba_vpool_test - This night's builds started throwing these for this test on both of the unstable envs:
EE-PHY-3N-187.31-unstable - http://testrail.openvstorage.com/index.php?/plans/view/38378
EE-VIR-3N-199.181-unstable - http://testrail.openvstorage.com/index.php?/plans/view/38374
`Removal of preset 'ciaddremovevpool' for backend 'mybackend' failed with 'Albamgr exception(Albamgr_protocol.Protocol.Error.Unknown,Protocol_common.XException(4, "/alba/preset/propagation/ciaddremovevpool"))'`
Most likely related to https://github.com/openvstorage/integrationtests/issues/556 | process | failed test add remove alba vpool test this night s builds started throwing these for this test on both of the unstable envs ee phy unstable ee vir unstable removal of preset ciaddremovevpool for backend mybackend failed with albamgr exception albamgr protocol protocol error unknown protocol common xexception alba preset propagation ciaddremovevpool most likely related to | 1 |
6,593 | 9,668,169,216 | IssuesEvent | 2019-05-21 14:39:28 | CGAL/cgal | https://api.github.com/repos/CGAL/cgal | closed | Point Set Processing: cleanup | Pkg::Point_set_processing_3 cleaning |
## Issue Details
For 5.0 we should remove functionality that was deprecated long ago, such as [this](https://doc.cgal.org/latest/Point_set_processing_3/index.html#title4).
| 1.0 | Point Set Processing: cleanup -
## Issue Details
For 5.0 we should remove functionality that was deprecated long ago, such as [this](https://doc.cgal.org/latest/Point_set_processing_3/index.html#title4).
| process | point set processing cleanup issue details for we should remove functionality that was deprecated long ago such as | 1 |
286,816 | 21,610,402,241 | IssuesEvent | 2022-05-04 09:30:25 | hitontology/ontology | https://api.github.com/repos/hitontology/ontology | closed | rename outcome criteria in ontology | documentation needsDiscussion refactor | Outcome criteria muss zu outcome criterion werden - sonst haben wir auch nirgendwo die Mehrzahl. | 1.0 | rename outcome criteria in ontology - Outcome criteria muss zu outcome criterion werden - sonst haben wir auch nirgendwo die Mehrzahl. | non_process | rename outcome criteria in ontology outcome criteria muss zu outcome criterion werden sonst haben wir auch nirgendwo die mehrzahl | 0 |
47,171 | 13,056,046,314 | IssuesEvent | 2020-07-30 03:29:32 | icecube-trac/tix2 | https://api.github.com/repos/icecube-trac/tix2 | closed | python-not-found message wrong (Trac #92) | Migrated from Trac cmake defect | it says "probably available via the ports system"
Migrated from https://code.icecube.wisc.edu/ticket/92
```json
{
"status": "closed",
"changetime": "2007-11-09T22:26:46",
"description": "it says \"probably available via the ports system\"\n\n",
"reporter": "troy",
"cc": "",
"resolution": "fixed",
"_ts": "1194647206000000",
"component": "cmake",
"summary": "python-not-found message wrong",
"priority": "normal",
"keywords": "",
"time": "2007-08-22T12:31:29",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
| 1.0 | python-not-found message wrong (Trac #92) - it says "probably available via the ports system"
Migrated from https://code.icecube.wisc.edu/ticket/92
```json
{
"status": "closed",
"changetime": "2007-11-09T22:26:46",
"description": "it says \"probably available via the ports system\"\n\n",
"reporter": "troy",
"cc": "",
"resolution": "fixed",
"_ts": "1194647206000000",
"component": "cmake",
"summary": "python-not-found message wrong",
"priority": "normal",
"keywords": "",
"time": "2007-08-22T12:31:29",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
| non_process | python not found message wrong trac it says probably available via the ports system migrated from json status closed changetime description it says probably available via the ports system n n reporter troy cc resolution fixed ts component cmake summary python not found message wrong priority normal keywords time milestone owner troy type defect | 0 |
419,693 | 12,227,115,491 | IssuesEvent | 2020-05-03 13:59:25 | eladve/omnitool | https://api.github.com/repos/eladve/omnitool | opened | Improvement Request: improve search functionality! | high-priority | Some inspiration for the below can come from web search (e.g. google) , as well as from memex (https://getmemex.com/).
Track 1: Improve search logic:
right now I only implemented single-word search. To improve this add:
- multiword search, exact-pharse search, etc (a-la google, memex, ...). can do this by deploying an existing solution (either python library or database). Or can implement this from scratch. Shouldn't be too hard. E.g. to look for two words, just take the hit list of each of them, and intersect the lists.
- better prioritization/ranking of results. (Remember that screenshots overlap so there might be many hits that are very close in time. Need to prioritize so that only one of those is chosen for the "first page" -- maybe by using some clustering on the timestamps, something like k-means on the timestamps). Search result ranking is a common problem so there might be some open-sorce library that can be a good place to start.
- maybe add meta-data (e.g. which application is used, etc) and allow to search also on this.
Track 2: Better NLP/tokenization/etc:
- right now the text is not tokenized properly, and that hurts the quality of the inverted index. For example words with dashes are not-split, special characters are not removed properly, and in general, tokens are not recognized in the best way to enable them to be used in a future search. Should fix that. Should be easy to get a 10x improvement by using NLP tools like NLTK https://www.nltk.org/
- can do even more NLP to enrich with metadata: named entity recognition, etc. . (for each named entity that is recognized, an email that is recognized, etc, can just add it to the inverted index and that will allow future search).
Track 3: GUI and interactivity:
- once the query is executed and properly-ranked search results are returned, the user still needs to get value by scrolling through the hits, examining them, maybe refining the search, looking at the OCR results, and in general, looking for the hit that they wanted. Need to improve this via GUI and interactivity: allowing to scroll through the results, pick some, magnify them, scrolling through the timeline, etc
| 1.0 | Improvement Request: improve search functionality! - Some inspiration for the below can come from web search (e.g. google) , as well as from memex (https://getmemex.com/).
Track 1: Improve search logic:
right now I only implemented single-word search. To improve this add:
- multiword search, exact-pharse search, etc (a-la google, memex, ...). can do this by deploying an existing solution (either python library or database). Or can implement this from scratch. Shouldn't be too hard. E.g. to look for two words, just take the hit list of each of them, and intersect the lists.
- better prioritization/ranking of results. (Remember that screenshots overlap so there might be many hits that are very close in time. Need to prioritize so that only one of those is chosen for the "first page" -- maybe by using some clustering on the timestamps, something like k-means on the timestamps). Search result ranking is a common problem so there might be some open-sorce library that can be a good place to start.
- maybe add meta-data (e.g. which application is used, etc) and allow to search also on this.
Track 2: Better NLP/tokenization/etc:
- right now the text is not tokenized properly, and that hurts the quality of the inverted index. For example words with dashes are not-split, special characters are not removed properly, and in general, tokens are not recognized in the best way to enable them to be used in a future search. Should fix that. Should be easy to get a 10x improvement by using NLP tools like NLTK https://www.nltk.org/
- can do even more NLP to enrich with metadata: named entity recognition, etc. . (for each named entity that is recognized, an email that is recognized, etc, can just add it to the inverted index and that will allow future search).
Track 3: GUI and interactivity:
- once the query is executed and properly-ranked search results are returned, the user still needs to get value by scrolling through the hits, examining them, maybe refining the search, looking at the OCR results, and in general, looking for the hit that they wanted. Need to improve this via GUI and interactivity: allowing to scroll through the results, pick some, magnify them, scrolling through the timeline, etc
| non_process | improvement request improve search functionality some inspiration for the below can come from web search e g google as well as from memex track improve search logic right now i only implemented single word search to improve this add multiword search exact pharse search etc a la google memex can do this by deploying an existing solution either python library or database or can implement this from scratch shouldn t be too hard e g to look for two words just take the hit list of each of them and intersect the lists better prioritization ranking of results remember that screenshots overlap so there might be many hits that are very close in time need to prioritize so that only one of those is chosen for the first page maybe by using some clustering on the timestamps something like k means on the timestamps search result ranking is a common problem so there might be some open sorce library that can be a good place to start maybe add meta data e g which application is used etc and allow to search also on this track better nlp tokenization etc right now the text is not tokenized properly and that hurts the quality of the inverted index for example words with dashes are not split special characters are not removed properly and in general tokens are not recognized in the best way to enable them to be used in a future search should fix that should be easy to get a improvement by using nlp tools like nltk can do even more nlp to enrich with metadata named entity recognition etc for each named entity that is recognized an email that is recognized etc can just add it to the inverted index and that will allow future search track gui and interactivity once the query is executed and properly ranked search results are returned the user still needs to get value by scrolling through the hits examining them maybe refining the search looking at the ocr results and in general looking for the hit that they wanted need to improve this via gui and interactivity allowing to scroll through the results pick some magnify them scrolling through the timeline etc | 0 |
21,480 | 29,515,033,867 | IssuesEvent | 2023-06-04 11:49:50 | firebase/firebase-cpp-sdk | https://api.github.com/repos/firebase/firebase-cpp-sdk | reopened | [C++] Nightly Integration Testing Report for Firestore | type: process nightly-testing |
<hidden value="integration-test-status-comment"></hidden>
### ✅ [build against repo] Integration test succeeded!
Requested by @sunmou99 on commit 70989bc3fb8cdf0476a97f6d1a60109cd9464b7d
Last updated: Sat Jun 3 04:52 PDT 2023
**[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/5163087638)**
<hidden value="integration-test-status-comment"></hidden>
***
### ✅ [build against SDK] Integration test succeeded!
Requested by @firebase-workflow-trigger[bot] on commit 70989bc3fb8cdf0476a97f6d1a60109cd9464b7d
Last updated: Sat Jun 3 07:05 PDT 2023
**[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/5163692145)**
<hidden value="integration-test-status-comment"></hidden>
***
### ✅ [build against tip] Integration test succeeded!
Requested by @sunmou99 on commit 70989bc3fb8cdf0476a97f6d1a60109cd9464b7d
Last updated: Sun Jun 4 04:39 PDT 2023
**[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/5168701586)**
| 1.0 | [C++] Nightly Integration Testing Report for Firestore -
<hidden value="integration-test-status-comment"></hidden>
### ✅ [build against repo] Integration test succeeded!
Requested by @sunmou99 on commit 70989bc3fb8cdf0476a97f6d1a60109cd9464b7d
Last updated: Sat Jun 3 04:52 PDT 2023
**[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/5163087638)**
<hidden value="integration-test-status-comment"></hidden>
***
### ✅ [build against SDK] Integration test succeeded!
Requested by @firebase-workflow-trigger[bot] on commit 70989bc3fb8cdf0476a97f6d1a60109cd9464b7d
Last updated: Sat Jun 3 07:05 PDT 2023
**[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/5163692145)**
<hidden value="integration-test-status-comment"></hidden>
***
### ✅ [build against tip] Integration test succeeded!
Requested by @sunmou99 on commit 70989bc3fb8cdf0476a97f6d1a60109cd9464b7d
Last updated: Sun Jun 4 04:39 PDT 2023
**[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/5168701586)**
| process | nightly integration testing report for firestore ✅ nbsp integration test succeeded requested by on commit last updated sat jun pdt ✅ nbsp integration test succeeded requested by firebase workflow trigger on commit last updated sat jun pdt ✅ nbsp integration test succeeded requested by on commit last updated sun jun pdt | 1 |
14,918 | 18,354,678,353 | IssuesEvent | 2021-10-08 16:27:46 | ESMValGroup/ESMValCore | https://api.github.com/repos/ESMValGroup/ESMValCore | closed | Issue in multimodel cube merge surfaced by running `recipe_preprocessor_test.yml` | bug preprocessor | ```
2021-10-01 10:35:40,900 UTC [16801] ERROR Failed to run multi_model_statistics({<esmvalcore.preprocessor.PreprocessorFile object at 0x7f95f92d1190>, <esmvalcore.preprocessor.PreprocessorFile object at 0x7f95f92d1340>, <esmvalcore.preprocessor.PreprocessorFile object at 0x7f95f92b2d60>, <esmvalcore.preprocessor.PreprocessorFile object at 0x7f95f92d1070>}, {'span': 'overlap', 'statistics': ['mean', 'median'], 'output_products': {'mean': <esmvalcore.preprocessor.PreprocessorFile object at 0x7f95f92b2e80>, 'median': <esmvalcore.preprocessor.PreprocessorFile object at 0x7f95f92b2eb0>}})
2021-10-01 10:35:41,867 UTC [16766] INFO Maximum memory used (estimate): 6.2 GB
2021-10-01 10:35:41,868 UTC [16766] INFO Sampled every second. It may be inaccurate if short but high spikes in memory consumption occur.
2021-10-01 10:35:41,870 UTC [16766] ERROR Program terminated abnormally, see stack trace below for more information:
multiprocessing.pool.RemoteTraceback:
"""
Traceback (most recent call last):
File "/home/users/valeriu/miniconda3-June2021/envs/esmvaltool-stock/lib/python3.9/multiprocessing/pool.py", line 125, in worker
result = (True, func(*args, **kwds))
File "/home/users/valeriu/esmvalcore/esmvalcore/_task.py", line 747, in _run_task
output_files = task.run()
File "/home/users/valeriu/esmvalcore/esmvalcore/_task.py", line 252, in run
self.output_files = self._run(input_files)
File "/home/users/valeriu/esmvalcore/esmvalcore/preprocessor/__init__.py", line 481, in _run
self.products = _apply_multimodel(self.products, step,
File "/home/users/valeriu/esmvalcore/esmvalcore/preprocessor/__init__.py", line 422, in _apply_multimodel
result = preprocess(products - exclude, step, **settings)
File "/home/users/valeriu/esmvalcore/esmvalcore/preprocessor/__init__.py", line 299, in preprocess
result.append(_run_preproc_function(function, items, settings))
File "/home/users/valeriu/esmvalcore/esmvalcore/preprocessor/__init__.py", line 285, in _run_preproc_function
return function(items, **kwargs)
File "/home/users/valeriu/esmvalcore/esmvalcore/preprocessor/_multimodel.py", line 414, in multi_model_statistics
return _multiproduct_statistics(
File "/home/users/valeriu/esmvalcore/esmvalcore/preprocessor/_multimodel.py", line 325, in _multiproduct_statistics
statistics_cubes = _multicube_statistics(cubes=cubes,
File "/home/users/valeriu/esmvalcore/esmvalcore/preprocessor/_multimodel.py", line 305, in _multicube_statistics
result_cube = _compute_eager(aligned_cubes,
File "/home/users/valeriu/esmvalcore/esmvalcore/preprocessor/_multimodel.py", line 250, in _compute_eager
combined_slice = _combine(single_model_slices)
File "/home/users/valeriu/esmvalcore/esmvalcore/preprocessor/_multimodel.py", line 220, in _combine
merged_cube = cubes.merge_cube()
File "/home/users/valeriu/miniconda3-June2021/envs/esmvaltool-stock/lib/python3.9/site-packages/iris/cube.py", line 402, in merge_cube
proto_cube.register(cube, error_on_mismatch=True)
File "/home/users/valeriu/miniconda3-June2021/envs/esmvaltool-stock/lib/python3.9/site-packages/iris/_merge.py", line 1361, in register
match = coord_payload.match_signature(
File "/home/users/valeriu/miniconda3-June2021/envs/esmvaltool-stock/lib/python3.9/site-packages/iris/_merge.py", line 284, in match_signature
raise iris.exceptions.MergeError(msgs)
iris.exceptions.MergeError: failed to merge into a single cube.
Coordinates in cube.dim_coords differ: air_pressure.
"""
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/home/users/valeriu/esmvalcore/esmvalcore/_main.py", line 438, in run
fire.Fire(ESMValTool())
File "/home/users/valeriu/miniconda3-June2021/envs/esmvaltool-stock/lib/python3.9/site-packages/fire/core.py", line 141, in Fire
component_trace = _Fire(component, args, parsed_flag_args, context, name)
File "/home/users/valeriu/miniconda3-June2021/envs/esmvaltool-stock/lib/python3.9/site-packages/fire/core.py", line 466, in _Fire
component, remaining_args = _CallAndUpdateTrace(
File "/home/users/valeriu/miniconda3-June2021/envs/esmvaltool-stock/lib/python3.9/site-packages/fire/core.py", line 681, in _CallAndUpdateTrace
component = fn(*varargs, **kwargs)
File "/home/users/valeriu/esmvalcore/esmvalcore/_main.py", line 415, in run
process_recipe(recipe_file=recipe, config_user=cfg)
File "/home/users/valeriu/esmvalcore/esmvalcore/_main.py", line 105, in process_recipe
recipe.run()
File "/home/users/valeriu/esmvalcore/esmvalcore/_recipe.py", line 1515, in run
self.tasks.run(max_parallel_tasks=self._cfg['max_parallel_tasks'])
File "/home/users/valeriu/esmvalcore/esmvalcore/_task.py", line 669, in run
self._run_parallel(max_parallel_tasks)
File "/home/users/valeriu/esmvalcore/esmvalcore/_task.py", line 712, in _run_parallel
_copy_results(task, running[task])
File "/home/users/valeriu/esmvalcore/esmvalcore/_task.py", line 735, in _copy_results
task.output_files, updated_products = future.get()
File "/home/users/valeriu/miniconda3-June2021/envs/esmvaltool-stock/lib/python3.9/multiprocessing/pool.py", line 771, in get
raise self._value
File "/home/users/valeriu/miniconda3-June2021/envs/esmvaltool-stock/lib/python3.9/multiprocessing/pool.py", line 125, in worker
result = (True, func(*args, **kwds))
File "/home/users/valeriu/esmvalcore/esmvalcore/_task.py", line 747, in _run_task
output_files = task.run()
File "/home/users/valeriu/esmvalcore/esmvalcore/_task.py", line 252, in run
self.output_files = self._run(input_files)
File "/home/users/valeriu/esmvalcore/esmvalcore/preprocessor/__init__.py", line 481, in _run
self.products = _apply_multimodel(self.products, step,
File "/home/users/valeriu/esmvalcore/esmvalcore/preprocessor/__init__.py", line 422, in _apply_multimodel
result = preprocess(products - exclude, step, **settings)
File "/home/users/valeriu/esmvalcore/esmvalcore/preprocessor/__init__.py", line 299, in preprocess
result.append(_run_preproc_function(function, items, settings))
File "/home/users/valeriu/esmvalcore/esmvalcore/preprocessor/__init__.py", line 285, in _run_preproc_function
return function(items, **kwargs)
File "/home/users/valeriu/esmvalcore/esmvalcore/preprocessor/_multimodel.py", line 414, in multi_model_statistics
return _multiproduct_statistics(
File "/home/users/valeriu/esmvalcore/esmvalcore/preprocessor/_multimodel.py", line 325, in _multiproduct_statistics
statistics_cubes = _multicube_statistics(cubes=cubes,
File "/home/users/valeriu/esmvalcore/esmvalcore/preprocessor/_multimodel.py", line 305, in _multicube_statistics
result_cube = _compute_eager(aligned_cubes,
File "/home/users/valeriu/esmvalcore/esmvalcore/preprocessor/_multimodel.py", line 250, in _compute_eager
combined_slice = _combine(single_model_slices)
File "/home/users/valeriu/esmvalcore/esmvalcore/preprocessor/_multimodel.py", line 220, in _combine
merged_cube = cubes.merge_cube()
File "/home/users/valeriu/miniconda3-June2021/envs/esmvaltool-stock/lib/python3.9/site-packages/iris/cube.py", line 402, in merge_cube
proto_cube.register(cube, error_on_mismatch=True)
File "/home/users/valeriu/miniconda3-June2021/envs/esmvaltool-stock/lib/python3.9/site-packages/iris/_merge.py", line 1361, in register
match = coord_payload.match_signature(
File "/home/users/valeriu/miniconda3-June2021/envs/esmvaltool-stock/lib/python3.9/site-packages/iris/_merge.py", line 284, in match_signature
raise iris.exceptions.MergeError(msgs)
iris.exceptions.MergeError: failed to merge into a single cube.
Coordinates in cube.dim_coords differ: air_pressure.
2021-10-01 10:35:41,885 UTC [16766] INFO
```
iris latest:
```
(esmvaltool-stock) [valeriu@sci2 esmvaltool_var_test]$ conda list iris
# packages in environment at /home/users/valeriu/miniconda3-June2021/envs/esmvaltool-stock:
#
# Name Version Build Channel
iris 3.1.0 py39hf3d152e_1 conda-forge
``` | 1.0 | Issue in multimodel cube merge surfaced by running `recipe_preprocessor_test.yml` - ```
2021-10-01 10:35:40,900 UTC [16801] ERROR Failed to run multi_model_statistics({<esmvalcore.preprocessor.PreprocessorFile object at 0x7f95f92d1190>, <esmvalcore.preprocessor.PreprocessorFile object at 0x7f95f92d1340>, <esmvalcore.preprocessor.PreprocessorFile object at 0x7f95f92b2d60>, <esmvalcore.preprocessor.PreprocessorFile object at 0x7f95f92d1070>}, {'span': 'overlap', 'statistics': ['mean', 'median'], 'output_products': {'mean': <esmvalcore.preprocessor.PreprocessorFile object at 0x7f95f92b2e80>, 'median': <esmvalcore.preprocessor.PreprocessorFile object at 0x7f95f92b2eb0>}})
2021-10-01 10:35:41,867 UTC [16766] INFO Maximum memory used (estimate): 6.2 GB
2021-10-01 10:35:41,868 UTC [16766] INFO Sampled every second. It may be inaccurate if short but high spikes in memory consumption occur.
2021-10-01 10:35:41,870 UTC [16766] ERROR Program terminated abnormally, see stack trace below for more information:
multiprocessing.pool.RemoteTraceback:
"""
Traceback (most recent call last):
File "/home/users/valeriu/miniconda3-June2021/envs/esmvaltool-stock/lib/python3.9/multiprocessing/pool.py", line 125, in worker
result = (True, func(*args, **kwds))
File "/home/users/valeriu/esmvalcore/esmvalcore/_task.py", line 747, in _run_task
output_files = task.run()
File "/home/users/valeriu/esmvalcore/esmvalcore/_task.py", line 252, in run
self.output_files = self._run(input_files)
File "/home/users/valeriu/esmvalcore/esmvalcore/preprocessor/__init__.py", line 481, in _run
self.products = _apply_multimodel(self.products, step,
File "/home/users/valeriu/esmvalcore/esmvalcore/preprocessor/__init__.py", line 422, in _apply_multimodel
result = preprocess(products - exclude, step, **settings)
File "/home/users/valeriu/esmvalcore/esmvalcore/preprocessor/__init__.py", line 299, in preprocess
result.append(_run_preproc_function(function, items, settings))
File "/home/users/valeriu/esmvalcore/esmvalcore/preprocessor/__init__.py", line 285, in _run_preproc_function
return function(items, **kwargs)
File "/home/users/valeriu/esmvalcore/esmvalcore/preprocessor/_multimodel.py", line 414, in multi_model_statistics
return _multiproduct_statistics(
File "/home/users/valeriu/esmvalcore/esmvalcore/preprocessor/_multimodel.py", line 325, in _multiproduct_statistics
statistics_cubes = _multicube_statistics(cubes=cubes,
File "/home/users/valeriu/esmvalcore/esmvalcore/preprocessor/_multimodel.py", line 305, in _multicube_statistics
result_cube = _compute_eager(aligned_cubes,
File "/home/users/valeriu/esmvalcore/esmvalcore/preprocessor/_multimodel.py", line 250, in _compute_eager
combined_slice = _combine(single_model_slices)
File "/home/users/valeriu/esmvalcore/esmvalcore/preprocessor/_multimodel.py", line 220, in _combine
merged_cube = cubes.merge_cube()
File "/home/users/valeriu/miniconda3-June2021/envs/esmvaltool-stock/lib/python3.9/site-packages/iris/cube.py", line 402, in merge_cube
proto_cube.register(cube, error_on_mismatch=True)
File "/home/users/valeriu/miniconda3-June2021/envs/esmvaltool-stock/lib/python3.9/site-packages/iris/_merge.py", line 1361, in register
match = coord_payload.match_signature(
File "/home/users/valeriu/miniconda3-June2021/envs/esmvaltool-stock/lib/python3.9/site-packages/iris/_merge.py", line 284, in match_signature
raise iris.exceptions.MergeError(msgs)
iris.exceptions.MergeError: failed to merge into a single cube.
Coordinates in cube.dim_coords differ: air_pressure.
"""
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/home/users/valeriu/esmvalcore/esmvalcore/_main.py", line 438, in run
fire.Fire(ESMValTool())
File "/home/users/valeriu/miniconda3-June2021/envs/esmvaltool-stock/lib/python3.9/site-packages/fire/core.py", line 141, in Fire
component_trace = _Fire(component, args, parsed_flag_args, context, name)
File "/home/users/valeriu/miniconda3-June2021/envs/esmvaltool-stock/lib/python3.9/site-packages/fire/core.py", line 466, in _Fire
component, remaining_args = _CallAndUpdateTrace(
File "/home/users/valeriu/miniconda3-June2021/envs/esmvaltool-stock/lib/python3.9/site-packages/fire/core.py", line 681, in _CallAndUpdateTrace
component = fn(*varargs, **kwargs)
File "/home/users/valeriu/esmvalcore/esmvalcore/_main.py", line 415, in run
process_recipe(recipe_file=recipe, config_user=cfg)
File "/home/users/valeriu/esmvalcore/esmvalcore/_main.py", line 105, in process_recipe
recipe.run()
File "/home/users/valeriu/esmvalcore/esmvalcore/_recipe.py", line 1515, in run
self.tasks.run(max_parallel_tasks=self._cfg['max_parallel_tasks'])
File "/home/users/valeriu/esmvalcore/esmvalcore/_task.py", line 669, in run
self._run_parallel(max_parallel_tasks)
File "/home/users/valeriu/esmvalcore/esmvalcore/_task.py", line 712, in _run_parallel
_copy_results(task, running[task])
File "/home/users/valeriu/esmvalcore/esmvalcore/_task.py", line 735, in _copy_results
task.output_files, updated_products = future.get()
File "/home/users/valeriu/miniconda3-June2021/envs/esmvaltool-stock/lib/python3.9/multiprocessing/pool.py", line 771, in get
raise self._value
File "/home/users/valeriu/miniconda3-June2021/envs/esmvaltool-stock/lib/python3.9/multiprocessing/pool.py", line 125, in worker
result = (True, func(*args, **kwds))
File "/home/users/valeriu/esmvalcore/esmvalcore/_task.py", line 747, in _run_task
output_files = task.run()
File "/home/users/valeriu/esmvalcore/esmvalcore/_task.py", line 252, in run
self.output_files = self._run(input_files)
File "/home/users/valeriu/esmvalcore/esmvalcore/preprocessor/__init__.py", line 481, in _run
self.products = _apply_multimodel(self.products, step,
File "/home/users/valeriu/esmvalcore/esmvalcore/preprocessor/__init__.py", line 422, in _apply_multimodel
result = preprocess(products - exclude, step, **settings)
File "/home/users/valeriu/esmvalcore/esmvalcore/preprocessor/__init__.py", line 299, in preprocess
result.append(_run_preproc_function(function, items, settings))
File "/home/users/valeriu/esmvalcore/esmvalcore/preprocessor/__init__.py", line 285, in _run_preproc_function
return function(items, **kwargs)
File "/home/users/valeriu/esmvalcore/esmvalcore/preprocessor/_multimodel.py", line 414, in multi_model_statistics
return _multiproduct_statistics(
File "/home/users/valeriu/esmvalcore/esmvalcore/preprocessor/_multimodel.py", line 325, in _multiproduct_statistics
statistics_cubes = _multicube_statistics(cubes=cubes,
File "/home/users/valeriu/esmvalcore/esmvalcore/preprocessor/_multimodel.py", line 305, in _multicube_statistics
result_cube = _compute_eager(aligned_cubes,
File "/home/users/valeriu/esmvalcore/esmvalcore/preprocessor/_multimodel.py", line 250, in _compute_eager
combined_slice = _combine(single_model_slices)
File "/home/users/valeriu/esmvalcore/esmvalcore/preprocessor/_multimodel.py", line 220, in _combine
merged_cube = cubes.merge_cube()
File "/home/users/valeriu/miniconda3-June2021/envs/esmvaltool-stock/lib/python3.9/site-packages/iris/cube.py", line 402, in merge_cube
proto_cube.register(cube, error_on_mismatch=True)
File "/home/users/valeriu/miniconda3-June2021/envs/esmvaltool-stock/lib/python3.9/site-packages/iris/_merge.py", line 1361, in register
match = coord_payload.match_signature(
File "/home/users/valeriu/miniconda3-June2021/envs/esmvaltool-stock/lib/python3.9/site-packages/iris/_merge.py", line 284, in match_signature
raise iris.exceptions.MergeError(msgs)
iris.exceptions.MergeError: failed to merge into a single cube.
Coordinates in cube.dim_coords differ: air_pressure.
2021-10-01 10:35:41,885 UTC [16766] INFO
```
iris latest:
```
(esmvaltool-stock) [valeriu@sci2 esmvaltool_var_test]$ conda list iris
# packages in environment at /home/users/valeriu/miniconda3-June2021/envs/esmvaltool-stock:
#
# Name Version Build Channel
iris 3.1.0 py39hf3d152e_1 conda-forge
``` | process | issue in multimodel cube merge surfaced by running recipe preprocessor test yml utc error failed to run multi model statistics span overlap statistics output products mean median utc info maximum memory used estimate gb utc info sampled every second it may be inaccurate if short but high spikes in memory consumption occur utc error program terminated abnormally see stack trace below for more information multiprocessing pool remotetraceback traceback most recent call last file home users valeriu envs esmvaltool stock lib multiprocessing pool py line in worker result true func args kwds file home users valeriu esmvalcore esmvalcore task py line in run task output files task run file home users valeriu esmvalcore esmvalcore task py line in run self output files self run input files file home users valeriu esmvalcore esmvalcore preprocessor init py line in run self products apply multimodel self products step file home users valeriu esmvalcore esmvalcore preprocessor init py line in apply multimodel result preprocess products exclude step settings file home users valeriu esmvalcore esmvalcore preprocessor init py line in preprocess result append run preproc function function items settings file home users valeriu esmvalcore esmvalcore preprocessor init py line in run preproc function return function items kwargs file home users valeriu esmvalcore esmvalcore preprocessor multimodel py line in multi model statistics return multiproduct statistics file home users valeriu esmvalcore esmvalcore preprocessor multimodel py line in multiproduct statistics statistics cubes multicube statistics cubes cubes file home users valeriu esmvalcore esmvalcore preprocessor multimodel py line in multicube statistics result cube compute eager aligned cubes file home users valeriu esmvalcore esmvalcore preprocessor multimodel py line in compute eager combined slice combine single model slices file home users valeriu esmvalcore esmvalcore preprocessor multimodel py line in combine merged cube cubes merge cube file home users valeriu envs esmvaltool stock lib site packages iris cube py line in merge cube proto cube register cube error on mismatch true file home users valeriu envs esmvaltool stock lib site packages iris merge py line in register match coord payload match signature file home users valeriu envs esmvaltool stock lib site packages iris merge py line in match signature raise iris exceptions mergeerror msgs iris exceptions mergeerror failed to merge into a single cube coordinates in cube dim coords differ air pressure the above exception was the direct cause of the following exception traceback most recent call last file home users valeriu esmvalcore esmvalcore main py line in run fire fire esmvaltool file home users valeriu envs esmvaltool stock lib site packages fire core py line in fire component trace fire component args parsed flag args context name file home users valeriu envs esmvaltool stock lib site packages fire core py line in fire component remaining args callandupdatetrace file home users valeriu envs esmvaltool stock lib site packages fire core py line in callandupdatetrace component fn varargs kwargs file home users valeriu esmvalcore esmvalcore main py line in run process recipe recipe file recipe config user cfg file home users valeriu esmvalcore esmvalcore main py line in process recipe recipe run file home users valeriu esmvalcore esmvalcore recipe py line in run self tasks run max parallel tasks self cfg file home users valeriu esmvalcore esmvalcore task py line in run self run parallel max parallel tasks file home users valeriu esmvalcore esmvalcore task py line in run parallel copy results task running file home users valeriu esmvalcore esmvalcore task py line in copy results task output files updated products future get file home users valeriu envs esmvaltool stock lib multiprocessing pool py line in get raise self value file home users valeriu envs esmvaltool stock lib multiprocessing pool py line in worker result true func args kwds file home users valeriu esmvalcore esmvalcore task py line in run task output files task run file home users valeriu esmvalcore esmvalcore task py line in run self output files self run input files file home users valeriu esmvalcore esmvalcore preprocessor init py line in run self products apply multimodel self products step file home users valeriu esmvalcore esmvalcore preprocessor init py line in apply multimodel result preprocess products exclude step settings file home users valeriu esmvalcore esmvalcore preprocessor init py line in preprocess result append run preproc function function items settings file home users valeriu esmvalcore esmvalcore preprocessor init py line in run preproc function return function items kwargs file home users valeriu esmvalcore esmvalcore preprocessor multimodel py line in multi model statistics return multiproduct statistics file home users valeriu esmvalcore esmvalcore preprocessor multimodel py line in multiproduct statistics statistics cubes multicube statistics cubes cubes file home users valeriu esmvalcore esmvalcore preprocessor multimodel py line in multicube statistics result cube compute eager aligned cubes file home users valeriu esmvalcore esmvalcore preprocessor multimodel py line in compute eager combined slice combine single model slices file home users valeriu esmvalcore esmvalcore preprocessor multimodel py line in combine merged cube cubes merge cube file home users valeriu envs esmvaltool stock lib site packages iris cube py line in merge cube proto cube register cube error on mismatch true file home users valeriu envs esmvaltool stock lib site packages iris merge py line in register match coord payload match signature file home users valeriu envs esmvaltool stock lib site packages iris merge py line in match signature raise iris exceptions mergeerror msgs iris exceptions mergeerror failed to merge into a single cube coordinates in cube dim coords differ air pressure utc info iris latest esmvaltool stock conda list iris packages in environment at home users valeriu envs esmvaltool stock name version build channel iris conda forge | 1 |
9,022 | 12,126,629,722 | IssuesEvent | 2020-04-22 17:20:43 | MicrosoftDocs/azure-devops-docs | https://api.github.com/repos/MicrosoftDocs/azure-devops-docs | reopened | System Variables do not seem to be available in the template condition despite what this documentation page claims | devops-cicd-process/tech devops/prod doc-enhancement | Hey,
This [page](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/templates?view=azure-devops#context) claims that:
> Template expressions
> ...
> Additionally, you have access to the variables context that contains all the variables specified in the YAML file plus the [system variables](https://docs.microsoft.com/en-us/azure/devops/pipelines/build/variables?view=azure-devops&tabs=yaml#system-variables).
I have a hard time trying to make it work. Consider the following pipeline definition. It uses template expression where it checks a value from `System.PullRequest.PullRequestId` system variable:
```yaml
variables:
systemVariableCondition: $[lt(0, variables['System.PullRequest.PullRequestId'])]
${{ if lt(0, variables['System.PullRequest.PullRequestId']) }}:
dependentVariable: "some value"
phases:
- phase: Build
displayName: Build
steps:
- powershell: |
Write-Host "dependentVariable: $(dependentVariable)"
```
Run-time variable `systemVariableCondition` is in this definition just to confirm that system variable in question has expected value.
Here is the output from the "Prepare job" step, please note that expected variable `dependentVariable` was *not* introduced:
```
##[section]Starting: Prepare job Build
Variables:
systemVariableCondition:
Parsing expression: <lt(0, variables['System.PullRequest.PullRequestId'])>
Evaluating: lt(0, variables['System.PullRequest.PullRequestId'])
Expanded: lt(0, '335')
Result: 'True'
ContinueOnError: False
TimeoutInMinutes: 60
CancelTimeoutInMinutes: 5
Execution:
##[section]Finishing: Prepare job Build
```
So the step fails:
```
+ Write-Host "dependentVariable: $(dependentVariable)"
+ ~~~~~~~~~~~~~~~~~
+ CategoryInfo : ObjectNotFound: (dependentVariable:String) [], ParentContainsErrorRecordException
+ FullyQualifiedErrorId : CommandNotFoundException
##[error]PowerShell exited with code '1'.
```
Please elaborate what exactly this page means by stating that one has access to system variables in a template expression.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 6724abea-bbdc-bf66-ed5e-3214fa6c3e66
* Version Independent ID: 4f8dab21-3f0e-da32-cc0e-1d85c13c0065
* Content: [Templates - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/templates?view=azure-devops)
* Content Source: [docs/pipelines/process/templates.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/templates.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam** | 1.0 | System Variables do not seem to be available in the template condition despite what this documentation page claims - Hey,
This [page](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/templates?view=azure-devops#context) claims that:
> Template expressions
> ...
> Additionally, you have access to the variables context that contains all the variables specified in the YAML file plus the [system variables](https://docs.microsoft.com/en-us/azure/devops/pipelines/build/variables?view=azure-devops&tabs=yaml#system-variables).
I have a hard time trying to make it work. Consider the following pipeline definition. It uses template expression where it checks a value from `System.PullRequest.PullRequestId` system variable:
```yaml
variables:
systemVariableCondition: $[lt(0, variables['System.PullRequest.PullRequestId'])]
${{ if lt(0, variables['System.PullRequest.PullRequestId']) }}:
dependentVariable: "some value"
phases:
- phase: Build
displayName: Build
steps:
- powershell: |
Write-Host "dependentVariable: $(dependentVariable)"
```
Run-time variable `systemVariableCondition` is in this definition just to confirm that system variable in question has expected value.
Here is the output from the "Prepare job" step, please note that expected variable `dependentVariable` was *not* introduced:
```
##[section]Starting: Prepare job Build
Variables:
systemVariableCondition:
Parsing expression: <lt(0, variables['System.PullRequest.PullRequestId'])>
Evaluating: lt(0, variables['System.PullRequest.PullRequestId'])
Expanded: lt(0, '335')
Result: 'True'
ContinueOnError: False
TimeoutInMinutes: 60
CancelTimeoutInMinutes: 5
Execution:
##[section]Finishing: Prepare job Build
```
So the step fails:
```
+ Write-Host "dependentVariable: $(dependentVariable)"
+ ~~~~~~~~~~~~~~~~~
+ CategoryInfo : ObjectNotFound: (dependentVariable:String) [], ParentContainsErrorRecordException
+ FullyQualifiedErrorId : CommandNotFoundException
##[error]PowerShell exited with code '1'.
```
Please elaborate what exactly this page means by stating that one has access to system variables in a template expression.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 6724abea-bbdc-bf66-ed5e-3214fa6c3e66
* Version Independent ID: 4f8dab21-3f0e-da32-cc0e-1d85c13c0065
* Content: [Templates - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/templates?view=azure-devops)
* Content Source: [docs/pipelines/process/templates.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/templates.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam** | process | system variables do not seem to be available in the template condition despite what this documentation page claims hey this claims that template expressions additionally you have access to the variables context that contains all the variables specified in the yaml file plus the i have a hard time trying to make it work consider the following pipeline definition it uses template expression where it checks a value from system pullrequest pullrequestid system variable yaml variables systemvariablecondition if lt variables dependentvariable some value phases phase build displayname build steps powershell write host dependentvariable dependentvariable run time variable systemvariablecondition is in this definition just to confirm that system variable in question has expected value here is the output from the prepare job step please note that expected variable dependentvariable was not introduced starting prepare job build variables systemvariablecondition parsing expression lt lt variables gt evaluating lt variables expanded lt result true continueonerror false timeoutinminutes canceltimeoutinminutes execution finishing prepare job build so the step fails write host dependentvariable dependentvariable categoryinfo objectnotfound dependentvariable string parentcontainserrorrecordexception fullyqualifiederrorid commandnotfoundexception powershell exited with code please elaborate what exactly this page means by stating that one has access to system variables in a template expression document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id bbdc version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam | 1 |
18,916 | 24,862,827,911 | IssuesEvent | 2022-10-27 09:34:54 | aiidateam/aiida-core | https://api.github.com/repos/aiidateam/aiida-core | closed | A lot of failed jobs in AiiDA 1.5 (DuplicateSubcriber) when machine is overloaded | type/bug priority/critical-blocking topic/rabbitmq topic/engine topic/processes | Duplicate of #3973
When submitting *a lot* of jobs in AiiDA 1.5.0, I get many failed (excepted) jobs with reports like the following one:
```
Traceback (most recent call last):
File "/home/pizzi/.virtualenvs/aiida-prod/codes/aiida-core/aiida/manage/external/rmq.py", line 201, in _continue
result = yield super()._continue(communicator, pid, nowait, tag)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1055, in run
value = future.result()
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/concurrent.py", line 238, in result
raise_exc_info(self._exc_info)
File "<string>", line 4, in raise_exc_info
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 307, in wrapper
yielded = next(result)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/process_comms.py", line 541, in _continue
proc = saved_state.unbundle(self._load_context)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/persistence.py", line 51, in unbundle
return Savable.load(self, load_context)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/persistence.py", line 442, in load
return load_cls.recreate_from(saved_state, load_context)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/processes.py", line 234, in recreate_from
base.call_with_super_check(process.init)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/base/utils.py", line 28, in call_with_super_check
wrapped(*args, **kwargs)
File "/home/pizzi/.virtualenvs/aiida-prod/codes/aiida-core/aiida/engine/processes/process.py", line 125, in init
super().init()
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/base/utils.py", line 15, in wrapper
wrapped(self, *args, **kwargs)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/processes.py", line 293, in init
self.broadcast_receive, identifier=str(self.pid)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/communications.py", line 125, in add_broadcast_subscriber
return self._communicator.add_broadcast_subscriber(converted, identifier)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/kiwipy/rmq/communicator.py", line 592, in add_broadcast_subscriber
return self._run_task(coro)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/kiwipy/rmq/communicator.py", line 677, in _run_task
return self.tornado_to_kiwi_future(self._create_task(coro)).result(timeout=self.TASK_TIMEOUT)
File "/usr/lib/python3.7/concurrent/futures/_base.py", line 435, in result
return self.__get_result()
File "/usr/lib/python3.7/concurrent/futures/_base.py", line 384, in __get_result
raise self._exception
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/kiwipy/rmq/communicator.py", line 646, in done
result = done_future.result()
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/concurrent.py", line 238, in result
raise_exc_info(self._exc_info)
File "<string>", line 4, in raise_exc_info
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/kiwipy/futures.py", line 54, in capture_exceptions
yield
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/kiwipy/rmq/utils.py", line 146, in run_task
future.set_result((yield coro()))
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1055, in run
value = future.result()
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/concurrent.py", line 238, in result
raise_exc_info(self._exc_info)
File "<string>", line 4, in raise_exc_info
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1063, in run
yielded = self.gen.throw(*exc_info)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/kiwipy/rmq/communicator.py", line 393, in add_broadcast_subscriber
identifier = yield self._message_subscriber.add_broadcast_subscriber(subscriber, identifier)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1055, in run
value = future.result()
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/concurrent.py", line 238, in result
raise_exc_info(self._exc_info)
File "<string>", line 4, in raise_exc_info
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 307, in wrapper
yielded = next(result)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/kiwipy/rmq/communicator.py", line 146, in add_broadcast_subscriber
raise kiwipy.DuplicateSubscriberIdentifier("Broadcast identifier '{}'".format(identifier))
kiwipy.communications.DuplicateSubscriberIdentifier: Broadcast identifier '52632'
```
or
```
Traceback (most recent call last):
File "/home/pizzi/.virtualenvs/aiida-prod/codes/aiida-core/aiida/manage/external/rmq.py", line 201, in _continue
result = yield super()._continue(communicator, pid, nowait, tag)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1055, in run
value = future.result()
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/concurrent.py", line 238, in result
raise_exc_info(self._exc_info)
File "<string>", line 4, in raise_exc_info
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 307, in wrapper
yielded = next(result)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/process_comms.py", line 541, in _continue
proc = saved_state.unbundle(self._load_context)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/persistence.py", line 51, in unbundle
return Savable.load(self, load_context)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/persistence.py", line 442, in load
return load_cls.recreate_from(saved_state, load_context)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/processes.py", line 234, in recreate_from
base.call_with_super_check(process.init)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/base/utils.py", line 28, in call_with_super_check
wrapped(*args, **kwargs)
File "/home/pizzi/.virtualenvs/aiida-prod/codes/aiida-core/aiida/engine/processes/process.py", line 125, in init
super().init()
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/base/utils.py", line 15, in wrapper
wrapped(self, *args, **kwargs)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/processes.py", line 293, in init
self.broadcast_receive, identifier=str(self.pid)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/communications.py", line 125, in add_broadcast_subscriber
return self._communicator.add_broadcast_subscriber(converted, identifier)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/kiwipy/rmq/communicator.py", line 592, in add_broadcast_subscriber
return self._run_task(coro)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/kiwipy/rmq/communicator.py", line 677, in _run_task
return self.tornado_to_kiwi_future(self._create_task(coro)).result(timeout=self.TASK_TIMEOUT)
File "/usr/lib/python3.7/concurrent/futures/_base.py", line 435, in result
return self.__get_result()
File "/usr/lib/python3.7/concurrent/futures/_base.py", line 384, in __get_result
raise self._exception
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/kiwipy/rmq/communicator.py", line 646, in done
result = done_future.result()
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/concurrent.py", line 238, in result
raise_exc_info(self._exc_info)
File "<string>", line 4, in raise_exc_info
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/kiwipy/futures.py", line 54, in capture_exceptions
yield
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/kiwipy/rmq/utils.py", line 146, in run_task
future.set_result((yield coro()))
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1055, in run
value = future.result()
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/concurrent.py", line 238, in result
raise_exc_info(self._exc_info)
File "<string>", line 4, in raise_exc_info
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1063, in run
yielded = self.gen.throw(*exc_info)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/kiwipy/rmq/communicator.py", line 393, in add_broadcast_subscriber
identifier = yield self._message_subscriber.add_broadcast_subscriber(subscriber, identifier)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1055, in run
value = future.result()
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/concurrent.py", line 238, in result
raise_exc_info(self._exc_info)
File "<string>", line 4, in raise_exc_info
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 307, in wrapper
yielded = next(result)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/kiwipy/rmq/communicator.py", line 146, in add_broadcast_subscriber
raise kiwipy.DuplicateSubscriberIdentifier("Broadcast identifier '{}'".format(identifier))
kiwipy.communications.DuplicateSubscriberIdentifier: Broadcast identifier '51047'
2020-11-29 11:22:39 [27009 | ERROR]: Traceback (most recent call last):
File "/home/pizzi/.virtualenvs/aiida-prod/codes/aiida-core/aiida/manage/external/rmq.py", line 201, in _continue
result = yield super()._continue(communicator, pid, nowait, tag)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1055, in run
value = future.result()
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/concurrent.py", line 238, in result
raise_exc_info(self._exc_info)
File "<string>", line 4, in raise_exc_info
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1063, in run
yielded = self.gen.throw(*exc_info)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/process_comms.py", line 547, in _continue
yield proc.step_until_terminated()
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1055, in run
value = future.result()
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/concurrent.py", line 238, in result
raise_exc_info(self._exc_info)
File "<string>", line 4, in raise_exc_info
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1063, in run
yielded = self.gen.throw(*exc_info)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/processes.py", line 1117, in step_until_terminated
yield self.step()
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1055, in run
value = future.result()
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/concurrent.py", line 238, in result
raise_exc_info(self._exc_info)
File "<string>", line 4, in raise_exc_info
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1069, in run
yielded = self.gen.send(value)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/processes.py", line 1108, in step
self.transition_to(next_state)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/base/state_machine.py", line 318, in transition_to
self.transition_failed(initial_state_label, label, *sys.exc_info()[1:])
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/base/state_machine.py", line 332, in transition_failed
raise exception.with_traceback(trace)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/base/state_machine.py", line 302, in transition_to
self._enter_next_state(new_state)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/base/state_machine.py", line 367, in _enter_next_state
self._fire_state_event(StateEventHook.ENTERED_STATE, last_state)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/base/state_machine.py", line 281, in _fire_state_event
callback(self, hook, state)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/processes.py", line 313, in <lambda>
state_machine.StateEventHook.ENTERED_STATE, lambda _s, _h, from_state: self.on_entered(from_state)
File "/home/pizzi/.virtualenvs/aiida-prod/codes/aiida-core/aiida/engine/processes/process.py", line 339, in on_entered
super().on_entered(from_state)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/processes.py", line 639, in on_entered
self._communicator.broadcast_send(body=None, sender=self.pid, subject=subject)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/communications.py", line 137, in broadcast_send
return self._communicator.broadcast_send(body, sender, subject, correlation_id)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/kiwipy/rmq/communicator.py", line 619, in broadcast_send
correlation_id=correlation_id))
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/kiwipy/rmq/communicator.py", line 630, in _send_message
return send_future.result()
File "/usr/lib/python3.7/concurrent/futures/_base.py", line 435, in result
return self.__get_result()
File "/usr/lib/python3.7/concurrent/futures/_base.py", line 384, in __get_result
raise self._exception
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/kiwipy/futures.py", line 54, in capture_exceptions
yield
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/kiwipy/rmq/communicator.py", line 627, in do_task
send_future.set_result((yield coro()))
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1055, in run
value = future.result()
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/concurrent.py", line 238, in result
raise_exc_info(self._exc_info)
File "<string>", line 4, in raise_exc_info
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1063, in run
yielded = self.gen.throw(*exc_info)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/kiwipy/rmq/communicator.py", line 417, in broadcast_send
result = yield self._message_publisher.broadcast_send(body, sender, subject, correlation_id)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1055, in run
value = future.result()
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/concurrent.py", line 238, in result
raise_exc_info(self._exc_info)
File "<string>", line 4, in raise_exc_info
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1063, in run
yielded = self.gen.throw(*exc_info)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/kiwipy/rmq/communicator.py", line 57, in broadcast_send
result = yield self.publish(message, routing_key=defaults.BROADCAST_TOPIC)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1055, in run
value = future.result()
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/concurrent.py", line 238, in result
raise_exc_info(self._exc_info)
File "<string>", line 4, in raise_exc_info
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1063, in run
yielded = self.gen.throw(*exc_info)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/kiwipy/rmq/messages.py", line 207, in publish
message, routing_key=routing_key, mandatory=mandatory)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1055, in run
value = future.result()
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/concurrent.py", line 238, in result
raise_exc_info(self._exc_info)
File "<string>", line 4, in raise_exc_info
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1063, in run
yielded = self.gen.throw(*exc_info)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/topika/common.py", line 174, in wrap
raise gen.Return((yield func(self, *args, **kwargs)))
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1055, in run
value = future.result()
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/concurrent.py", line 238, in result
raise_exc_info(self._exc_info)
File "<string>", line 4, in raise_exc_info
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1063, in run
yielded = self.gen.throw(*exc_info)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/topika/exchange.py", line 210, in publish
mandatory=mandatory)))
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1055, in run
value = future.result()
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/concurrent.py", line 238, in result
raise_exc_info(self._exc_info)
File "<string>", line 4, in raise_exc_info
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 307, in wrapper
yielded = next(result)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/topika/common.py", line 172, in wrap
raise RuntimeError("The channel is closed")
RuntimeError: The channel is closed
```
I think other people observed the same.
The other thing I noticed (not sure if a cause or a consequence) is that now `verdi process list` gives an empty list, **but** the daemon workers are (while not busy) still using *a lot* of memory:
```
$ verdi daemon status
Profile: dispero2020
Daemon is running as PID 62212 since 2020-11-28 23:41:36
Active workers [8]:
PID MEM % CPU % started
----- ------- ------- -------------------
62216 6.786 0 2020-11-28 23:41:37
62217 6.434 0 2020-11-28 23:41:37
62218 7.091 0 2020-11-28 23:41:37
62219 9.856 0 2020-11-28 23:41:37
62220 15.758 0 2020-11-28 23:41:37
62221 5.442 0 2020-11-28 23:41:37
62222 8.017 0 2020-11-28 23:41:37
62223 8.67 0 2020-11-28 23:41:37
Use verdi daemon [incr | decr] [num] to increase / decrease the amount of workers
```
I don't know if this is a memory leak. What I can say is that I was submitting quite a lot of processes/workflows and my machine was under heavy stress. Maybe everything became so slow that the two heartbeats were missed?
@sphuber @muhrin @unkcpz @chrisjsewell | 1.0 | A lot of failed jobs in AiiDA 1.5 (DuplicateSubcriber) when machine is overloaded - Duplicate of #3973
When submitting *a lot* of jobs in AiiDA 1.5.0, I get many failed (excepted) jobs with reports like the following one:
```
Traceback (most recent call last):
File "/home/pizzi/.virtualenvs/aiida-prod/codes/aiida-core/aiida/manage/external/rmq.py", line 201, in _continue
result = yield super()._continue(communicator, pid, nowait, tag)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1055, in run
value = future.result()
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/concurrent.py", line 238, in result
raise_exc_info(self._exc_info)
File "<string>", line 4, in raise_exc_info
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 307, in wrapper
yielded = next(result)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/process_comms.py", line 541, in _continue
proc = saved_state.unbundle(self._load_context)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/persistence.py", line 51, in unbundle
return Savable.load(self, load_context)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/persistence.py", line 442, in load
return load_cls.recreate_from(saved_state, load_context)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/processes.py", line 234, in recreate_from
base.call_with_super_check(process.init)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/base/utils.py", line 28, in call_with_super_check
wrapped(*args, **kwargs)
File "/home/pizzi/.virtualenvs/aiida-prod/codes/aiida-core/aiida/engine/processes/process.py", line 125, in init
super().init()
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/base/utils.py", line 15, in wrapper
wrapped(self, *args, **kwargs)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/processes.py", line 293, in init
self.broadcast_receive, identifier=str(self.pid)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/communications.py", line 125, in add_broadcast_subscriber
return self._communicator.add_broadcast_subscriber(converted, identifier)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/kiwipy/rmq/communicator.py", line 592, in add_broadcast_subscriber
return self._run_task(coro)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/kiwipy/rmq/communicator.py", line 677, in _run_task
return self.tornado_to_kiwi_future(self._create_task(coro)).result(timeout=self.TASK_TIMEOUT)
File "/usr/lib/python3.7/concurrent/futures/_base.py", line 435, in result
return self.__get_result()
File "/usr/lib/python3.7/concurrent/futures/_base.py", line 384, in __get_result
raise self._exception
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/kiwipy/rmq/communicator.py", line 646, in done
result = done_future.result()
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/concurrent.py", line 238, in result
raise_exc_info(self._exc_info)
File "<string>", line 4, in raise_exc_info
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/kiwipy/futures.py", line 54, in capture_exceptions
yield
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/kiwipy/rmq/utils.py", line 146, in run_task
future.set_result((yield coro()))
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1055, in run
value = future.result()
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/concurrent.py", line 238, in result
raise_exc_info(self._exc_info)
File "<string>", line 4, in raise_exc_info
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1063, in run
yielded = self.gen.throw(*exc_info)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/kiwipy/rmq/communicator.py", line 393, in add_broadcast_subscriber
identifier = yield self._message_subscriber.add_broadcast_subscriber(subscriber, identifier)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1055, in run
value = future.result()
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/concurrent.py", line 238, in result
raise_exc_info(self._exc_info)
File "<string>", line 4, in raise_exc_info
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 307, in wrapper
yielded = next(result)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/kiwipy/rmq/communicator.py", line 146, in add_broadcast_subscriber
raise kiwipy.DuplicateSubscriberIdentifier("Broadcast identifier '{}'".format(identifier))
kiwipy.communications.DuplicateSubscriberIdentifier: Broadcast identifier '52632'
```
or
```
Traceback (most recent call last):
File "/home/pizzi/.virtualenvs/aiida-prod/codes/aiida-core/aiida/manage/external/rmq.py", line 201, in _continue
result = yield super()._continue(communicator, pid, nowait, tag)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1055, in run
value = future.result()
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/concurrent.py", line 238, in result
raise_exc_info(self._exc_info)
File "<string>", line 4, in raise_exc_info
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 307, in wrapper
yielded = next(result)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/process_comms.py", line 541, in _continue
proc = saved_state.unbundle(self._load_context)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/persistence.py", line 51, in unbundle
return Savable.load(self, load_context)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/persistence.py", line 442, in load
return load_cls.recreate_from(saved_state, load_context)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/processes.py", line 234, in recreate_from
base.call_with_super_check(process.init)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/base/utils.py", line 28, in call_with_super_check
wrapped(*args, **kwargs)
File "/home/pizzi/.virtualenvs/aiida-prod/codes/aiida-core/aiida/engine/processes/process.py", line 125, in init
super().init()
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/base/utils.py", line 15, in wrapper
wrapped(self, *args, **kwargs)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/processes.py", line 293, in init
self.broadcast_receive, identifier=str(self.pid)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/communications.py", line 125, in add_broadcast_subscriber
return self._communicator.add_broadcast_subscriber(converted, identifier)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/kiwipy/rmq/communicator.py", line 592, in add_broadcast_subscriber
return self._run_task(coro)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/kiwipy/rmq/communicator.py", line 677, in _run_task
return self.tornado_to_kiwi_future(self._create_task(coro)).result(timeout=self.TASK_TIMEOUT)
File "/usr/lib/python3.7/concurrent/futures/_base.py", line 435, in result
return self.__get_result()
File "/usr/lib/python3.7/concurrent/futures/_base.py", line 384, in __get_result
raise self._exception
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/kiwipy/rmq/communicator.py", line 646, in done
result = done_future.result()
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/concurrent.py", line 238, in result
raise_exc_info(self._exc_info)
File "<string>", line 4, in raise_exc_info
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/kiwipy/futures.py", line 54, in capture_exceptions
yield
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/kiwipy/rmq/utils.py", line 146, in run_task
future.set_result((yield coro()))
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1055, in run
value = future.result()
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/concurrent.py", line 238, in result
raise_exc_info(self._exc_info)
File "<string>", line 4, in raise_exc_info
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1063, in run
yielded = self.gen.throw(*exc_info)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/kiwipy/rmq/communicator.py", line 393, in add_broadcast_subscriber
identifier = yield self._message_subscriber.add_broadcast_subscriber(subscriber, identifier)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1055, in run
value = future.result()
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/concurrent.py", line 238, in result
raise_exc_info(self._exc_info)
File "<string>", line 4, in raise_exc_info
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 307, in wrapper
yielded = next(result)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/kiwipy/rmq/communicator.py", line 146, in add_broadcast_subscriber
raise kiwipy.DuplicateSubscriberIdentifier("Broadcast identifier '{}'".format(identifier))
kiwipy.communications.DuplicateSubscriberIdentifier: Broadcast identifier '51047'
2020-11-29 11:22:39 [27009 | ERROR]: Traceback (most recent call last):
File "/home/pizzi/.virtualenvs/aiida-prod/codes/aiida-core/aiida/manage/external/rmq.py", line 201, in _continue
result = yield super()._continue(communicator, pid, nowait, tag)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1055, in run
value = future.result()
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/concurrent.py", line 238, in result
raise_exc_info(self._exc_info)
File "<string>", line 4, in raise_exc_info
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1063, in run
yielded = self.gen.throw(*exc_info)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/process_comms.py", line 547, in _continue
yield proc.step_until_terminated()
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1055, in run
value = future.result()
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/concurrent.py", line 238, in result
raise_exc_info(self._exc_info)
File "<string>", line 4, in raise_exc_info
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1063, in run
yielded = self.gen.throw(*exc_info)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/processes.py", line 1117, in step_until_terminated
yield self.step()
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1055, in run
value = future.result()
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/concurrent.py", line 238, in result
raise_exc_info(self._exc_info)
File "<string>", line 4, in raise_exc_info
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1069, in run
yielded = self.gen.send(value)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/processes.py", line 1108, in step
self.transition_to(next_state)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/base/state_machine.py", line 318, in transition_to
self.transition_failed(initial_state_label, label, *sys.exc_info()[1:])
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/base/state_machine.py", line 332, in transition_failed
raise exception.with_traceback(trace)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/base/state_machine.py", line 302, in transition_to
self._enter_next_state(new_state)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/base/state_machine.py", line 367, in _enter_next_state
self._fire_state_event(StateEventHook.ENTERED_STATE, last_state)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/base/state_machine.py", line 281, in _fire_state_event
callback(self, hook, state)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/processes.py", line 313, in <lambda>
state_machine.StateEventHook.ENTERED_STATE, lambda _s, _h, from_state: self.on_entered(from_state)
File "/home/pizzi/.virtualenvs/aiida-prod/codes/aiida-core/aiida/engine/processes/process.py", line 339, in on_entered
super().on_entered(from_state)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/processes.py", line 639, in on_entered
self._communicator.broadcast_send(body=None, sender=self.pid, subject=subject)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/plumpy/communications.py", line 137, in broadcast_send
return self._communicator.broadcast_send(body, sender, subject, correlation_id)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/kiwipy/rmq/communicator.py", line 619, in broadcast_send
correlation_id=correlation_id))
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/kiwipy/rmq/communicator.py", line 630, in _send_message
return send_future.result()
File "/usr/lib/python3.7/concurrent/futures/_base.py", line 435, in result
return self.__get_result()
File "/usr/lib/python3.7/concurrent/futures/_base.py", line 384, in __get_result
raise self._exception
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/kiwipy/futures.py", line 54, in capture_exceptions
yield
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/kiwipy/rmq/communicator.py", line 627, in do_task
send_future.set_result((yield coro()))
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1055, in run
value = future.result()
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/concurrent.py", line 238, in result
raise_exc_info(self._exc_info)
File "<string>", line 4, in raise_exc_info
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1063, in run
yielded = self.gen.throw(*exc_info)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/kiwipy/rmq/communicator.py", line 417, in broadcast_send
result = yield self._message_publisher.broadcast_send(body, sender, subject, correlation_id)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1055, in run
value = future.result()
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/concurrent.py", line 238, in result
raise_exc_info(self._exc_info)
File "<string>", line 4, in raise_exc_info
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1063, in run
yielded = self.gen.throw(*exc_info)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/kiwipy/rmq/communicator.py", line 57, in broadcast_send
result = yield self.publish(message, routing_key=defaults.BROADCAST_TOPIC)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1055, in run
value = future.result()
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/concurrent.py", line 238, in result
raise_exc_info(self._exc_info)
File "<string>", line 4, in raise_exc_info
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1063, in run
yielded = self.gen.throw(*exc_info)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/kiwipy/rmq/messages.py", line 207, in publish
message, routing_key=routing_key, mandatory=mandatory)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1055, in run
value = future.result()
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/concurrent.py", line 238, in result
raise_exc_info(self._exc_info)
File "<string>", line 4, in raise_exc_info
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1063, in run
yielded = self.gen.throw(*exc_info)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/topika/common.py", line 174, in wrap
raise gen.Return((yield func(self, *args, **kwargs)))
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1055, in run
value = future.result()
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/concurrent.py", line 238, in result
raise_exc_info(self._exc_info)
File "<string>", line 4, in raise_exc_info
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1063, in run
yielded = self.gen.throw(*exc_info)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/topika/exchange.py", line 210, in publish
mandatory=mandatory)))
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 1055, in run
value = future.result()
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/concurrent.py", line 238, in result
raise_exc_info(self._exc_info)
File "<string>", line 4, in raise_exc_info
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/tornado/gen.py", line 307, in wrapper
yielded = next(result)
File "/home/pizzi/.virtualenvs/aiida-prod/lib/python3.7/site-packages/topika/common.py", line 172, in wrap
raise RuntimeError("The channel is closed")
RuntimeError: The channel is closed
```
I think other people observed the same.
The other thing I noticed (not sure if a cause or a consequence) is that now `verdi process list` gives an empty list, **but** the daemon workers are (while not busy) still using *a lot* of memory:
```
$ verdi daemon status
Profile: dispero2020
Daemon is running as PID 62212 since 2020-11-28 23:41:36
Active workers [8]:
PID MEM % CPU % started
----- ------- ------- -------------------
62216 6.786 0 2020-11-28 23:41:37
62217 6.434 0 2020-11-28 23:41:37
62218 7.091 0 2020-11-28 23:41:37
62219 9.856 0 2020-11-28 23:41:37
62220 15.758 0 2020-11-28 23:41:37
62221 5.442 0 2020-11-28 23:41:37
62222 8.017 0 2020-11-28 23:41:37
62223 8.67 0 2020-11-28 23:41:37
Use verdi daemon [incr | decr] [num] to increase / decrease the amount of workers
```
I don't know if this is a memory leak. What I can say is that I was submitting quite a lot of processes/workflows and my machine was under heavy stress. Maybe everything became so slow that the two heartbeats were missed?
@sphuber @muhrin @unkcpz @chrisjsewell | process | a lot of failed jobs in aiida duplicatesubcriber when machine is overloaded duplicate of when submitting a lot of jobs in aiida i get many failed excepted jobs with reports like the following one traceback most recent call last file home pizzi virtualenvs aiida prod codes aiida core aiida manage external rmq py line in continue result yield super continue communicator pid nowait tag file home pizzi virtualenvs aiida prod lib site packages tornado gen py line in run value future result file home pizzi virtualenvs aiida prod lib site packages tornado concurrent py line in result raise exc info self exc info file line in raise exc info file home pizzi virtualenvs aiida prod lib site packages tornado gen py line in wrapper yielded next result file home pizzi virtualenvs aiida prod lib site packages plumpy process comms py line in continue proc saved state unbundle self load context file home pizzi virtualenvs aiida prod lib site packages plumpy persistence py line in unbundle return savable load self load context file home pizzi virtualenvs aiida prod lib site packages plumpy persistence py line in load return load cls recreate from saved state load context file home pizzi virtualenvs aiida prod lib site packages plumpy processes py line in recreate from base call with super check process init file home pizzi virtualenvs aiida prod lib site packages plumpy base utils py line in call with super check wrapped args kwargs file home pizzi virtualenvs aiida prod codes aiida core aiida engine processes process py line in init super init file home pizzi virtualenvs aiida prod lib site packages plumpy base utils py line in wrapper wrapped self args kwargs file home pizzi virtualenvs aiida prod lib site packages plumpy processes py line in init self broadcast receive identifier str self pid file home pizzi virtualenvs aiida prod lib site packages plumpy communications py line in add broadcast subscriber return self communicator add broadcast subscriber converted identifier file home pizzi virtualenvs aiida prod lib site packages kiwipy rmq communicator py line in add broadcast subscriber return self run task coro file home pizzi virtualenvs aiida prod lib site packages kiwipy rmq communicator py line in run task return self tornado to kiwi future self create task coro result timeout self task timeout file usr lib concurrent futures base py line in result return self get result file usr lib concurrent futures base py line in get result raise self exception file home pizzi virtualenvs aiida prod lib site packages kiwipy rmq communicator py line in done result done future result file home pizzi virtualenvs aiida prod lib site packages tornado concurrent py line in result raise exc info self exc info file line in raise exc info file home pizzi virtualenvs aiida prod lib site packages kiwipy futures py line in capture exceptions yield file home pizzi virtualenvs aiida prod lib site packages kiwipy rmq utils py line in run task future set result yield coro file home pizzi virtualenvs aiida prod lib site packages tornado gen py line in run value future result file home pizzi virtualenvs aiida prod lib site packages tornado concurrent py line in result raise exc info self exc info file line in raise exc info file home pizzi virtualenvs aiida prod lib site packages tornado gen py line in run yielded self gen throw exc info file home pizzi virtualenvs aiida prod lib site packages kiwipy rmq communicator py line in add broadcast subscriber identifier yield self message subscriber add broadcast subscriber subscriber identifier file home pizzi virtualenvs aiida prod lib site packages tornado gen py line in run value future result file home pizzi virtualenvs aiida prod lib site packages tornado concurrent py line in result raise exc info self exc info file line in raise exc info file home pizzi virtualenvs aiida prod lib site packages tornado gen py line in wrapper yielded next result file home pizzi virtualenvs aiida prod lib site packages kiwipy rmq communicator py line in add broadcast subscriber raise kiwipy duplicatesubscriberidentifier broadcast identifier format identifier kiwipy communications duplicatesubscriberidentifier broadcast identifier or traceback most recent call last file home pizzi virtualenvs aiida prod codes aiida core aiida manage external rmq py line in continue result yield super continue communicator pid nowait tag file home pizzi virtualenvs aiida prod lib site packages tornado gen py line in run value future result file home pizzi virtualenvs aiida prod lib site packages tornado concurrent py line in result raise exc info self exc info file line in raise exc info file home pizzi virtualenvs aiida prod lib site packages tornado gen py line in wrapper yielded next result file home pizzi virtualenvs aiida prod lib site packages plumpy process comms py line in continue proc saved state unbundle self load context file home pizzi virtualenvs aiida prod lib site packages plumpy persistence py line in unbundle return savable load self load context file home pizzi virtualenvs aiida prod lib site packages plumpy persistence py line in load return load cls recreate from saved state load context file home pizzi virtualenvs aiida prod lib site packages plumpy processes py line in recreate from base call with super check process init file home pizzi virtualenvs aiida prod lib site packages plumpy base utils py line in call with super check wrapped args kwargs file home pizzi virtualenvs aiida prod codes aiida core aiida engine processes process py line in init super init file home pizzi virtualenvs aiida prod lib site packages plumpy base utils py line in wrapper wrapped self args kwargs file home pizzi virtualenvs aiida prod lib site packages plumpy processes py line in init self broadcast receive identifier str self pid file home pizzi virtualenvs aiida prod lib site packages plumpy communications py line in add broadcast subscriber return self communicator add broadcast subscriber converted identifier file home pizzi virtualenvs aiida prod lib site packages kiwipy rmq communicator py line in add broadcast subscriber return self run task coro file home pizzi virtualenvs aiida prod lib site packages kiwipy rmq communicator py line in run task return self tornado to kiwi future self create task coro result timeout self task timeout file usr lib concurrent futures base py line in result return self get result file usr lib concurrent futures base py line in get result raise self exception file home pizzi virtualenvs aiida prod lib site packages kiwipy rmq communicator py line in done result done future result file home pizzi virtualenvs aiida prod lib site packages tornado concurrent py line in result raise exc info self exc info file line in raise exc info file home pizzi virtualenvs aiida prod lib site packages kiwipy futures py line in capture exceptions yield file home pizzi virtualenvs aiida prod lib site packages kiwipy rmq utils py line in run task future set result yield coro file home pizzi virtualenvs aiida prod lib site packages tornado gen py line in run value future result file home pizzi virtualenvs aiida prod lib site packages tornado concurrent py line in result raise exc info self exc info file line in raise exc info file home pizzi virtualenvs aiida prod lib site packages tornado gen py line in run yielded self gen throw exc info file home pizzi virtualenvs aiida prod lib site packages kiwipy rmq communicator py line in add broadcast subscriber identifier yield self message subscriber add broadcast subscriber subscriber identifier file home pizzi virtualenvs aiida prod lib site packages tornado gen py line in run value future result file home pizzi virtualenvs aiida prod lib site packages tornado concurrent py line in result raise exc info self exc info file line in raise exc info file home pizzi virtualenvs aiida prod lib site packages tornado gen py line in wrapper yielded next result file home pizzi virtualenvs aiida prod lib site packages kiwipy rmq communicator py line in add broadcast subscriber raise kiwipy duplicatesubscriberidentifier broadcast identifier format identifier kiwipy communications duplicatesubscriberidentifier broadcast identifier traceback most recent call last file home pizzi virtualenvs aiida prod codes aiida core aiida manage external rmq py line in continue result yield super continue communicator pid nowait tag file home pizzi virtualenvs aiida prod lib site packages tornado gen py line in run value future result file home pizzi virtualenvs aiida prod lib site packages tornado concurrent py line in result raise exc info self exc info file line in raise exc info file home pizzi virtualenvs aiida prod lib site packages tornado gen py line in run yielded self gen throw exc info file home pizzi virtualenvs aiida prod lib site packages plumpy process comms py line in continue yield proc step until terminated file home pizzi virtualenvs aiida prod lib site packages tornado gen py line in run value future result file home pizzi virtualenvs aiida prod lib site packages tornado concurrent py line in result raise exc info self exc info file line in raise exc info file home pizzi virtualenvs aiida prod lib site packages tornado gen py line in run yielded self gen throw exc info file home pizzi virtualenvs aiida prod lib site packages plumpy processes py line in step until terminated yield self step file home pizzi virtualenvs aiida prod lib site packages tornado gen py line in run value future result file home pizzi virtualenvs aiida prod lib site packages tornado concurrent py line in result raise exc info self exc info file line in raise exc info file home pizzi virtualenvs aiida prod lib site packages tornado gen py line in run yielded self gen send value file home pizzi virtualenvs aiida prod lib site packages plumpy processes py line in step self transition to next state file home pizzi virtualenvs aiida prod lib site packages plumpy base state machine py line in transition to self transition failed initial state label label sys exc info file home pizzi virtualenvs aiida prod lib site packages plumpy base state machine py line in transition failed raise exception with traceback trace file home pizzi virtualenvs aiida prod lib site packages plumpy base state machine py line in transition to self enter next state new state file home pizzi virtualenvs aiida prod lib site packages plumpy base state machine py line in enter next state self fire state event stateeventhook entered state last state file home pizzi virtualenvs aiida prod lib site packages plumpy base state machine py line in fire state event callback self hook state file home pizzi virtualenvs aiida prod lib site packages plumpy processes py line in state machine stateeventhook entered state lambda s h from state self on entered from state file home pizzi virtualenvs aiida prod codes aiida core aiida engine processes process py line in on entered super on entered from state file home pizzi virtualenvs aiida prod lib site packages plumpy processes py line in on entered self communicator broadcast send body none sender self pid subject subject file home pizzi virtualenvs aiida prod lib site packages plumpy communications py line in broadcast send return self communicator broadcast send body sender subject correlation id file home pizzi virtualenvs aiida prod lib site packages kiwipy rmq communicator py line in broadcast send correlation id correlation id file home pizzi virtualenvs aiida prod lib site packages kiwipy rmq communicator py line in send message return send future result file usr lib concurrent futures base py line in result return self get result file usr lib concurrent futures base py line in get result raise self exception file home pizzi virtualenvs aiida prod lib site packages kiwipy futures py line in capture exceptions yield file home pizzi virtualenvs aiida prod lib site packages kiwipy rmq communicator py line in do task send future set result yield coro file home pizzi virtualenvs aiida prod lib site packages tornado gen py line in run value future result file home pizzi virtualenvs aiida prod lib site packages tornado concurrent py line in result raise exc info self exc info file line in raise exc info file home pizzi virtualenvs aiida prod lib site packages tornado gen py line in run yielded self gen throw exc info file home pizzi virtualenvs aiida prod lib site packages kiwipy rmq communicator py line in broadcast send result yield self message publisher broadcast send body sender subject correlation id file home pizzi virtualenvs aiida prod lib site packages tornado gen py line in run value future result file home pizzi virtualenvs aiida prod lib site packages tornado concurrent py line in result raise exc info self exc info file line in raise exc info file home pizzi virtualenvs aiida prod lib site packages tornado gen py line in run yielded self gen throw exc info file home pizzi virtualenvs aiida prod lib site packages kiwipy rmq communicator py line in broadcast send result yield self publish message routing key defaults broadcast topic file home pizzi virtualenvs aiida prod lib site packages tornado gen py line in run value future result file home pizzi virtualenvs aiida prod lib site packages tornado concurrent py line in result raise exc info self exc info file line in raise exc info file home pizzi virtualenvs aiida prod lib site packages tornado gen py line in run yielded self gen throw exc info file home pizzi virtualenvs aiida prod lib site packages kiwipy rmq messages py line in publish message routing key routing key mandatory mandatory file home pizzi virtualenvs aiida prod lib site packages tornado gen py line in run value future result file home pizzi virtualenvs aiida prod lib site packages tornado concurrent py line in result raise exc info self exc info file line in raise exc info file home pizzi virtualenvs aiida prod lib site packages tornado gen py line in run yielded self gen throw exc info file home pizzi virtualenvs aiida prod lib site packages topika common py line in wrap raise gen return yield func self args kwargs file home pizzi virtualenvs aiida prod lib site packages tornado gen py line in run value future result file home pizzi virtualenvs aiida prod lib site packages tornado concurrent py line in result raise exc info self exc info file line in raise exc info file home pizzi virtualenvs aiida prod lib site packages tornado gen py line in run yielded self gen throw exc info file home pizzi virtualenvs aiida prod lib site packages topika exchange py line in publish mandatory mandatory file home pizzi virtualenvs aiida prod lib site packages tornado gen py line in run value future result file home pizzi virtualenvs aiida prod lib site packages tornado concurrent py line in result raise exc info self exc info file line in raise exc info file home pizzi virtualenvs aiida prod lib site packages tornado gen py line in wrapper yielded next result file home pizzi virtualenvs aiida prod lib site packages topika common py line in wrap raise runtimeerror the channel is closed runtimeerror the channel is closed i think other people observed the same the other thing i noticed not sure if a cause or a consequence is that now verdi process list gives an empty list but the daemon workers are while not busy still using a lot of memory verdi daemon status profile daemon is running as pid since active workers pid mem cpu started use verdi daemon to increase decrease the amount of workers i don t know if this is a memory leak what i can say is that i was submitting quite a lot of processes workflows and my machine was under heavy stress maybe everything became so slow that the two heartbeats were missed sphuber muhrin unkcpz chrisjsewell | 1 |
61,261 | 3,143,290,063 | IssuesEvent | 2015-09-14 05:19:30 | nim-lang/Nim | https://api.github.com/repos/nim-lang/Nim | closed | Slow lines reading (even slower than python) | Medium Priority Stdlib | I tested only naive implementations for C, Nim, Python in Linux x86_64:
```C
/* C */
#include <stdio.h>
#include <stdlib.h>
int main(int argc, char **argv)
{
char *buf = NULL;
size_t i = 0, len;
while(getline(&buf, &len, stdin) != -1) i++;
printf("%d\n", i);
free(buf);
}
```
```Python
# Python
import sys
i = 0
for line in sys.stdin:
i+=1
print i
```
```Nimrod
# Nim
var i = 0
# ugly but faster because of string copying
proc main()=
for line in stdin.lines:
inc i
echo i
main()
```
Generate test data (100MB):
```Bash
tr -c [:print:] ' ' </dev/urandom | fold -w1000 | head -100000 > test.data
```
And test like this: `./test < test.data`
Results:
* Nim (`gcc -O2`): 0.62
* Nim (`gcc -O3`): 0.31
* Python: 0.059
* C: 0.031 | 1.0 | Slow lines reading (even slower than python) - I tested only naive implementations for C, Nim, Python in Linux x86_64:
```C
/* C */
#include <stdio.h>
#include <stdlib.h>
int main(int argc, char **argv)
{
char *buf = NULL;
size_t i = 0, len;
while(getline(&buf, &len, stdin) != -1) i++;
printf("%d\n", i);
free(buf);
}
```
```Python
# Python
import sys
i = 0
for line in sys.stdin:
i+=1
print i
```
```Nimrod
# Nim
var i = 0
# ugly but faster because of string copying
proc main()=
for line in stdin.lines:
inc i
echo i
main()
```
Generate test data (100MB):
```Bash
tr -c [:print:] ' ' </dev/urandom | fold -w1000 | head -100000 > test.data
```
And test like this: `./test < test.data`
Results:
* Nim (`gcc -O2`): 0.62
* Nim (`gcc -O3`): 0.31
* Python: 0.059
* C: 0.031 | non_process | slow lines reading even slower than python i tested only naive implementations for c nim python in linux c c include include int main int argc char argv char buf null size t i len while getline buf len stdin i printf d n i free buf python python import sys i for line in sys stdin i print i nimrod nim var i ugly but faster because of string copying proc main for line in stdin lines inc i echo i main generate test data bash tr c test data and test like this test test data results nim gcc nim gcc python c | 0 |
2,090 | 4,927,361,631 | IssuesEvent | 2016-11-26 18:00:52 | mitchellh/packer | https://api.github.com/repos/mitchellh/packer | closed | Uploading Docker container to AWS ECR using shell-local post-processor | post-processor/docker question | _Latest Packer in Atlas_
I've been using packer-tag and packer-push as it suggested in [docs](https://www.packer.io/docs/builders/docker.html) to upload a Docker container built with Packer to Amazon EC2 Container Registry. It works fine but it requires me to manually update `login_password` using `aws ec2 ecr-login` every 12 hours since the token provided by amazon is temporary.
Since I can provide Packer with AWS keys there should be a way for it to obtain a new token on each execution.
I've tried to use `shell-local` instead of `packer-tag` and `packer-push` to get this working:
```
[
{
"type": "shell-local",
"inline_shebang": "/bin/sh -ex",
"inline": [
"echo 'Uploading to AWS ECR'",
"eval $(aws ecr get-login --region us-east-1)",
"docker tag {{ user `repo_name` }}/0.1:latest 123.dkr.ecr.us-east-1.amazonaws.com/{{ user `repo_name` }}/0.1:latest",
"docker push 123.dkr.ecr.us-east-1.amazonaws.com/{{ user `repo_name` }}/0.1"
]
},
```
Packer build completes with no errors. However the image is not being uploaded and there's no related output even with `PACKER_LOG=debug`. The only thing I see in logs is
```
2016/06/23 00:42:37 ui: ==> docker: Running post-processor: shell-local
==> docker: Running post-processor: shell-local
2016/06/23 00:42:37 Flagging to keep original artifact from post-processor 'shell-local'
2016/06/23 00:42:37 ui: Build 'docker' finished.
2016/06/23 00:42:37 Builds completed. Waiting on interrupt barrier...
```
Any suggestions on how to get it working?
Better way would be to keep docker-tag and docker-push post-processors but provide the output of `AWS_ACCESS_KEY_ID=123 AWS_SECRET_ACCESS_KEY=123 aws ecr get-authorization-token --output text --query "authorizationData[].authorizationToken"` to `login_password` field somehow. The question is how to do that.
| 1.0 | Uploading Docker container to AWS ECR using shell-local post-processor - _Latest Packer in Atlas_
I've been using packer-tag and packer-push as it suggested in [docs](https://www.packer.io/docs/builders/docker.html) to upload a Docker container built with Packer to Amazon EC2 Container Registry. It works fine but it requires me to manually update `login_password` using `aws ec2 ecr-login` every 12 hours since the token provided by amazon is temporary.
Since I can provide Packer with AWS keys there should be a way for it to obtain a new token on each execution.
I've tried to use `shell-local` instead of `packer-tag` and `packer-push` to get this working:
```
[
{
"type": "shell-local",
"inline_shebang": "/bin/sh -ex",
"inline": [
"echo 'Uploading to AWS ECR'",
"eval $(aws ecr get-login --region us-east-1)",
"docker tag {{ user `repo_name` }}/0.1:latest 123.dkr.ecr.us-east-1.amazonaws.com/{{ user `repo_name` }}/0.1:latest",
"docker push 123.dkr.ecr.us-east-1.amazonaws.com/{{ user `repo_name` }}/0.1"
]
},
```
Packer build completes with no errors. However the image is not being uploaded and there's no related output even with `PACKER_LOG=debug`. The only thing I see in logs is
```
2016/06/23 00:42:37 ui: ==> docker: Running post-processor: shell-local
==> docker: Running post-processor: shell-local
2016/06/23 00:42:37 Flagging to keep original artifact from post-processor 'shell-local'
2016/06/23 00:42:37 ui: Build 'docker' finished.
2016/06/23 00:42:37 Builds completed. Waiting on interrupt barrier...
```
Any suggestions on how to get it working?
Better way would be to keep docker-tag and docker-push post-processors but provide the output of `AWS_ACCESS_KEY_ID=123 AWS_SECRET_ACCESS_KEY=123 aws ecr get-authorization-token --output text --query "authorizationData[].authorizationToken"` to `login_password` field somehow. The question is how to do that.
| process | uploading docker container to aws ecr using shell local post processor latest packer in atlas i ve been using packer tag and packer push as it suggested in to upload a docker container built with packer to amazon container registry it works fine but it requires me to manually update login password using aws ecr login every hours since the token provided by amazon is temporary since i can provide packer with aws keys there should be a way for it to obtain a new token on each execution i ve tried to use shell local instead of packer tag and packer push to get this working type shell local inline shebang bin sh ex inline echo uploading to aws ecr eval aws ecr get login region us east docker tag user repo name latest dkr ecr us east amazonaws com user repo name latest docker push dkr ecr us east amazonaws com user repo name packer build completes with no errors however the image is not being uploaded and there s no related output even with packer log debug the only thing i see in logs is ui docker running post processor shell local docker running post processor shell local flagging to keep original artifact from post processor shell local ui build docker finished builds completed waiting on interrupt barrier any suggestions on how to get it working better way would be to keep docker tag and docker push post processors but provide the output of aws access key id aws secret access key aws ecr get authorization token output text query authorizationdata authorizationtoken to login password field somehow the question is how to do that | 1 |
627,283 | 19,900,415,572 | IssuesEvent | 2022-01-25 07:07:50 | microsoftgraph/microsoft-graph-explorer-v4 | https://api.github.com/repos/microsoftgraph/microsoft-graph-explorer-v4 | opened | Keyboard focus is going to top of the page after activating the 'Copy access token' icon button. | Priority 3: Normal Area: Accessibility | **Test Environment:**
OS Version: Windows 11 Build 22504.1010
Browser: Edge Dev Version 98.0.1093.6 (Official build) dev (64-bit)
URL: Graph Explorer - Microsoft Graph
Screen Reader: Narrator
**Repro Steps:**
1. Open above URL in Edge browser.
2. Login with valid credential.
3. Navigate to the 'Action token' button using tab key and activate it using enter key.
4. Now, navigate to the 'Copy access token' icon button using tab key and activate it.
5. Now navigate with tab key.
6. Observe if focus goes to the top of the page.
**Actual Result:**
Keyboard focus is going to top of the page after activating the 'Copy access token' icon button.
**Expected Result:**
Keyboard focus should be present on the 'Copy access token' button after activating the button.
**User Impact:**
Keyboard users will be impacted as they will have to extra navigate to reach the specified control.
**MAS Reference:**
[Focus Order - Liquid (microsoft.com)](https://liquid.microsoft.com/Web/Object/Read/MS.Accessibility/Requirements/02.04.03) | 1.0 | Keyboard focus is going to top of the page after activating the 'Copy access token' icon button. - **Test Environment:**
OS Version: Windows 11 Build 22504.1010
Browser: Edge Dev Version 98.0.1093.6 (Official build) dev (64-bit)
URL: Graph Explorer - Microsoft Graph
Screen Reader: Narrator
**Repro Steps:**
1. Open above URL in Edge browser.
2. Login with valid credential.
3. Navigate to the 'Action token' button using tab key and activate it using enter key.
4. Now, navigate to the 'Copy access token' icon button using tab key and activate it.
5. Now navigate with tab key.
6. Observe if focus goes to the top of the page.
**Actual Result:**
Keyboard focus is going to top of the page after activating the 'Copy access token' icon button.
**Expected Result:**
Keyboard focus should be present on the 'Copy access token' button after activating the button.
**User Impact:**
Keyboard users will be impacted as they will have to extra navigate to reach the specified control.
**MAS Reference:**
[Focus Order - Liquid (microsoft.com)](https://liquid.microsoft.com/Web/Object/Read/MS.Accessibility/Requirements/02.04.03) | non_process | keyboard focus is going to top of the page after activating the copy access token icon button test environment os version windows build browser edge dev version official build dev bit url graph explorer microsoft graph screen reader narrator repro steps open above url in edge browser login with valid credential navigate to the action token button using tab key and activate it using enter key now navigate to the copy access token icon button using tab key and activate it now navigate with tab key observe if focus goes to the top of the page actual result keyboard focus is going to top of the page after activating the copy access token icon button expected result keyboard focus should be present on the copy access token button after activating the button user impact keyboard users will be impacted as they will have to extra navigate to reach the specified control mas reference | 0 |
11,818 | 14,632,897,815 | IssuesEvent | 2020-12-23 23:55:58 | ncbo/bioportal-project | https://api.github.com/repos/ncbo/bioportal-project | opened | FDC-GDMT wrongly shown as no submissions | bug ontology processing problem | The FDC-GDMT ontology has a number of submissions, as shown below. But the Admin page thinks it has no submissions, as shown in the second screen shot.
The problem started when a new submission was uploaded 12/21. The new submission can be parsed by Protege (I'm told), but has two root concepts so does not do well as a SKOS file.

 | 1.0 | FDC-GDMT wrongly shown as no submissions - The FDC-GDMT ontology has a number of submissions, as shown below. But the Admin page thinks it has no submissions, as shown in the second screen shot.
The problem started when a new submission was uploaded 12/21. The new submission can be parsed by Protege (I'm told), but has two root concepts so does not do well as a SKOS file.

 | process | fdc gdmt wrongly shown as no submissions the fdc gdmt ontology has a number of submissions as shown below but the admin page thinks it has no submissions as shown in the second screen shot the problem started when a new submission was uploaded the new submission can be parsed by protege i m told but has two root concepts so does not do well as a skos file | 1 |
4,076 | 2,544,842,606 | IssuesEvent | 2015-01-29 13:26:18 | olga-jane/prizm | https://api.github.com/repos/olga-jane/prizm | closed | crash Trassing report for conected componentry | bug bug - crash/performance/leak Coding HIGH priority Reports | An unhandled exception of type 'System.Collections.Generic.KeyNotFoundException' occurred in mscorlib.dll
Additional information: The given key was not present in the dictionary. | 1.0 | crash Trassing report for conected componentry - An unhandled exception of type 'System.Collections.Generic.KeyNotFoundException' occurred in mscorlib.dll
Additional information: The given key was not present in the dictionary. | non_process | crash trassing report for conected componentry an unhandled exception of type system collections generic keynotfoundexception occurred in mscorlib dll additional information the given key was not present in the dictionary | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.