Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
16,775
| 10,564,614,219
|
IssuesEvent
|
2019-10-05 03:35:03
|
Azure/azure-cli
|
https://api.github.com/repos/Azure/azure-cli
|
closed
|
webapp:ISO 8601 timestamp web app setting is being converted
|
App Services Web Apps
|
**Describe the bug**
I am attempting to set an app setting to an ISO 8601 timestamp, and the value is being converted to a different format and timezone.
Example of the command and setting:
```bash
az webapp config appsettings set -n my-app \
-g my-app-rq \
--settings MAINTENANCE_END="2019-08-12T08:00:00.000-04:00"
```
Azure is saving this setting with the following value instead: `"08/12/2019 12:00:00"`
When I set the same value in the Azure Portal, the value _does not_ get converted, which is the expected behavior.
**To Reproduce**
- Run the command above
- Edit the value in the Azure Portal and observe it _not_ doing the same type of conversion
**Expected behavior**
- When I run `az webapp config appsettings set`, I expect the value (`"2019-08-12T08:00:00.000-04:00"`) to be set in Azure without being converted to a different format (`"08/12/2019 12:00:00"`).
## Environment Summary
```
Darwin-18.7.0-x86_64-i386-64bit
Python 3.7.4
Shell: bash
azure-cli 2.0.69 *
|
1.0
|
webapp:ISO 8601 timestamp web app setting is being converted - **Describe the bug**
I am attempting to set an app setting to an ISO 8601 timestamp, and the value is being converted to a different format and timezone.
Example of the command and setting:
```bash
az webapp config appsettings set -n my-app \
-g my-app-rq \
--settings MAINTENANCE_END="2019-08-12T08:00:00.000-04:00"
```
Azure is saving this setting with the following value instead: `"08/12/2019 12:00:00"`
When I set the same value in the Azure Portal, the value _does not_ get converted, which is the expected behavior.
**To Reproduce**
- Run the command above
- Edit the value in the Azure Portal and observe it _not_ doing the same type of conversion
**Expected behavior**
- When I run `az webapp config appsettings set`, I expect the value (`"2019-08-12T08:00:00.000-04:00"`) to be set in Azure without being converted to a different format (`"08/12/2019 12:00:00"`).
## Environment Summary
```
Darwin-18.7.0-x86_64-i386-64bit
Python 3.7.4
Shell: bash
azure-cli 2.0.69 *
|
non_process
|
webapp iso timestamp web app setting is being converted describe the bug i am attempting to set an app setting to an iso timestamp and the value is being converted to a different format and timezone example of the command and setting bash az webapp config appsettings set n my app g my app rq settings maintenance end azure is saving this setting with the following value instead when i set the same value in the azure portal the value does not get converted which is the expected behavior to reproduce run the command above edit the value in the azure portal and observe it not doing the same type of conversion expected behavior when i run az webapp config appsettings set i expect the value to be set in azure without being converted to a different format environment summary darwin python shell bash azure cli
| 0
|
55,705
| 11,460,350,736
|
IssuesEvent
|
2020-02-07 09:34:02
|
jOOQ/jOOQ
|
https://api.github.com/repos/jOOQ/jOOQ
|
closed
|
Apply Configuration.onError flag also to SQL errors from AbstractDatabase
|
C: Code Generation E: All Editions P: Medium R: Fixed T: Enhancement
|
The current implementation of the code generator logs and skips all SQL errors produced by `jOOQ-meta`, when reverse engineering the meta data. The reason for this is that we occasionally have bugs in our own SQL queries, which should not necessarily fail the build because the bug might be cosmetic for most users. An example is when the query producing the `CHECK` constraints fails, most users will be able to continue working with that problem, as opposed to when the query producing the columns per table fails.
However, some queries fail because of user errors, e.g. insufficient privileges to read system views and other objects. In those cases, it would sometimes be desireable to fail the build instead of skipping the functionality.
We should have a flag that allows for specifying whether any SQL errors (or other exceptions) should fail the build.
----
See also:
https://stackoverflow.com/q/60075399/521799
|
1.0
|
Apply Configuration.onError flag also to SQL errors from AbstractDatabase - The current implementation of the code generator logs and skips all SQL errors produced by `jOOQ-meta`, when reverse engineering the meta data. The reason for this is that we occasionally have bugs in our own SQL queries, which should not necessarily fail the build because the bug might be cosmetic for most users. An example is when the query producing the `CHECK` constraints fails, most users will be able to continue working with that problem, as opposed to when the query producing the columns per table fails.
However, some queries fail because of user errors, e.g. insufficient privileges to read system views and other objects. In those cases, it would sometimes be desireable to fail the build instead of skipping the functionality.
We should have a flag that allows for specifying whether any SQL errors (or other exceptions) should fail the build.
----
See also:
https://stackoverflow.com/q/60075399/521799
|
non_process
|
apply configuration onerror flag also to sql errors from abstractdatabase the current implementation of the code generator logs and skips all sql errors produced by jooq meta when reverse engineering the meta data the reason for this is that we occasionally have bugs in our own sql queries which should not necessarily fail the build because the bug might be cosmetic for most users an example is when the query producing the check constraints fails most users will be able to continue working with that problem as opposed to when the query producing the columns per table fails however some queries fail because of user errors e g insufficient privileges to read system views and other objects in those cases it would sometimes be desireable to fail the build instead of skipping the functionality we should have a flag that allows for specifying whether any sql errors or other exceptions should fail the build see also
| 0
|
19,535
| 25,848,636,377
|
IssuesEvent
|
2022-12-13 08:49:33
|
bazelbuild/bazel
|
https://api.github.com/repos/bazelbuild/bazel
|
closed
|
`-std=c++14` in `tools/bazel.rc` ignored when building external repos
|
P3 type: support / not a bug (process) team-Rules-CPP
|
### Description of the bug:
Recently gRPC bumped the version of C++ to 14 and its [tools/bazel.rc](https://github.com/grpc/grpc/blob/6fdbe6794695be853cdb87095a6b90868517f985/tools/bazel.rc#L5) got `--cxxopt='-std=c++14'` option to handle this. This worked and gRPC started having more C++14 code. But this began to have an issue when gRPC tried to upgrade protobuf to the version which started using C++14 code.
```
$ bazel build :grpc++
bazel-out/k8-opt-exec-2B5CBBC6/bin/external/com_google_protobuf/src/google/protobuf/io/_virtual_includes/printer/google/protobuf/io/printer.h:539:3:
error: 'auto' return without trailing return type; deduced return types are a C++14 extension
```
But when I built this with `BAZEL_CXXOPTS="-std=c++14"`, it worked.
```
$ export BAZEL_CXXOPTS="-std=c++14"
$ bazel build :grpc++
successfully built it.
```
If I'm doing wrong, what would be the right way to configure to use C++14 with bazel?
### What's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible.
$ git clone https://github.com/veblush/grpc.git
$ git checkout 10bbf8e50dcbed1401d4ea8ddf08eb2151cc3d5f
$ bazel build :grpc++
### Which operating system are you running Bazel on?
Linux/Debian
### What is the output of `bazel info release`?
release 5.3.0
### If `bazel info release` returns `development version` or `(@non-git)`, tell us how you built Bazel.
_No response_
### What's the output of `git remote get-url origin; git rev-parse master; git rev-parse HEAD` ?
_No response_
### Have you found anything relevant by searching the web?
_No response_
### Any other information, logs, or outputs that you want to share?
_No response_
|
1.0
|
`-std=c++14` in `tools/bazel.rc` ignored when building external repos - ### Description of the bug:
Recently gRPC bumped the version of C++ to 14 and its [tools/bazel.rc](https://github.com/grpc/grpc/blob/6fdbe6794695be853cdb87095a6b90868517f985/tools/bazel.rc#L5) got `--cxxopt='-std=c++14'` option to handle this. This worked and gRPC started having more C++14 code. But this began to have an issue when gRPC tried to upgrade protobuf to the version which started using C++14 code.
```
$ bazel build :grpc++
bazel-out/k8-opt-exec-2B5CBBC6/bin/external/com_google_protobuf/src/google/protobuf/io/_virtual_includes/printer/google/protobuf/io/printer.h:539:3:
error: 'auto' return without trailing return type; deduced return types are a C++14 extension
```
But when I built this with `BAZEL_CXXOPTS="-std=c++14"`, it worked.
```
$ export BAZEL_CXXOPTS="-std=c++14"
$ bazel build :grpc++
successfully built it.
```
If I'm doing wrong, what would be the right way to configure to use C++14 with bazel?
### What's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible.
$ git clone https://github.com/veblush/grpc.git
$ git checkout 10bbf8e50dcbed1401d4ea8ddf08eb2151cc3d5f
$ bazel build :grpc++
### Which operating system are you running Bazel on?
Linux/Debian
### What is the output of `bazel info release`?
release 5.3.0
### If `bazel info release` returns `development version` or `(@non-git)`, tell us how you built Bazel.
_No response_
### What's the output of `git remote get-url origin; git rev-parse master; git rev-parse HEAD` ?
_No response_
### Have you found anything relevant by searching the web?
_No response_
### Any other information, logs, or outputs that you want to share?
_No response_
|
process
|
std c in tools bazel rc ignored when building external repos description of the bug recently grpc bumped the version of c to and its got cxxopt std c option to handle this this worked and grpc started having more c code but this began to have an issue when grpc tried to upgrade protobuf to the version which started using c code bazel build grpc bazel out opt exec bin external com google protobuf src google protobuf io virtual includes printer google protobuf io printer h error auto return without trailing return type deduced return types are a c extension but when i built this with bazel cxxopts std c it worked export bazel cxxopts std c bazel build grpc successfully built it if i m doing wrong what would be the right way to configure to use c with bazel what s the simplest easiest way to reproduce this bug please provide a minimal example if possible git clone git checkout bazel build grpc which operating system are you running bazel on linux debian what is the output of bazel info release release if bazel info release returns development version or non git tell us how you built bazel no response what s the output of git remote get url origin git rev parse master git rev parse head no response have you found anything relevant by searching the web no response any other information logs or outputs that you want to share no response
| 1
|
235,351
| 7,737,114,207
|
IssuesEvent
|
2018-05-28 06:57:58
|
kubernetes/test-infra
|
https://api.github.com/repos/kubernetes/test-infra
|
opened
|
Publishing bot cannot push
|
kind/bug priority/important-soon
|
Our publishing bot https://github.com/k8s-publishing-bot cannot push due to a `cla/linuxfoundation` check:
```
[28 May 18 05:54 UTC]: /publish_scripts/push.sh /etc/secret-volume/token master
remote: error: GH006: Protected branch update failed for refs/heads/master.
remote: error: Required status check "cla/linuxfoundation" is expected.
To https://github.com/kubernetes/client-go
! [remote rejected] master -> master (protected branch hook declined)
error: failed to push some refs to 'https://github.com/kubernetes/client-go'
```
|
1.0
|
Publishing bot cannot push - Our publishing bot https://github.com/k8s-publishing-bot cannot push due to a `cla/linuxfoundation` check:
```
[28 May 18 05:54 UTC]: /publish_scripts/push.sh /etc/secret-volume/token master
remote: error: GH006: Protected branch update failed for refs/heads/master.
remote: error: Required status check "cla/linuxfoundation" is expected.
To https://github.com/kubernetes/client-go
! [remote rejected] master -> master (protected branch hook declined)
error: failed to push some refs to 'https://github.com/kubernetes/client-go'
```
|
non_process
|
publishing bot cannot push our publishing bot cannot push due to a cla linuxfoundation check publish scripts push sh etc secret volume token master remote error protected branch update failed for refs heads master remote error required status check cla linuxfoundation is expected to master master protected branch hook declined error failed to push some refs to
| 0
|
6,302
| 9,308,335,573
|
IssuesEvent
|
2019-03-25 14:22:04
|
brandon1roadgears/Interpreter-of-programming-language-of-Turing-Machine
|
https://api.github.com/repos/brandon1roadgears/Interpreter-of-programming-language-of-Turing-Machine
|
opened
|
Реализация третьей части программы.
|
C++ Work in process
|
## В этой части необходимо написать фрагмент вывода изменённой в ходе работы интерпретатора строки (@goldmen4ik).
***
### Думаю, тут подробности не нужны. вывести обработанную строку хотя бы через cout.
### `cout << ishodnik;`.
|
1.0
|
Реализация третьей части программы. - ## В этой части необходимо написать фрагмент вывода изменённой в ходе работы интерпретатора строки (@goldmen4ik).
***
### Думаю, тут подробности не нужны. вывести обработанную строку хотя бы через cout.
### `cout << ishodnik;`.
|
process
|
реализация третьей части программы в этой части необходимо написать фрагмент вывода изменённой в ходе работы интерпретатора строки думаю тут подробности не нужны вывести обработанную строку хотя бы через cout cout ishodnik
| 1
|
185,168
| 21,785,095,786
|
IssuesEvent
|
2022-05-14 02:28:34
|
prashantgodhwani/phonefriend
|
https://api.github.com/repos/prashantgodhwani/phonefriend
|
closed
|
WS-2021-0079 (Medium) detected in laravel/framework-v5.5.40 - autoclosed
|
security vulnerability
|
## WS-2021-0079 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>laravel/framework-v5.5.40</b></p></summary>
<p>The Laravel Framework.</p>
<p>Library home page: <a href="https://api.github.com/repos/laravel/framework/zipball/d724ce0aa61bbd9adf658215eec484f5dd6711d6">https://api.github.com/repos/laravel/framework/zipball/d724ce0aa61bbd9adf658215eec484f5dd6711d6</a></p>
<p>
Dependency Hierarchy:
- :x: **laravel/framework-v5.5.40** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Those using SQL Server with Laravel and allowing user input to be passed directly to the limit and offset functions are vulnerable to SQL injection. Other database drivers such as MySQL and Postgres are not affected by this vulnerability.
This problem has been patched on Laravel versions 6.20.26 and 8.40.0.
<p>Publish Date: 2021-04-30
<p>URL: <a href=https://github.com/advisories/GHSA-4mg9-vhxq-vm7j>WS-2021-0079</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-4mg9-vhxq-vm7j">https://github.com/advisories/GHSA-4mg9-vhxq-vm7j</a></p>
<p>Release Date: 2021-04-30</p>
<p>Fix Resolution: laravel/framework - v6.20.26,v8.40.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
WS-2021-0079 (Medium) detected in laravel/framework-v5.5.40 - autoclosed - ## WS-2021-0079 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>laravel/framework-v5.5.40</b></p></summary>
<p>The Laravel Framework.</p>
<p>Library home page: <a href="https://api.github.com/repos/laravel/framework/zipball/d724ce0aa61bbd9adf658215eec484f5dd6711d6">https://api.github.com/repos/laravel/framework/zipball/d724ce0aa61bbd9adf658215eec484f5dd6711d6</a></p>
<p>
Dependency Hierarchy:
- :x: **laravel/framework-v5.5.40** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Those using SQL Server with Laravel and allowing user input to be passed directly to the limit and offset functions are vulnerable to SQL injection. Other database drivers such as MySQL and Postgres are not affected by this vulnerability.
This problem has been patched on Laravel versions 6.20.26 and 8.40.0.
<p>Publish Date: 2021-04-30
<p>URL: <a href=https://github.com/advisories/GHSA-4mg9-vhxq-vm7j>WS-2021-0079</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-4mg9-vhxq-vm7j">https://github.com/advisories/GHSA-4mg9-vhxq-vm7j</a></p>
<p>Release Date: 2021-04-30</p>
<p>Fix Resolution: laravel/framework - v6.20.26,v8.40.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
ws medium detected in laravel framework autoclosed ws medium severity vulnerability vulnerable library laravel framework the laravel framework library home page a href dependency hierarchy x laravel framework vulnerable library vulnerability details those using sql server with laravel and allowing user input to be passed directly to the limit and offset functions are vulnerable to sql injection other database drivers such as mysql and postgres are not affected by this vulnerability this problem has been patched on laravel versions and publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution laravel framework step up your open source security game with whitesource
| 0
|
321,868
| 9,810,144,078
|
IssuesEvent
|
2019-06-12 19:45:11
|
HabitRPG/habitica-android
|
https://api.github.com/repos/HabitRPG/habitica-android
|
closed
|
Party: Switching tabs quickly causes blank members screen and crashes app
|
Help wanted Priority: minor Type: Bug Type: Crash
|
To replicate: go to a party and quickly tap between tabs. It seems to happen especially fast switching between chat and members screens. I submitted crash feedback when i did it once.
|
1.0
|
Party: Switching tabs quickly causes blank members screen and crashes app - To replicate: go to a party and quickly tap between tabs. It seems to happen especially fast switching between chat and members screens. I submitted crash feedback when i did it once.
|
non_process
|
party switching tabs quickly causes blank members screen and crashes app to replicate go to a party and quickly tap between tabs it seems to happen especially fast switching between chat and members screens i submitted crash feedback when i did it once
| 0
|
211,906
| 16,468,637,140
|
IssuesEvent
|
2021-05-23 00:13:14
|
pystatgen/sgkit
|
https://api.github.com/repos/pystatgen/sgkit
|
opened
|
Remove mention of CuPy in docs
|
documentation
|
We have two mentions of CuPy in our docs but to the best of my knowledge we don't use CuPy yet. Should we remove those mentions until we make use of CuPy?
|
1.0
|
Remove mention of CuPy in docs - We have two mentions of CuPy in our docs but to the best of my knowledge we don't use CuPy yet. Should we remove those mentions until we make use of CuPy?
|
non_process
|
remove mention of cupy in docs we have two mentions of cupy in our docs but to the best of my knowledge we don t use cupy yet should we remove those mentions until we make use of cupy
| 0
|
16,572
| 21,580,092,080
|
IssuesEvent
|
2022-05-02 17:44:47
|
microsoft/vscode
|
https://api.github.com/repos/microsoft/vscode
|
closed
|
Terminal activate conda base environment just after activating the right environment
|
bug confirmation-pending terminal-process
|
<!-- ⚠️⚠️ Do Not Delete This! bug_report_template ⚠️⚠️ -->
<!-- Please read our Rules of Conduct: https://opensource.microsoft.com/codeofconduct/ -->
<!-- 🕮 Read our guide about submitting issues: https://github.com/microsoft/vscode/wiki/Submitting-Bugs-and-Suggestions -->
<!-- 🔎 Search existing issues to avoid creating duplicates. -->
<!-- 🧪 Test using the latest Insiders build to see if your issue has already been fixed: https://code.visualstudio.com/insiders/ -->
<!-- 💡 Instead of creating your report here, use 'Report Issue' from the 'Help' menu in VS Code to pre-fill useful information. -->
<!-- 🔧 Launch with `code --disable-extensions` to check. -->
Does this issue occur when all extensions are disabled?: Yes/No
<!-- 🪓 If you answered No above, use 'Help: Start Extension Bisect' from Command Palette to try to identify the cause. -->
<!-- 📣 Issues caused by an extension need to be reported directly to the extension publisher. The 'Help > Report Issue' <!-- **NOTE**: Everything below is auto-generated; no editing required. -->
# Diagnostic data
- Python version (& distribution if applicable, e.g. Anaconda): 3.9.9
- Type of virtual environment used (e.g. conda, venv, virtualenv, etc.): Conda
- Value of the `python.languageServer` setting: Default
<details>
<summary>User Settings</summary>
<p>
```
defaultLS: {"defaultLSType":"Pylance"}
downloadLanguageServer: true
envFile: "<placeholder>"
venvPath: "<placeholder>"
venvFolders: "<placeholder>"
condaPath: "<placeholder>"
pipenvPath: "<placeholder>"
poetryPath: "<placeholder>"
languageServer: "Pylance"
linting
• enabled: true
• cwd: "<placeholder>"
• Flake8Args: "<placeholder>"
• flake8Enabled: false
• flake8Path: "<placeholder>"
• lintOnSave: true
• banditArgs: "<placeholder>"
• banditEnabled: false
• banditPath: "<placeholder>"
• mypyArgs: "<placeholder>"
• mypyEnabled: false
• mypyPath: "<placeholder>"
• pycodestyleArgs: "<placeholder>"
• pycodestyleEnabled: false
• pycodestylePath: "<placeholder>"
• prospectorArgs: "<placeholder>"
• prospectorEnabled: false
• prospectorPath: "<placeholder>"
• pydocstyleArgs: "<placeholder>"
• pydocstyleEnabled: false
• pydocstylePath: "<placeholder>"
• pylamaArgs: "<placeholder>"
• pylamaEnabled: false
• pylamaPath: "<placeholder>"
• pylintArgs: "<placeholder>"
• pylintPath: "<placeholder>"
sortImports
• args: "<placeholder>"
• path: "<placeholder>"
formatting
• autopep8Args: "<placeholder>"
• autopep8Path: "<placeholder>"
• provider: "black"
• blackArgs: "<placeholder>"
• blackPath: "<placeholder>"
• yapfArgs: "<placeholder>"
• yapfPath: "<placeholder>"
testing
• cwd: "<placeholder>"
• debugPort: 3000
• nosetestArgs: "<placeholder>"
• nosetestsEnabled: undefined
• nosetestPath: "<placeholder>"
• promptToConfigure: true
• pytestArgs: "<placeholder>"
• pytestEnabled: false
• pytestPath: "<placeholder>"
• unittestArgs: "<placeholder>"
• unittestEnabled: false
• autoTestDiscoverOnSaveEnabled: true
terminal
• activateEnvironment: true
• executeInFileDir: "<placeholder>"
• launchArgs: "<placeholder>"
experiments
• enabled: true
• optInto: []
• optOutFrom: []
tensorBoard
• logDirectory: "<placeholder>"
```
</p>
</details>
Extension version: 2022.4.0
VS Code version: Code 1.66.0 (Universal) (e18005f0f1b33c29e81d732535d8c0e47cafb0b5, 2022-03-30T05:50:04.419Z)
OS version: Darwin arm64 21.4.0
Restricted Mode: No
<details>
<summary>System Info</summary>
|Item|Value|
|---|---|
|CPUs|Apple M1 Pro (8 x 24)|
|GPU Status|2d_canvas: enabled<br>canvas_oop_rasterization: disabled_off<br>direct_rendering_display_compositor: disabled_off_ok<br>gpu_compositing: enabled<br>metal: disabled_off<br>multiple_raster_threads: enabled_on<br>oop_rasterization: enabled<br>opengl: enabled_on<br>rasterization: enabled<br>raw_draw: disabled_off_ok<br>skia_renderer: enabled_on<br>video_decode: enabled<br>video_encode: enabled<br>webgl: enabled<br>webgl2: enabled|
|Load (avg)|5, 5, 5|
|Memory (System)|16.00GB (0.08GB free)|
|Process Argv|--crash-reporter-id a129d0b2-ecd6-4f0a-8ff3-d67e5b6507fa|
|Screen Reader|yes|
|VM|0%|
</details><details>
<summary>A/B Experiments</summary>
```
vsliv368:30146709
vsreu685:30147344
python383cf:30185419
vspor879:30202332
vspor708:30202333
vspor363:30204092
pythonvspyl392:30443607
pythontb:30283811
pythonvspyt551cf:30345471
pythonptprofiler:30281270
vshan820:30294714
vstes263:30335439
vscoreces:30445986
pythondataviewer:30285071
vscod805cf:30301675
pythonvspyt200:30340761
binariesv615:30325510
bridge0708:30335490
bridge0723:30353136
vsaa593:30376534
testflight:30433670
vsc1dst:30438360
pythonvs932:30410667
wslgetstarted:30449410
pythonvsnew555:30457759
vscscmwlcmt:30465135
cppdebug:30466689
pynewfile477:30463512
```
</details>
<!-- generated by issue reporter -->
# Behaviour
## Expected vs. Actual
Hello,
When I select my conda env for the workspace, I expect that each terminal session activate this environment. In the past it worked well. But since a few days I have a strange behaviour : the terminal execute the command `conda activate myenv` then just after `source /Users/luis/miniforge3/bin/activate` reactivating base env. (see in screenshoot)
## Steps to reproduce:

|
1.0
|
Terminal activate conda base environment just after activating the right environment - <!-- ⚠️⚠️ Do Not Delete This! bug_report_template ⚠️⚠️ -->
<!-- Please read our Rules of Conduct: https://opensource.microsoft.com/codeofconduct/ -->
<!-- 🕮 Read our guide about submitting issues: https://github.com/microsoft/vscode/wiki/Submitting-Bugs-and-Suggestions -->
<!-- 🔎 Search existing issues to avoid creating duplicates. -->
<!-- 🧪 Test using the latest Insiders build to see if your issue has already been fixed: https://code.visualstudio.com/insiders/ -->
<!-- 💡 Instead of creating your report here, use 'Report Issue' from the 'Help' menu in VS Code to pre-fill useful information. -->
<!-- 🔧 Launch with `code --disable-extensions` to check. -->
Does this issue occur when all extensions are disabled?: Yes/No
<!-- 🪓 If you answered No above, use 'Help: Start Extension Bisect' from Command Palette to try to identify the cause. -->
<!-- 📣 Issues caused by an extension need to be reported directly to the extension publisher. The 'Help > Report Issue' <!-- **NOTE**: Everything below is auto-generated; no editing required. -->
# Diagnostic data
- Python version (& distribution if applicable, e.g. Anaconda): 3.9.9
- Type of virtual environment used (e.g. conda, venv, virtualenv, etc.): Conda
- Value of the `python.languageServer` setting: Default
<details>
<summary>User Settings</summary>
<p>
```
defaultLS: {"defaultLSType":"Pylance"}
downloadLanguageServer: true
envFile: "<placeholder>"
venvPath: "<placeholder>"
venvFolders: "<placeholder>"
condaPath: "<placeholder>"
pipenvPath: "<placeholder>"
poetryPath: "<placeholder>"
languageServer: "Pylance"
linting
• enabled: true
• cwd: "<placeholder>"
• Flake8Args: "<placeholder>"
• flake8Enabled: false
• flake8Path: "<placeholder>"
• lintOnSave: true
• banditArgs: "<placeholder>"
• banditEnabled: false
• banditPath: "<placeholder>"
• mypyArgs: "<placeholder>"
• mypyEnabled: false
• mypyPath: "<placeholder>"
• pycodestyleArgs: "<placeholder>"
• pycodestyleEnabled: false
• pycodestylePath: "<placeholder>"
• prospectorArgs: "<placeholder>"
• prospectorEnabled: false
• prospectorPath: "<placeholder>"
• pydocstyleArgs: "<placeholder>"
• pydocstyleEnabled: false
• pydocstylePath: "<placeholder>"
• pylamaArgs: "<placeholder>"
• pylamaEnabled: false
• pylamaPath: "<placeholder>"
• pylintArgs: "<placeholder>"
• pylintPath: "<placeholder>"
sortImports
• args: "<placeholder>"
• path: "<placeholder>"
formatting
• autopep8Args: "<placeholder>"
• autopep8Path: "<placeholder>"
• provider: "black"
• blackArgs: "<placeholder>"
• blackPath: "<placeholder>"
• yapfArgs: "<placeholder>"
• yapfPath: "<placeholder>"
testing
• cwd: "<placeholder>"
• debugPort: 3000
• nosetestArgs: "<placeholder>"
• nosetestsEnabled: undefined
• nosetestPath: "<placeholder>"
• promptToConfigure: true
• pytestArgs: "<placeholder>"
• pytestEnabled: false
• pytestPath: "<placeholder>"
• unittestArgs: "<placeholder>"
• unittestEnabled: false
• autoTestDiscoverOnSaveEnabled: true
terminal
• activateEnvironment: true
• executeInFileDir: "<placeholder>"
• launchArgs: "<placeholder>"
experiments
• enabled: true
• optInto: []
• optOutFrom: []
tensorBoard
• logDirectory: "<placeholder>"
```
</p>
</details>
Extension version: 2022.4.0
VS Code version: Code 1.66.0 (Universal) (e18005f0f1b33c29e81d732535d8c0e47cafb0b5, 2022-03-30T05:50:04.419Z)
OS version: Darwin arm64 21.4.0
Restricted Mode: No
<details>
<summary>System Info</summary>
|Item|Value|
|---|---|
|CPUs|Apple M1 Pro (8 x 24)|
|GPU Status|2d_canvas: enabled<br>canvas_oop_rasterization: disabled_off<br>direct_rendering_display_compositor: disabled_off_ok<br>gpu_compositing: enabled<br>metal: disabled_off<br>multiple_raster_threads: enabled_on<br>oop_rasterization: enabled<br>opengl: enabled_on<br>rasterization: enabled<br>raw_draw: disabled_off_ok<br>skia_renderer: enabled_on<br>video_decode: enabled<br>video_encode: enabled<br>webgl: enabled<br>webgl2: enabled|
|Load (avg)|5, 5, 5|
|Memory (System)|16.00GB (0.08GB free)|
|Process Argv|--crash-reporter-id a129d0b2-ecd6-4f0a-8ff3-d67e5b6507fa|
|Screen Reader|yes|
|VM|0%|
</details><details>
<summary>A/B Experiments</summary>
```
vsliv368:30146709
vsreu685:30147344
python383cf:30185419
vspor879:30202332
vspor708:30202333
vspor363:30204092
pythonvspyl392:30443607
pythontb:30283811
pythonvspyt551cf:30345471
pythonptprofiler:30281270
vshan820:30294714
vstes263:30335439
vscoreces:30445986
pythondataviewer:30285071
vscod805cf:30301675
pythonvspyt200:30340761
binariesv615:30325510
bridge0708:30335490
bridge0723:30353136
vsaa593:30376534
testflight:30433670
vsc1dst:30438360
pythonvs932:30410667
wslgetstarted:30449410
pythonvsnew555:30457759
vscscmwlcmt:30465135
cppdebug:30466689
pynewfile477:30463512
```
</details>
<!-- generated by issue reporter -->
# Behaviour
## Expected vs. Actual
Hello,
When I select my conda env for the workspace, I expect that each terminal session activate this environment. In the past it worked well. But since a few days I have a strange behaviour : the terminal execute the command `conda activate myenv` then just after `source /Users/luis/miniforge3/bin/activate` reactivating base env. (see in screenshoot)
## Steps to reproduce:

|
process
|
terminal activate conda base environment just after activating the right environment does this issue occur when all extensions are disabled yes no report issue diagnostic data python version distribution if applicable e g anaconda type of virtual environment used e g conda venv virtualenv etc conda value of the python languageserver setting default user settings defaultls defaultlstype pylance downloadlanguageserver true envfile venvpath venvfolders condapath pipenvpath poetrypath languageserver pylance linting • enabled true • cwd • • false • • lintonsave true • banditargs • banditenabled false • banditpath • mypyargs • mypyenabled false • mypypath • pycodestyleargs • pycodestyleenabled false • pycodestylepath • prospectorargs • prospectorenabled false • prospectorpath • pydocstyleargs • pydocstyleenabled false • pydocstylepath • pylamaargs • pylamaenabled false • pylamapath • pylintargs • pylintpath sortimports • args • path formatting • • • provider black • blackargs • blackpath • yapfargs • yapfpath testing • cwd • debugport • nosetestargs • nosetestsenabled undefined • nosetestpath • prompttoconfigure true • pytestargs • pytestenabled false • pytestpath • unittestargs • unittestenabled false • autotestdiscoveronsaveenabled true terminal • activateenvironment true • executeinfiledir • launchargs experiments • enabled true • optinto • optoutfrom tensorboard • logdirectory extension version vs code version code universal os version darwin restricted mode no system info item value cpus apple pro x gpu status canvas enabled canvas oop rasterization disabled off direct rendering display compositor disabled off ok gpu compositing enabled metal disabled off multiple raster threads enabled on oop rasterization enabled opengl enabled on rasterization enabled raw draw disabled off ok skia renderer enabled on video decode enabled video encode enabled webgl enabled enabled load avg memory system free process argv crash reporter id screen reader yes vm a b experiments pythontb pythonptprofiler vscoreces pythondataviewer testflight wslgetstarted vscscmwlcmt cppdebug behaviour expected vs actual hello when i select my conda env for the workspace i expect that each terminal session activate this environment in the past it worked well but since a few days i have a strange behaviour the terminal execute the command conda activate myenv then just after source users luis bin activate reactivating base env see in screenshoot steps to reproduce
| 1
|
11,714
| 14,546,541,572
|
IssuesEvent
|
2020-12-15 21:24:17
|
MicrosoftDocs/azure-devops-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
|
closed
|
No Support for Generic Environment Resources
|
Pri2 devops-cicd-process/tech devops/prod product-feedback
|
I see that Azure DevOps yaml supports Kubernetes and Virtual Machines for deployment tracability in the form of "resources". Is there a reason there doesnt just exists a "generic" resource that is created if the resource is referenced through yml but doesnt exists yet? The same as environments are managed. That way jobs and commits will be recorded by resource based on the name of the resources.
Example:
`
environment: environment_name.resource_name
`
And the environment **environment_name** is created and the "generic" resource is created **resource_name**
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 77d95db6-9983-7346-d0eb-4b7443e4e252
* Version Independent ID: 0a22cccc-318d-592f-d1ab-09ec01d88087
* Content: [Environment - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/environments?view=azure-devops)
* Content Source: [docs/pipelines/process/environments.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/environments.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
1.0
|
No Support for Generic Environment Resources -
I see that Azure DevOps yaml supports Kubernetes and Virtual Machines for deployment tracability in the form of "resources". Is there a reason there doesnt just exists a "generic" resource that is created if the resource is referenced through yml but doesnt exists yet? The same as environments are managed. That way jobs and commits will be recorded by resource based on the name of the resources.
Example:
`
environment: environment_name.resource_name
`
And the environment **environment_name** is created and the "generic" resource is created **resource_name**
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 77d95db6-9983-7346-d0eb-4b7443e4e252
* Version Independent ID: 0a22cccc-318d-592f-d1ab-09ec01d88087
* Content: [Environment - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/environments?view=azure-devops)
* Content Source: [docs/pipelines/process/environments.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/environments.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
process
|
no support for generic environment resources i see that azure devops yaml supports kubernetes and virtual machines for deployment tracability in the form of resources is there a reason there doesnt just exists a generic resource that is created if the resource is referenced through yml but doesnt exists yet the same as environments are managed that way jobs and commits will be recorded by resource based on the name of the resources example environment environment name resource name and the environment environment name is created and the generic resource is created resource name document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
| 1
|
4,508
| 7,353,507,171
|
IssuesEvent
|
2018-03-09 01:05:20
|
dotnet/corefx
|
https://api.github.com/repos/dotnet/corefx
|
closed
|
ServiceProcess test failures, including PauseAndContinue
|
area-System.ServiceProcess disabled-test os-windows tenet-reliability test bug test-run-core
|
This has failed the last three builds (2/11 - 2/12) - [MC test results link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180212.05/workItem/System.ServiceProcess.ServiceController.Tests/analysis/xunit/System.ServiceProcess.Tests.ServiceControllerTests~2FPauseAndContinue)
The test `System.ServiceProcess.Tests.ServiceControllerTests/PauseAndContinue` has failed.
System.ComponentModel.Win32Exception : Could not find service
Stack Trace:
at System.ServiceProcess.Tests.TestServiceInstaller.DeleteService() in E:\A\_work\1424\s\corefx\src\System.ServiceProcess.ServiceController\tests\System.ServiceProcess.ServiceController.TestService\TestServiceInstaller.cs:line 186
at System.ServiceProcess.Tests.TestServiceInstaller.RemoveService() in E:\A\_work\1424\s\corefx\src\System.ServiceProcess.ServiceController\tests\System.ServiceProcess.ServiceController.TestService\TestServiceInstaller.cs:line 133
at System.ServiceProcess.Tests.TestServiceProvider.DeleteTestServices() in E:\A\_work\1424\s\corefx\src\System.ServiceProcess.ServiceController\tests\TestServiceProvider.cs:line 139
at System.ServiceProcess.Tests.TestServiceProvider.DeleteTestServices() in E:\A\_work\1424\s\corefx\src\System.ServiceProcess.ServiceController\tests\TestServiceProvider.cs:line 149
at System.ServiceProcess.Tests.ServiceControllerTests.Dispose() in E:\A\_work\1424\s\corefx\src\System.ServiceProcess.ServiceController\tests\ServiceControllerTests.cs:line 184
at ReflectionAbstractionExtensions.DisposeTestClass(ITest test, Object testClass, IMessageBus messageBus, ExecutionTimer timer, CancellationTokenSource cancellationTokenSource) in C:\BuildAgent\work\cb37e9acf085d108\src\xunit.execution\Extensions\ReflectionAbstractionExtensions.cs:line 76
Build : Master - 20180212.05 (Core Tests)
Failing configurations:
- Windows.10.Amd64-x64
- Release
|
1.0
|
ServiceProcess test failures, including PauseAndContinue - This has failed the last three builds (2/11 - 2/12) - [MC test results link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180212.05/workItem/System.ServiceProcess.ServiceController.Tests/analysis/xunit/System.ServiceProcess.Tests.ServiceControllerTests~2FPauseAndContinue)
The test `System.ServiceProcess.Tests.ServiceControllerTests/PauseAndContinue` has failed.
System.ComponentModel.Win32Exception : Could not find service
Stack Trace:
at System.ServiceProcess.Tests.TestServiceInstaller.DeleteService() in E:\A\_work\1424\s\corefx\src\System.ServiceProcess.ServiceController\tests\System.ServiceProcess.ServiceController.TestService\TestServiceInstaller.cs:line 186
at System.ServiceProcess.Tests.TestServiceInstaller.RemoveService() in E:\A\_work\1424\s\corefx\src\System.ServiceProcess.ServiceController\tests\System.ServiceProcess.ServiceController.TestService\TestServiceInstaller.cs:line 133
at System.ServiceProcess.Tests.TestServiceProvider.DeleteTestServices() in E:\A\_work\1424\s\corefx\src\System.ServiceProcess.ServiceController\tests\TestServiceProvider.cs:line 139
at System.ServiceProcess.Tests.TestServiceProvider.DeleteTestServices() in E:\A\_work\1424\s\corefx\src\System.ServiceProcess.ServiceController\tests\TestServiceProvider.cs:line 149
at System.ServiceProcess.Tests.ServiceControllerTests.Dispose() in E:\A\_work\1424\s\corefx\src\System.ServiceProcess.ServiceController\tests\ServiceControllerTests.cs:line 184
at ReflectionAbstractionExtensions.DisposeTestClass(ITest test, Object testClass, IMessageBus messageBus, ExecutionTimer timer, CancellationTokenSource cancellationTokenSource) in C:\BuildAgent\work\cb37e9acf085d108\src\xunit.execution\Extensions\ReflectionAbstractionExtensions.cs:line 76
Build : Master - 20180212.05 (Core Tests)
Failing configurations:
- Windows.10.Amd64-x64
- Release
|
process
|
serviceprocess test failures including pauseandcontinue this has failed the last three builds the test system serviceprocess tests servicecontrollertests pauseandcontinue has failed system componentmodel could not find service stack trace at system serviceprocess tests testserviceinstaller deleteservice in e a work s corefx src system serviceprocess servicecontroller tests system serviceprocess servicecontroller testservice testserviceinstaller cs line at system serviceprocess tests testserviceinstaller removeservice in e a work s corefx src system serviceprocess servicecontroller tests system serviceprocess servicecontroller testservice testserviceinstaller cs line at system serviceprocess tests testserviceprovider deletetestservices in e a work s corefx src system serviceprocess servicecontroller tests testserviceprovider cs line at system serviceprocess tests testserviceprovider deletetestservices in e a work s corefx src system serviceprocess servicecontroller tests testserviceprovider cs line at system serviceprocess tests servicecontrollertests dispose in e a work s corefx src system serviceprocess servicecontroller tests servicecontrollertests cs line at reflectionabstractionextensions disposetestclass itest test object testclass imessagebus messagebus executiontimer timer cancellationtokensource cancellationtokensource in c buildagent work src xunit execution extensions reflectionabstractionextensions cs line build master core tests failing configurations windows release
| 1
|
161,797
| 20,155,364,456
|
IssuesEvent
|
2022-02-09 16:00:01
|
jgeraigery/experian-java
|
https://api.github.com/repos/jgeraigery/experian-java
|
closed
|
CVE-2019-14892 (High) detected in jackson-databind-2.9.2.jar - autoclosed
|
security vulnerability
|
## CVE-2019-14892 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.2.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: experian-java/MavenWorkspace/bis-services-lib/bis-services-base/pom.xml</p>
<p>Path to vulnerable library: ository/com/fasterxml/jackson/core/jackson-databind/2.9.2/jackson-databind-2.9.2.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.2.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/jgeraigery/experian-java/commit/9c836fb688a6591b590fe6154680b73ceed58629">9c836fb688a6591b590fe6154680b73ceed58629</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A flaw was discovered in jackson-databind in versions before 2.9.10, 2.8.11.5 and 2.6.7.3, where it would permit polymorphic deserialization of a malicious object using commons-configuration 1 and 2 JNDI classes. An attacker could use this flaw to execute arbitrary code.
<p>Publish Date: 2020-03-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14892>CVE-2019-14892</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2462">https://github.com/FasterXML/jackson-databind/issues/2462</a></p>
<p>Release Date: 2020-09-04</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.6.7.3,2.7.9.7,2.8.11.5,2.9.10</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.2","packageFilePaths":["/MavenWorkspace/bis-services-lib/bis-services-base/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.6.7.3,2.7.9.7,2.8.11.5,2.9.10","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-14892","vulnerabilityDetails":"A flaw was discovered in jackson-databind in versions before 2.9.10, 2.8.11.5 and 2.6.7.3, where it would permit polymorphic deserialization of a malicious object using commons-configuration 1 and 2 JNDI classes. An attacker could use this flaw to execute arbitrary code.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14892","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2019-14892 (High) detected in jackson-databind-2.9.2.jar - autoclosed - ## CVE-2019-14892 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.2.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: experian-java/MavenWorkspace/bis-services-lib/bis-services-base/pom.xml</p>
<p>Path to vulnerable library: ository/com/fasterxml/jackson/core/jackson-databind/2.9.2/jackson-databind-2.9.2.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.2.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/jgeraigery/experian-java/commit/9c836fb688a6591b590fe6154680b73ceed58629">9c836fb688a6591b590fe6154680b73ceed58629</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A flaw was discovered in jackson-databind in versions before 2.9.10, 2.8.11.5 and 2.6.7.3, where it would permit polymorphic deserialization of a malicious object using commons-configuration 1 and 2 JNDI classes. An attacker could use this flaw to execute arbitrary code.
<p>Publish Date: 2020-03-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14892>CVE-2019-14892</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2462">https://github.com/FasterXML/jackson-databind/issues/2462</a></p>
<p>Release Date: 2020-09-04</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.6.7.3,2.7.9.7,2.8.11.5,2.9.10</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.2","packageFilePaths":["/MavenWorkspace/bis-services-lib/bis-services-base/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.6.7.3,2.7.9.7,2.8.11.5,2.9.10","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-14892","vulnerabilityDetails":"A flaw was discovered in jackson-databind in versions before 2.9.10, 2.8.11.5 and 2.6.7.3, where it would permit polymorphic deserialization of a malicious object using commons-configuration 1 and 2 JNDI classes. An attacker could use this flaw to execute arbitrary code.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14892","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
non_process
|
cve high detected in jackson databind jar autoclosed cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file experian java mavenworkspace bis services lib bis services base pom xml path to vulnerable library ository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details a flaw was discovered in jackson databind in versions before and where it would permit polymorphic deserialization of a malicious object using commons configuration and jndi classes an attacker could use this flaw to execute arbitrary code publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails a flaw was discovered in jackson databind in versions before and where it would permit polymorphic deserialization of a malicious object using commons configuration and jndi classes an attacker could use this flaw to execute arbitrary code vulnerabilityurl
| 0
|
2,465
| 5,242,967,585
|
IssuesEvent
|
2017-01-31 19:26:48
|
opentrials/opentrials
|
https://api.github.com/repos/opentrials/opentrials
|
closed
|
Add records.source_url_type like field
|
API Explorer Processors refactoring
|
Our `source_url` could be human-readable url or API endpoint or link to secure space. So we need some instruction to frontend added on processors level how to show it to the user.
|
1.0
|
Add records.source_url_type like field - Our `source_url` could be human-readable url or API endpoint or link to secure space. So we need some instruction to frontend added on processors level how to show it to the user.
|
process
|
add records source url type like field our source url could be human readable url or api endpoint or link to secure space so we need some instruction to frontend added on processors level how to show it to the user
| 1
|
832,178
| 32,075,345,508
|
IssuesEvent
|
2023-09-25 10:37:08
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
www.nytimes.com - see bug description
|
priority-critical browser-fenix engine-gecko
|
<!-- @browser: Firefox Mobile 118.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 13; Mobile; rv:109.0) Gecko/118.0 Firefox/118.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/127499 -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://www.nytimes.com/2023/09/23/world/asia/china-sea-philippines-us.html
**Browser / Version**: Firefox Mobile 118.0
**Operating System**: Android 13
**Tested Another Browser**: Yes Chrome
**Problem type**: Something else
**Description**: page doesn't render completely. scrolling down leads to complete whitespaces and it takes multiple many as l refreshes to view
**Steps to Reproduce**:
Didn't do anything unusual before I visited the page
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>buildID: 20230914180032</li><li>channel: beta</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2023/9/dc01d5be-7683-4308-92a4-87141e63bfe2)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
www.nytimes.com - see bug description - <!-- @browser: Firefox Mobile 118.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 13; Mobile; rv:109.0) Gecko/118.0 Firefox/118.0 -->
<!-- @reported_with: android-components-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/127499 -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://www.nytimes.com/2023/09/23/world/asia/china-sea-philippines-us.html
**Browser / Version**: Firefox Mobile 118.0
**Operating System**: Android 13
**Tested Another Browser**: Yes Chrome
**Problem type**: Something else
**Description**: page doesn't render completely. scrolling down leads to complete whitespaces and it takes multiple many as l refreshes to view
**Steps to Reproduce**:
Didn't do anything unusual before I visited the page
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>buildID: 20230914180032</li><li>channel: beta</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2023/9/dc01d5be-7683-4308-92a4-87141e63bfe2)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_process
|
see bug description url browser version firefox mobile operating system android tested another browser yes chrome problem type something else description page doesn t render completely scrolling down leads to complete whitespaces and it takes multiple many as l refreshes to view steps to reproduce didn t do anything unusual before i visited the page browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false buildid channel beta hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
| 0
|
11,919
| 14,702,541,081
|
IssuesEvent
|
2021-01-04 13:45:55
|
modi-w/AutoVersionsDB
|
https://api.github.com/repos/modi-w/AutoVersionsDB
|
opened
|
Add a numeric identification for the filenames to allow ordering
|
area-Core process-ready-for-implementation type-enhancement
|
**The Problem**
The repeatable scripts files list and the DevDummyData scripts files are sorted by alphanumeric order, we don't have any chronological or logical order for those files.
This cause the following problems:
1. When the user adds a new script file, the file is not always added to the end of the list, so sometimes it is not so clear where is the new file in the list.
2. Sometimes, it is important the order of those files execution, we cant do that right now unless the user gave appropriate names to those script files.
**Solution**
Add a numeric identification for the filenames to allow ordering.
**Action Items:**
**Updates**
|
1.0
|
Add a numeric identification for the filenames to allow ordering - **The Problem**
The repeatable scripts files list and the DevDummyData scripts files are sorted by alphanumeric order, we don't have any chronological or logical order for those files.
This cause the following problems:
1. When the user adds a new script file, the file is not always added to the end of the list, so sometimes it is not so clear where is the new file in the list.
2. Sometimes, it is important the order of those files execution, we cant do that right now unless the user gave appropriate names to those script files.
**Solution**
Add a numeric identification for the filenames to allow ordering.
**Action Items:**
**Updates**
|
process
|
add a numeric identification for the filenames to allow ordering the problem the repeatable scripts files list and the devdummydata scripts files are sorted by alphanumeric order we don t have any chronological or logical order for those files this cause the following problems when the user adds a new script file the file is not always added to the end of the list so sometimes it is not so clear where is the new file in the list sometimes it is important the order of those files execution we cant do that right now unless the user gave appropriate names to those script files solution add a numeric identification for the filenames to allow ordering action items updates
| 1
|
7,658
| 10,743,040,593
|
IssuesEvent
|
2019-10-30 00:35:23
|
HouraiTeahouse/FantasyCrescendo
|
https://api.github.com/repos/HouraiTeahouse/FantasyCrescendo
|
closed
|
Migrate submodule dependencies into Unity Package Manager format to be fetched remotely
|
Category:Game Engine Priority:0 Severity:0 Status:Assigned Type:Experimental Type:Process
|
The new Unity Package Manager should be able to support downloading remote 3rd party packages sometime in the near future (2018/2019), we should migrate the dependencies we currently have held as git submodules into their own repos to be fetched by the package manager. This should offer an easier method of dependency management and lower the barrier for new contributors to join the project.
|
1.0
|
Migrate submodule dependencies into Unity Package Manager format to be fetched remotely - The new Unity Package Manager should be able to support downloading remote 3rd party packages sometime in the near future (2018/2019), we should migrate the dependencies we currently have held as git submodules into their own repos to be fetched by the package manager. This should offer an easier method of dependency management and lower the barrier for new contributors to join the project.
|
process
|
migrate submodule dependencies into unity package manager format to be fetched remotely the new unity package manager should be able to support downloading remote party packages sometime in the near future we should migrate the dependencies we currently have held as git submodules into their own repos to be fetched by the package manager this should offer an easier method of dependency management and lower the barrier for new contributors to join the project
| 1
|
10,108
| 13,044,162,160
|
IssuesEvent
|
2020-07-29 03:47:30
|
tikv/tikv
|
https://api.github.com/repos/tikv/tikv
|
closed
|
UCP: Migrate scalar function `AddDatetimeAndDuration` from TiDB
|
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
|
## Description
Port the scalar function `AddDatetimeAndDuration` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @sticnarf
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
2.0
|
UCP: Migrate scalar function `AddDatetimeAndDuration` from TiDB -
## Description
Port the scalar function `AddDatetimeAndDuration` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @sticnarf
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
process
|
ucp migrate scalar function adddatetimeandduration from tidb description port the scalar function adddatetimeandduration from tidb to coprocessor score mentor s sticnarf recommended skills rust programming learning materials already implemented expressions ported from tidb
| 1
|
81,533
| 23,487,602,541
|
IssuesEvent
|
2022-08-17 15:35:47
|
microsoft/PowerToys
|
https://api.github.com/repos/microsoft/PowerToys
|
closed
|
ARM64 verification script for CI
|
Help Wanted Status-In progress Area-Build Area-ARM64
|
### Description of the new feature / enhancement
Let's have a CI verification script for ARM64.
Thoughts CSProj and SLN. Props files maybe?
### Scenario when this would be used?
ARM support being sure we are consistant
### Supporting information
_No response_
|
1.0
|
ARM64 verification script for CI - ### Description of the new feature / enhancement
Let's have a CI verification script for ARM64.
Thoughts CSProj and SLN. Props files maybe?
### Scenario when this would be used?
ARM support being sure we are consistant
### Supporting information
_No response_
|
non_process
|
verification script for ci description of the new feature enhancement let s have a ci verification script for thoughts csproj and sln props files maybe scenario when this would be used arm support being sure we are consistant supporting information no response
| 0
|
18,702
| 24,596,205,283
|
IssuesEvent
|
2022-10-14 08:33:16
|
prisma/prisma
|
https://api.github.com/repos/prisma/prisma
|
opened
|
Error: [libs/dml/src/datamodel.rs:178:14] Every RelationInfo should have a complementary RelationInfo on the opposite relation field.
|
bug/1-unconfirmed kind/bug process/candidate topic: error reporting team/schema
|
<!-- If required, please update the title to be clear and descriptive -->
Command: `prisma db pull`
Version: `4.4.0`
Binary Version: `f352a33b70356f46311da8b00d83386dd9f145d6`
Report: https://prisma-errors.netlify.app/report/14372
OS: `arm64 darwin 21.5.0`
|
1.0
|
Error: [libs/dml/src/datamodel.rs:178:14] Every RelationInfo should have a complementary RelationInfo on the opposite relation field. - <!-- If required, please update the title to be clear and descriptive -->
Command: `prisma db pull`
Version: `4.4.0`
Binary Version: `f352a33b70356f46311da8b00d83386dd9f145d6`
Report: https://prisma-errors.netlify.app/report/14372
OS: `arm64 darwin 21.5.0`
|
process
|
error every relationinfo should have a complementary relationinfo on the opposite relation field command prisma db pull version binary version report os darwin
| 1
|
154,218
| 19,710,927,730
|
IssuesEvent
|
2022-01-13 05:10:40
|
ChoeMinji/redis-6.2.3
|
https://api.github.com/repos/ChoeMinji/redis-6.2.3
|
opened
|
CVE-2021-32762 (High) detected in redis6.2.6
|
security vulnerability
|
## CVE-2021-32762 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>redis6.2.6</b></p></summary>
<p>
<p>Redis is an in-memory database that persists on disk. The data model is key-value, but many different kind of values are supported: Strings, Lists, Sets, Sorted Sets, Hashes, Streams, HyperLogLogs, Bitmaps.</p>
<p>Library home page: <a href=https://github.com/redis/redis.git>https://github.com/redis/redis.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/ChoeMinji/redis-6.2.3/commit/5b7aad699eab111b003a1a58082434041ce3e008">5b7aad699eab111b003a1a58082434041ce3e008</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/deps/hiredis/hiredis.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Redis is an open source, in-memory database that persists on disk. The redis-cli command line tool and redis-sentinel service may be vulnerable to integer overflow when parsing specially crafted large multi-bulk network replies. This is a result of a vulnerability in the underlying hiredis library which does not perform an overflow check before calling the calloc() heap allocation function. This issue only impacts systems with heap allocators that do not perform their own overflow checks. Most modern systems do and are therefore not likely to be affected. Furthermore, by default redis-sentinel uses the jemalloc allocator which is also not vulnerable. The problem is fixed in Redis versions 6.2.6, 6.0.16 and 5.0.14.
<p>Publish Date: 2021-10-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32762>CVE-2021-32762</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/redis/redis/security/advisories/GHSA-833w-8v3m-8wwr">https://github.com/redis/redis/security/advisories/GHSA-833w-8v3m-8wwr</a></p>
<p>Release Date: 2021-10-04</p>
<p>Fix Resolution: 5.0.14, 6.0.16, 6.2.6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-32762 (High) detected in redis6.2.6 - ## CVE-2021-32762 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>redis6.2.6</b></p></summary>
<p>
<p>Redis is an in-memory database that persists on disk. The data model is key-value, but many different kind of values are supported: Strings, Lists, Sets, Sorted Sets, Hashes, Streams, HyperLogLogs, Bitmaps.</p>
<p>Library home page: <a href=https://github.com/redis/redis.git>https://github.com/redis/redis.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/ChoeMinji/redis-6.2.3/commit/5b7aad699eab111b003a1a58082434041ce3e008">5b7aad699eab111b003a1a58082434041ce3e008</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/deps/hiredis/hiredis.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Redis is an open source, in-memory database that persists on disk. The redis-cli command line tool and redis-sentinel service may be vulnerable to integer overflow when parsing specially crafted large multi-bulk network replies. This is a result of a vulnerability in the underlying hiredis library which does not perform an overflow check before calling the calloc() heap allocation function. This issue only impacts systems with heap allocators that do not perform their own overflow checks. Most modern systems do and are therefore not likely to be affected. Furthermore, by default redis-sentinel uses the jemalloc allocator which is also not vulnerable. The problem is fixed in Redis versions 6.2.6, 6.0.16 and 5.0.14.
<p>Publish Date: 2021-10-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32762>CVE-2021-32762</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/redis/redis/security/advisories/GHSA-833w-8v3m-8wwr">https://github.com/redis/redis/security/advisories/GHSA-833w-8v3m-8wwr</a></p>
<p>Release Date: 2021-10-04</p>
<p>Fix Resolution: 5.0.14, 6.0.16, 6.2.6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in cve high severity vulnerability vulnerable library redis is an in memory database that persists on disk the data model is key value but many different kind of values are supported strings lists sets sorted sets hashes streams hyperloglogs bitmaps library home page a href found in head commit a href found in base branch master vulnerable source files deps hiredis hiredis c vulnerability details redis is an open source in memory database that persists on disk the redis cli command line tool and redis sentinel service may be vulnerable to integer overflow when parsing specially crafted large multi bulk network replies this is a result of a vulnerability in the underlying hiredis library which does not perform an overflow check before calling the calloc heap allocation function this issue only impacts systems with heap allocators that do not perform their own overflow checks most modern systems do and are therefore not likely to be affected furthermore by default redis sentinel uses the jemalloc allocator which is also not vulnerable the problem is fixed in redis versions and publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
627,538
| 19,908,105,021
|
IssuesEvent
|
2022-01-25 14:44:16
|
xournalpp/xournalpp
|
https://api.github.com/repos/xournalpp/xournalpp
|
closed
|
Crush session when close xournal++
|
bug priority::high confirmed wayland
|
(Please complete the following information, and then delete this line)
**Affects versions :**
- OS: manjaro
- (Linux only) gnome wayland
- libgtk 3.24.30
- xournal++ 1.1.0+dev
**Describe the bug**
If I use xournalpp for some time maybe hours my de restart
**To Reproduce**
Steps to reproduce the behavior:
1. Step 1 open xournal and use it for hours
2. Step 2 close it
3. etc.
**Expected behavior**
A clear and concise description of what you expected to happen.
**Screenshots of Problem**
(Optional) If applicable, add screenshots to help explain your problem.

**Additional context**
Add any other information or context about the problem here.
|
1.0
|
Crush session when close xournal++ - (Please complete the following information, and then delete this line)
**Affects versions :**
- OS: manjaro
- (Linux only) gnome wayland
- libgtk 3.24.30
- xournal++ 1.1.0+dev
**Describe the bug**
If I use xournalpp for some time maybe hours my de restart
**To Reproduce**
Steps to reproduce the behavior:
1. Step 1 open xournal and use it for hours
2. Step 2 close it
3. etc.
**Expected behavior**
A clear and concise description of what you expected to happen.
**Screenshots of Problem**
(Optional) If applicable, add screenshots to help explain your problem.

**Additional context**
Add any other information or context about the problem here.
|
non_process
|
crush session when close xournal please complete the following information and then delete this line affects versions os manjaro linux only gnome wayland libgtk xournal dev describe the bug if i use xournalpp for some time maybe hours my de restart to reproduce steps to reproduce the behavior step open xournal and use it for hours step close it etc expected behavior a clear and concise description of what you expected to happen screenshots of problem optional if applicable add screenshots to help explain your problem additional context add any other information or context about the problem here
| 0
|
10,310
| 13,156,020,143
|
IssuesEvent
|
2020-08-10 09:57:51
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
child_process spawn does not pass environment key-value pairs on Windows
|
child_process question windows
|
<!--
Thank you for reporting an issue.
This issue tracker is for bugs and issues found within Node.js core.
If you require more general support please file an issue on our help
repo. https://github.com/nodejs/help
Please fill in as much of the template below as you're able.
Version: output of `node -v`
Platform: output of `uname -a` (UNIX), or version and 32 or 64-bit (Windows)
Subsystem: if known, please specify affected core module name
-->
* **Version**: 12.18.3
* **Platform**: 64-bit Windows
### What steps will reproduce the bug?
```
require('child_process').spawn('node', ['-pe', 'process.env.PATH'], {
stdio: 'inherit',
shell: true,
env: {
...process.env,
PATH: process.env.PATH + require('path').delimiter + __dirname,
}
});
```
Actual behavior: PATH environment variable is not updated in the spawned process.
### How often does it reproduce? Is there a required condition?
Always on Windows. It is not an issue on OSX.
### What is the expected behavior?
PATH environment variable should be updated in the spawned process.
### What do you see instead?
add a fake env variable to the env object passed to spawn
```
require('child_process').spawn('node', ['-pe', 'process.env.PATH'], {
stdio: 'inherit',
shell: true,
env: {
...process.env,
PATH: process.env.PATH + require('path').delimiter + __dirname,
FAKEENV: 'FAKEVALUE',
}
});
```
|
1.0
|
child_process spawn does not pass environment key-value pairs on Windows - <!--
Thank you for reporting an issue.
This issue tracker is for bugs and issues found within Node.js core.
If you require more general support please file an issue on our help
repo. https://github.com/nodejs/help
Please fill in as much of the template below as you're able.
Version: output of `node -v`
Platform: output of `uname -a` (UNIX), or version and 32 or 64-bit (Windows)
Subsystem: if known, please specify affected core module name
-->
* **Version**: 12.18.3
* **Platform**: 64-bit Windows
### What steps will reproduce the bug?
```
require('child_process').spawn('node', ['-pe', 'process.env.PATH'], {
stdio: 'inherit',
shell: true,
env: {
...process.env,
PATH: process.env.PATH + require('path').delimiter + __dirname,
}
});
```
Actual behavior: PATH environment variable is not updated in the spawned process.
### How often does it reproduce? Is there a required condition?
Always on Windows. It is not an issue on OSX.
### What is the expected behavior?
PATH environment variable should be updated in the spawned process.
### What do you see instead?
add a fake env variable to the env object passed to spawn
```
require('child_process').spawn('node', ['-pe', 'process.env.PATH'], {
stdio: 'inherit',
shell: true,
env: {
...process.env,
PATH: process.env.PATH + require('path').delimiter + __dirname,
FAKEENV: 'FAKEVALUE',
}
});
```
|
process
|
child process spawn does not pass environment key value pairs on windows thank you for reporting an issue this issue tracker is for bugs and issues found within node js core if you require more general support please file an issue on our help repo please fill in as much of the template below as you re able version output of node v platform output of uname a unix or version and or bit windows subsystem if known please specify affected core module name version platform bit windows what steps will reproduce the bug require child process spawn node stdio inherit shell true env process env path process env path require path delimiter dirname actual behavior path environment variable is not updated in the spawned process how often does it reproduce is there a required condition always on windows it is not an issue on osx what is the expected behavior path environment variable should be updated in the spawned process what do you see instead add a fake env variable to the env object passed to spawn require child process spawn node stdio inherit shell true env process env path process env path require path delimiter dirname fakeenv fakevalue
| 1
|
10,279
| 13,131,969,874
|
IssuesEvent
|
2020-08-06 18:00:31
|
ORNL-AMO/AMO-Tools-Desktop
|
https://api.github.com/repos/ORNL-AMO/AMO-Tools-Desktop
|
closed
|
O2 enrichment tweaks
|
Calculator Process Heating
|
Change headers to "Baseline" and "Modification"
Maybe reformat to be like other BL vs Mod calcs (almost any calc from the TH) {this might be better as a separate issue}
Once the user hits "Plot" the BL is "locked" (unlocks when user hits "Reset data")
then the user can change various fields in the mod and "plot" more lines
New graphs:
y always = Savings
Different X axis:
Combustion preheat temp, flue gas temp, O2 in flue gas
Maybe switch between them like steam properties (a dropdown near the X axis?) or different tabs?
|
1.0
|
O2 enrichment tweaks - Change headers to "Baseline" and "Modification"
Maybe reformat to be like other BL vs Mod calcs (almost any calc from the TH) {this might be better as a separate issue}
Once the user hits "Plot" the BL is "locked" (unlocks when user hits "Reset data")
then the user can change various fields in the mod and "plot" more lines
New graphs:
y always = Savings
Different X axis:
Combustion preheat temp, flue gas temp, O2 in flue gas
Maybe switch between them like steam properties (a dropdown near the X axis?) or different tabs?
|
process
|
enrichment tweaks change headers to baseline and modification maybe reformat to be like other bl vs mod calcs almost any calc from the th this might be better as a separate issue once the user hits plot the bl is locked unlocks when user hits reset data then the user can change various fields in the mod and plot more lines new graphs y always savings different x axis combustion preheat temp flue gas temp in flue gas maybe switch between them like steam properties a dropdown near the x axis or different tabs
| 1
|
1,020
| 3,480,679,523
|
IssuesEvent
|
2015-12-29 10:04:44
|
DevExpress/testcafe-hammerhead
|
https://api.github.com/repos/DevExpress/testcafe-hammerhead
|
opened
|
urlUtils.getProxyUrl always add the trailing slash
|
AREA: client SYSTEM: URL processing TYPE: bug
|
For example:
urlUtils.getProxyUrl('http://google.com') -> "http://<host>:<port>/<sessionId>/http://google.com/"
Reproduced in IE, Chrome, FF
|
1.0
|
urlUtils.getProxyUrl always add the trailing slash - For example:
urlUtils.getProxyUrl('http://google.com') -> "http://<host>:<port>/<sessionId>/http://google.com/"
Reproduced in IE, Chrome, FF
|
process
|
urlutils getproxyurl always add the trailing slash for example urlutils getproxyurl reproduced in ie chrome ff
| 1
|
240,907
| 7,807,254,702
|
IssuesEvent
|
2018-06-11 16:16:15
|
fac-13/GP_ProjectBernadette
|
https://api.github.com/repos/fac-13/GP_ProjectBernadette
|
closed
|
Manage focus with auto scroll
|
priority-3
|
When using auto scroll, the page does scroll to the bottom correctly however, focus remains at the top. This is problematic for keyboard users
|
1.0
|
Manage focus with auto scroll - When using auto scroll, the page does scroll to the bottom correctly however, focus remains at the top. This is problematic for keyboard users
|
non_process
|
manage focus with auto scroll when using auto scroll the page does scroll to the bottom correctly however focus remains at the top this is problematic for keyboard users
| 0
|
2,456
| 5,240,609,797
|
IssuesEvent
|
2017-01-31 13:39:54
|
openvstorage/alba
|
https://api.github.com/repos/openvstorage/alba
|
closed
|
Namespace names in OSD backend or Cache backend are not UTF-8
|
priority_normal process_wontfix type_bug
|
Namespace names in OSD backend or Cache backend are not UTF-8. While not strictly required, UTF-8 is advised for JSON, and the python json parser seems to depend on this advise as it can't parse the JSON output returned by e.g. show-namespaces.
It would be nice if the counter could be added in some more readable way, e.g. `<prefix>_<counter>` where counter is `00000001`, `00000002`, and so on.
|
1.0
|
Namespace names in OSD backend or Cache backend are not UTF-8 - Namespace names in OSD backend or Cache backend are not UTF-8. While not strictly required, UTF-8 is advised for JSON, and the python json parser seems to depend on this advise as it can't parse the JSON output returned by e.g. show-namespaces.
It would be nice if the counter could be added in some more readable way, e.g. `<prefix>_<counter>` where counter is `00000001`, `00000002`, and so on.
|
process
|
namespace names in osd backend or cache backend are not utf namespace names in osd backend or cache backend are not utf while not strictly required utf is advised for json and the python json parser seems to depend on this advise as it can t parse the json output returned by e g show namespaces it would be nice if the counter could be added in some more readable way e g where counter is and so on
| 1
|
33,710
| 7,200,534,791
|
IssuesEvent
|
2018-02-05 19:23:23
|
Openki/Openki
|
https://api.github.com/repos/Openki/Openki
|
opened
|
Notifications: Mails classified as spam by gmx
|
Defect
|
related: #874, #802
Html mails from Openki go into spam with reason: `6 (nemesis text pattern profiler); Detail=V3;` (most other mails recognized as Junk have same text there)
I think all html mails go into spam in all gmx accounts. Can someone confirm in order to define priority of this problem?
```
Content-Type: text/html; charset=utf-8
Subject: [Openki] Private message from greg
DKIM-Signature: a=rsa-sha256; v=1; c=relaxed/relaxed; d=mail.openki.net; q=dns/txt; s=smtp; t=1517857287; h=MIME-Version: Date: Content-Transfer-Encoding: Message-ID: Subject: To: Sender: From: Content-Type; bh=rmWxmmQe7gHbsUbtiXjvXLNmqJnU8gQFKyTl0NBLYCY=; b=itu4bRnST8koq9etAgycBkphHij7mfQJSBVOXnocs0tIhdpvUcoZKTLn8Ng/sSQvLmlePCyb wZb8C7tM0A6dcj6PUj7CykYL2UModPwrktMjF9zZ1LgBIaMWha8s6OBAOs3PKFvB4WGf/Xyq xSL6T2IIJuN63NazVxV4zq5ptVA=
X-GMX-Antispam: 6 (nemesis text pattern profiler); Detail=V3;
X-UI-Filterresults: junk:10;V01:K0:2hp2TGwjjHg=:G+F7Ms/vrRNhTTZ+BtF13Q8nrw7b oojOC2iGwPE14uAjctORiCdUOEY5zXa/enanXqfC5SgJjvN0+ty9kYk0Sa+HQyHeVWOcuQd6r jz6Ac38HpwaKOQgqQ7ZitNK5cZ6v623ncE1q7p820w6W+hC7/T3mfcXQO5kUUUVpgF6zzo5jK dnxiD+ayRKlBOJdNC/BNFdKP72zgQZlWA7QwGec/0b4EWtYa6BQY3FAhzmpTQAFn3Rs7BSxha +QCeqvzwm1fqE85wVBJpSxPns4LEGOGn9FaEc2V35uG4MTHVU84+NAd2FIHc896qjRQGSXFBK hwc/+D+gnUU+utSx31NzdjgltDnM7Gy8w4j/PDhTDS5NSci0OC7/oYNTZ5/IsgHcvjwIM9pDQ hJyHxQImiiBmQp0JHsf5vXvIMcSGpj1igxIFzYTqu0aDHVjEuWu1qLX7Wu1YpZQCjkQWWKoUs wgY02wKvfZhIJ2aby/NOnfLZqBAgvyBeV23EVPMHccr3hxYC+AX2bZlCZrj6NhXxW1VMRUiOK 0KmU6TOveDEZgQS/d2qTMGEV0jYid8y8TJgRmVhT6mz+3QjojbCIOSTFKwa5yWBRJ+GI3mLvz 0ugOWmRPX622tU7BVLGN/KdjsWj13k5WnBigcYldcYWFgMiT+E75rdzCqFeP/nJYLSSWbUvrC LJGENknwftnpJNnCP2WFNY65PqlgEKo4zetIXPHsp2JpjoH70T37MSbv25w7KaBCvWUdOrQHs J33jQzN6eRF3Wvvl7xCFlWdB37Sohq0rLhzTwupglKUc+fU8js785aIv3jc7ZK7uniehaxDAL /JVS05lBqAfaxs9fphGGJKTqN4iODRJq+VjhFv0V2TIdeTXdew/ho+6/BJBfyiOv4XkumPrhA AqiBkDZRIG3VlZ2IpL6HAWRJ9RZZPSY5eDsvpOeIfNYgLFfJB92yYpOZ5kFEoZno9I1K12aHq IxK9F2krKnWbijsGVy1YVJPidLESqiM8kWBJAGk5xdW12CGhVPweFgDiwcXYh0KK9SDlVp5xn bxYQw7TJjiaVsXL47DHbdvlWWM55CUzsgT1v8g5SqI4FWW2UmdVoYdAXr9e67SHPdY5D9NAEV 5yL49utwf2wi6NsCKx/oZhqxcIDFzxCaPkJaRi09GCZ/uRMI8f1FbWs556isoULiCf5r7xQfr nFUVqJb5VlhVeVJChHk8S6IxNjx37Cf5Wcr6ImU6RqgzMA6DFM6OhzGyohldiKSEeYamS7YP4 C95rnKNkhgLmRd01ELAUuA/1puz82R56y516ZFCucDmoQZrRD4TDY8TRJo4J3il3Mf5Jn+az2 OnEn7d5SHSyE4fKCO6B7vX7pIScKj0ZBOVhrgbztC17Rti7reiQtZyblSbS9AqX+vkqXVhP0B o+O0Bn5GXWmFNMjtMyRztOHAEyRA8Jo24yBIc7QprCU5k5YORpCOdLPS6+SNtJY8P6mrdnzu2 HULf5gTl4uOxPJmeVbL/FplmvrLyrz4dQExtIt6b0w83+abkRlIxJeGzpR29OIiiEz5i3q+mS YV33HCrzpRjxLYQM18GafaMX6VWaKlz3JbhCFgwaXQ6Pb5z7ZU6O3y/ngbKVsE2fgYwv/rbnw hUkg80hKGGMyPH9qp8DnxjJmi9BwsIUHQ9bulS9KdzeVUTp0+1tr1SMkcjagO0iQamVWl6gsW /UB4TaSUVVaF3ixiYDLlWqcAlbMnGmnihgV37BAJfb/nQVfu7DhIT1ocVdFx8LXq43JdE3wUB pScnPZwTXnXGbyLXnyqbU/k3J+KSiu96NA==
```
this one didn't:
```
Content-Type: text/plain
Subject: Reset your password on Openki
DKIM-Signature: a=rsa-sha256; v=1; c=relaxed/relaxed; d=mail.openki.net; q=dns/txt; s=smtp; t=1517857782; h=MIME-Version: Date: Content-Transfer-Encoding: Message-ID: Subject: To: From: Content-Type: Sender; bh=oKUJN/Zp/axIt1TxpObfXxRaDXYHFFxDfStAEWtaZjs=; b=FX1PFwSPC0pzPgAuqHJeRMhFcjpj4gvTYc+HRaJgHiDBocQI+TfwFy+9tDjfQ4uLEIMQD8Vp gDs02VzFwyXVE5gV8G8/aSoASesN3KgQMgRpyt3LTxL717ogfj4Bry+pPxOHKJMu3s6lFAdF xkhfs5XMhVVNdzd/8gqJKybgroE=
X-GMX-Antispam: 0 (Mail was not recognized as spam); Detail=V3;
X-UI-Filterresults: notjunk:1;V01:K0:E+67RFQNlC8=:wfCyEvRQIf1Qg1xCghnW2ks5V8 U0pkQ9zA7JmK7Trq4Qjg3oTOOqKb70H3q/5DmPDcXuaU8yfoWaZWb0V659TVDbGp437jNcVGm p9k6O/owYlLtmcQbGjvRsNlvDJGZMCGjJILDLIyiKVMIE2rxzAaYUPuhtnKaSx19RbnrAnZsQ aQHgnJQrT243vL1QPkbulfuBoS1kQ8/r8q7cUDD6rxvg74hxM+EyVaqPh1WyguUwrNOWF3tZN K+O5yvp0YTq8cciAY7tNYzu7xVvqo/BLVGph2CSMe+azRrNd1tP627GvOoDrJUoP1Afg/JwFY 69xhBLUJ+VmBmRe37IYkTshdRUM+nJcno/XHeTIUKb/27S9A2zfFht/lCe9gfPLXCWpnPaoKe 77OVcqYHrRObb7nE2T1PTLYTkZh9SFBtyqfWBkzBYMUgVOwcERj51WsjNdt0IZT2168tvuFu6 2kRF4KAg+I4vSOFOd14eSpFy+6zpgkCNvWDkq83p8KmCc3Nx1cKBm1hRK5OKaE5f3MqG8uSCd Ew+4XfJClVK7xGe8/b9yZ04EaTFzJ3435nc4IXaHaAB9XcSUSJuUHsLzblxWHAG7GwyL6HfHK peJnMk8utaKljvxHrz3MFfaXEJlyLGsVe3phNeLX/2R5FDoEBZt+3TYJx5AwNsHNktjGRhLud 8/8iNd9/0SOrsY/xFu55YjCe9iAdi9FSiZSDeo6WfqWHKb75px7NMUN2iKO+YF4aRjBfTB3EK Al8hVJF1+eFDd0dz65bzbcV9UCvosNp4PZ/SLT19cugz2eiV0RDnyCPBaMpAPwYjzYpAkmhjm DBCN4HloHTwJEo7MbkygUyR299O+wTDhiNR9Z8fDht3hwSfwS2j9SY+4qsmswRf4Id94O10NJ cwZDEm4SCziaGFV70a9C9ADuFzAW1jrfcQkvpRRe/Zm966CwZ0DW7eDCWmefovc6LXdHITBhv kASG36BSW+zzw4cNneBnKaIgACWXFH37Cp7shkVg/kuOYbV2Tu+wxO7Xc5yG2Z5G8HbLfvyJ1 OxjTFJGZjNAsgmr0pVq0gQY6Sbt0TYfQS3x3RBvMnp+MYumwfIshOJehDyj1zbvnBx4/5aTTg xuuJhOviQ2FSDNkWMc8sA3SUo6EmPydv0dMqntnKHmwyHEtsKnjXg6BLmqmunpJbpQkxgleb9 QAP0we5LoJyQw2VoFL/FuTsaKwbWxTcPAPJLVLT3ErYRsAkTnVdSLwuBjM92R8uBbVDJWxJ3F DMrKbsuDSax4C03IAEStz6jbCIRgM/BfXNq0LmdBRwelViHszBCqcp9Les+Iu1J5D0w9Qzf4C yrl5pte8/PWrbfQtpnAK+omQODQDxqtjO3qHoCEagryD3Ym1Qi92/O3TN+ccHIg6iFPwRNvcz wSUjlCydXu25FIg0rxiy/k0LHjFLrM9t6QuLLpV5ET38se5sFqa6Ufy84K41RoCB1h8Rv6tL8 NDaeEK1mJT2KvEliYH6IalUrfJjXHG/g+KLUfUSRrGhzAINCUZCh5bjAMO1O21MBLBw3wNvod gk6GWGC+4QlP32UcWI4BU+poNYYBKQu/rdKrQ5KM9tDwW41ZKd4lWqqEQnfmSNhBzNaFJrWiG WnCKG/a9x1PesPgY+uEQ4XT8ImE8/6rO9EzssBmWzJaueeih+gUh9UpuT8iybCP6alxiT/ky5 R3ZK9qy2oT9gpzUiCiuCegEIBmDE455wzk87FbDXvWXd1kyAHcvUwucM52a7aTzkBqmnH4B4F U1eFxftRLxAmFdkF5Hcx61AYfNZSo3evrhLQ==
```
|
1.0
|
Notifications: Mails classified as spam by gmx - related: #874, #802
Html mails from Openki go into spam with reason: `6 (nemesis text pattern profiler); Detail=V3;` (most other mails recognized as Junk have same text there)
I think all html mails go into spam in all gmx accounts. Can someone confirm in order to define priority of this problem?
```
Content-Type: text/html; charset=utf-8
Subject: [Openki] Private message from greg
DKIM-Signature: a=rsa-sha256; v=1; c=relaxed/relaxed; d=mail.openki.net; q=dns/txt; s=smtp; t=1517857287; h=MIME-Version: Date: Content-Transfer-Encoding: Message-ID: Subject: To: Sender: From: Content-Type; bh=rmWxmmQe7gHbsUbtiXjvXLNmqJnU8gQFKyTl0NBLYCY=; b=itu4bRnST8koq9etAgycBkphHij7mfQJSBVOXnocs0tIhdpvUcoZKTLn8Ng/sSQvLmlePCyb wZb8C7tM0A6dcj6PUj7CykYL2UModPwrktMjF9zZ1LgBIaMWha8s6OBAOs3PKFvB4WGf/Xyq xSL6T2IIJuN63NazVxV4zq5ptVA=
X-GMX-Antispam: 6 (nemesis text pattern profiler); Detail=V3;
X-UI-Filterresults: junk:10;V01:K0:2hp2TGwjjHg=:G+F7Ms/vrRNhTTZ+BtF13Q8nrw7b oojOC2iGwPE14uAjctORiCdUOEY5zXa/enanXqfC5SgJjvN0+ty9kYk0Sa+HQyHeVWOcuQd6r jz6Ac38HpwaKOQgqQ7ZitNK5cZ6v623ncE1q7p820w6W+hC7/T3mfcXQO5kUUUVpgF6zzo5jK dnxiD+ayRKlBOJdNC/BNFdKP72zgQZlWA7QwGec/0b4EWtYa6BQY3FAhzmpTQAFn3Rs7BSxha +QCeqvzwm1fqE85wVBJpSxPns4LEGOGn9FaEc2V35uG4MTHVU84+NAd2FIHc896qjRQGSXFBK hwc/+D+gnUU+utSx31NzdjgltDnM7Gy8w4j/PDhTDS5NSci0OC7/oYNTZ5/IsgHcvjwIM9pDQ hJyHxQImiiBmQp0JHsf5vXvIMcSGpj1igxIFzYTqu0aDHVjEuWu1qLX7Wu1YpZQCjkQWWKoUs wgY02wKvfZhIJ2aby/NOnfLZqBAgvyBeV23EVPMHccr3hxYC+AX2bZlCZrj6NhXxW1VMRUiOK 0KmU6TOveDEZgQS/d2qTMGEV0jYid8y8TJgRmVhT6mz+3QjojbCIOSTFKwa5yWBRJ+GI3mLvz 0ugOWmRPX622tU7BVLGN/KdjsWj13k5WnBigcYldcYWFgMiT+E75rdzCqFeP/nJYLSSWbUvrC LJGENknwftnpJNnCP2WFNY65PqlgEKo4zetIXPHsp2JpjoH70T37MSbv25w7KaBCvWUdOrQHs J33jQzN6eRF3Wvvl7xCFlWdB37Sohq0rLhzTwupglKUc+fU8js785aIv3jc7ZK7uniehaxDAL /JVS05lBqAfaxs9fphGGJKTqN4iODRJq+VjhFv0V2TIdeTXdew/ho+6/BJBfyiOv4XkumPrhA AqiBkDZRIG3VlZ2IpL6HAWRJ9RZZPSY5eDsvpOeIfNYgLFfJB92yYpOZ5kFEoZno9I1K12aHq IxK9F2krKnWbijsGVy1YVJPidLESqiM8kWBJAGk5xdW12CGhVPweFgDiwcXYh0KK9SDlVp5xn bxYQw7TJjiaVsXL47DHbdvlWWM55CUzsgT1v8g5SqI4FWW2UmdVoYdAXr9e67SHPdY5D9NAEV 5yL49utwf2wi6NsCKx/oZhqxcIDFzxCaPkJaRi09GCZ/uRMI8f1FbWs556isoULiCf5r7xQfr nFUVqJb5VlhVeVJChHk8S6IxNjx37Cf5Wcr6ImU6RqgzMA6DFM6OhzGyohldiKSEeYamS7YP4 C95rnKNkhgLmRd01ELAUuA/1puz82R56y516ZFCucDmoQZrRD4TDY8TRJo4J3il3Mf5Jn+az2 OnEn7d5SHSyE4fKCO6B7vX7pIScKj0ZBOVhrgbztC17Rti7reiQtZyblSbS9AqX+vkqXVhP0B o+O0Bn5GXWmFNMjtMyRztOHAEyRA8Jo24yBIc7QprCU5k5YORpCOdLPS6+SNtJY8P6mrdnzu2 HULf5gTl4uOxPJmeVbL/FplmvrLyrz4dQExtIt6b0w83+abkRlIxJeGzpR29OIiiEz5i3q+mS YV33HCrzpRjxLYQM18GafaMX6VWaKlz3JbhCFgwaXQ6Pb5z7ZU6O3y/ngbKVsE2fgYwv/rbnw hUkg80hKGGMyPH9qp8DnxjJmi9BwsIUHQ9bulS9KdzeVUTp0+1tr1SMkcjagO0iQamVWl6gsW /UB4TaSUVVaF3ixiYDLlWqcAlbMnGmnihgV37BAJfb/nQVfu7DhIT1ocVdFx8LXq43JdE3wUB pScnPZwTXnXGbyLXnyqbU/k3J+KSiu96NA==
```
this one didn't:
```
Content-Type: text/plain
Subject: Reset your password on Openki
DKIM-Signature: a=rsa-sha256; v=1; c=relaxed/relaxed; d=mail.openki.net; q=dns/txt; s=smtp; t=1517857782; h=MIME-Version: Date: Content-Transfer-Encoding: Message-ID: Subject: To: From: Content-Type: Sender; bh=oKUJN/Zp/axIt1TxpObfXxRaDXYHFFxDfStAEWtaZjs=; b=FX1PFwSPC0pzPgAuqHJeRMhFcjpj4gvTYc+HRaJgHiDBocQI+TfwFy+9tDjfQ4uLEIMQD8Vp gDs02VzFwyXVE5gV8G8/aSoASesN3KgQMgRpyt3LTxL717ogfj4Bry+pPxOHKJMu3s6lFAdF xkhfs5XMhVVNdzd/8gqJKybgroE=
X-GMX-Antispam: 0 (Mail was not recognized as spam); Detail=V3;
X-UI-Filterresults: notjunk:1;V01:K0:E+67RFQNlC8=:wfCyEvRQIf1Qg1xCghnW2ks5V8 U0pkQ9zA7JmK7Trq4Qjg3oTOOqKb70H3q/5DmPDcXuaU8yfoWaZWb0V659TVDbGp437jNcVGm p9k6O/owYlLtmcQbGjvRsNlvDJGZMCGjJILDLIyiKVMIE2rxzAaYUPuhtnKaSx19RbnrAnZsQ aQHgnJQrT243vL1QPkbulfuBoS1kQ8/r8q7cUDD6rxvg74hxM+EyVaqPh1WyguUwrNOWF3tZN K+O5yvp0YTq8cciAY7tNYzu7xVvqo/BLVGph2CSMe+azRrNd1tP627GvOoDrJUoP1Afg/JwFY 69xhBLUJ+VmBmRe37IYkTshdRUM+nJcno/XHeTIUKb/27S9A2zfFht/lCe9gfPLXCWpnPaoKe 77OVcqYHrRObb7nE2T1PTLYTkZh9SFBtyqfWBkzBYMUgVOwcERj51WsjNdt0IZT2168tvuFu6 2kRF4KAg+I4vSOFOd14eSpFy+6zpgkCNvWDkq83p8KmCc3Nx1cKBm1hRK5OKaE5f3MqG8uSCd Ew+4XfJClVK7xGe8/b9yZ04EaTFzJ3435nc4IXaHaAB9XcSUSJuUHsLzblxWHAG7GwyL6HfHK peJnMk8utaKljvxHrz3MFfaXEJlyLGsVe3phNeLX/2R5FDoEBZt+3TYJx5AwNsHNktjGRhLud 8/8iNd9/0SOrsY/xFu55YjCe9iAdi9FSiZSDeo6WfqWHKb75px7NMUN2iKO+YF4aRjBfTB3EK Al8hVJF1+eFDd0dz65bzbcV9UCvosNp4PZ/SLT19cugz2eiV0RDnyCPBaMpAPwYjzYpAkmhjm DBCN4HloHTwJEo7MbkygUyR299O+wTDhiNR9Z8fDht3hwSfwS2j9SY+4qsmswRf4Id94O10NJ cwZDEm4SCziaGFV70a9C9ADuFzAW1jrfcQkvpRRe/Zm966CwZ0DW7eDCWmefovc6LXdHITBhv kASG36BSW+zzw4cNneBnKaIgACWXFH37Cp7shkVg/kuOYbV2Tu+wxO7Xc5yG2Z5G8HbLfvyJ1 OxjTFJGZjNAsgmr0pVq0gQY6Sbt0TYfQS3x3RBvMnp+MYumwfIshOJehDyj1zbvnBx4/5aTTg xuuJhOviQ2FSDNkWMc8sA3SUo6EmPydv0dMqntnKHmwyHEtsKnjXg6BLmqmunpJbpQkxgleb9 QAP0we5LoJyQw2VoFL/FuTsaKwbWxTcPAPJLVLT3ErYRsAkTnVdSLwuBjM92R8uBbVDJWxJ3F DMrKbsuDSax4C03IAEStz6jbCIRgM/BfXNq0LmdBRwelViHszBCqcp9Les+Iu1J5D0w9Qzf4C yrl5pte8/PWrbfQtpnAK+omQODQDxqtjO3qHoCEagryD3Ym1Qi92/O3TN+ccHIg6iFPwRNvcz wSUjlCydXu25FIg0rxiy/k0LHjFLrM9t6QuLLpV5ET38se5sFqa6Ufy84K41RoCB1h8Rv6tL8 NDaeEK1mJT2KvEliYH6IalUrfJjXHG/g+KLUfUSRrGhzAINCUZCh5bjAMO1O21MBLBw3wNvod gk6GWGC+4QlP32UcWI4BU+poNYYBKQu/rdKrQ5KM9tDwW41ZKd4lWqqEQnfmSNhBzNaFJrWiG WnCKG/a9x1PesPgY+uEQ4XT8ImE8/6rO9EzssBmWzJaueeih+gUh9UpuT8iybCP6alxiT/ky5 R3ZK9qy2oT9gpzUiCiuCegEIBmDE455wzk87FbDXvWXd1kyAHcvUwucM52a7aTzkBqmnH4B4F U1eFxftRLxAmFdkF5Hcx61AYfNZSo3evrhLQ==
```
|
non_process
|
notifications mails classified as spam by gmx related html mails from openki go into spam with reason nemesis text pattern profiler detail most other mails recognized as junk have same text there i think all html mails go into spam in all gmx accounts can someone confirm in order to define priority of this problem content type text html charset utf subject private message from greg dkim signature a rsa v c relaxed relaxed d mail openki net q dns txt s smtp t h mime version date content transfer encoding message id subject to sender from content type bh b ssqvlmlepcyb xyq x gmx antispam nemesis text pattern profiler detail x ui filterresults junk g vrrnhttz dnxid ayrklbojdnc hwc d gnuu njylsswbuvrc ho o ms rbnw pscnpzwtxnxgbylxnyqbu this one didn t content type text plain subject reset your password on openki dkim signature a rsa v c relaxed relaxed d mail openki net q dns txt s smtp t h mime version date content transfer encoding message id subject to from content type sender bh okujn zp b hrajghidbocqi tfwfy x gmx antispam mail was not recognized as spam detail x ui filterresults notjunk e k jwfy njcno xhetiukb ew pwrbfqtpnak g ponyybkqu wnckg
| 0
|
22,188
| 30,737,262,412
|
IssuesEvent
|
2023-07-28 08:39:31
|
dita-ot/dita-ot
|
https://api.github.com/repos/dita-ot/dita-ot
|
closed
|
Preprocess2 does not report validation error
|
bug priority/medium preprocess preprocess2
|
This problem is from following Stack Overflow question:
minOccurs maxOccurs don't work with dita?
https://stackoverflow.com/questions/55454803/minoccurs-maxoccurs-dont-work-with-dita
The reproducing step:
1. Modify OASIS XSD file: dita-ot-3.3.1/plugins/org.oasis-open.dita.v1_3/schema/bookmap/xsd/bookmapMod.xsd to `minOccurs="1"`.
````xml
<xs:group name="bookmeta.content">
<xs:sequence>
<xs:group ref="linktext" minOccurs="1"/>
<xs:group ref="searchtitle" minOccurs="1"/>
<xs:group ref="shortdesc" minOccurs="1"/>
<xs:group ref="author" minOccurs="1" maxOccurs="unbounded"/>
<xs:group ref="source" minOccurs="1"/>
<xs:group ref="publisherinformation" minOccurs="1" maxOccurs="unbounded"/>
<xs:group ref="critdates" minOccurs="1"/>
<xs:group ref="permissions" minOccurs="1"/>
<xs:group ref="metadata" minOccurs="1" maxOccurs="unbounded"/>
<xs:group ref="audience" minOccurs="1" maxOccurs="unbounded"/>
<xs:group ref="category" minOccurs="1" maxOccurs="unbounded"/>
<xs:group ref="keywords" minOccurs="1" maxOccurs="unbounded"/>
<xs:group ref="prodinfo" minOccurs="1" maxOccurs="unbounded"/>
<xs:group ref="othermeta" minOccurs="1" maxOccurs="unbounded"/>
<xs:group ref="resourceid" minOccurs="1" maxOccurs="unbounded"/>
<xs:group ref="ux-window" minOccurs="1" maxOccurs="unbounded"/>
<xs:group ref="bookid" minOccurs="1"/>
<xs:group ref="bookchangehistory" minOccurs="1" maxOccurs="unbounded"/>
<xs:group ref="bookrights" minOccurs="1" maxOccurs="unbounded"/>
<xs:group ref="data" minOccurs="1" maxOccurs="unbounded"/>
</xs:sequence>
</xs:group>
````
2. Modify the sample file: dita-ot-3.3.1/docsrc/samples/taskbook.ditamap to use XML Schema
````xml
<bookmap xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
xsi:noNamespaceSchemaLocation="urn:oasis:names:tc:dita:xsd:bookmap.xsd" id="taskbook">
<!--DOCTYPE bookmap PUBLIC "-//OASIS//DTD DITA BookMap//EN" "bookmap.dtd"-->
<!-- This file is part of the DITA Open Toolkit project. See the accompanying LICENSE file for applicable license. -->
<!-- (c) Copyright IBM Corp. 2004, 2007 All Rights Reserved. -->
<!--bookmap id="taskbook"-->
<booktitle>
<booklibrary>Retro Tools</booklibrary>
<mainbooktitle>Product tasks</mainbooktitle>
<booktitlealt>Tasks and what they can do</booktitlealt>
</booktitle>
<bookmeta>
<author>Howe Tuduit</author>
<bookid>
<isbn>071271271X</isbn>
<booknumber>SG99-9999-00</booknumber>
<maintainer>
<organization>Retro Tools</organization>
<person/>
</maintainer>
</bookid>
<bookrights>
<copyrfirst>
<year>2004</year>
</copyrfirst>
<copyrlast>
<year>2007</year>
</copyrlast>
<bookowner>
<organization>Retro Tools, Inc.</organization>
</bookowner>
</bookrights>
</bookmeta>
...
</bookmap>
````
3. Run dita command
`bin/dita -f pdf -i docsrc/samples/taskbook.ditamap -o out -l out/log.txt -v`
The result:
`preprocess2` does not report any validation error, but if I change `preprocess2` to `preprocess` in dita-ot-3.3.1/plugins/pdf2/build.xml, the validation error is reported.
> Error: Failed to run pipeline: [DOTJ012F][FATAL] Failed to parse the input file 'file:/D:/DITA-OT/dita-ot-3.3.1/docsrc/samples/taskbook.ditamap'.: file:/D:/DITA-OT/dita-ot-3.3.1/docsrc/samples/taskbook.ditamap Line 13:cvc-complex-type.2.4.a: Invalid content was found starting with element 'author'. One of '{linktext}' is expected.
[log-preprocess2.txt](https://github.com/dita-ot/dita-ot/files/3064208/log-preprocess2.txt)
[log-preprocess.txt](https://github.com/dita-ot/dita-ot/files/3064209/log-preprocess.txt)
I think `preprocess2` should also report validation error.
|
2.0
|
Preprocess2 does not report validation error - This problem is from following Stack Overflow question:
minOccurs maxOccurs don't work with dita?
https://stackoverflow.com/questions/55454803/minoccurs-maxoccurs-dont-work-with-dita
The reproducing step:
1. Modify OASIS XSD file: dita-ot-3.3.1/plugins/org.oasis-open.dita.v1_3/schema/bookmap/xsd/bookmapMod.xsd to `minOccurs="1"`.
````xml
<xs:group name="bookmeta.content">
<xs:sequence>
<xs:group ref="linktext" minOccurs="1"/>
<xs:group ref="searchtitle" minOccurs="1"/>
<xs:group ref="shortdesc" minOccurs="1"/>
<xs:group ref="author" minOccurs="1" maxOccurs="unbounded"/>
<xs:group ref="source" minOccurs="1"/>
<xs:group ref="publisherinformation" minOccurs="1" maxOccurs="unbounded"/>
<xs:group ref="critdates" minOccurs="1"/>
<xs:group ref="permissions" minOccurs="1"/>
<xs:group ref="metadata" minOccurs="1" maxOccurs="unbounded"/>
<xs:group ref="audience" minOccurs="1" maxOccurs="unbounded"/>
<xs:group ref="category" minOccurs="1" maxOccurs="unbounded"/>
<xs:group ref="keywords" minOccurs="1" maxOccurs="unbounded"/>
<xs:group ref="prodinfo" minOccurs="1" maxOccurs="unbounded"/>
<xs:group ref="othermeta" minOccurs="1" maxOccurs="unbounded"/>
<xs:group ref="resourceid" minOccurs="1" maxOccurs="unbounded"/>
<xs:group ref="ux-window" minOccurs="1" maxOccurs="unbounded"/>
<xs:group ref="bookid" minOccurs="1"/>
<xs:group ref="bookchangehistory" minOccurs="1" maxOccurs="unbounded"/>
<xs:group ref="bookrights" minOccurs="1" maxOccurs="unbounded"/>
<xs:group ref="data" minOccurs="1" maxOccurs="unbounded"/>
</xs:sequence>
</xs:group>
````
2. Modify the sample file: dita-ot-3.3.1/docsrc/samples/taskbook.ditamap to use XML Schema
````xml
<bookmap xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
xsi:noNamespaceSchemaLocation="urn:oasis:names:tc:dita:xsd:bookmap.xsd" id="taskbook">
<!--DOCTYPE bookmap PUBLIC "-//OASIS//DTD DITA BookMap//EN" "bookmap.dtd"-->
<!-- This file is part of the DITA Open Toolkit project. See the accompanying LICENSE file for applicable license. -->
<!-- (c) Copyright IBM Corp. 2004, 2007 All Rights Reserved. -->
<!--bookmap id="taskbook"-->
<booktitle>
<booklibrary>Retro Tools</booklibrary>
<mainbooktitle>Product tasks</mainbooktitle>
<booktitlealt>Tasks and what they can do</booktitlealt>
</booktitle>
<bookmeta>
<author>Howe Tuduit</author>
<bookid>
<isbn>071271271X</isbn>
<booknumber>SG99-9999-00</booknumber>
<maintainer>
<organization>Retro Tools</organization>
<person/>
</maintainer>
</bookid>
<bookrights>
<copyrfirst>
<year>2004</year>
</copyrfirst>
<copyrlast>
<year>2007</year>
</copyrlast>
<bookowner>
<organization>Retro Tools, Inc.</organization>
</bookowner>
</bookrights>
</bookmeta>
...
</bookmap>
````
3. Run dita command
`bin/dita -f pdf -i docsrc/samples/taskbook.ditamap -o out -l out/log.txt -v`
The result:
`preprocess2` does not report any validation error, but if I change `preprocess2` to `preprocess` in dita-ot-3.3.1/plugins/pdf2/build.xml, the validation error is reported.
> Error: Failed to run pipeline: [DOTJ012F][FATAL] Failed to parse the input file 'file:/D:/DITA-OT/dita-ot-3.3.1/docsrc/samples/taskbook.ditamap'.: file:/D:/DITA-OT/dita-ot-3.3.1/docsrc/samples/taskbook.ditamap Line 13:cvc-complex-type.2.4.a: Invalid content was found starting with element 'author'. One of '{linktext}' is expected.
[log-preprocess2.txt](https://github.com/dita-ot/dita-ot/files/3064208/log-preprocess2.txt)
[log-preprocess.txt](https://github.com/dita-ot/dita-ot/files/3064209/log-preprocess.txt)
I think `preprocess2` should also report validation error.
|
process
|
does not report validation error this problem is from following stack overflow question minoccurs maxoccurs don t work with dita the reproducing step modify oasis xsd file dita ot plugins org oasis open dita schema bookmap xsd bookmapmod xsd to minoccurs xml modify the sample file dita ot docsrc samples taskbook ditamap to use xml schema xml bookmap xmlns xsi xsi nonamespaceschemalocation urn oasis names tc dita xsd bookmap xsd id taskbook retro tools product tasks tasks and what they can do howe tuduit retro tools retro tools inc run dita command bin dita f pdf i docsrc samples taskbook ditamap o out l out log txt v the result does not report any validation error but if i change to preprocess in dita ot plugins build xml the validation error is reported error failed to run pipeline failed to parse the input file file d dita ot dita ot docsrc samples taskbook ditamap file d dita ot dita ot docsrc samples taskbook ditamap line cvc complex type a invalid content was found starting with element author one of linktext is expected i think should also report validation error
| 1
|
22,750
| 32,067,682,147
|
IssuesEvent
|
2023-09-25 05:18:47
|
open-telemetry/opentelemetry-collector-contrib
|
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
|
closed
|
Resource detection for EC2 partially works if collector is running as a systemd service on an EC2
|
bug Stale priority:p2 processor/resourcedetection closed as inactive
|
**Describe the bug**
Hello team, I'm running the Otel collector as a systemd service on an ec2 instance. I have resource detection processor enabled on the collector with `detectors: ["ec2"]` and added configuration to gather tags for the EC2 instance. However it does not detect the EC2 information.
- If I run the collector plainly on the instance host and not as a systemd service
or
- If I run the collector as a systemd service and only have `detectors: ["ec2"]` without the `tags` parameter
the resource detection for EC2 starts working again.
**Steps to reproduce**
Enable detectors: ["ec2"] under resource detection processor with the `tags` parameter in config.yaml
Run Otel collector as a systemd service
**What did you expect to see?**
I expected to see something similar to this in the collector startup log
`info internal/resourcedetection.go:150 detected resource information {"kind": "processor", "name": "resourcedetection", "pipeline": "traces", "resource": {"cloud.account.id":"XXXXXXXXXXXX","cloud.availability_zone":"us-east-1c","cloud.platform":"aws_ec2","cloud.provider":"aws","cloud.region":"us-east-1","host.id":"i-XXXXXXXXXXXXXXXXX","host.image.id":"ami-XXXXXXXXXXXXXXXXX","host.name":"ip-XX-XX-XX-XX","host.type":"m5.xlarge"}}`
**What did you see instead?**
`warn internal/resourcedetection.go:141 failed to detect resource {"kind": "processor", "name": "resourcedetection", "pipeline": "traces", "error": "failed fetching ec2 instance tags: RequestError: send request failed\ncaused by: Post \"https://ec2.us-east-1.amazonaws.com/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)"}`
**What version did you use?**
Version: v0.58.0
**What config did you use?**
Config:
```
receivers:
hostmetrics:
collection_interval: 10m
scrapers:
cpu:
disk:
load:
filesystem:
memory:
network:
paging:
processes:
process:
docker_stats:
collection_interval: 5m
otlp:
protocols:
grpc:
http:
exporters:
logging:
loglevel: debug
otlp:
endpoint: $OTLP_ENDPOINT
processors:
batch:
resourcedetection:
detectors: ["ec2"]
ec2:
tags:
- ^ASV$
- ^BA$
extensions:
health_check:
pprof:
zpages:
service:
extensions: [pprof, zpages, health_check]
pipelines:
traces:
receivers: [otlp]
exporters: [logging, otlp]
processors: [batch, resourcedetection]
metrics:
receivers: [otlp, hostmetrics, docker_stats]
exporters: [otlp]
processors: [batch]
logs:
receivers: [otlp, filelog]
exporters: [otlp]
processors: [batch]
```
**Environment**
OS: Linux 5.10.130-118.517.amzn2.x86_64
Compiler(if manually compiled): (e.g., "go 14.2")
**Additional context**
Add any other context about the problem here.
|
1.0
|
Resource detection for EC2 partially works if collector is running as a systemd service on an EC2 - **Describe the bug**
Hello team, I'm running the Otel collector as a systemd service on an ec2 instance. I have resource detection processor enabled on the collector with `detectors: ["ec2"]` and added configuration to gather tags for the EC2 instance. However it does not detect the EC2 information.
- If I run the collector plainly on the instance host and not as a systemd service
or
- If I run the collector as a systemd service and only have `detectors: ["ec2"]` without the `tags` parameter
the resource detection for EC2 starts working again.
**Steps to reproduce**
Enable detectors: ["ec2"] under resource detection processor with the `tags` parameter in config.yaml
Run Otel collector as a systemd service
**What did you expect to see?**
I expected to see something similar to this in the collector startup log
`info internal/resourcedetection.go:150 detected resource information {"kind": "processor", "name": "resourcedetection", "pipeline": "traces", "resource": {"cloud.account.id":"XXXXXXXXXXXX","cloud.availability_zone":"us-east-1c","cloud.platform":"aws_ec2","cloud.provider":"aws","cloud.region":"us-east-1","host.id":"i-XXXXXXXXXXXXXXXXX","host.image.id":"ami-XXXXXXXXXXXXXXXXX","host.name":"ip-XX-XX-XX-XX","host.type":"m5.xlarge"}}`
**What did you see instead?**
`warn internal/resourcedetection.go:141 failed to detect resource {"kind": "processor", "name": "resourcedetection", "pipeline": "traces", "error": "failed fetching ec2 instance tags: RequestError: send request failed\ncaused by: Post \"https://ec2.us-east-1.amazonaws.com/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)"}`
**What version did you use?**
Version: v0.58.0
**What config did you use?**
Config:
```
receivers:
hostmetrics:
collection_interval: 10m
scrapers:
cpu:
disk:
load:
filesystem:
memory:
network:
paging:
processes:
process:
docker_stats:
collection_interval: 5m
otlp:
protocols:
grpc:
http:
exporters:
logging:
loglevel: debug
otlp:
endpoint: $OTLP_ENDPOINT
processors:
batch:
resourcedetection:
detectors: ["ec2"]
ec2:
tags:
- ^ASV$
- ^BA$
extensions:
health_check:
pprof:
zpages:
service:
extensions: [pprof, zpages, health_check]
pipelines:
traces:
receivers: [otlp]
exporters: [logging, otlp]
processors: [batch, resourcedetection]
metrics:
receivers: [otlp, hostmetrics, docker_stats]
exporters: [otlp]
processors: [batch]
logs:
receivers: [otlp, filelog]
exporters: [otlp]
processors: [batch]
```
**Environment**
OS: Linux 5.10.130-118.517.amzn2.x86_64
Compiler(if manually compiled): (e.g., "go 14.2")
**Additional context**
Add any other context about the problem here.
|
process
|
resource detection for partially works if collector is running as a systemd service on an describe the bug hello team i m running the otel collector as a systemd service on an instance i have resource detection processor enabled on the collector with detectors and added configuration to gather tags for the instance however it does not detect the information if i run the collector plainly on the instance host and not as a systemd service or if i run the collector as a systemd service and only have detectors without the tags parameter the resource detection for starts working again steps to reproduce enable detectors under resource detection processor with the tags parameter in config yaml run otel collector as a systemd service what did you expect to see i expected to see something similar to this in the collector startup log info internal resourcedetection go detected resource information kind processor name resourcedetection pipeline traces resource cloud account id xxxxxxxxxxxx cloud availability zone us east cloud platform aws cloud provider aws cloud region us east host id i xxxxxxxxxxxxxxxxx host image id ami xxxxxxxxxxxxxxxxx host name ip xx xx xx xx host type xlarge what did you see instead warn internal resourcedetection go failed to detect resource kind processor name resourcedetection pipeline traces error failed fetching instance tags requesterror send request failed ncaused by post context deadline exceeded client timeout exceeded while awaiting headers what version did you use version what config did you use config receivers hostmetrics collection interval scrapers cpu disk load filesystem memory network paging processes process docker stats collection interval otlp protocols grpc http exporters logging loglevel debug otlp endpoint otlp endpoint processors batch resourcedetection detectors tags asv ba extensions health check pprof zpages service extensions pipelines traces receivers exporters processors metrics receivers exporters processors logs receivers exporters processors environment os linux compiler if manually compiled e g go additional context add any other context about the problem here
| 1
|
105,763
| 4,241,468,486
|
IssuesEvent
|
2016-07-06 16:23:55
|
CascadesCarnivoreProject/Timelapse
|
https://api.github.com/repos/CascadesCarnivoreProject/Timelapse
|
closed
|
UX: video dialog can only be displayed once
|
High Priority fix
|
The following scenario has undesirable behaviour:
- user pops video dialog through menu
- user closes video dialog
- user selects the video dialog menu item a second time and nothing happens
This is a state management bug in MenuItemVideoViewer_Click() causing the call to Show() to be one time only.
|
1.0
|
UX: video dialog can only be displayed once - The following scenario has undesirable behaviour:
- user pops video dialog through menu
- user closes video dialog
- user selects the video dialog menu item a second time and nothing happens
This is a state management bug in MenuItemVideoViewer_Click() causing the call to Show() to be one time only.
|
non_process
|
ux video dialog can only be displayed once the following scenario has undesirable behaviour user pops video dialog through menu user closes video dialog user selects the video dialog menu item a second time and nothing happens this is a state management bug in menuitemvideoviewer click causing the call to show to be one time only
| 0
|
1,573
| 4,167,451,307
|
IssuesEvent
|
2016-06-20 09:34:16
|
e-government-ua/iBP
|
https://api.github.com/repos/e-government-ua/iBP
|
closed
|
Видача дозволу на розміщення зовнішньої реклами - м.Первомайськ - Миколаївська область
|
In process of testing in work test
|
[Видача дозволу на розміщення зовнішньої реклами.pdf](https://github.com/e-government-ua/iBP/files/262917/default.pdf)
**Координатор города:**
Корой Виталий - 0991962269 - somati.orlik@gmail.com
И еще просьба, в получатели добавить еще два контакта, для более быстрого реагирования )
kukharenko.vsevolod@gmail.com - Всеволод Кухаренко
sergey.donchenko@gmail.com - Сергей Донченко
**Контактное лицо ЦНАП:**
Наталія Миколаївна Петрущак, natashadc@mail.ru, (099) 277-41-42
на тест отправляем в ЦНАП, но обязательно ставим в копию 3 координаторов выше.
|
1.0
|
Видача дозволу на розміщення зовнішньої реклами - м.Первомайськ - Миколаївська область - [Видача дозволу на розміщення зовнішньої реклами.pdf](https://github.com/e-government-ua/iBP/files/262917/default.pdf)
**Координатор города:**
Корой Виталий - 0991962269 - somati.orlik@gmail.com
И еще просьба, в получатели добавить еще два контакта, для более быстрого реагирования )
kukharenko.vsevolod@gmail.com - Всеволод Кухаренко
sergey.donchenko@gmail.com - Сергей Донченко
**Контактное лицо ЦНАП:**
Наталія Миколаївна Петрущак, natashadc@mail.ru, (099) 277-41-42
на тест отправляем в ЦНАП, но обязательно ставим в копию 3 координаторов выше.
|
process
|
видача дозволу на розміщення зовнішньої реклами м первомайськ миколаївська область координатор города корой виталий somati orlik gmail com и еще просьба в получатели добавить еще два контакта для более быстрого реагирования kukharenko vsevolod gmail com всеволод кухаренко sergey donchenko gmail com сергей донченко контактное лицо цнап наталія миколаївна петрущак natashadc mail ru на тест отправляем в цнап но обязательно ставим в копию координаторов выше
| 1
|
247,247
| 7,915,697,252
|
IssuesEvent
|
2018-07-04 01:03:10
|
StrangeLoopGames/EcoIssues
|
https://api.github.com/repos/StrangeLoopGames/EcoIssues
|
closed
|
USER ISSUE: cant start any save data
|
High Priority
|
**Version:** 0.7.5.1 beta
**Steps to Reproduce:**
i try to load any game and nothing happens
**Expected behavior:**
play
**Actual behavior:**
nothing
|
1.0
|
USER ISSUE: cant start any save data - **Version:** 0.7.5.1 beta
**Steps to Reproduce:**
i try to load any game and nothing happens
**Expected behavior:**
play
**Actual behavior:**
nothing
|
non_process
|
user issue cant start any save data version beta steps to reproduce i try to load any game and nothing happens expected behavior play actual behavior nothing
| 0
|
3,676
| 6,711,771,538
|
IssuesEvent
|
2017-10-13 06:21:54
|
zeebe-io/zeebe
|
https://api.github.com/repos/zeebe-io/zeebe
|
closed
|
Test broker restart without index savepoints
|
broker ci stream processor
|
Write tests where the last log entries are API requests that no follow-up events point to (e.g. task locked events that are not completed yet) and where these events have not been indexed yet, such that the recovery logic processes them again.
Fix problems if any.
Should be both, integration and unit tested.
|
1.0
|
Test broker restart without index savepoints - Write tests where the last log entries are API requests that no follow-up events point to (e.g. task locked events that are not completed yet) and where these events have not been indexed yet, such that the recovery logic processes them again.
Fix problems if any.
Should be both, integration and unit tested.
|
process
|
test broker restart without index savepoints write tests where the last log entries are api requests that no follow up events point to e g task locked events that are not completed yet and where these events have not been indexed yet such that the recovery logic processes them again fix problems if any should be both integration and unit tested
| 1
|
87,686
| 17,360,181,345
|
IssuesEvent
|
2021-07-29 19:25:05
|
EDCD/EDMarketConnector
|
https://api.github.com/repos/EDCD/EDMarketConnector
|
closed
|
Treat Journal event names case-insensitively
|
code cleanup enhancement
|
I know of no instance of two distinct Journal event names only differing in their case.
Things like `BackPack`, but `Backpack.json` are a pain to keep straight.
So we should refactor to specifically use a `lower()`'d copy of the event name in all matching code.
We should *not* change the actual event name, plugins get to decide how they treat it.
This entails making the `event_name = entry['name'].tolower()` and then using it instead in matches (maybe in logging too?).
This issue is the latest in the series entitled "Ath is getting sick of this shit" ;) .
|
1.0
|
Treat Journal event names case-insensitively - I know of no instance of two distinct Journal event names only differing in their case.
Things like `BackPack`, but `Backpack.json` are a pain to keep straight.
So we should refactor to specifically use a `lower()`'d copy of the event name in all matching code.
We should *not* change the actual event name, plugins get to decide how they treat it.
This entails making the `event_name = entry['name'].tolower()` and then using it instead in matches (maybe in logging too?).
This issue is the latest in the series entitled "Ath is getting sick of this shit" ;) .
|
non_process
|
treat journal event names case insensitively i know of no instance of two distinct journal event names only differing in their case things like backpack but backpack json are a pain to keep straight so we should refactor to specifically use a lower d copy of the event name in all matching code we should not change the actual event name plugins get to decide how they treat it this entails making the event name entry tolower and then using it instead in matches maybe in logging too this issue is the latest in the series entitled ath is getting sick of this shit
| 0
|
436,726
| 12,552,221,505
|
IssuesEvent
|
2020-06-06 17:21:57
|
Eastrall/Rhisis
|
https://api.github.com/repos/Eastrall/Rhisis
|
closed
|
Disconnecting while dead makes you a zombie
|
bug priority: critical srv: world sys: battle v0.4.x
|
# :beetle: Bug Report
<!-- Enter rhisis version here with format "vX.Y.Z" -->
**Rhisis version:** v0.4.x
## Expected Behavior
<!--- Tell us what should happen -->
When disconnecting when dead and relogging afterwards, the player should revive at the nearest resurection point.
## Current Behavior
<!--- Tell us what happens instead of the expected behavior -->
The player is standing at the place he died and has 0 HP.
## Steps to Reproduce
<!--- Provide a link to a live example, or an unambiguous set of steps to -->
<!--- reproduce this bug. Include code to reproduce, if relevant -->
1. Die in battle versus a monster
2. Disconnect
3. Reconnect
|
1.0
|
Disconnecting while dead makes you a zombie - # :beetle: Bug Report
<!-- Enter rhisis version here with format "vX.Y.Z" -->
**Rhisis version:** v0.4.x
## Expected Behavior
<!--- Tell us what should happen -->
When disconnecting when dead and relogging afterwards, the player should revive at the nearest resurection point.
## Current Behavior
<!--- Tell us what happens instead of the expected behavior -->
The player is standing at the place he died and has 0 HP.
## Steps to Reproduce
<!--- Provide a link to a live example, or an unambiguous set of steps to -->
<!--- reproduce this bug. Include code to reproduce, if relevant -->
1. Die in battle versus a monster
2. Disconnect
3. Reconnect
|
non_process
|
disconnecting while dead makes you a zombie beetle bug report rhisis version x expected behavior when disconnecting when dead and relogging afterwards the player should revive at the nearest resurection point current behavior the player is standing at the place he died and has hp steps to reproduce die in battle versus a monster disconnect reconnect
| 0
|
25,597
| 4,163,400,411
|
IssuesEvent
|
2016-06-18 03:02:21
|
alexrj/Slic3r
|
https://api.github.com/repos/alexrj/Slic3r
|
closed
|
stable giving segfault error
|
Needs testing with current dev version or next release
|
on mint 16, 64, running stable 64.
Unknown option interface_shells
The program 'slic3r' received an X Window System error.
This probably reflects a bug in the program.
The error was 'RenderBadFormat (invalid PictFormat parameter)'.
(Details: serial 50139 error_code 142 request_code 139 minor_code 2)
(Note to programmers: normally, X errors are reported asynchronously;
that is, you will receive the error a while after causing it.
To debug your program, run it with the --sync command line
option to change this behavior. You can then get a meaningful
backtrace from your debugger if you break on the gdk_x_error() function.)
The program 'slic3r' received an X Window System error.
This probably reflects a bug in the program.
The error was 'BadRequest (invalid request code or no such operation)'.
(Details: serial 50158 error_code 1 request_code 255 minor_code 0)
(Note to programmers: normally, X errors are reported asynchronously;
that is, you will receive the error a while after causing it.
To debug your program, run it with the --sync command line
option to change this behavior. You can then get a meaningful
backtrace from your debugger if you break on the gdk_x_error() function.)
Segmentation fault
This happened when exporting to g-code, although no gcode or gcode file was generated.
config.ini is
# generated by Slic3r 1.1.0 on Thu Apr 3 22:30:38 2014
avoid_crossing_perimeters = 1
bed_size = 284.2,301.88
bed_temperature = 75
bottom_solid_layers = 2
bridge_acceleration = 0
bridge_fan_speed = 100
bridge_flow_ratio = 1
bridge_speed = 40
brim_width = 0
complete_objects = 0
cooling = 1
default_acceleration = 0
disable_fan_first_layers = 1
duplicate_distance = 6
end_gcode = G92 E0; reset extruder\nG1 E-1 F1000; retract extruder to prevent final string\nG1 E-2 F1000; retract extruder to prevent final string\nG1 E-3 F1000; retract extruder to prevent final string\nG1 E-4 F1000; retract extruder to prevent final string\nG1 E-5 F1000; retract extruder to prevent final string\nT0;\nM104 S0 ; turn off temperature\nT1;\nM104 S0 ; turn off temperature\nT2;\nM104 S0 ; turn off temperature\nT3;\nM104 S0 ; turn off temperature\nT4;\nM104 S0 ; turn off temperature\nM140 S0 ; Turn off heated print bed\nG28 X0 ; home X axis\nM84 ; disable motors
external_perimeter_speed = 40
external_perimeters_first = 0
extra_perimeters = 1
extruder_clearance_height = 20
extruder_clearance_radius = 20
extruder_offset = 10x10,72.95x12.05,9.35x37.7,73.18x38.65,41.6x38.25
extrusion_axis = E
extrusion_multiplier = 1
extrusion_width = 0
fan_always_on = 0
fan_below_layer_time = 60
filament_diameter = 1.8
fill_angle = 45
fill_density = 30%
fill_pattern = rectilinear
first_layer_acceleration = 0
first_layer_bed_temperature = 75
first_layer_extrusion_width = 175%
first_layer_height = 0.3
first_layer_speed = 20
first_layer_temperature = 215
g0 = 0
gap_fill_speed = 40
gcode_arcs = 0
gcode_comments = 0
gcode_flavor = reprap
infill_acceleration = 0
infill_every_layers = 1
infill_extruder = 1
infill_extrusion_width = 0
infill_first = 1
infill_only_where_needed = 0
infill_speed = 40
interface_shells = 0
layer_gcode =
layer_height = 0.3
max_fan_speed = 100
min_fan_speed = 35
min_print_speed = 5
min_skirt_length = 0
notes =
nozzle_diameter = 0.5,0.5,0.5,0.5,0.5
only_retract_when_crossing_perimeters = 1
ooze_prevention = 1
output_filename_format = [input_filename_base].gcode
overhangs = 1
perimeter_acceleration = 0
perimeter_extruder = 1
perimeter_extrusion_width = 0
perimeter_speed = 40
perimeters = 3
post_process =
print_center = 142.1,150.94
raft_layers = 0
randomize_start = 0
resolution = 0
retract_before_travel = 20,20,20,20,20
retract_layer_change = 0,0,0,0,0
retract_length = 26,26,26,26,21
retract_length_toolchange = 26,26,26,26,21
retract_lift = 0,0,0,0,0
retract_restart_extra = 0,0,0,0,0
retract_restart_extra_toolchange = 0.5,1,0,0,0
retract_speed = 5,5,3,5,3
skirt_distance = 10
skirt_height = 1
skirts = 2
slowdown_below_layer_time = 30
small_perimeter_speed = 40
solid_fill_pattern = rectilinear
solid_infill_below_area = 1
solid_infill_every_layers = 0
solid_infill_extrusion_width = 0
solid_infill_speed = 40
spiral_vase = 0
standby_temperature_delta = -1
start_gcode = M106 S255; turn fans on\nG28 ; home all axes\nT0; //set the 1st extruder to be active
start_perimeters_at_concave_points = 0
start_perimeters_at_non_overhang = 0
support_material = 0
support_material_angle = 0
support_material_enforce_layers = 0
support_material_extruder = 1
support_material_extrusion_width = 0
support_material_interface_extruder = 1
support_material_interface_layers = 1
support_material_interface_spacing = 2.5
support_material_pattern = rectilinear
support_material_spacing = 4
support_material_speed = 40
support_material_threshold = 0
temperature = 215
thin_walls = 1
threads = 4
toolchange_gcode =
top_infill_extrusion_width = 0
top_solid_infill_speed = 40
top_solid_layers = 3
travel_speed = 12
use_firmware_retraction = 0
use_relative_e_distances = 0
vibration_limit = 0
wipe = 1,1,1,1,1
z_offset = 0
Making this test object
http://www.thingiverse.com/thing:38096
|
1.0
|
stable giving segfault error - on mint 16, 64, running stable 64.
Unknown option interface_shells
The program 'slic3r' received an X Window System error.
This probably reflects a bug in the program.
The error was 'RenderBadFormat (invalid PictFormat parameter)'.
(Details: serial 50139 error_code 142 request_code 139 minor_code 2)
(Note to programmers: normally, X errors are reported asynchronously;
that is, you will receive the error a while after causing it.
To debug your program, run it with the --sync command line
option to change this behavior. You can then get a meaningful
backtrace from your debugger if you break on the gdk_x_error() function.)
The program 'slic3r' received an X Window System error.
This probably reflects a bug in the program.
The error was 'BadRequest (invalid request code or no such operation)'.
(Details: serial 50158 error_code 1 request_code 255 minor_code 0)
(Note to programmers: normally, X errors are reported asynchronously;
that is, you will receive the error a while after causing it.
To debug your program, run it with the --sync command line
option to change this behavior. You can then get a meaningful
backtrace from your debugger if you break on the gdk_x_error() function.)
Segmentation fault
This happened when exporting to g-code, although no gcode or gcode file was generated.
config.ini is
# generated by Slic3r 1.1.0 on Thu Apr 3 22:30:38 2014
avoid_crossing_perimeters = 1
bed_size = 284.2,301.88
bed_temperature = 75
bottom_solid_layers = 2
bridge_acceleration = 0
bridge_fan_speed = 100
bridge_flow_ratio = 1
bridge_speed = 40
brim_width = 0
complete_objects = 0
cooling = 1
default_acceleration = 0
disable_fan_first_layers = 1
duplicate_distance = 6
end_gcode = G92 E0; reset extruder\nG1 E-1 F1000; retract extruder to prevent final string\nG1 E-2 F1000; retract extruder to prevent final string\nG1 E-3 F1000; retract extruder to prevent final string\nG1 E-4 F1000; retract extruder to prevent final string\nG1 E-5 F1000; retract extruder to prevent final string\nT0;\nM104 S0 ; turn off temperature\nT1;\nM104 S0 ; turn off temperature\nT2;\nM104 S0 ; turn off temperature\nT3;\nM104 S0 ; turn off temperature\nT4;\nM104 S0 ; turn off temperature\nM140 S0 ; Turn off heated print bed\nG28 X0 ; home X axis\nM84 ; disable motors
external_perimeter_speed = 40
external_perimeters_first = 0
extra_perimeters = 1
extruder_clearance_height = 20
extruder_clearance_radius = 20
extruder_offset = 10x10,72.95x12.05,9.35x37.7,73.18x38.65,41.6x38.25
extrusion_axis = E
extrusion_multiplier = 1
extrusion_width = 0
fan_always_on = 0
fan_below_layer_time = 60
filament_diameter = 1.8
fill_angle = 45
fill_density = 30%
fill_pattern = rectilinear
first_layer_acceleration = 0
first_layer_bed_temperature = 75
first_layer_extrusion_width = 175%
first_layer_height = 0.3
first_layer_speed = 20
first_layer_temperature = 215
g0 = 0
gap_fill_speed = 40
gcode_arcs = 0
gcode_comments = 0
gcode_flavor = reprap
infill_acceleration = 0
infill_every_layers = 1
infill_extruder = 1
infill_extrusion_width = 0
infill_first = 1
infill_only_where_needed = 0
infill_speed = 40
interface_shells = 0
layer_gcode =
layer_height = 0.3
max_fan_speed = 100
min_fan_speed = 35
min_print_speed = 5
min_skirt_length = 0
notes =
nozzle_diameter = 0.5,0.5,0.5,0.5,0.5
only_retract_when_crossing_perimeters = 1
ooze_prevention = 1
output_filename_format = [input_filename_base].gcode
overhangs = 1
perimeter_acceleration = 0
perimeter_extruder = 1
perimeter_extrusion_width = 0
perimeter_speed = 40
perimeters = 3
post_process =
print_center = 142.1,150.94
raft_layers = 0
randomize_start = 0
resolution = 0
retract_before_travel = 20,20,20,20,20
retract_layer_change = 0,0,0,0,0
retract_length = 26,26,26,26,21
retract_length_toolchange = 26,26,26,26,21
retract_lift = 0,0,0,0,0
retract_restart_extra = 0,0,0,0,0
retract_restart_extra_toolchange = 0.5,1,0,0,0
retract_speed = 5,5,3,5,3
skirt_distance = 10
skirt_height = 1
skirts = 2
slowdown_below_layer_time = 30
small_perimeter_speed = 40
solid_fill_pattern = rectilinear
solid_infill_below_area = 1
solid_infill_every_layers = 0
solid_infill_extrusion_width = 0
solid_infill_speed = 40
spiral_vase = 0
standby_temperature_delta = -1
start_gcode = M106 S255; turn fans on\nG28 ; home all axes\nT0; //set the 1st extruder to be active
start_perimeters_at_concave_points = 0
start_perimeters_at_non_overhang = 0
support_material = 0
support_material_angle = 0
support_material_enforce_layers = 0
support_material_extruder = 1
support_material_extrusion_width = 0
support_material_interface_extruder = 1
support_material_interface_layers = 1
support_material_interface_spacing = 2.5
support_material_pattern = rectilinear
support_material_spacing = 4
support_material_speed = 40
support_material_threshold = 0
temperature = 215
thin_walls = 1
threads = 4
toolchange_gcode =
top_infill_extrusion_width = 0
top_solid_infill_speed = 40
top_solid_layers = 3
travel_speed = 12
use_firmware_retraction = 0
use_relative_e_distances = 0
vibration_limit = 0
wipe = 1,1,1,1,1
z_offset = 0
Making this test object
http://www.thingiverse.com/thing:38096
|
non_process
|
stable giving segfault error on mint running stable unknown option interface shells the program received an x window system error this probably reflects a bug in the program the error was renderbadformat invalid pictformat parameter details serial error code request code minor code note to programmers normally x errors are reported asynchronously that is you will receive the error a while after causing it to debug your program run it with the sync command line option to change this behavior you can then get a meaningful backtrace from your debugger if you break on the gdk x error function the program received an x window system error this probably reflects a bug in the program the error was badrequest invalid request code or no such operation details serial error code request code minor code note to programmers normally x errors are reported asynchronously that is you will receive the error a while after causing it to debug your program run it with the sync command line option to change this behavior you can then get a meaningful backtrace from your debugger if you break on the gdk x error function segmentation fault this happened when exporting to g code although no gcode or gcode file was generated config ini is generated by on thu apr avoid crossing perimeters bed size bed temperature bottom solid layers bridge acceleration bridge fan speed bridge flow ratio bridge speed brim width complete objects cooling default acceleration disable fan first layers duplicate distance end gcode reset extruder e retract extruder to prevent final string e retract extruder to prevent final string e retract extruder to prevent final string e retract extruder to prevent final string e retract extruder to prevent final string turn off temperature turn off temperature turn off temperature turn off temperature turn off temperature turn off heated print bed home x axis disable motors external perimeter speed external perimeters first extra perimeters extruder clearance height extruder clearance radius extruder offset extrusion axis e extrusion multiplier extrusion width fan always on fan below layer time filament diameter fill angle fill density fill pattern rectilinear first layer acceleration first layer bed temperature first layer extrusion width first layer height first layer speed first layer temperature gap fill speed gcode arcs gcode comments gcode flavor reprap infill acceleration infill every layers infill extruder infill extrusion width infill first infill only where needed infill speed interface shells layer gcode layer height max fan speed min fan speed min print speed min skirt length notes nozzle diameter only retract when crossing perimeters ooze prevention output filename format gcode overhangs perimeter acceleration perimeter extruder perimeter extrusion width perimeter speed perimeters post process print center raft layers randomize start resolution retract before travel retract layer change retract length retract length toolchange retract lift retract restart extra retract restart extra toolchange retract speed skirt distance skirt height skirts slowdown below layer time small perimeter speed solid fill pattern rectilinear solid infill below area solid infill every layers solid infill extrusion width solid infill speed spiral vase standby temperature delta start gcode turn fans on home all axes set the extruder to be active start perimeters at concave points start perimeters at non overhang support material support material angle support material enforce layers support material extruder support material extrusion width support material interface extruder support material interface layers support material interface spacing support material pattern rectilinear support material spacing support material speed support material threshold temperature thin walls threads toolchange gcode top infill extrusion width top solid infill speed top solid layers travel speed use firmware retraction use relative e distances vibration limit wipe z offset making this test object
| 0
|
37,788
| 8,518,431,208
|
IssuesEvent
|
2018-11-01 11:35:29
|
jccastillo0007/eFacturaT
|
https://api.github.com/repos/jccastillo0007/eFacturaT
|
opened
|
Cancelación escritorio - una factura cancelada previamente no le cambia el status
|
bug defect
|
Cuando existe una factura que ha sido cancelada previamente, si te lo indica como respuesta.
Lo que no hace es cambiar el status
|
1.0
|
Cancelación escritorio - una factura cancelada previamente no le cambia el status - Cuando existe una factura que ha sido cancelada previamente, si te lo indica como respuesta.
Lo que no hace es cambiar el status
|
non_process
|
cancelación escritorio una factura cancelada previamente no le cambia el status cuando existe una factura que ha sido cancelada previamente si te lo indica como respuesta lo que no hace es cambiar el status
| 0
|
20,582
| 27,243,704,477
|
IssuesEvent
|
2023-02-21 23:05:56
|
open-telemetry/opentelemetry-collector-contrib
|
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
|
closed
|
k8sattributes tests failing
|
processor/k8sattributes needs triage
|
### Component(s)
processor/k8sattributes
### Describe the issue you're reporting
e2e tests for kubernetes are failing. See
https://github.com/open-telemetry/opentelemetry-collector-contrib/actions/runs/4236650784/jobs/7361708703#step:24:16
Likely related to https://github.com/open-telemetry/opentelemetry-collector-contrib/pull/18767
|
1.0
|
k8sattributes tests failing - ### Component(s)
processor/k8sattributes
### Describe the issue you're reporting
e2e tests for kubernetes are failing. See
https://github.com/open-telemetry/opentelemetry-collector-contrib/actions/runs/4236650784/jobs/7361708703#step:24:16
Likely related to https://github.com/open-telemetry/opentelemetry-collector-contrib/pull/18767
|
process
|
tests failing component s processor describe the issue you re reporting tests for kubernetes are failing see likely related to
| 1
|
9,059
| 12,133,623,008
|
IssuesEvent
|
2020-04-23 09:20:34
|
arcus-azure/arcus.messaging
|
https://api.github.com/repos/arcus-azure/arcus.messaging
|
closed
|
Provide more thorough integration scenarios for message pump
|
area:message-processing enhancement
|
Provide more thorough integration scenarios for message pump which interacts more with the messages such as abandon, complete, et al.
|
1.0
|
Provide more thorough integration scenarios for message pump - Provide more thorough integration scenarios for message pump which interacts more with the messages such as abandon, complete, et al.
|
process
|
provide more thorough integration scenarios for message pump provide more thorough integration scenarios for message pump which interacts more with the messages such as abandon complete et al
| 1
|
7,655
| 10,740,449,797
|
IssuesEvent
|
2019-10-29 18:12:32
|
allinurl/goaccess
|
https://api.github.com/repos/allinurl/goaccess
|
closed
|
[Help]Is there a way to only view data from a set range date?
|
duplicate log-processing
|
I want to be able to see how many "200" status codes in a month. Is there any way for me to get that kind of data? I saw there was a date parameter but it looks like it only takes one date and not a date range.
|
1.0
|
[Help]Is there a way to only view data from a set range date? - I want to be able to see how many "200" status codes in a month. Is there any way for me to get that kind of data? I saw there was a date parameter but it looks like it only takes one date and not a date range.
|
process
|
is there a way to only view data from a set range date i want to be able to see how many status codes in a month is there any way for me to get that kind of data i saw there was a date parameter but it looks like it only takes one date and not a date range
| 1
|
73,768
| 7,354,595,618
|
IssuesEvent
|
2018-03-09 07:42:48
|
AAA-Intelligence/eve
|
https://api.github.com/repos/AAA-Intelligence/eve
|
opened
|
[TF-ANT-2] Generated Answer is shown in the UI #2
|
test
|
Die Antwort des Bots soll auf der Benutzeroberfläche angezeigt werden.
|
1.0
|
[TF-ANT-2] Generated Answer is shown in the UI #2 - Die Antwort des Bots soll auf der Benutzeroberfläche angezeigt werden.
|
non_process
|
generated answer is shown in the ui die antwort des bots soll auf der benutzeroberfläche angezeigt werden
| 0
|
16,966
| 22,330,800,193
|
IssuesEvent
|
2022-06-14 14:22:53
|
UserOfficeProject/user-office-project-issue-tracker
|
https://api.github.com/repos/UserOfficeProject/user-office-project-issue-tracker
|
closed
|
Train Emma Gozzard to modify the questionnaire
|
type: process area: uop/stfc
|
Train Emma Gozzard on how to modify the questionnaire. Training for creation of it can be done at a later date.
We may want to include other people in this training, but we can ask Emma who she would like included.
We probably want this to be one-to-one training rather than documentation, as written documentation will likely be time consuming and not as effective.
|
1.0
|
Train Emma Gozzard to modify the questionnaire - Train Emma Gozzard on how to modify the questionnaire. Training for creation of it can be done at a later date.
We may want to include other people in this training, but we can ask Emma who she would like included.
We probably want this to be one-to-one training rather than documentation, as written documentation will likely be time consuming and not as effective.
|
process
|
train emma gozzard to modify the questionnaire train emma gozzard on how to modify the questionnaire training for creation of it can be done at a later date we may want to include other people in this training but we can ask emma who she would like included we probably want this to be one to one training rather than documentation as written documentation will likely be time consuming and not as effective
| 1
|
10,340
| 13,169,446,102
|
IssuesEvent
|
2020-08-11 13:45:53
|
MicrosoftDocs/azure-devops-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
|
closed
|
If condition with variable missing on template conditions page
|
Pri1 devops-cicd-process/tech devops/prod doc-bug
|
I read the documentation on this page and didn't need parameters but variables. It was really confusing that there is no example on this page - so my intuition was that if checks with variables are not possible when including templates.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 6724abea-bbdc-bf66-ed5e-3214fa6c3e66
* Version Independent ID: 4f8dab21-3f0e-da32-cc0e-1d85c13c0065
* Content: [Templates - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/templates?view=azure-devops)
* Content Source: [docs/pipelines/process/templates.md](https://github.com/MicrosoftDocs/vsts-docs/blob/master/docs/pipelines/process/templates.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
1.0
|
If condition with variable missing on template conditions page - I read the documentation on this page and didn't need parameters but variables. It was really confusing that there is no example on this page - so my intuition was that if checks with variables are not possible when including templates.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 6724abea-bbdc-bf66-ed5e-3214fa6c3e66
* Version Independent ID: 4f8dab21-3f0e-da32-cc0e-1d85c13c0065
* Content: [Templates - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/templates?view=azure-devops)
* Content Source: [docs/pipelines/process/templates.md](https://github.com/MicrosoftDocs/vsts-docs/blob/master/docs/pipelines/process/templates.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
process
|
if condition with variable missing on template conditions page i read the documentation on this page and didn t need parameters but variables it was really confusing that there is no example on this page so my intuition was that if checks with variables are not possible when including templates document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id bbdc version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
| 1
|
338,663
| 30,312,864,613
|
IssuesEvent
|
2023-07-10 13:52:14
|
AIDASoft/DD4hep
|
https://api.github.com/repos/AIDASoft/DD4hep
|
closed
|
Ambiguous assignment operator in Geant4Output2EDM4hep.cpp on LCG macOS platforms
|
Testing Fixed Waiting for caller
|
Hello,
Building **master** of DD4Hep we encountered a compilation error related to an ambiguous assignment operator on LCG macOS platforms. The error occurs in the following code snippet:
```
/Users/sftnight/build/workspace/lcg_nightly_pipeline/build/frameworks/DD4hep-master/src/DD4hep/master/DDG4/edm4hep/Geant4Output2EDM4hep.cpp:476:45: error: use of overloaded operator '=' is ambiguous (with operand types 'std::map<std::string, std::pair<edm4hep::SimCalorimeterHitCollection, edm4hep::CaloHitContributionCollection>>::mapped_type' (aka 'std::pair<edm4hep::SimCalorimeterHitCollection, edm4hep::CaloHitContributionCollection>') and 'void')
auto& hits = m_calorimeterHits[colName] = {};
~~~~~~~~~~~~~~~~~~~~~~~~~~ ^ ~~
/Library/Developer/CommandLineTools/SDKs/MacOSX13.3.sdk/usr/include/c++/v1/__utility/pair.h:262:11: note: candidate function
pair& operator=(typename conditional<
^
/Library/Developer/CommandLineTools/SDKs/MacOSX13.3.sdk/usr/include/c++/v1/__utility/pair.h:275:11: note: candidate function
pair& operator=(typename conditional<
^
/Library/Developer/CommandLineTools/SDKs/MacOSX13.3.sdk/usr/include/c++/v1/__utility/pair.h:40:29: note: candidate function (the implicit copy assignment operator) has been implicitly deleted
struct _LIBCPP_TEMPLATE_VIS pair
^
```
https://github.com/AIDASoft/DD4hep/blob/3347996361696054fc32b1515bc277be731c724f/DDG4/edm4hep/Geant4Output2EDM4hep.cpp#L476
|
1.0
|
Ambiguous assignment operator in Geant4Output2EDM4hep.cpp on LCG macOS platforms - Hello,
Building **master** of DD4Hep we encountered a compilation error related to an ambiguous assignment operator on LCG macOS platforms. The error occurs in the following code snippet:
```
/Users/sftnight/build/workspace/lcg_nightly_pipeline/build/frameworks/DD4hep-master/src/DD4hep/master/DDG4/edm4hep/Geant4Output2EDM4hep.cpp:476:45: error: use of overloaded operator '=' is ambiguous (with operand types 'std::map<std::string, std::pair<edm4hep::SimCalorimeterHitCollection, edm4hep::CaloHitContributionCollection>>::mapped_type' (aka 'std::pair<edm4hep::SimCalorimeterHitCollection, edm4hep::CaloHitContributionCollection>') and 'void')
auto& hits = m_calorimeterHits[colName] = {};
~~~~~~~~~~~~~~~~~~~~~~~~~~ ^ ~~
/Library/Developer/CommandLineTools/SDKs/MacOSX13.3.sdk/usr/include/c++/v1/__utility/pair.h:262:11: note: candidate function
pair& operator=(typename conditional<
^
/Library/Developer/CommandLineTools/SDKs/MacOSX13.3.sdk/usr/include/c++/v1/__utility/pair.h:275:11: note: candidate function
pair& operator=(typename conditional<
^
/Library/Developer/CommandLineTools/SDKs/MacOSX13.3.sdk/usr/include/c++/v1/__utility/pair.h:40:29: note: candidate function (the implicit copy assignment operator) has been implicitly deleted
struct _LIBCPP_TEMPLATE_VIS pair
^
```
https://github.com/AIDASoft/DD4hep/blob/3347996361696054fc32b1515bc277be731c724f/DDG4/edm4hep/Geant4Output2EDM4hep.cpp#L476
|
non_process
|
ambiguous assignment operator in cpp on lcg macos platforms hello building master of we encountered a compilation error related to an ambiguous assignment operator on lcg macos platforms the error occurs in the following code snippet users sftnight build workspace lcg nightly pipeline build frameworks master src master cpp error use of overloaded operator is ambiguous with operand types std map mapped type aka std pair and void auto hits m calorimeterhits library developer commandlinetools sdks sdk usr include c utility pair h note candidate function pair operator typename conditional library developer commandlinetools sdks sdk usr include c utility pair h note candidate function pair operator typename conditional library developer commandlinetools sdks sdk usr include c utility pair h note candidate function the implicit copy assignment operator has been implicitly deleted struct libcpp template vis pair
| 0
|
155,694
| 12,267,364,354
|
IssuesEvent
|
2020-05-07 10:32:41
|
hyperledger-labs/blockchain-automation-framework
|
https://api.github.com/repos/hyperledger-labs/blockchain-automation-framework
|
closed
|
Add default molecule test scenario in Corda
|
corda molecule-test
|
### Description
As a developer I want to run default molecule test so that I can test my ansible scripts.
### Acceptance Criteria
1.Molecule tests should cover `create/storageclass` and `create/namespace_serviceaccount` and related roles
2. Verification step should be configured.
3. Tests should complete successfully
|
1.0
|
Add default molecule test scenario in Corda - ### Description
As a developer I want to run default molecule test so that I can test my ansible scripts.
### Acceptance Criteria
1.Molecule tests should cover `create/storageclass` and `create/namespace_serviceaccount` and related roles
2. Verification step should be configured.
3. Tests should complete successfully
|
non_process
|
add default molecule test scenario in corda description as a developer i want to run default molecule test so that i can test my ansible scripts acceptance criteria molecule tests should cover create storageclass and create namespace serviceaccount and related roles verification step should be configured tests should complete successfully
| 0
|
430,825
| 12,466,571,289
|
IssuesEvent
|
2020-05-28 15:40:32
|
Damgt-Studios/GolemGame
|
https://api.github.com/repos/Damgt-Studios/GolemGame
|
closed
|
FBX Animation Data Exporter
|
Priority: High Type: Tools / Support
|
- [ ] Make an FBX Exporter that has mesh data in the wobj animation mesh format
- [ ] Make an adanim file that contains animation clip data in the format described in the following filespec
## Notes
- Binormals don't need to be in the filespec, the layout will be different regardless so it can be excluded here
# Wobject Filespec
A 3d file format created by Joseph Whittignton to store 3d mesh, material & animation data into one compact binary file.
# File composition
The file is composed of a header, vertex data, index, data, animation keyframe data and then bind pose data.
```
[Header]
[VertexData]
[IndexData]
[KeyframeData]
[BindPoseData]
```
## Header
This header defines the metadata for the binary data encoded into the file
```C++
// Mesh header struct
struct MeshHeader
{
int indexcount, vertexcount;
int indexstart, vertexstart;
int animation_start, animation_count;
int joint_count, keyframe_count;
char t_diffuse[256];
char t_normal[256];
char t_mixed[256];
};
```
## Mesh Data
The mesh data is defined using int32s as indices and the following structure defines the data for each vertex
```C++
struct
{
float pos[3];
float Normal[3];
float Tangent[3];
int joints[4];
float weights[4];
float tex[2];
};
```
The file will contain the all the vertex data in order then all the indices in order.
## Animation Data
The animation data is defined using the following structures
```C++
struct joint
{
float transform[16];
int parent_index;
};
typedef float transform[16]
struct keyframe
{
float time;
joints[j]
}
```
The animation data contains only one onimation clip. The clip length is first written as a 32 bit float then for each keyframe, the keyframe time is written as a 32 bit float then an array of joint structs is written and this continues until all keyframes have been written.
The bindpose data is then written as an array of 32 bit floats representing the joint transforms for the bind pose.
The animation data is written as the following. n is the number of keyframes, j is the number of joints
```
[animation_clip_duration]
[keyframe_array[n]]
[transform[j]]
```
|
1.0
|
FBX Animation Data Exporter - - [ ] Make an FBX Exporter that has mesh data in the wobj animation mesh format
- [ ] Make an adanim file that contains animation clip data in the format described in the following filespec
## Notes
- Binormals don't need to be in the filespec, the layout will be different regardless so it can be excluded here
# Wobject Filespec
A 3d file format created by Joseph Whittignton to store 3d mesh, material & animation data into one compact binary file.
# File composition
The file is composed of a header, vertex data, index, data, animation keyframe data and then bind pose data.
```
[Header]
[VertexData]
[IndexData]
[KeyframeData]
[BindPoseData]
```
## Header
This header defines the metadata for the binary data encoded into the file
```C++
// Mesh header struct
struct MeshHeader
{
int indexcount, vertexcount;
int indexstart, vertexstart;
int animation_start, animation_count;
int joint_count, keyframe_count;
char t_diffuse[256];
char t_normal[256];
char t_mixed[256];
};
```
## Mesh Data
The mesh data is defined using int32s as indices and the following structure defines the data for each vertex
```C++
struct
{
float pos[3];
float Normal[3];
float Tangent[3];
int joints[4];
float weights[4];
float tex[2];
};
```
The file will contain the all the vertex data in order then all the indices in order.
## Animation Data
The animation data is defined using the following structures
```C++
struct joint
{
float transform[16];
int parent_index;
};
typedef float transform[16]
struct keyframe
{
float time;
joints[j]
}
```
The animation data contains only one onimation clip. The clip length is first written as a 32 bit float then for each keyframe, the keyframe time is written as a 32 bit float then an array of joint structs is written and this continues until all keyframes have been written.
The bindpose data is then written as an array of 32 bit floats representing the joint transforms for the bind pose.
The animation data is written as the following. n is the number of keyframes, j is the number of joints
```
[animation_clip_duration]
[keyframe_array[n]]
[transform[j]]
```
|
non_process
|
fbx animation data exporter make an fbx exporter that has mesh data in the wobj animation mesh format make an adanim file that contains animation clip data in the format described in the following filespec notes binormals don t need to be in the filespec the layout will be different regardless so it can be excluded here wobject filespec a file format created by joseph whittignton to store mesh material animation data into one compact binary file file composition the file is composed of a header vertex data index data animation keyframe data and then bind pose data header this header defines the metadata for the binary data encoded into the file c mesh header struct struct meshheader int indexcount vertexcount int indexstart vertexstart int animation start animation count int joint count keyframe count char t diffuse char t normal char t mixed mesh data the mesh data is defined using as indices and the following structure defines the data for each vertex c struct float pos float normal float tangent int joints float weights float tex the file will contain the all the vertex data in order then all the indices in order animation data the animation data is defined using the following structures c struct joint float transform int parent index typedef float transform struct keyframe float time joints the animation data contains only one onimation clip the clip length is first written as a bit float then for each keyframe the keyframe time is written as a bit float then an array of joint structs is written and this continues until all keyframes have been written the bindpose data is then written as an array of bit floats representing the joint transforms for the bind pose the animation data is written as the following n is the number of keyframes j is the number of joints
| 0
|
21,755
| 30,274,142,678
|
IssuesEvent
|
2023-07-07 17:57:28
|
core-wg/oscore-key-update
|
https://api.github.com/repos/core-wg/oscore-key-update
|
closed
|
Considerations on nonce lengths
|
KUDOS Processing
|
Motivate recommended nonce lengths, and tradeoffs if using a nonce length less than 8 (which is recommended).
Should we allow a zero-length nonce? In either case do not disallow a small nonce size, it can be used as long as awareness exists about potential security implications.
|
1.0
|
Considerations on nonce lengths - Motivate recommended nonce lengths, and tradeoffs if using a nonce length less than 8 (which is recommended).
Should we allow a zero-length nonce? In either case do not disallow a small nonce size, it can be used as long as awareness exists about potential security implications.
|
process
|
considerations on nonce lengths motivate recommended nonce lengths and tradeoffs if using a nonce length less than which is recommended should we allow a zero length nonce in either case do not disallow a small nonce size it can be used as long as awareness exists about potential security implications
| 1
|
127,024
| 5,011,809,654
|
IssuesEvent
|
2016-12-13 09:19:09
|
openvstorage/framework
|
https://api.github.com/repos/openvstorage/framework
|
closed
|
Remove the vpool if the volumedriver is not running
|
priority_normal state_verification type_enhancement
|
We should be able to remove the vpool if the vpool is not running. We experienced this error during a test:
```
2016-08-24 17:29:03 43500 +0200 - stor-04.be-g8-3 - 4854/139859400984384 - celery/celery.worker.job - 930 - ERROR - Task ovs.storagerouter.remove_storagedriver[7c3022c9-0447-4a78-aa1e-029b1b6dcaab] raised unexpected: RuntimeError('Not all StorageDrivers are reachable, please (re)start them and try again',)
Traceback (most recent call last):
File "/usr/lib/python2.7/dist-packages/celery/app/trace.py", line 240, in trace_task
R = retval = fun(*args, **kwargs)
File "/usr/lib/python2.7/dist-packages/celery/app/trace.py", line 438, in __protected_call__
return self.run(*args, **kwargs)
File "/opt/OpenvStorage/ovs/lib/storagerouter.py", line 1102, in remove_storagedriver
raise RuntimeError('Not all StorageDrivers are reachable, please (re)start them and try again')
RuntimeError: Not all StorageDrivers are reachable, please (re)start them and try again
```
|
1.0
|
Remove the vpool if the volumedriver is not running - We should be able to remove the vpool if the vpool is not running. We experienced this error during a test:
```
2016-08-24 17:29:03 43500 +0200 - stor-04.be-g8-3 - 4854/139859400984384 - celery/celery.worker.job - 930 - ERROR - Task ovs.storagerouter.remove_storagedriver[7c3022c9-0447-4a78-aa1e-029b1b6dcaab] raised unexpected: RuntimeError('Not all StorageDrivers are reachable, please (re)start them and try again',)
Traceback (most recent call last):
File "/usr/lib/python2.7/dist-packages/celery/app/trace.py", line 240, in trace_task
R = retval = fun(*args, **kwargs)
File "/usr/lib/python2.7/dist-packages/celery/app/trace.py", line 438, in __protected_call__
return self.run(*args, **kwargs)
File "/opt/OpenvStorage/ovs/lib/storagerouter.py", line 1102, in remove_storagedriver
raise RuntimeError('Not all StorageDrivers are reachable, please (re)start them and try again')
RuntimeError: Not all StorageDrivers are reachable, please (re)start them and try again
```
|
non_process
|
remove the vpool if the volumedriver is not running we should be able to remove the vpool if the vpool is not running we experienced this error during a test stor be celery celery worker job error task ovs storagerouter remove storagedriver raised unexpected runtimeerror not all storagedrivers are reachable please re start them and try again traceback most recent call last file usr lib dist packages celery app trace py line in trace task r retval fun args kwargs file usr lib dist packages celery app trace py line in protected call return self run args kwargs file opt openvstorage ovs lib storagerouter py line in remove storagedriver raise runtimeerror not all storagedrivers are reachable please re start them and try again runtimeerror not all storagedrivers are reachable please re start them and try again
| 0
|
5,703
| 8,564,049,582
|
IssuesEvent
|
2018-11-09 15:43:23
|
mitre/inspec_tools
|
https://api.github.com/repos/mitre/inspec_tools
|
opened
|
add rubocop integration or PRs
|
public-release release_process
|
add rubocop testing and rake testing as part of the PR process
|
1.0
|
add rubocop integration or PRs - add rubocop testing and rake testing as part of the PR process
|
process
|
add rubocop integration or prs add rubocop testing and rake testing as part of the pr process
| 1
|
154,881
| 24,362,978,003
|
IssuesEvent
|
2022-10-03 13:13:17
|
equinor/design-system
|
https://api.github.com/repos/equinor/design-system
|
closed
|
Compact density for pagination
|
storefront core-react design figma storybook
|
We need to look into how the Pagination is solved and how we can handle several digit numbers. We use buttons today. Find examples in use in apps today.
|
1.0
|
Compact density for pagination - We need to look into how the Pagination is solved and how we can handle several digit numbers. We use buttons today. Find examples in use in apps today.
|
non_process
|
compact density for pagination we need to look into how the pagination is solved and how we can handle several digit numbers we use buttons today find examples in use in apps today
| 0
|
99,795
| 8,711,814,330
|
IssuesEvent
|
2018-12-06 20:18:22
|
morissonmaciel/vscode-tsc-compiler
|
https://api.github.com/repos/morissonmaciel/vscode-tsc-compiler
|
closed
|
Windows only?
|
testing
|
Am I correct in that this extension is for Windows only (or not tested in any other platforms)? It can't find the global installation of Typescript on a Mac and running a local installation in node_modules fails because it uses a Windows-style backslash directory separators (`\`) instead of Unix-style forward slashes (`/`). It tries to find `/node_modules/.bin\tsc` and fails because `\t` is interpreted as a control character.
If this is Windows only, it might be reasonable to mention it in the readme.
|
1.0
|
Windows only? - Am I correct in that this extension is for Windows only (or not tested in any other platforms)? It can't find the global installation of Typescript on a Mac and running a local installation in node_modules fails because it uses a Windows-style backslash directory separators (`\`) instead of Unix-style forward slashes (`/`). It tries to find `/node_modules/.bin\tsc` and fails because `\t` is interpreted as a control character.
If this is Windows only, it might be reasonable to mention it in the readme.
|
non_process
|
windows only am i correct in that this extension is for windows only or not tested in any other platforms it can t find the global installation of typescript on a mac and running a local installation in node modules fails because it uses a windows style backslash directory separators instead of unix style forward slashes it tries to find node modules bin tsc and fails because t is interpreted as a control character if this is windows only it might be reasonable to mention it in the readme
| 0
|
32,297
| 13,784,972,458
|
IssuesEvent
|
2020-10-08 21:51:16
|
MicrosoftDocs/powerbi-docs
|
https://api.github.com/repos/MicrosoftDocs/powerbi-docs
|
closed
|
Where do you set the enhanced compute engine at dataflow level?
|
Pri2 doc-enhancement powerbi-service/subsvc powerbi/svc question
|
How do you access the screen shown by the screenshot under "Enable DirectQuery for dataflows"? I have the enhanced compute engine turned on at capacity level but cannot find this option to set the enhanced compute engine at dataflow level - it needs to be explained where this is accessed.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 369c84b0-97ef-a750-86cd-5330626c75ac
* Version Independent ID: 3af79c25-1e82-ab08-7604-bfaa9ace6128
* Content: [Use the enhanced compute engine with dataflows - Power BI](https://docs.microsoft.com/en-us/power-bi/transform-model/service-dataflows-enhanced-compute-engine)
* Content Source: [powerbi-docs/transform-model/service-dataflows-enhanced-compute-engine.md](https://github.com/MicrosoftDocs/powerbi-docs/blob/live/powerbi-docs/transform-model/service-dataflows-enhanced-compute-engine.md)
* Service: **powerbi**
* Sub-service: **powerbi-service**
* GitHub Login: @davidiseminger
* Microsoft Alias: **davidi**
|
1.0
|
Where do you set the enhanced compute engine at dataflow level? - How do you access the screen shown by the screenshot under "Enable DirectQuery for dataflows"? I have the enhanced compute engine turned on at capacity level but cannot find this option to set the enhanced compute engine at dataflow level - it needs to be explained where this is accessed.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 369c84b0-97ef-a750-86cd-5330626c75ac
* Version Independent ID: 3af79c25-1e82-ab08-7604-bfaa9ace6128
* Content: [Use the enhanced compute engine with dataflows - Power BI](https://docs.microsoft.com/en-us/power-bi/transform-model/service-dataflows-enhanced-compute-engine)
* Content Source: [powerbi-docs/transform-model/service-dataflows-enhanced-compute-engine.md](https://github.com/MicrosoftDocs/powerbi-docs/blob/live/powerbi-docs/transform-model/service-dataflows-enhanced-compute-engine.md)
* Service: **powerbi**
* Sub-service: **powerbi-service**
* GitHub Login: @davidiseminger
* Microsoft Alias: **davidi**
|
non_process
|
where do you set the enhanced compute engine at dataflow level how do you access the screen shown by the screenshot under enable directquery for dataflows i have the enhanced compute engine turned on at capacity level but cannot find this option to set the enhanced compute engine at dataflow level it needs to be explained where this is accessed document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service powerbi sub service powerbi service github login davidiseminger microsoft alias davidi
| 0
|
185,347
| 6,721,137,993
|
IssuesEvent
|
2017-10-16 10:28:03
|
CS2103AUG2017-T17-B1/main
|
https://api.github.com/repos/CS2103AUG2017-T17-B1/main
|
closed
|
As a clumsy user I want to have automatic backups
|
priority.medium
|
so that I can not worry about accidentally deleting things
|
1.0
|
As a clumsy user I want to have automatic backups - so that I can not worry about accidentally deleting things
|
non_process
|
as a clumsy user i want to have automatic backups so that i can not worry about accidentally deleting things
| 0
|
20,902
| 27,741,879,672
|
IssuesEvent
|
2023-03-15 14:46:23
|
bazelbuild/bazel
|
https://api.github.com/repos/bazelbuild/bazel
|
closed
|
GitHub codeowners group not working
|
P2 type: process team-OSS
|
It looks like the rules apple team who is supposed to be an owner of some of the objc logic is not working correctly
https://github.com/bazelbuild/bazel/pull/14844#issuecomment-1061239960
You can see github reporting errors on https://github.com/bazelbuild/bazel/pull/14614/files
I believe it's because the team doesn't include the bazel repo https://github.com/orgs/bazelbuild/teams/rules_apple-team/repositories, and I wonder what level of access the team would have to have for this to work, if just "read" would be enough or not.
@aiuto can you debug? I assume you need admin access to fix anything here
|
1.0
|
GitHub codeowners group not working - It looks like the rules apple team who is supposed to be an owner of some of the objc logic is not working correctly
https://github.com/bazelbuild/bazel/pull/14844#issuecomment-1061239960
You can see github reporting errors on https://github.com/bazelbuild/bazel/pull/14614/files
I believe it's because the team doesn't include the bazel repo https://github.com/orgs/bazelbuild/teams/rules_apple-team/repositories, and I wonder what level of access the team would have to have for this to work, if just "read" would be enough or not.
@aiuto can you debug? I assume you need admin access to fix anything here
|
process
|
github codeowners group not working it looks like the rules apple team who is supposed to be an owner of some of the objc logic is not working correctly you can see github reporting errors on i believe it s because the team doesn t include the bazel repo and i wonder what level of access the team would have to have for this to work if just read would be enough or not aiuto can you debug i assume you need admin access to fix anything here
| 1
|
1,890
| 4,716,816,061
|
IssuesEvent
|
2016-10-16 08:41:39
|
symfony/symfony
|
https://api.github.com/repos/symfony/symfony
|
closed
|
Sub process output not being returned 'real-time'
|
Bug Process Status: Needs Review Unconfirmed
|
I'm just testing out the Process class and it doesn't seem to do what it says on the tin.
Docs state
> When executing a long running command (like rsync-ing files to a remote server), you can give feedback to the end user in real-time by passing an anonymous function to the run() method:
I have 1 file with the following
```php
$process = new Process('php terminal.php');
$process->setTimeout(null);
$process->run( function ( $type, $buffer ) {
if (Process::ERR === $type) {
echo 'ERR > '.$buffer;
} else {
echo 'OUT > '.$buffer;
}
});
```
and another with the following
```php
$x = 10;
while( $x ) {
echo "{$x}\n";
sleep(1);
$x--;
}
```
While running this via terminal after a **10 second wait** the output is
```cli
OUT > 10
OUT > 9
OUT > 8
OUT > 7
OUT > 6
OUT > 5
OUT > 4
OUT > 3
OUT > 2
OUT > 1
```
which shows there are 10 iterations of closure above but there is no output until the process is finished.
Am I missing something here?
Regards, Luke
|
1.0
|
Sub process output not being returned 'real-time' - I'm just testing out the Process class and it doesn't seem to do what it says on the tin.
Docs state
> When executing a long running command (like rsync-ing files to a remote server), you can give feedback to the end user in real-time by passing an anonymous function to the run() method:
I have 1 file with the following
```php
$process = new Process('php terminal.php');
$process->setTimeout(null);
$process->run( function ( $type, $buffer ) {
if (Process::ERR === $type) {
echo 'ERR > '.$buffer;
} else {
echo 'OUT > '.$buffer;
}
});
```
and another with the following
```php
$x = 10;
while( $x ) {
echo "{$x}\n";
sleep(1);
$x--;
}
```
While running this via terminal after a **10 second wait** the output is
```cli
OUT > 10
OUT > 9
OUT > 8
OUT > 7
OUT > 6
OUT > 5
OUT > 4
OUT > 3
OUT > 2
OUT > 1
```
which shows there are 10 iterations of closure above but there is no output until the process is finished.
Am I missing something here?
Regards, Luke
|
process
|
sub process output not being returned real time i m just testing out the process class and it doesn t seem to do what it says on the tin docs state when executing a long running command like rsync ing files to a remote server you can give feedback to the end user in real time by passing an anonymous function to the run method i have file with the following php process new process php terminal php process settimeout null process run function type buffer if process err type echo err buffer else echo out buffer and another with the following php x while x echo x n sleep x while running this via terminal after a second wait the output is cli out out out out out out out out out out which shows there are iterations of closure above but there is no output until the process is finished am i missing something here regards luke
| 1
|
6,781
| 9,916,317,701
|
IssuesEvent
|
2019-06-28 19:33:02
|
HumanCellAtlas/dcp-community
|
https://api.github.com/repos/HumanCellAtlas/dcp-community
|
closed
|
The template for renaming approved RFCs is incorrect.
|
rfc-process
|
Reported by @mweiden
#### Author(s):
- Rename the RFC from `0000-my-feature.md` to `rfc####-my-feature.md` (with leading zeros) where `####` is the next available RFC number
should be:
- Rename the RFC from `0000-my-feature.md` to `####-my-feature.md` (with leading zeros) where `####` is the next available RFC number
|
1.0
|
The template for renaming approved RFCs is incorrect. - Reported by @mweiden
#### Author(s):
- Rename the RFC from `0000-my-feature.md` to `rfc####-my-feature.md` (with leading zeros) where `####` is the next available RFC number
should be:
- Rename the RFC from `0000-my-feature.md` to `####-my-feature.md` (with leading zeros) where `####` is the next available RFC number
|
process
|
the template for renaming approved rfcs is incorrect reported by mweiden author s rename the rfc from my feature md to rfc my feature md with leading zeros where is the next available rfc number should be rename the rfc from my feature md to my feature md with leading zeros where is the next available rfc number
| 1
|
386,605
| 26,692,556,612
|
IssuesEvent
|
2023-01-27 07:10:45
|
OSCA-Kampala-Chapter/Windows-Musawo
|
https://api.github.com/repos/OSCA-Kampala-Chapter/Windows-Musawo
|
opened
|
Document Font styles used for the Design Assets
|
documentation
|
Document the Font styles used for all the Design Assets as we intend to have a common style used through out all our future designs. It'd be great to have all the design decisions well documented for future designers to follow.
- [ ] Add the name of the font style/typeface used in the [design assets README file]()
- [ ] Similarly, add the color codes used in the designs.
|
1.0
|
Document Font styles used for the Design Assets - Document the Font styles used for all the Design Assets as we intend to have a common style used through out all our future designs. It'd be great to have all the design decisions well documented for future designers to follow.
- [ ] Add the name of the font style/typeface used in the [design assets README file]()
- [ ] Similarly, add the color codes used in the designs.
|
non_process
|
document font styles used for the design assets document the font styles used for all the design assets as we intend to have a common style used through out all our future designs it d be great to have all the design decisions well documented for future designers to follow add the name of the font style typeface used in the similarly add the color codes used in the designs
| 0
|
1,621
| 4,236,291,028
|
IssuesEvent
|
2016-07-05 17:56:17
|
AnalyticalGraphicsInc/cesium
|
https://api.github.com/repos/AnalyticalGraphicsInc/cesium
|
closed
|
Cesium does not build with npm 1.3.10
|
dev process doc
|
There is no compilation error, but the build only takes half a minute and Cesium.js is missing from Build/Cesium. The [build guide](https://github.com/AnalyticalGraphicsInc/cesium/blob/master/Documentation/Contributors/BuildGuide/README.md) does not mention a minimal version of npm.
With npm 1.3.10
```
$ npm run minifyRelease
> cesium@1.23.0 minifyRelease /home/gberaudo/cesium
> gulp minifyRelease
[08:31:22] Using gulpfile ~/cesium/gulpfile.js
[08:31:22] Starting 'build'...
[08:31:23] Finished 'build' after 1.25 s
[08:31:23] Starting 'generateStubs'...
[08:31:24] Finished 'generateStubs' after 268 ms
[08:31:24] Starting 'minifyRelease'...
Building Cesium.js
Building Source/Workers/cesiumWorkerBootstrapper.js
Finished Cesium.js
Finished Source/Workers/cesiumWorkerBootstrapper.js
Building Source/ThirdParty/Workers/inflate.js
Finished Source/ThirdParty/Workers/inflate.js
...
[08:32:00] Finished 'minifyRelease' after 36 s
```
Then
$ ls Build/Cesium/
Assets Widgets
With npm 2.15.5
```
$ npm run minifyRelease
> cesium@1.23.0 minifyRelease /tmp/cesium
> gulp minifyRelease
[08:43:33] Using gulpfile /tmp/cesium/gulpfile.js
[08:43:33] Starting 'build'...
[08:43:33] Finished 'build' after 339 ms
[08:43:33] Starting 'generateStubs'...
[08:43:33] Finished 'generateStubs' after 46 ms
[08:43:33] Starting 'minifyRelease'...
Building Cesium.js
Building Source/Workers/cesiumWorkerBootstrapper.js
Building Source/Workers/transferTypedArrayTest.js
Building Source/ThirdParty/Workers/deflate.js
Building Source/ThirdParty/Workers/inflate.js
Finished Source/Workers/transferTypedArrayTest.js
Finished Source/Workers/cesiumWorkerBootstrapper.js
Finished Source/ThirdParty/Workers/deflate.js
Finished Source/ThirdParty/Workers/inflate.js
Building Source/Workers/combineGeometry.js
Building Source/Workers/createBoxGeometry.js
Building Source/Workers/createBoxOutlineGeometry.js
Building Source/Workers/createCircleGeometry.js
...
```
|
1.0
|
Cesium does not build with npm 1.3.10 - There is no compilation error, but the build only takes half a minute and Cesium.js is missing from Build/Cesium. The [build guide](https://github.com/AnalyticalGraphicsInc/cesium/blob/master/Documentation/Contributors/BuildGuide/README.md) does not mention a minimal version of npm.
With npm 1.3.10
```
$ npm run minifyRelease
> cesium@1.23.0 minifyRelease /home/gberaudo/cesium
> gulp minifyRelease
[08:31:22] Using gulpfile ~/cesium/gulpfile.js
[08:31:22] Starting 'build'...
[08:31:23] Finished 'build' after 1.25 s
[08:31:23] Starting 'generateStubs'...
[08:31:24] Finished 'generateStubs' after 268 ms
[08:31:24] Starting 'minifyRelease'...
Building Cesium.js
Building Source/Workers/cesiumWorkerBootstrapper.js
Finished Cesium.js
Finished Source/Workers/cesiumWorkerBootstrapper.js
Building Source/ThirdParty/Workers/inflate.js
Finished Source/ThirdParty/Workers/inflate.js
...
[08:32:00] Finished 'minifyRelease' after 36 s
```
Then
$ ls Build/Cesium/
Assets Widgets
With npm 2.15.5
```
$ npm run minifyRelease
> cesium@1.23.0 minifyRelease /tmp/cesium
> gulp minifyRelease
[08:43:33] Using gulpfile /tmp/cesium/gulpfile.js
[08:43:33] Starting 'build'...
[08:43:33] Finished 'build' after 339 ms
[08:43:33] Starting 'generateStubs'...
[08:43:33] Finished 'generateStubs' after 46 ms
[08:43:33] Starting 'minifyRelease'...
Building Cesium.js
Building Source/Workers/cesiumWorkerBootstrapper.js
Building Source/Workers/transferTypedArrayTest.js
Building Source/ThirdParty/Workers/deflate.js
Building Source/ThirdParty/Workers/inflate.js
Finished Source/Workers/transferTypedArrayTest.js
Finished Source/Workers/cesiumWorkerBootstrapper.js
Finished Source/ThirdParty/Workers/deflate.js
Finished Source/ThirdParty/Workers/inflate.js
Building Source/Workers/combineGeometry.js
Building Source/Workers/createBoxGeometry.js
Building Source/Workers/createBoxOutlineGeometry.js
Building Source/Workers/createCircleGeometry.js
...
```
|
process
|
cesium does not build with npm there is no compilation error but the build only takes half a minute and cesium js is missing from build cesium the does not mention a minimal version of npm with npm npm run minifyrelease cesium minifyrelease home gberaudo cesium gulp minifyrelease using gulpfile cesium gulpfile js starting build finished build after s starting generatestubs finished generatestubs after ms starting minifyrelease building cesium js building source workers cesiumworkerbootstrapper js finished cesium js finished source workers cesiumworkerbootstrapper js building source thirdparty workers inflate js finished source thirdparty workers inflate js finished minifyrelease after s then ls build cesium assets widgets with npm npm run minifyrelease cesium minifyrelease tmp cesium gulp minifyrelease using gulpfile tmp cesium gulpfile js starting build finished build after ms starting generatestubs finished generatestubs after ms starting minifyrelease building cesium js building source workers cesiumworkerbootstrapper js building source workers transfertypedarraytest js building source thirdparty workers deflate js building source thirdparty workers inflate js finished source workers transfertypedarraytest js finished source workers cesiumworkerbootstrapper js finished source thirdparty workers deflate js finished source thirdparty workers inflate js building source workers combinegeometry js building source workers createboxgeometry js building source workers createboxoutlinegeometry js building source workers createcirclegeometry js
| 1
|
99,101
| 20,883,167,587
|
IssuesEvent
|
2022-03-23 00:05:31
|
alefragnani/vscode-separators
|
https://api.github.com/repos/alefragnani/vscode-separators
|
closed
|
[FEATURE] - Support Settings Editor Extension Categories API
|
enhancement vscode api adoption
|
<!-- Please search existing issues to avoid creating duplicates. -->
<!-- Describe the feature you'd like. -->
https://code.visualstudio.com/updates/v1_61#_settings-editor-extension-categories
Possible Categories:
* Each kind of symbol
|
1.0
|
[FEATURE] - Support Settings Editor Extension Categories API - <!-- Please search existing issues to avoid creating duplicates. -->
<!-- Describe the feature you'd like. -->
https://code.visualstudio.com/updates/v1_61#_settings-editor-extension-categories
Possible Categories:
* Each kind of symbol
|
non_process
|
support settings editor extension categories api possible categories each kind of symbol
| 0
|
21,979
| 30,470,942,946
|
IssuesEvent
|
2023-07-17 13:36:33
|
USGS-WiM/StreamStats
|
https://api.github.com/repos/USGS-WiM/StreamStats
|
closed
|
Select server for Batch Processor Manage Queue tab
|
Batch Processor
|
Create a way for the user to specify which Batch Processor queue they would like to manage:
At least 2 options for this:
1. On the Manage Queue tab, add a drop down with options for Development Server or Production Server. When the user selects the server, the service URLs will need to be changed
2. Use separate URLs for each server: examples
- https://apps-int.usgs.gov/streamstats/dev/bp/
- https://apps-int.usgs.gov/streamstats/bp/
|
1.0
|
Select server for Batch Processor Manage Queue tab - Create a way for the user to specify which Batch Processor queue they would like to manage:
At least 2 options for this:
1. On the Manage Queue tab, add a drop down with options for Development Server or Production Server. When the user selects the server, the service URLs will need to be changed
2. Use separate URLs for each server: examples
- https://apps-int.usgs.gov/streamstats/dev/bp/
- https://apps-int.usgs.gov/streamstats/bp/
|
process
|
select server for batch processor manage queue tab create a way for the user to specify which batch processor queue they would like to manage at least options for this on the manage queue tab add a drop down with options for development server or production server when the user selects the server the service urls will need to be changed use separate urls for each server examples
| 1
|
11,476
| 14,343,515,246
|
IssuesEvent
|
2020-11-28 09:50:10
|
emacs-ess/ESS
|
https://api.github.com/repos/emacs-ess/ESS
|
closed
|
ess-eval-region-or-line-and-step : Failing in multiline cases in `*.Rd` \examples{..} section when "namespace mode"
|
process:eval
|
Related to #725 . Also related to the entry in its help
> Note that when inside a package and namespaced evaluation is in
> place (see `ess-r-set-evaluation-env'), the evaluation of multiline input will fail."
I find this very "unfriendly", and also wrong design: Inside the \example{} section I do *NOT* want namespaced evaluation.
I will have a version of the package to which the help page belongs *attach*ed in my `search()` path, and want the examples to behave as for a user; i.e., things should not happen in the package namespace, but all in globalenv.
|
1.0
|
ess-eval-region-or-line-and-step : Failing in multiline cases in `*.Rd` \examples{..} section when "namespace mode" - Related to #725 . Also related to the entry in its help
> Note that when inside a package and namespaced evaluation is in
> place (see `ess-r-set-evaluation-env'), the evaluation of multiline input will fail."
I find this very "unfriendly", and also wrong design: Inside the \example{} section I do *NOT* want namespaced evaluation.
I will have a version of the package to which the help page belongs *attach*ed in my `search()` path, and want the examples to behave as for a user; i.e., things should not happen in the package namespace, but all in globalenv.
|
process
|
ess eval region or line and step failing in multiline cases in rd examples section when namespace mode related to also related to the entry in its help note that when inside a package and namespaced evaluation is in place see ess r set evaluation env the evaluation of multiline input will fail i find this very unfriendly and also wrong design inside the example section i do not want namespaced evaluation i will have a version of the package to which the help page belongs attach ed in my search path and want the examples to behave as for a user i e things should not happen in the package namespace but all in globalenv
| 1
|
20,528
| 27,189,080,556
|
IssuesEvent
|
2023-02-19 15:27:50
|
gchq/stroom
|
https://api.github.com/repos/gchq/stroom
|
closed
|
Allow enumeration of meta keys in XSLT
|
enhancement f:processing
|
Currently there is no way of listing all meta values within an XSLT translation.
A new XSLT function should be created to allow a user to retrieve the distinct keys in a stream's meta and then get the value of each using `stroom:meta(.)`.
|
1.0
|
Allow enumeration of meta keys in XSLT - Currently there is no way of listing all meta values within an XSLT translation.
A new XSLT function should be created to allow a user to retrieve the distinct keys in a stream's meta and then get the value of each using `stroom:meta(.)`.
|
process
|
allow enumeration of meta keys in xslt currently there is no way of listing all meta values within an xslt translation a new xslt function should be created to allow a user to retrieve the distinct keys in a stream s meta and then get the value of each using stroom meta
| 1
|
23,119
| 15,832,103,515
|
IssuesEvent
|
2021-04-06 14:17:36
|
ILLIXR/ILLIXR
|
https://api.github.com/repos/ILLIXR/ILLIXR
|
closed
|
The Runner `clean` action should remove targets
|
bug infrastructure
|
- `RM` should be defined in `ILLIXR/common/common.mk`.
- `common` should be in the plugins list for `ILLIXR/configs/clean.yaml`.
|
1.0
|
The Runner `clean` action should remove targets - - `RM` should be defined in `ILLIXR/common/common.mk`.
- `common` should be in the plugins list for `ILLIXR/configs/clean.yaml`.
|
non_process
|
the runner clean action should remove targets rm should be defined in illixr common common mk common should be in the plugins list for illixr configs clean yaml
| 0
|
515,864
| 14,970,527,075
|
IssuesEvent
|
2021-01-27 19:44:59
|
wso2/product-is
|
https://api.github.com/repos/wso2/product-is
|
closed
|
About the Upgrading guide
|
Affected/5.2.0 Component/Migration Priority/Normal Severity/Minor docs
|
1. Lost some configuration file when copy `conf` floder
2. Should add a step to change the `Version` in `repository/conf/carbon.xml`
|
1.0
|
About the Upgrading guide - 1. Lost some configuration file when copy `conf` floder
2. Should add a step to change the `Version` in `repository/conf/carbon.xml`
|
non_process
|
about the upgrading guide lost some configuration file when copy conf floder should add a step to change the version in repository conf carbon xml
| 0
|
3,012
| 6,016,613,404
|
IssuesEvent
|
2017-06-07 07:30:55
|
openvstorage/framework
|
https://api.github.com/repos/openvstorage/framework
|
closed
|
2 Scrub roles got assigned by using the API
|
process_wontfix type_bug
|
Version: openvstorage 2.8.2-1
2 Scrub roles got assigned by using the API.
|
1.0
|
2 Scrub roles got assigned by using the API - Version: openvstorage 2.8.2-1
2 Scrub roles got assigned by using the API.
|
process
|
scrub roles got assigned by using the api version openvstorage scrub roles got assigned by using the api
| 1
|
20,299
| 26,937,849,144
|
IssuesEvent
|
2023-02-07 22:22:52
|
python/cpython
|
https://api.github.com/repos/python/cpython
|
closed
|
Multiprocessing shared memory duplicated with changed size
|
type-bug pending expert-multiprocessing
|
# What I do:
Inside class create shared mem and store it on the list:
```
self.inner_np_mem=[]
shm = shared_memory.SharedMemory(name='zxcv', create=True, size=r_arr.nbytes) #r_arr is numpy array
self.inner_np_mem.append(shm)
```
few lines down yet in the same function inside the same class I load the mem:
`shm = shared_memory.SharedMemory(name='zxcv')`
and compare to the stored one in the list self.inner_np_mem this is what I got:
```
print(190, 'shm.size',shm.size,len(shm.buf),shm.buf,self.inner_np_mem[0].buf)
print(191, 'shm.size',shm.size,len(shm.buf),shm ,self.inner_np_mem[0] )
```
> line 190 shm.size 81920 81920 <memory at 0x0000016EF3EA5400> <memory at 0x0000016EF3EA51C0>
> line 191 shm.size 81920 81920 SharedMemory('zxcv', size=81920) SharedMemory('zxcv', size=80000)
# Bug report
So I have 2 shared mem objects with the same name and different size.
# Your environment
Python 3.9.10
|
1.0
|
Multiprocessing shared memory duplicated with changed size - # What I do:
Inside class create shared mem and store it on the list:
```
self.inner_np_mem=[]
shm = shared_memory.SharedMemory(name='zxcv', create=True, size=r_arr.nbytes) #r_arr is numpy array
self.inner_np_mem.append(shm)
```
few lines down yet in the same function inside the same class I load the mem:
`shm = shared_memory.SharedMemory(name='zxcv')`
and compare to the stored one in the list self.inner_np_mem this is what I got:
```
print(190, 'shm.size',shm.size,len(shm.buf),shm.buf,self.inner_np_mem[0].buf)
print(191, 'shm.size',shm.size,len(shm.buf),shm ,self.inner_np_mem[0] )
```
> line 190 shm.size 81920 81920 <memory at 0x0000016EF3EA5400> <memory at 0x0000016EF3EA51C0>
> line 191 shm.size 81920 81920 SharedMemory('zxcv', size=81920) SharedMemory('zxcv', size=80000)
# Bug report
So I have 2 shared mem objects with the same name and different size.
# Your environment
Python 3.9.10
|
process
|
multiprocessing shared memory duplicated with changed size what i do inside class create shared mem and store it on the list self inner np mem shm shared memory sharedmemory name zxcv create true size r arr nbytes r arr is numpy array self inner np mem append shm few lines down yet in the same function inside the same class i load the mem shm shared memory sharedmemory name zxcv and compare to the stored one in the list self inner np mem this is what i got print shm size shm size len shm buf shm buf self inner np mem buf print shm size shm size len shm buf shm self inner np mem line shm size line shm size sharedmemory zxcv size sharedmemory zxcv size bug report so i have shared mem objects with the same name and different size your environment python
| 1
|
166,508
| 6,305,823,695
|
IssuesEvent
|
2017-07-21 19:20:33
|
craftercms/craftercms
|
https://api.github.com/repos/craftercms/craftercms
|
opened
|
[deployer] Change log level to debug for cache invalidation
|
enhancement Priority: Low
|
```
[INFO] 2017-07-21 19:14:32,994 [http-nio-8080-exec-8] [load-test] [rest.SiteCacheRestController] | Content cache and Freemarker cache have been cleared for site 'load-test'
```
|
1.0
|
[deployer] Change log level to debug for cache invalidation - ```
[INFO] 2017-07-21 19:14:32,994 [http-nio-8080-exec-8] [load-test] [rest.SiteCacheRestController] | Content cache and Freemarker cache have been cleared for site 'load-test'
```
|
non_process
|
change log level to debug for cache invalidation content cache and freemarker cache have been cleared for site load test
| 0
|
251,283
| 27,151,071,430
|
IssuesEvent
|
2023-02-17 01:20:15
|
EdHomeH/fitter-happier-docker
|
https://api.github.com/repos/EdHomeH/fitter-happier-docker
|
opened
|
CVE-2023-25577 (High) detected in Werkzeug-1.0.1-py2.py3-none-any.whl
|
security vulnerability
|
## CVE-2023-25577 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Werkzeug-1.0.1-py2.py3-none-any.whl</b></p></summary>
<p>The comprehensive WSGI web application library.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/cc/94/5f7079a0e00bd6863ef8f1da638721e9da21e5bacee597595b318f71d62e/Werkzeug-1.0.1-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/cc/94/5f7079a0e00bd6863ef8f1da638721e9da21e5bacee597595b318f71d62e/Werkzeug-1.0.1-py2.py3-none-any.whl</a></p>
<p>Path to dependency file: /web/requirements.txt</p>
<p>Path to vulnerable library: /web/requirements.txt</p>
<p>
Dependency Hierarchy:
- Flask-1.1.2-py2.py3-none-any.whl (Root Library)
- :x: **Werkzeug-1.0.1-py2.py3-none-any.whl** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Werkzeug is a comprehensive WSGI web application library. Prior to version 2.2.3, Werkzeug's multipart form data parser will parse an unlimited number of parts, including file parts. Parts can be a small amount of bytes, but each requires CPU time to parse and may use more memory as Python data. If a request can be made to an endpoint that accesses `request.data`, `request.form`, `request.files`, or `request.get_data(parse_form_data=False)`, it can cause unexpectedly high resource usage. This allows an attacker to cause a denial of service by sending crafted multipart data to an endpoint that will parse it. The amount of CPU time required can block worker processes from handling legitimate requests. The amount of RAM required can trigger an out of memory kill of the process. Unlimited file parts can use up memory and file handles. If many concurrent requests are sent continuously, this can exhaust or kill all available workers. Version 2.2.3 contains a patch for this issue.
<p>Publish Date: 2023-02-14
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-25577>CVE-2023-25577</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2023-25577">https://www.cve.org/CVERecord?id=CVE-2023-25577</a></p>
<p>Release Date: 2023-02-14</p>
<p>Fix Resolution: Werkzeug - 2.2.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2023-25577 (High) detected in Werkzeug-1.0.1-py2.py3-none-any.whl - ## CVE-2023-25577 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Werkzeug-1.0.1-py2.py3-none-any.whl</b></p></summary>
<p>The comprehensive WSGI web application library.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/cc/94/5f7079a0e00bd6863ef8f1da638721e9da21e5bacee597595b318f71d62e/Werkzeug-1.0.1-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/cc/94/5f7079a0e00bd6863ef8f1da638721e9da21e5bacee597595b318f71d62e/Werkzeug-1.0.1-py2.py3-none-any.whl</a></p>
<p>Path to dependency file: /web/requirements.txt</p>
<p>Path to vulnerable library: /web/requirements.txt</p>
<p>
Dependency Hierarchy:
- Flask-1.1.2-py2.py3-none-any.whl (Root Library)
- :x: **Werkzeug-1.0.1-py2.py3-none-any.whl** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Werkzeug is a comprehensive WSGI web application library. Prior to version 2.2.3, Werkzeug's multipart form data parser will parse an unlimited number of parts, including file parts. Parts can be a small amount of bytes, but each requires CPU time to parse and may use more memory as Python data. If a request can be made to an endpoint that accesses `request.data`, `request.form`, `request.files`, or `request.get_data(parse_form_data=False)`, it can cause unexpectedly high resource usage. This allows an attacker to cause a denial of service by sending crafted multipart data to an endpoint that will parse it. The amount of CPU time required can block worker processes from handling legitimate requests. The amount of RAM required can trigger an out of memory kill of the process. Unlimited file parts can use up memory and file handles. If many concurrent requests are sent continuously, this can exhaust or kill all available workers. Version 2.2.3 contains a patch for this issue.
<p>Publish Date: 2023-02-14
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-25577>CVE-2023-25577</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2023-25577">https://www.cve.org/CVERecord?id=CVE-2023-25577</a></p>
<p>Release Date: 2023-02-14</p>
<p>Fix Resolution: Werkzeug - 2.2.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in werkzeug none any whl cve high severity vulnerability vulnerable library werkzeug none any whl the comprehensive wsgi web application library library home page a href path to dependency file web requirements txt path to vulnerable library web requirements txt dependency hierarchy flask none any whl root library x werkzeug none any whl vulnerable library vulnerability details werkzeug is a comprehensive wsgi web application library prior to version werkzeug s multipart form data parser will parse an unlimited number of parts including file parts parts can be a small amount of bytes but each requires cpu time to parse and may use more memory as python data if a request can be made to an endpoint that accesses request data request form request files or request get data parse form data false it can cause unexpectedly high resource usage this allows an attacker to cause a denial of service by sending crafted multipart data to an endpoint that will parse it the amount of cpu time required can block worker processes from handling legitimate requests the amount of ram required can trigger an out of memory kill of the process unlimited file parts can use up memory and file handles if many concurrent requests are sent continuously this can exhaust or kill all available workers version contains a patch for this issue publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution werkzeug step up your open source security game with mend
| 0
|
20,682
| 27,354,410,637
|
IssuesEvent
|
2023-02-27 11:51:04
|
camunda/issues
|
https://api.github.com/repos/camunda/issues
|
opened
|
BPMN Signal Events(3): Broadcast signal event using throw signal event
|
component:desktopModeler component:operate component:optimize component:webModeler component:zeebe-process-automation public kind:epic feature-parity
|
### Value Proposition Statement
Use BPMN Throw Signal Events to easily start or continue instances that wait for a signal - without any coding.
### User Problem
Users can use BPMN Catch Signal Events (e.g. Start Event or Intermediate Events), but they have to be triggered via gRPC or one of our Clients. This means using Signal Events requires writing code and also different BPMN symbols have to be used than signals for throwing signals (e.g. I cannot use Signal Throw Events and attach a job worker, but I would have to use a Service Task instead).
### User Stories
I can model signal end event and intermediate signal throw events and linting works correctly.
I can deploy the models with such symbols to the engine and the engine triggers all signals correctly without me having to use the API.
I can see the symbols in other tools like Operate, Optimize.
### Implementation Notes
In the third stage, we'll increase support to all BPMN signal symbols. Specifically, it adds support for the Signal Intermediate Throw Event, and the Signal End Event.

Model highlighting that all the signal events will be supported at this stage, including the signal throw events
When a process instance arrives at a signal throw event, we'll broadcast a signal in the same way as the gateway can broadcast a signal: write a Signal:Broadcast command with relaying to the current partition. The implementation of stage 1 will then relay the command to other partitions, and all signal (start event) subscriptions will be triggered, without additional implementation efforts needed.
### Breakdown
> This section links to various sub-issues / -tasks contributing to respective epic phase or phase results where appropriate.
#### Discovery phase ##
<!-- Example: link to "Conduct customer interview with xyz" -->
#### Define phase ##
<!-- Consider: UI, UX, technical design, documentation design -->
<!-- Example: link to "Define User-Journey Flow" or "Define target architecture" -->
Design Planning
* Reviewed by design: {date}
* Designer assigned: {Yes, No Design Necessary, or No Designer Available}
* Assignee:
* Design Brief - {link to design brief }
* Research Brief - {link to research brief }
Design Deliverables
* {Deliverable Name} {Link to GH Issue}
Documentation Planning
<!-- Complex changes must be reviewed during the Define phase by the DRI of Documentation or technical writer. -->
<!-- Briefly describe the anticipated impact to documentation. -->
<!-- Example: "Creates structural changes in docs as UX is reworked." _Add docs reviewer to Epic for feedback._ -->
Risk Management <!-- add link to risk management issue -->
* Risk Class: <!-- e.g. very low | low | medium | high | very high -->
* Risk Treatment: <!-- e.g. avoid | mitigate | transfer | accept -->
#### Implement phase ##
<!-- Example: link to "Implement User Story xyz". Should not only include core implementation, but also documentation. -->
#### Validate phase ##
<!-- Example: link to "Evaluate usage data of last quarter" -->
### Links to additional collateral
<!-- Example: link to relevant support cases -->
|
1.0
|
BPMN Signal Events(3): Broadcast signal event using throw signal event - ### Value Proposition Statement
Use BPMN Throw Signal Events to easily start or continue instances that wait for a signal - without any coding.
### User Problem
Users can use BPMN Catch Signal Events (e.g. Start Event or Intermediate Events), but they have to be triggered via gRPC or one of our Clients. This means using Signal Events requires writing code and also different BPMN symbols have to be used than signals for throwing signals (e.g. I cannot use Signal Throw Events and attach a job worker, but I would have to use a Service Task instead).
### User Stories
I can model signal end event and intermediate signal throw events and linting works correctly.
I can deploy the models with such symbols to the engine and the engine triggers all signals correctly without me having to use the API.
I can see the symbols in other tools like Operate, Optimize.
### Implementation Notes
In the third stage, we'll increase support to all BPMN signal symbols. Specifically, it adds support for the Signal Intermediate Throw Event, and the Signal End Event.

Model highlighting that all the signal events will be supported at this stage, including the signal throw events
When a process instance arrives at a signal throw event, we'll broadcast a signal in the same way as the gateway can broadcast a signal: write a Signal:Broadcast command with relaying to the current partition. The implementation of stage 1 will then relay the command to other partitions, and all signal (start event) subscriptions will be triggered, without additional implementation efforts needed.
### Breakdown
> This section links to various sub-issues / -tasks contributing to respective epic phase or phase results where appropriate.
#### Discovery phase ##
<!-- Example: link to "Conduct customer interview with xyz" -->
#### Define phase ##
<!-- Consider: UI, UX, technical design, documentation design -->
<!-- Example: link to "Define User-Journey Flow" or "Define target architecture" -->
Design Planning
* Reviewed by design: {date}
* Designer assigned: {Yes, No Design Necessary, or No Designer Available}
* Assignee:
* Design Brief - {link to design brief }
* Research Brief - {link to research brief }
Design Deliverables
* {Deliverable Name} {Link to GH Issue}
Documentation Planning
<!-- Complex changes must be reviewed during the Define phase by the DRI of Documentation or technical writer. -->
<!-- Briefly describe the anticipated impact to documentation. -->
<!-- Example: "Creates structural changes in docs as UX is reworked." _Add docs reviewer to Epic for feedback._ -->
Risk Management <!-- add link to risk management issue -->
* Risk Class: <!-- e.g. very low | low | medium | high | very high -->
* Risk Treatment: <!-- e.g. avoid | mitigate | transfer | accept -->
#### Implement phase ##
<!-- Example: link to "Implement User Story xyz". Should not only include core implementation, but also documentation. -->
#### Validate phase ##
<!-- Example: link to "Evaluate usage data of last quarter" -->
### Links to additional collateral
<!-- Example: link to relevant support cases -->
|
process
|
bpmn signal events broadcast signal event using throw signal event value proposition statement use bpmn throw signal events to easily start or continue instances that wait for a signal without any coding user problem users can use bpmn catch signal events e g start event or intermediate events but they have to be triggered via grpc or one of our clients this means using signal events requires writing code and also different bpmn symbols have to be used than signals for throwing signals e g i cannot use signal throw events and attach a job worker but i would have to use a service task instead user stories i can model signal end event and intermediate signal throw events and linting works correctly i can deploy the models with such symbols to the engine and the engine triggers all signals correctly without me having to use the api i can see the symbols in other tools like operate optimize implementation notes in the third stage we ll increase support to all bpmn signal symbols specifically it adds support for the signal intermediate throw event and the signal end event model highlighting that all the signal events will be supported at this stage including the signal throw events when a process instance arrives at a signal throw event we ll broadcast a signal in the same way as the gateway can broadcast a signal write a signal broadcast command with relaying to the current partition the implementation of stage will then relay the command to other partitions and all signal start event subscriptions will be triggered without additional implementation efforts needed breakdown this section links to various sub issues tasks contributing to respective epic phase or phase results where appropriate discovery phase define phase design planning reviewed by design date designer assigned yes no design necessary or no designer available assignee design brief link to design brief research brief link to research brief design deliverables deliverable name link to gh issue documentation planning risk management risk class risk treatment implement phase validate phase links to additional collateral
| 1
|
5,548
| 8,393,232,952
|
IssuesEvent
|
2018-10-09 19:56:29
|
hashicorp/packer
|
https://api.github.com/repos/hashicorp/packer
|
closed
|
amazon-import: If the import job fails Packer wait until timeout instead of failing immediately
|
bug post-processor/amazon-import
|
There are three issues:
1) If the import job fails Packer still waits for success until the timeout is reached instead of detecting the failure and exit immediately.
2) Interrupt is not handled by the post-processor.
3) We should hardcode the `Format` to `OVA`
|
1.0
|
amazon-import: If the import job fails Packer wait until timeout instead of failing immediately - There are three issues:
1) If the import job fails Packer still waits for success until the timeout is reached instead of detecting the failure and exit immediately.
2) Interrupt is not handled by the post-processor.
3) We should hardcode the `Format` to `OVA`
|
process
|
amazon import if the import job fails packer wait until timeout instead of failing immediately there are three issues if the import job fails packer still waits for success until the timeout is reached instead of detecting the failure and exit immediately interrupt is not handled by the post processor we should hardcode the format to ova
| 1
|
848
| 3,315,586,554
|
IssuesEvent
|
2015-11-06 13:00:26
|
pwittchen/prefser
|
https://api.github.com/repos/pwittchen/prefser
|
closed
|
Release 2.0.2
|
release process
|
**Initial release notes**:
- fixed bug reported in issue #70: get(...) method now returns a null value instead of "null" string when setting default value to null of String type
- fixed RxJava usage in sample app
- fixed RxJava usage in code snippets in `README.md`
- changed code formatting to `SquareAndroid`
- added static code analysis
- improved code according to static code analysis suggestions
**Things to do**:
- [x] bump library version: PR #75
- [x] upload archives to Maven Central
- [x] close and release artifact on Maven Central
- [x] update `CHANGELOG.md` after Maven Sync
- [x] bump library version in `README.md`
- [x] create new GitHub release
|
1.0
|
Release 2.0.2 - **Initial release notes**:
- fixed bug reported in issue #70: get(...) method now returns a null value instead of "null" string when setting default value to null of String type
- fixed RxJava usage in sample app
- fixed RxJava usage in code snippets in `README.md`
- changed code formatting to `SquareAndroid`
- added static code analysis
- improved code according to static code analysis suggestions
**Things to do**:
- [x] bump library version: PR #75
- [x] upload archives to Maven Central
- [x] close and release artifact on Maven Central
- [x] update `CHANGELOG.md` after Maven Sync
- [x] bump library version in `README.md`
- [x] create new GitHub release
|
process
|
release initial release notes fixed bug reported in issue get method now returns a null value instead of null string when setting default value to null of string type fixed rxjava usage in sample app fixed rxjava usage in code snippets in readme md changed code formatting to squareandroid added static code analysis improved code according to static code analysis suggestions things to do bump library version pr upload archives to maven central close and release artifact on maven central update changelog md after maven sync bump library version in readme md create new github release
| 1
|
16,453
| 2,899,800,500
|
IssuesEvent
|
2015-06-17 13:33:43
|
hazelcast/hazelcast
|
https://api.github.com/repos/hazelcast/hazelcast
|
reopened
|
[TEST-FAILURE] RingbufferAddAllReadManyStressTest failures
|
Team: QuSP Type: Defect
|
com.hazelcast.ringbuffer.impl.RingbufferAddAllReadManyStressTest.whenNoTTL
com.hazelcast.ringbuffer.impl.RingbufferAddReadOneStressTest.whenLongTTLAndSmallBuffer
com.hazelcast.ringbuffer.impl.RingbufferAsyncAddWithBackoffStressTest.whenTTLEnabled
```
java.lang.AssertionError: No error should have been thrown, but ConsumeThread-1 completed error expected null, but was:<java.util.concurrent.ExecutionException: java.lang.IllegalArgumentException: sequence:1 is too large. The current tailSequence is:-1>
```
https://hazelcast-l337.ci.cloudbees.com/job/Hazelcast-3.x-nightly/563/#showFailuresLink
|
1.0
|
[TEST-FAILURE] RingbufferAddAllReadManyStressTest failures - com.hazelcast.ringbuffer.impl.RingbufferAddAllReadManyStressTest.whenNoTTL
com.hazelcast.ringbuffer.impl.RingbufferAddReadOneStressTest.whenLongTTLAndSmallBuffer
com.hazelcast.ringbuffer.impl.RingbufferAsyncAddWithBackoffStressTest.whenTTLEnabled
```
java.lang.AssertionError: No error should have been thrown, but ConsumeThread-1 completed error expected null, but was:<java.util.concurrent.ExecutionException: java.lang.IllegalArgumentException: sequence:1 is too large. The current tailSequence is:-1>
```
https://hazelcast-l337.ci.cloudbees.com/job/Hazelcast-3.x-nightly/563/#showFailuresLink
|
non_process
|
ringbufferaddallreadmanystresstest failures com hazelcast ringbuffer impl ringbufferaddallreadmanystresstest whennottl com hazelcast ringbuffer impl ringbufferaddreadonestresstest whenlongttlandsmallbuffer com hazelcast ringbuffer impl ringbufferasyncaddwithbackoffstresstest whenttlenabled java lang assertionerror no error should have been thrown but consumethread completed error expected null but was
| 0
|
769,723
| 27,017,805,503
|
IssuesEvent
|
2023-02-10 21:15:44
|
DDMAL/CantusDB
|
https://api.github.com/repos/DDMAL/CantusDB
|
closed
|
OldCantus Chant Detail page now has "view facsimile" link in sidebar
|
priority: medium
|
(See red rectangle, below). Points to the same URL as the Image Link (blue rectangle), at least on this page.

|
1.0
|
OldCantus Chant Detail page now has "view facsimile" link in sidebar - (See red rectangle, below). Points to the same URL as the Image Link (blue rectangle), at least on this page.

|
non_process
|
oldcantus chant detail page now has view facsimile link in sidebar see red rectangle below points to the same url as the image link blue rectangle at least on this page
| 0
|
48,440
| 20,152,517,993
|
IssuesEvent
|
2022-02-09 13:45:51
|
elastic/kibana
|
https://api.github.com/repos/elastic/kibana
|
closed
|
[Screenshotting] Limit number of concurrently running Chromium instances
|
Feature:Reporting loe:days Team:AppServicesSv impact:medium Team:Reporting Services
|
Currently, if we run the `getScreenshot` method twice, we will likely get two Chromium instances. That may cause some performance issues, especially in an environment running with low resources.
The number of concurrent calls should be artificially limited from the code per process. It can be done by using Mutex (or Semaphore).
|
2.0
|
[Screenshotting] Limit number of concurrently running Chromium instances - Currently, if we run the `getScreenshot` method twice, we will likely get two Chromium instances. That may cause some performance issues, especially in an environment running with low resources.
The number of concurrent calls should be artificially limited from the code per process. It can be done by using Mutex (or Semaphore).
|
non_process
|
limit number of concurrently running chromium instances currently if we run the getscreenshot method twice we will likely get two chromium instances that may cause some performance issues especially in an environment running with low resources the number of concurrent calls should be artificially limited from the code per process it can be done by using mutex or semaphore
| 0
|
440,140
| 12,693,890,979
|
IssuesEvent
|
2020-06-22 04:59:12
|
zeebe-io/zeebe
|
https://api.github.com/repos/zeebe-io/zeebe
|
opened
|
Do not send Swim requests to local node
|
Priority: High Status: Ready Type: Maintenance
|
**Description**
Currently Swim sends probe requests and other requests to the local node, since it tries to compare the local id with the member id. The problem is that we use the bootstrap service, which assigns each node an unique id, which is not the same as our local node id.
|
1.0
|
Do not send Swim requests to local node - **Description**
Currently Swim sends probe requests and other requests to the local node, since it tries to compare the local id with the member id. The problem is that we use the bootstrap service, which assigns each node an unique id, which is not the same as our local node id.
|
non_process
|
do not send swim requests to local node description currently swim sends probe requests and other requests to the local node since it tries to compare the local id with the member id the problem is that we use the bootstrap service which assigns each node an unique id which is not the same as our local node id
| 0
|
63,963
| 14,656,816,586
|
IssuesEvent
|
2020-12-28 14:15:48
|
fu1771695yongxie/odoo
|
https://api.github.com/repos/fu1771695yongxie/odoo
|
opened
|
CVE-2019-19911 (High) detected in Pillow-5.4.1-cp27-cp27mu-manylinux1_x86_64.whl
|
security vulnerability
|
## CVE-2019-19911 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Pillow-5.4.1-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary>
<p>Python Imaging Library (Fork)</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/0d/f3/421598450cb9503f4565d936860763b5af413a61009d87a5ab1e34139672/Pillow-5.4.1-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/0d/f3/421598450cb9503f4565d936860763b5af413a61009d87a5ab1e34139672/Pillow-5.4.1-cp27-cp27mu-manylinux1_x86_64.whl</a></p>
<p>Path to dependency file: odoo</p>
<p>Path to vulnerable library: odoo,odoo/requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **Pillow-5.4.1-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/fu1771695yongxie/odoo/commit/d2f4766e5250a653951e7d3f6813f6da8973da84">d2f4766e5250a653951e7d3f6813f6da8973da84</a></p>
<p>Found in base branch: <b>14.0</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
There is a DoS vulnerability in Pillow before 6.2.2 caused by FpxImagePlugin.py calling the range function on an unvalidated 32-bit integer if the number of bands is large. On Windows running 32-bit Python, this results in an OverflowError or MemoryError due to the 2 GB limit. However, on Linux running 64-bit Python this results in the process being terminated by the OOM killer.
<p>Publish Date: 2020-01-05
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-19911>CVE-2019-19911</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/python-pillow/Pillow/tree/6.2.2">https://github.com/python-pillow/Pillow/tree/6.2.2</a></p>
<p>Release Date: 2020-01-05</p>
<p>Fix Resolution: Pillow - 6.2.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2019-19911 (High) detected in Pillow-5.4.1-cp27-cp27mu-manylinux1_x86_64.whl - ## CVE-2019-19911 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Pillow-5.4.1-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary>
<p>Python Imaging Library (Fork)</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/0d/f3/421598450cb9503f4565d936860763b5af413a61009d87a5ab1e34139672/Pillow-5.4.1-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/0d/f3/421598450cb9503f4565d936860763b5af413a61009d87a5ab1e34139672/Pillow-5.4.1-cp27-cp27mu-manylinux1_x86_64.whl</a></p>
<p>Path to dependency file: odoo</p>
<p>Path to vulnerable library: odoo,odoo/requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **Pillow-5.4.1-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/fu1771695yongxie/odoo/commit/d2f4766e5250a653951e7d3f6813f6da8973da84">d2f4766e5250a653951e7d3f6813f6da8973da84</a></p>
<p>Found in base branch: <b>14.0</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
There is a DoS vulnerability in Pillow before 6.2.2 caused by FpxImagePlugin.py calling the range function on an unvalidated 32-bit integer if the number of bands is large. On Windows running 32-bit Python, this results in an OverflowError or MemoryError due to the 2 GB limit. However, on Linux running 64-bit Python this results in the process being terminated by the OOM killer.
<p>Publish Date: 2020-01-05
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-19911>CVE-2019-19911</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/python-pillow/Pillow/tree/6.2.2">https://github.com/python-pillow/Pillow/tree/6.2.2</a></p>
<p>Release Date: 2020-01-05</p>
<p>Fix Resolution: Pillow - 6.2.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in pillow whl cve high severity vulnerability vulnerable library pillow whl python imaging library fork library home page a href path to dependency file odoo path to vulnerable library odoo odoo requirements txt dependency hierarchy x pillow whl vulnerable library found in head commit a href found in base branch vulnerability details there is a dos vulnerability in pillow before caused by fpximageplugin py calling the range function on an unvalidated bit integer if the number of bands is large on windows running bit python this results in an overflowerror or memoryerror due to the gb limit however on linux running bit python this results in the process being terminated by the oom killer publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution pillow step up your open source security game with whitesource
| 0
|
531,123
| 15,441,423,926
|
IssuesEvent
|
2021-03-08 05:52:38
|
Nelson-Gon/urlfix
|
https://api.github.com/repos/Nelson-Gon/urlfix
|
opened
|
Some links in MarkDown files are not replaced
|
bug good first issue help wanted priority:high
|
**Describe the bug**
For markdown in the format `[[]]()()`, links in the second `()` are not replaced.
**To Reproduce**
Run `replace_urls` on Markdown files.
**Expected behavior**
I expected an output file with all links replaced.
**Unexpected behavior**
Only links in the first `()` are matched and replaced.
**System Details**
0.2.1 Dev: https://github.com/Nelson-Gon/urlfix/tree/4212bad4503cca5ae53f694c7bf623ca6d66fbf5/urlfix
OS independent.
---
TODO:
- [x] Ensure text is kept
- [ ] Match all URLs for Markdown in the form `[[]]()()`
- [ ] Add tests for known string existence
|
1.0
|
Some links in MarkDown files are not replaced - **Describe the bug**
For markdown in the format `[[]]()()`, links in the second `()` are not replaced.
**To Reproduce**
Run `replace_urls` on Markdown files.
**Expected behavior**
I expected an output file with all links replaced.
**Unexpected behavior**
Only links in the first `()` are matched and replaced.
**System Details**
0.2.1 Dev: https://github.com/Nelson-Gon/urlfix/tree/4212bad4503cca5ae53f694c7bf623ca6d66fbf5/urlfix
OS independent.
---
TODO:
- [x] Ensure text is kept
- [ ] Match all URLs for Markdown in the form `[[]]()()`
- [ ] Add tests for known string existence
|
non_process
|
some links in markdown files are not replaced describe the bug for markdown in the format links in the second are not replaced to reproduce run replace urls on markdown files expected behavior i expected an output file with all links replaced unexpected behavior only links in the first are matched and replaced system details dev os independent todo ensure text is kept match all urls for markdown in the form add tests for known string existence
| 0
|
7,479
| 10,571,644,491
|
IssuesEvent
|
2019-10-07 07:43:20
|
Altinn/altinn-studio
|
https://api.github.com/repos/Altinn/altinn-studio
|
closed
|
Analyze and define process handling and authorization
|
analysis process
|
## Description
As part of #243 we have defined that process API is separate from data API.
This means that moving the process forward would trigger a separate API call.
This API calls need to be authorized.
The big question is. What kind of authorization "action" is Next on the process. In current platform we always move the process part of the main action on the task.
If you do sign on the signing step we authorize for sign and move forward in the same operation.
As I see it there is two options here
## Introduce a common "action" for Next used for all steps
In this scenario we can define the "Next" operation in the XACML rule file and for all process changes we authorize user for "Next" operation on any type of task
### Pro
- Makes it more flexible to add new types of task
- Can configure PEP to use on action
### Con
- More rules in policy file
- Makes delegation more complex
## A logic in Next operation that identifies the action based on type of Task we currently are in
### Pro
- Less rules
- Matches more current solution
### Cons
- Add complexity to Next operation
- Limits flexibility on process
- Cant use standard PEP
We need to decide
## Conclusion
## Tasks
- [ ] Analyze and discuss
- [ ] Decide
- [ ] Update relevant issues in backlog
|
1.0
|
Analyze and define process handling and authorization - ## Description
As part of #243 we have defined that process API is separate from data API.
This means that moving the process forward would trigger a separate API call.
This API calls need to be authorized.
The big question is. What kind of authorization "action" is Next on the process. In current platform we always move the process part of the main action on the task.
If you do sign on the signing step we authorize for sign and move forward in the same operation.
As I see it there is two options here
## Introduce a common "action" for Next used for all steps
In this scenario we can define the "Next" operation in the XACML rule file and for all process changes we authorize user for "Next" operation on any type of task
### Pro
- Makes it more flexible to add new types of task
- Can configure PEP to use on action
### Con
- More rules in policy file
- Makes delegation more complex
## A logic in Next operation that identifies the action based on type of Task we currently are in
### Pro
- Less rules
- Matches more current solution
### Cons
- Add complexity to Next operation
- Limits flexibility on process
- Cant use standard PEP
We need to decide
## Conclusion
## Tasks
- [ ] Analyze and discuss
- [ ] Decide
- [ ] Update relevant issues in backlog
|
process
|
analyze and define process handling and authorization description as part of we have defined that process api is separate from data api this means that moving the process forward would trigger a separate api call this api calls need to be authorized the big question is what kind of authorization action is next on the process in current platform we always move the process part of the main action on the task if you do sign on the signing step we authorize for sign and move forward in the same operation as i see it there is two options here introduce a common action for next used for all steps in this scenario we can define the next operation in the xacml rule file and for all process changes we authorize user for next operation on any type of task pro makes it more flexible to add new types of task can configure pep to use on action con more rules in policy file makes delegation more complex a logic in next operation that identifies the action based on type of task we currently are in pro less rules matches more current solution cons add complexity to next operation limits flexibility on process cant use standard pep we need to decide conclusion tasks analyze and discuss decide update relevant issues in backlog
| 1
|
11,142
| 13,957,692,339
|
IssuesEvent
|
2020-10-24 08:10:40
|
alexanderkotsev/geoportal
|
https://api.github.com/repos/alexanderkotsev/geoportal
|
opened
|
PT: Harvesting
|
Geoportal Harvesting process PT - Portugal
|
Geoportal Team,
Can you please start a harvesting to the Portuguese catalogue?
Thank you!
Vanda
|
1.0
|
PT: Harvesting - Geoportal Team,
Can you please start a harvesting to the Portuguese catalogue?
Thank you!
Vanda
|
process
|
pt harvesting geoportal team can you please start a harvesting to the portuguese catalogue thank you vanda
| 1
|
187,601
| 14,428,533,842
|
IssuesEvent
|
2020-12-06 10:16:38
|
software-engineering-labs-group-5/image-similarity
|
https://api.github.com/repos/software-engineering-labs-group-5/image-similarity
|
opened
|
[Tests] Time computing
|
test
|
Check the time of modifications and metrics calculations for different sizes of images.
The calculations of ones smaller than 512x512 should not take more than 1s.
|
1.0
|
[Tests] Time computing - Check the time of modifications and metrics calculations for different sizes of images.
The calculations of ones smaller than 512x512 should not take more than 1s.
|
non_process
|
time computing check the time of modifications and metrics calculations for different sizes of images the calculations of ones smaller than should not take more than
| 0
|
11,177
| 13,957,695,276
|
IssuesEvent
|
2020-10-24 08:11:29
|
alexanderkotsev/geoportal
|
https://api.github.com/repos/alexanderkotsev/geoportal
|
opened
|
RO: New harvesting from Romanian Geoportal
|
Geoportal Harvesting process RO - Romania
|
Dear Angelo,
Can you please start a new havesting from Romanian Geoportal? We did some changes and we want to see the outcome.
Thank you in advance for your help.
Beest Regards,
Simona Bunea
|
1.0
|
RO: New harvesting from Romanian Geoportal - Dear Angelo,
Can you please start a new havesting from Romanian Geoportal? We did some changes and we want to see the outcome.
Thank you in advance for your help.
Beest Regards,
Simona Bunea
|
process
|
ro new harvesting from romanian geoportal dear angelo can you please start a new havesting from romanian geoportal we did some changes and we want to see the outcome thank you in advance for your help beest regards simona bunea
| 1
|
31,759
| 13,630,583,149
|
IssuesEvent
|
2020-09-24 16:39:33
|
microsoft/botbuilder-dotnet
|
https://api.github.com/repos/microsoft/botbuilder-dotnet
|
closed
|
Avoid sending typing activity when bot is invoked as skill
|
Area: Skills Bot Services P1 Size: S bug customer-replied-to customer-reported
|
## Version
4.9.4
## Describe the bug
The ShowTypingMiddleware doesn't consider if it is running in the context of a root bot or a skill. This will cause unnecessary typing events if both the root bot and the skill are sending typing events. To avoid this the ShowTypingMiddleware should check the context using the `SkillValidation.IsSkillClaim` and only send typing events when running as a root bot.
I'll be submitting a PR with a fix for this issue.
## To Reproduce
Steps to reproduce the behavior:
1. Create a bot and a skill
2. Configure ShowTypingMiddleware for both the bot and the skill
3. Invoke the skill through the bot
4. Bot the bot and the skill send typing events
## Expected behavior
Only the root bot should send typing events. Skills should not unless they are invoked as a root bot.
## Additional context
Created parent issue to track this gets ported to the other languages [here](https://github.com/microsoft/botframework-sdk/issues/6049)
|
1.0
|
Avoid sending typing activity when bot is invoked as skill - ## Version
4.9.4
## Describe the bug
The ShowTypingMiddleware doesn't consider if it is running in the context of a root bot or a skill. This will cause unnecessary typing events if both the root bot and the skill are sending typing events. To avoid this the ShowTypingMiddleware should check the context using the `SkillValidation.IsSkillClaim` and only send typing events when running as a root bot.
I'll be submitting a PR with a fix for this issue.
## To Reproduce
Steps to reproduce the behavior:
1. Create a bot and a skill
2. Configure ShowTypingMiddleware for both the bot and the skill
3. Invoke the skill through the bot
4. Bot the bot and the skill send typing events
## Expected behavior
Only the root bot should send typing events. Skills should not unless they are invoked as a root bot.
## Additional context
Created parent issue to track this gets ported to the other languages [here](https://github.com/microsoft/botframework-sdk/issues/6049)
|
non_process
|
avoid sending typing activity when bot is invoked as skill version describe the bug the showtypingmiddleware doesn t consider if it is running in the context of a root bot or a skill this will cause unnecessary typing events if both the root bot and the skill are sending typing events to avoid this the showtypingmiddleware should check the context using the skillvalidation isskillclaim and only send typing events when running as a root bot i ll be submitting a pr with a fix for this issue to reproduce steps to reproduce the behavior create a bot and a skill configure showtypingmiddleware for both the bot and the skill invoke the skill through the bot bot the bot and the skill send typing events expected behavior only the root bot should send typing events skills should not unless they are invoked as a root bot additional context created parent issue to track this gets ported to the other languages
| 0
|
13,347
| 15,806,330,507
|
IssuesEvent
|
2021-04-04 04:33:44
|
gfx-rs/naga
|
https://api.github.com/repos/gfx-rs/naga
|
opened
|
Incorrectly placed Return
|
area: processing kind: bug
|
From https://github.com/gfx-rs/wgpu-rs/issues/838
```rust
if (vm * vn > 0.0) {
const cosThetaV = dot(n, v);
const sinThetaV2 = 1.0 - cosThetaV * cosThetaV;
const tanThetaV2 = sinThetaV2 / cosThetaV / cosThetaV;
return 2.0 / (1.0 + sqrt(1.0 + alpha * alpha * tanThetaV2));
} else {
return 0.0;
}
```
This code ends up with a bare `return` at the end.
|
1.0
|
Incorrectly placed Return - From https://github.com/gfx-rs/wgpu-rs/issues/838
```rust
if (vm * vn > 0.0) {
const cosThetaV = dot(n, v);
const sinThetaV2 = 1.0 - cosThetaV * cosThetaV;
const tanThetaV2 = sinThetaV2 / cosThetaV / cosThetaV;
return 2.0 / (1.0 + sqrt(1.0 + alpha * alpha * tanThetaV2));
} else {
return 0.0;
}
```
This code ends up with a bare `return` at the end.
|
process
|
incorrectly placed return from rust if vm vn const costhetav dot n v const costhetav costhetav const costhetav costhetav return sqrt alpha alpha else return this code ends up with a bare return at the end
| 1
|
541
| 3,002,466,967
|
IssuesEvent
|
2015-07-24 17:24:20
|
HazyResearch/dd-genomics
|
https://api.github.com/repos/HazyResearch/dd-genomics
|
closed
|
Remove sentences from current sentences_input that do not contain a gene name
|
Data source Preprocessing
|
Basically for speed. This should cut our sentences_input size by a factor of 8.
|
1.0
|
Remove sentences from current sentences_input that do not contain a gene name - Basically for speed. This should cut our sentences_input size by a factor of 8.
|
process
|
remove sentences from current sentences input that do not contain a gene name basically for speed this should cut our sentences input size by a factor of
| 1
|
243,915
| 18,734,047,021
|
IssuesEvent
|
2021-11-04 03:32:46
|
ihumanable/patch
|
https://api.github.com/repos/ihumanable/patch
|
closed
|
Cross Reference `expose/2` and `private/1`
|
:blue_book: Documentation :muscle: Enhancement
|
Common issue when using `expose/2` is that the compiler starts producing nuisance warnings. This is because as far as the compiler is concerned you are calling a function that doesn't exists, since the exposure happens at runtime.
There's already a solution in the `private/1` macro but these two functions don't formally cross-reference each other.
- [x] Add cross-references so test authors using `expose/2` know about `private/1` and vice versa.
- [x] The super-powers doc should also be updated to show correct usage
- [x] This is a good candidate for adding to the guide book.
|
1.0
|
Cross Reference `expose/2` and `private/1` - Common issue when using `expose/2` is that the compiler starts producing nuisance warnings. This is because as far as the compiler is concerned you are calling a function that doesn't exists, since the exposure happens at runtime.
There's already a solution in the `private/1` macro but these two functions don't formally cross-reference each other.
- [x] Add cross-references so test authors using `expose/2` know about `private/1` and vice versa.
- [x] The super-powers doc should also be updated to show correct usage
- [x] This is a good candidate for adding to the guide book.
|
non_process
|
cross reference expose and private common issue when using expose is that the compiler starts producing nuisance warnings this is because as far as the compiler is concerned you are calling a function that doesn t exists since the exposure happens at runtime there s already a solution in the private macro but these two functions don t formally cross reference each other add cross references so test authors using expose know about private and vice versa the super powers doc should also be updated to show correct usage this is a good candidate for adding to the guide book
| 0
|
311,017
| 26,761,797,525
|
IssuesEvent
|
2023-01-31 07:37:40
|
softeerbootcamp/Team7-Amatta
|
https://api.github.com/repos/softeerbootcamp/Team7-Amatta
|
closed
|
[기능 개발] DB 연결(BE)
|
✨ Feature ✅ Test
|
#### 설명
Spring boot 프로젝트에 DB 연결
개발 브랜치 : BE-feature/DBConnection
#### 🗒: 체크리스트
- [x] DB 연결
- [x] .yml or .properties 파일로 연결 관련 정보 관리
- [x] 연결 테스트 코드 작성
|
1.0
|
[기능 개발] DB 연결(BE) - #### 설명
Spring boot 프로젝트에 DB 연결
개발 브랜치 : BE-feature/DBConnection
#### 🗒: 체크리스트
- [x] DB 연결
- [x] .yml or .properties 파일로 연결 관련 정보 관리
- [x] 연결 테스트 코드 작성
|
non_process
|
db 연결 be 설명 spring boot 프로젝트에 db 연결 개발 브랜치 be feature dbconnection 🗒 체크리스트 db 연결 yml or properties 파일로 연결 관련 정보 관리 연결 테스트 코드 작성
| 0
|
414,887
| 12,120,983,067
|
IssuesEvent
|
2020-04-22 08:34:51
|
microsoft/STL
|
https://api.github.com/repos/microsoft/STL
|
closed
|
<functional>: std::function doesn't handle over-aligned types
|
bug high priority work in progress
|
The Small Functor Optimization in `std::function` was implemented before we started to respect over-aligned types. Currently, we emit silent bad codegen when we decide to locally store an over-aligned function object.
```
C:\Temp>type meow.cpp
#include <functional>
#include <stdint.h>
#include <stdio.h>
using namespace std;
struct alignas(16) Meow {
int i;
void operator()() {
printf("This should be 0: %d\n", static_cast<int>(reinterpret_cast<uintptr_t>(this) % alignof(Meow)));
}
};
int main() {
function<void()> f(Meow{});
int x = 1729;
function<void()> g(Meow{});
f();
g();
(void) x;
}
C:\Temp>cl /EHsc /nologo /W4 /wd4324 meow.cpp
meow.cpp
C:\Temp>meow
This should be 0: 0
This should be 0: 8
```
@CaseyCarter observed: "I think all that's required here is to add `|| alignof(_Impl) > alignof(max_align_t)` to `_Is_large` and write a quick regression test."
https://github.com/microsoft/STL/blob/22391019cba3863a978ea2d7d3f5ca240f38c576/stl/inc/functional#L797-L798
https://github.com/microsoft/STL/blob/22391019cba3863a978ea2d7d3f5ca240f38c576/stl/inc/functional#L1065-L1069
Tracked by Microsoft-internal VSO-157296 and VSO-1062649.
|
1.0
|
<functional>: std::function doesn't handle over-aligned types - The Small Functor Optimization in `std::function` was implemented before we started to respect over-aligned types. Currently, we emit silent bad codegen when we decide to locally store an over-aligned function object.
```
C:\Temp>type meow.cpp
#include <functional>
#include <stdint.h>
#include <stdio.h>
using namespace std;
struct alignas(16) Meow {
int i;
void operator()() {
printf("This should be 0: %d\n", static_cast<int>(reinterpret_cast<uintptr_t>(this) % alignof(Meow)));
}
};
int main() {
function<void()> f(Meow{});
int x = 1729;
function<void()> g(Meow{});
f();
g();
(void) x;
}
C:\Temp>cl /EHsc /nologo /W4 /wd4324 meow.cpp
meow.cpp
C:\Temp>meow
This should be 0: 0
This should be 0: 8
```
@CaseyCarter observed: "I think all that's required here is to add `|| alignof(_Impl) > alignof(max_align_t)` to `_Is_large` and write a quick regression test."
https://github.com/microsoft/STL/blob/22391019cba3863a978ea2d7d3f5ca240f38c576/stl/inc/functional#L797-L798
https://github.com/microsoft/STL/blob/22391019cba3863a978ea2d7d3f5ca240f38c576/stl/inc/functional#L1065-L1069
Tracked by Microsoft-internal VSO-157296 and VSO-1062649.
|
non_process
|
std function doesn t handle over aligned types the small functor optimization in std function was implemented before we started to respect over aligned types currently we emit silent bad codegen when we decide to locally store an over aligned function object c temp type meow cpp include include include using namespace std struct alignas meow int i void operator printf this should be d n static cast reinterpret cast this alignof meow int main function f meow int x function g meow f g void x c temp cl ehsc nologo meow cpp meow cpp c temp meow this should be this should be caseycarter observed i think all that s required here is to add alignof impl alignof max align t to is large and write a quick regression test tracked by microsoft internal vso and vso
| 0
|
34,884
| 14,538,137,168
|
IssuesEvent
|
2020-12-15 10:04:34
|
elastic/kibana
|
https://api.github.com/repos/elastic/kibana
|
closed
|
[APM] Service overview: instances table
|
Team:apm [zube]: In Progress apm:service-overview v7.11.0
|
(Part of the new APM service overview. See elastic/kibana#81135 for more.)
Show table of instances and metrics for the service.
The table will be sorted by **traffic** by default

|
1.0
|
[APM] Service overview: instances table - (Part of the new APM service overview. See elastic/kibana#81135 for more.)
Show table of instances and metrics for the service.
The table will be sorted by **traffic** by default

|
non_process
|
service overview instances table part of the new apm service overview see elastic kibana for more show table of instances and metrics for the service the table will be sorted by traffic by default
| 0
|
188,885
| 6,786,895,051
|
IssuesEvent
|
2017-10-31 00:27:36
|
uva-slp/meals
|
https://api.github.com/repos/uva-slp/meals
|
opened
|
Fix Login Issues on Course Server
|
bug priority-very-high story-points-3
|
Fix the issue with users registering but not being able to login.
|
1.0
|
Fix Login Issues on Course Server - Fix the issue with users registering but not being able to login.
|
non_process
|
fix login issues on course server fix the issue with users registering but not being able to login
| 0
|
19,518
| 25,829,638,464
|
IssuesEvent
|
2022-12-12 15:15:31
|
microsoft/vscode
|
https://api.github.com/repos/microsoft/vscode
|
closed
|
Local terminal uses remote shell path
|
bug remote confirmation-pending terminal-profiles terminal-process
|
Issue Type: <b>Bug</b>
I'm SSH'd to a remote Linux machine where zsh is the default shell. It's at /usr/bin/zsh.
When I try to launch a local shell on my Mac via the command palette, I get the error:
The terminal Process failed to launch: Path to shell executable "/usr/bin/zsh" does not exist.
On my local Mac the path is /bin/zsh, so it won't work. Local shells should launch with different expectations about the shell and path to it.
VS Code version: Code 1.65.2 (c722ca6c7eed3d7987c0d5c3df5c45f6b15e77d1, 2022-03-10T14:33:49.188Z)
OS version: Darwin x64 19.6.0
Restricted Mode: No
Remote OS version: Linux x64 5.13.0-30-generic
<details>
<summary>System Info</summary>
|Item|Value|
|---|---|
|CPUs|Intel(R) Core(TM) i9-9980HK CPU @ 2.40GHz (16 x 2400)|
|GPU Status|2d_canvas: enabled<br>gpu_compositing: enabled<br>metal: disabled_off<br>multiple_raster_threads: enabled_on<br>oop_rasterization: enabled<br>opengl: enabled_on<br>rasterization: enabled<br>skia_renderer: disabled_off_ok<br>video_decode: enabled<br>webgl: enabled<br>webgl2: enabled|
|Load (avg)|1, 2, 2|
|Memory (System)|16.00GB (0.24GB free)|
|Process Argv|-psn_0_13114497 --crash-reporter-id 2e2b7f5a-8a45-4967-bac0-e501e2043bd4|
|Screen Reader|no|
|VM|0%|
|Item|Value|
|---|---|
|Remote|SSH: REDACTED|
|OS|Linux x64 5.13.0-30-generic|
|CPUs|AMD Ryzen 9 5950X 16-Core Processor (32 x 2200)|
|Memory (System)|62.71GB (0.46GB free)|
|VM|0%|
</details><details><summary>Extensions (24)</summary>
Extension|Author (truncated)|Version
---|---|---
better-cpp-syntax|jef|1.15.13
kconfig|luv|1.0.3
autoconf|mae|0.1.0
jupyter-keymap|ms-|1.0.0
remote-containers|ms-|0.231.0
remote-ssh|ms-|0.76.1
remote-ssh-edit|ms-|0.78.0
remote-wsl|ms-|0.64.2
doxdocgen|csc|1.4.0
xml|Dot|2.5.1
gitlens|eam|12.0.5
vscode-pull-request-github|Git|0.38.1
better-cpp-syntax|jef|1.15.13
python|ms-|2022.2.1924087327
vscode-pylance|ms-|2022.3.3
jupyter|ms-|2022.2.1030672458
cmake-tools|ms-|1.10.5
cpptools|ms-|1.9.7
cpptools-extension-pack|ms-|1.1.0
vsliveshare|ms-|1.0.5449
vsliveshare-audio|ms-|0.1.91
perforce|sle|3.2.0
vscode-hexdump|sle|1.8.1
cmake|twx|0.0.17
(2 theme extensions excluded)
</details><details>
<summary>A/B Experiments</summary>
```
vsliv368cf:30146710
vsreu685:30147344
python383:30185418
vspor879:30202332
vspor708:30202333
vspor363:30204092
pythonvspyl392:30443607
pythontb:30283811
pythonvspyt551cf:30345471
pythonptprofiler:30281270
vshan820:30294714
vstes263:30335439
vscorecescf:30445987
pythondataviewer:30285071
vscod805:30301674
pythonvspyt200:30340761
binariesv615:30325510
bridge0708:30335490
bridge0723:30353136
vsaa593cf:30376535
vsc1dst:30438360
pythonvs932:30410667
wslgetstarted:30449410
vsclayoutctrc:30451274
dsvsc009:30452663
pythonvsnew555:30457759
vscscmwlcmt:30438805
vscgsvid1:30447480
pynewfile477:30463512
```
</details>
<!-- generated by issue reporter -->
|
1.0
|
Local terminal uses remote shell path -
Issue Type: <b>Bug</b>
I'm SSH'd to a remote Linux machine where zsh is the default shell. It's at /usr/bin/zsh.
When I try to launch a local shell on my Mac via the command palette, I get the error:
The terminal Process failed to launch: Path to shell executable "/usr/bin/zsh" does not exist.
On my local Mac the path is /bin/zsh, so it won't work. Local shells should launch with different expectations about the shell and path to it.
VS Code version: Code 1.65.2 (c722ca6c7eed3d7987c0d5c3df5c45f6b15e77d1, 2022-03-10T14:33:49.188Z)
OS version: Darwin x64 19.6.0
Restricted Mode: No
Remote OS version: Linux x64 5.13.0-30-generic
<details>
<summary>System Info</summary>
|Item|Value|
|---|---|
|CPUs|Intel(R) Core(TM) i9-9980HK CPU @ 2.40GHz (16 x 2400)|
|GPU Status|2d_canvas: enabled<br>gpu_compositing: enabled<br>metal: disabled_off<br>multiple_raster_threads: enabled_on<br>oop_rasterization: enabled<br>opengl: enabled_on<br>rasterization: enabled<br>skia_renderer: disabled_off_ok<br>video_decode: enabled<br>webgl: enabled<br>webgl2: enabled|
|Load (avg)|1, 2, 2|
|Memory (System)|16.00GB (0.24GB free)|
|Process Argv|-psn_0_13114497 --crash-reporter-id 2e2b7f5a-8a45-4967-bac0-e501e2043bd4|
|Screen Reader|no|
|VM|0%|
|Item|Value|
|---|---|
|Remote|SSH: REDACTED|
|OS|Linux x64 5.13.0-30-generic|
|CPUs|AMD Ryzen 9 5950X 16-Core Processor (32 x 2200)|
|Memory (System)|62.71GB (0.46GB free)|
|VM|0%|
</details><details><summary>Extensions (24)</summary>
Extension|Author (truncated)|Version
---|---|---
better-cpp-syntax|jef|1.15.13
kconfig|luv|1.0.3
autoconf|mae|0.1.0
jupyter-keymap|ms-|1.0.0
remote-containers|ms-|0.231.0
remote-ssh|ms-|0.76.1
remote-ssh-edit|ms-|0.78.0
remote-wsl|ms-|0.64.2
doxdocgen|csc|1.4.0
xml|Dot|2.5.1
gitlens|eam|12.0.5
vscode-pull-request-github|Git|0.38.1
better-cpp-syntax|jef|1.15.13
python|ms-|2022.2.1924087327
vscode-pylance|ms-|2022.3.3
jupyter|ms-|2022.2.1030672458
cmake-tools|ms-|1.10.5
cpptools|ms-|1.9.7
cpptools-extension-pack|ms-|1.1.0
vsliveshare|ms-|1.0.5449
vsliveshare-audio|ms-|0.1.91
perforce|sle|3.2.0
vscode-hexdump|sle|1.8.1
cmake|twx|0.0.17
(2 theme extensions excluded)
</details><details>
<summary>A/B Experiments</summary>
```
vsliv368cf:30146710
vsreu685:30147344
python383:30185418
vspor879:30202332
vspor708:30202333
vspor363:30204092
pythonvspyl392:30443607
pythontb:30283811
pythonvspyt551cf:30345471
pythonptprofiler:30281270
vshan820:30294714
vstes263:30335439
vscorecescf:30445987
pythondataviewer:30285071
vscod805:30301674
pythonvspyt200:30340761
binariesv615:30325510
bridge0708:30335490
bridge0723:30353136
vsaa593cf:30376535
vsc1dst:30438360
pythonvs932:30410667
wslgetstarted:30449410
vsclayoutctrc:30451274
dsvsc009:30452663
pythonvsnew555:30457759
vscscmwlcmt:30438805
vscgsvid1:30447480
pynewfile477:30463512
```
</details>
<!-- generated by issue reporter -->
|
process
|
local terminal uses remote shell path issue type bug i m ssh d to a remote linux machine where zsh is the default shell it s at usr bin zsh when i try to launch a local shell on my mac via the command palette i get the error the terminal process failed to launch path to shell executable usr bin zsh does not exist on my local mac the path is bin zsh so it won t work local shells should launch with different expectations about the shell and path to it vs code version code os version darwin restricted mode no remote os version linux generic system info item value cpus intel r core tm cpu x gpu status canvas enabled gpu compositing enabled metal disabled off multiple raster threads enabled on oop rasterization enabled opengl enabled on rasterization enabled skia renderer disabled off ok video decode enabled webgl enabled enabled load avg memory system free process argv psn crash reporter id screen reader no vm item value remote ssh redacted os linux generic cpus amd ryzen core processor x memory system free vm extensions extension author truncated version better cpp syntax jef kconfig luv autoconf mae jupyter keymap ms remote containers ms remote ssh ms remote ssh edit ms remote wsl ms doxdocgen csc xml dot gitlens eam vscode pull request github git better cpp syntax jef python ms vscode pylance ms jupyter ms cmake tools ms cpptools ms cpptools extension pack ms vsliveshare ms vsliveshare audio ms perforce sle vscode hexdump sle cmake twx theme extensions excluded a b experiments pythontb pythonptprofiler vscorecescf pythondataviewer wslgetstarted vsclayoutctrc vscscmwlcmt
| 1
|
72,863
| 15,251,714,701
|
IssuesEvent
|
2021-02-20 00:13:34
|
wss-demo/ImportedNodeGoat
|
https://api.github.com/repos/wss-demo/ImportedNodeGoat
|
opened
|
CVE-2020-7608 (Medium) detected in yargs-parser-2.4.1.tgz
|
security vulnerability
|
## CVE-2020-7608 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>yargs-parser-2.4.1.tgz</b></p></summary>
<p>the mighty option parser used by yargs</p>
<p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-2.4.1.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-2.4.1.tgz</a></p>
<p>Path to dependency file: ImportedNodeGoat/package.json</p>
<p>Path to vulnerable library: ImportedNodeGoat/node_modules/nyc/node_modules/yargs-parser/package.json</p>
<p>
Dependency Hierarchy:
- grunt-if-0.2.0.tgz (Root Library)
- grunt-contrib-nodeunit-1.0.0.tgz
- nodeunit-0.9.5.tgz
- tap-7.1.2.tgz
- nyc-7.1.0.tgz
- :x: **yargs-parser-2.4.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://api.github.com/repos/wss-demo/ImportedNodeGoat/commits/8a9930d26492d27fecc32157284fbe70021fc725">8a9930d26492d27fecc32157284fbe70021fc725</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
yargs-parser could be tricked into adding or modifying properties of Object.prototype using a "__proto__" payload.
<p>Publish Date: 2020-03-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7608>CVE-2020-7608</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7608">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7608</a></p>
<p>Release Date: 2020-03-16</p>
<p>Fix Resolution: v18.1.1;13.1.2;15.0.1</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"yargs-parser","packageVersion":"2.4.1","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-if:0.2.0;grunt-contrib-nodeunit:1.0.0;nodeunit:0.9.5;tap:7.1.2;nyc:7.1.0;yargs-parser:2.4.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v18.1.1;13.1.2;15.0.1"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-7608","vulnerabilityDetails":"yargs-parser could be tricked into adding or modifying properties of Object.prototype using a \"__proto__\" payload.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7608","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"Low","S":"Unchanged","C":"Low","UI":"None","AV":"Local","I":"Low"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2020-7608 (Medium) detected in yargs-parser-2.4.1.tgz - ## CVE-2020-7608 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>yargs-parser-2.4.1.tgz</b></p></summary>
<p>the mighty option parser used by yargs</p>
<p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-2.4.1.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-2.4.1.tgz</a></p>
<p>Path to dependency file: ImportedNodeGoat/package.json</p>
<p>Path to vulnerable library: ImportedNodeGoat/node_modules/nyc/node_modules/yargs-parser/package.json</p>
<p>
Dependency Hierarchy:
- grunt-if-0.2.0.tgz (Root Library)
- grunt-contrib-nodeunit-1.0.0.tgz
- nodeunit-0.9.5.tgz
- tap-7.1.2.tgz
- nyc-7.1.0.tgz
- :x: **yargs-parser-2.4.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://api.github.com/repos/wss-demo/ImportedNodeGoat/commits/8a9930d26492d27fecc32157284fbe70021fc725">8a9930d26492d27fecc32157284fbe70021fc725</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
yargs-parser could be tricked into adding or modifying properties of Object.prototype using a "__proto__" payload.
<p>Publish Date: 2020-03-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7608>CVE-2020-7608</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7608">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7608</a></p>
<p>Release Date: 2020-03-16</p>
<p>Fix Resolution: v18.1.1;13.1.2;15.0.1</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"yargs-parser","packageVersion":"2.4.1","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-if:0.2.0;grunt-contrib-nodeunit:1.0.0;nodeunit:0.9.5;tap:7.1.2;nyc:7.1.0;yargs-parser:2.4.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v18.1.1;13.1.2;15.0.1"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-7608","vulnerabilityDetails":"yargs-parser could be tricked into adding or modifying properties of Object.prototype using a \"__proto__\" payload.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7608","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"Low","S":"Unchanged","C":"Low","UI":"None","AV":"Local","I":"Low"},"extraData":{}}</REMEDIATE> -->
|
non_process
|
cve medium detected in yargs parser tgz cve medium severity vulnerability vulnerable library yargs parser tgz the mighty option parser used by yargs library home page a href path to dependency file importednodegoat package json path to vulnerable library importednodegoat node modules nyc node modules yargs parser package json dependency hierarchy grunt if tgz root library grunt contrib nodeunit tgz nodeunit tgz tap tgz nyc tgz x yargs parser tgz vulnerable library found in head commit a href found in base branch master vulnerability details yargs parser could be tricked into adding or modifying properties of object prototype using a proto payload publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree grunt if grunt contrib nodeunit nodeunit tap nyc yargs parser isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails yargs parser could be tricked into adding or modifying properties of object prototype using a proto payload vulnerabilityurl
| 0
|
8,234
| 11,415,575,457
|
IssuesEvent
|
2020-02-02 12:02:16
|
parcel-bundler/parcel
|
https://api.github.com/repos/parcel-bundler/parcel
|
closed
|
CSS files being compiled with PostCSS before they're imported
|
:bug: Bug CSS Preprocessing Stale
|
# 🐛
CSS files appear to be compiling with PostCSS before they're imported. This will lead to unexpected behavior in most PostCSS builds, as imports are typically the first thing that should run. See:
https://github.com/postcss/postcss-import
> This plugin should probably be used as the first plugin of your list. This way, other plugins will work on the AST as if there were only a single file to process, and will probably work as you can expect.
Possibly related to [#329](https://github.com/parcel-bundler/parcel/issues/329) and [#593](https://github.com/parcel-bundler/parcel/issues/593).
### 🎛 Configuration
package.json
```
{
"main": "index.js",
"scripts": {
"start": "parcel src/index.html"
},
"devDependencies": {
"parcel-bundler": "^1.4.1",
"postcss-easy-import": "^3.0.0",
"postcss-simple-vars": "^4.1.0"
}
}
```
postcss.config.js
```
module.exports = {
plugins: [
require("postcss-easy-import"),
require("postcss-simple-vars")
]
};
```
index.html
```
<html>
<head>
<link rel="stylesheet" href="./assets/css/index.css"></link>
</head>
<body>
<script src="./index.js"></script>
</body>
</html>
```
assets/css/index.css
```
@import "./global/colors.css";
body {
background: $red;
}
```
assets/css/global/colors.css
```
$red: #f00;
```
### 🤔 Expected Behavior
```
body {
background: #f00;
}
```
### 😯 Current Behavior
```
🚨 /Users/jon/git/parcel-test/src/assets/css/index.css:4:3: Undefined variable $red
2 |
3 | body {
> 4 | background: $red;
| ^
5 | }
6 |
```
### 💁 Possible Solution
It appears that CSS imports will work even without requiring a PostCSS plugin like `postcss-easy-import`. Because different import plugins can behave differently, and because the order of PostCSS plugins is very important, I believe CSS import behavior should be left completely up to the PostCSS config and plugin.
### 🔦 Context
This config works with webpack + postcss-loader, and gulp.
### 💻 Code Sample
https://github.com/jonmilner/parcel-test
### 🌍 Your Environment
<!--- Include as many relevant details about the environment you experienced the bug in -->
| Software | Version(s) |
| ---------------- | ---------- |
| Parcel | 1.4.1
| Node | 8.2.1
| npm/Yarn | 1.3.2
| Operating System | macOS 10.13.3
|
1.0
|
CSS files being compiled with PostCSS before they're imported - # 🐛
CSS files appear to be compiling with PostCSS before they're imported. This will lead to unexpected behavior in most PostCSS builds, as imports are typically the first thing that should run. See:
https://github.com/postcss/postcss-import
> This plugin should probably be used as the first plugin of your list. This way, other plugins will work on the AST as if there were only a single file to process, and will probably work as you can expect.
Possibly related to [#329](https://github.com/parcel-bundler/parcel/issues/329) and [#593](https://github.com/parcel-bundler/parcel/issues/593).
### 🎛 Configuration
package.json
```
{
"main": "index.js",
"scripts": {
"start": "parcel src/index.html"
},
"devDependencies": {
"parcel-bundler": "^1.4.1",
"postcss-easy-import": "^3.0.0",
"postcss-simple-vars": "^4.1.0"
}
}
```
postcss.config.js
```
module.exports = {
plugins: [
require("postcss-easy-import"),
require("postcss-simple-vars")
]
};
```
index.html
```
<html>
<head>
<link rel="stylesheet" href="./assets/css/index.css"></link>
</head>
<body>
<script src="./index.js"></script>
</body>
</html>
```
assets/css/index.css
```
@import "./global/colors.css";
body {
background: $red;
}
```
assets/css/global/colors.css
```
$red: #f00;
```
### 🤔 Expected Behavior
```
body {
background: #f00;
}
```
### 😯 Current Behavior
```
🚨 /Users/jon/git/parcel-test/src/assets/css/index.css:4:3: Undefined variable $red
2 |
3 | body {
> 4 | background: $red;
| ^
5 | }
6 |
```
### 💁 Possible Solution
It appears that CSS imports will work even without requiring a PostCSS plugin like `postcss-easy-import`. Because different import plugins can behave differently, and because the order of PostCSS plugins is very important, I believe CSS import behavior should be left completely up to the PostCSS config and plugin.
### 🔦 Context
This config works with webpack + postcss-loader, and gulp.
### 💻 Code Sample
https://github.com/jonmilner/parcel-test
### 🌍 Your Environment
<!--- Include as many relevant details about the environment you experienced the bug in -->
| Software | Version(s) |
| ---------------- | ---------- |
| Parcel | 1.4.1
| Node | 8.2.1
| npm/Yarn | 1.3.2
| Operating System | macOS 10.13.3
|
process
|
css files being compiled with postcss before they re imported 🐛 css files appear to be compiling with postcss before they re imported this will lead to unexpected behavior in most postcss builds as imports are typically the first thing that should run see this plugin should probably be used as the first plugin of your list this way other plugins will work on the ast as if there were only a single file to process and will probably work as you can expect possibly related to and 🎛 configuration package json main index js scripts start parcel src index html devdependencies parcel bundler postcss easy import postcss simple vars postcss config js module exports plugins require postcss easy import require postcss simple vars index html assets css index css import global colors css body background red assets css global colors css red 🤔 expected behavior body background 😯 current behavior 🚨 users jon git parcel test src assets css index css undefined variable red body background red 💁 possible solution it appears that css imports will work even without requiring a postcss plugin like postcss easy import because different import plugins can behave differently and because the order of postcss plugins is very important i believe css import behavior should be left completely up to the postcss config and plugin 🔦 context this config works with webpack postcss loader and gulp 💻 code sample 🌍 your environment software version s parcel node npm yarn operating system macos
| 1
|
10,770
| 13,594,677,372
|
IssuesEvent
|
2020-09-22 00:37:55
|
kubernetes/minikube
|
https://api.github.com/repos/kubernetes/minikube
|
closed
|
Release process: run brew bump-formula-pr in docker
|
help wanted kind/process lifecycle/rotten priority/important-longterm
|
we need to run this command in a docker so it doesnt die of gem dependency
```
brew bump-formula-pr \
--strict minikube \
--revision="${revision}" \
--message="This PR was automatically created by minikube release scripts. Contact @tstromberg with any questions." \
--no-browse \
--tag="${TAG}" \
&& status=0 || status=$?
```
https://github.com/kubernetes/minikube/blob/eb13446e786c9ef70cb0a9f85a633194e62396a1/hack/jenkins/release_update_brew.sh#L48
here is the current gem dep error from out latest 1.8.2 release
<details>
<pre>
14:33:14 + brew bump-formula-pr --strict minikube --revision=eb13446e786c9ef70cb0a9f85a633194e62396a1 '--message=This PR was automatically created by minikube release scripts. Contact @tstromberg with any questions.' --no-browse --tag=v1.8.2
14:33:18 Updating Homebrew...
14:33:22 ==> Auto-updated Homebrew!
14:33:22 Updated Homebrew from 78140cef4 to db121c7b1.
14:35:02 Updated 1 tap (homebrew/core).
14:35:02 ==> New Formulae
14:35:02 adoptopenjdk
14:35:02 adoptopenjdk@11
14:35:02 alsa-lib
14:35:02 apt
14:35:02 archiver
14:35:02 at-spi2-atk
14:35:02 at-spi2-core
14:35:02 attr
14:35:02 avahi
14:35:02 aws-vault
14:35:02 bsdmainutils
14:35:02 cpio
14:35:02 device-mapper
14:35:02 elfutils
14:35:02 gcc@9
14:35:02 git-trim
14:35:02 glibc
14:35:02 hdf5@1.10
14:35:02 hdt
14:35:02 isl@0.18
14:35:02 jdk
14:35:02 jdk@7
14:35:02 jdk@8
14:35:02 ladspa-sdk
14:35:02 libaio
14:35:02 libbsd
14:35:02 libcap
14:35:02 libcap-ng
14:35:02 libfuse
14:35:02 libmnl
14:35:02 libnetfilter-queue
14:35:02 libnfnetlink
14:35:02 libnsl
14:35:02 libprelude
14:35:02 libsasl2
14:35:02 libseccomp
14:35:02 libtirpc
14:35:02 linux-headers
14:35:02 linux-pam
14:35:02 lm-sensors
14:35:02 minimal-python
14:35:02 mockolo
14:35:02 net-tools
14:35:02 newrelic-cli
14:35:02 onedrive
14:35:02 openjdk@10
14:35:02 openjdk@9
14:35:02 procps
14:35:02 python-dbus
14:35:02 python@2
14:35:02 qp
14:35:02 singularity
14:35:02 strace
14:35:02 sysstat
14:35:02 systemd
14:35:02 texlive
14:35:02 tlx
14:35:02 xclogparser
14:35:02 ==> Updated Formulae
14:35:02 abseil
14:35:02 advancemenu
14:35:02 adwaita-icon-theme
14:35:02 afuse
14:35:02 agda
14:35:02 algol68g
14:35:02 angular-cli
14:35:02 ansible
14:35:02 ansible-lint
14:35:02 apache-drill
14:35:02 apibuilder-cli
14:35:02 apng2gif
14:35:02 apr-util
14:35:02 arangodb
14:35:02 argon2
14:35:02 aria2
14:35:02 armadillo
14:35:02 arping
14:35:02 artifactory
14:35:02 asciidoc
14:35:02 asciinema
14:35:02 astrometry-net
14:35:02 atdtool
14:35:02 autoconf
14:35:02 aws-cdk
14:35:02 aws-elasticbeanstalk
14:35:02 aws-shell
14:35:02 awscli
14:35:02 awscli@1
14:35:02 awscurl
14:35:02 awslogs
14:35:02 awsume
14:35:02 azure-cli
14:35:02 azure-storage-cpp
14:35:02 bacula-fd
14:35:02 balena-cli
14:35:02 bandcamp-dl
14:35:02 bazaar
14:35:02 bde
14:35:02 benthos
14:35:02 biosig
14:35:02 bison
14:35:02 bitcoin
14:35:02 black
14:35:02 blazegraph
14:35:02 bonnie++
14:35:02 buildifier
14:35:02 bundletool
14:35:02 bzt
14:35:02 caffe
14:35:02 camlp5
14:35:02 cargo-completion
14:35:02 cartridge-cli
14:35:02 ceres-solver
14:35:02 certbot
14:35:02 cfn-lint
14:35:02 cfr-decompiler
14:35:02 cgns
14:35:02 cgrep
14:35:02 chakra
14:35:02 chapel
14:35:02 charm-tools
14:35:02 cheat
14:35:02 cli53
14:35:02 clojure
14:35:02 clojure-lsp
14:35:02 clutter
14:35:02 cobalt
14:35:02 cocoapods
14:35:02 codec2
14:35:02 codemod
14:35:02 codespell
14:35:02 cogl
14:35:02 comby
14:35:02 composer
14:35:02 conan
14:35:02 confluent-platform
14:35:02 conjure-up
14:35:02 cookiecutter
14:35:02 cpl
14:35:02 cppman
14:35:02 cquery
14:35:02 csound
14:35:02 csvkit
14:35:02 csvtomd
14:35:02 curl
14:35:02 curl-openssl
14:35:02 cvs
14:35:02 dartsim
14:35:02 davmail
14:35:02 db-vcs
14:35:02 dependency-check
14:35:02 derby
14:35:02 detekt
14:35:02 dialog
14:35:02 diceware
14:35:02 diff-pdf
14:35:02 diffoscope
14:35:02 dmd
14:35:02 docbook-xsl
14:35:02 docbook2x
14:35:02 docfx
14:35:02 docker
14:35:02 docker-completion
14:35:02 docker-compose
14:35:02 docker-slim
14:35:02 docker-squash
14:35:02 dosbox-x
14:35:02 double-conversion
14:35:02 dovecot
14:35:02 dtc
14:35:02 dub
14:35:02 dune
14:35:02 duplicity
14:35:02 dust
14:35:02 dxpy
14:35:02 dynare
14:35:02 e2fsprogs
14:35:02 easyrpg-player
14:35:02 ecasound
14:35:02 eccodes
14:35:02 eg-examples
14:35:02 ejdb
14:35:02 elasticsearch
14:35:02 embree
14:35:02 emscripten
14:35:02 ensmallen
14:35:02 esptool
14:35:02 etsh
14:35:02 evince
14:35:02 exim
14:35:02 expect
14:35:02 exploitdb
14:35:02 ext2fuse
14:35:02 ezstream
14:35:02 fairymax
14:35:02 faiss
14:35:02 fantom
14:35:02 fastlane
14:35:02 faudio
14:35:02 fcl
14:35:02 fdroidserver
14:35:02 file-roller
14:35:02 findomain
14:35:02 flann
14:35:02 flatbuffers
14:35:02 flintrock
14:35:02 fluent-bit
14:35:02 flyway
14:35:02 fn
14:35:02 folly
14:35:02 foma
14:35:02 fontconfig
14:35:02 fonttools
14:35:02 fq
14:35:02 frpc
14:35:02 frps
14:35:02 frugal
14:35:02 fselect
14:35:02 fzf
14:35:02 gambit
14:35:02 gambit-scheme
14:35:02 gammaray
14:35:02 gandi.cli
14:35:02 gatsby-cli
14:35:02 gcc
14:35:02 gcc@8
14:35:02 gdal
14:35:02 gdcm
14:35:02 geeqie
14:35:02 geocode-glib
14:35:02 geos
14:35:02 gerbil-scheme
14:35:02 gettext
14:35:02 gimme-aws-creds
14:35:02 git
14:35:02 git-annex
14:35:02 git-plus
14:35:02 git-review
14:35:02 git-revise
14:35:02 git-standup
14:35:02 gitmoji
14:35:02 gitup
14:35:02 gjs
14:35:02 glances
14:35:02 glib
14:35:02 glog
14:35:02 gmsh
14:35:02 gmt
14:35:02 gmt@5
14:35:02 gobject-introspection
14:35:02 google-sparsehash
14:35:02 gpgme
14:35:02 grin
14:35:02 grin-wallet
14:35:02 gromacs
14:35:02 groovy
14:35:02 groovysdk
14:35:02 gsettings-desktop-schemas
14:35:02 gti
14:35:02 gtkmm
14:35:02 gtksourceview4
14:35:02 hdf5
14:35:02 helm
14:35:02 hexyl
14:35:02 homeassistant-cli
14:35:02 hostess
14:35:02 howdoi
14:35:02 htop
14:35:02 httpie
14:35:02 hugo
14:35:02 hyperkit
14:35:02 ibex
14:35:02 ilmbase
14:35:02 imagemagick
14:35:02 imagemagick@6
14:35:02 imgproxy
14:35:02 inadyn
14:35:02 inlets
14:35:02 instalooter
14:35:02 internetarchive
14:35:02 intltool
14:35:02 io
14:35:02 ios-sim
14:35:02 ipmiutil
14:35:02 ipython
14:35:02 isl
14:35:02 itk
14:35:02 jbig2dec
14:35:02 jed
14:35:02 jenkins
14:35:02 jenkins-job-builder
14:35:02 jenkins-lts
14:35:02 jhead
14:35:02 jhipster
14:35:02 jinja2-cli
14:35:02 joplin
14:35:02 jrnl
14:35:02 jruby
14:35:02 juise
14:35:02 juju
14:35:02 juju-wait
14:35:02 jupyterlab
14:35:02 kallisto
14:35:02 keepkey-agent
14:35:02 khard
14:35:02 kibana
14:35:02 krew
14:35:02 ktoblzcheck
14:35:02 kubeprod
14:35:02 kubernetes-cli
14:35:02 kubeseal
14:35:02 kumo
14:35:02 lablgtk
14:35:02 lammps
14:35:02 ldc
14:35:02 legit
14:35:02 lgogdownloader
14:35:02 libbi
14:35:02 libcouchbase
14:35:02 libdazzle
14:35:02 libewf
14:35:02 libexosip
14:35:02 libfabric
14:35:02 libgetdata
14:35:02 libgosu
14:35:02 libgweather
14:35:02 libhttpseverywhere
14:35:02 libical
14:35:02 libmatio
14:35:02 libowfat
14:35:02 libpeas
14:35:02 libproxy
14:35:02 libpsl
14:35:02 libqalculate
14:35:02 librealsense
14:35:02 libsecret
14:35:02 libserdes
14:35:02 libsoup
14:35:02 libssh
14:35:02 libsvm
14:35:02 libtcod
14:35:02 libtool
14:35:02 libtorrent-rasterbar
14:35:02 libupnp
14:35:02 libwebsockets
14:35:02 libxml++
14:35:02 libzdb
14:35:02 liquibase
14:35:02 literate-git
14:35:02 llvm
14:35:02 llvm@6
14:35:02 llvm@7
14:35:02 llvm@8
14:35:02 lolcat
14:35:02 lua
14:35:02 lua@5.1
14:35:02 luaradio
14:35:02 luarocks
14:35:02 luit
14:35:02 mackup
14:35:02 magic-wormhole
14:35:02 mapserver
14:35:02 mbedtls
14:35:02 mdds
14:35:02 mdv
14:35:02 memcached
14:35:02 menhir
14:35:02 mercurial
14:35:02 mercury
14:35:02 mesa
14:35:02 meson-internal
14:35:02 metabase
14:35:02 micronaut
14:35:02 minikube
14:35:02 minio
14:35:02 minizinc
14:35:02 minizip2
14:35:02 mitmproxy
14:35:02 mkdocs
14:35:02 mkl-dnn
14:35:02 mkvtoolnix
14:35:02 mlpack
14:35:02 molecule
14:35:02 mongo-cxx-driver
14:35:02 mongo-orchestration
14:35:02 mosh
14:35:02 mosquitto
14:35:02 mpd
14:35:02 mruby
14:35:02 mujs
14:35:02 mycli
14:35:02 mypy
14:35:02 mysql-connector-c++
14:35:02 n
14:35:02 nagios-plugins
14:35:02 nbdime
14:35:02 nco
14:35:02 ncview
14:35:02 neofetch
14:35:02 neovim
14:35:02 netcdf
14:35:02 nethack
14:35:02 ninja
14:35:02 nng
14:35:02 node-sass
14:35:02 nodeenv
14:35:02 nss
14:35:02 nudoku
14:35:02 nushell
14:35:02 oauth2l
14:35:02 ocaml
14:35:02 ocaml-findlib
14:35:02 ocaml-num
14:35:02 ocamlbuild
14:35:02 ocamlsdl
14:35:02 ocrmypdf
14:35:02 octave
14:35:02 ode
14:35:02 odpi
14:35:02 onnxruntime
14:35:02 open-image-denoise
14:35:02 open-mpi
14:35:02 openexr
14:35:02 openssl@1.1
14:35:02 pastel
14:35:02 pcl
14:35:02 pdal
14:35:02 pgcli
14:35:02 pipenv
14:35:02 pipx
14:35:02 pkg-config
14:35:02 platformio
14:35:02 plotutils
14:35:02 pmd
14:35:02 pnpm
14:35:02 points2grid
14:35:02 pre-commit
14:35:02 procs
14:35:02 prodigal
14:35:02 proselint
14:35:02 protoc-gen-go
14:35:02 psc-package
14:35:02 pssh
14:35:02 pulseaudio
14:35:02 pulumi
14:35:02 pushpin
14:35:02 pwntools
14:35:02 pyenv
14:35:02 pygitup
14:35:02 pygments
14:35:02 pygobject3
14:35:02 pyinvoke
14:35:02 pylint
14:35:02 python
14:35:02 python@3.8
14:35:02 pyvim
14:35:02 qalculate-gtk
14:35:02 qt
14:35:02 rabbitmq
14:35:02 radamsa
14:35:02 radare2
14:35:02 rakudo-star
14:35:02 re-flex
14:35:02 remarshal
14:35:02 rex
14:35:02 rst-lint
14:35:02 rtv
14:35:02 rubberband
14:35:02 rust
14:35:02 s3cmd
14:35:02 s3fs
14:35:02 s6
14:35:02 sbjson
14:35:02 sceptre
14:35:02 scons
14:35:02 scour
14:35:02 sdcc
14:35:02 sdedit
14:35:02 sdl2
14:35:02 serf
14:35:02 serverless
14:35:02 sfk
14:35:02 shogun
14:35:02 simg2img
14:35:02 simple-tiles
14:35:02 sjk
14:35:02 skinny
14:35:02 sn0int
14:35:02 sonar-scanner
14:35:02 sonarqube
14:35:02 sphinx-doc
14:35:02 spice-protocol
14:35:02 spidermonkey
14:35:02 spigot
14:35:02 spirv-cross
14:35:02 spirv-tools
14:35:02 sqliteodbc
14:35:02 sqlparse
14:35:02 ssdb
14:35:02 ssh-audit
14:35:02 sshuttle
14:35:02 statik
14:35:02 stdman
14:35:02 stellar-core
14:35:02 stgit
14:35:02 stolon
14:35:02 streamlink
14:35:02 stress-ng
14:35:02 stubby
14:35:02 subversion
14:35:02 suite-sparse
14:35:02 sundials
14:35:02 supervisor
14:35:02 svtplay-dl
14:35:02 swagger-codegen
14:35:02 swift-protobuf
14:35:02 swift-sh
14:35:02 swiftformat
14:35:02 switch-lan-play
14:35:02 talisman
14:35:02 tasksh
14:35:02 tbox
14:35:02 tektoncd-cli
14:35:02 terragrunt
14:35:02 terrahelp
14:35:02 terrahub
14:35:02 texinfo
14:35:02 texmath
14:35:02 tgui
14:35:02 thefuck
14:35:02 tin
14:35:02 tinyxml
14:35:02 tinyxml2
14:35:02 todoman
14:35:02 tox
14:35:02 trafficserver
14:35:02 translate-toolkit
14:35:02 trash-cli
14:35:02 ttyd
14:35:02 twarc
14:35:02 twine-pypi
14:35:02 txr
14:35:02 txt2tags
14:35:02 u-boot-tools
14:35:02 urh
14:35:02 v2ray-plugin
14:35:02 v8
14:35:02 vala
14:35:02 vale
14:35:02 valgrind
14:35:02 vault
14:35:02 velero
14:35:02 verilator
14:35:02 vips
14:35:02 visp
14:35:02 vowpal-wabbit
14:35:02 vsts-cli
14:35:02 vtk
14:35:02 vttest
14:35:02 vulkan-headers
14:35:02 wabt
14:35:02 wal2json
14:35:02 wapm
14:35:02 wartremover
14:35:02 wasmer
14:35:02 watson
14:35:02 wcslib
14:35:02 webdis
14:35:02 websocat
14:35:02 webtorrent-cli
14:35:02 wp-cli-completion
14:35:02 wpscan
14:35:02 wxmaxima
14:35:02 x3270
14:35:02 xclip
14:35:02 xdotool
14:35:02 xmlto
14:35:02 xmrig
14:35:02 xonsh
14:35:02 xtensor
14:35:02 xxhash
14:35:02 yamllint
14:35:02 yapf
14:35:02 yarn
14:35:02 ydcv
14:35:02 ykman
14:35:02 you-get
14:35:02 youtube-dl
14:35:02 z
14:35:02 zabbix-cli
14:35:02 zeek
14:35:02 zimg
14:35:02 znc
14:35:02 zola
14:35:02 ==> Deleted Formulae
14:35:02 camlp4
14:35:02 ck
14:35:02 djmount
14:35:02 field3d
14:35:02 gmediaserver
14:35:02
14:35:03 ==> Fetching homebrew/master
14:35:06 From https://github.com/Homebrew/homebrew-core
14:35:06 * branch master -> FETCH_HEAD
14:35:06 257bb3e8cb..d3eb71b109 master -> homebrew/master
14:35:06 ==> minikube exists in homebrew/master
14:35:07 Previous HEAD position was a99a5e418f... Merge branch homebrew/master into linuxbrew/master
14:35:07 HEAD is now at d3eb71b109... composer 1.10.1 (#51704)
14:35:07 ==> replace "v1.8.1" with "v1.8.2"
14:35:07 ==> replace "cbda04cf6bbe65e987ae52bb393c10099ab62014" with "eb13446e786c9ef70cb0a9f85a633194e62396a1"
14:35:09 Fetching gem metadata from https://rubygems.org/.........
14:35:10 Using concurrent-ruby 1.1.6
14:35:10 Using minitest 5.14.0
14:35:10 Using thread_safe 0.3.6
14:35:10 Using zeitwerk 2.3.0
14:35:10 Using ast 2.4.0
14:35:10 Using bundler 1.17.2
14:35:10 Using connection_pool 2.2.2
14:35:10 Using json 2.3.0
14:35:10 Using docile 1.3.2
14:35:10 Using simplecov-html 0.10.2
14:35:10 Using sync 0.5.0
14:35:10 Using thor 1.0.1
14:35:10 Using diff-lcs 1.3
14:35:10 Using unf_ext 0.0.7.6
14:35:10 Using hpricot 0.8.6
14:35:10 Using jaro_winkler 1.5.4
14:35:10 Using mime-types-data 3.2019.1009
14:35:10 Using net-http-digest_auth 1.4.1
14:35:10 Using mini_portile2 2.4.0
14:35:10 Using ntlm-http 0.1.1
14:35:10 Using webrobots 0.1.2
14:35:10 Using mustache 1.1.1
14:35:10 Using parallel 1.19.1
14:35:10 Using plist 3.5.0
14:35:10 Using rainbow 3.0.0
14:35:10 Using rdiscount 2.2.0.1
14:35:10 Using rexml 3.2.4
14:35:10 Using rspec-support 3.9.2
14:35:10 Using ruby-progressbar 1.10.1
14:35:10 Using unicode-display_width 1.6.1
14:35:10 Using ruby-macho 2.2.0
14:35:10 Using i18n 1.8.2
14:35:10 Using tzinfo 1.2.6
14:35:10 Using parser 2.7.0.4
14:35:10 Using tins 1.24.1
14:35:10 Using net-http-persistent 3.1.0
14:35:10 Using unf 0.1.4
14:35:10 Using mime-types 3.3.1
14:35:10 Using simplecov 0.16.1
14:35:10 Using ronn 0.7.3
14:35:10 Using rspec-core 3.9.1
14:35:10 Using rspec-expectations 3.9.0
14:35:10 Using rspec-mocks 3.9.1
14:35:10 Using activesupport 6.0.2.1
14:35:10 Using parallel_tests 2.31.0
14:35:10 Using rubocop 0.80.1
14:35:10 Using term-ansicolor 1.7.1
14:35:10 Fetching nokogiri 1.10.9
14:35:10 Using coveralls 0.8.23
14:35:10 Using domain_name 0.5.20190701
14:35:10 Using rspec 3.9.0
14:35:10 Using rspec-its 1.3.0
14:35:10 Using rspec-retry 0.6.2
14:35:10 Using rubocop-performance 1.5.2
14:35:10 Using rubocop-rspec 1.38.1
14:35:10 Using http-cookie 1.0.3
14:35:10 Using rspec-wait 0.0.9
14:35:10 Installing nokogiri 1.10.9 with native extensions
14:35:12 Gem::Ext::BuildError: ERROR: Failed to build gem native extension.
14:35:12
14:35:12 current directory:
14:35:12 /home/linuxbrew/.linuxbrew/Homebrew/Library/Homebrew/vendor/bundle/ruby/2.6.0/gems/nokogiri-1.10.9/ext/nokogiri
14:35:12 /home/linuxbrew/.linuxbrew/Homebrew/Library/Homebrew/vendor/portable-ruby/2.6.3/bin/ruby
14:35:12 -I
14:35:12 /home/linuxbrew/.linuxbrew/Homebrew/Library/Homebrew/vendor/portable-ruby/2.6.3/lib/ruby/2.6.0
14:35:12 -r ./siteconf20200313-30573-18g5kc4.rb extconf.rb
14:35:12 checking if the C compiler accepts ... yes
14:35:12 Building nokogiri using packaged libraries.
14:35:12 Using mini_portile version 2.4.0
14:35:12 checking for gzdopen() in -lz... no
14:35:12 zlib is missing; necessary for building libxml2
14:35:12 *** extconf.rb failed ***
14:35:12 Could not create Makefile due to some reason, probably lack of necessary
14:35:12 libraries and/or headers. Check the mkmf.log file for more details. You may
14:35:12 need configuration options.
14:35:12
14:35:12 Provided configuration options:
14:35:12 --with-opt-dir
14:35:12 --with-opt-include
14:35:12 --without-opt-include=${opt-dir}/include
14:35:12 --with-opt-lib
14:35:12 --without-opt-lib=${opt-dir}/lib
14:35:12 --with-make-prog
14:35:12 --without-make-prog
14:35:12 --srcdir=.
14:35:12 --curdir
14:35:12 --ruby=/home/linuxbrew/.linuxbrew/Homebrew/Library/Homebrew/vendor/portable-ruby/2.6.3/bin/$(RUBY_BASE_NAME)
14:35:12 --help
14:35:12 --clean
14:35:12 --use-system-libraries
14:35:12 --enable-static
14:35:12 --disable-static
14:35:12 --with-zlib-dir
14:35:12 --without-zlib-dir
14:35:12 --with-zlib-include
14:35:12 --without-zlib-include=${zlib-dir}/include
14:35:12 --with-zlib-lib
14:35:12 --without-zlib-lib=${zlib-dir}/lib
14:35:12 --enable-cross-build
14:35:12 --disable-cross-build
14:35:12
14:35:12 To see why this extension failed to compile, please check the mkmf.log which can
14:35:12 be found here:
14:35:12
14:35:12 /home/linuxbrew/.linuxbrew/Homebrew/Library/Homebrew/vendor/bundle/ruby/2.6.0/extensions/x86_64-linux/2.6.0-static/nokogiri-1.10.9/mkmf.log
14:35:12
14:35:12 extconf failed, exit code 1
14:35:12
14:35:12 Gem files will remain installed in
14:35:12 /home/linuxbrew/.linuxbrew/Homebrew/Library/Homebrew/vendor/bundle/ruby/2.6.0/gems/nokogiri-1.10.9
14:35:12 for inspection.
14:35:12 Results logged to
14:35:12 /home/linuxbrew/.linuxbrew/Homebrew/Library/Homebrew/vendor/bundle/ruby/2.6.0/extensions/x86_64-linux/2.6.0-static/nokogiri-1.10.9/gem_make.out
14:35:12
14:35:12 An error occurred while installing nokogiri (1.10.9), and Bundler cannot
14:35:12 continue.
14:35:12 Make sure that `gem install nokogiri -v '1.10.9' --source
14:35:12 'https://rubygems.org/'` succeeds before bundling.
14:35:12
14:35:12 In Gemfile:
14:35:12 mechanize was resolved to 2.7.6, which depends on
14:35:12 nokogiri
14:35:12 Error: failed to run `/home/linuxbrew/.linuxbrew/Homebrew/Library/Homebrew/vendor/portable-ruby/2.6.3/bin/bundle install`!
14:35:12 Error: `brew audit` failed!
</pre>
</details>
|
1.0
|
Release process: run brew bump-formula-pr in docker - we need to run this command in a docker so it doesnt die of gem dependency
```
brew bump-formula-pr \
--strict minikube \
--revision="${revision}" \
--message="This PR was automatically created by minikube release scripts. Contact @tstromberg with any questions." \
--no-browse \
--tag="${TAG}" \
&& status=0 || status=$?
```
https://github.com/kubernetes/minikube/blob/eb13446e786c9ef70cb0a9f85a633194e62396a1/hack/jenkins/release_update_brew.sh#L48
here is the current gem dep error from out latest 1.8.2 release
<details>
<pre>
14:33:14 + brew bump-formula-pr --strict minikube --revision=eb13446e786c9ef70cb0a9f85a633194e62396a1 '--message=This PR was automatically created by minikube release scripts. Contact @tstromberg with any questions.' --no-browse --tag=v1.8.2
14:33:18 Updating Homebrew...
14:33:22 ==> Auto-updated Homebrew!
14:33:22 Updated Homebrew from 78140cef4 to db121c7b1.
14:35:02 Updated 1 tap (homebrew/core).
14:35:02 ==> New Formulae
14:35:02 adoptopenjdk
14:35:02 adoptopenjdk@11
14:35:02 alsa-lib
14:35:02 apt
14:35:02 archiver
14:35:02 at-spi2-atk
14:35:02 at-spi2-core
14:35:02 attr
14:35:02 avahi
14:35:02 aws-vault
14:35:02 bsdmainutils
14:35:02 cpio
14:35:02 device-mapper
14:35:02 elfutils
14:35:02 gcc@9
14:35:02 git-trim
14:35:02 glibc
14:35:02 hdf5@1.10
14:35:02 hdt
14:35:02 isl@0.18
14:35:02 jdk
14:35:02 jdk@7
14:35:02 jdk@8
14:35:02 ladspa-sdk
14:35:02 libaio
14:35:02 libbsd
14:35:02 libcap
14:35:02 libcap-ng
14:35:02 libfuse
14:35:02 libmnl
14:35:02 libnetfilter-queue
14:35:02 libnfnetlink
14:35:02 libnsl
14:35:02 libprelude
14:35:02 libsasl2
14:35:02 libseccomp
14:35:02 libtirpc
14:35:02 linux-headers
14:35:02 linux-pam
14:35:02 lm-sensors
14:35:02 minimal-python
14:35:02 mockolo
14:35:02 net-tools
14:35:02 newrelic-cli
14:35:02 onedrive
14:35:02 openjdk@10
14:35:02 openjdk@9
14:35:02 procps
14:35:02 python-dbus
14:35:02 python@2
14:35:02 qp
14:35:02 singularity
14:35:02 strace
14:35:02 sysstat
14:35:02 systemd
14:35:02 texlive
14:35:02 tlx
14:35:02 xclogparser
14:35:02 ==> Updated Formulae
14:35:02 abseil
14:35:02 advancemenu
14:35:02 adwaita-icon-theme
14:35:02 afuse
14:35:02 agda
14:35:02 algol68g
14:35:02 angular-cli
14:35:02 ansible
14:35:02 ansible-lint
14:35:02 apache-drill
14:35:02 apibuilder-cli
14:35:02 apng2gif
14:35:02 apr-util
14:35:02 arangodb
14:35:02 argon2
14:35:02 aria2
14:35:02 armadillo
14:35:02 arping
14:35:02 artifactory
14:35:02 asciidoc
14:35:02 asciinema
14:35:02 astrometry-net
14:35:02 atdtool
14:35:02 autoconf
14:35:02 aws-cdk
14:35:02 aws-elasticbeanstalk
14:35:02 aws-shell
14:35:02 awscli
14:35:02 awscli@1
14:35:02 awscurl
14:35:02 awslogs
14:35:02 awsume
14:35:02 azure-cli
14:35:02 azure-storage-cpp
14:35:02 bacula-fd
14:35:02 balena-cli
14:35:02 bandcamp-dl
14:35:02 bazaar
14:35:02 bde
14:35:02 benthos
14:35:02 biosig
14:35:02 bison
14:35:02 bitcoin
14:35:02 black
14:35:02 blazegraph
14:35:02 bonnie++
14:35:02 buildifier
14:35:02 bundletool
14:35:02 bzt
14:35:02 caffe
14:35:02 camlp5
14:35:02 cargo-completion
14:35:02 cartridge-cli
14:35:02 ceres-solver
14:35:02 certbot
14:35:02 cfn-lint
14:35:02 cfr-decompiler
14:35:02 cgns
14:35:02 cgrep
14:35:02 chakra
14:35:02 chapel
14:35:02 charm-tools
14:35:02 cheat
14:35:02 cli53
14:35:02 clojure
14:35:02 clojure-lsp
14:35:02 clutter
14:35:02 cobalt
14:35:02 cocoapods
14:35:02 codec2
14:35:02 codemod
14:35:02 codespell
14:35:02 cogl
14:35:02 comby
14:35:02 composer
14:35:02 conan
14:35:02 confluent-platform
14:35:02 conjure-up
14:35:02 cookiecutter
14:35:02 cpl
14:35:02 cppman
14:35:02 cquery
14:35:02 csound
14:35:02 csvkit
14:35:02 csvtomd
14:35:02 curl
14:35:02 curl-openssl
14:35:02 cvs
14:35:02 dartsim
14:35:02 davmail
14:35:02 db-vcs
14:35:02 dependency-check
14:35:02 derby
14:35:02 detekt
14:35:02 dialog
14:35:02 diceware
14:35:02 diff-pdf
14:35:02 diffoscope
14:35:02 dmd
14:35:02 docbook-xsl
14:35:02 docbook2x
14:35:02 docfx
14:35:02 docker
14:35:02 docker-completion
14:35:02 docker-compose
14:35:02 docker-slim
14:35:02 docker-squash
14:35:02 dosbox-x
14:35:02 double-conversion
14:35:02 dovecot
14:35:02 dtc
14:35:02 dub
14:35:02 dune
14:35:02 duplicity
14:35:02 dust
14:35:02 dxpy
14:35:02 dynare
14:35:02 e2fsprogs
14:35:02 easyrpg-player
14:35:02 ecasound
14:35:02 eccodes
14:35:02 eg-examples
14:35:02 ejdb
14:35:02 elasticsearch
14:35:02 embree
14:35:02 emscripten
14:35:02 ensmallen
14:35:02 esptool
14:35:02 etsh
14:35:02 evince
14:35:02 exim
14:35:02 expect
14:35:02 exploitdb
14:35:02 ext2fuse
14:35:02 ezstream
14:35:02 fairymax
14:35:02 faiss
14:35:02 fantom
14:35:02 fastlane
14:35:02 faudio
14:35:02 fcl
14:35:02 fdroidserver
14:35:02 file-roller
14:35:02 findomain
14:35:02 flann
14:35:02 flatbuffers
14:35:02 flintrock
14:35:02 fluent-bit
14:35:02 flyway
14:35:02 fn
14:35:02 folly
14:35:02 foma
14:35:02 fontconfig
14:35:02 fonttools
14:35:02 fq
14:35:02 frpc
14:35:02 frps
14:35:02 frugal
14:35:02 fselect
14:35:02 fzf
14:35:02 gambit
14:35:02 gambit-scheme
14:35:02 gammaray
14:35:02 gandi.cli
14:35:02 gatsby-cli
14:35:02 gcc
14:35:02 gcc@8
14:35:02 gdal
14:35:02 gdcm
14:35:02 geeqie
14:35:02 geocode-glib
14:35:02 geos
14:35:02 gerbil-scheme
14:35:02 gettext
14:35:02 gimme-aws-creds
14:35:02 git
14:35:02 git-annex
14:35:02 git-plus
14:35:02 git-review
14:35:02 git-revise
14:35:02 git-standup
14:35:02 gitmoji
14:35:02 gitup
14:35:02 gjs
14:35:02 glances
14:35:02 glib
14:35:02 glog
14:35:02 gmsh
14:35:02 gmt
14:35:02 gmt@5
14:35:02 gobject-introspection
14:35:02 google-sparsehash
14:35:02 gpgme
14:35:02 grin
14:35:02 grin-wallet
14:35:02 gromacs
14:35:02 groovy
14:35:02 groovysdk
14:35:02 gsettings-desktop-schemas
14:35:02 gti
14:35:02 gtkmm
14:35:02 gtksourceview4
14:35:02 hdf5
14:35:02 helm
14:35:02 hexyl
14:35:02 homeassistant-cli
14:35:02 hostess
14:35:02 howdoi
14:35:02 htop
14:35:02 httpie
14:35:02 hugo
14:35:02 hyperkit
14:35:02 ibex
14:35:02 ilmbase
14:35:02 imagemagick
14:35:02 imagemagick@6
14:35:02 imgproxy
14:35:02 inadyn
14:35:02 inlets
14:35:02 instalooter
14:35:02 internetarchive
14:35:02 intltool
14:35:02 io
14:35:02 ios-sim
14:35:02 ipmiutil
14:35:02 ipython
14:35:02 isl
14:35:02 itk
14:35:02 jbig2dec
14:35:02 jed
14:35:02 jenkins
14:35:02 jenkins-job-builder
14:35:02 jenkins-lts
14:35:02 jhead
14:35:02 jhipster
14:35:02 jinja2-cli
14:35:02 joplin
14:35:02 jrnl
14:35:02 jruby
14:35:02 juise
14:35:02 juju
14:35:02 juju-wait
14:35:02 jupyterlab
14:35:02 kallisto
14:35:02 keepkey-agent
14:35:02 khard
14:35:02 kibana
14:35:02 krew
14:35:02 ktoblzcheck
14:35:02 kubeprod
14:35:02 kubernetes-cli
14:35:02 kubeseal
14:35:02 kumo
14:35:02 lablgtk
14:35:02 lammps
14:35:02 ldc
14:35:02 legit
14:35:02 lgogdownloader
14:35:02 libbi
14:35:02 libcouchbase
14:35:02 libdazzle
14:35:02 libewf
14:35:02 libexosip
14:35:02 libfabric
14:35:02 libgetdata
14:35:02 libgosu
14:35:02 libgweather
14:35:02 libhttpseverywhere
14:35:02 libical
14:35:02 libmatio
14:35:02 libowfat
14:35:02 libpeas
14:35:02 libproxy
14:35:02 libpsl
14:35:02 libqalculate
14:35:02 librealsense
14:35:02 libsecret
14:35:02 libserdes
14:35:02 libsoup
14:35:02 libssh
14:35:02 libsvm
14:35:02 libtcod
14:35:02 libtool
14:35:02 libtorrent-rasterbar
14:35:02 libupnp
14:35:02 libwebsockets
14:35:02 libxml++
14:35:02 libzdb
14:35:02 liquibase
14:35:02 literate-git
14:35:02 llvm
14:35:02 llvm@6
14:35:02 llvm@7
14:35:02 llvm@8
14:35:02 lolcat
14:35:02 lua
14:35:02 lua@5.1
14:35:02 luaradio
14:35:02 luarocks
14:35:02 luit
14:35:02 mackup
14:35:02 magic-wormhole
14:35:02 mapserver
14:35:02 mbedtls
14:35:02 mdds
14:35:02 mdv
14:35:02 memcached
14:35:02 menhir
14:35:02 mercurial
14:35:02 mercury
14:35:02 mesa
14:35:02 meson-internal
14:35:02 metabase
14:35:02 micronaut
14:35:02 minikube
14:35:02 minio
14:35:02 minizinc
14:35:02 minizip2
14:35:02 mitmproxy
14:35:02 mkdocs
14:35:02 mkl-dnn
14:35:02 mkvtoolnix
14:35:02 mlpack
14:35:02 molecule
14:35:02 mongo-cxx-driver
14:35:02 mongo-orchestration
14:35:02 mosh
14:35:02 mosquitto
14:35:02 mpd
14:35:02 mruby
14:35:02 mujs
14:35:02 mycli
14:35:02 mypy
14:35:02 mysql-connector-c++
14:35:02 n
14:35:02 nagios-plugins
14:35:02 nbdime
14:35:02 nco
14:35:02 ncview
14:35:02 neofetch
14:35:02 neovim
14:35:02 netcdf
14:35:02 nethack
14:35:02 ninja
14:35:02 nng
14:35:02 node-sass
14:35:02 nodeenv
14:35:02 nss
14:35:02 nudoku
14:35:02 nushell
14:35:02 oauth2l
14:35:02 ocaml
14:35:02 ocaml-findlib
14:35:02 ocaml-num
14:35:02 ocamlbuild
14:35:02 ocamlsdl
14:35:02 ocrmypdf
14:35:02 octave
14:35:02 ode
14:35:02 odpi
14:35:02 onnxruntime
14:35:02 open-image-denoise
14:35:02 open-mpi
14:35:02 openexr
14:35:02 openssl@1.1
14:35:02 pastel
14:35:02 pcl
14:35:02 pdal
14:35:02 pgcli
14:35:02 pipenv
14:35:02 pipx
14:35:02 pkg-config
14:35:02 platformio
14:35:02 plotutils
14:35:02 pmd
14:35:02 pnpm
14:35:02 points2grid
14:35:02 pre-commit
14:35:02 procs
14:35:02 prodigal
14:35:02 proselint
14:35:02 protoc-gen-go
14:35:02 psc-package
14:35:02 pssh
14:35:02 pulseaudio
14:35:02 pulumi
14:35:02 pushpin
14:35:02 pwntools
14:35:02 pyenv
14:35:02 pygitup
14:35:02 pygments
14:35:02 pygobject3
14:35:02 pyinvoke
14:35:02 pylint
14:35:02 python
14:35:02 python@3.8
14:35:02 pyvim
14:35:02 qalculate-gtk
14:35:02 qt
14:35:02 rabbitmq
14:35:02 radamsa
14:35:02 radare2
14:35:02 rakudo-star
14:35:02 re-flex
14:35:02 remarshal
14:35:02 rex
14:35:02 rst-lint
14:35:02 rtv
14:35:02 rubberband
14:35:02 rust
14:35:02 s3cmd
14:35:02 s3fs
14:35:02 s6
14:35:02 sbjson
14:35:02 sceptre
14:35:02 scons
14:35:02 scour
14:35:02 sdcc
14:35:02 sdedit
14:35:02 sdl2
14:35:02 serf
14:35:02 serverless
14:35:02 sfk
14:35:02 shogun
14:35:02 simg2img
14:35:02 simple-tiles
14:35:02 sjk
14:35:02 skinny
14:35:02 sn0int
14:35:02 sonar-scanner
14:35:02 sonarqube
14:35:02 sphinx-doc
14:35:02 spice-protocol
14:35:02 spidermonkey
14:35:02 spigot
14:35:02 spirv-cross
14:35:02 spirv-tools
14:35:02 sqliteodbc
14:35:02 sqlparse
14:35:02 ssdb
14:35:02 ssh-audit
14:35:02 sshuttle
14:35:02 statik
14:35:02 stdman
14:35:02 stellar-core
14:35:02 stgit
14:35:02 stolon
14:35:02 streamlink
14:35:02 stress-ng
14:35:02 stubby
14:35:02 subversion
14:35:02 suite-sparse
14:35:02 sundials
14:35:02 supervisor
14:35:02 svtplay-dl
14:35:02 swagger-codegen
14:35:02 swift-protobuf
14:35:02 swift-sh
14:35:02 swiftformat
14:35:02 switch-lan-play
14:35:02 talisman
14:35:02 tasksh
14:35:02 tbox
14:35:02 tektoncd-cli
14:35:02 terragrunt
14:35:02 terrahelp
14:35:02 terrahub
14:35:02 texinfo
14:35:02 texmath
14:35:02 tgui
14:35:02 thefuck
14:35:02 tin
14:35:02 tinyxml
14:35:02 tinyxml2
14:35:02 todoman
14:35:02 tox
14:35:02 trafficserver
14:35:02 translate-toolkit
14:35:02 trash-cli
14:35:02 ttyd
14:35:02 twarc
14:35:02 twine-pypi
14:35:02 txr
14:35:02 txt2tags
14:35:02 u-boot-tools
14:35:02 urh
14:35:02 v2ray-plugin
14:35:02 v8
14:35:02 vala
14:35:02 vale
14:35:02 valgrind
14:35:02 vault
14:35:02 velero
14:35:02 verilator
14:35:02 vips
14:35:02 visp
14:35:02 vowpal-wabbit
14:35:02 vsts-cli
14:35:02 vtk
14:35:02 vttest
14:35:02 vulkan-headers
14:35:02 wabt
14:35:02 wal2json
14:35:02 wapm
14:35:02 wartremover
14:35:02 wasmer
14:35:02 watson
14:35:02 wcslib
14:35:02 webdis
14:35:02 websocat
14:35:02 webtorrent-cli
14:35:02 wp-cli-completion
14:35:02 wpscan
14:35:02 wxmaxima
14:35:02 x3270
14:35:02 xclip
14:35:02 xdotool
14:35:02 xmlto
14:35:02 xmrig
14:35:02 xonsh
14:35:02 xtensor
14:35:02 xxhash
14:35:02 yamllint
14:35:02 yapf
14:35:02 yarn
14:35:02 ydcv
14:35:02 ykman
14:35:02 you-get
14:35:02 youtube-dl
14:35:02 z
14:35:02 zabbix-cli
14:35:02 zeek
14:35:02 zimg
14:35:02 znc
14:35:02 zola
14:35:02 ==> Deleted Formulae
14:35:02 camlp4
14:35:02 ck
14:35:02 djmount
14:35:02 field3d
14:35:02 gmediaserver
14:35:02
14:35:03 ==> Fetching homebrew/master
14:35:06 From https://github.com/Homebrew/homebrew-core
14:35:06 * branch master -> FETCH_HEAD
14:35:06 257bb3e8cb..d3eb71b109 master -> homebrew/master
14:35:06 ==> minikube exists in homebrew/master
14:35:07 Previous HEAD position was a99a5e418f... Merge branch homebrew/master into linuxbrew/master
14:35:07 HEAD is now at d3eb71b109... composer 1.10.1 (#51704)
14:35:07 ==> replace "v1.8.1" with "v1.8.2"
14:35:07 ==> replace "cbda04cf6bbe65e987ae52bb393c10099ab62014" with "eb13446e786c9ef70cb0a9f85a633194e62396a1"
14:35:09 Fetching gem metadata from https://rubygems.org/.........
14:35:10 Using concurrent-ruby 1.1.6
14:35:10 Using minitest 5.14.0
14:35:10 Using thread_safe 0.3.6
14:35:10 Using zeitwerk 2.3.0
14:35:10 Using ast 2.4.0
14:35:10 Using bundler 1.17.2
14:35:10 Using connection_pool 2.2.2
14:35:10 Using json 2.3.0
14:35:10 Using docile 1.3.2
14:35:10 Using simplecov-html 0.10.2
14:35:10 Using sync 0.5.0
14:35:10 Using thor 1.0.1
14:35:10 Using diff-lcs 1.3
14:35:10 Using unf_ext 0.0.7.6
14:35:10 Using hpricot 0.8.6
14:35:10 Using jaro_winkler 1.5.4
14:35:10 Using mime-types-data 3.2019.1009
14:35:10 Using net-http-digest_auth 1.4.1
14:35:10 Using mini_portile2 2.4.0
14:35:10 Using ntlm-http 0.1.1
14:35:10 Using webrobots 0.1.2
14:35:10 Using mustache 1.1.1
14:35:10 Using parallel 1.19.1
14:35:10 Using plist 3.5.0
14:35:10 Using rainbow 3.0.0
14:35:10 Using rdiscount 2.2.0.1
14:35:10 Using rexml 3.2.4
14:35:10 Using rspec-support 3.9.2
14:35:10 Using ruby-progressbar 1.10.1
14:35:10 Using unicode-display_width 1.6.1
14:35:10 Using ruby-macho 2.2.0
14:35:10 Using i18n 1.8.2
14:35:10 Using tzinfo 1.2.6
14:35:10 Using parser 2.7.0.4
14:35:10 Using tins 1.24.1
14:35:10 Using net-http-persistent 3.1.0
14:35:10 Using unf 0.1.4
14:35:10 Using mime-types 3.3.1
14:35:10 Using simplecov 0.16.1
14:35:10 Using ronn 0.7.3
14:35:10 Using rspec-core 3.9.1
14:35:10 Using rspec-expectations 3.9.0
14:35:10 Using rspec-mocks 3.9.1
14:35:10 Using activesupport 6.0.2.1
14:35:10 Using parallel_tests 2.31.0
14:35:10 Using rubocop 0.80.1
14:35:10 Using term-ansicolor 1.7.1
14:35:10 Fetching nokogiri 1.10.9
14:35:10 Using coveralls 0.8.23
14:35:10 Using domain_name 0.5.20190701
14:35:10 Using rspec 3.9.0
14:35:10 Using rspec-its 1.3.0
14:35:10 Using rspec-retry 0.6.2
14:35:10 Using rubocop-performance 1.5.2
14:35:10 Using rubocop-rspec 1.38.1
14:35:10 Using http-cookie 1.0.3
14:35:10 Using rspec-wait 0.0.9
14:35:10 Installing nokogiri 1.10.9 with native extensions
14:35:12 Gem::Ext::BuildError: ERROR: Failed to build gem native extension.
14:35:12
14:35:12 current directory:
14:35:12 /home/linuxbrew/.linuxbrew/Homebrew/Library/Homebrew/vendor/bundle/ruby/2.6.0/gems/nokogiri-1.10.9/ext/nokogiri
14:35:12 /home/linuxbrew/.linuxbrew/Homebrew/Library/Homebrew/vendor/portable-ruby/2.6.3/bin/ruby
14:35:12 -I
14:35:12 /home/linuxbrew/.linuxbrew/Homebrew/Library/Homebrew/vendor/portable-ruby/2.6.3/lib/ruby/2.6.0
14:35:12 -r ./siteconf20200313-30573-18g5kc4.rb extconf.rb
14:35:12 checking if the C compiler accepts ... yes
14:35:12 Building nokogiri using packaged libraries.
14:35:12 Using mini_portile version 2.4.0
14:35:12 checking for gzdopen() in -lz... no
14:35:12 zlib is missing; necessary for building libxml2
14:35:12 *** extconf.rb failed ***
14:35:12 Could not create Makefile due to some reason, probably lack of necessary
14:35:12 libraries and/or headers. Check the mkmf.log file for more details. You may
14:35:12 need configuration options.
14:35:12
14:35:12 Provided configuration options:
14:35:12 --with-opt-dir
14:35:12 --with-opt-include
14:35:12 --without-opt-include=${opt-dir}/include
14:35:12 --with-opt-lib
14:35:12 --without-opt-lib=${opt-dir}/lib
14:35:12 --with-make-prog
14:35:12 --without-make-prog
14:35:12 --srcdir=.
14:35:12 --curdir
14:35:12 --ruby=/home/linuxbrew/.linuxbrew/Homebrew/Library/Homebrew/vendor/portable-ruby/2.6.3/bin/$(RUBY_BASE_NAME)
14:35:12 --help
14:35:12 --clean
14:35:12 --use-system-libraries
14:35:12 --enable-static
14:35:12 --disable-static
14:35:12 --with-zlib-dir
14:35:12 --without-zlib-dir
14:35:12 --with-zlib-include
14:35:12 --without-zlib-include=${zlib-dir}/include
14:35:12 --with-zlib-lib
14:35:12 --without-zlib-lib=${zlib-dir}/lib
14:35:12 --enable-cross-build
14:35:12 --disable-cross-build
14:35:12
14:35:12 To see why this extension failed to compile, please check the mkmf.log which can
14:35:12 be found here:
14:35:12
14:35:12 /home/linuxbrew/.linuxbrew/Homebrew/Library/Homebrew/vendor/bundle/ruby/2.6.0/extensions/x86_64-linux/2.6.0-static/nokogiri-1.10.9/mkmf.log
14:35:12
14:35:12 extconf failed, exit code 1
14:35:12
14:35:12 Gem files will remain installed in
14:35:12 /home/linuxbrew/.linuxbrew/Homebrew/Library/Homebrew/vendor/bundle/ruby/2.6.0/gems/nokogiri-1.10.9
14:35:12 for inspection.
14:35:12 Results logged to
14:35:12 /home/linuxbrew/.linuxbrew/Homebrew/Library/Homebrew/vendor/bundle/ruby/2.6.0/extensions/x86_64-linux/2.6.0-static/nokogiri-1.10.9/gem_make.out
14:35:12
14:35:12 An error occurred while installing nokogiri (1.10.9), and Bundler cannot
14:35:12 continue.
14:35:12 Make sure that `gem install nokogiri -v '1.10.9' --source
14:35:12 'https://rubygems.org/'` succeeds before bundling.
14:35:12
14:35:12 In Gemfile:
14:35:12 mechanize was resolved to 2.7.6, which depends on
14:35:12 nokogiri
14:35:12 Error: failed to run `/home/linuxbrew/.linuxbrew/Homebrew/Library/Homebrew/vendor/portable-ruby/2.6.3/bin/bundle install`!
14:35:12 Error: `brew audit` failed!
</pre>
</details>
|
process
|
release process run brew bump formula pr in docker we need to run this command in a docker so it doesnt die of gem dependency brew bump formula pr strict minikube revision revision message this pr was automatically created by minikube release scripts contact tstromberg with any questions no browse tag tag status status here is the current gem dep error from out latest release brew bump formula pr strict minikube revision message this pr was automatically created by minikube release scripts contact tstromberg with any questions no browse tag updating homebrew auto updated homebrew updated homebrew from to updated tap homebrew core new formulae adoptopenjdk adoptopenjdk alsa lib apt archiver at atk at core attr avahi aws vault bsdmainutils cpio device mapper elfutils gcc git trim glibc hdt isl jdk jdk jdk ladspa sdk libaio libbsd libcap libcap ng libfuse libmnl libnetfilter queue libnfnetlink libnsl libprelude libseccomp libtirpc linux headers linux pam lm sensors minimal python mockolo net tools newrelic cli onedrive openjdk openjdk procps python dbus python qp singularity strace sysstat systemd texlive tlx xclogparser updated formulae abseil advancemenu adwaita icon theme afuse agda angular cli ansible ansible lint apache drill apibuilder cli apr util arangodb armadillo arping artifactory asciidoc asciinema astrometry net atdtool autoconf aws cdk aws elasticbeanstalk aws shell awscli awscli awscurl awslogs awsume azure cli azure storage cpp bacula fd balena cli bandcamp dl bazaar bde benthos biosig bison bitcoin black blazegraph bonnie buildifier bundletool bzt caffe cargo completion cartridge cli ceres solver certbot cfn lint cfr decompiler cgns cgrep chakra chapel charm tools cheat clojure clojure lsp clutter cobalt cocoapods codemod codespell cogl comby composer conan confluent platform conjure up cookiecutter cpl cppman cquery csound csvkit csvtomd curl curl openssl cvs dartsim davmail db vcs dependency check derby detekt dialog diceware diff pdf diffoscope dmd docbook xsl docfx docker docker completion docker compose docker slim docker squash dosbox x double conversion dovecot dtc dub dune duplicity dust dxpy dynare easyrpg player ecasound eccodes eg examples ejdb elasticsearch embree emscripten ensmallen esptool etsh evince exim expect exploitdb ezstream fairymax faiss fantom fastlane faudio fcl fdroidserver file roller findomain flann flatbuffers flintrock fluent bit flyway fn folly foma fontconfig fonttools fq frpc frps frugal fselect fzf gambit gambit scheme gammaray gandi cli gatsby cli gcc gcc gdal gdcm geeqie geocode glib geos gerbil scheme gettext gimme aws creds git git annex git plus git review git revise git standup gitmoji gitup gjs glances glib glog gmsh gmt gmt gobject introspection google sparsehash gpgme grin grin wallet gromacs groovy groovysdk gsettings desktop schemas gti gtkmm helm hexyl homeassistant cli hostess howdoi htop httpie hugo hyperkit ibex ilmbase imagemagick imagemagick imgproxy inadyn inlets instalooter internetarchive intltool io ios sim ipmiutil ipython isl itk jed jenkins jenkins job builder jenkins lts jhead jhipster cli joplin jrnl jruby juise juju juju wait jupyterlab kallisto keepkey agent khard kibana krew ktoblzcheck kubeprod kubernetes cli kubeseal kumo lablgtk lammps ldc legit lgogdownloader libbi libcouchbase libdazzle libewf libexosip libfabric libgetdata libgosu libgweather libhttpseverywhere libical libmatio libowfat libpeas libproxy libpsl libqalculate librealsense libsecret libserdes libsoup libssh libsvm libtcod libtool libtorrent rasterbar libupnp libwebsockets libxml libzdb liquibase literate git llvm llvm llvm llvm lolcat lua lua luaradio luarocks luit mackup magic wormhole mapserver mbedtls mdds mdv memcached menhir mercurial mercury mesa meson internal metabase micronaut minikube minio minizinc mitmproxy mkdocs mkl dnn mkvtoolnix mlpack molecule mongo cxx driver mongo orchestration mosh mosquitto mpd mruby mujs mycli mypy mysql connector c n nagios plugins nbdime nco ncview neofetch neovim netcdf nethack ninja nng node sass nodeenv nss nudoku nushell ocaml ocaml findlib ocaml num ocamlbuild ocamlsdl ocrmypdf octave ode odpi onnxruntime open image denoise open mpi openexr openssl pastel pcl pdal pgcli pipenv pipx pkg config platformio plotutils pmd pnpm pre commit procs prodigal proselint protoc gen go psc package pssh pulseaudio pulumi pushpin pwntools pyenv pygitup pygments pyinvoke pylint python python pyvim qalculate gtk qt rabbitmq radamsa rakudo star re flex remarshal rex rst lint rtv rubberband rust sbjson sceptre scons scour sdcc sdedit serf serverless sfk shogun simple tiles sjk skinny sonar scanner sonarqube sphinx doc spice protocol spidermonkey spigot spirv cross spirv tools sqliteodbc sqlparse ssdb ssh audit sshuttle statik stdman stellar core stgit stolon streamlink stress ng stubby subversion suite sparse sundials supervisor svtplay dl swagger codegen swift protobuf swift sh swiftformat switch lan play talisman tasksh tbox tektoncd cli terragrunt terrahelp terrahub texinfo texmath tgui thefuck tin tinyxml todoman tox trafficserver translate toolkit trash cli ttyd twarc twine pypi txr u boot tools urh plugin vala vale valgrind vault velero verilator vips visp vowpal wabbit vsts cli vtk vttest vulkan headers wabt wapm wartremover wasmer watson wcslib webdis websocat webtorrent cli wp cli completion wpscan wxmaxima xclip xdotool xmlto xmrig xonsh xtensor xxhash yamllint yapf yarn ydcv ykman you get youtube dl z zabbix cli zeek zimg znc zola deleted formulae ck djmount gmediaserver fetching homebrew master from branch master fetch head master homebrew master minikube exists in homebrew master previous head position was merge branch homebrew master into linuxbrew master head is now at composer replace with replace with fetching gem metadata from using concurrent ruby using minitest using thread safe using zeitwerk using ast using bundler using connection pool using json using docile using simplecov html using sync using thor using diff lcs using unf ext using hpricot using jaro winkler using mime types data using net http digest auth using mini using ntlm http using webrobots using mustache using parallel using plist using rainbow using rdiscount using rexml using rspec support using ruby progressbar using unicode display width using ruby macho using using tzinfo using parser using tins using net http persistent using unf using mime types using simplecov using ronn using rspec core using rspec expectations using rspec mocks using activesupport using parallel tests using rubocop using term ansicolor fetching nokogiri using coveralls using domain name using rspec using rspec its using rspec retry using rubocop performance using rubocop rspec using http cookie using rspec wait installing nokogiri with native extensions gem ext builderror error failed to build gem native extension current directory home linuxbrew linuxbrew homebrew library homebrew vendor bundle ruby gems nokogiri ext nokogiri home linuxbrew linuxbrew homebrew library homebrew vendor portable ruby bin ruby i home linuxbrew linuxbrew homebrew library homebrew vendor portable ruby lib ruby r rb extconf rb checking if the c compiler accepts yes building nokogiri using packaged libraries using mini portile version checking for gzdopen in lz no zlib is missing necessary for building extconf rb failed could not create makefile due to some reason probably lack of necessary libraries and or headers check the mkmf log file for more details you may need configuration options provided configuration options with opt dir with opt include without opt include opt dir include with opt lib without opt lib opt dir lib with make prog without make prog srcdir curdir ruby home linuxbrew linuxbrew homebrew library homebrew vendor portable ruby bin ruby base name help clean use system libraries enable static disable static with zlib dir without zlib dir with zlib include without zlib include zlib dir include with zlib lib without zlib lib zlib dir lib enable cross build disable cross build to see why this extension failed to compile please check the mkmf log which can be found here home linuxbrew linuxbrew homebrew library homebrew vendor bundle ruby extensions linux static nokogiri mkmf log extconf failed exit code gem files will remain installed in home linuxbrew linuxbrew homebrew library homebrew vendor bundle ruby gems nokogiri for inspection results logged to home linuxbrew linuxbrew homebrew library homebrew vendor bundle ruby extensions linux static nokogiri gem make out an error occurred while installing nokogiri and bundler cannot continue make sure that gem install nokogiri v source succeeds before bundling in gemfile mechanize was resolved to which depends on nokogiri error failed to run home linuxbrew linuxbrew homebrew library homebrew vendor portable ruby bin bundle install error brew audit failed
| 1
|
70,877
| 18,289,823,649
|
IssuesEvent
|
2021-10-05 14:11:27
|
google/or-tools
|
https://api.github.com/repos/google/or-tools
|
closed
|
Fedora 34 build: `CMake Error at cmake/FindSCIP.cmake:68 (set_property): set_property not given a PROPERTY <name> argument.`
|
Bug Build: CMake Solver: SCIP
|
**What version of OR-Tools and what language are you using?**
Version: stable
Language: C++
**Which solver are you using (e.g. CP-SAT, Routing Solver, GLOP, BOP, Gurobi)**
SCIP
**What operating system (Linux, Windows, ...) and version?**
Fedora 34, cmake version 3.20.5
**What did you do?**
Steps to reproduce the behavior:
I tried to build the `stable` branch like so:
```
git clone https://github.com/google/or-tools.git && cd or-tools
mkdir build && cd build && SCIP_ROOT=/usr/local/ cmake ..
```
**What did you expect to see**
A successful build
**What did you see instead?**
The build fails with the following error messages:
<pre>
-- SCIP_ROOT: /usr/local/
-- TARGET_PROPERTY:
CMake Error at cmake/FindSCIP.cmake:68 (set_property):
<b>set_property not given a PROPERTY <name> argument. </b>
Call Stack (most recent call first):
cmake/deps.cmake:53 (find_package)
CMakeLists.txt:246 (include)
</pre>
Note: I posted this issue as a problem about linking the pthreads library before. I edited it now, because I think the issue is in fact a different one - namely the Error above.
|
1.0
|
Fedora 34 build: `CMake Error at cmake/FindSCIP.cmake:68 (set_property): set_property not given a PROPERTY <name> argument.` - **What version of OR-Tools and what language are you using?**
Version: stable
Language: C++
**Which solver are you using (e.g. CP-SAT, Routing Solver, GLOP, BOP, Gurobi)**
SCIP
**What operating system (Linux, Windows, ...) and version?**
Fedora 34, cmake version 3.20.5
**What did you do?**
Steps to reproduce the behavior:
I tried to build the `stable` branch like so:
```
git clone https://github.com/google/or-tools.git && cd or-tools
mkdir build && cd build && SCIP_ROOT=/usr/local/ cmake ..
```
**What did you expect to see**
A successful build
**What did you see instead?**
The build fails with the following error messages:
<pre>
-- SCIP_ROOT: /usr/local/
-- TARGET_PROPERTY:
CMake Error at cmake/FindSCIP.cmake:68 (set_property):
<b>set_property not given a PROPERTY <name> argument. </b>
Call Stack (most recent call first):
cmake/deps.cmake:53 (find_package)
CMakeLists.txt:246 (include)
</pre>
Note: I posted this issue as a problem about linking the pthreads library before. I edited it now, because I think the issue is in fact a different one - namely the Error above.
|
non_process
|
fedora build cmake error at cmake findscip cmake set property set property not given a property argument what version of or tools and what language are you using version stable language c which solver are you using e g cp sat routing solver glop bop gurobi scip what operating system linux windows and version fedora cmake version what did you do steps to reproduce the behavior i tried to build the stable branch like so git clone cd or tools mkdir build cd build scip root usr local cmake what did you expect to see a successful build what did you see instead the build fails with the following error messages scip root usr local target property cmake error at cmake findscip cmake set property set property not given a property argument call stack most recent call first cmake deps cmake find package cmakelists txt include note i posted this issue as a problem about linking the pthreads library before i edited it now because i think the issue is in fact a different one namely the error above
| 0
|
273,754
| 29,831,080,086
|
IssuesEvent
|
2023-06-18 09:28:42
|
RG4421/ampere-centos-kernel
|
https://api.github.com/repos/RG4421/ampere-centos-kernel
|
closed
|
CVE-2019-11599 (High) detected in linuxv5.2 - autoclosed
|
Mend: dependency security vulnerability
|
## CVE-2019-11599 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv5.2</b></p></summary>
<p>
<p>Linux kernel source tree</p>
<p>Library home page: <a href=https://github.com/torvalds/linux.git>https://github.com/torvalds/linux.git</a></p>
<p>Found in base branch: <b>amp-centos-8.0-kernel</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/proc/task_mmu.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/proc/task_mmu.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
The coredump implementation in the Linux kernel before 5.0.10 does not use locking or other mechanisms to prevent vma layout or vma flags changes while it runs, which allows local users to obtain sensitive information, cause a denial of service, or possibly have unspecified other impact by triggering a race condition with mmget_not_zero or get_task_mm calls. This is related to fs/userfaultfd.c, mm/mmap.c, fs/proc/task_mmu.c, and drivers/infiniband/core/uverbs_main.c.
<p>Publish Date: 2019-04-29
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-11599>CVE-2019-11599</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.0</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11599">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11599</a></p>
<p>Release Date: 2019-04-29</p>
<p>Fix Resolution: v5.1-rc6</p>
</p>
</details>
<p></p>
|
True
|
CVE-2019-11599 (High) detected in linuxv5.2 - autoclosed - ## CVE-2019-11599 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv5.2</b></p></summary>
<p>
<p>Linux kernel source tree</p>
<p>Library home page: <a href=https://github.com/torvalds/linux.git>https://github.com/torvalds/linux.git</a></p>
<p>Found in base branch: <b>amp-centos-8.0-kernel</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/proc/task_mmu.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/proc/task_mmu.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
The coredump implementation in the Linux kernel before 5.0.10 does not use locking or other mechanisms to prevent vma layout or vma flags changes while it runs, which allows local users to obtain sensitive information, cause a denial of service, or possibly have unspecified other impact by triggering a race condition with mmget_not_zero or get_task_mm calls. This is related to fs/userfaultfd.c, mm/mmap.c, fs/proc/task_mmu.c, and drivers/infiniband/core/uverbs_main.c.
<p>Publish Date: 2019-04-29
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-11599>CVE-2019-11599</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.0</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11599">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11599</a></p>
<p>Release Date: 2019-04-29</p>
<p>Fix Resolution: v5.1-rc6</p>
</p>
</details>
<p></p>
|
non_process
|
cve high detected in autoclosed cve high severity vulnerability vulnerable library linux kernel source tree library home page a href found in base branch amp centos kernel vulnerable source files fs proc task mmu c fs proc task mmu c vulnerability details the coredump implementation in the linux kernel before does not use locking or other mechanisms to prevent vma layout or vma flags changes while it runs which allows local users to obtain sensitive information cause a denial of service or possibly have unspecified other impact by triggering a race condition with mmget not zero or get task mm calls this is related to fs userfaultfd c mm mmap c fs proc task mmu c and drivers infiniband core uverbs main c publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution
| 0
|
20,255
| 26,874,022,498
|
IssuesEvent
|
2023-02-04 20:31:17
|
SatDump/SatDump
|
https://api.github.com/repos/SatDump/SatDump
|
opened
|
L1b product quality flags
|
enhancement Processing
|
We need quality flags to make any L2 products, or trust the generated products for research purposes.
- [ ] AVHRR
- [ ] AMSU-A
- [ ] MHS
|
1.0
|
L1b product quality flags - We need quality flags to make any L2 products, or trust the generated products for research purposes.
- [ ] AVHRR
- [ ] AMSU-A
- [ ] MHS
|
process
|
product quality flags we need quality flags to make any products or trust the generated products for research purposes avhrr amsu a mhs
| 1
|
21,923
| 30,446,460,024
|
IssuesEvent
|
2023-07-15 18:31:24
|
h4sh5/pypi-auto-scanner
|
https://api.github.com/repos/h4sh5/pypi-auto-scanner
|
opened
|
pyutils 0.0.1b18 has 2 GuardDog issues
|
guarddog typosquatting silent-process-execution
|
https://pypi.org/project/pyutils
https://inspector.pypi.io/project/pyutils
```{
"dependency": "pyutils",
"version": "0.0.1b18",
"result": {
"issues": 2,
"errors": {},
"results": {
"typosquatting": "This package closely ressembles the following package names, and might be a typosquatting attempt: pytils, python-utils",
"silent-process-execution": [
{
"location": "pyutils/exec_utils.py/pyutils/exec_utils.py:214",
"code": " subproc = subprocess.Popen(\n args,\n stdin=subprocess.DEVNULL,\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n )",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
}
]
},
"path": "/tmp/tmpooezcz_t/pyutils"
}
}```
|
1.0
|
pyutils 0.0.1b18 has 2 GuardDog issues - https://pypi.org/project/pyutils
https://inspector.pypi.io/project/pyutils
```{
"dependency": "pyutils",
"version": "0.0.1b18",
"result": {
"issues": 2,
"errors": {},
"results": {
"typosquatting": "This package closely ressembles the following package names, and might be a typosquatting attempt: pytils, python-utils",
"silent-process-execution": [
{
"location": "pyutils/exec_utils.py/pyutils/exec_utils.py:214",
"code": " subproc = subprocess.Popen(\n args,\n stdin=subprocess.DEVNULL,\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n )",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
}
]
},
"path": "/tmp/tmpooezcz_t/pyutils"
}
}```
|
process
|
pyutils has guarddog issues dependency pyutils version result issues errors results typosquatting this package closely ressembles the following package names and might be a typosquatting attempt pytils python utils silent process execution location pyutils exec utils py pyutils exec utils py code subproc subprocess popen n args n stdin subprocess devnull n stdout subprocess devnull n stderr subprocess devnull n message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null path tmp tmpooezcz t pyutils
| 1
|
1,701
| 4,349,710,064
|
IssuesEvent
|
2016-07-30 19:07:13
|
pwittchen/ReactiveNetwork
|
https://api.github.com/repos/pwittchen/ReactiveNetwork
|
closed
|
Release 0.5.1
|
release process
|
**Initial release notes**:
- bumped RxJava to v. 1.1.8
**Things to do**:
- [x] bump library version to 0.5.1
- [x] upload Archives to Maven Central Repository
- [x] close and release artifact on Nexus
- [x] update `CHANGELOG.md` after Maven Sync
- [x] bump library version in `README.md` after Maven Sync
- [x] create new GitHub release
|
1.0
|
Release 0.5.1 - **Initial release notes**:
- bumped RxJava to v. 1.1.8
**Things to do**:
- [x] bump library version to 0.5.1
- [x] upload Archives to Maven Central Repository
- [x] close and release artifact on Nexus
- [x] update `CHANGELOG.md` after Maven Sync
- [x] bump library version in `README.md` after Maven Sync
- [x] create new GitHub release
|
process
|
release initial release notes bumped rxjava to v things to do bump library version to upload archives to maven central repository close and release artifact on nexus update changelog md after maven sync bump library version in readme md after maven sync create new github release
| 1
|
13,251
| 15,722,798,243
|
IssuesEvent
|
2021-03-29 06:28:10
|
Ultimate-Hosts-Blacklist/whitelist
|
https://api.github.com/repos/Ultimate-Hosts-Blacklist/whitelist
|
closed
|
[FALSE-POSITIVE?] fossbytes.com
|
whitelisting process
|
**Domains or links**
- fossbytes.com
**More Information**
```
Exact match for fossbytes.com found in:
- https://raw.githubusercontent.com/Ultimate-Hosts-Blacklist/Ultimate.Hosts.Blacklist/master/domains/domains0.list
```
**Have you requested removal from other sources?**
Nope.
**Additional context**
N/A, but happy to provide if necessary
|
1.0
|
[FALSE-POSITIVE?] fossbytes.com - **Domains or links**
- fossbytes.com
**More Information**
```
Exact match for fossbytes.com found in:
- https://raw.githubusercontent.com/Ultimate-Hosts-Blacklist/Ultimate.Hosts.Blacklist/master/domains/domains0.list
```
**Have you requested removal from other sources?**
Nope.
**Additional context**
N/A, but happy to provide if necessary
|
process
|
fossbytes com domains or links fossbytes com more information exact match for fossbytes com found in have you requested removal from other sources nope additional context n a but happy to provide if necessary
| 1
|
9,894
| 12,891,758,134
|
IssuesEvent
|
2020-07-13 18:19:21
|
google/ground-android
|
https://api.github.com/repos/google/ground-android
|
opened
|
[UXR] Research open questions during pilot
|
priority: p2 type: process
|
- [ ] Is context of view/add/edit observation clear (i.e. which layer and which point is selected), or do we need breadcrumbs, or some other affordance in the header, etc.?
@sergeydolgov1 please add other open questions above.
|
1.0
|
[UXR] Research open questions during pilot - - [ ] Is context of view/add/edit observation clear (i.e. which layer and which point is selected), or do we need breadcrumbs, or some other affordance in the header, etc.?
@sergeydolgov1 please add other open questions above.
|
process
|
research open questions during pilot is context of view add edit observation clear i e which layer and which point is selected or do we need breadcrumbs or some other affordance in the header etc please add other open questions above
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.