Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
16,775
10,564,614,219
IssuesEvent
2019-10-05 03:35:03
Azure/azure-cli
https://api.github.com/repos/Azure/azure-cli
closed
webapp:ISO 8601 timestamp web app setting is being converted
App Services Web Apps
**Describe the bug** I am attempting to set an app setting to an ISO 8601 timestamp, and the value is being converted to a different format and timezone. Example of the command and setting: ```bash az webapp config appsettings set -n my-app \ -g my-app-rq \ --settings MAINTENANCE_END="2019-08-12T08:00:00.000-04:00" ``` Azure is saving this setting with the following value instead: `"08/12/2019 12:00:00"` When I set the same value in the Azure Portal, the value _does not_ get converted, which is the expected behavior. **To Reproduce** - Run the command above - Edit the value in the Azure Portal and observe it _not_ doing the same type of conversion **Expected behavior** - When I run `az webapp config appsettings set`, I expect the value (`"2019-08-12T08:00:00.000-04:00"`) to be set in Azure without being converted to a different format (`"08/12/2019 12:00:00"`). ## Environment Summary ``` Darwin-18.7.0-x86_64-i386-64bit Python 3.7.4 Shell: bash azure-cli 2.0.69 *
1.0
webapp:ISO 8601 timestamp web app setting is being converted - **Describe the bug** I am attempting to set an app setting to an ISO 8601 timestamp, and the value is being converted to a different format and timezone. Example of the command and setting: ```bash az webapp config appsettings set -n my-app \ -g my-app-rq \ --settings MAINTENANCE_END="2019-08-12T08:00:00.000-04:00" ``` Azure is saving this setting with the following value instead: `"08/12/2019 12:00:00"` When I set the same value in the Azure Portal, the value _does not_ get converted, which is the expected behavior. **To Reproduce** - Run the command above - Edit the value in the Azure Portal and observe it _not_ doing the same type of conversion **Expected behavior** - When I run `az webapp config appsettings set`, I expect the value (`"2019-08-12T08:00:00.000-04:00"`) to be set in Azure without being converted to a different format (`"08/12/2019 12:00:00"`). ## Environment Summary ``` Darwin-18.7.0-x86_64-i386-64bit Python 3.7.4 Shell: bash azure-cli 2.0.69 *
non_process
webapp iso timestamp web app setting is being converted describe the bug i am attempting to set an app setting to an iso timestamp and the value is being converted to a different format and timezone example of the command and setting bash az webapp config appsettings set n my app g my app rq settings maintenance end azure is saving this setting with the following value instead when i set the same value in the azure portal the value does not get converted which is the expected behavior to reproduce run the command above edit the value in the azure portal and observe it not doing the same type of conversion expected behavior when i run az webapp config appsettings set i expect the value to be set in azure without being converted to a different format environment summary darwin python shell bash azure cli
0
55,705
11,460,350,736
IssuesEvent
2020-02-07 09:34:02
jOOQ/jOOQ
https://api.github.com/repos/jOOQ/jOOQ
closed
Apply Configuration.onError flag also to SQL errors from AbstractDatabase
C: Code Generation E: All Editions P: Medium R: Fixed T: Enhancement
The current implementation of the code generator logs and skips all SQL errors produced by `jOOQ-meta`, when reverse engineering the meta data. The reason for this is that we occasionally have bugs in our own SQL queries, which should not necessarily fail the build because the bug might be cosmetic for most users. An example is when the query producing the `CHECK` constraints fails, most users will be able to continue working with that problem, as opposed to when the query producing the columns per table fails. However, some queries fail because of user errors, e.g. insufficient privileges to read system views and other objects. In those cases, it would sometimes be desireable to fail the build instead of skipping the functionality. We should have a flag that allows for specifying whether any SQL errors (or other exceptions) should fail the build. ---- See also: https://stackoverflow.com/q/60075399/521799
1.0
Apply Configuration.onError flag also to SQL errors from AbstractDatabase - The current implementation of the code generator logs and skips all SQL errors produced by `jOOQ-meta`, when reverse engineering the meta data. The reason for this is that we occasionally have bugs in our own SQL queries, which should not necessarily fail the build because the bug might be cosmetic for most users. An example is when the query producing the `CHECK` constraints fails, most users will be able to continue working with that problem, as opposed to when the query producing the columns per table fails. However, some queries fail because of user errors, e.g. insufficient privileges to read system views and other objects. In those cases, it would sometimes be desireable to fail the build instead of skipping the functionality. We should have a flag that allows for specifying whether any SQL errors (or other exceptions) should fail the build. ---- See also: https://stackoverflow.com/q/60075399/521799
non_process
apply configuration onerror flag also to sql errors from abstractdatabase the current implementation of the code generator logs and skips all sql errors produced by jooq meta when reverse engineering the meta data the reason for this is that we occasionally have bugs in our own sql queries which should not necessarily fail the build because the bug might be cosmetic for most users an example is when the query producing the check constraints fails most users will be able to continue working with that problem as opposed to when the query producing the columns per table fails however some queries fail because of user errors e g insufficient privileges to read system views and other objects in those cases it would sometimes be desireable to fail the build instead of skipping the functionality we should have a flag that allows for specifying whether any sql errors or other exceptions should fail the build see also
0
19,535
25,848,636,377
IssuesEvent
2022-12-13 08:49:33
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
`-std=c++14` in `tools/bazel.rc` ignored when building external repos
P3 type: support / not a bug (process) team-Rules-CPP
### Description of the bug: Recently gRPC bumped the version of C++ to 14 and its [tools/bazel.rc](https://github.com/grpc/grpc/blob/6fdbe6794695be853cdb87095a6b90868517f985/tools/bazel.rc#L5) got `--cxxopt='-std=c++14'` option to handle this. This worked and gRPC started having more C++14 code. But this began to have an issue when gRPC tried to upgrade protobuf to the version which started using C++14 code. ``` $ bazel build :grpc++ bazel-out/k8-opt-exec-2B5CBBC6/bin/external/com_google_protobuf/src/google/protobuf/io/_virtual_includes/printer/google/protobuf/io/printer.h:539:3: error: 'auto' return without trailing return type; deduced return types are a C++14 extension ``` But when I built this with `BAZEL_CXXOPTS="-std=c++14"`, it worked. ``` $ export BAZEL_CXXOPTS="-std=c++14" $ bazel build :grpc++ successfully built it. ``` If I'm doing wrong, what would be the right way to configure to use C++14 with bazel? ### What's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible. $ git clone https://github.com/veblush/grpc.git $ git checkout 10bbf8e50dcbed1401d4ea8ddf08eb2151cc3d5f $ bazel build :grpc++ ### Which operating system are you running Bazel on? Linux/Debian ### What is the output of `bazel info release`? release 5.3.0 ### If `bazel info release` returns `development version` or `(@non-git)`, tell us how you built Bazel. _No response_ ### What's the output of `git remote get-url origin; git rev-parse master; git rev-parse HEAD` ? _No response_ ### Have you found anything relevant by searching the web? _No response_ ### Any other information, logs, or outputs that you want to share? _No response_
1.0
`-std=c++14` in `tools/bazel.rc` ignored when building external repos - ### Description of the bug: Recently gRPC bumped the version of C++ to 14 and its [tools/bazel.rc](https://github.com/grpc/grpc/blob/6fdbe6794695be853cdb87095a6b90868517f985/tools/bazel.rc#L5) got `--cxxopt='-std=c++14'` option to handle this. This worked and gRPC started having more C++14 code. But this began to have an issue when gRPC tried to upgrade protobuf to the version which started using C++14 code. ``` $ bazel build :grpc++ bazel-out/k8-opt-exec-2B5CBBC6/bin/external/com_google_protobuf/src/google/protobuf/io/_virtual_includes/printer/google/protobuf/io/printer.h:539:3: error: 'auto' return without trailing return type; deduced return types are a C++14 extension ``` But when I built this with `BAZEL_CXXOPTS="-std=c++14"`, it worked. ``` $ export BAZEL_CXXOPTS="-std=c++14" $ bazel build :grpc++ successfully built it. ``` If I'm doing wrong, what would be the right way to configure to use C++14 with bazel? ### What's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible. $ git clone https://github.com/veblush/grpc.git $ git checkout 10bbf8e50dcbed1401d4ea8ddf08eb2151cc3d5f $ bazel build :grpc++ ### Which operating system are you running Bazel on? Linux/Debian ### What is the output of `bazel info release`? release 5.3.0 ### If `bazel info release` returns `development version` or `(@non-git)`, tell us how you built Bazel. _No response_ ### What's the output of `git remote get-url origin; git rev-parse master; git rev-parse HEAD` ? _No response_ ### Have you found anything relevant by searching the web? _No response_ ### Any other information, logs, or outputs that you want to share? _No response_
process
std c in tools bazel rc ignored when building external repos description of the bug recently grpc bumped the version of c to and its got cxxopt std c option to handle this this worked and grpc started having more c code but this began to have an issue when grpc tried to upgrade protobuf to the version which started using c code bazel build grpc bazel out opt exec bin external com google protobuf src google protobuf io virtual includes printer google protobuf io printer h error auto return without trailing return type deduced return types are a c extension but when i built this with bazel cxxopts std c it worked export bazel cxxopts std c bazel build grpc successfully built it if i m doing wrong what would be the right way to configure to use c with bazel what s the simplest easiest way to reproduce this bug please provide a minimal example if possible git clone git checkout bazel build grpc which operating system are you running bazel on linux debian what is the output of bazel info release release if bazel info release returns development version or non git tell us how you built bazel no response what s the output of git remote get url origin git rev parse master git rev parse head no response have you found anything relevant by searching the web no response any other information logs or outputs that you want to share no response
1
235,351
7,737,114,207
IssuesEvent
2018-05-28 06:57:58
kubernetes/test-infra
https://api.github.com/repos/kubernetes/test-infra
opened
Publishing bot cannot push
kind/bug priority/important-soon
Our publishing bot https://github.com/k8s-publishing-bot cannot push due to a `cla/linuxfoundation` check: ``` [28 May 18 05:54 UTC]: /publish_scripts/push.sh /etc/secret-volume/token master remote: error: GH006: Protected branch update failed for refs/heads/master. remote: error: Required status check "cla/linuxfoundation" is expected. To https://github.com/kubernetes/client-go ! [remote rejected] master -> master (protected branch hook declined) error: failed to push some refs to 'https://github.com/kubernetes/client-go' ```
1.0
Publishing bot cannot push - Our publishing bot https://github.com/k8s-publishing-bot cannot push due to a `cla/linuxfoundation` check: ``` [28 May 18 05:54 UTC]: /publish_scripts/push.sh /etc/secret-volume/token master remote: error: GH006: Protected branch update failed for refs/heads/master. remote: error: Required status check "cla/linuxfoundation" is expected. To https://github.com/kubernetes/client-go ! [remote rejected] master -> master (protected branch hook declined) error: failed to push some refs to 'https://github.com/kubernetes/client-go' ```
non_process
publishing bot cannot push our publishing bot cannot push due to a cla linuxfoundation check publish scripts push sh etc secret volume token master remote error protected branch update failed for refs heads master remote error required status check cla linuxfoundation is expected to master master protected branch hook declined error failed to push some refs to
0
6,302
9,308,335,573
IssuesEvent
2019-03-25 14:22:04
brandon1roadgears/Interpreter-of-programming-language-of-Turing-Machine
https://api.github.com/repos/brandon1roadgears/Interpreter-of-programming-language-of-Turing-Machine
opened
Реализация третьей части программы.
C++ Work in process
## В этой части необходимо написать фрагмент вывода изменённой в ходе работы интерпретатора строки (@goldmen4ik). *** ### Думаю, тут подробности не нужны. вывести обработанную строку хотя бы через cout. ### `cout << ishodnik;`.
1.0
Реализация третьей части программы. - ## В этой части необходимо написать фрагмент вывода изменённой в ходе работы интерпретатора строки (@goldmen4ik). *** ### Думаю, тут подробности не нужны. вывести обработанную строку хотя бы через cout. ### `cout << ishodnik;`.
process
реализация третьей части программы в этой части необходимо написать фрагмент вывода изменённой в ходе работы интерпретатора строки думаю тут подробности не нужны вывести обработанную строку хотя бы через cout cout ishodnik
1
185,168
21,785,095,786
IssuesEvent
2022-05-14 02:28:34
prashantgodhwani/phonefriend
https://api.github.com/repos/prashantgodhwani/phonefriend
closed
WS-2021-0079 (Medium) detected in laravel/framework-v5.5.40 - autoclosed
security vulnerability
## WS-2021-0079 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>laravel/framework-v5.5.40</b></p></summary> <p>The Laravel Framework.</p> <p>Library home page: <a href="https://api.github.com/repos/laravel/framework/zipball/d724ce0aa61bbd9adf658215eec484f5dd6711d6">https://api.github.com/repos/laravel/framework/zipball/d724ce0aa61bbd9adf658215eec484f5dd6711d6</a></p> <p> Dependency Hierarchy: - :x: **laravel/framework-v5.5.40** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Those using SQL Server with Laravel and allowing user input to be passed directly to the limit and offset functions are vulnerable to SQL injection. Other database drivers such as MySQL and Postgres are not affected by this vulnerability. This problem has been patched on Laravel versions 6.20.26 and 8.40.0. <p>Publish Date: 2021-04-30 <p>URL: <a href=https://github.com/advisories/GHSA-4mg9-vhxq-vm7j>WS-2021-0079</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-4mg9-vhxq-vm7j">https://github.com/advisories/GHSA-4mg9-vhxq-vm7j</a></p> <p>Release Date: 2021-04-30</p> <p>Fix Resolution: laravel/framework - v6.20.26,v8.40.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2021-0079 (Medium) detected in laravel/framework-v5.5.40 - autoclosed - ## WS-2021-0079 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>laravel/framework-v5.5.40</b></p></summary> <p>The Laravel Framework.</p> <p>Library home page: <a href="https://api.github.com/repos/laravel/framework/zipball/d724ce0aa61bbd9adf658215eec484f5dd6711d6">https://api.github.com/repos/laravel/framework/zipball/d724ce0aa61bbd9adf658215eec484f5dd6711d6</a></p> <p> Dependency Hierarchy: - :x: **laravel/framework-v5.5.40** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Those using SQL Server with Laravel and allowing user input to be passed directly to the limit and offset functions are vulnerable to SQL injection. Other database drivers such as MySQL and Postgres are not affected by this vulnerability. This problem has been patched on Laravel versions 6.20.26 and 8.40.0. <p>Publish Date: 2021-04-30 <p>URL: <a href=https://github.com/advisories/GHSA-4mg9-vhxq-vm7j>WS-2021-0079</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-4mg9-vhxq-vm7j">https://github.com/advisories/GHSA-4mg9-vhxq-vm7j</a></p> <p>Release Date: 2021-04-30</p> <p>Fix Resolution: laravel/framework - v6.20.26,v8.40.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
ws medium detected in laravel framework autoclosed ws medium severity vulnerability vulnerable library laravel framework the laravel framework library home page a href dependency hierarchy x laravel framework vulnerable library vulnerability details those using sql server with laravel and allowing user input to be passed directly to the limit and offset functions are vulnerable to sql injection other database drivers such as mysql and postgres are not affected by this vulnerability this problem has been patched on laravel versions and publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution laravel framework step up your open source security game with whitesource
0
321,868
9,810,144,078
IssuesEvent
2019-06-12 19:45:11
HabitRPG/habitica-android
https://api.github.com/repos/HabitRPG/habitica-android
closed
Party: Switching tabs quickly causes blank members screen and crashes app
Help wanted Priority: minor Type: Bug Type: Crash
To replicate: go to a party and quickly tap between tabs. It seems to happen especially fast switching between chat and members screens. I submitted crash feedback when i did it once.
1.0
Party: Switching tabs quickly causes blank members screen and crashes app - To replicate: go to a party and quickly tap between tabs. It seems to happen especially fast switching between chat and members screens. I submitted crash feedback when i did it once.
non_process
party switching tabs quickly causes blank members screen and crashes app to replicate go to a party and quickly tap between tabs it seems to happen especially fast switching between chat and members screens i submitted crash feedback when i did it once
0
211,906
16,468,637,140
IssuesEvent
2021-05-23 00:13:14
pystatgen/sgkit
https://api.github.com/repos/pystatgen/sgkit
opened
Remove mention of CuPy in docs
documentation
We have two mentions of CuPy in our docs but to the best of my knowledge we don't use CuPy yet. Should we remove those mentions until we make use of CuPy?
1.0
Remove mention of CuPy in docs - We have two mentions of CuPy in our docs but to the best of my knowledge we don't use CuPy yet. Should we remove those mentions until we make use of CuPy?
non_process
remove mention of cupy in docs we have two mentions of cupy in our docs but to the best of my knowledge we don t use cupy yet should we remove those mentions until we make use of cupy
0
16,572
21,580,092,080
IssuesEvent
2022-05-02 17:44:47
microsoft/vscode
https://api.github.com/repos/microsoft/vscode
closed
Terminal activate conda base environment just after activating the right environment
bug confirmation-pending terminal-process
<!-- ⚠️⚠️ Do Not Delete This! bug_report_template ⚠️⚠️ --> <!-- Please read our Rules of Conduct: https://opensource.microsoft.com/codeofconduct/ --> <!-- 🕮 Read our guide about submitting issues: https://github.com/microsoft/vscode/wiki/Submitting-Bugs-and-Suggestions --> <!-- 🔎 Search existing issues to avoid creating duplicates. --> <!-- 🧪 Test using the latest Insiders build to see if your issue has already been fixed: https://code.visualstudio.com/insiders/ --> <!-- 💡 Instead of creating your report here, use 'Report Issue' from the 'Help' menu in VS Code to pre-fill useful information. --> <!-- 🔧 Launch with `code --disable-extensions` to check. --> Does this issue occur when all extensions are disabled?: Yes/No <!-- 🪓 If you answered No above, use 'Help: Start Extension Bisect' from Command Palette to try to identify the cause. --> <!-- 📣 Issues caused by an extension need to be reported directly to the extension publisher. The 'Help > Report Issue' <!-- **NOTE**: Everything below is auto-generated; no editing required. --> # Diagnostic data - Python version (& distribution if applicable, e.g. Anaconda): 3.9.9 - Type of virtual environment used (e.g. conda, venv, virtualenv, etc.): Conda - Value of the `python.languageServer` setting: Default <details> <summary>User Settings</summary> <p> ``` defaultLS: {"defaultLSType":"Pylance"} downloadLanguageServer: true envFile: "<placeholder>" venvPath: "<placeholder>" venvFolders: "<placeholder>" condaPath: "<placeholder>" pipenvPath: "<placeholder>" poetryPath: "<placeholder>" languageServer: "Pylance" linting • enabled: true • cwd: "<placeholder>" • Flake8Args: "<placeholder>" • flake8Enabled: false • flake8Path: "<placeholder>" • lintOnSave: true • banditArgs: "<placeholder>" • banditEnabled: false • banditPath: "<placeholder>" • mypyArgs: "<placeholder>" • mypyEnabled: false • mypyPath: "<placeholder>" • pycodestyleArgs: "<placeholder>" • pycodestyleEnabled: false • pycodestylePath: "<placeholder>" • prospectorArgs: "<placeholder>" • prospectorEnabled: false • prospectorPath: "<placeholder>" • pydocstyleArgs: "<placeholder>" • pydocstyleEnabled: false • pydocstylePath: "<placeholder>" • pylamaArgs: "<placeholder>" • pylamaEnabled: false • pylamaPath: "<placeholder>" • pylintArgs: "<placeholder>" • pylintPath: "<placeholder>" sortImports • args: "<placeholder>" • path: "<placeholder>" formatting • autopep8Args: "<placeholder>" • autopep8Path: "<placeholder>" • provider: "black" • blackArgs: "<placeholder>" • blackPath: "<placeholder>" • yapfArgs: "<placeholder>" • yapfPath: "<placeholder>" testing • cwd: "<placeholder>" • debugPort: 3000 • nosetestArgs: "<placeholder>" • nosetestsEnabled: undefined • nosetestPath: "<placeholder>" • promptToConfigure: true • pytestArgs: "<placeholder>" • pytestEnabled: false • pytestPath: "<placeholder>" • unittestArgs: "<placeholder>" • unittestEnabled: false • autoTestDiscoverOnSaveEnabled: true terminal • activateEnvironment: true • executeInFileDir: "<placeholder>" • launchArgs: "<placeholder>" experiments • enabled: true • optInto: [] • optOutFrom: [] tensorBoard • logDirectory: "<placeholder>" ``` </p> </details> Extension version: 2022.4.0 VS Code version: Code 1.66.0 (Universal) (e18005f0f1b33c29e81d732535d8c0e47cafb0b5, 2022-03-30T05:50:04.419Z) OS version: Darwin arm64 21.4.0 Restricted Mode: No <details> <summary>System Info</summary> |Item|Value| |---|---| |CPUs|Apple M1 Pro (8 x 24)| |GPU Status|2d_canvas: enabled<br>canvas_oop_rasterization: disabled_off<br>direct_rendering_display_compositor: disabled_off_ok<br>gpu_compositing: enabled<br>metal: disabled_off<br>multiple_raster_threads: enabled_on<br>oop_rasterization: enabled<br>opengl: enabled_on<br>rasterization: enabled<br>raw_draw: disabled_off_ok<br>skia_renderer: enabled_on<br>video_decode: enabled<br>video_encode: enabled<br>webgl: enabled<br>webgl2: enabled| |Load (avg)|5, 5, 5| |Memory (System)|16.00GB (0.08GB free)| |Process Argv|--crash-reporter-id a129d0b2-ecd6-4f0a-8ff3-d67e5b6507fa| |Screen Reader|yes| |VM|0%| </details><details> <summary>A/B Experiments</summary> ``` vsliv368:30146709 vsreu685:30147344 python383cf:30185419 vspor879:30202332 vspor708:30202333 vspor363:30204092 pythonvspyl392:30443607 pythontb:30283811 pythonvspyt551cf:30345471 pythonptprofiler:30281270 vshan820:30294714 vstes263:30335439 vscoreces:30445986 pythondataviewer:30285071 vscod805cf:30301675 pythonvspyt200:30340761 binariesv615:30325510 bridge0708:30335490 bridge0723:30353136 vsaa593:30376534 testflight:30433670 vsc1dst:30438360 pythonvs932:30410667 wslgetstarted:30449410 pythonvsnew555:30457759 vscscmwlcmt:30465135 cppdebug:30466689 pynewfile477:30463512 ``` </details> <!-- generated by issue reporter --> # Behaviour ## Expected vs. Actual Hello, When I select my conda env for the workspace, I expect that each terminal session activate this environment. In the past it worked well. But since a few days I have a strange behaviour : the terminal execute the command `conda activate myenv` then just after `source /Users/luis/miniforge3/bin/activate` reactivating base env. (see in screenshoot) ## Steps to reproduce: ![Enregistrement de l’écran 2022-04-06 à 12 03 22(1)](https://user-images.githubusercontent.com/30115537/161980990-9f437397-0e65-4023-a666-0fb3217c20ec.gif)
1.0
Terminal activate conda base environment just after activating the right environment - <!-- ⚠️⚠️ Do Not Delete This! bug_report_template ⚠️⚠️ --> <!-- Please read our Rules of Conduct: https://opensource.microsoft.com/codeofconduct/ --> <!-- 🕮 Read our guide about submitting issues: https://github.com/microsoft/vscode/wiki/Submitting-Bugs-and-Suggestions --> <!-- 🔎 Search existing issues to avoid creating duplicates. --> <!-- 🧪 Test using the latest Insiders build to see if your issue has already been fixed: https://code.visualstudio.com/insiders/ --> <!-- 💡 Instead of creating your report here, use 'Report Issue' from the 'Help' menu in VS Code to pre-fill useful information. --> <!-- 🔧 Launch with `code --disable-extensions` to check. --> Does this issue occur when all extensions are disabled?: Yes/No <!-- 🪓 If you answered No above, use 'Help: Start Extension Bisect' from Command Palette to try to identify the cause. --> <!-- 📣 Issues caused by an extension need to be reported directly to the extension publisher. The 'Help > Report Issue' <!-- **NOTE**: Everything below is auto-generated; no editing required. --> # Diagnostic data - Python version (& distribution if applicable, e.g. Anaconda): 3.9.9 - Type of virtual environment used (e.g. conda, venv, virtualenv, etc.): Conda - Value of the `python.languageServer` setting: Default <details> <summary>User Settings</summary> <p> ``` defaultLS: {"defaultLSType":"Pylance"} downloadLanguageServer: true envFile: "<placeholder>" venvPath: "<placeholder>" venvFolders: "<placeholder>" condaPath: "<placeholder>" pipenvPath: "<placeholder>" poetryPath: "<placeholder>" languageServer: "Pylance" linting • enabled: true • cwd: "<placeholder>" • Flake8Args: "<placeholder>" • flake8Enabled: false • flake8Path: "<placeholder>" • lintOnSave: true • banditArgs: "<placeholder>" • banditEnabled: false • banditPath: "<placeholder>" • mypyArgs: "<placeholder>" • mypyEnabled: false • mypyPath: "<placeholder>" • pycodestyleArgs: "<placeholder>" • pycodestyleEnabled: false • pycodestylePath: "<placeholder>" • prospectorArgs: "<placeholder>" • prospectorEnabled: false • prospectorPath: "<placeholder>" • pydocstyleArgs: "<placeholder>" • pydocstyleEnabled: false • pydocstylePath: "<placeholder>" • pylamaArgs: "<placeholder>" • pylamaEnabled: false • pylamaPath: "<placeholder>" • pylintArgs: "<placeholder>" • pylintPath: "<placeholder>" sortImports • args: "<placeholder>" • path: "<placeholder>" formatting • autopep8Args: "<placeholder>" • autopep8Path: "<placeholder>" • provider: "black" • blackArgs: "<placeholder>" • blackPath: "<placeholder>" • yapfArgs: "<placeholder>" • yapfPath: "<placeholder>" testing • cwd: "<placeholder>" • debugPort: 3000 • nosetestArgs: "<placeholder>" • nosetestsEnabled: undefined • nosetestPath: "<placeholder>" • promptToConfigure: true • pytestArgs: "<placeholder>" • pytestEnabled: false • pytestPath: "<placeholder>" • unittestArgs: "<placeholder>" • unittestEnabled: false • autoTestDiscoverOnSaveEnabled: true terminal • activateEnvironment: true • executeInFileDir: "<placeholder>" • launchArgs: "<placeholder>" experiments • enabled: true • optInto: [] • optOutFrom: [] tensorBoard • logDirectory: "<placeholder>" ``` </p> </details> Extension version: 2022.4.0 VS Code version: Code 1.66.0 (Universal) (e18005f0f1b33c29e81d732535d8c0e47cafb0b5, 2022-03-30T05:50:04.419Z) OS version: Darwin arm64 21.4.0 Restricted Mode: No <details> <summary>System Info</summary> |Item|Value| |---|---| |CPUs|Apple M1 Pro (8 x 24)| |GPU Status|2d_canvas: enabled<br>canvas_oop_rasterization: disabled_off<br>direct_rendering_display_compositor: disabled_off_ok<br>gpu_compositing: enabled<br>metal: disabled_off<br>multiple_raster_threads: enabled_on<br>oop_rasterization: enabled<br>opengl: enabled_on<br>rasterization: enabled<br>raw_draw: disabled_off_ok<br>skia_renderer: enabled_on<br>video_decode: enabled<br>video_encode: enabled<br>webgl: enabled<br>webgl2: enabled| |Load (avg)|5, 5, 5| |Memory (System)|16.00GB (0.08GB free)| |Process Argv|--crash-reporter-id a129d0b2-ecd6-4f0a-8ff3-d67e5b6507fa| |Screen Reader|yes| |VM|0%| </details><details> <summary>A/B Experiments</summary> ``` vsliv368:30146709 vsreu685:30147344 python383cf:30185419 vspor879:30202332 vspor708:30202333 vspor363:30204092 pythonvspyl392:30443607 pythontb:30283811 pythonvspyt551cf:30345471 pythonptprofiler:30281270 vshan820:30294714 vstes263:30335439 vscoreces:30445986 pythondataviewer:30285071 vscod805cf:30301675 pythonvspyt200:30340761 binariesv615:30325510 bridge0708:30335490 bridge0723:30353136 vsaa593:30376534 testflight:30433670 vsc1dst:30438360 pythonvs932:30410667 wslgetstarted:30449410 pythonvsnew555:30457759 vscscmwlcmt:30465135 cppdebug:30466689 pynewfile477:30463512 ``` </details> <!-- generated by issue reporter --> # Behaviour ## Expected vs. Actual Hello, When I select my conda env for the workspace, I expect that each terminal session activate this environment. In the past it worked well. But since a few days I have a strange behaviour : the terminal execute the command `conda activate myenv` then just after `source /Users/luis/miniforge3/bin/activate` reactivating base env. (see in screenshoot) ## Steps to reproduce: ![Enregistrement de l’écran 2022-04-06 à 12 03 22(1)](https://user-images.githubusercontent.com/30115537/161980990-9f437397-0e65-4023-a666-0fb3217c20ec.gif)
process
terminal activate conda base environment just after activating the right environment does this issue occur when all extensions are disabled yes no report issue diagnostic data python version distribution if applicable e g anaconda type of virtual environment used e g conda venv virtualenv etc conda value of the python languageserver setting default user settings defaultls defaultlstype pylance downloadlanguageserver true envfile venvpath venvfolders condapath pipenvpath poetrypath languageserver pylance linting • enabled true • cwd • • false • • lintonsave true • banditargs • banditenabled false • banditpath • mypyargs • mypyenabled false • mypypath • pycodestyleargs • pycodestyleenabled false • pycodestylepath • prospectorargs • prospectorenabled false • prospectorpath • pydocstyleargs • pydocstyleenabled false • pydocstylepath • pylamaargs • pylamaenabled false • pylamapath • pylintargs • pylintpath sortimports • args • path formatting • • • provider black • blackargs • blackpath • yapfargs • yapfpath testing • cwd • debugport • nosetestargs • nosetestsenabled undefined • nosetestpath • prompttoconfigure true • pytestargs • pytestenabled false • pytestpath • unittestargs • unittestenabled false • autotestdiscoveronsaveenabled true terminal • activateenvironment true • executeinfiledir • launchargs experiments • enabled true • optinto • optoutfrom tensorboard • logdirectory extension version vs code version code universal os version darwin restricted mode no system info item value cpus apple pro x gpu status canvas enabled canvas oop rasterization disabled off direct rendering display compositor disabled off ok gpu compositing enabled metal disabled off multiple raster threads enabled on oop rasterization enabled opengl enabled on rasterization enabled raw draw disabled off ok skia renderer enabled on video decode enabled video encode enabled webgl enabled enabled load avg memory system free process argv crash reporter id screen reader yes vm a b experiments pythontb pythonptprofiler vscoreces pythondataviewer testflight wslgetstarted vscscmwlcmt cppdebug behaviour expected vs actual hello when i select my conda env for the workspace i expect that each terminal session activate this environment in the past it worked well but since a few days i have a strange behaviour the terminal execute the command conda activate myenv then just after source users luis bin activate reactivating base env see in screenshoot steps to reproduce
1
11,714
14,546,541,572
IssuesEvent
2020-12-15 21:24:17
MicrosoftDocs/azure-devops-docs
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
closed
No Support for Generic Environment Resources
Pri2 devops-cicd-process/tech devops/prod product-feedback
I see that Azure DevOps yaml supports Kubernetes and Virtual Machines for deployment tracability in the form of "resources". Is there a reason there doesnt just exists a "generic" resource that is created if the resource is referenced through yml but doesnt exists yet? The same as environments are managed. That way jobs and commits will be recorded by resource based on the name of the resources. Example: ` environment: environment_name.resource_name ` And the environment **environment_name** is created and the "generic" resource is created **resource_name** --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 77d95db6-9983-7346-d0eb-4b7443e4e252 * Version Independent ID: 0a22cccc-318d-592f-d1ab-09ec01d88087 * Content: [Environment - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/environments?view=azure-devops) * Content Source: [docs/pipelines/process/environments.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/environments.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
1.0
No Support for Generic Environment Resources - I see that Azure DevOps yaml supports Kubernetes and Virtual Machines for deployment tracability in the form of "resources". Is there a reason there doesnt just exists a "generic" resource that is created if the resource is referenced through yml but doesnt exists yet? The same as environments are managed. That way jobs and commits will be recorded by resource based on the name of the resources. Example: ` environment: environment_name.resource_name ` And the environment **environment_name** is created and the "generic" resource is created **resource_name** --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 77d95db6-9983-7346-d0eb-4b7443e4e252 * Version Independent ID: 0a22cccc-318d-592f-d1ab-09ec01d88087 * Content: [Environment - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/environments?view=azure-devops) * Content Source: [docs/pipelines/process/environments.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/environments.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
process
no support for generic environment resources i see that azure devops yaml supports kubernetes and virtual machines for deployment tracability in the form of resources is there a reason there doesnt just exists a generic resource that is created if the resource is referenced through yml but doesnt exists yet the same as environments are managed that way jobs and commits will be recorded by resource based on the name of the resources example environment environment name resource name and the environment environment name is created and the generic resource is created resource name document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
1
4,508
7,353,507,171
IssuesEvent
2018-03-09 01:05:20
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
closed
ServiceProcess test failures, including PauseAndContinue
area-System.ServiceProcess disabled-test os-windows tenet-reliability test bug test-run-core
This has failed the last three builds (2/11 - 2/12) - [MC test results link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180212.05/workItem/System.ServiceProcess.ServiceController.Tests/analysis/xunit/System.ServiceProcess.Tests.ServiceControllerTests~2FPauseAndContinue) The test `System.ServiceProcess.Tests.ServiceControllerTests/PauseAndContinue` has failed. System.ComponentModel.Win32Exception : Could not find service Stack Trace: at System.ServiceProcess.Tests.TestServiceInstaller.DeleteService() in E:\A\_work\1424\s\corefx\src\System.ServiceProcess.ServiceController\tests\System.ServiceProcess.ServiceController.TestService\TestServiceInstaller.cs:line 186 at System.ServiceProcess.Tests.TestServiceInstaller.RemoveService() in E:\A\_work\1424\s\corefx\src\System.ServiceProcess.ServiceController\tests\System.ServiceProcess.ServiceController.TestService\TestServiceInstaller.cs:line 133 at System.ServiceProcess.Tests.TestServiceProvider.DeleteTestServices() in E:\A\_work\1424\s\corefx\src\System.ServiceProcess.ServiceController\tests\TestServiceProvider.cs:line 139 at System.ServiceProcess.Tests.TestServiceProvider.DeleteTestServices() in E:\A\_work\1424\s\corefx\src\System.ServiceProcess.ServiceController\tests\TestServiceProvider.cs:line 149 at System.ServiceProcess.Tests.ServiceControllerTests.Dispose() in E:\A\_work\1424\s\corefx\src\System.ServiceProcess.ServiceController\tests\ServiceControllerTests.cs:line 184 at ReflectionAbstractionExtensions.DisposeTestClass(ITest test, Object testClass, IMessageBus messageBus, ExecutionTimer timer, CancellationTokenSource cancellationTokenSource) in C:\BuildAgent\work\cb37e9acf085d108\src\xunit.execution\Extensions\ReflectionAbstractionExtensions.cs:line 76 Build : Master - 20180212.05 (Core Tests) Failing configurations: - Windows.10.Amd64-x64 - Release
1.0
ServiceProcess test failures, including PauseAndContinue - This has failed the last three builds (2/11 - 2/12) - [MC test results link](https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180212.05/workItem/System.ServiceProcess.ServiceController.Tests/analysis/xunit/System.ServiceProcess.Tests.ServiceControllerTests~2FPauseAndContinue) The test `System.ServiceProcess.Tests.ServiceControllerTests/PauseAndContinue` has failed. System.ComponentModel.Win32Exception : Could not find service Stack Trace: at System.ServiceProcess.Tests.TestServiceInstaller.DeleteService() in E:\A\_work\1424\s\corefx\src\System.ServiceProcess.ServiceController\tests\System.ServiceProcess.ServiceController.TestService\TestServiceInstaller.cs:line 186 at System.ServiceProcess.Tests.TestServiceInstaller.RemoveService() in E:\A\_work\1424\s\corefx\src\System.ServiceProcess.ServiceController\tests\System.ServiceProcess.ServiceController.TestService\TestServiceInstaller.cs:line 133 at System.ServiceProcess.Tests.TestServiceProvider.DeleteTestServices() in E:\A\_work\1424\s\corefx\src\System.ServiceProcess.ServiceController\tests\TestServiceProvider.cs:line 139 at System.ServiceProcess.Tests.TestServiceProvider.DeleteTestServices() in E:\A\_work\1424\s\corefx\src\System.ServiceProcess.ServiceController\tests\TestServiceProvider.cs:line 149 at System.ServiceProcess.Tests.ServiceControllerTests.Dispose() in E:\A\_work\1424\s\corefx\src\System.ServiceProcess.ServiceController\tests\ServiceControllerTests.cs:line 184 at ReflectionAbstractionExtensions.DisposeTestClass(ITest test, Object testClass, IMessageBus messageBus, ExecutionTimer timer, CancellationTokenSource cancellationTokenSource) in C:\BuildAgent\work\cb37e9acf085d108\src\xunit.execution\Extensions\ReflectionAbstractionExtensions.cs:line 76 Build : Master - 20180212.05 (Core Tests) Failing configurations: - Windows.10.Amd64-x64 - Release
process
serviceprocess test failures including pauseandcontinue this has failed the last three builds the test system serviceprocess tests servicecontrollertests pauseandcontinue has failed system componentmodel could not find service stack trace at system serviceprocess tests testserviceinstaller deleteservice in e a work s corefx src system serviceprocess servicecontroller tests system serviceprocess servicecontroller testservice testserviceinstaller cs line at system serviceprocess tests testserviceinstaller removeservice in e a work s corefx src system serviceprocess servicecontroller tests system serviceprocess servicecontroller testservice testserviceinstaller cs line at system serviceprocess tests testserviceprovider deletetestservices in e a work s corefx src system serviceprocess servicecontroller tests testserviceprovider cs line at system serviceprocess tests testserviceprovider deletetestservices in e a work s corefx src system serviceprocess servicecontroller tests testserviceprovider cs line at system serviceprocess tests servicecontrollertests dispose in e a work s corefx src system serviceprocess servicecontroller tests servicecontrollertests cs line at reflectionabstractionextensions disposetestclass itest test object testclass imessagebus messagebus executiontimer timer cancellationtokensource cancellationtokensource in c buildagent work src xunit execution extensions reflectionabstractionextensions cs line build master core tests failing configurations windows release
1
161,797
20,155,364,456
IssuesEvent
2022-02-09 16:00:01
jgeraigery/experian-java
https://api.github.com/repos/jgeraigery/experian-java
closed
CVE-2019-14892 (High) detected in jackson-databind-2.9.2.jar - autoclosed
security vulnerability
## CVE-2019-14892 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.2.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: experian-java/MavenWorkspace/bis-services-lib/bis-services-base/pom.xml</p> <p>Path to vulnerable library: ository/com/fasterxml/jackson/core/jackson-databind/2.9.2/jackson-databind-2.9.2.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.9.2.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jgeraigery/experian-java/commit/9c836fb688a6591b590fe6154680b73ceed58629">9c836fb688a6591b590fe6154680b73ceed58629</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A flaw was discovered in jackson-databind in versions before 2.9.10, 2.8.11.5 and 2.6.7.3, where it would permit polymorphic deserialization of a malicious object using commons-configuration 1 and 2 JNDI classes. An attacker could use this flaw to execute arbitrary code. <p>Publish Date: 2020-03-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14892>CVE-2019-14892</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2462">https://github.com/FasterXML/jackson-databind/issues/2462</a></p> <p>Release Date: 2020-09-04</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.6.7.3,2.7.9.7,2.8.11.5,2.9.10</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.2","packageFilePaths":["/MavenWorkspace/bis-services-lib/bis-services-base/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.6.7.3,2.7.9.7,2.8.11.5,2.9.10","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-14892","vulnerabilityDetails":"A flaw was discovered in jackson-databind in versions before 2.9.10, 2.8.11.5 and 2.6.7.3, where it would permit polymorphic deserialization of a malicious object using commons-configuration 1 and 2 JNDI classes. An attacker could use this flaw to execute arbitrary code.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14892","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2019-14892 (High) detected in jackson-databind-2.9.2.jar - autoclosed - ## CVE-2019-14892 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.2.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: experian-java/MavenWorkspace/bis-services-lib/bis-services-base/pom.xml</p> <p>Path to vulnerable library: ository/com/fasterxml/jackson/core/jackson-databind/2.9.2/jackson-databind-2.9.2.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.9.2.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jgeraigery/experian-java/commit/9c836fb688a6591b590fe6154680b73ceed58629">9c836fb688a6591b590fe6154680b73ceed58629</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A flaw was discovered in jackson-databind in versions before 2.9.10, 2.8.11.5 and 2.6.7.3, where it would permit polymorphic deserialization of a malicious object using commons-configuration 1 and 2 JNDI classes. An attacker could use this flaw to execute arbitrary code. <p>Publish Date: 2020-03-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14892>CVE-2019-14892</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2462">https://github.com/FasterXML/jackson-databind/issues/2462</a></p> <p>Release Date: 2020-09-04</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.6.7.3,2.7.9.7,2.8.11.5,2.9.10</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.2","packageFilePaths":["/MavenWorkspace/bis-services-lib/bis-services-base/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.6.7.3,2.7.9.7,2.8.11.5,2.9.10","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-14892","vulnerabilityDetails":"A flaw was discovered in jackson-databind in versions before 2.9.10, 2.8.11.5 and 2.6.7.3, where it would permit polymorphic deserialization of a malicious object using commons-configuration 1 and 2 JNDI classes. An attacker could use this flaw to execute arbitrary code.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14892","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_process
cve high detected in jackson databind jar autoclosed cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file experian java mavenworkspace bis services lib bis services base pom xml path to vulnerable library ository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details a flaw was discovered in jackson databind in versions before and where it would permit polymorphic deserialization of a malicious object using commons configuration and jndi classes an attacker could use this flaw to execute arbitrary code publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails a flaw was discovered in jackson databind in versions before and where it would permit polymorphic deserialization of a malicious object using commons configuration and jndi classes an attacker could use this flaw to execute arbitrary code vulnerabilityurl
0
2,465
5,242,967,585
IssuesEvent
2017-01-31 19:26:48
opentrials/opentrials
https://api.github.com/repos/opentrials/opentrials
closed
Add records.source_url_type like field
API Explorer Processors refactoring
Our `source_url` could be human-readable url or API endpoint or link to secure space. So we need some instruction to frontend added on processors level how to show it to the user.
1.0
Add records.source_url_type like field - Our `source_url` could be human-readable url or API endpoint or link to secure space. So we need some instruction to frontend added on processors level how to show it to the user.
process
add records source url type like field our source url could be human readable url or api endpoint or link to secure space so we need some instruction to frontend added on processors level how to show it to the user
1
832,178
32,075,345,508
IssuesEvent
2023-09-25 10:37:08
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.nytimes.com - see bug description
priority-critical browser-fenix engine-gecko
<!-- @browser: Firefox Mobile 118.0 --> <!-- @ua_header: Mozilla/5.0 (Android 13; Mobile; rv:109.0) Gecko/118.0 Firefox/118.0 --> <!-- @reported_with: android-components-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/127499 --> <!-- @extra_labels: browser-fenix --> **URL**: https://www.nytimes.com/2023/09/23/world/asia/china-sea-philippines-us.html **Browser / Version**: Firefox Mobile 118.0 **Operating System**: Android 13 **Tested Another Browser**: Yes Chrome **Problem type**: Something else **Description**: page doesn't render completely. scrolling down leads to complete whitespaces and it takes multiple many as l refreshes to view **Steps to Reproduce**: Didn't do anything unusual before I visited the page <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>buildID: 20230914180032</li><li>channel: beta</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2023/9/dc01d5be-7683-4308-92a4-87141e63bfe2) _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.nytimes.com - see bug description - <!-- @browser: Firefox Mobile 118.0 --> <!-- @ua_header: Mozilla/5.0 (Android 13; Mobile; rv:109.0) Gecko/118.0 Firefox/118.0 --> <!-- @reported_with: android-components-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/127499 --> <!-- @extra_labels: browser-fenix --> **URL**: https://www.nytimes.com/2023/09/23/world/asia/china-sea-philippines-us.html **Browser / Version**: Firefox Mobile 118.0 **Operating System**: Android 13 **Tested Another Browser**: Yes Chrome **Problem type**: Something else **Description**: page doesn't render completely. scrolling down leads to complete whitespaces and it takes multiple many as l refreshes to view **Steps to Reproduce**: Didn't do anything unusual before I visited the page <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>buildID: 20230914180032</li><li>channel: beta</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2023/9/dc01d5be-7683-4308-92a4-87141e63bfe2) _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
see bug description url browser version firefox mobile operating system android tested another browser yes chrome problem type something else description page doesn t render completely scrolling down leads to complete whitespaces and it takes multiple many as l refreshes to view steps to reproduce didn t do anything unusual before i visited the page browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false buildid channel beta hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
0
11,919
14,702,541,081
IssuesEvent
2021-01-04 13:45:55
modi-w/AutoVersionsDB
https://api.github.com/repos/modi-w/AutoVersionsDB
opened
Add a numeric identification for the filenames to allow ordering
area-Core process-ready-for-implementation type-enhancement
**The Problem** The repeatable scripts files list and the DevDummyData scripts files are sorted by alphanumeric order, we don't have any chronological or logical order for those files. This cause the following problems: 1. When the user adds a new script file, the file is not always added to the end of the list, so sometimes it is not so clear where is the new file in the list. 2. Sometimes, it is important the order of those files execution, we cant do that right now unless the user gave appropriate names to those script files. **Solution** Add a numeric identification for the filenames to allow ordering. **Action Items:** **Updates**
1.0
Add a numeric identification for the filenames to allow ordering - **The Problem** The repeatable scripts files list and the DevDummyData scripts files are sorted by alphanumeric order, we don't have any chronological or logical order for those files. This cause the following problems: 1. When the user adds a new script file, the file is not always added to the end of the list, so sometimes it is not so clear where is the new file in the list. 2. Sometimes, it is important the order of those files execution, we cant do that right now unless the user gave appropriate names to those script files. **Solution** Add a numeric identification for the filenames to allow ordering. **Action Items:** **Updates**
process
add a numeric identification for the filenames to allow ordering the problem the repeatable scripts files list and the devdummydata scripts files are sorted by alphanumeric order we don t have any chronological or logical order for those files this cause the following problems when the user adds a new script file the file is not always added to the end of the list so sometimes it is not so clear where is the new file in the list sometimes it is important the order of those files execution we cant do that right now unless the user gave appropriate names to those script files solution add a numeric identification for the filenames to allow ordering action items updates
1
7,658
10,743,040,593
IssuesEvent
2019-10-30 00:35:23
HouraiTeahouse/FantasyCrescendo
https://api.github.com/repos/HouraiTeahouse/FantasyCrescendo
closed
Migrate submodule dependencies into Unity Package Manager format to be fetched remotely
Category:Game Engine Priority:0 Severity:0 Status:Assigned Type:Experimental Type:Process
The new Unity Package Manager should be able to support downloading remote 3rd party packages sometime in the near future (2018/2019), we should migrate the dependencies we currently have held as git submodules into their own repos to be fetched by the package manager. This should offer an easier method of dependency management and lower the barrier for new contributors to join the project.
1.0
Migrate submodule dependencies into Unity Package Manager format to be fetched remotely - The new Unity Package Manager should be able to support downloading remote 3rd party packages sometime in the near future (2018/2019), we should migrate the dependencies we currently have held as git submodules into their own repos to be fetched by the package manager. This should offer an easier method of dependency management and lower the barrier for new contributors to join the project.
process
migrate submodule dependencies into unity package manager format to be fetched remotely the new unity package manager should be able to support downloading remote party packages sometime in the near future we should migrate the dependencies we currently have held as git submodules into their own repos to be fetched by the package manager this should offer an easier method of dependency management and lower the barrier for new contributors to join the project
1
10,108
13,044,162,160
IssuesEvent
2020-07-29 03:47:30
tikv/tikv
https://api.github.com/repos/tikv/tikv
closed
UCP: Migrate scalar function `AddDatetimeAndDuration` from TiDB
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
## Description Port the scalar function `AddDatetimeAndDuration` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @sticnarf ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
2.0
UCP: Migrate scalar function `AddDatetimeAndDuration` from TiDB - ## Description Port the scalar function `AddDatetimeAndDuration` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @sticnarf ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
process
ucp migrate scalar function adddatetimeandduration from tidb description port the scalar function adddatetimeandduration from tidb to coprocessor score mentor s sticnarf recommended skills rust programming learning materials already implemented expressions ported from tidb
1
81,533
23,487,602,541
IssuesEvent
2022-08-17 15:35:47
microsoft/PowerToys
https://api.github.com/repos/microsoft/PowerToys
closed
ARM64 verification script for CI
Help Wanted Status-In progress Area-Build Area-ARM64
### Description of the new feature / enhancement Let's have a CI verification script for ARM64. Thoughts CSProj and SLN. Props files maybe? ### Scenario when this would be used? ARM support being sure we are consistant ### Supporting information _No response_
1.0
ARM64 verification script for CI - ### Description of the new feature / enhancement Let's have a CI verification script for ARM64. Thoughts CSProj and SLN. Props files maybe? ### Scenario when this would be used? ARM support being sure we are consistant ### Supporting information _No response_
non_process
verification script for ci description of the new feature enhancement let s have a ci verification script for thoughts csproj and sln props files maybe scenario when this would be used arm support being sure we are consistant supporting information no response
0
18,702
24,596,205,283
IssuesEvent
2022-10-14 08:33:16
prisma/prisma
https://api.github.com/repos/prisma/prisma
opened
Error: [libs/dml/src/datamodel.rs:178:14] Every RelationInfo should have a complementary RelationInfo on the opposite relation field.
bug/1-unconfirmed kind/bug process/candidate topic: error reporting team/schema
<!-- If required, please update the title to be clear and descriptive --> Command: `prisma db pull` Version: `4.4.0` Binary Version: `f352a33b70356f46311da8b00d83386dd9f145d6` Report: https://prisma-errors.netlify.app/report/14372 OS: `arm64 darwin 21.5.0`
1.0
Error: [libs/dml/src/datamodel.rs:178:14] Every RelationInfo should have a complementary RelationInfo on the opposite relation field. - <!-- If required, please update the title to be clear and descriptive --> Command: `prisma db pull` Version: `4.4.0` Binary Version: `f352a33b70356f46311da8b00d83386dd9f145d6` Report: https://prisma-errors.netlify.app/report/14372 OS: `arm64 darwin 21.5.0`
process
error every relationinfo should have a complementary relationinfo on the opposite relation field command prisma db pull version binary version report os darwin
1
154,218
19,710,927,730
IssuesEvent
2022-01-13 05:10:40
ChoeMinji/redis-6.2.3
https://api.github.com/repos/ChoeMinji/redis-6.2.3
opened
CVE-2021-32762 (High) detected in redis6.2.6
security vulnerability
## CVE-2021-32762 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>redis6.2.6</b></p></summary> <p> <p>Redis is an in-memory database that persists on disk. The data model is key-value, but many different kind of values are supported: Strings, Lists, Sets, Sorted Sets, Hashes, Streams, HyperLogLogs, Bitmaps.</p> <p>Library home page: <a href=https://github.com/redis/redis.git>https://github.com/redis/redis.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/ChoeMinji/redis-6.2.3/commit/5b7aad699eab111b003a1a58082434041ce3e008">5b7aad699eab111b003a1a58082434041ce3e008</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/deps/hiredis/hiredis.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Redis is an open source, in-memory database that persists on disk. The redis-cli command line tool and redis-sentinel service may be vulnerable to integer overflow when parsing specially crafted large multi-bulk network replies. This is a result of a vulnerability in the underlying hiredis library which does not perform an overflow check before calling the calloc() heap allocation function. This issue only impacts systems with heap allocators that do not perform their own overflow checks. Most modern systems do and are therefore not likely to be affected. Furthermore, by default redis-sentinel uses the jemalloc allocator which is also not vulnerable. The problem is fixed in Redis versions 6.2.6, 6.0.16 and 5.0.14. <p>Publish Date: 2021-10-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32762>CVE-2021-32762</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/redis/redis/security/advisories/GHSA-833w-8v3m-8wwr">https://github.com/redis/redis/security/advisories/GHSA-833w-8v3m-8wwr</a></p> <p>Release Date: 2021-10-04</p> <p>Fix Resolution: 5.0.14, 6.0.16, 6.2.6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-32762 (High) detected in redis6.2.6 - ## CVE-2021-32762 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>redis6.2.6</b></p></summary> <p> <p>Redis is an in-memory database that persists on disk. The data model is key-value, but many different kind of values are supported: Strings, Lists, Sets, Sorted Sets, Hashes, Streams, HyperLogLogs, Bitmaps.</p> <p>Library home page: <a href=https://github.com/redis/redis.git>https://github.com/redis/redis.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/ChoeMinji/redis-6.2.3/commit/5b7aad699eab111b003a1a58082434041ce3e008">5b7aad699eab111b003a1a58082434041ce3e008</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/deps/hiredis/hiredis.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Redis is an open source, in-memory database that persists on disk. The redis-cli command line tool and redis-sentinel service may be vulnerable to integer overflow when parsing specially crafted large multi-bulk network replies. This is a result of a vulnerability in the underlying hiredis library which does not perform an overflow check before calling the calloc() heap allocation function. This issue only impacts systems with heap allocators that do not perform their own overflow checks. Most modern systems do and are therefore not likely to be affected. Furthermore, by default redis-sentinel uses the jemalloc allocator which is also not vulnerable. The problem is fixed in Redis versions 6.2.6, 6.0.16 and 5.0.14. <p>Publish Date: 2021-10-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32762>CVE-2021-32762</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/redis/redis/security/advisories/GHSA-833w-8v3m-8wwr">https://github.com/redis/redis/security/advisories/GHSA-833w-8v3m-8wwr</a></p> <p>Release Date: 2021-10-04</p> <p>Fix Resolution: 5.0.14, 6.0.16, 6.2.6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in cve high severity vulnerability vulnerable library redis is an in memory database that persists on disk the data model is key value but many different kind of values are supported strings lists sets sorted sets hashes streams hyperloglogs bitmaps library home page a href found in head commit a href found in base branch master vulnerable source files deps hiredis hiredis c vulnerability details redis is an open source in memory database that persists on disk the redis cli command line tool and redis sentinel service may be vulnerable to integer overflow when parsing specially crafted large multi bulk network replies this is a result of a vulnerability in the underlying hiredis library which does not perform an overflow check before calling the calloc heap allocation function this issue only impacts systems with heap allocators that do not perform their own overflow checks most modern systems do and are therefore not likely to be affected furthermore by default redis sentinel uses the jemalloc allocator which is also not vulnerable the problem is fixed in redis versions and publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
627,538
19,908,105,021
IssuesEvent
2022-01-25 14:44:16
xournalpp/xournalpp
https://api.github.com/repos/xournalpp/xournalpp
closed
Crush session when close xournal++
bug priority::high confirmed wayland
(Please complete the following information, and then delete this line) **Affects versions :** - OS: manjaro - (Linux only) gnome wayland - libgtk 3.24.30 - xournal++ 1.1.0+dev **Describe the bug** If I use xournalpp for some time maybe hours my de restart **To Reproduce** Steps to reproduce the behavior: 1. Step 1 open xournal and use it for hours 2. Step 2 close it 3. etc. **Expected behavior** A clear and concise description of what you expected to happen. **Screenshots of Problem** (Optional) If applicable, add screenshots to help explain your problem. ![IMG_20211111_125351.jpg](https://user-images.githubusercontent.com/50950554/141294422-582b80d2-7043-4d6c-af47-babddab22d87.jpg) **Additional context** Add any other information or context about the problem here.
1.0
Crush session when close xournal++ - (Please complete the following information, and then delete this line) **Affects versions :** - OS: manjaro - (Linux only) gnome wayland - libgtk 3.24.30 - xournal++ 1.1.0+dev **Describe the bug** If I use xournalpp for some time maybe hours my de restart **To Reproduce** Steps to reproduce the behavior: 1. Step 1 open xournal and use it for hours 2. Step 2 close it 3. etc. **Expected behavior** A clear and concise description of what you expected to happen. **Screenshots of Problem** (Optional) If applicable, add screenshots to help explain your problem. ![IMG_20211111_125351.jpg](https://user-images.githubusercontent.com/50950554/141294422-582b80d2-7043-4d6c-af47-babddab22d87.jpg) **Additional context** Add any other information or context about the problem here.
non_process
crush session when close xournal please complete the following information and then delete this line affects versions os manjaro linux only gnome wayland libgtk xournal dev describe the bug if i use xournalpp for some time maybe hours my de restart to reproduce steps to reproduce the behavior step open xournal and use it for hours step close it etc expected behavior a clear and concise description of what you expected to happen screenshots of problem optional if applicable add screenshots to help explain your problem additional context add any other information or context about the problem here
0
10,310
13,156,020,143
IssuesEvent
2020-08-10 09:57:51
nodejs/node
https://api.github.com/repos/nodejs/node
closed
child_process spawn does not pass environment key-value pairs on Windows
child_process question windows
<!-- Thank you for reporting an issue. This issue tracker is for bugs and issues found within Node.js core. If you require more general support please file an issue on our help repo. https://github.com/nodejs/help Please fill in as much of the template below as you're able. Version: output of `node -v` Platform: output of `uname -a` (UNIX), or version and 32 or 64-bit (Windows) Subsystem: if known, please specify affected core module name --> * **Version**: 12.18.3 * **Platform**: 64-bit Windows ### What steps will reproduce the bug? ``` require('child_process').spawn('node', ['-pe', 'process.env.PATH'], { stdio: 'inherit', shell: true, env: { ...process.env, PATH: process.env.PATH + require('path').delimiter + __dirname, } }); ``` Actual behavior: PATH environment variable is not updated in the spawned process. ### How often does it reproduce? Is there a required condition? Always on Windows. It is not an issue on OSX. ### What is the expected behavior? PATH environment variable should be updated in the spawned process. ### What do you see instead? add a fake env variable to the env object passed to spawn ``` require('child_process').spawn('node', ['-pe', 'process.env.PATH'], { stdio: 'inherit', shell: true, env: { ...process.env, PATH: process.env.PATH + require('path').delimiter + __dirname, FAKEENV: 'FAKEVALUE', } }); ```
1.0
child_process spawn does not pass environment key-value pairs on Windows - <!-- Thank you for reporting an issue. This issue tracker is for bugs and issues found within Node.js core. If you require more general support please file an issue on our help repo. https://github.com/nodejs/help Please fill in as much of the template below as you're able. Version: output of `node -v` Platform: output of `uname -a` (UNIX), or version and 32 or 64-bit (Windows) Subsystem: if known, please specify affected core module name --> * **Version**: 12.18.3 * **Platform**: 64-bit Windows ### What steps will reproduce the bug? ``` require('child_process').spawn('node', ['-pe', 'process.env.PATH'], { stdio: 'inherit', shell: true, env: { ...process.env, PATH: process.env.PATH + require('path').delimiter + __dirname, } }); ``` Actual behavior: PATH environment variable is not updated in the spawned process. ### How often does it reproduce? Is there a required condition? Always on Windows. It is not an issue on OSX. ### What is the expected behavior? PATH environment variable should be updated in the spawned process. ### What do you see instead? add a fake env variable to the env object passed to spawn ``` require('child_process').spawn('node', ['-pe', 'process.env.PATH'], { stdio: 'inherit', shell: true, env: { ...process.env, PATH: process.env.PATH + require('path').delimiter + __dirname, FAKEENV: 'FAKEVALUE', } }); ```
process
child process spawn does not pass environment key value pairs on windows thank you for reporting an issue this issue tracker is for bugs and issues found within node js core if you require more general support please file an issue on our help repo please fill in as much of the template below as you re able version output of node v platform output of uname a unix or version and or bit windows subsystem if known please specify affected core module name version platform bit windows what steps will reproduce the bug require child process spawn node stdio inherit shell true env process env path process env path require path delimiter dirname actual behavior path environment variable is not updated in the spawned process how often does it reproduce is there a required condition always on windows it is not an issue on osx what is the expected behavior path environment variable should be updated in the spawned process what do you see instead add a fake env variable to the env object passed to spawn require child process spawn node stdio inherit shell true env process env path process env path require path delimiter dirname fakeenv fakevalue
1
10,279
13,131,969,874
IssuesEvent
2020-08-06 18:00:31
ORNL-AMO/AMO-Tools-Desktop
https://api.github.com/repos/ORNL-AMO/AMO-Tools-Desktop
closed
O2 enrichment tweaks
Calculator Process Heating
Change headers to "Baseline" and "Modification" Maybe reformat to be like other BL vs Mod calcs (almost any calc from the TH) {this might be better as a separate issue} Once the user hits "Plot" the BL is "locked" (unlocks when user hits "Reset data") then the user can change various fields in the mod and "plot" more lines New graphs: y always = Savings Different X axis: Combustion preheat temp, flue gas temp, O2 in flue gas Maybe switch between them like steam properties (a dropdown near the X axis?) or different tabs?
1.0
O2 enrichment tweaks - Change headers to "Baseline" and "Modification" Maybe reformat to be like other BL vs Mod calcs (almost any calc from the TH) {this might be better as a separate issue} Once the user hits "Plot" the BL is "locked" (unlocks when user hits "Reset data") then the user can change various fields in the mod and "plot" more lines New graphs: y always = Savings Different X axis: Combustion preheat temp, flue gas temp, O2 in flue gas Maybe switch between them like steam properties (a dropdown near the X axis?) or different tabs?
process
enrichment tweaks change headers to baseline and modification maybe reformat to be like other bl vs mod calcs almost any calc from the th this might be better as a separate issue once the user hits plot the bl is locked unlocks when user hits reset data then the user can change various fields in the mod and plot more lines new graphs y always savings different x axis combustion preheat temp flue gas temp in flue gas maybe switch between them like steam properties a dropdown near the x axis or different tabs
1
1,020
3,480,679,523
IssuesEvent
2015-12-29 10:04:44
DevExpress/testcafe-hammerhead
https://api.github.com/repos/DevExpress/testcafe-hammerhead
opened
urlUtils.getProxyUrl always add the trailing slash
AREA: client SYSTEM: URL processing TYPE: bug
For example: urlUtils.getProxyUrl('http://google.com') -> "http://<host>:<port>/<sessionId>/http://google.com/" Reproduced in IE, Chrome, FF
1.0
urlUtils.getProxyUrl always add the trailing slash - For example: urlUtils.getProxyUrl('http://google.com') -> "http://<host>:<port>/<sessionId>/http://google.com/" Reproduced in IE, Chrome, FF
process
urlutils getproxyurl always add the trailing slash for example urlutils getproxyurl reproduced in ie chrome ff
1
240,907
7,807,254,702
IssuesEvent
2018-06-11 16:16:15
fac-13/GP_ProjectBernadette
https://api.github.com/repos/fac-13/GP_ProjectBernadette
closed
Manage focus with auto scroll
priority-3
When using auto scroll, the page does scroll to the bottom correctly however, focus remains at the top. This is problematic for keyboard users
1.0
Manage focus with auto scroll - When using auto scroll, the page does scroll to the bottom correctly however, focus remains at the top. This is problematic for keyboard users
non_process
manage focus with auto scroll when using auto scroll the page does scroll to the bottom correctly however focus remains at the top this is problematic for keyboard users
0
2,456
5,240,609,797
IssuesEvent
2017-01-31 13:39:54
openvstorage/alba
https://api.github.com/repos/openvstorage/alba
closed
Namespace names in OSD backend or Cache backend are not UTF-8
priority_normal process_wontfix type_bug
Namespace names in OSD backend or Cache backend are not UTF-8. While not strictly required, UTF-8 is advised for JSON, and the python json parser seems to depend on this advise as it can't parse the JSON output returned by e.g. show-namespaces. It would be nice if the counter could be added in some more readable way, e.g. `<prefix>_<counter>` where counter is `00000001`, `00000002`, and so on.
1.0
Namespace names in OSD backend or Cache backend are not UTF-8 - Namespace names in OSD backend or Cache backend are not UTF-8. While not strictly required, UTF-8 is advised for JSON, and the python json parser seems to depend on this advise as it can't parse the JSON output returned by e.g. show-namespaces. It would be nice if the counter could be added in some more readable way, e.g. `<prefix>_<counter>` where counter is `00000001`, `00000002`, and so on.
process
namespace names in osd backend or cache backend are not utf namespace names in osd backend or cache backend are not utf while not strictly required utf is advised for json and the python json parser seems to depend on this advise as it can t parse the json output returned by e g show namespaces it would be nice if the counter could be added in some more readable way e g where counter is and so on
1
33,710
7,200,534,791
IssuesEvent
2018-02-05 19:23:23
Openki/Openki
https://api.github.com/repos/Openki/Openki
opened
Notifications: Mails classified as spam by gmx
Defect
related: #874, #802 Html mails from Openki go into spam with reason: `6 (nemesis text pattern profiler); Detail=V3;` (most other mails recognized as Junk have same text there) I think all html mails go into spam in all gmx accounts. Can someone confirm in order to define priority of this problem? ``` Content-Type: text/html; charset=utf-8 Subject: [Openki] Private message from greg DKIM-Signature: a=rsa-sha256; v=1; c=relaxed/relaxed; d=mail.openki.net; q=dns/txt; s=smtp; t=1517857287; h=MIME-Version: Date: Content-Transfer-Encoding: Message-ID: Subject: To: Sender: From: Content-Type; bh=rmWxmmQe7gHbsUbtiXjvXLNmqJnU8gQFKyTl0NBLYCY=; b=itu4bRnST8koq9etAgycBkphHij7mfQJSBVOXnocs0tIhdpvUcoZKTLn8Ng/sSQvLmlePCyb wZb8C7tM0A6dcj6PUj7CykYL2UModPwrktMjF9zZ1LgBIaMWha8s6OBAOs3PKFvB4WGf/Xyq xSL6T2IIJuN63NazVxV4zq5ptVA= X-GMX-Antispam: 6 (nemesis text pattern profiler); Detail=V3; X-UI-Filterresults: junk:10;V01:K0:2hp2TGwjjHg=:G+F7Ms/vrRNhTTZ+BtF13Q8nrw7b oojOC2iGwPE14uAjctORiCdUOEY5zXa/enanXqfC5SgJjvN0+ty9kYk0Sa+HQyHeVWOcuQd6r jz6Ac38HpwaKOQgqQ7ZitNK5cZ6v623ncE1q7p820w6W+hC7/T3mfcXQO5kUUUVpgF6zzo5jK dnxiD+ayRKlBOJdNC/BNFdKP72zgQZlWA7QwGec/0b4EWtYa6BQY3FAhzmpTQAFn3Rs7BSxha +QCeqvzwm1fqE85wVBJpSxPns4LEGOGn9FaEc2V35uG4MTHVU84+NAd2FIHc896qjRQGSXFBK hwc/+D+gnUU+utSx31NzdjgltDnM7Gy8w4j/PDhTDS5NSci0OC7/oYNTZ5/IsgHcvjwIM9pDQ hJyHxQImiiBmQp0JHsf5vXvIMcSGpj1igxIFzYTqu0aDHVjEuWu1qLX7Wu1YpZQCjkQWWKoUs wgY02wKvfZhIJ2aby/NOnfLZqBAgvyBeV23EVPMHccr3hxYC+AX2bZlCZrj6NhXxW1VMRUiOK 0KmU6TOveDEZgQS/d2qTMGEV0jYid8y8TJgRmVhT6mz+3QjojbCIOSTFKwa5yWBRJ+GI3mLvz 0ugOWmRPX622tU7BVLGN/KdjsWj13k5WnBigcYldcYWFgMiT+E75rdzCqFeP/nJYLSSWbUvrC LJGENknwftnpJNnCP2WFNY65PqlgEKo4zetIXPHsp2JpjoH70T37MSbv25w7KaBCvWUdOrQHs J33jQzN6eRF3Wvvl7xCFlWdB37Sohq0rLhzTwupglKUc+fU8js785aIv3jc7ZK7uniehaxDAL /JVS05lBqAfaxs9fphGGJKTqN4iODRJq+VjhFv0V2TIdeTXdew/ho+6/BJBfyiOv4XkumPrhA AqiBkDZRIG3VlZ2IpL6HAWRJ9RZZPSY5eDsvpOeIfNYgLFfJB92yYpOZ5kFEoZno9I1K12aHq IxK9F2krKnWbijsGVy1YVJPidLESqiM8kWBJAGk5xdW12CGhVPweFgDiwcXYh0KK9SDlVp5xn bxYQw7TJjiaVsXL47DHbdvlWWM55CUzsgT1v8g5SqI4FWW2UmdVoYdAXr9e67SHPdY5D9NAEV 5yL49utwf2wi6NsCKx/oZhqxcIDFzxCaPkJaRi09GCZ/uRMI8f1FbWs556isoULiCf5r7xQfr nFUVqJb5VlhVeVJChHk8S6IxNjx37Cf5Wcr6ImU6RqgzMA6DFM6OhzGyohldiKSEeYamS7YP4 C95rnKNkhgLmRd01ELAUuA/1puz82R56y516ZFCucDmoQZrRD4TDY8TRJo4J3il3Mf5Jn+az2 OnEn7d5SHSyE4fKCO6B7vX7pIScKj0ZBOVhrgbztC17Rti7reiQtZyblSbS9AqX+vkqXVhP0B o+O0Bn5GXWmFNMjtMyRztOHAEyRA8Jo24yBIc7QprCU5k5YORpCOdLPS6+SNtJY8P6mrdnzu2 HULf5gTl4uOxPJmeVbL/FplmvrLyrz4dQExtIt6b0w83+abkRlIxJeGzpR29OIiiEz5i3q+mS YV33HCrzpRjxLYQM18GafaMX6VWaKlz3JbhCFgwaXQ6Pb5z7ZU6O3y/ngbKVsE2fgYwv/rbnw hUkg80hKGGMyPH9qp8DnxjJmi9BwsIUHQ9bulS9KdzeVUTp0+1tr1SMkcjagO0iQamVWl6gsW /UB4TaSUVVaF3ixiYDLlWqcAlbMnGmnihgV37BAJfb/nQVfu7DhIT1ocVdFx8LXq43JdE3wUB pScnPZwTXnXGbyLXnyqbU/k3J+KSiu96NA== ``` this one didn't: ``` Content-Type: text/plain Subject: Reset your password on Openki DKIM-Signature: a=rsa-sha256; v=1; c=relaxed/relaxed; d=mail.openki.net; q=dns/txt; s=smtp; t=1517857782; h=MIME-Version: Date: Content-Transfer-Encoding: Message-ID: Subject: To: From: Content-Type: Sender; bh=oKUJN/Zp/axIt1TxpObfXxRaDXYHFFxDfStAEWtaZjs=; b=FX1PFwSPC0pzPgAuqHJeRMhFcjpj4gvTYc+HRaJgHiDBocQI+TfwFy+9tDjfQ4uLEIMQD8Vp gDs02VzFwyXVE5gV8G8/aSoASesN3KgQMgRpyt3LTxL717ogfj4Bry+pPxOHKJMu3s6lFAdF xkhfs5XMhVVNdzd/8gqJKybgroE= X-GMX-Antispam: 0 (Mail was not recognized as spam); Detail=V3; X-UI-Filterresults: notjunk:1;V01:K0:E+67RFQNlC8=:wfCyEvRQIf1Qg1xCghnW2ks5V8 U0pkQ9zA7JmK7Trq4Qjg3oTOOqKb70H3q/5DmPDcXuaU8yfoWaZWb0V659TVDbGp437jNcVGm p9k6O/owYlLtmcQbGjvRsNlvDJGZMCGjJILDLIyiKVMIE2rxzAaYUPuhtnKaSx19RbnrAnZsQ aQHgnJQrT243vL1QPkbulfuBoS1kQ8/r8q7cUDD6rxvg74hxM+EyVaqPh1WyguUwrNOWF3tZN K+O5yvp0YTq8cciAY7tNYzu7xVvqo/BLVGph2CSMe+azRrNd1tP627GvOoDrJUoP1Afg/JwFY 69xhBLUJ+VmBmRe37IYkTshdRUM+nJcno/XHeTIUKb/27S9A2zfFht/lCe9gfPLXCWpnPaoKe 77OVcqYHrRObb7nE2T1PTLYTkZh9SFBtyqfWBkzBYMUgVOwcERj51WsjNdt0IZT2168tvuFu6 2kRF4KAg+I4vSOFOd14eSpFy+6zpgkCNvWDkq83p8KmCc3Nx1cKBm1hRK5OKaE5f3MqG8uSCd Ew+4XfJClVK7xGe8/b9yZ04EaTFzJ3435nc4IXaHaAB9XcSUSJuUHsLzblxWHAG7GwyL6HfHK peJnMk8utaKljvxHrz3MFfaXEJlyLGsVe3phNeLX/2R5FDoEBZt+3TYJx5AwNsHNktjGRhLud 8/8iNd9/0SOrsY/xFu55YjCe9iAdi9FSiZSDeo6WfqWHKb75px7NMUN2iKO+YF4aRjBfTB3EK Al8hVJF1+eFDd0dz65bzbcV9UCvosNp4PZ/SLT19cugz2eiV0RDnyCPBaMpAPwYjzYpAkmhjm DBCN4HloHTwJEo7MbkygUyR299O+wTDhiNR9Z8fDht3hwSfwS2j9SY+4qsmswRf4Id94O10NJ cwZDEm4SCziaGFV70a9C9ADuFzAW1jrfcQkvpRRe/Zm966CwZ0DW7eDCWmefovc6LXdHITBhv kASG36BSW+zzw4cNneBnKaIgACWXFH37Cp7shkVg/kuOYbV2Tu+wxO7Xc5yG2Z5G8HbLfvyJ1 OxjTFJGZjNAsgmr0pVq0gQY6Sbt0TYfQS3x3RBvMnp+MYumwfIshOJehDyj1zbvnBx4/5aTTg xuuJhOviQ2FSDNkWMc8sA3SUo6EmPydv0dMqntnKHmwyHEtsKnjXg6BLmqmunpJbpQkxgleb9 QAP0we5LoJyQw2VoFL/FuTsaKwbWxTcPAPJLVLT3ErYRsAkTnVdSLwuBjM92R8uBbVDJWxJ3F DMrKbsuDSax4C03IAEStz6jbCIRgM/BfXNq0LmdBRwelViHszBCqcp9Les+Iu1J5D0w9Qzf4C yrl5pte8/PWrbfQtpnAK+omQODQDxqtjO3qHoCEagryD3Ym1Qi92/O3TN+ccHIg6iFPwRNvcz wSUjlCydXu25FIg0rxiy/k0LHjFLrM9t6QuLLpV5ET38se5sFqa6Ufy84K41RoCB1h8Rv6tL8 NDaeEK1mJT2KvEliYH6IalUrfJjXHG/g+KLUfUSRrGhzAINCUZCh5bjAMO1O21MBLBw3wNvod gk6GWGC+4QlP32UcWI4BU+poNYYBKQu/rdKrQ5KM9tDwW41ZKd4lWqqEQnfmSNhBzNaFJrWiG WnCKG/a9x1PesPgY+uEQ4XT8ImE8/6rO9EzssBmWzJaueeih+gUh9UpuT8iybCP6alxiT/ky5 R3ZK9qy2oT9gpzUiCiuCegEIBmDE455wzk87FbDXvWXd1kyAHcvUwucM52a7aTzkBqmnH4B4F U1eFxftRLxAmFdkF5Hcx61AYfNZSo3evrhLQ== ```
1.0
Notifications: Mails classified as spam by gmx - related: #874, #802 Html mails from Openki go into spam with reason: `6 (nemesis text pattern profiler); Detail=V3;` (most other mails recognized as Junk have same text there) I think all html mails go into spam in all gmx accounts. Can someone confirm in order to define priority of this problem? ``` Content-Type: text/html; charset=utf-8 Subject: [Openki] Private message from greg DKIM-Signature: a=rsa-sha256; v=1; c=relaxed/relaxed; d=mail.openki.net; q=dns/txt; s=smtp; t=1517857287; h=MIME-Version: Date: Content-Transfer-Encoding: Message-ID: Subject: To: Sender: From: Content-Type; bh=rmWxmmQe7gHbsUbtiXjvXLNmqJnU8gQFKyTl0NBLYCY=; b=itu4bRnST8koq9etAgycBkphHij7mfQJSBVOXnocs0tIhdpvUcoZKTLn8Ng/sSQvLmlePCyb wZb8C7tM0A6dcj6PUj7CykYL2UModPwrktMjF9zZ1LgBIaMWha8s6OBAOs3PKFvB4WGf/Xyq xSL6T2IIJuN63NazVxV4zq5ptVA= X-GMX-Antispam: 6 (nemesis text pattern profiler); Detail=V3; X-UI-Filterresults: junk:10;V01:K0:2hp2TGwjjHg=:G+F7Ms/vrRNhTTZ+BtF13Q8nrw7b oojOC2iGwPE14uAjctORiCdUOEY5zXa/enanXqfC5SgJjvN0+ty9kYk0Sa+HQyHeVWOcuQd6r jz6Ac38HpwaKOQgqQ7ZitNK5cZ6v623ncE1q7p820w6W+hC7/T3mfcXQO5kUUUVpgF6zzo5jK dnxiD+ayRKlBOJdNC/BNFdKP72zgQZlWA7QwGec/0b4EWtYa6BQY3FAhzmpTQAFn3Rs7BSxha +QCeqvzwm1fqE85wVBJpSxPns4LEGOGn9FaEc2V35uG4MTHVU84+NAd2FIHc896qjRQGSXFBK hwc/+D+gnUU+utSx31NzdjgltDnM7Gy8w4j/PDhTDS5NSci0OC7/oYNTZ5/IsgHcvjwIM9pDQ hJyHxQImiiBmQp0JHsf5vXvIMcSGpj1igxIFzYTqu0aDHVjEuWu1qLX7Wu1YpZQCjkQWWKoUs wgY02wKvfZhIJ2aby/NOnfLZqBAgvyBeV23EVPMHccr3hxYC+AX2bZlCZrj6NhXxW1VMRUiOK 0KmU6TOveDEZgQS/d2qTMGEV0jYid8y8TJgRmVhT6mz+3QjojbCIOSTFKwa5yWBRJ+GI3mLvz 0ugOWmRPX622tU7BVLGN/KdjsWj13k5WnBigcYldcYWFgMiT+E75rdzCqFeP/nJYLSSWbUvrC LJGENknwftnpJNnCP2WFNY65PqlgEKo4zetIXPHsp2JpjoH70T37MSbv25w7KaBCvWUdOrQHs J33jQzN6eRF3Wvvl7xCFlWdB37Sohq0rLhzTwupglKUc+fU8js785aIv3jc7ZK7uniehaxDAL /JVS05lBqAfaxs9fphGGJKTqN4iODRJq+VjhFv0V2TIdeTXdew/ho+6/BJBfyiOv4XkumPrhA AqiBkDZRIG3VlZ2IpL6HAWRJ9RZZPSY5eDsvpOeIfNYgLFfJB92yYpOZ5kFEoZno9I1K12aHq IxK9F2krKnWbijsGVy1YVJPidLESqiM8kWBJAGk5xdW12CGhVPweFgDiwcXYh0KK9SDlVp5xn bxYQw7TJjiaVsXL47DHbdvlWWM55CUzsgT1v8g5SqI4FWW2UmdVoYdAXr9e67SHPdY5D9NAEV 5yL49utwf2wi6NsCKx/oZhqxcIDFzxCaPkJaRi09GCZ/uRMI8f1FbWs556isoULiCf5r7xQfr nFUVqJb5VlhVeVJChHk8S6IxNjx37Cf5Wcr6ImU6RqgzMA6DFM6OhzGyohldiKSEeYamS7YP4 C95rnKNkhgLmRd01ELAUuA/1puz82R56y516ZFCucDmoQZrRD4TDY8TRJo4J3il3Mf5Jn+az2 OnEn7d5SHSyE4fKCO6B7vX7pIScKj0ZBOVhrgbztC17Rti7reiQtZyblSbS9AqX+vkqXVhP0B o+O0Bn5GXWmFNMjtMyRztOHAEyRA8Jo24yBIc7QprCU5k5YORpCOdLPS6+SNtJY8P6mrdnzu2 HULf5gTl4uOxPJmeVbL/FplmvrLyrz4dQExtIt6b0w83+abkRlIxJeGzpR29OIiiEz5i3q+mS YV33HCrzpRjxLYQM18GafaMX6VWaKlz3JbhCFgwaXQ6Pb5z7ZU6O3y/ngbKVsE2fgYwv/rbnw hUkg80hKGGMyPH9qp8DnxjJmi9BwsIUHQ9bulS9KdzeVUTp0+1tr1SMkcjagO0iQamVWl6gsW /UB4TaSUVVaF3ixiYDLlWqcAlbMnGmnihgV37BAJfb/nQVfu7DhIT1ocVdFx8LXq43JdE3wUB pScnPZwTXnXGbyLXnyqbU/k3J+KSiu96NA== ``` this one didn't: ``` Content-Type: text/plain Subject: Reset your password on Openki DKIM-Signature: a=rsa-sha256; v=1; c=relaxed/relaxed; d=mail.openki.net; q=dns/txt; s=smtp; t=1517857782; h=MIME-Version: Date: Content-Transfer-Encoding: Message-ID: Subject: To: From: Content-Type: Sender; bh=oKUJN/Zp/axIt1TxpObfXxRaDXYHFFxDfStAEWtaZjs=; b=FX1PFwSPC0pzPgAuqHJeRMhFcjpj4gvTYc+HRaJgHiDBocQI+TfwFy+9tDjfQ4uLEIMQD8Vp gDs02VzFwyXVE5gV8G8/aSoASesN3KgQMgRpyt3LTxL717ogfj4Bry+pPxOHKJMu3s6lFAdF xkhfs5XMhVVNdzd/8gqJKybgroE= X-GMX-Antispam: 0 (Mail was not recognized as spam); Detail=V3; X-UI-Filterresults: notjunk:1;V01:K0:E+67RFQNlC8=:wfCyEvRQIf1Qg1xCghnW2ks5V8 U0pkQ9zA7JmK7Trq4Qjg3oTOOqKb70H3q/5DmPDcXuaU8yfoWaZWb0V659TVDbGp437jNcVGm p9k6O/owYlLtmcQbGjvRsNlvDJGZMCGjJILDLIyiKVMIE2rxzAaYUPuhtnKaSx19RbnrAnZsQ aQHgnJQrT243vL1QPkbulfuBoS1kQ8/r8q7cUDD6rxvg74hxM+EyVaqPh1WyguUwrNOWF3tZN K+O5yvp0YTq8cciAY7tNYzu7xVvqo/BLVGph2CSMe+azRrNd1tP627GvOoDrJUoP1Afg/JwFY 69xhBLUJ+VmBmRe37IYkTshdRUM+nJcno/XHeTIUKb/27S9A2zfFht/lCe9gfPLXCWpnPaoKe 77OVcqYHrRObb7nE2T1PTLYTkZh9SFBtyqfWBkzBYMUgVOwcERj51WsjNdt0IZT2168tvuFu6 2kRF4KAg+I4vSOFOd14eSpFy+6zpgkCNvWDkq83p8KmCc3Nx1cKBm1hRK5OKaE5f3MqG8uSCd Ew+4XfJClVK7xGe8/b9yZ04EaTFzJ3435nc4IXaHaAB9XcSUSJuUHsLzblxWHAG7GwyL6HfHK peJnMk8utaKljvxHrz3MFfaXEJlyLGsVe3phNeLX/2R5FDoEBZt+3TYJx5AwNsHNktjGRhLud 8/8iNd9/0SOrsY/xFu55YjCe9iAdi9FSiZSDeo6WfqWHKb75px7NMUN2iKO+YF4aRjBfTB3EK Al8hVJF1+eFDd0dz65bzbcV9UCvosNp4PZ/SLT19cugz2eiV0RDnyCPBaMpAPwYjzYpAkmhjm DBCN4HloHTwJEo7MbkygUyR299O+wTDhiNR9Z8fDht3hwSfwS2j9SY+4qsmswRf4Id94O10NJ cwZDEm4SCziaGFV70a9C9ADuFzAW1jrfcQkvpRRe/Zm966CwZ0DW7eDCWmefovc6LXdHITBhv kASG36BSW+zzw4cNneBnKaIgACWXFH37Cp7shkVg/kuOYbV2Tu+wxO7Xc5yG2Z5G8HbLfvyJ1 OxjTFJGZjNAsgmr0pVq0gQY6Sbt0TYfQS3x3RBvMnp+MYumwfIshOJehDyj1zbvnBx4/5aTTg xuuJhOviQ2FSDNkWMc8sA3SUo6EmPydv0dMqntnKHmwyHEtsKnjXg6BLmqmunpJbpQkxgleb9 QAP0we5LoJyQw2VoFL/FuTsaKwbWxTcPAPJLVLT3ErYRsAkTnVdSLwuBjM92R8uBbVDJWxJ3F DMrKbsuDSax4C03IAEStz6jbCIRgM/BfXNq0LmdBRwelViHszBCqcp9Les+Iu1J5D0w9Qzf4C yrl5pte8/PWrbfQtpnAK+omQODQDxqtjO3qHoCEagryD3Ym1Qi92/O3TN+ccHIg6iFPwRNvcz wSUjlCydXu25FIg0rxiy/k0LHjFLrM9t6QuLLpV5ET38se5sFqa6Ufy84K41RoCB1h8Rv6tL8 NDaeEK1mJT2KvEliYH6IalUrfJjXHG/g+KLUfUSRrGhzAINCUZCh5bjAMO1O21MBLBw3wNvod gk6GWGC+4QlP32UcWI4BU+poNYYBKQu/rdKrQ5KM9tDwW41ZKd4lWqqEQnfmSNhBzNaFJrWiG WnCKG/a9x1PesPgY+uEQ4XT8ImE8/6rO9EzssBmWzJaueeih+gUh9UpuT8iybCP6alxiT/ky5 R3ZK9qy2oT9gpzUiCiuCegEIBmDE455wzk87FbDXvWXd1kyAHcvUwucM52a7aTzkBqmnH4B4F U1eFxftRLxAmFdkF5Hcx61AYfNZSo3evrhLQ== ```
non_process
notifications mails classified as spam by gmx related html mails from openki go into spam with reason nemesis text pattern profiler detail most other mails recognized as junk have same text there i think all html mails go into spam in all gmx accounts can someone confirm in order to define priority of this problem content type text html charset utf subject private message from greg dkim signature a rsa v c relaxed relaxed d mail openki net q dns txt s smtp t h mime version date content transfer encoding message id subject to sender from content type bh b ssqvlmlepcyb xyq x gmx antispam nemesis text pattern profiler detail x ui filterresults junk g vrrnhttz dnxid ayrklbojdnc hwc d gnuu njylsswbuvrc ho o ms rbnw pscnpzwtxnxgbylxnyqbu this one didn t content type text plain subject reset your password on openki dkim signature a rsa v c relaxed relaxed d mail openki net q dns txt s smtp t h mime version date content transfer encoding message id subject to from content type sender bh okujn zp b hrajghidbocqi tfwfy x gmx antispam mail was not recognized as spam detail x ui filterresults notjunk e k jwfy njcno xhetiukb ew pwrbfqtpnak g ponyybkqu wnckg
0
22,188
30,737,262,412
IssuesEvent
2023-07-28 08:39:31
dita-ot/dita-ot
https://api.github.com/repos/dita-ot/dita-ot
closed
Preprocess2 does not report validation error
bug priority/medium preprocess preprocess2
This problem is from following Stack Overflow question: minOccurs maxOccurs don't work with dita? https://stackoverflow.com/questions/55454803/minoccurs-maxoccurs-dont-work-with-dita The reproducing step: 1. Modify OASIS XSD file: dita-ot-3.3.1/plugins/org.oasis-open.dita.v1_3/schema/bookmap/xsd/bookmapMod.xsd to `minOccurs="1"`. ````xml <xs:group name="bookmeta.content"> <xs:sequence> <xs:group ref="linktext" minOccurs="1"/> <xs:group ref="searchtitle" minOccurs="1"/> <xs:group ref="shortdesc" minOccurs="1"/> <xs:group ref="author" minOccurs="1" maxOccurs="unbounded"/> <xs:group ref="source" minOccurs="1"/> <xs:group ref="publisherinformation" minOccurs="1" maxOccurs="unbounded"/> <xs:group ref="critdates" minOccurs="1"/> <xs:group ref="permissions" minOccurs="1"/> <xs:group ref="metadata" minOccurs="1" maxOccurs="unbounded"/> <xs:group ref="audience" minOccurs="1" maxOccurs="unbounded"/> <xs:group ref="category" minOccurs="1" maxOccurs="unbounded"/> <xs:group ref="keywords" minOccurs="1" maxOccurs="unbounded"/> <xs:group ref="prodinfo" minOccurs="1" maxOccurs="unbounded"/> <xs:group ref="othermeta" minOccurs="1" maxOccurs="unbounded"/> <xs:group ref="resourceid" minOccurs="1" maxOccurs="unbounded"/> <xs:group ref="ux-window" minOccurs="1" maxOccurs="unbounded"/> <xs:group ref="bookid" minOccurs="1"/> <xs:group ref="bookchangehistory" minOccurs="1" maxOccurs="unbounded"/> <xs:group ref="bookrights" minOccurs="1" maxOccurs="unbounded"/> <xs:group ref="data" minOccurs="1" maxOccurs="unbounded"/> </xs:sequence> </xs:group> ```` 2. Modify the sample file: dita-ot-3.3.1/docsrc/samples/taskbook.ditamap to use XML Schema ````xml <bookmap xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:noNamespaceSchemaLocation="urn:oasis:names:tc:dita:xsd:bookmap.xsd" id="taskbook"> <!--DOCTYPE bookmap PUBLIC "-//OASIS//DTD DITA BookMap//EN" "bookmap.dtd"--> <!-- This file is part of the DITA Open Toolkit project. See the accompanying LICENSE file for applicable license. --> <!-- (c) Copyright IBM Corp. 2004, 2007 All Rights Reserved. --> <!--bookmap id="taskbook"--> <booktitle> <booklibrary>Retro Tools</booklibrary> <mainbooktitle>Product tasks</mainbooktitle> <booktitlealt>Tasks and what they can do</booktitlealt> </booktitle> <bookmeta> <author>Howe Tuduit</author> <bookid> <isbn>071271271X</isbn> <booknumber>SG99-9999-00</booknumber> <maintainer> <organization>Retro Tools</organization> <person/> </maintainer> </bookid> <bookrights> <copyrfirst> <year>2004</year> </copyrfirst> <copyrlast> <year>2007</year> </copyrlast> <bookowner> <organization>Retro Tools, Inc.</organization> </bookowner> </bookrights> </bookmeta> ... </bookmap> ```` 3. Run dita command `bin/dita -f pdf -i docsrc/samples/taskbook.ditamap -o out -l out/log.txt -v` The result: `preprocess2` does not report any validation error, but if I change `preprocess2` to `preprocess` in dita-ot-3.3.1/plugins/pdf2/build.xml, the validation error is reported. > Error: Failed to run pipeline: [DOTJ012F][FATAL] Failed to parse the input file 'file:/D:/DITA-OT/dita-ot-3.3.1/docsrc/samples/taskbook.ditamap'.: file:/D:/DITA-OT/dita-ot-3.3.1/docsrc/samples/taskbook.ditamap Line 13:cvc-complex-type.2.4.a: Invalid content was found starting with element 'author'. One of '{linktext}' is expected. [log-preprocess2.txt](https://github.com/dita-ot/dita-ot/files/3064208/log-preprocess2.txt) [log-preprocess.txt](https://github.com/dita-ot/dita-ot/files/3064209/log-preprocess.txt) I think `preprocess2` should also report validation error.
2.0
Preprocess2 does not report validation error - This problem is from following Stack Overflow question: minOccurs maxOccurs don't work with dita? https://stackoverflow.com/questions/55454803/minoccurs-maxoccurs-dont-work-with-dita The reproducing step: 1. Modify OASIS XSD file: dita-ot-3.3.1/plugins/org.oasis-open.dita.v1_3/schema/bookmap/xsd/bookmapMod.xsd to `minOccurs="1"`. ````xml <xs:group name="bookmeta.content"> <xs:sequence> <xs:group ref="linktext" minOccurs="1"/> <xs:group ref="searchtitle" minOccurs="1"/> <xs:group ref="shortdesc" minOccurs="1"/> <xs:group ref="author" minOccurs="1" maxOccurs="unbounded"/> <xs:group ref="source" minOccurs="1"/> <xs:group ref="publisherinformation" minOccurs="1" maxOccurs="unbounded"/> <xs:group ref="critdates" minOccurs="1"/> <xs:group ref="permissions" minOccurs="1"/> <xs:group ref="metadata" minOccurs="1" maxOccurs="unbounded"/> <xs:group ref="audience" minOccurs="1" maxOccurs="unbounded"/> <xs:group ref="category" minOccurs="1" maxOccurs="unbounded"/> <xs:group ref="keywords" minOccurs="1" maxOccurs="unbounded"/> <xs:group ref="prodinfo" minOccurs="1" maxOccurs="unbounded"/> <xs:group ref="othermeta" minOccurs="1" maxOccurs="unbounded"/> <xs:group ref="resourceid" minOccurs="1" maxOccurs="unbounded"/> <xs:group ref="ux-window" minOccurs="1" maxOccurs="unbounded"/> <xs:group ref="bookid" minOccurs="1"/> <xs:group ref="bookchangehistory" minOccurs="1" maxOccurs="unbounded"/> <xs:group ref="bookrights" minOccurs="1" maxOccurs="unbounded"/> <xs:group ref="data" minOccurs="1" maxOccurs="unbounded"/> </xs:sequence> </xs:group> ```` 2. Modify the sample file: dita-ot-3.3.1/docsrc/samples/taskbook.ditamap to use XML Schema ````xml <bookmap xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:noNamespaceSchemaLocation="urn:oasis:names:tc:dita:xsd:bookmap.xsd" id="taskbook"> <!--DOCTYPE bookmap PUBLIC "-//OASIS//DTD DITA BookMap//EN" "bookmap.dtd"--> <!-- This file is part of the DITA Open Toolkit project. See the accompanying LICENSE file for applicable license. --> <!-- (c) Copyright IBM Corp. 2004, 2007 All Rights Reserved. --> <!--bookmap id="taskbook"--> <booktitle> <booklibrary>Retro Tools</booklibrary> <mainbooktitle>Product tasks</mainbooktitle> <booktitlealt>Tasks and what they can do</booktitlealt> </booktitle> <bookmeta> <author>Howe Tuduit</author> <bookid> <isbn>071271271X</isbn> <booknumber>SG99-9999-00</booknumber> <maintainer> <organization>Retro Tools</organization> <person/> </maintainer> </bookid> <bookrights> <copyrfirst> <year>2004</year> </copyrfirst> <copyrlast> <year>2007</year> </copyrlast> <bookowner> <organization>Retro Tools, Inc.</organization> </bookowner> </bookrights> </bookmeta> ... </bookmap> ```` 3. Run dita command `bin/dita -f pdf -i docsrc/samples/taskbook.ditamap -o out -l out/log.txt -v` The result: `preprocess2` does not report any validation error, but if I change `preprocess2` to `preprocess` in dita-ot-3.3.1/plugins/pdf2/build.xml, the validation error is reported. > Error: Failed to run pipeline: [DOTJ012F][FATAL] Failed to parse the input file 'file:/D:/DITA-OT/dita-ot-3.3.1/docsrc/samples/taskbook.ditamap'.: file:/D:/DITA-OT/dita-ot-3.3.1/docsrc/samples/taskbook.ditamap Line 13:cvc-complex-type.2.4.a: Invalid content was found starting with element 'author'. One of '{linktext}' is expected. [log-preprocess2.txt](https://github.com/dita-ot/dita-ot/files/3064208/log-preprocess2.txt) [log-preprocess.txt](https://github.com/dita-ot/dita-ot/files/3064209/log-preprocess.txt) I think `preprocess2` should also report validation error.
process
does not report validation error this problem is from following stack overflow question minoccurs maxoccurs don t work with dita the reproducing step modify oasis xsd file dita ot plugins org oasis open dita schema bookmap xsd bookmapmod xsd to minoccurs xml modify the sample file dita ot docsrc samples taskbook ditamap to use xml schema xml bookmap xmlns xsi xsi nonamespaceschemalocation urn oasis names tc dita xsd bookmap xsd id taskbook retro tools product tasks tasks and what they can do howe tuduit retro tools retro tools inc run dita command bin dita f pdf i docsrc samples taskbook ditamap o out l out log txt v the result does not report any validation error but if i change to preprocess in dita ot plugins build xml the validation error is reported error failed to run pipeline failed to parse the input file file d dita ot dita ot docsrc samples taskbook ditamap file d dita ot dita ot docsrc samples taskbook ditamap line cvc complex type a invalid content was found starting with element author one of linktext is expected i think should also report validation error
1
22,750
32,067,682,147
IssuesEvent
2023-09-25 05:18:47
open-telemetry/opentelemetry-collector-contrib
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
closed
Resource detection for EC2 partially works if collector is running as a systemd service on an EC2
bug Stale priority:p2 processor/resourcedetection closed as inactive
**Describe the bug** Hello team, I'm running the Otel collector as a systemd service on an ec2 instance. I have resource detection processor enabled on the collector with `detectors: ["ec2"]` and added configuration to gather tags for the EC2 instance. However it does not detect the EC2 information. - If I run the collector plainly on the instance host and not as a systemd service or - If I run the collector as a systemd service and only have `detectors: ["ec2"]` without the `tags` parameter the resource detection for EC2 starts working again. **Steps to reproduce** Enable detectors: ["ec2"] under resource detection processor with the `tags` parameter in config.yaml Run Otel collector as a systemd service **What did you expect to see?** I expected to see something similar to this in the collector startup log `info internal/resourcedetection.go:150 detected resource information {"kind": "processor", "name": "resourcedetection", "pipeline": "traces", "resource": {"cloud.account.id":"XXXXXXXXXXXX","cloud.availability_zone":"us-east-1c","cloud.platform":"aws_ec2","cloud.provider":"aws","cloud.region":"us-east-1","host.id":"i-XXXXXXXXXXXXXXXXX","host.image.id":"ami-XXXXXXXXXXXXXXXXX","host.name":"ip-XX-XX-XX-XX","host.type":"m5.xlarge"}}` **What did you see instead?** `warn internal/resourcedetection.go:141 failed to detect resource {"kind": "processor", "name": "resourcedetection", "pipeline": "traces", "error": "failed fetching ec2 instance tags: RequestError: send request failed\ncaused by: Post \"https://ec2.us-east-1.amazonaws.com/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)"}` **What version did you use?** Version: v0.58.0 **What config did you use?** Config: ``` receivers: hostmetrics: collection_interval: 10m scrapers: cpu: disk: load: filesystem: memory: network: paging: processes: process: docker_stats: collection_interval: 5m otlp: protocols: grpc: http: exporters: logging: loglevel: debug otlp: endpoint: $OTLP_ENDPOINT processors: batch: resourcedetection: detectors: ["ec2"] ec2: tags: - ^ASV$ - ^BA$ extensions: health_check: pprof: zpages: service: extensions: [pprof, zpages, health_check] pipelines: traces: receivers: [otlp] exporters: [logging, otlp] processors: [batch, resourcedetection] metrics: receivers: [otlp, hostmetrics, docker_stats] exporters: [otlp] processors: [batch] logs: receivers: [otlp, filelog] exporters: [otlp] processors: [batch] ``` **Environment** OS: Linux 5.10.130-118.517.amzn2.x86_64 Compiler(if manually compiled): (e.g., "go 14.2") **Additional context** Add any other context about the problem here.
1.0
Resource detection for EC2 partially works if collector is running as a systemd service on an EC2 - **Describe the bug** Hello team, I'm running the Otel collector as a systemd service on an ec2 instance. I have resource detection processor enabled on the collector with `detectors: ["ec2"]` and added configuration to gather tags for the EC2 instance. However it does not detect the EC2 information. - If I run the collector plainly on the instance host and not as a systemd service or - If I run the collector as a systemd service and only have `detectors: ["ec2"]` without the `tags` parameter the resource detection for EC2 starts working again. **Steps to reproduce** Enable detectors: ["ec2"] under resource detection processor with the `tags` parameter in config.yaml Run Otel collector as a systemd service **What did you expect to see?** I expected to see something similar to this in the collector startup log `info internal/resourcedetection.go:150 detected resource information {"kind": "processor", "name": "resourcedetection", "pipeline": "traces", "resource": {"cloud.account.id":"XXXXXXXXXXXX","cloud.availability_zone":"us-east-1c","cloud.platform":"aws_ec2","cloud.provider":"aws","cloud.region":"us-east-1","host.id":"i-XXXXXXXXXXXXXXXXX","host.image.id":"ami-XXXXXXXXXXXXXXXXX","host.name":"ip-XX-XX-XX-XX","host.type":"m5.xlarge"}}` **What did you see instead?** `warn internal/resourcedetection.go:141 failed to detect resource {"kind": "processor", "name": "resourcedetection", "pipeline": "traces", "error": "failed fetching ec2 instance tags: RequestError: send request failed\ncaused by: Post \"https://ec2.us-east-1.amazonaws.com/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)"}` **What version did you use?** Version: v0.58.0 **What config did you use?** Config: ``` receivers: hostmetrics: collection_interval: 10m scrapers: cpu: disk: load: filesystem: memory: network: paging: processes: process: docker_stats: collection_interval: 5m otlp: protocols: grpc: http: exporters: logging: loglevel: debug otlp: endpoint: $OTLP_ENDPOINT processors: batch: resourcedetection: detectors: ["ec2"] ec2: tags: - ^ASV$ - ^BA$ extensions: health_check: pprof: zpages: service: extensions: [pprof, zpages, health_check] pipelines: traces: receivers: [otlp] exporters: [logging, otlp] processors: [batch, resourcedetection] metrics: receivers: [otlp, hostmetrics, docker_stats] exporters: [otlp] processors: [batch] logs: receivers: [otlp, filelog] exporters: [otlp] processors: [batch] ``` **Environment** OS: Linux 5.10.130-118.517.amzn2.x86_64 Compiler(if manually compiled): (e.g., "go 14.2") **Additional context** Add any other context about the problem here.
process
resource detection for partially works if collector is running as a systemd service on an describe the bug hello team i m running the otel collector as a systemd service on an instance i have resource detection processor enabled on the collector with detectors and added configuration to gather tags for the instance however it does not detect the information if i run the collector plainly on the instance host and not as a systemd service or if i run the collector as a systemd service and only have detectors without the tags parameter the resource detection for starts working again steps to reproduce enable detectors under resource detection processor with the tags parameter in config yaml run otel collector as a systemd service what did you expect to see i expected to see something similar to this in the collector startup log info internal resourcedetection go detected resource information kind processor name resourcedetection pipeline traces resource cloud account id xxxxxxxxxxxx cloud availability zone us east cloud platform aws cloud provider aws cloud region us east host id i xxxxxxxxxxxxxxxxx host image id ami xxxxxxxxxxxxxxxxx host name ip xx xx xx xx host type xlarge what did you see instead warn internal resourcedetection go failed to detect resource kind processor name resourcedetection pipeline traces error failed fetching instance tags requesterror send request failed ncaused by post context deadline exceeded client timeout exceeded while awaiting headers what version did you use version what config did you use config receivers hostmetrics collection interval scrapers cpu disk load filesystem memory network paging processes process docker stats collection interval otlp protocols grpc http exporters logging loglevel debug otlp endpoint otlp endpoint processors batch resourcedetection detectors tags asv ba extensions health check pprof zpages service extensions pipelines traces receivers exporters processors metrics receivers exporters processors logs receivers exporters processors environment os linux compiler if manually compiled e g go additional context add any other context about the problem here
1
105,763
4,241,468,486
IssuesEvent
2016-07-06 16:23:55
CascadesCarnivoreProject/Timelapse
https://api.github.com/repos/CascadesCarnivoreProject/Timelapse
closed
UX: video dialog can only be displayed once
High Priority fix
The following scenario has undesirable behaviour: - user pops video dialog through menu - user closes video dialog - user selects the video dialog menu item a second time and nothing happens This is a state management bug in MenuItemVideoViewer_Click() causing the call to Show() to be one time only.
1.0
UX: video dialog can only be displayed once - The following scenario has undesirable behaviour: - user pops video dialog through menu - user closes video dialog - user selects the video dialog menu item a second time and nothing happens This is a state management bug in MenuItemVideoViewer_Click() causing the call to Show() to be one time only.
non_process
ux video dialog can only be displayed once the following scenario has undesirable behaviour user pops video dialog through menu user closes video dialog user selects the video dialog menu item a second time and nothing happens this is a state management bug in menuitemvideoviewer click causing the call to show to be one time only
0
1,573
4,167,451,307
IssuesEvent
2016-06-20 09:34:16
e-government-ua/iBP
https://api.github.com/repos/e-government-ua/iBP
closed
Видача дозволу на розміщення зовнішньої реклами - м.Первомайськ - Миколаївська область
In process of testing in work test
[Видача дозволу на розміщення зовнішньої реклами.pdf](https://github.com/e-government-ua/iBP/files/262917/default.pdf) **Координатор города:** Корой Виталий - 0991962269 - somati.orlik@gmail.com И еще просьба, в получатели добавить еще два контакта, для более быстрого реагирования ) kukharenko.vsevolod@gmail.com - Всеволод Кухаренко sergey.donchenko@gmail.com - Сергей Донченко **Контактное лицо ЦНАП:** Наталія Миколаївна Петрущак, natashadc@mail.ru, (099) 277-41-42 на тест отправляем в ЦНАП, но обязательно ставим в копию 3 координаторов выше.
1.0
Видача дозволу на розміщення зовнішньої реклами - м.Первомайськ - Миколаївська область - [Видача дозволу на розміщення зовнішньої реклами.pdf](https://github.com/e-government-ua/iBP/files/262917/default.pdf) **Координатор города:** Корой Виталий - 0991962269 - somati.orlik@gmail.com И еще просьба, в получатели добавить еще два контакта, для более быстрого реагирования ) kukharenko.vsevolod@gmail.com - Всеволод Кухаренко sergey.donchenko@gmail.com - Сергей Донченко **Контактное лицо ЦНАП:** Наталія Миколаївна Петрущак, natashadc@mail.ru, (099) 277-41-42 на тест отправляем в ЦНАП, но обязательно ставим в копию 3 координаторов выше.
process
видача дозволу на розміщення зовнішньої реклами м первомайськ миколаївська область координатор города корой виталий somati orlik gmail com и еще просьба в получатели добавить еще два контакта для более быстрого реагирования kukharenko vsevolod gmail com всеволод кухаренко sergey donchenko gmail com сергей донченко контактное лицо цнап наталія миколаївна петрущак natashadc mail ru на тест отправляем в цнап но обязательно ставим в копию координаторов выше
1
247,247
7,915,697,252
IssuesEvent
2018-07-04 01:03:10
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
closed
USER ISSUE: cant start any save data
High Priority
**Version:** 0.7.5.1 beta **Steps to Reproduce:** i try to load any game and nothing happens **Expected behavior:** play **Actual behavior:** nothing
1.0
USER ISSUE: cant start any save data - **Version:** 0.7.5.1 beta **Steps to Reproduce:** i try to load any game and nothing happens **Expected behavior:** play **Actual behavior:** nothing
non_process
user issue cant start any save data version beta steps to reproduce i try to load any game and nothing happens expected behavior play actual behavior nothing
0
3,676
6,711,771,538
IssuesEvent
2017-10-13 06:21:54
zeebe-io/zeebe
https://api.github.com/repos/zeebe-io/zeebe
closed
Test broker restart without index savepoints
broker ci stream processor
Write tests where the last log entries are API requests that no follow-up events point to (e.g. task locked events that are not completed yet) and where these events have not been indexed yet, such that the recovery logic processes them again. Fix problems if any. Should be both, integration and unit tested.
1.0
Test broker restart without index savepoints - Write tests where the last log entries are API requests that no follow-up events point to (e.g. task locked events that are not completed yet) and where these events have not been indexed yet, such that the recovery logic processes them again. Fix problems if any. Should be both, integration and unit tested.
process
test broker restart without index savepoints write tests where the last log entries are api requests that no follow up events point to e g task locked events that are not completed yet and where these events have not been indexed yet such that the recovery logic processes them again fix problems if any should be both integration and unit tested
1
87,686
17,360,181,345
IssuesEvent
2021-07-29 19:25:05
EDCD/EDMarketConnector
https://api.github.com/repos/EDCD/EDMarketConnector
closed
Treat Journal event names case-insensitively
code cleanup enhancement
I know of no instance of two distinct Journal event names only differing in their case. Things like `BackPack`, but `Backpack.json` are a pain to keep straight. So we should refactor to specifically use a `lower()`'d copy of the event name in all matching code. We should *not* change the actual event name, plugins get to decide how they treat it. This entails making the `event_name = entry['name'].tolower()` and then using it instead in matches (maybe in logging too?). This issue is the latest in the series entitled "Ath is getting sick of this shit" ;) .
1.0
Treat Journal event names case-insensitively - I know of no instance of two distinct Journal event names only differing in their case. Things like `BackPack`, but `Backpack.json` are a pain to keep straight. So we should refactor to specifically use a `lower()`'d copy of the event name in all matching code. We should *not* change the actual event name, plugins get to decide how they treat it. This entails making the `event_name = entry['name'].tolower()` and then using it instead in matches (maybe in logging too?). This issue is the latest in the series entitled "Ath is getting sick of this shit" ;) .
non_process
treat journal event names case insensitively i know of no instance of two distinct journal event names only differing in their case things like backpack but backpack json are a pain to keep straight so we should refactor to specifically use a lower d copy of the event name in all matching code we should not change the actual event name plugins get to decide how they treat it this entails making the event name entry tolower and then using it instead in matches maybe in logging too this issue is the latest in the series entitled ath is getting sick of this shit
0
436,726
12,552,221,505
IssuesEvent
2020-06-06 17:21:57
Eastrall/Rhisis
https://api.github.com/repos/Eastrall/Rhisis
closed
Disconnecting while dead makes you a zombie
bug priority: critical srv: world sys: battle v0.4.x
# :beetle: Bug Report <!-- Enter rhisis version here with format "vX.Y.Z" --> **Rhisis version:** v0.4.x ## Expected Behavior <!--- Tell us what should happen --> When disconnecting when dead and relogging afterwards, the player should revive at the nearest resurection point. ## Current Behavior <!--- Tell us what happens instead of the expected behavior --> The player is standing at the place he died and has 0 HP. ## Steps to Reproduce <!--- Provide a link to a live example, or an unambiguous set of steps to --> <!--- reproduce this bug. Include code to reproduce, if relevant --> 1. Die in battle versus a monster 2. Disconnect 3. Reconnect
1.0
Disconnecting while dead makes you a zombie - # :beetle: Bug Report <!-- Enter rhisis version here with format "vX.Y.Z" --> **Rhisis version:** v0.4.x ## Expected Behavior <!--- Tell us what should happen --> When disconnecting when dead and relogging afterwards, the player should revive at the nearest resurection point. ## Current Behavior <!--- Tell us what happens instead of the expected behavior --> The player is standing at the place he died and has 0 HP. ## Steps to Reproduce <!--- Provide a link to a live example, or an unambiguous set of steps to --> <!--- reproduce this bug. Include code to reproduce, if relevant --> 1. Die in battle versus a monster 2. Disconnect 3. Reconnect
non_process
disconnecting while dead makes you a zombie beetle bug report rhisis version x expected behavior when disconnecting when dead and relogging afterwards the player should revive at the nearest resurection point current behavior the player is standing at the place he died and has hp steps to reproduce die in battle versus a monster disconnect reconnect
0
25,597
4,163,400,411
IssuesEvent
2016-06-18 03:02:21
alexrj/Slic3r
https://api.github.com/repos/alexrj/Slic3r
closed
stable giving segfault error
Needs testing with current dev version or next release
on mint 16, 64, running stable 64. Unknown option interface_shells The program 'slic3r' received an X Window System error. This probably reflects a bug in the program. The error was 'RenderBadFormat (invalid PictFormat parameter)'. (Details: serial 50139 error_code 142 request_code 139 minor_code 2) (Note to programmers: normally, X errors are reported asynchronously; that is, you will receive the error a while after causing it. To debug your program, run it with the --sync command line option to change this behavior. You can then get a meaningful backtrace from your debugger if you break on the gdk_x_error() function.) The program 'slic3r' received an X Window System error. This probably reflects a bug in the program. The error was 'BadRequest (invalid request code or no such operation)'. (Details: serial 50158 error_code 1 request_code 255 minor_code 0) (Note to programmers: normally, X errors are reported asynchronously; that is, you will receive the error a while after causing it. To debug your program, run it with the --sync command line option to change this behavior. You can then get a meaningful backtrace from your debugger if you break on the gdk_x_error() function.) Segmentation fault This happened when exporting to g-code, although no gcode or gcode file was generated. config.ini is # generated by Slic3r 1.1.0 on Thu Apr 3 22:30:38 2014 avoid_crossing_perimeters = 1 bed_size = 284.2,301.88 bed_temperature = 75 bottom_solid_layers = 2 bridge_acceleration = 0 bridge_fan_speed = 100 bridge_flow_ratio = 1 bridge_speed = 40 brim_width = 0 complete_objects = 0 cooling = 1 default_acceleration = 0 disable_fan_first_layers = 1 duplicate_distance = 6 end_gcode = G92 E0; reset extruder\nG1 E-1 F1000; retract extruder to prevent final string\nG1 E-2 F1000; retract extruder to prevent final string\nG1 E-3 F1000; retract extruder to prevent final string\nG1 E-4 F1000; retract extruder to prevent final string\nG1 E-5 F1000; retract extruder to prevent final string\nT0;\nM104 S0 ; turn off temperature\nT1;\nM104 S0 ; turn off temperature\nT2;\nM104 S0 ; turn off temperature\nT3;\nM104 S0 ; turn off temperature\nT4;\nM104 S0 ; turn off temperature\nM140 S0 ; Turn off heated print bed\nG28 X0 ; home X axis\nM84 ; disable motors external_perimeter_speed = 40 external_perimeters_first = 0 extra_perimeters = 1 extruder_clearance_height = 20 extruder_clearance_radius = 20 extruder_offset = 10x10,72.95x12.05,9.35x37.7,73.18x38.65,41.6x38.25 extrusion_axis = E extrusion_multiplier = 1 extrusion_width = 0 fan_always_on = 0 fan_below_layer_time = 60 filament_diameter = 1.8 fill_angle = 45 fill_density = 30% fill_pattern = rectilinear first_layer_acceleration = 0 first_layer_bed_temperature = 75 first_layer_extrusion_width = 175% first_layer_height = 0.3 first_layer_speed = 20 first_layer_temperature = 215 g0 = 0 gap_fill_speed = 40 gcode_arcs = 0 gcode_comments = 0 gcode_flavor = reprap infill_acceleration = 0 infill_every_layers = 1 infill_extruder = 1 infill_extrusion_width = 0 infill_first = 1 infill_only_where_needed = 0 infill_speed = 40 interface_shells = 0 layer_gcode = layer_height = 0.3 max_fan_speed = 100 min_fan_speed = 35 min_print_speed = 5 min_skirt_length = 0 notes = nozzle_diameter = 0.5,0.5,0.5,0.5,0.5 only_retract_when_crossing_perimeters = 1 ooze_prevention = 1 output_filename_format = [input_filename_base].gcode overhangs = 1 perimeter_acceleration = 0 perimeter_extruder = 1 perimeter_extrusion_width = 0 perimeter_speed = 40 perimeters = 3 post_process = print_center = 142.1,150.94 raft_layers = 0 randomize_start = 0 resolution = 0 retract_before_travel = 20,20,20,20,20 retract_layer_change = 0,0,0,0,0 retract_length = 26,26,26,26,21 retract_length_toolchange = 26,26,26,26,21 retract_lift = 0,0,0,0,0 retract_restart_extra = 0,0,0,0,0 retract_restart_extra_toolchange = 0.5,1,0,0,0 retract_speed = 5,5,3,5,3 skirt_distance = 10 skirt_height = 1 skirts = 2 slowdown_below_layer_time = 30 small_perimeter_speed = 40 solid_fill_pattern = rectilinear solid_infill_below_area = 1 solid_infill_every_layers = 0 solid_infill_extrusion_width = 0 solid_infill_speed = 40 spiral_vase = 0 standby_temperature_delta = -1 start_gcode = M106 S255; turn fans on\nG28 ; home all axes\nT0; //set the 1st extruder to be active start_perimeters_at_concave_points = 0 start_perimeters_at_non_overhang = 0 support_material = 0 support_material_angle = 0 support_material_enforce_layers = 0 support_material_extruder = 1 support_material_extrusion_width = 0 support_material_interface_extruder = 1 support_material_interface_layers = 1 support_material_interface_spacing = 2.5 support_material_pattern = rectilinear support_material_spacing = 4 support_material_speed = 40 support_material_threshold = 0 temperature = 215 thin_walls = 1 threads = 4 toolchange_gcode = top_infill_extrusion_width = 0 top_solid_infill_speed = 40 top_solid_layers = 3 travel_speed = 12 use_firmware_retraction = 0 use_relative_e_distances = 0 vibration_limit = 0 wipe = 1,1,1,1,1 z_offset = 0 Making this test object http://www.thingiverse.com/thing:38096
1.0
stable giving segfault error - on mint 16, 64, running stable 64. Unknown option interface_shells The program 'slic3r' received an X Window System error. This probably reflects a bug in the program. The error was 'RenderBadFormat (invalid PictFormat parameter)'. (Details: serial 50139 error_code 142 request_code 139 minor_code 2) (Note to programmers: normally, X errors are reported asynchronously; that is, you will receive the error a while after causing it. To debug your program, run it with the --sync command line option to change this behavior. You can then get a meaningful backtrace from your debugger if you break on the gdk_x_error() function.) The program 'slic3r' received an X Window System error. This probably reflects a bug in the program. The error was 'BadRequest (invalid request code or no such operation)'. (Details: serial 50158 error_code 1 request_code 255 minor_code 0) (Note to programmers: normally, X errors are reported asynchronously; that is, you will receive the error a while after causing it. To debug your program, run it with the --sync command line option to change this behavior. You can then get a meaningful backtrace from your debugger if you break on the gdk_x_error() function.) Segmentation fault This happened when exporting to g-code, although no gcode or gcode file was generated. config.ini is # generated by Slic3r 1.1.0 on Thu Apr 3 22:30:38 2014 avoid_crossing_perimeters = 1 bed_size = 284.2,301.88 bed_temperature = 75 bottom_solid_layers = 2 bridge_acceleration = 0 bridge_fan_speed = 100 bridge_flow_ratio = 1 bridge_speed = 40 brim_width = 0 complete_objects = 0 cooling = 1 default_acceleration = 0 disable_fan_first_layers = 1 duplicate_distance = 6 end_gcode = G92 E0; reset extruder\nG1 E-1 F1000; retract extruder to prevent final string\nG1 E-2 F1000; retract extruder to prevent final string\nG1 E-3 F1000; retract extruder to prevent final string\nG1 E-4 F1000; retract extruder to prevent final string\nG1 E-5 F1000; retract extruder to prevent final string\nT0;\nM104 S0 ; turn off temperature\nT1;\nM104 S0 ; turn off temperature\nT2;\nM104 S0 ; turn off temperature\nT3;\nM104 S0 ; turn off temperature\nT4;\nM104 S0 ; turn off temperature\nM140 S0 ; Turn off heated print bed\nG28 X0 ; home X axis\nM84 ; disable motors external_perimeter_speed = 40 external_perimeters_first = 0 extra_perimeters = 1 extruder_clearance_height = 20 extruder_clearance_radius = 20 extruder_offset = 10x10,72.95x12.05,9.35x37.7,73.18x38.65,41.6x38.25 extrusion_axis = E extrusion_multiplier = 1 extrusion_width = 0 fan_always_on = 0 fan_below_layer_time = 60 filament_diameter = 1.8 fill_angle = 45 fill_density = 30% fill_pattern = rectilinear first_layer_acceleration = 0 first_layer_bed_temperature = 75 first_layer_extrusion_width = 175% first_layer_height = 0.3 first_layer_speed = 20 first_layer_temperature = 215 g0 = 0 gap_fill_speed = 40 gcode_arcs = 0 gcode_comments = 0 gcode_flavor = reprap infill_acceleration = 0 infill_every_layers = 1 infill_extruder = 1 infill_extrusion_width = 0 infill_first = 1 infill_only_where_needed = 0 infill_speed = 40 interface_shells = 0 layer_gcode = layer_height = 0.3 max_fan_speed = 100 min_fan_speed = 35 min_print_speed = 5 min_skirt_length = 0 notes = nozzle_diameter = 0.5,0.5,0.5,0.5,0.5 only_retract_when_crossing_perimeters = 1 ooze_prevention = 1 output_filename_format = [input_filename_base].gcode overhangs = 1 perimeter_acceleration = 0 perimeter_extruder = 1 perimeter_extrusion_width = 0 perimeter_speed = 40 perimeters = 3 post_process = print_center = 142.1,150.94 raft_layers = 0 randomize_start = 0 resolution = 0 retract_before_travel = 20,20,20,20,20 retract_layer_change = 0,0,0,0,0 retract_length = 26,26,26,26,21 retract_length_toolchange = 26,26,26,26,21 retract_lift = 0,0,0,0,0 retract_restart_extra = 0,0,0,0,0 retract_restart_extra_toolchange = 0.5,1,0,0,0 retract_speed = 5,5,3,5,3 skirt_distance = 10 skirt_height = 1 skirts = 2 slowdown_below_layer_time = 30 small_perimeter_speed = 40 solid_fill_pattern = rectilinear solid_infill_below_area = 1 solid_infill_every_layers = 0 solid_infill_extrusion_width = 0 solid_infill_speed = 40 spiral_vase = 0 standby_temperature_delta = -1 start_gcode = M106 S255; turn fans on\nG28 ; home all axes\nT0; //set the 1st extruder to be active start_perimeters_at_concave_points = 0 start_perimeters_at_non_overhang = 0 support_material = 0 support_material_angle = 0 support_material_enforce_layers = 0 support_material_extruder = 1 support_material_extrusion_width = 0 support_material_interface_extruder = 1 support_material_interface_layers = 1 support_material_interface_spacing = 2.5 support_material_pattern = rectilinear support_material_spacing = 4 support_material_speed = 40 support_material_threshold = 0 temperature = 215 thin_walls = 1 threads = 4 toolchange_gcode = top_infill_extrusion_width = 0 top_solid_infill_speed = 40 top_solid_layers = 3 travel_speed = 12 use_firmware_retraction = 0 use_relative_e_distances = 0 vibration_limit = 0 wipe = 1,1,1,1,1 z_offset = 0 Making this test object http://www.thingiverse.com/thing:38096
non_process
stable giving segfault error on mint running stable unknown option interface shells the program received an x window system error this probably reflects a bug in the program the error was renderbadformat invalid pictformat parameter details serial error code request code minor code note to programmers normally x errors are reported asynchronously that is you will receive the error a while after causing it to debug your program run it with the sync command line option to change this behavior you can then get a meaningful backtrace from your debugger if you break on the gdk x error function the program received an x window system error this probably reflects a bug in the program the error was badrequest invalid request code or no such operation details serial error code request code minor code note to programmers normally x errors are reported asynchronously that is you will receive the error a while after causing it to debug your program run it with the sync command line option to change this behavior you can then get a meaningful backtrace from your debugger if you break on the gdk x error function segmentation fault this happened when exporting to g code although no gcode or gcode file was generated config ini is generated by on thu apr avoid crossing perimeters bed size bed temperature bottom solid layers bridge acceleration bridge fan speed bridge flow ratio bridge speed brim width complete objects cooling default acceleration disable fan first layers duplicate distance end gcode reset extruder e retract extruder to prevent final string e retract extruder to prevent final string e retract extruder to prevent final string e retract extruder to prevent final string e retract extruder to prevent final string turn off temperature turn off temperature turn off temperature turn off temperature turn off temperature turn off heated print bed home x axis disable motors external perimeter speed external perimeters first extra perimeters extruder clearance height extruder clearance radius extruder offset extrusion axis e extrusion multiplier extrusion width fan always on fan below layer time filament diameter fill angle fill density fill pattern rectilinear first layer acceleration first layer bed temperature first layer extrusion width first layer height first layer speed first layer temperature gap fill speed gcode arcs gcode comments gcode flavor reprap infill acceleration infill every layers infill extruder infill extrusion width infill first infill only where needed infill speed interface shells layer gcode layer height max fan speed min fan speed min print speed min skirt length notes nozzle diameter only retract when crossing perimeters ooze prevention output filename format gcode overhangs perimeter acceleration perimeter extruder perimeter extrusion width perimeter speed perimeters post process print center raft layers randomize start resolution retract before travel retract layer change retract length retract length toolchange retract lift retract restart extra retract restart extra toolchange retract speed skirt distance skirt height skirts slowdown below layer time small perimeter speed solid fill pattern rectilinear solid infill below area solid infill every layers solid infill extrusion width solid infill speed spiral vase standby temperature delta start gcode turn fans on home all axes set the extruder to be active start perimeters at concave points start perimeters at non overhang support material support material angle support material enforce layers support material extruder support material extrusion width support material interface extruder support material interface layers support material interface spacing support material pattern rectilinear support material spacing support material speed support material threshold temperature thin walls threads toolchange gcode top infill extrusion width top solid infill speed top solid layers travel speed use firmware retraction use relative e distances vibration limit wipe z offset making this test object
0
37,788
8,518,431,208
IssuesEvent
2018-11-01 11:35:29
jccastillo0007/eFacturaT
https://api.github.com/repos/jccastillo0007/eFacturaT
opened
Cancelación escritorio - una factura cancelada previamente no le cambia el status
bug defect
Cuando existe una factura que ha sido cancelada previamente, si te lo indica como respuesta. Lo que no hace es cambiar el status
1.0
Cancelación escritorio - una factura cancelada previamente no le cambia el status - Cuando existe una factura que ha sido cancelada previamente, si te lo indica como respuesta. Lo que no hace es cambiar el status
non_process
cancelación escritorio una factura cancelada previamente no le cambia el status cuando existe una factura que ha sido cancelada previamente si te lo indica como respuesta lo que no hace es cambiar el status
0
20,582
27,243,704,477
IssuesEvent
2023-02-21 23:05:56
open-telemetry/opentelemetry-collector-contrib
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
closed
k8sattributes tests failing
processor/k8sattributes needs triage
### Component(s) processor/k8sattributes ### Describe the issue you're reporting e2e tests for kubernetes are failing. See https://github.com/open-telemetry/opentelemetry-collector-contrib/actions/runs/4236650784/jobs/7361708703#step:24:16 Likely related to https://github.com/open-telemetry/opentelemetry-collector-contrib/pull/18767
1.0
k8sattributes tests failing - ### Component(s) processor/k8sattributes ### Describe the issue you're reporting e2e tests for kubernetes are failing. See https://github.com/open-telemetry/opentelemetry-collector-contrib/actions/runs/4236650784/jobs/7361708703#step:24:16 Likely related to https://github.com/open-telemetry/opentelemetry-collector-contrib/pull/18767
process
tests failing component s processor describe the issue you re reporting tests for kubernetes are failing see likely related to
1
9,059
12,133,623,008
IssuesEvent
2020-04-23 09:20:34
arcus-azure/arcus.messaging
https://api.github.com/repos/arcus-azure/arcus.messaging
closed
Provide more thorough integration scenarios for message pump
area:message-processing enhancement
Provide more thorough integration scenarios for message pump which interacts more with the messages such as abandon, complete, et al.
1.0
Provide more thorough integration scenarios for message pump - Provide more thorough integration scenarios for message pump which interacts more with the messages such as abandon, complete, et al.
process
provide more thorough integration scenarios for message pump provide more thorough integration scenarios for message pump which interacts more with the messages such as abandon complete et al
1
7,655
10,740,449,797
IssuesEvent
2019-10-29 18:12:32
allinurl/goaccess
https://api.github.com/repos/allinurl/goaccess
closed
[Help]Is there a way to only view data from a set range date?
duplicate log-processing
I want to be able to see how many "200" status codes in a month. Is there any way for me to get that kind of data? I saw there was a date parameter but it looks like it only takes one date and not a date range.
1.0
[Help]Is there a way to only view data from a set range date? - I want to be able to see how many "200" status codes in a month. Is there any way for me to get that kind of data? I saw there was a date parameter but it looks like it only takes one date and not a date range.
process
is there a way to only view data from a set range date i want to be able to see how many status codes in a month is there any way for me to get that kind of data i saw there was a date parameter but it looks like it only takes one date and not a date range
1
73,768
7,354,595,618
IssuesEvent
2018-03-09 07:42:48
AAA-Intelligence/eve
https://api.github.com/repos/AAA-Intelligence/eve
opened
[TF-ANT-2] Generated Answer is shown in the UI #2
test
Die Antwort des Bots soll auf der Benutzeroberfläche angezeigt werden.
1.0
[TF-ANT-2] Generated Answer is shown in the UI #2 - Die Antwort des Bots soll auf der Benutzeroberfläche angezeigt werden.
non_process
generated answer is shown in the ui die antwort des bots soll auf der benutzeroberfläche angezeigt werden
0
16,966
22,330,800,193
IssuesEvent
2022-06-14 14:22:53
UserOfficeProject/user-office-project-issue-tracker
https://api.github.com/repos/UserOfficeProject/user-office-project-issue-tracker
closed
Train Emma Gozzard to modify the questionnaire
type: process area: uop/stfc
Train Emma Gozzard on how to modify the questionnaire. Training for creation of it can be done at a later date. We may want to include other people in this training, but we can ask Emma who she would like included. We probably want this to be one-to-one training rather than documentation, as written documentation will likely be time consuming and not as effective.
1.0
Train Emma Gozzard to modify the questionnaire - Train Emma Gozzard on how to modify the questionnaire. Training for creation of it can be done at a later date. We may want to include other people in this training, but we can ask Emma who she would like included. We probably want this to be one-to-one training rather than documentation, as written documentation will likely be time consuming and not as effective.
process
train emma gozzard to modify the questionnaire train emma gozzard on how to modify the questionnaire training for creation of it can be done at a later date we may want to include other people in this training but we can ask emma who she would like included we probably want this to be one to one training rather than documentation as written documentation will likely be time consuming and not as effective
1
10,340
13,169,446,102
IssuesEvent
2020-08-11 13:45:53
MicrosoftDocs/azure-devops-docs
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
closed
If condition with variable missing on template conditions page
Pri1 devops-cicd-process/tech devops/prod doc-bug
I read the documentation on this page and didn't need parameters but variables. It was really confusing that there is no example on this page - so my intuition was that if checks with variables are not possible when including templates. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 6724abea-bbdc-bf66-ed5e-3214fa6c3e66 * Version Independent ID: 4f8dab21-3f0e-da32-cc0e-1d85c13c0065 * Content: [Templates - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/templates?view=azure-devops) * Content Source: [docs/pipelines/process/templates.md](https://github.com/MicrosoftDocs/vsts-docs/blob/master/docs/pipelines/process/templates.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
1.0
If condition with variable missing on template conditions page - I read the documentation on this page and didn't need parameters but variables. It was really confusing that there is no example on this page - so my intuition was that if checks with variables are not possible when including templates. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 6724abea-bbdc-bf66-ed5e-3214fa6c3e66 * Version Independent ID: 4f8dab21-3f0e-da32-cc0e-1d85c13c0065 * Content: [Templates - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/templates?view=azure-devops) * Content Source: [docs/pipelines/process/templates.md](https://github.com/MicrosoftDocs/vsts-docs/blob/master/docs/pipelines/process/templates.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
process
if condition with variable missing on template conditions page i read the documentation on this page and didn t need parameters but variables it was really confusing that there is no example on this page so my intuition was that if checks with variables are not possible when including templates document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id bbdc version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
1
338,663
30,312,864,613
IssuesEvent
2023-07-10 13:52:14
AIDASoft/DD4hep
https://api.github.com/repos/AIDASoft/DD4hep
closed
Ambiguous assignment operator in Geant4Output2EDM4hep.cpp on LCG macOS platforms
Testing Fixed Waiting for caller
Hello, Building **master** of DD4Hep we encountered a compilation error related to an ambiguous assignment operator on LCG macOS platforms. The error occurs in the following code snippet: ``` /Users/sftnight/build/workspace/lcg_nightly_pipeline/build/frameworks/DD4hep-master/src/DD4hep/master/DDG4/edm4hep/Geant4Output2EDM4hep.cpp:476:45: error: use of overloaded operator '=' is ambiguous (with operand types 'std::map<std::string, std::pair<edm4hep::SimCalorimeterHitCollection, edm4hep::CaloHitContributionCollection>>::mapped_type' (aka 'std::pair<edm4hep::SimCalorimeterHitCollection, edm4hep::CaloHitContributionCollection>') and 'void') auto& hits = m_calorimeterHits[colName] = {}; ~~~~~~~~~~~~~~~~~~~~~~~~~~ ^ ~~ /Library/Developer/CommandLineTools/SDKs/MacOSX13.3.sdk/usr/include/c++/v1/__utility/pair.h:262:11: note: candidate function pair& operator=(typename conditional< ^ /Library/Developer/CommandLineTools/SDKs/MacOSX13.3.sdk/usr/include/c++/v1/__utility/pair.h:275:11: note: candidate function pair& operator=(typename conditional< ^ /Library/Developer/CommandLineTools/SDKs/MacOSX13.3.sdk/usr/include/c++/v1/__utility/pair.h:40:29: note: candidate function (the implicit copy assignment operator) has been implicitly deleted struct _LIBCPP_TEMPLATE_VIS pair ^ ``` https://github.com/AIDASoft/DD4hep/blob/3347996361696054fc32b1515bc277be731c724f/DDG4/edm4hep/Geant4Output2EDM4hep.cpp#L476
1.0
Ambiguous assignment operator in Geant4Output2EDM4hep.cpp on LCG macOS platforms - Hello, Building **master** of DD4Hep we encountered a compilation error related to an ambiguous assignment operator on LCG macOS platforms. The error occurs in the following code snippet: ``` /Users/sftnight/build/workspace/lcg_nightly_pipeline/build/frameworks/DD4hep-master/src/DD4hep/master/DDG4/edm4hep/Geant4Output2EDM4hep.cpp:476:45: error: use of overloaded operator '=' is ambiguous (with operand types 'std::map<std::string, std::pair<edm4hep::SimCalorimeterHitCollection, edm4hep::CaloHitContributionCollection>>::mapped_type' (aka 'std::pair<edm4hep::SimCalorimeterHitCollection, edm4hep::CaloHitContributionCollection>') and 'void') auto& hits = m_calorimeterHits[colName] = {}; ~~~~~~~~~~~~~~~~~~~~~~~~~~ ^ ~~ /Library/Developer/CommandLineTools/SDKs/MacOSX13.3.sdk/usr/include/c++/v1/__utility/pair.h:262:11: note: candidate function pair& operator=(typename conditional< ^ /Library/Developer/CommandLineTools/SDKs/MacOSX13.3.sdk/usr/include/c++/v1/__utility/pair.h:275:11: note: candidate function pair& operator=(typename conditional< ^ /Library/Developer/CommandLineTools/SDKs/MacOSX13.3.sdk/usr/include/c++/v1/__utility/pair.h:40:29: note: candidate function (the implicit copy assignment operator) has been implicitly deleted struct _LIBCPP_TEMPLATE_VIS pair ^ ``` https://github.com/AIDASoft/DD4hep/blob/3347996361696054fc32b1515bc277be731c724f/DDG4/edm4hep/Geant4Output2EDM4hep.cpp#L476
non_process
ambiguous assignment operator in cpp on lcg macos platforms hello building master of we encountered a compilation error related to an ambiguous assignment operator on lcg macos platforms the error occurs in the following code snippet users sftnight build workspace lcg nightly pipeline build frameworks master src master cpp error use of overloaded operator is ambiguous with operand types std map mapped type aka std pair and void auto hits m calorimeterhits library developer commandlinetools sdks sdk usr include c utility pair h note candidate function pair operator typename conditional library developer commandlinetools sdks sdk usr include c utility pair h note candidate function pair operator typename conditional library developer commandlinetools sdks sdk usr include c utility pair h note candidate function the implicit copy assignment operator has been implicitly deleted struct libcpp template vis pair
0
155,694
12,267,364,354
IssuesEvent
2020-05-07 10:32:41
hyperledger-labs/blockchain-automation-framework
https://api.github.com/repos/hyperledger-labs/blockchain-automation-framework
closed
Add default molecule test scenario in Corda
corda molecule-test
### Description As a developer I want to run default molecule test so that I can test my ansible scripts. ### Acceptance Criteria 1.Molecule tests should cover `create/storageclass` and `create/namespace_serviceaccount` and related roles 2. Verification step should be configured. 3. Tests should complete successfully
1.0
Add default molecule test scenario in Corda - ### Description As a developer I want to run default molecule test so that I can test my ansible scripts. ### Acceptance Criteria 1.Molecule tests should cover `create/storageclass` and `create/namespace_serviceaccount` and related roles 2. Verification step should be configured. 3. Tests should complete successfully
non_process
add default molecule test scenario in corda description as a developer i want to run default molecule test so that i can test my ansible scripts acceptance criteria molecule tests should cover create storageclass and create namespace serviceaccount and related roles verification step should be configured tests should complete successfully
0
430,825
12,466,571,289
IssuesEvent
2020-05-28 15:40:32
Damgt-Studios/GolemGame
https://api.github.com/repos/Damgt-Studios/GolemGame
closed
FBX Animation Data Exporter
Priority: High Type: Tools / Support
- [ ] Make an FBX Exporter that has mesh data in the wobj animation mesh format - [ ] Make an adanim file that contains animation clip data in the format described in the following filespec ## Notes - Binormals don't need to be in the filespec, the layout will be different regardless so it can be excluded here # Wobject Filespec A 3d file format created by Joseph Whittignton to store 3d mesh, material & animation data into one compact binary file. # File composition The file is composed of a header, vertex data, index, data, animation keyframe data and then bind pose data. ``` [Header] [VertexData] [IndexData] [KeyframeData] [BindPoseData] ``` ## Header This header defines the metadata for the binary data encoded into the file ```C++ // Mesh header struct struct MeshHeader { int indexcount, vertexcount; int indexstart, vertexstart; int animation_start, animation_count; int joint_count, keyframe_count; char t_diffuse[256]; char t_normal[256]; char t_mixed[256]; }; ``` ## Mesh Data The mesh data is defined using int32s as indices and the following structure defines the data for each vertex ```C++ struct { float pos[3]; float Normal[3]; float Tangent[3]; int joints[4]; float weights[4]; float tex[2]; }; ``` The file will contain the all the vertex data in order then all the indices in order. ## Animation Data The animation data is defined using the following structures ```C++ struct joint { float transform[16]; int parent_index; }; typedef float transform[16] struct keyframe { float time; joints[j] } ``` The animation data contains only one onimation clip. The clip length is first written as a 32 bit float then for each keyframe, the keyframe time is written as a 32 bit float then an array of joint structs is written and this continues until all keyframes have been written. The bindpose data is then written as an array of 32 bit floats representing the joint transforms for the bind pose. The animation data is written as the following. n is the number of keyframes, j is the number of joints ``` [animation_clip_duration] [keyframe_array[n]] [transform[j]] ```
1.0
FBX Animation Data Exporter - - [ ] Make an FBX Exporter that has mesh data in the wobj animation mesh format - [ ] Make an adanim file that contains animation clip data in the format described in the following filespec ## Notes - Binormals don't need to be in the filespec, the layout will be different regardless so it can be excluded here # Wobject Filespec A 3d file format created by Joseph Whittignton to store 3d mesh, material & animation data into one compact binary file. # File composition The file is composed of a header, vertex data, index, data, animation keyframe data and then bind pose data. ``` [Header] [VertexData] [IndexData] [KeyframeData] [BindPoseData] ``` ## Header This header defines the metadata for the binary data encoded into the file ```C++ // Mesh header struct struct MeshHeader { int indexcount, vertexcount; int indexstart, vertexstart; int animation_start, animation_count; int joint_count, keyframe_count; char t_diffuse[256]; char t_normal[256]; char t_mixed[256]; }; ``` ## Mesh Data The mesh data is defined using int32s as indices and the following structure defines the data for each vertex ```C++ struct { float pos[3]; float Normal[3]; float Tangent[3]; int joints[4]; float weights[4]; float tex[2]; }; ``` The file will contain the all the vertex data in order then all the indices in order. ## Animation Data The animation data is defined using the following structures ```C++ struct joint { float transform[16]; int parent_index; }; typedef float transform[16] struct keyframe { float time; joints[j] } ``` The animation data contains only one onimation clip. The clip length is first written as a 32 bit float then for each keyframe, the keyframe time is written as a 32 bit float then an array of joint structs is written and this continues until all keyframes have been written. The bindpose data is then written as an array of 32 bit floats representing the joint transforms for the bind pose. The animation data is written as the following. n is the number of keyframes, j is the number of joints ``` [animation_clip_duration] [keyframe_array[n]] [transform[j]] ```
non_process
fbx animation data exporter make an fbx exporter that has mesh data in the wobj animation mesh format make an adanim file that contains animation clip data in the format described in the following filespec notes binormals don t need to be in the filespec the layout will be different regardless so it can be excluded here wobject filespec a file format created by joseph whittignton to store mesh material animation data into one compact binary file file composition the file is composed of a header vertex data index data animation keyframe data and then bind pose data header this header defines the metadata for the binary data encoded into the file c mesh header struct struct meshheader int indexcount vertexcount int indexstart vertexstart int animation start animation count int joint count keyframe count char t diffuse char t normal char t mixed mesh data the mesh data is defined using as indices and the following structure defines the data for each vertex c struct float pos float normal float tangent int joints float weights float tex the file will contain the all the vertex data in order then all the indices in order animation data the animation data is defined using the following structures c struct joint float transform int parent index typedef float transform struct keyframe float time joints the animation data contains only one onimation clip the clip length is first written as a bit float then for each keyframe the keyframe time is written as a bit float then an array of joint structs is written and this continues until all keyframes have been written the bindpose data is then written as an array of bit floats representing the joint transforms for the bind pose the animation data is written as the following n is the number of keyframes j is the number of joints
0
21,755
30,274,142,678
IssuesEvent
2023-07-07 17:57:28
core-wg/oscore-key-update
https://api.github.com/repos/core-wg/oscore-key-update
closed
Considerations on nonce lengths
KUDOS Processing
Motivate recommended nonce lengths, and tradeoffs if using a nonce length less than 8 (which is recommended). Should we allow a zero-length nonce? In either case do not disallow a small nonce size, it can be used as long as awareness exists about potential security implications.
1.0
Considerations on nonce lengths - Motivate recommended nonce lengths, and tradeoffs if using a nonce length less than 8 (which is recommended). Should we allow a zero-length nonce? In either case do not disallow a small nonce size, it can be used as long as awareness exists about potential security implications.
process
considerations on nonce lengths motivate recommended nonce lengths and tradeoffs if using a nonce length less than which is recommended should we allow a zero length nonce in either case do not disallow a small nonce size it can be used as long as awareness exists about potential security implications
1
127,024
5,011,809,654
IssuesEvent
2016-12-13 09:19:09
openvstorage/framework
https://api.github.com/repos/openvstorage/framework
closed
Remove the vpool if the volumedriver is not running
priority_normal state_verification type_enhancement
We should be able to remove the vpool if the vpool is not running. We experienced this error during a test: ``` 2016-08-24 17:29:03 43500 +0200 - stor-04.be-g8-3 - 4854/139859400984384 - celery/celery.worker.job - 930 - ERROR - Task ovs.storagerouter.remove_storagedriver[7c3022c9-0447-4a78-aa1e-029b1b6dcaab] raised unexpected: RuntimeError('Not all StorageDrivers are reachable, please (re)start them and try again',) Traceback (most recent call last): File "/usr/lib/python2.7/dist-packages/celery/app/trace.py", line 240, in trace_task R = retval = fun(*args, **kwargs) File "/usr/lib/python2.7/dist-packages/celery/app/trace.py", line 438, in __protected_call__ return self.run(*args, **kwargs) File "/opt/OpenvStorage/ovs/lib/storagerouter.py", line 1102, in remove_storagedriver raise RuntimeError('Not all StorageDrivers are reachable, please (re)start them and try again') RuntimeError: Not all StorageDrivers are reachable, please (re)start them and try again ```
1.0
Remove the vpool if the volumedriver is not running - We should be able to remove the vpool if the vpool is not running. We experienced this error during a test: ``` 2016-08-24 17:29:03 43500 +0200 - stor-04.be-g8-3 - 4854/139859400984384 - celery/celery.worker.job - 930 - ERROR - Task ovs.storagerouter.remove_storagedriver[7c3022c9-0447-4a78-aa1e-029b1b6dcaab] raised unexpected: RuntimeError('Not all StorageDrivers are reachable, please (re)start them and try again',) Traceback (most recent call last): File "/usr/lib/python2.7/dist-packages/celery/app/trace.py", line 240, in trace_task R = retval = fun(*args, **kwargs) File "/usr/lib/python2.7/dist-packages/celery/app/trace.py", line 438, in __protected_call__ return self.run(*args, **kwargs) File "/opt/OpenvStorage/ovs/lib/storagerouter.py", line 1102, in remove_storagedriver raise RuntimeError('Not all StorageDrivers are reachable, please (re)start them and try again') RuntimeError: Not all StorageDrivers are reachable, please (re)start them and try again ```
non_process
remove the vpool if the volumedriver is not running we should be able to remove the vpool if the vpool is not running we experienced this error during a test stor be celery celery worker job error task ovs storagerouter remove storagedriver raised unexpected runtimeerror not all storagedrivers are reachable please re start them and try again traceback most recent call last file usr lib dist packages celery app trace py line in trace task r retval fun args kwargs file usr lib dist packages celery app trace py line in protected call return self run args kwargs file opt openvstorage ovs lib storagerouter py line in remove storagedriver raise runtimeerror not all storagedrivers are reachable please re start them and try again runtimeerror not all storagedrivers are reachable please re start them and try again
0
5,703
8,564,049,582
IssuesEvent
2018-11-09 15:43:23
mitre/inspec_tools
https://api.github.com/repos/mitre/inspec_tools
opened
add rubocop integration or PRs
public-release release_process
add rubocop testing and rake testing as part of the PR process
1.0
add rubocop integration or PRs - add rubocop testing and rake testing as part of the PR process
process
add rubocop integration or prs add rubocop testing and rake testing as part of the pr process
1
154,881
24,362,978,003
IssuesEvent
2022-10-03 13:13:17
equinor/design-system
https://api.github.com/repos/equinor/design-system
closed
Compact density for pagination
storefront core-react design figma storybook
We need to look into how the Pagination is solved and how we can handle several digit numbers. We use buttons today. Find examples in use in apps today.
1.0
Compact density for pagination - We need to look into how the Pagination is solved and how we can handle several digit numbers. We use buttons today. Find examples in use in apps today.
non_process
compact density for pagination we need to look into how the pagination is solved and how we can handle several digit numbers we use buttons today find examples in use in apps today
0
99,795
8,711,814,330
IssuesEvent
2018-12-06 20:18:22
morissonmaciel/vscode-tsc-compiler
https://api.github.com/repos/morissonmaciel/vscode-tsc-compiler
closed
Windows only?
testing
Am I correct in that this extension is for Windows only (or not tested in any other platforms)? It can't find the global installation of Typescript on a Mac and running a local installation in node_modules fails because it uses a Windows-style backslash directory separators (`\`) instead of Unix-style forward slashes (`/`). It tries to find `/node_modules/.bin\tsc` and fails because `\t` is interpreted as a control character. If this is Windows only, it might be reasonable to mention it in the readme.
1.0
Windows only? - Am I correct in that this extension is for Windows only (or not tested in any other platforms)? It can't find the global installation of Typescript on a Mac and running a local installation in node_modules fails because it uses a Windows-style backslash directory separators (`\`) instead of Unix-style forward slashes (`/`). It tries to find `/node_modules/.bin\tsc` and fails because `\t` is interpreted as a control character. If this is Windows only, it might be reasonable to mention it in the readme.
non_process
windows only am i correct in that this extension is for windows only or not tested in any other platforms it can t find the global installation of typescript on a mac and running a local installation in node modules fails because it uses a windows style backslash directory separators instead of unix style forward slashes it tries to find node modules bin tsc and fails because t is interpreted as a control character if this is windows only it might be reasonable to mention it in the readme
0
32,297
13,784,972,458
IssuesEvent
2020-10-08 21:51:16
MicrosoftDocs/powerbi-docs
https://api.github.com/repos/MicrosoftDocs/powerbi-docs
closed
Where do you set the enhanced compute engine at dataflow level?
Pri2 doc-enhancement powerbi-service/subsvc powerbi/svc question
How do you access the screen shown by the screenshot under "Enable DirectQuery for dataflows"? I have the enhanced compute engine turned on at capacity level but cannot find this option to set the enhanced compute engine at dataflow level - it needs to be explained where this is accessed. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 369c84b0-97ef-a750-86cd-5330626c75ac * Version Independent ID: 3af79c25-1e82-ab08-7604-bfaa9ace6128 * Content: [Use the enhanced compute engine with dataflows - Power BI](https://docs.microsoft.com/en-us/power-bi/transform-model/service-dataflows-enhanced-compute-engine) * Content Source: [powerbi-docs/transform-model/service-dataflows-enhanced-compute-engine.md](https://github.com/MicrosoftDocs/powerbi-docs/blob/live/powerbi-docs/transform-model/service-dataflows-enhanced-compute-engine.md) * Service: **powerbi** * Sub-service: **powerbi-service** * GitHub Login: @davidiseminger * Microsoft Alias: **davidi**
1.0
Where do you set the enhanced compute engine at dataflow level? - How do you access the screen shown by the screenshot under "Enable DirectQuery for dataflows"? I have the enhanced compute engine turned on at capacity level but cannot find this option to set the enhanced compute engine at dataflow level - it needs to be explained where this is accessed. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 369c84b0-97ef-a750-86cd-5330626c75ac * Version Independent ID: 3af79c25-1e82-ab08-7604-bfaa9ace6128 * Content: [Use the enhanced compute engine with dataflows - Power BI](https://docs.microsoft.com/en-us/power-bi/transform-model/service-dataflows-enhanced-compute-engine) * Content Source: [powerbi-docs/transform-model/service-dataflows-enhanced-compute-engine.md](https://github.com/MicrosoftDocs/powerbi-docs/blob/live/powerbi-docs/transform-model/service-dataflows-enhanced-compute-engine.md) * Service: **powerbi** * Sub-service: **powerbi-service** * GitHub Login: @davidiseminger * Microsoft Alias: **davidi**
non_process
where do you set the enhanced compute engine at dataflow level how do you access the screen shown by the screenshot under enable directquery for dataflows i have the enhanced compute engine turned on at capacity level but cannot find this option to set the enhanced compute engine at dataflow level it needs to be explained where this is accessed document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service powerbi sub service powerbi service github login davidiseminger microsoft alias davidi
0
185,347
6,721,137,993
IssuesEvent
2017-10-16 10:28:03
CS2103AUG2017-T17-B1/main
https://api.github.com/repos/CS2103AUG2017-T17-B1/main
closed
As a clumsy user I want to have automatic backups
priority.medium
so that I can not worry about accidentally deleting things
1.0
As a clumsy user I want to have automatic backups - so that I can not worry about accidentally deleting things
non_process
as a clumsy user i want to have automatic backups so that i can not worry about accidentally deleting things
0
20,902
27,741,879,672
IssuesEvent
2023-03-15 14:46:23
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
GitHub codeowners group not working
P2 type: process team-OSS
It looks like the rules apple team who is supposed to be an owner of some of the objc logic is not working correctly https://github.com/bazelbuild/bazel/pull/14844#issuecomment-1061239960 You can see github reporting errors on https://github.com/bazelbuild/bazel/pull/14614/files I believe it's because the team doesn't include the bazel repo https://github.com/orgs/bazelbuild/teams/rules_apple-team/repositories, and I wonder what level of access the team would have to have for this to work, if just "read" would be enough or not. @aiuto can you debug? I assume you need admin access to fix anything here
1.0
GitHub codeowners group not working - It looks like the rules apple team who is supposed to be an owner of some of the objc logic is not working correctly https://github.com/bazelbuild/bazel/pull/14844#issuecomment-1061239960 You can see github reporting errors on https://github.com/bazelbuild/bazel/pull/14614/files I believe it's because the team doesn't include the bazel repo https://github.com/orgs/bazelbuild/teams/rules_apple-team/repositories, and I wonder what level of access the team would have to have for this to work, if just "read" would be enough or not. @aiuto can you debug? I assume you need admin access to fix anything here
process
github codeowners group not working it looks like the rules apple team who is supposed to be an owner of some of the objc logic is not working correctly you can see github reporting errors on i believe it s because the team doesn t include the bazel repo and i wonder what level of access the team would have to have for this to work if just read would be enough or not aiuto can you debug i assume you need admin access to fix anything here
1
1,890
4,716,816,061
IssuesEvent
2016-10-16 08:41:39
symfony/symfony
https://api.github.com/repos/symfony/symfony
closed
Sub process output not being returned 'real-time'
Bug Process Status: Needs Review Unconfirmed
I'm just testing out the Process class and it doesn't seem to do what it says on the tin. Docs state > When executing a long running command (like rsync-ing files to a remote server), you can give feedback to the end user in real-time by passing an anonymous function to the run() method: I have 1 file with the following ```php $process = new Process('php terminal.php'); $process->setTimeout(null); $process->run( function ( $type, $buffer ) { if (Process::ERR === $type) { echo 'ERR > '.$buffer; } else { echo 'OUT > '.$buffer; } }); ``` and another with the following ```php $x = 10; while( $x ) { echo "{$x}\n"; sleep(1); $x--; } ``` While running this via terminal after a **10 second wait** the output is ```cli OUT > 10 OUT > 9 OUT > 8 OUT > 7 OUT > 6 OUT > 5 OUT > 4 OUT > 3 OUT > 2 OUT > 1 ``` which shows there are 10 iterations of closure above but there is no output until the process is finished. Am I missing something here? Regards, Luke
1.0
Sub process output not being returned 'real-time' - I'm just testing out the Process class and it doesn't seem to do what it says on the tin. Docs state > When executing a long running command (like rsync-ing files to a remote server), you can give feedback to the end user in real-time by passing an anonymous function to the run() method: I have 1 file with the following ```php $process = new Process('php terminal.php'); $process->setTimeout(null); $process->run( function ( $type, $buffer ) { if (Process::ERR === $type) { echo 'ERR > '.$buffer; } else { echo 'OUT > '.$buffer; } }); ``` and another with the following ```php $x = 10; while( $x ) { echo "{$x}\n"; sleep(1); $x--; } ``` While running this via terminal after a **10 second wait** the output is ```cli OUT > 10 OUT > 9 OUT > 8 OUT > 7 OUT > 6 OUT > 5 OUT > 4 OUT > 3 OUT > 2 OUT > 1 ``` which shows there are 10 iterations of closure above but there is no output until the process is finished. Am I missing something here? Regards, Luke
process
sub process output not being returned real time i m just testing out the process class and it doesn t seem to do what it says on the tin docs state when executing a long running command like rsync ing files to a remote server you can give feedback to the end user in real time by passing an anonymous function to the run method i have file with the following php process new process php terminal php process settimeout null process run function type buffer if process err type echo err buffer else echo out buffer and another with the following php x while x echo x n sleep x while running this via terminal after a second wait the output is cli out out out out out out out out out out which shows there are iterations of closure above but there is no output until the process is finished am i missing something here regards luke
1
6,781
9,916,317,701
IssuesEvent
2019-06-28 19:33:02
HumanCellAtlas/dcp-community
https://api.github.com/repos/HumanCellAtlas/dcp-community
closed
The template for renaming approved RFCs is incorrect.
rfc-process
Reported by @mweiden #### Author(s): - Rename the RFC from `0000-my-feature.md` to `rfc####-my-feature.md` (with leading zeros) where `####` is the next available RFC number should be: - Rename the RFC from `0000-my-feature.md` to `####-my-feature.md` (with leading zeros) where `####` is the next available RFC number
1.0
The template for renaming approved RFCs is incorrect. - Reported by @mweiden #### Author(s): - Rename the RFC from `0000-my-feature.md` to `rfc####-my-feature.md` (with leading zeros) where `####` is the next available RFC number should be: - Rename the RFC from `0000-my-feature.md` to `####-my-feature.md` (with leading zeros) where `####` is the next available RFC number
process
the template for renaming approved rfcs is incorrect reported by mweiden author s rename the rfc from my feature md to rfc my feature md with leading zeros where is the next available rfc number should be rename the rfc from my feature md to my feature md with leading zeros where is the next available rfc number
1
386,605
26,692,556,612
IssuesEvent
2023-01-27 07:10:45
OSCA-Kampala-Chapter/Windows-Musawo
https://api.github.com/repos/OSCA-Kampala-Chapter/Windows-Musawo
opened
Document Font styles used for the Design Assets
documentation
Document the Font styles used for all the Design Assets as we intend to have a common style used through out all our future designs. It'd be great to have all the design decisions well documented for future designers to follow. - [ ] Add the name of the font style/typeface used in the [design assets README file]() - [ ] Similarly, add the color codes used in the designs.
1.0
Document Font styles used for the Design Assets - Document the Font styles used for all the Design Assets as we intend to have a common style used through out all our future designs. It'd be great to have all the design decisions well documented for future designers to follow. - [ ] Add the name of the font style/typeface used in the [design assets README file]() - [ ] Similarly, add the color codes used in the designs.
non_process
document font styles used for the design assets document the font styles used for all the design assets as we intend to have a common style used through out all our future designs it d be great to have all the design decisions well documented for future designers to follow add the name of the font style typeface used in the similarly add the color codes used in the designs
0
1,621
4,236,291,028
IssuesEvent
2016-07-05 17:56:17
AnalyticalGraphicsInc/cesium
https://api.github.com/repos/AnalyticalGraphicsInc/cesium
closed
Cesium does not build with npm 1.3.10
dev process doc
There is no compilation error, but the build only takes half a minute and Cesium.js is missing from Build/Cesium. The [build guide](https://github.com/AnalyticalGraphicsInc/cesium/blob/master/Documentation/Contributors/BuildGuide/README.md) does not mention a minimal version of npm. With npm 1.3.10 ``` $ npm run minifyRelease > cesium@1.23.0 minifyRelease /home/gberaudo/cesium > gulp minifyRelease [08:31:22] Using gulpfile ~/cesium/gulpfile.js [08:31:22] Starting 'build'... [08:31:23] Finished 'build' after 1.25 s [08:31:23] Starting 'generateStubs'... [08:31:24] Finished 'generateStubs' after 268 ms [08:31:24] Starting 'minifyRelease'... Building Cesium.js Building Source/Workers/cesiumWorkerBootstrapper.js Finished Cesium.js Finished Source/Workers/cesiumWorkerBootstrapper.js Building Source/ThirdParty/Workers/inflate.js Finished Source/ThirdParty/Workers/inflate.js ... [08:32:00] Finished 'minifyRelease' after 36 s ``` Then $ ls Build/Cesium/ Assets Widgets With npm 2.15.5 ``` $ npm run minifyRelease > cesium@1.23.0 minifyRelease /tmp/cesium > gulp minifyRelease [08:43:33] Using gulpfile /tmp/cesium/gulpfile.js [08:43:33] Starting 'build'... [08:43:33] Finished 'build' after 339 ms [08:43:33] Starting 'generateStubs'... [08:43:33] Finished 'generateStubs' after 46 ms [08:43:33] Starting 'minifyRelease'... Building Cesium.js Building Source/Workers/cesiumWorkerBootstrapper.js Building Source/Workers/transferTypedArrayTest.js Building Source/ThirdParty/Workers/deflate.js Building Source/ThirdParty/Workers/inflate.js Finished Source/Workers/transferTypedArrayTest.js Finished Source/Workers/cesiumWorkerBootstrapper.js Finished Source/ThirdParty/Workers/deflate.js Finished Source/ThirdParty/Workers/inflate.js Building Source/Workers/combineGeometry.js Building Source/Workers/createBoxGeometry.js Building Source/Workers/createBoxOutlineGeometry.js Building Source/Workers/createCircleGeometry.js ... ```
1.0
Cesium does not build with npm 1.3.10 - There is no compilation error, but the build only takes half a minute and Cesium.js is missing from Build/Cesium. The [build guide](https://github.com/AnalyticalGraphicsInc/cesium/blob/master/Documentation/Contributors/BuildGuide/README.md) does not mention a minimal version of npm. With npm 1.3.10 ``` $ npm run minifyRelease > cesium@1.23.0 minifyRelease /home/gberaudo/cesium > gulp minifyRelease [08:31:22] Using gulpfile ~/cesium/gulpfile.js [08:31:22] Starting 'build'... [08:31:23] Finished 'build' after 1.25 s [08:31:23] Starting 'generateStubs'... [08:31:24] Finished 'generateStubs' after 268 ms [08:31:24] Starting 'minifyRelease'... Building Cesium.js Building Source/Workers/cesiumWorkerBootstrapper.js Finished Cesium.js Finished Source/Workers/cesiumWorkerBootstrapper.js Building Source/ThirdParty/Workers/inflate.js Finished Source/ThirdParty/Workers/inflate.js ... [08:32:00] Finished 'minifyRelease' after 36 s ``` Then $ ls Build/Cesium/ Assets Widgets With npm 2.15.5 ``` $ npm run minifyRelease > cesium@1.23.0 minifyRelease /tmp/cesium > gulp minifyRelease [08:43:33] Using gulpfile /tmp/cesium/gulpfile.js [08:43:33] Starting 'build'... [08:43:33] Finished 'build' after 339 ms [08:43:33] Starting 'generateStubs'... [08:43:33] Finished 'generateStubs' after 46 ms [08:43:33] Starting 'minifyRelease'... Building Cesium.js Building Source/Workers/cesiumWorkerBootstrapper.js Building Source/Workers/transferTypedArrayTest.js Building Source/ThirdParty/Workers/deflate.js Building Source/ThirdParty/Workers/inflate.js Finished Source/Workers/transferTypedArrayTest.js Finished Source/Workers/cesiumWorkerBootstrapper.js Finished Source/ThirdParty/Workers/deflate.js Finished Source/ThirdParty/Workers/inflate.js Building Source/Workers/combineGeometry.js Building Source/Workers/createBoxGeometry.js Building Source/Workers/createBoxOutlineGeometry.js Building Source/Workers/createCircleGeometry.js ... ```
process
cesium does not build with npm there is no compilation error but the build only takes half a minute and cesium js is missing from build cesium the does not mention a minimal version of npm with npm npm run minifyrelease cesium minifyrelease home gberaudo cesium gulp minifyrelease using gulpfile cesium gulpfile js starting build finished build after s starting generatestubs finished generatestubs after ms starting minifyrelease building cesium js building source workers cesiumworkerbootstrapper js finished cesium js finished source workers cesiumworkerbootstrapper js building source thirdparty workers inflate js finished source thirdparty workers inflate js finished minifyrelease after s then ls build cesium assets widgets with npm npm run minifyrelease cesium minifyrelease tmp cesium gulp minifyrelease using gulpfile tmp cesium gulpfile js starting build finished build after ms starting generatestubs finished generatestubs after ms starting minifyrelease building cesium js building source workers cesiumworkerbootstrapper js building source workers transfertypedarraytest js building source thirdparty workers deflate js building source thirdparty workers inflate js finished source workers transfertypedarraytest js finished source workers cesiumworkerbootstrapper js finished source thirdparty workers deflate js finished source thirdparty workers inflate js building source workers combinegeometry js building source workers createboxgeometry js building source workers createboxoutlinegeometry js building source workers createcirclegeometry js
1
99,101
20,883,167,587
IssuesEvent
2022-03-23 00:05:31
alefragnani/vscode-separators
https://api.github.com/repos/alefragnani/vscode-separators
closed
[FEATURE] - Support Settings Editor Extension Categories API
enhancement vscode api adoption
<!-- Please search existing issues to avoid creating duplicates. --> <!-- Describe the feature you'd like. --> https://code.visualstudio.com/updates/v1_61#_settings-editor-extension-categories Possible Categories: * Each kind of symbol
1.0
[FEATURE] - Support Settings Editor Extension Categories API - <!-- Please search existing issues to avoid creating duplicates. --> <!-- Describe the feature you'd like. --> https://code.visualstudio.com/updates/v1_61#_settings-editor-extension-categories Possible Categories: * Each kind of symbol
non_process
support settings editor extension categories api possible categories each kind of symbol
0
21,979
30,470,942,946
IssuesEvent
2023-07-17 13:36:33
USGS-WiM/StreamStats
https://api.github.com/repos/USGS-WiM/StreamStats
closed
Select server for Batch Processor Manage Queue tab
Batch Processor
Create a way for the user to specify which Batch Processor queue they would like to manage: At least 2 options for this: 1. On the Manage Queue tab, add a drop down with options for Development Server or Production Server. When the user selects the server, the service URLs will need to be changed 2. Use separate URLs for each server: examples - https://apps-int.usgs.gov/streamstats/dev/bp/ - https://apps-int.usgs.gov/streamstats/bp/
1.0
Select server for Batch Processor Manage Queue tab - Create a way for the user to specify which Batch Processor queue they would like to manage: At least 2 options for this: 1. On the Manage Queue tab, add a drop down with options for Development Server or Production Server. When the user selects the server, the service URLs will need to be changed 2. Use separate URLs for each server: examples - https://apps-int.usgs.gov/streamstats/dev/bp/ - https://apps-int.usgs.gov/streamstats/bp/
process
select server for batch processor manage queue tab create a way for the user to specify which batch processor queue they would like to manage at least options for this on the manage queue tab add a drop down with options for development server or production server when the user selects the server the service urls will need to be changed use separate urls for each server examples
1
11,476
14,343,515,246
IssuesEvent
2020-11-28 09:50:10
emacs-ess/ESS
https://api.github.com/repos/emacs-ess/ESS
closed
ess-eval-region-or-line-and-step : Failing in multiline cases in `*.Rd` \examples{..} section when "namespace mode"
process:eval
Related to #725 . Also related to the entry in its help > Note that when inside a package and namespaced evaluation is in > place (see `ess-r-set-evaluation-env'), the evaluation of multiline input will fail." I find this very "unfriendly", and also wrong design: Inside the \example{} section I do *NOT* want namespaced evaluation. I will have a version of the package to which the help page belongs *attach*ed in my `search()` path, and want the examples to behave as for a user; i.e., things should not happen in the package namespace, but all in globalenv.
1.0
ess-eval-region-or-line-and-step : Failing in multiline cases in `*.Rd` \examples{..} section when "namespace mode" - Related to #725 . Also related to the entry in its help > Note that when inside a package and namespaced evaluation is in > place (see `ess-r-set-evaluation-env'), the evaluation of multiline input will fail." I find this very "unfriendly", and also wrong design: Inside the \example{} section I do *NOT* want namespaced evaluation. I will have a version of the package to which the help page belongs *attach*ed in my `search()` path, and want the examples to behave as for a user; i.e., things should not happen in the package namespace, but all in globalenv.
process
ess eval region or line and step failing in multiline cases in rd examples section when namespace mode related to also related to the entry in its help note that when inside a package and namespaced evaluation is in place see ess r set evaluation env the evaluation of multiline input will fail i find this very unfriendly and also wrong design inside the example section i do not want namespaced evaluation i will have a version of the package to which the help page belongs attach ed in my search path and want the examples to behave as for a user i e things should not happen in the package namespace but all in globalenv
1
20,528
27,189,080,556
IssuesEvent
2023-02-19 15:27:50
gchq/stroom
https://api.github.com/repos/gchq/stroom
closed
Allow enumeration of meta keys in XSLT
enhancement f:processing
Currently there is no way of listing all meta values within an XSLT translation. A new XSLT function should be created to allow a user to retrieve the distinct keys in a stream's meta and then get the value of each using `stroom:meta(.)`.
1.0
Allow enumeration of meta keys in XSLT - Currently there is no way of listing all meta values within an XSLT translation. A new XSLT function should be created to allow a user to retrieve the distinct keys in a stream's meta and then get the value of each using `stroom:meta(.)`.
process
allow enumeration of meta keys in xslt currently there is no way of listing all meta values within an xslt translation a new xslt function should be created to allow a user to retrieve the distinct keys in a stream s meta and then get the value of each using stroom meta
1
23,119
15,832,103,515
IssuesEvent
2021-04-06 14:17:36
ILLIXR/ILLIXR
https://api.github.com/repos/ILLIXR/ILLIXR
closed
The Runner `clean` action should remove targets
bug infrastructure
- `RM` should be defined in `ILLIXR/common/common.mk`. - `common` should be in the plugins list for `ILLIXR/configs/clean.yaml`.
1.0
The Runner `clean` action should remove targets - - `RM` should be defined in `ILLIXR/common/common.mk`. - `common` should be in the plugins list for `ILLIXR/configs/clean.yaml`.
non_process
the runner clean action should remove targets rm should be defined in illixr common common mk common should be in the plugins list for illixr configs clean yaml
0
515,864
14,970,527,075
IssuesEvent
2021-01-27 19:44:59
wso2/product-is
https://api.github.com/repos/wso2/product-is
closed
About the Upgrading guide
Affected/5.2.0 Component/Migration Priority/Normal Severity/Minor docs
1. Lost some configuration file when copy `conf` floder 2. Should add a step to change the `Version` in `repository/conf/carbon.xml`
1.0
About the Upgrading guide - 1. Lost some configuration file when copy `conf` floder 2. Should add a step to change the `Version` in `repository/conf/carbon.xml`
non_process
about the upgrading guide lost some configuration file when copy conf floder should add a step to change the version in repository conf carbon xml
0
3,012
6,016,613,404
IssuesEvent
2017-06-07 07:30:55
openvstorage/framework
https://api.github.com/repos/openvstorage/framework
closed
2 Scrub roles got assigned by using the API
process_wontfix type_bug
Version: openvstorage 2.8.2-1 2 Scrub roles got assigned by using the API.
1.0
2 Scrub roles got assigned by using the API - Version: openvstorage 2.8.2-1 2 Scrub roles got assigned by using the API.
process
scrub roles got assigned by using the api version openvstorage scrub roles got assigned by using the api
1
20,299
26,937,849,144
IssuesEvent
2023-02-07 22:22:52
python/cpython
https://api.github.com/repos/python/cpython
closed
Multiprocessing shared memory duplicated with changed size
type-bug pending expert-multiprocessing
# What I do: Inside class create shared mem and store it on the list: ``` self.inner_np_mem=[] shm = shared_memory.SharedMemory(name='zxcv', create=True, size=r_arr.nbytes) #r_arr is numpy array self.inner_np_mem.append(shm) ``` few lines down yet in the same function inside the same class I load the mem: `shm = shared_memory.SharedMemory(name='zxcv')` and compare to the stored one in the list self.inner_np_mem this is what I got: ``` print(190, 'shm.size',shm.size,len(shm.buf),shm.buf,self.inner_np_mem[0].buf) print(191, 'shm.size',shm.size,len(shm.buf),shm ,self.inner_np_mem[0] ) ``` > line 190 shm.size 81920 81920 <memory at 0x0000016EF3EA5400> <memory at 0x0000016EF3EA51C0> > line 191 shm.size 81920 81920 SharedMemory('zxcv', size=81920) SharedMemory('zxcv', size=80000) # Bug report So I have 2 shared mem objects with the same name and different size. # Your environment Python 3.9.10
1.0
Multiprocessing shared memory duplicated with changed size - # What I do: Inside class create shared mem and store it on the list: ``` self.inner_np_mem=[] shm = shared_memory.SharedMemory(name='zxcv', create=True, size=r_arr.nbytes) #r_arr is numpy array self.inner_np_mem.append(shm) ``` few lines down yet in the same function inside the same class I load the mem: `shm = shared_memory.SharedMemory(name='zxcv')` and compare to the stored one in the list self.inner_np_mem this is what I got: ``` print(190, 'shm.size',shm.size,len(shm.buf),shm.buf,self.inner_np_mem[0].buf) print(191, 'shm.size',shm.size,len(shm.buf),shm ,self.inner_np_mem[0] ) ``` > line 190 shm.size 81920 81920 <memory at 0x0000016EF3EA5400> <memory at 0x0000016EF3EA51C0> > line 191 shm.size 81920 81920 SharedMemory('zxcv', size=81920) SharedMemory('zxcv', size=80000) # Bug report So I have 2 shared mem objects with the same name and different size. # Your environment Python 3.9.10
process
multiprocessing shared memory duplicated with changed size what i do inside class create shared mem and store it on the list self inner np mem shm shared memory sharedmemory name zxcv create true size r arr nbytes r arr is numpy array self inner np mem append shm few lines down yet in the same function inside the same class i load the mem shm shared memory sharedmemory name zxcv and compare to the stored one in the list self inner np mem this is what i got print shm size shm size len shm buf shm buf self inner np mem buf print shm size shm size len shm buf shm self inner np mem line shm size line shm size sharedmemory zxcv size sharedmemory zxcv size bug report so i have shared mem objects with the same name and different size your environment python
1
166,508
6,305,823,695
IssuesEvent
2017-07-21 19:20:33
craftercms/craftercms
https://api.github.com/repos/craftercms/craftercms
opened
[deployer] Change log level to debug for cache invalidation
enhancement Priority: Low
``` [INFO] 2017-07-21 19:14:32,994 [http-nio-8080-exec-8] [load-test] [rest.SiteCacheRestController] | Content cache and Freemarker cache have been cleared for site 'load-test' ```
1.0
[deployer] Change log level to debug for cache invalidation - ``` [INFO] 2017-07-21 19:14:32,994 [http-nio-8080-exec-8] [load-test] [rest.SiteCacheRestController] | Content cache and Freemarker cache have been cleared for site 'load-test' ```
non_process
change log level to debug for cache invalidation content cache and freemarker cache have been cleared for site load test
0
251,283
27,151,071,430
IssuesEvent
2023-02-17 01:20:15
EdHomeH/fitter-happier-docker
https://api.github.com/repos/EdHomeH/fitter-happier-docker
opened
CVE-2023-25577 (High) detected in Werkzeug-1.0.1-py2.py3-none-any.whl
security vulnerability
## CVE-2023-25577 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Werkzeug-1.0.1-py2.py3-none-any.whl</b></p></summary> <p>The comprehensive WSGI web application library.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/cc/94/5f7079a0e00bd6863ef8f1da638721e9da21e5bacee597595b318f71d62e/Werkzeug-1.0.1-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/cc/94/5f7079a0e00bd6863ef8f1da638721e9da21e5bacee597595b318f71d62e/Werkzeug-1.0.1-py2.py3-none-any.whl</a></p> <p>Path to dependency file: /web/requirements.txt</p> <p>Path to vulnerable library: /web/requirements.txt</p> <p> Dependency Hierarchy: - Flask-1.1.2-py2.py3-none-any.whl (Root Library) - :x: **Werkzeug-1.0.1-py2.py3-none-any.whl** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Werkzeug is a comprehensive WSGI web application library. Prior to version 2.2.3, Werkzeug's multipart form data parser will parse an unlimited number of parts, including file parts. Parts can be a small amount of bytes, but each requires CPU time to parse and may use more memory as Python data. If a request can be made to an endpoint that accesses `request.data`, `request.form`, `request.files`, or `request.get_data(parse_form_data=False)`, it can cause unexpectedly high resource usage. This allows an attacker to cause a denial of service by sending crafted multipart data to an endpoint that will parse it. The amount of CPU time required can block worker processes from handling legitimate requests. The amount of RAM required can trigger an out of memory kill of the process. Unlimited file parts can use up memory and file handles. If many concurrent requests are sent continuously, this can exhaust or kill all available workers. Version 2.2.3 contains a patch for this issue. <p>Publish Date: 2023-02-14 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-25577>CVE-2023-25577</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2023-25577">https://www.cve.org/CVERecord?id=CVE-2023-25577</a></p> <p>Release Date: 2023-02-14</p> <p>Fix Resolution: Werkzeug - 2.2.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2023-25577 (High) detected in Werkzeug-1.0.1-py2.py3-none-any.whl - ## CVE-2023-25577 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Werkzeug-1.0.1-py2.py3-none-any.whl</b></p></summary> <p>The comprehensive WSGI web application library.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/cc/94/5f7079a0e00bd6863ef8f1da638721e9da21e5bacee597595b318f71d62e/Werkzeug-1.0.1-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/cc/94/5f7079a0e00bd6863ef8f1da638721e9da21e5bacee597595b318f71d62e/Werkzeug-1.0.1-py2.py3-none-any.whl</a></p> <p>Path to dependency file: /web/requirements.txt</p> <p>Path to vulnerable library: /web/requirements.txt</p> <p> Dependency Hierarchy: - Flask-1.1.2-py2.py3-none-any.whl (Root Library) - :x: **Werkzeug-1.0.1-py2.py3-none-any.whl** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Werkzeug is a comprehensive WSGI web application library. Prior to version 2.2.3, Werkzeug's multipart form data parser will parse an unlimited number of parts, including file parts. Parts can be a small amount of bytes, but each requires CPU time to parse and may use more memory as Python data. If a request can be made to an endpoint that accesses `request.data`, `request.form`, `request.files`, or `request.get_data(parse_form_data=False)`, it can cause unexpectedly high resource usage. This allows an attacker to cause a denial of service by sending crafted multipart data to an endpoint that will parse it. The amount of CPU time required can block worker processes from handling legitimate requests. The amount of RAM required can trigger an out of memory kill of the process. Unlimited file parts can use up memory and file handles. If many concurrent requests are sent continuously, this can exhaust or kill all available workers. Version 2.2.3 contains a patch for this issue. <p>Publish Date: 2023-02-14 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-25577>CVE-2023-25577</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2023-25577">https://www.cve.org/CVERecord?id=CVE-2023-25577</a></p> <p>Release Date: 2023-02-14</p> <p>Fix Resolution: Werkzeug - 2.2.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in werkzeug none any whl cve high severity vulnerability vulnerable library werkzeug none any whl the comprehensive wsgi web application library library home page a href path to dependency file web requirements txt path to vulnerable library web requirements txt dependency hierarchy flask none any whl root library x werkzeug none any whl vulnerable library vulnerability details werkzeug is a comprehensive wsgi web application library prior to version werkzeug s multipart form data parser will parse an unlimited number of parts including file parts parts can be a small amount of bytes but each requires cpu time to parse and may use more memory as python data if a request can be made to an endpoint that accesses request data request form request files or request get data parse form data false it can cause unexpectedly high resource usage this allows an attacker to cause a denial of service by sending crafted multipart data to an endpoint that will parse it the amount of cpu time required can block worker processes from handling legitimate requests the amount of ram required can trigger an out of memory kill of the process unlimited file parts can use up memory and file handles if many concurrent requests are sent continuously this can exhaust or kill all available workers version contains a patch for this issue publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution werkzeug step up your open source security game with mend
0
20,682
27,354,410,637
IssuesEvent
2023-02-27 11:51:04
camunda/issues
https://api.github.com/repos/camunda/issues
opened
BPMN Signal Events(3): Broadcast signal event using throw signal event
component:desktopModeler component:operate component:optimize component:webModeler component:zeebe-process-automation public kind:epic feature-parity
### Value Proposition Statement Use BPMN Throw Signal Events to easily start or continue instances that wait for a signal - without any coding. ### User Problem Users can use BPMN Catch Signal Events (e.g. Start Event or Intermediate Events), but they have to be triggered via gRPC or one of our Clients. This means using Signal Events requires writing code and also different BPMN symbols have to be used than signals for throwing signals (e.g. I cannot use Signal Throw Events and attach a job worker, but I would have to use a Service Task instead). ### User Stories I can model signal end event and intermediate signal throw events and linting works correctly. I can deploy the models with such symbols to the engine and the engine triggers all signals correctly without me having to use the API. I can see the symbols in other tools like Operate, Optimize. ### Implementation Notes In the third stage, we'll increase support to all BPMN signal symbols. Specifically, it adds support for the Signal Intermediate Throw Event, and the Signal End Event. ![image](https://user-images.githubusercontent.com/20283848/202135285-a709ae46-efee-4b90-a324-4a3d87b98e94.png) Model highlighting that all the signal events will be supported at this stage, including the signal throw events When a process instance arrives at a signal throw event, we'll broadcast a signal in the same way as the gateway can broadcast a signal: write a Signal:Broadcast command with relaying to the current partition. The implementation of stage 1 will then relay the command to other partitions, and all signal (start event) subscriptions will be triggered, without additional implementation efforts needed. ### Breakdown > This section links to various sub-issues / -tasks contributing to respective epic phase or phase results where appropriate. #### Discovery phase ## <!-- Example: link to "Conduct customer interview with xyz" --> #### Define phase ## <!-- Consider: UI, UX, technical design, documentation design --> <!-- Example: link to "Define User-Journey Flow" or "Define target architecture" --> Design Planning * Reviewed by design: {date} * Designer assigned: {Yes, No Design Necessary, or No Designer Available} * Assignee: * Design Brief - {link to design brief } * Research Brief - {link to research brief } Design Deliverables * {Deliverable Name} {Link to GH Issue} Documentation Planning <!-- Complex changes must be reviewed during the Define phase by the DRI of Documentation or technical writer. --> <!-- Briefly describe the anticipated impact to documentation. --> <!-- Example: "Creates structural changes in docs as UX is reworked." _Add docs reviewer to Epic for feedback._ --> Risk Management <!-- add link to risk management issue --> * Risk Class: <!-- e.g. very low | low | medium | high | very high --> * Risk Treatment: <!-- e.g. avoid | mitigate | transfer | accept --> #### Implement phase ## <!-- Example: link to "Implement User Story xyz". Should not only include core implementation, but also documentation. --> #### Validate phase ## <!-- Example: link to "Evaluate usage data of last quarter" --> ### Links to additional collateral <!-- Example: link to relevant support cases -->
1.0
BPMN Signal Events(3): Broadcast signal event using throw signal event - ### Value Proposition Statement Use BPMN Throw Signal Events to easily start or continue instances that wait for a signal - without any coding. ### User Problem Users can use BPMN Catch Signal Events (e.g. Start Event or Intermediate Events), but they have to be triggered via gRPC or one of our Clients. This means using Signal Events requires writing code and also different BPMN symbols have to be used than signals for throwing signals (e.g. I cannot use Signal Throw Events and attach a job worker, but I would have to use a Service Task instead). ### User Stories I can model signal end event and intermediate signal throw events and linting works correctly. I can deploy the models with such symbols to the engine and the engine triggers all signals correctly without me having to use the API. I can see the symbols in other tools like Operate, Optimize. ### Implementation Notes In the third stage, we'll increase support to all BPMN signal symbols. Specifically, it adds support for the Signal Intermediate Throw Event, and the Signal End Event. ![image](https://user-images.githubusercontent.com/20283848/202135285-a709ae46-efee-4b90-a324-4a3d87b98e94.png) Model highlighting that all the signal events will be supported at this stage, including the signal throw events When a process instance arrives at a signal throw event, we'll broadcast a signal in the same way as the gateway can broadcast a signal: write a Signal:Broadcast command with relaying to the current partition. The implementation of stage 1 will then relay the command to other partitions, and all signal (start event) subscriptions will be triggered, without additional implementation efforts needed. ### Breakdown > This section links to various sub-issues / -tasks contributing to respective epic phase or phase results where appropriate. #### Discovery phase ## <!-- Example: link to "Conduct customer interview with xyz" --> #### Define phase ## <!-- Consider: UI, UX, technical design, documentation design --> <!-- Example: link to "Define User-Journey Flow" or "Define target architecture" --> Design Planning * Reviewed by design: {date} * Designer assigned: {Yes, No Design Necessary, or No Designer Available} * Assignee: * Design Brief - {link to design brief } * Research Brief - {link to research brief } Design Deliverables * {Deliverable Name} {Link to GH Issue} Documentation Planning <!-- Complex changes must be reviewed during the Define phase by the DRI of Documentation or technical writer. --> <!-- Briefly describe the anticipated impact to documentation. --> <!-- Example: "Creates structural changes in docs as UX is reworked." _Add docs reviewer to Epic for feedback._ --> Risk Management <!-- add link to risk management issue --> * Risk Class: <!-- e.g. very low | low | medium | high | very high --> * Risk Treatment: <!-- e.g. avoid | mitigate | transfer | accept --> #### Implement phase ## <!-- Example: link to "Implement User Story xyz". Should not only include core implementation, but also documentation. --> #### Validate phase ## <!-- Example: link to "Evaluate usage data of last quarter" --> ### Links to additional collateral <!-- Example: link to relevant support cases -->
process
bpmn signal events broadcast signal event using throw signal event value proposition statement use bpmn throw signal events to easily start or continue instances that wait for a signal without any coding user problem users can use bpmn catch signal events e g start event or intermediate events but they have to be triggered via grpc or one of our clients this means using signal events requires writing code and also different bpmn symbols have to be used than signals for throwing signals e g i cannot use signal throw events and attach a job worker but i would have to use a service task instead user stories i can model signal end event and intermediate signal throw events and linting works correctly i can deploy the models with such symbols to the engine and the engine triggers all signals correctly without me having to use the api i can see the symbols in other tools like operate optimize implementation notes in the third stage we ll increase support to all bpmn signal symbols specifically it adds support for the signal intermediate throw event and the signal end event model highlighting that all the signal events will be supported at this stage including the signal throw events when a process instance arrives at a signal throw event we ll broadcast a signal in the same way as the gateway can broadcast a signal write a signal broadcast command with relaying to the current partition the implementation of stage will then relay the command to other partitions and all signal start event subscriptions will be triggered without additional implementation efforts needed breakdown this section links to various sub issues tasks contributing to respective epic phase or phase results where appropriate discovery phase define phase design planning reviewed by design date designer assigned yes no design necessary or no designer available assignee design brief link to design brief research brief link to research brief design deliverables deliverable name link to gh issue documentation planning risk management risk class risk treatment implement phase validate phase links to additional collateral
1
5,548
8,393,232,952
IssuesEvent
2018-10-09 19:56:29
hashicorp/packer
https://api.github.com/repos/hashicorp/packer
closed
amazon-import: If the import job fails Packer wait until timeout instead of failing immediately
bug post-processor/amazon-import
There are three issues: 1) If the import job fails Packer still waits for success until the timeout is reached instead of detecting the failure and exit immediately. 2) Interrupt is not handled by the post-processor. 3) We should hardcode the `Format` to `OVA`
1.0
amazon-import: If the import job fails Packer wait until timeout instead of failing immediately - There are three issues: 1) If the import job fails Packer still waits for success until the timeout is reached instead of detecting the failure and exit immediately. 2) Interrupt is not handled by the post-processor. 3) We should hardcode the `Format` to `OVA`
process
amazon import if the import job fails packer wait until timeout instead of failing immediately there are three issues if the import job fails packer still waits for success until the timeout is reached instead of detecting the failure and exit immediately interrupt is not handled by the post processor we should hardcode the format to ova
1
848
3,315,586,554
IssuesEvent
2015-11-06 13:00:26
pwittchen/prefser
https://api.github.com/repos/pwittchen/prefser
closed
Release 2.0.2
release process
**Initial release notes**: - fixed bug reported in issue #70: get(...) method now returns a null value instead of "null" string when setting default value to null of String type - fixed RxJava usage in sample app - fixed RxJava usage in code snippets in `README.md` - changed code formatting to `SquareAndroid` - added static code analysis - improved code according to static code analysis suggestions **Things to do**: - [x] bump library version: PR #75 - [x] upload archives to Maven Central - [x] close and release artifact on Maven Central - [x] update `CHANGELOG.md` after Maven Sync - [x] bump library version in `README.md` - [x] create new GitHub release
1.0
Release 2.0.2 - **Initial release notes**: - fixed bug reported in issue #70: get(...) method now returns a null value instead of "null" string when setting default value to null of String type - fixed RxJava usage in sample app - fixed RxJava usage in code snippets in `README.md` - changed code formatting to `SquareAndroid` - added static code analysis - improved code according to static code analysis suggestions **Things to do**: - [x] bump library version: PR #75 - [x] upload archives to Maven Central - [x] close and release artifact on Maven Central - [x] update `CHANGELOG.md` after Maven Sync - [x] bump library version in `README.md` - [x] create new GitHub release
process
release initial release notes fixed bug reported in issue get method now returns a null value instead of null string when setting default value to null of string type fixed rxjava usage in sample app fixed rxjava usage in code snippets in readme md changed code formatting to squareandroid added static code analysis improved code according to static code analysis suggestions things to do bump library version pr upload archives to maven central close and release artifact on maven central update changelog md after maven sync bump library version in readme md create new github release
1
16,453
2,899,800,500
IssuesEvent
2015-06-17 13:33:43
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
reopened
[TEST-FAILURE] RingbufferAddAllReadManyStressTest failures
Team: QuSP Type: Defect
com.hazelcast.ringbuffer.impl.RingbufferAddAllReadManyStressTest.whenNoTTL com.hazelcast.ringbuffer.impl.RingbufferAddReadOneStressTest.whenLongTTLAndSmallBuffer com.hazelcast.ringbuffer.impl.RingbufferAsyncAddWithBackoffStressTest.whenTTLEnabled ``` java.lang.AssertionError: No error should have been thrown, but ConsumeThread-1 completed error expected null, but was:<java.util.concurrent.ExecutionException: java.lang.IllegalArgumentException: sequence:1 is too large. The current tailSequence is:-1> ``` https://hazelcast-l337.ci.cloudbees.com/job/Hazelcast-3.x-nightly/563/#showFailuresLink
1.0
[TEST-FAILURE] RingbufferAddAllReadManyStressTest failures - com.hazelcast.ringbuffer.impl.RingbufferAddAllReadManyStressTest.whenNoTTL com.hazelcast.ringbuffer.impl.RingbufferAddReadOneStressTest.whenLongTTLAndSmallBuffer com.hazelcast.ringbuffer.impl.RingbufferAsyncAddWithBackoffStressTest.whenTTLEnabled ``` java.lang.AssertionError: No error should have been thrown, but ConsumeThread-1 completed error expected null, but was:<java.util.concurrent.ExecutionException: java.lang.IllegalArgumentException: sequence:1 is too large. The current tailSequence is:-1> ``` https://hazelcast-l337.ci.cloudbees.com/job/Hazelcast-3.x-nightly/563/#showFailuresLink
non_process
ringbufferaddallreadmanystresstest failures com hazelcast ringbuffer impl ringbufferaddallreadmanystresstest whennottl com hazelcast ringbuffer impl ringbufferaddreadonestresstest whenlongttlandsmallbuffer com hazelcast ringbuffer impl ringbufferasyncaddwithbackoffstresstest whenttlenabled java lang assertionerror no error should have been thrown but consumethread completed error expected null but was
0
769,723
27,017,805,503
IssuesEvent
2023-02-10 21:15:44
DDMAL/CantusDB
https://api.github.com/repos/DDMAL/CantusDB
closed
OldCantus Chant Detail page now has "view facsimile" link in sidebar
priority: medium
(See red rectangle, below). Points to the same URL as the Image Link (blue rectangle), at least on this page. ![Screen Shot 2023-02-10 at 4 03 04 PM](https://user-images.githubusercontent.com/58090591/218197277-7af530b4-32b4-4b90-aebc-b59258939f65.png)
1.0
OldCantus Chant Detail page now has "view facsimile" link in sidebar - (See red rectangle, below). Points to the same URL as the Image Link (blue rectangle), at least on this page. ![Screen Shot 2023-02-10 at 4 03 04 PM](https://user-images.githubusercontent.com/58090591/218197277-7af530b4-32b4-4b90-aebc-b59258939f65.png)
non_process
oldcantus chant detail page now has view facsimile link in sidebar see red rectangle below points to the same url as the image link blue rectangle at least on this page
0
48,440
20,152,517,993
IssuesEvent
2022-02-09 13:45:51
elastic/kibana
https://api.github.com/repos/elastic/kibana
closed
[Screenshotting] Limit number of concurrently running Chromium instances
Feature:Reporting loe:days Team:AppServicesSv impact:medium Team:Reporting Services
Currently, if we run the `getScreenshot` method twice, we will likely get two Chromium instances. That may cause some performance issues, especially in an environment running with low resources. The number of concurrent calls should be artificially limited from the code per process. It can be done by using Mutex (or Semaphore).
2.0
[Screenshotting] Limit number of concurrently running Chromium instances - Currently, if we run the `getScreenshot` method twice, we will likely get two Chromium instances. That may cause some performance issues, especially in an environment running with low resources. The number of concurrent calls should be artificially limited from the code per process. It can be done by using Mutex (or Semaphore).
non_process
limit number of concurrently running chromium instances currently if we run the getscreenshot method twice we will likely get two chromium instances that may cause some performance issues especially in an environment running with low resources the number of concurrent calls should be artificially limited from the code per process it can be done by using mutex or semaphore
0
440,140
12,693,890,979
IssuesEvent
2020-06-22 04:59:12
zeebe-io/zeebe
https://api.github.com/repos/zeebe-io/zeebe
opened
Do not send Swim requests to local node
Priority: High Status: Ready Type: Maintenance
**Description** Currently Swim sends probe requests and other requests to the local node, since it tries to compare the local id with the member id. The problem is that we use the bootstrap service, which assigns each node an unique id, which is not the same as our local node id.
1.0
Do not send Swim requests to local node - **Description** Currently Swim sends probe requests and other requests to the local node, since it tries to compare the local id with the member id. The problem is that we use the bootstrap service, which assigns each node an unique id, which is not the same as our local node id.
non_process
do not send swim requests to local node description currently swim sends probe requests and other requests to the local node since it tries to compare the local id with the member id the problem is that we use the bootstrap service which assigns each node an unique id which is not the same as our local node id
0
63,963
14,656,816,586
IssuesEvent
2020-12-28 14:15:48
fu1771695yongxie/odoo
https://api.github.com/repos/fu1771695yongxie/odoo
opened
CVE-2019-19911 (High) detected in Pillow-5.4.1-cp27-cp27mu-manylinux1_x86_64.whl
security vulnerability
## CVE-2019-19911 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Pillow-5.4.1-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary> <p>Python Imaging Library (Fork)</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/0d/f3/421598450cb9503f4565d936860763b5af413a61009d87a5ab1e34139672/Pillow-5.4.1-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/0d/f3/421598450cb9503f4565d936860763b5af413a61009d87a5ab1e34139672/Pillow-5.4.1-cp27-cp27mu-manylinux1_x86_64.whl</a></p> <p>Path to dependency file: odoo</p> <p>Path to vulnerable library: odoo,odoo/requirements.txt</p> <p> Dependency Hierarchy: - :x: **Pillow-5.4.1-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/fu1771695yongxie/odoo/commit/d2f4766e5250a653951e7d3f6813f6da8973da84">d2f4766e5250a653951e7d3f6813f6da8973da84</a></p> <p>Found in base branch: <b>14.0</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> There is a DoS vulnerability in Pillow before 6.2.2 caused by FpxImagePlugin.py calling the range function on an unvalidated 32-bit integer if the number of bands is large. On Windows running 32-bit Python, this results in an OverflowError or MemoryError due to the 2 GB limit. However, on Linux running 64-bit Python this results in the process being terminated by the OOM killer. <p>Publish Date: 2020-01-05 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-19911>CVE-2019-19911</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/python-pillow/Pillow/tree/6.2.2">https://github.com/python-pillow/Pillow/tree/6.2.2</a></p> <p>Release Date: 2020-01-05</p> <p>Fix Resolution: Pillow - 6.2.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-19911 (High) detected in Pillow-5.4.1-cp27-cp27mu-manylinux1_x86_64.whl - ## CVE-2019-19911 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Pillow-5.4.1-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary> <p>Python Imaging Library (Fork)</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/0d/f3/421598450cb9503f4565d936860763b5af413a61009d87a5ab1e34139672/Pillow-5.4.1-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/0d/f3/421598450cb9503f4565d936860763b5af413a61009d87a5ab1e34139672/Pillow-5.4.1-cp27-cp27mu-manylinux1_x86_64.whl</a></p> <p>Path to dependency file: odoo</p> <p>Path to vulnerable library: odoo,odoo/requirements.txt</p> <p> Dependency Hierarchy: - :x: **Pillow-5.4.1-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/fu1771695yongxie/odoo/commit/d2f4766e5250a653951e7d3f6813f6da8973da84">d2f4766e5250a653951e7d3f6813f6da8973da84</a></p> <p>Found in base branch: <b>14.0</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> There is a DoS vulnerability in Pillow before 6.2.2 caused by FpxImagePlugin.py calling the range function on an unvalidated 32-bit integer if the number of bands is large. On Windows running 32-bit Python, this results in an OverflowError or MemoryError due to the 2 GB limit. However, on Linux running 64-bit Python this results in the process being terminated by the OOM killer. <p>Publish Date: 2020-01-05 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-19911>CVE-2019-19911</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/python-pillow/Pillow/tree/6.2.2">https://github.com/python-pillow/Pillow/tree/6.2.2</a></p> <p>Release Date: 2020-01-05</p> <p>Fix Resolution: Pillow - 6.2.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in pillow whl cve high severity vulnerability vulnerable library pillow whl python imaging library fork library home page a href path to dependency file odoo path to vulnerable library odoo odoo requirements txt dependency hierarchy x pillow whl vulnerable library found in head commit a href found in base branch vulnerability details there is a dos vulnerability in pillow before caused by fpximageplugin py calling the range function on an unvalidated bit integer if the number of bands is large on windows running bit python this results in an overflowerror or memoryerror due to the gb limit however on linux running bit python this results in the process being terminated by the oom killer publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution pillow step up your open source security game with whitesource
0
531,123
15,441,423,926
IssuesEvent
2021-03-08 05:52:38
Nelson-Gon/urlfix
https://api.github.com/repos/Nelson-Gon/urlfix
opened
Some links in MarkDown files are not replaced
bug good first issue help wanted priority:high
**Describe the bug** For markdown in the format `[[]]()()`, links in the second `()` are not replaced. **To Reproduce** Run `replace_urls` on Markdown files. **Expected behavior** I expected an output file with all links replaced. **Unexpected behavior** Only links in the first `()` are matched and replaced. **System Details** 0.2.1 Dev: https://github.com/Nelson-Gon/urlfix/tree/4212bad4503cca5ae53f694c7bf623ca6d66fbf5/urlfix OS independent. --- TODO: - [x] Ensure text is kept - [ ] Match all URLs for Markdown in the form `[[]]()()` - [ ] Add tests for known string existence
1.0
Some links in MarkDown files are not replaced - **Describe the bug** For markdown in the format `[[]]()()`, links in the second `()` are not replaced. **To Reproduce** Run `replace_urls` on Markdown files. **Expected behavior** I expected an output file with all links replaced. **Unexpected behavior** Only links in the first `()` are matched and replaced. **System Details** 0.2.1 Dev: https://github.com/Nelson-Gon/urlfix/tree/4212bad4503cca5ae53f694c7bf623ca6d66fbf5/urlfix OS independent. --- TODO: - [x] Ensure text is kept - [ ] Match all URLs for Markdown in the form `[[]]()()` - [ ] Add tests for known string existence
non_process
some links in markdown files are not replaced describe the bug for markdown in the format links in the second are not replaced to reproduce run replace urls on markdown files expected behavior i expected an output file with all links replaced unexpected behavior only links in the first are matched and replaced system details dev os independent todo ensure text is kept match all urls for markdown in the form add tests for known string existence
0
7,479
10,571,644,491
IssuesEvent
2019-10-07 07:43:20
Altinn/altinn-studio
https://api.github.com/repos/Altinn/altinn-studio
closed
Analyze and define process handling and authorization
analysis process
## Description As part of #243 we have defined that process API is separate from data API. This means that moving the process forward would trigger a separate API call. This API calls need to be authorized. The big question is. What kind of authorization "action" is Next on the process. In current platform we always move the process part of the main action on the task. If you do sign on the signing step we authorize for sign and move forward in the same operation. As I see it there is two options here ## Introduce a common "action" for Next used for all steps In this scenario we can define the "Next" operation in the XACML rule file and for all process changes we authorize user for "Next" operation on any type of task ### Pro - Makes it more flexible to add new types of task - Can configure PEP to use on action ### Con - More rules in policy file - Makes delegation more complex ## A logic in Next operation that identifies the action based on type of Task we currently are in ### Pro - Less rules - Matches more current solution ### Cons - Add complexity to Next operation - Limits flexibility on process - Cant use standard PEP We need to decide ## Conclusion ## Tasks - [ ] Analyze and discuss - [ ] Decide - [ ] Update relevant issues in backlog
1.0
Analyze and define process handling and authorization - ## Description As part of #243 we have defined that process API is separate from data API. This means that moving the process forward would trigger a separate API call. This API calls need to be authorized. The big question is. What kind of authorization "action" is Next on the process. In current platform we always move the process part of the main action on the task. If you do sign on the signing step we authorize for sign and move forward in the same operation. As I see it there is two options here ## Introduce a common "action" for Next used for all steps In this scenario we can define the "Next" operation in the XACML rule file and for all process changes we authorize user for "Next" operation on any type of task ### Pro - Makes it more flexible to add new types of task - Can configure PEP to use on action ### Con - More rules in policy file - Makes delegation more complex ## A logic in Next operation that identifies the action based on type of Task we currently are in ### Pro - Less rules - Matches more current solution ### Cons - Add complexity to Next operation - Limits flexibility on process - Cant use standard PEP We need to decide ## Conclusion ## Tasks - [ ] Analyze and discuss - [ ] Decide - [ ] Update relevant issues in backlog
process
analyze and define process handling and authorization description as part of we have defined that process api is separate from data api this means that moving the process forward would trigger a separate api call this api calls need to be authorized the big question is what kind of authorization action is next on the process in current platform we always move the process part of the main action on the task if you do sign on the signing step we authorize for sign and move forward in the same operation as i see it there is two options here introduce a common action for next used for all steps in this scenario we can define the next operation in the xacml rule file and for all process changes we authorize user for next operation on any type of task pro makes it more flexible to add new types of task can configure pep to use on action con more rules in policy file makes delegation more complex a logic in next operation that identifies the action based on type of task we currently are in pro less rules matches more current solution cons add complexity to next operation limits flexibility on process cant use standard pep we need to decide conclusion tasks analyze and discuss decide update relevant issues in backlog
1
11,142
13,957,692,339
IssuesEvent
2020-10-24 08:10:40
alexanderkotsev/geoportal
https://api.github.com/repos/alexanderkotsev/geoportal
opened
PT: Harvesting
Geoportal Harvesting process PT - Portugal
Geoportal Team, Can you please start a harvesting to the Portuguese catalogue? Thank you! Vanda
1.0
PT: Harvesting - Geoportal Team, Can you please start a harvesting to the Portuguese catalogue? Thank you! Vanda
process
pt harvesting geoportal team can you please start a harvesting to the portuguese catalogue thank you vanda
1
187,601
14,428,533,842
IssuesEvent
2020-12-06 10:16:38
software-engineering-labs-group-5/image-similarity
https://api.github.com/repos/software-engineering-labs-group-5/image-similarity
opened
[Tests] Time computing
test
Check the time of modifications and metrics calculations for different sizes of images. The calculations of ones smaller than 512x512 should not take more than 1s.
1.0
[Tests] Time computing - Check the time of modifications and metrics calculations for different sizes of images. The calculations of ones smaller than 512x512 should not take more than 1s.
non_process
time computing check the time of modifications and metrics calculations for different sizes of images the calculations of ones smaller than should not take more than
0
11,177
13,957,695,276
IssuesEvent
2020-10-24 08:11:29
alexanderkotsev/geoportal
https://api.github.com/repos/alexanderkotsev/geoportal
opened
RO: New harvesting from Romanian Geoportal
Geoportal Harvesting process RO - Romania
Dear Angelo, Can you please start a new havesting from Romanian Geoportal? We did some changes and we want to see the outcome. Thank you in advance for your help. Beest Regards, Simona Bunea
1.0
RO: New harvesting from Romanian Geoportal - Dear Angelo, Can you please start a new havesting from Romanian Geoportal? We did some changes and we want to see the outcome. Thank you in advance for your help. Beest Regards, Simona Bunea
process
ro new harvesting from romanian geoportal dear angelo can you please start a new havesting from romanian geoportal we did some changes and we want to see the outcome thank you in advance for your help beest regards simona bunea
1
31,759
13,630,583,149
IssuesEvent
2020-09-24 16:39:33
microsoft/botbuilder-dotnet
https://api.github.com/repos/microsoft/botbuilder-dotnet
closed
Avoid sending typing activity when bot is invoked as skill
Area: Skills Bot Services P1 Size: S bug customer-replied-to customer-reported
## Version 4.9.4 ## Describe the bug The ShowTypingMiddleware doesn't consider if it is running in the context of a root bot or a skill. This will cause unnecessary typing events if both the root bot and the skill are sending typing events. To avoid this the ShowTypingMiddleware should check the context using the `SkillValidation.IsSkillClaim` and only send typing events when running as a root bot. I'll be submitting a PR with a fix for this issue. ## To Reproduce Steps to reproduce the behavior: 1. Create a bot and a skill 2. Configure ShowTypingMiddleware for both the bot and the skill 3. Invoke the skill through the bot 4. Bot the bot and the skill send typing events ## Expected behavior Only the root bot should send typing events. Skills should not unless they are invoked as a root bot. ## Additional context Created parent issue to track this gets ported to the other languages [here](https://github.com/microsoft/botframework-sdk/issues/6049)
1.0
Avoid sending typing activity when bot is invoked as skill - ## Version 4.9.4 ## Describe the bug The ShowTypingMiddleware doesn't consider if it is running in the context of a root bot or a skill. This will cause unnecessary typing events if both the root bot and the skill are sending typing events. To avoid this the ShowTypingMiddleware should check the context using the `SkillValidation.IsSkillClaim` and only send typing events when running as a root bot. I'll be submitting a PR with a fix for this issue. ## To Reproduce Steps to reproduce the behavior: 1. Create a bot and a skill 2. Configure ShowTypingMiddleware for both the bot and the skill 3. Invoke the skill through the bot 4. Bot the bot and the skill send typing events ## Expected behavior Only the root bot should send typing events. Skills should not unless they are invoked as a root bot. ## Additional context Created parent issue to track this gets ported to the other languages [here](https://github.com/microsoft/botframework-sdk/issues/6049)
non_process
avoid sending typing activity when bot is invoked as skill version describe the bug the showtypingmiddleware doesn t consider if it is running in the context of a root bot or a skill this will cause unnecessary typing events if both the root bot and the skill are sending typing events to avoid this the showtypingmiddleware should check the context using the skillvalidation isskillclaim and only send typing events when running as a root bot i ll be submitting a pr with a fix for this issue to reproduce steps to reproduce the behavior create a bot and a skill configure showtypingmiddleware for both the bot and the skill invoke the skill through the bot bot the bot and the skill send typing events expected behavior only the root bot should send typing events skills should not unless they are invoked as a root bot additional context created parent issue to track this gets ported to the other languages
0
13,347
15,806,330,507
IssuesEvent
2021-04-04 04:33:44
gfx-rs/naga
https://api.github.com/repos/gfx-rs/naga
opened
Incorrectly placed Return
area: processing kind: bug
From https://github.com/gfx-rs/wgpu-rs/issues/838 ```rust if (vm * vn > 0.0) { const cosThetaV = dot(n, v); const sinThetaV2 = 1.0 - cosThetaV * cosThetaV; const tanThetaV2 = sinThetaV2 / cosThetaV / cosThetaV; return 2.0 / (1.0 + sqrt(1.0 + alpha * alpha * tanThetaV2)); } else { return 0.0; } ``` This code ends up with a bare `return` at the end.
1.0
Incorrectly placed Return - From https://github.com/gfx-rs/wgpu-rs/issues/838 ```rust if (vm * vn > 0.0) { const cosThetaV = dot(n, v); const sinThetaV2 = 1.0 - cosThetaV * cosThetaV; const tanThetaV2 = sinThetaV2 / cosThetaV / cosThetaV; return 2.0 / (1.0 + sqrt(1.0 + alpha * alpha * tanThetaV2)); } else { return 0.0; } ``` This code ends up with a bare `return` at the end.
process
incorrectly placed return from rust if vm vn const costhetav dot n v const costhetav costhetav const costhetav costhetav return sqrt alpha alpha else return this code ends up with a bare return at the end
1
541
3,002,466,967
IssuesEvent
2015-07-24 17:24:20
HazyResearch/dd-genomics
https://api.github.com/repos/HazyResearch/dd-genomics
closed
Remove sentences from current sentences_input that do not contain a gene name
Data source Preprocessing
Basically for speed. This should cut our sentences_input size by a factor of 8.
1.0
Remove sentences from current sentences_input that do not contain a gene name - Basically for speed. This should cut our sentences_input size by a factor of 8.
process
remove sentences from current sentences input that do not contain a gene name basically for speed this should cut our sentences input size by a factor of
1
243,915
18,734,047,021
IssuesEvent
2021-11-04 03:32:46
ihumanable/patch
https://api.github.com/repos/ihumanable/patch
closed
Cross Reference `expose/2` and `private/1`
:blue_book: Documentation :muscle: Enhancement
Common issue when using `expose/2` is that the compiler starts producing nuisance warnings. This is because as far as the compiler is concerned you are calling a function that doesn't exists, since the exposure happens at runtime. There's already a solution in the `private/1` macro but these two functions don't formally cross-reference each other. - [x] Add cross-references so test authors using `expose/2` know about `private/1` and vice versa. - [x] The super-powers doc should also be updated to show correct usage - [x] This is a good candidate for adding to the guide book.
1.0
Cross Reference `expose/2` and `private/1` - Common issue when using `expose/2` is that the compiler starts producing nuisance warnings. This is because as far as the compiler is concerned you are calling a function that doesn't exists, since the exposure happens at runtime. There's already a solution in the `private/1` macro but these two functions don't formally cross-reference each other. - [x] Add cross-references so test authors using `expose/2` know about `private/1` and vice versa. - [x] The super-powers doc should also be updated to show correct usage - [x] This is a good candidate for adding to the guide book.
non_process
cross reference expose and private common issue when using expose is that the compiler starts producing nuisance warnings this is because as far as the compiler is concerned you are calling a function that doesn t exists since the exposure happens at runtime there s already a solution in the private macro but these two functions don t formally cross reference each other add cross references so test authors using expose know about private and vice versa the super powers doc should also be updated to show correct usage this is a good candidate for adding to the guide book
0
311,017
26,761,797,525
IssuesEvent
2023-01-31 07:37:40
softeerbootcamp/Team7-Amatta
https://api.github.com/repos/softeerbootcamp/Team7-Amatta
closed
[기능 개발] DB 연결(BE)
✨ Feature ✅ Test
#### 설명 Spring boot 프로젝트에 DB 연결 개발 브랜치 : BE-feature/DBConnection #### 🗒: 체크리스트 - [x] DB 연결 - [x] .yml or .properties 파일로 연결 관련 정보 관리 - [x] 연결 테스트 코드 작성
1.0
[기능 개발] DB 연결(BE) - #### 설명 Spring boot 프로젝트에 DB 연결 개발 브랜치 : BE-feature/DBConnection #### 🗒: 체크리스트 - [x] DB 연결 - [x] .yml or .properties 파일로 연결 관련 정보 관리 - [x] 연결 테스트 코드 작성
non_process
db 연결 be 설명 spring boot 프로젝트에 db 연결 개발 브랜치 be feature dbconnection 🗒 체크리스트 db 연결 yml or properties 파일로 연결 관련 정보 관리 연결 테스트 코드 작성
0
414,887
12,120,983,067
IssuesEvent
2020-04-22 08:34:51
microsoft/STL
https://api.github.com/repos/microsoft/STL
closed
<functional>: std::function doesn't handle over-aligned types
bug high priority work in progress
The Small Functor Optimization in `std::function` was implemented before we started to respect over-aligned types. Currently, we emit silent bad codegen when we decide to locally store an over-aligned function object. ``` C:\Temp>type meow.cpp #include <functional> #include <stdint.h> #include <stdio.h> using namespace std; struct alignas(16) Meow { int i; void operator()() { printf("This should be 0: %d\n", static_cast<int>(reinterpret_cast<uintptr_t>(this) % alignof(Meow))); } }; int main() { function<void()> f(Meow{}); int x = 1729; function<void()> g(Meow{}); f(); g(); (void) x; } C:\Temp>cl /EHsc /nologo /W4 /wd4324 meow.cpp meow.cpp C:\Temp>meow This should be 0: 0 This should be 0: 8 ``` @CaseyCarter observed: "I think all that's required here is to add `|| alignof(_Impl) > alignof(max_align_t)` to `_Is_large` and write a quick regression test." https://github.com/microsoft/STL/blob/22391019cba3863a978ea2d7d3f5ca240f38c576/stl/inc/functional#L797-L798 https://github.com/microsoft/STL/blob/22391019cba3863a978ea2d7d3f5ca240f38c576/stl/inc/functional#L1065-L1069 Tracked by Microsoft-internal VSO-157296 and VSO-1062649.
1.0
<functional>: std::function doesn't handle over-aligned types - The Small Functor Optimization in `std::function` was implemented before we started to respect over-aligned types. Currently, we emit silent bad codegen when we decide to locally store an over-aligned function object. ``` C:\Temp>type meow.cpp #include <functional> #include <stdint.h> #include <stdio.h> using namespace std; struct alignas(16) Meow { int i; void operator()() { printf("This should be 0: %d\n", static_cast<int>(reinterpret_cast<uintptr_t>(this) % alignof(Meow))); } }; int main() { function<void()> f(Meow{}); int x = 1729; function<void()> g(Meow{}); f(); g(); (void) x; } C:\Temp>cl /EHsc /nologo /W4 /wd4324 meow.cpp meow.cpp C:\Temp>meow This should be 0: 0 This should be 0: 8 ``` @CaseyCarter observed: "I think all that's required here is to add `|| alignof(_Impl) > alignof(max_align_t)` to `_Is_large` and write a quick regression test." https://github.com/microsoft/STL/blob/22391019cba3863a978ea2d7d3f5ca240f38c576/stl/inc/functional#L797-L798 https://github.com/microsoft/STL/blob/22391019cba3863a978ea2d7d3f5ca240f38c576/stl/inc/functional#L1065-L1069 Tracked by Microsoft-internal VSO-157296 and VSO-1062649.
non_process
std function doesn t handle over aligned types the small functor optimization in std function was implemented before we started to respect over aligned types currently we emit silent bad codegen when we decide to locally store an over aligned function object c temp type meow cpp include include include using namespace std struct alignas meow int i void operator printf this should be d n static cast reinterpret cast this alignof meow int main function f meow int x function g meow f g void x c temp cl ehsc nologo meow cpp meow cpp c temp meow this should be this should be caseycarter observed i think all that s required here is to add alignof impl alignof max align t to is large and write a quick regression test tracked by microsoft internal vso and vso
0
34,884
14,538,137,168
IssuesEvent
2020-12-15 10:04:34
elastic/kibana
https://api.github.com/repos/elastic/kibana
closed
[APM] Service overview: instances table
Team:apm [zube]: In Progress apm:service-overview v7.11.0
(Part of the new APM service overview. See elastic/kibana#81135 for more.) Show table of instances and metrics for the service. The table will be sorted by **traffic** by default ![image](https://user-images.githubusercontent.com/9912/97252270-98915180-17d7-11eb-8316-e7f4b02fc087.png)
1.0
[APM] Service overview: instances table - (Part of the new APM service overview. See elastic/kibana#81135 for more.) Show table of instances and metrics for the service. The table will be sorted by **traffic** by default ![image](https://user-images.githubusercontent.com/9912/97252270-98915180-17d7-11eb-8316-e7f4b02fc087.png)
non_process
service overview instances table part of the new apm service overview see elastic kibana for more show table of instances and metrics for the service the table will be sorted by traffic by default
0
188,885
6,786,895,051
IssuesEvent
2017-10-31 00:27:36
uva-slp/meals
https://api.github.com/repos/uva-slp/meals
opened
Fix Login Issues on Course Server
bug priority-very-high story-points-3
Fix the issue with users registering but not being able to login.
1.0
Fix Login Issues on Course Server - Fix the issue with users registering but not being able to login.
non_process
fix login issues on course server fix the issue with users registering but not being able to login
0
19,518
25,829,638,464
IssuesEvent
2022-12-12 15:15:31
microsoft/vscode
https://api.github.com/repos/microsoft/vscode
closed
Local terminal uses remote shell path
bug remote confirmation-pending terminal-profiles terminal-process
Issue Type: <b>Bug</b> I'm SSH'd to a remote Linux machine where zsh is the default shell. It's at /usr/bin/zsh. When I try to launch a local shell on my Mac via the command palette, I get the error: The terminal Process failed to launch: Path to shell executable "/usr/bin/zsh" does not exist. On my local Mac the path is /bin/zsh, so it won't work. Local shells should launch with different expectations about the shell and path to it. VS Code version: Code 1.65.2 (c722ca6c7eed3d7987c0d5c3df5c45f6b15e77d1, 2022-03-10T14:33:49.188Z) OS version: Darwin x64 19.6.0 Restricted Mode: No Remote OS version: Linux x64 5.13.0-30-generic <details> <summary>System Info</summary> |Item|Value| |---|---| |CPUs|Intel(R) Core(TM) i9-9980HK CPU @ 2.40GHz (16 x 2400)| |GPU Status|2d_canvas: enabled<br>gpu_compositing: enabled<br>metal: disabled_off<br>multiple_raster_threads: enabled_on<br>oop_rasterization: enabled<br>opengl: enabled_on<br>rasterization: enabled<br>skia_renderer: disabled_off_ok<br>video_decode: enabled<br>webgl: enabled<br>webgl2: enabled| |Load (avg)|1, 2, 2| |Memory (System)|16.00GB (0.24GB free)| |Process Argv|-psn_0_13114497 --crash-reporter-id 2e2b7f5a-8a45-4967-bac0-e501e2043bd4| |Screen Reader|no| |VM|0%| |Item|Value| |---|---| |Remote|SSH: REDACTED| |OS|Linux x64 5.13.0-30-generic| |CPUs|AMD Ryzen 9 5950X 16-Core Processor (32 x 2200)| |Memory (System)|62.71GB (0.46GB free)| |VM|0%| </details><details><summary>Extensions (24)</summary> Extension|Author (truncated)|Version ---|---|--- better-cpp-syntax|jef|1.15.13 kconfig|luv|1.0.3 autoconf|mae|0.1.0 jupyter-keymap|ms-|1.0.0 remote-containers|ms-|0.231.0 remote-ssh|ms-|0.76.1 remote-ssh-edit|ms-|0.78.0 remote-wsl|ms-|0.64.2 doxdocgen|csc|1.4.0 xml|Dot|2.5.1 gitlens|eam|12.0.5 vscode-pull-request-github|Git|0.38.1 better-cpp-syntax|jef|1.15.13 python|ms-|2022.2.1924087327 vscode-pylance|ms-|2022.3.3 jupyter|ms-|2022.2.1030672458 cmake-tools|ms-|1.10.5 cpptools|ms-|1.9.7 cpptools-extension-pack|ms-|1.1.0 vsliveshare|ms-|1.0.5449 vsliveshare-audio|ms-|0.1.91 perforce|sle|3.2.0 vscode-hexdump|sle|1.8.1 cmake|twx|0.0.17 (2 theme extensions excluded) </details><details> <summary>A/B Experiments</summary> ``` vsliv368cf:30146710 vsreu685:30147344 python383:30185418 vspor879:30202332 vspor708:30202333 vspor363:30204092 pythonvspyl392:30443607 pythontb:30283811 pythonvspyt551cf:30345471 pythonptprofiler:30281270 vshan820:30294714 vstes263:30335439 vscorecescf:30445987 pythondataviewer:30285071 vscod805:30301674 pythonvspyt200:30340761 binariesv615:30325510 bridge0708:30335490 bridge0723:30353136 vsaa593cf:30376535 vsc1dst:30438360 pythonvs932:30410667 wslgetstarted:30449410 vsclayoutctrc:30451274 dsvsc009:30452663 pythonvsnew555:30457759 vscscmwlcmt:30438805 vscgsvid1:30447480 pynewfile477:30463512 ``` </details> <!-- generated by issue reporter -->
1.0
Local terminal uses remote shell path - Issue Type: <b>Bug</b> I'm SSH'd to a remote Linux machine where zsh is the default shell. It's at /usr/bin/zsh. When I try to launch a local shell on my Mac via the command palette, I get the error: The terminal Process failed to launch: Path to shell executable "/usr/bin/zsh" does not exist. On my local Mac the path is /bin/zsh, so it won't work. Local shells should launch with different expectations about the shell and path to it. VS Code version: Code 1.65.2 (c722ca6c7eed3d7987c0d5c3df5c45f6b15e77d1, 2022-03-10T14:33:49.188Z) OS version: Darwin x64 19.6.0 Restricted Mode: No Remote OS version: Linux x64 5.13.0-30-generic <details> <summary>System Info</summary> |Item|Value| |---|---| |CPUs|Intel(R) Core(TM) i9-9980HK CPU @ 2.40GHz (16 x 2400)| |GPU Status|2d_canvas: enabled<br>gpu_compositing: enabled<br>metal: disabled_off<br>multiple_raster_threads: enabled_on<br>oop_rasterization: enabled<br>opengl: enabled_on<br>rasterization: enabled<br>skia_renderer: disabled_off_ok<br>video_decode: enabled<br>webgl: enabled<br>webgl2: enabled| |Load (avg)|1, 2, 2| |Memory (System)|16.00GB (0.24GB free)| |Process Argv|-psn_0_13114497 --crash-reporter-id 2e2b7f5a-8a45-4967-bac0-e501e2043bd4| |Screen Reader|no| |VM|0%| |Item|Value| |---|---| |Remote|SSH: REDACTED| |OS|Linux x64 5.13.0-30-generic| |CPUs|AMD Ryzen 9 5950X 16-Core Processor (32 x 2200)| |Memory (System)|62.71GB (0.46GB free)| |VM|0%| </details><details><summary>Extensions (24)</summary> Extension|Author (truncated)|Version ---|---|--- better-cpp-syntax|jef|1.15.13 kconfig|luv|1.0.3 autoconf|mae|0.1.0 jupyter-keymap|ms-|1.0.0 remote-containers|ms-|0.231.0 remote-ssh|ms-|0.76.1 remote-ssh-edit|ms-|0.78.0 remote-wsl|ms-|0.64.2 doxdocgen|csc|1.4.0 xml|Dot|2.5.1 gitlens|eam|12.0.5 vscode-pull-request-github|Git|0.38.1 better-cpp-syntax|jef|1.15.13 python|ms-|2022.2.1924087327 vscode-pylance|ms-|2022.3.3 jupyter|ms-|2022.2.1030672458 cmake-tools|ms-|1.10.5 cpptools|ms-|1.9.7 cpptools-extension-pack|ms-|1.1.0 vsliveshare|ms-|1.0.5449 vsliveshare-audio|ms-|0.1.91 perforce|sle|3.2.0 vscode-hexdump|sle|1.8.1 cmake|twx|0.0.17 (2 theme extensions excluded) </details><details> <summary>A/B Experiments</summary> ``` vsliv368cf:30146710 vsreu685:30147344 python383:30185418 vspor879:30202332 vspor708:30202333 vspor363:30204092 pythonvspyl392:30443607 pythontb:30283811 pythonvspyt551cf:30345471 pythonptprofiler:30281270 vshan820:30294714 vstes263:30335439 vscorecescf:30445987 pythondataviewer:30285071 vscod805:30301674 pythonvspyt200:30340761 binariesv615:30325510 bridge0708:30335490 bridge0723:30353136 vsaa593cf:30376535 vsc1dst:30438360 pythonvs932:30410667 wslgetstarted:30449410 vsclayoutctrc:30451274 dsvsc009:30452663 pythonvsnew555:30457759 vscscmwlcmt:30438805 vscgsvid1:30447480 pynewfile477:30463512 ``` </details> <!-- generated by issue reporter -->
process
local terminal uses remote shell path issue type bug i m ssh d to a remote linux machine where zsh is the default shell it s at usr bin zsh when i try to launch a local shell on my mac via the command palette i get the error the terminal process failed to launch path to shell executable usr bin zsh does not exist on my local mac the path is bin zsh so it won t work local shells should launch with different expectations about the shell and path to it vs code version code os version darwin restricted mode no remote os version linux generic system info item value cpus intel r core tm cpu x gpu status canvas enabled gpu compositing enabled metal disabled off multiple raster threads enabled on oop rasterization enabled opengl enabled on rasterization enabled skia renderer disabled off ok video decode enabled webgl enabled enabled load avg memory system free process argv psn crash reporter id screen reader no vm item value remote ssh redacted os linux generic cpus amd ryzen core processor x memory system free vm extensions extension author truncated version better cpp syntax jef kconfig luv autoconf mae jupyter keymap ms remote containers ms remote ssh ms remote ssh edit ms remote wsl ms doxdocgen csc xml dot gitlens eam vscode pull request github git better cpp syntax jef python ms vscode pylance ms jupyter ms cmake tools ms cpptools ms cpptools extension pack ms vsliveshare ms vsliveshare audio ms perforce sle vscode hexdump sle cmake twx theme extensions excluded a b experiments pythontb pythonptprofiler vscorecescf pythondataviewer wslgetstarted vsclayoutctrc vscscmwlcmt
1
72,863
15,251,714,701
IssuesEvent
2021-02-20 00:13:34
wss-demo/ImportedNodeGoat
https://api.github.com/repos/wss-demo/ImportedNodeGoat
opened
CVE-2020-7608 (Medium) detected in yargs-parser-2.4.1.tgz
security vulnerability
## CVE-2020-7608 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>yargs-parser-2.4.1.tgz</b></p></summary> <p>the mighty option parser used by yargs</p> <p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-2.4.1.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-2.4.1.tgz</a></p> <p>Path to dependency file: ImportedNodeGoat/package.json</p> <p>Path to vulnerable library: ImportedNodeGoat/node_modules/nyc/node_modules/yargs-parser/package.json</p> <p> Dependency Hierarchy: - grunt-if-0.2.0.tgz (Root Library) - grunt-contrib-nodeunit-1.0.0.tgz - nodeunit-0.9.5.tgz - tap-7.1.2.tgz - nyc-7.1.0.tgz - :x: **yargs-parser-2.4.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://api.github.com/repos/wss-demo/ImportedNodeGoat/commits/8a9930d26492d27fecc32157284fbe70021fc725">8a9930d26492d27fecc32157284fbe70021fc725</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> yargs-parser could be tricked into adding or modifying properties of Object.prototype using a "__proto__" payload. <p>Publish Date: 2020-03-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7608>CVE-2020-7608</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7608">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7608</a></p> <p>Release Date: 2020-03-16</p> <p>Fix Resolution: v18.1.1;13.1.2;15.0.1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"yargs-parser","packageVersion":"2.4.1","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-if:0.2.0;grunt-contrib-nodeunit:1.0.0;nodeunit:0.9.5;tap:7.1.2;nyc:7.1.0;yargs-parser:2.4.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v18.1.1;13.1.2;15.0.1"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-7608","vulnerabilityDetails":"yargs-parser could be tricked into adding or modifying properties of Object.prototype using a \"__proto__\" payload.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7608","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"Low","S":"Unchanged","C":"Low","UI":"None","AV":"Local","I":"Low"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-7608 (Medium) detected in yargs-parser-2.4.1.tgz - ## CVE-2020-7608 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>yargs-parser-2.4.1.tgz</b></p></summary> <p>the mighty option parser used by yargs</p> <p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-2.4.1.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-2.4.1.tgz</a></p> <p>Path to dependency file: ImportedNodeGoat/package.json</p> <p>Path to vulnerable library: ImportedNodeGoat/node_modules/nyc/node_modules/yargs-parser/package.json</p> <p> Dependency Hierarchy: - grunt-if-0.2.0.tgz (Root Library) - grunt-contrib-nodeunit-1.0.0.tgz - nodeunit-0.9.5.tgz - tap-7.1.2.tgz - nyc-7.1.0.tgz - :x: **yargs-parser-2.4.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://api.github.com/repos/wss-demo/ImportedNodeGoat/commits/8a9930d26492d27fecc32157284fbe70021fc725">8a9930d26492d27fecc32157284fbe70021fc725</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> yargs-parser could be tricked into adding or modifying properties of Object.prototype using a "__proto__" payload. <p>Publish Date: 2020-03-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7608>CVE-2020-7608</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7608">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7608</a></p> <p>Release Date: 2020-03-16</p> <p>Fix Resolution: v18.1.1;13.1.2;15.0.1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"yargs-parser","packageVersion":"2.4.1","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-if:0.2.0;grunt-contrib-nodeunit:1.0.0;nodeunit:0.9.5;tap:7.1.2;nyc:7.1.0;yargs-parser:2.4.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v18.1.1;13.1.2;15.0.1"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-7608","vulnerabilityDetails":"yargs-parser could be tricked into adding or modifying properties of Object.prototype using a \"__proto__\" payload.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7608","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"Low","S":"Unchanged","C":"Low","UI":"None","AV":"Local","I":"Low"},"extraData":{}}</REMEDIATE> -->
non_process
cve medium detected in yargs parser tgz cve medium severity vulnerability vulnerable library yargs parser tgz the mighty option parser used by yargs library home page a href path to dependency file importednodegoat package json path to vulnerable library importednodegoat node modules nyc node modules yargs parser package json dependency hierarchy grunt if tgz root library grunt contrib nodeunit tgz nodeunit tgz tap tgz nyc tgz x yargs parser tgz vulnerable library found in head commit a href found in base branch master vulnerability details yargs parser could be tricked into adding or modifying properties of object prototype using a proto payload publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree grunt if grunt contrib nodeunit nodeunit tap nyc yargs parser isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails yargs parser could be tricked into adding or modifying properties of object prototype using a proto payload vulnerabilityurl
0
8,234
11,415,575,457
IssuesEvent
2020-02-02 12:02:16
parcel-bundler/parcel
https://api.github.com/repos/parcel-bundler/parcel
closed
CSS files being compiled with PostCSS before they're imported
:bug: Bug CSS Preprocessing Stale
# 🐛 CSS files appear to be compiling with PostCSS before they're imported. This will lead to unexpected behavior in most PostCSS builds, as imports are typically the first thing that should run. See: https://github.com/postcss/postcss-import > This plugin should probably be used as the first plugin of your list. This way, other plugins will work on the AST as if there were only a single file to process, and will probably work as you can expect. Possibly related to [#329](https://github.com/parcel-bundler/parcel/issues/329) and [#593](https://github.com/parcel-bundler/parcel/issues/593). ### 🎛 Configuration package.json ``` { "main": "index.js", "scripts": { "start": "parcel src/index.html" }, "devDependencies": { "parcel-bundler": "^1.4.1", "postcss-easy-import": "^3.0.0", "postcss-simple-vars": "^4.1.0" } } ``` postcss.config.js ``` module.exports = { plugins: [ require("postcss-easy-import"), require("postcss-simple-vars") ] }; ``` index.html ``` <html> <head> <link rel="stylesheet" href="./assets/css/index.css"></link> </head> <body> <script src="./index.js"></script> </body> </html> ``` assets/css/index.css ``` @import "./global/colors.css"; body { background: $red; } ``` assets/css/global/colors.css ``` $red: #f00; ``` ### 🤔 Expected Behavior ``` body { background: #f00; } ``` ### 😯 Current Behavior ``` 🚨 /Users/jon/git/parcel-test/src/assets/css/index.css:4:3: Undefined variable $red 2 | 3 | body { > 4 | background: $red; | ^ 5 | } 6 | ``` ### 💁 Possible Solution It appears that CSS imports will work even without requiring a PostCSS plugin like `postcss-easy-import`. Because different import plugins can behave differently, and because the order of PostCSS plugins is very important, I believe CSS import behavior should be left completely up to the PostCSS config and plugin. ### 🔦 Context This config works with webpack + postcss-loader, and gulp. ### 💻 Code Sample https://github.com/jonmilner/parcel-test ### 🌍 Your Environment <!--- Include as many relevant details about the environment you experienced the bug in --> | Software | Version(s) | | ---------------- | ---------- | | Parcel | 1.4.1 | Node | 8.2.1 | npm/Yarn | 1.3.2 | Operating System | macOS 10.13.3
1.0
CSS files being compiled with PostCSS before they're imported - # 🐛 CSS files appear to be compiling with PostCSS before they're imported. This will lead to unexpected behavior in most PostCSS builds, as imports are typically the first thing that should run. See: https://github.com/postcss/postcss-import > This plugin should probably be used as the first plugin of your list. This way, other plugins will work on the AST as if there were only a single file to process, and will probably work as you can expect. Possibly related to [#329](https://github.com/parcel-bundler/parcel/issues/329) and [#593](https://github.com/parcel-bundler/parcel/issues/593). ### 🎛 Configuration package.json ``` { "main": "index.js", "scripts": { "start": "parcel src/index.html" }, "devDependencies": { "parcel-bundler": "^1.4.1", "postcss-easy-import": "^3.0.0", "postcss-simple-vars": "^4.1.0" } } ``` postcss.config.js ``` module.exports = { plugins: [ require("postcss-easy-import"), require("postcss-simple-vars") ] }; ``` index.html ``` <html> <head> <link rel="stylesheet" href="./assets/css/index.css"></link> </head> <body> <script src="./index.js"></script> </body> </html> ``` assets/css/index.css ``` @import "./global/colors.css"; body { background: $red; } ``` assets/css/global/colors.css ``` $red: #f00; ``` ### 🤔 Expected Behavior ``` body { background: #f00; } ``` ### 😯 Current Behavior ``` 🚨 /Users/jon/git/parcel-test/src/assets/css/index.css:4:3: Undefined variable $red 2 | 3 | body { > 4 | background: $red; | ^ 5 | } 6 | ``` ### 💁 Possible Solution It appears that CSS imports will work even without requiring a PostCSS plugin like `postcss-easy-import`. Because different import plugins can behave differently, and because the order of PostCSS plugins is very important, I believe CSS import behavior should be left completely up to the PostCSS config and plugin. ### 🔦 Context This config works with webpack + postcss-loader, and gulp. ### 💻 Code Sample https://github.com/jonmilner/parcel-test ### 🌍 Your Environment <!--- Include as many relevant details about the environment you experienced the bug in --> | Software | Version(s) | | ---------------- | ---------- | | Parcel | 1.4.1 | Node | 8.2.1 | npm/Yarn | 1.3.2 | Operating System | macOS 10.13.3
process
css files being compiled with postcss before they re imported 🐛 css files appear to be compiling with postcss before they re imported this will lead to unexpected behavior in most postcss builds as imports are typically the first thing that should run see this plugin should probably be used as the first plugin of your list this way other plugins will work on the ast as if there were only a single file to process and will probably work as you can expect possibly related to and 🎛 configuration package json main index js scripts start parcel src index html devdependencies parcel bundler postcss easy import postcss simple vars postcss config js module exports plugins require postcss easy import require postcss simple vars index html assets css index css import global colors css body background red assets css global colors css red 🤔 expected behavior body background 😯 current behavior 🚨 users jon git parcel test src assets css index css undefined variable red body background red 💁 possible solution it appears that css imports will work even without requiring a postcss plugin like postcss easy import because different import plugins can behave differently and because the order of postcss plugins is very important i believe css import behavior should be left completely up to the postcss config and plugin 🔦 context this config works with webpack postcss loader and gulp 💻 code sample 🌍 your environment software version s parcel node npm yarn operating system macos
1
10,770
13,594,677,372
IssuesEvent
2020-09-22 00:37:55
kubernetes/minikube
https://api.github.com/repos/kubernetes/minikube
closed
Release process: run brew bump-formula-pr in docker
help wanted kind/process lifecycle/rotten priority/important-longterm
we need to run this command in a docker so it doesnt die of gem dependency ``` brew bump-formula-pr \ --strict minikube \ --revision="${revision}" \ --message="This PR was automatically created by minikube release scripts. Contact @tstromberg with any questions." \ --no-browse \ --tag="${TAG}" \ && status=0 || status=$? ``` https://github.com/kubernetes/minikube/blob/eb13446e786c9ef70cb0a9f85a633194e62396a1/hack/jenkins/release_update_brew.sh#L48 here is the current gem dep error from out latest 1.8.2 release <details> <pre> 14:33:14 + brew bump-formula-pr --strict minikube --revision=eb13446e786c9ef70cb0a9f85a633194e62396a1 '--message=This PR was automatically created by minikube release scripts. Contact @tstromberg with any questions.' --no-browse --tag=v1.8.2 14:33:18 Updating Homebrew... 14:33:22 ==> Auto-updated Homebrew! 14:33:22 Updated Homebrew from 78140cef4 to db121c7b1. 14:35:02 Updated 1 tap (homebrew/core). 14:35:02 ==> New Formulae 14:35:02 adoptopenjdk 14:35:02 adoptopenjdk@11 14:35:02 alsa-lib 14:35:02 apt 14:35:02 archiver 14:35:02 at-spi2-atk 14:35:02 at-spi2-core 14:35:02 attr 14:35:02 avahi 14:35:02 aws-vault 14:35:02 bsdmainutils 14:35:02 cpio 14:35:02 device-mapper 14:35:02 elfutils 14:35:02 gcc@9 14:35:02 git-trim 14:35:02 glibc 14:35:02 hdf5@1.10 14:35:02 hdt 14:35:02 isl@0.18 14:35:02 jdk 14:35:02 jdk@7 14:35:02 jdk@8 14:35:02 ladspa-sdk 14:35:02 libaio 14:35:02 libbsd 14:35:02 libcap 14:35:02 libcap-ng 14:35:02 libfuse 14:35:02 libmnl 14:35:02 libnetfilter-queue 14:35:02 libnfnetlink 14:35:02 libnsl 14:35:02 libprelude 14:35:02 libsasl2 14:35:02 libseccomp 14:35:02 libtirpc 14:35:02 linux-headers 14:35:02 linux-pam 14:35:02 lm-sensors 14:35:02 minimal-python 14:35:02 mockolo 14:35:02 net-tools 14:35:02 newrelic-cli 14:35:02 onedrive 14:35:02 openjdk@10 14:35:02 openjdk@9 14:35:02 procps 14:35:02 python-dbus 14:35:02 python@2 14:35:02 qp 14:35:02 singularity 14:35:02 strace 14:35:02 sysstat 14:35:02 systemd 14:35:02 texlive 14:35:02 tlx 14:35:02 xclogparser 14:35:02 ==> Updated Formulae 14:35:02 abseil 14:35:02 advancemenu 14:35:02 adwaita-icon-theme 14:35:02 afuse 14:35:02 agda 14:35:02 algol68g 14:35:02 angular-cli 14:35:02 ansible 14:35:02 ansible-lint 14:35:02 apache-drill 14:35:02 apibuilder-cli 14:35:02 apng2gif 14:35:02 apr-util 14:35:02 arangodb 14:35:02 argon2 14:35:02 aria2 14:35:02 armadillo 14:35:02 arping 14:35:02 artifactory 14:35:02 asciidoc 14:35:02 asciinema 14:35:02 astrometry-net 14:35:02 atdtool 14:35:02 autoconf 14:35:02 aws-cdk 14:35:02 aws-elasticbeanstalk 14:35:02 aws-shell 14:35:02 awscli 14:35:02 awscli@1 14:35:02 awscurl 14:35:02 awslogs 14:35:02 awsume 14:35:02 azure-cli 14:35:02 azure-storage-cpp 14:35:02 bacula-fd 14:35:02 balena-cli 14:35:02 bandcamp-dl 14:35:02 bazaar 14:35:02 bde 14:35:02 benthos 14:35:02 biosig 14:35:02 bison 14:35:02 bitcoin 14:35:02 black 14:35:02 blazegraph 14:35:02 bonnie++ 14:35:02 buildifier 14:35:02 bundletool 14:35:02 bzt 14:35:02 caffe 14:35:02 camlp5 14:35:02 cargo-completion 14:35:02 cartridge-cli 14:35:02 ceres-solver 14:35:02 certbot 14:35:02 cfn-lint 14:35:02 cfr-decompiler 14:35:02 cgns 14:35:02 cgrep 14:35:02 chakra 14:35:02 chapel 14:35:02 charm-tools 14:35:02 cheat 14:35:02 cli53 14:35:02 clojure 14:35:02 clojure-lsp 14:35:02 clutter 14:35:02 cobalt 14:35:02 cocoapods 14:35:02 codec2 14:35:02 codemod 14:35:02 codespell 14:35:02 cogl 14:35:02 comby 14:35:02 composer 14:35:02 conan 14:35:02 confluent-platform 14:35:02 conjure-up 14:35:02 cookiecutter 14:35:02 cpl 14:35:02 cppman 14:35:02 cquery 14:35:02 csound 14:35:02 csvkit 14:35:02 csvtomd 14:35:02 curl 14:35:02 curl-openssl 14:35:02 cvs 14:35:02 dartsim 14:35:02 davmail 14:35:02 db-vcs 14:35:02 dependency-check 14:35:02 derby 14:35:02 detekt 14:35:02 dialog 14:35:02 diceware 14:35:02 diff-pdf 14:35:02 diffoscope 14:35:02 dmd 14:35:02 docbook-xsl 14:35:02 docbook2x 14:35:02 docfx 14:35:02 docker 14:35:02 docker-completion 14:35:02 docker-compose 14:35:02 docker-slim 14:35:02 docker-squash 14:35:02 dosbox-x 14:35:02 double-conversion 14:35:02 dovecot 14:35:02 dtc 14:35:02 dub 14:35:02 dune 14:35:02 duplicity 14:35:02 dust 14:35:02 dxpy 14:35:02 dynare 14:35:02 e2fsprogs 14:35:02 easyrpg-player 14:35:02 ecasound 14:35:02 eccodes 14:35:02 eg-examples 14:35:02 ejdb 14:35:02 elasticsearch 14:35:02 embree 14:35:02 emscripten 14:35:02 ensmallen 14:35:02 esptool 14:35:02 etsh 14:35:02 evince 14:35:02 exim 14:35:02 expect 14:35:02 exploitdb 14:35:02 ext2fuse 14:35:02 ezstream 14:35:02 fairymax 14:35:02 faiss 14:35:02 fantom 14:35:02 fastlane 14:35:02 faudio 14:35:02 fcl 14:35:02 fdroidserver 14:35:02 file-roller 14:35:02 findomain 14:35:02 flann 14:35:02 flatbuffers 14:35:02 flintrock 14:35:02 fluent-bit 14:35:02 flyway 14:35:02 fn 14:35:02 folly 14:35:02 foma 14:35:02 fontconfig 14:35:02 fonttools 14:35:02 fq 14:35:02 frpc 14:35:02 frps 14:35:02 frugal 14:35:02 fselect 14:35:02 fzf 14:35:02 gambit 14:35:02 gambit-scheme 14:35:02 gammaray 14:35:02 gandi.cli 14:35:02 gatsby-cli 14:35:02 gcc 14:35:02 gcc@8 14:35:02 gdal 14:35:02 gdcm 14:35:02 geeqie 14:35:02 geocode-glib 14:35:02 geos 14:35:02 gerbil-scheme 14:35:02 gettext 14:35:02 gimme-aws-creds 14:35:02 git 14:35:02 git-annex 14:35:02 git-plus 14:35:02 git-review 14:35:02 git-revise 14:35:02 git-standup 14:35:02 gitmoji 14:35:02 gitup 14:35:02 gjs 14:35:02 glances 14:35:02 glib 14:35:02 glog 14:35:02 gmsh 14:35:02 gmt 14:35:02 gmt@5 14:35:02 gobject-introspection 14:35:02 google-sparsehash 14:35:02 gpgme 14:35:02 grin 14:35:02 grin-wallet 14:35:02 gromacs 14:35:02 groovy 14:35:02 groovysdk 14:35:02 gsettings-desktop-schemas 14:35:02 gti 14:35:02 gtkmm 14:35:02 gtksourceview4 14:35:02 hdf5 14:35:02 helm 14:35:02 hexyl 14:35:02 homeassistant-cli 14:35:02 hostess 14:35:02 howdoi 14:35:02 htop 14:35:02 httpie 14:35:02 hugo 14:35:02 hyperkit 14:35:02 ibex 14:35:02 ilmbase 14:35:02 imagemagick 14:35:02 imagemagick@6 14:35:02 imgproxy 14:35:02 inadyn 14:35:02 inlets 14:35:02 instalooter 14:35:02 internetarchive 14:35:02 intltool 14:35:02 io 14:35:02 ios-sim 14:35:02 ipmiutil 14:35:02 ipython 14:35:02 isl 14:35:02 itk 14:35:02 jbig2dec 14:35:02 jed 14:35:02 jenkins 14:35:02 jenkins-job-builder 14:35:02 jenkins-lts 14:35:02 jhead 14:35:02 jhipster 14:35:02 jinja2-cli 14:35:02 joplin 14:35:02 jrnl 14:35:02 jruby 14:35:02 juise 14:35:02 juju 14:35:02 juju-wait 14:35:02 jupyterlab 14:35:02 kallisto 14:35:02 keepkey-agent 14:35:02 khard 14:35:02 kibana 14:35:02 krew 14:35:02 ktoblzcheck 14:35:02 kubeprod 14:35:02 kubernetes-cli 14:35:02 kubeseal 14:35:02 kumo 14:35:02 lablgtk 14:35:02 lammps 14:35:02 ldc 14:35:02 legit 14:35:02 lgogdownloader 14:35:02 libbi 14:35:02 libcouchbase 14:35:02 libdazzle 14:35:02 libewf 14:35:02 libexosip 14:35:02 libfabric 14:35:02 libgetdata 14:35:02 libgosu 14:35:02 libgweather 14:35:02 libhttpseverywhere 14:35:02 libical 14:35:02 libmatio 14:35:02 libowfat 14:35:02 libpeas 14:35:02 libproxy 14:35:02 libpsl 14:35:02 libqalculate 14:35:02 librealsense 14:35:02 libsecret 14:35:02 libserdes 14:35:02 libsoup 14:35:02 libssh 14:35:02 libsvm 14:35:02 libtcod 14:35:02 libtool 14:35:02 libtorrent-rasterbar 14:35:02 libupnp 14:35:02 libwebsockets 14:35:02 libxml++ 14:35:02 libzdb 14:35:02 liquibase 14:35:02 literate-git 14:35:02 llvm 14:35:02 llvm@6 14:35:02 llvm@7 14:35:02 llvm@8 14:35:02 lolcat 14:35:02 lua 14:35:02 lua@5.1 14:35:02 luaradio 14:35:02 luarocks 14:35:02 luit 14:35:02 mackup 14:35:02 magic-wormhole 14:35:02 mapserver 14:35:02 mbedtls 14:35:02 mdds 14:35:02 mdv 14:35:02 memcached 14:35:02 menhir 14:35:02 mercurial 14:35:02 mercury 14:35:02 mesa 14:35:02 meson-internal 14:35:02 metabase 14:35:02 micronaut 14:35:02 minikube 14:35:02 minio 14:35:02 minizinc 14:35:02 minizip2 14:35:02 mitmproxy 14:35:02 mkdocs 14:35:02 mkl-dnn 14:35:02 mkvtoolnix 14:35:02 mlpack 14:35:02 molecule 14:35:02 mongo-cxx-driver 14:35:02 mongo-orchestration 14:35:02 mosh 14:35:02 mosquitto 14:35:02 mpd 14:35:02 mruby 14:35:02 mujs 14:35:02 mycli 14:35:02 mypy 14:35:02 mysql-connector-c++ 14:35:02 n 14:35:02 nagios-plugins 14:35:02 nbdime 14:35:02 nco 14:35:02 ncview 14:35:02 neofetch 14:35:02 neovim 14:35:02 netcdf 14:35:02 nethack 14:35:02 ninja 14:35:02 nng 14:35:02 node-sass 14:35:02 nodeenv 14:35:02 nss 14:35:02 nudoku 14:35:02 nushell 14:35:02 oauth2l 14:35:02 ocaml 14:35:02 ocaml-findlib 14:35:02 ocaml-num 14:35:02 ocamlbuild 14:35:02 ocamlsdl 14:35:02 ocrmypdf 14:35:02 octave 14:35:02 ode 14:35:02 odpi 14:35:02 onnxruntime 14:35:02 open-image-denoise 14:35:02 open-mpi 14:35:02 openexr 14:35:02 openssl@1.1 14:35:02 pastel 14:35:02 pcl 14:35:02 pdal 14:35:02 pgcli 14:35:02 pipenv 14:35:02 pipx 14:35:02 pkg-config 14:35:02 platformio 14:35:02 plotutils 14:35:02 pmd 14:35:02 pnpm 14:35:02 points2grid 14:35:02 pre-commit 14:35:02 procs 14:35:02 prodigal 14:35:02 proselint 14:35:02 protoc-gen-go 14:35:02 psc-package 14:35:02 pssh 14:35:02 pulseaudio 14:35:02 pulumi 14:35:02 pushpin 14:35:02 pwntools 14:35:02 pyenv 14:35:02 pygitup 14:35:02 pygments 14:35:02 pygobject3 14:35:02 pyinvoke 14:35:02 pylint 14:35:02 python 14:35:02 python@3.8 14:35:02 pyvim 14:35:02 qalculate-gtk 14:35:02 qt 14:35:02 rabbitmq 14:35:02 radamsa 14:35:02 radare2 14:35:02 rakudo-star 14:35:02 re-flex 14:35:02 remarshal 14:35:02 rex 14:35:02 rst-lint 14:35:02 rtv 14:35:02 rubberband 14:35:02 rust 14:35:02 s3cmd 14:35:02 s3fs 14:35:02 s6 14:35:02 sbjson 14:35:02 sceptre 14:35:02 scons 14:35:02 scour 14:35:02 sdcc 14:35:02 sdedit 14:35:02 sdl2 14:35:02 serf 14:35:02 serverless 14:35:02 sfk 14:35:02 shogun 14:35:02 simg2img 14:35:02 simple-tiles 14:35:02 sjk 14:35:02 skinny 14:35:02 sn0int 14:35:02 sonar-scanner 14:35:02 sonarqube 14:35:02 sphinx-doc 14:35:02 spice-protocol 14:35:02 spidermonkey 14:35:02 spigot 14:35:02 spirv-cross 14:35:02 spirv-tools 14:35:02 sqliteodbc 14:35:02 sqlparse 14:35:02 ssdb 14:35:02 ssh-audit 14:35:02 sshuttle 14:35:02 statik 14:35:02 stdman 14:35:02 stellar-core 14:35:02 stgit 14:35:02 stolon 14:35:02 streamlink 14:35:02 stress-ng 14:35:02 stubby 14:35:02 subversion 14:35:02 suite-sparse 14:35:02 sundials 14:35:02 supervisor 14:35:02 svtplay-dl 14:35:02 swagger-codegen 14:35:02 swift-protobuf 14:35:02 swift-sh 14:35:02 swiftformat 14:35:02 switch-lan-play 14:35:02 talisman 14:35:02 tasksh 14:35:02 tbox 14:35:02 tektoncd-cli 14:35:02 terragrunt 14:35:02 terrahelp 14:35:02 terrahub 14:35:02 texinfo 14:35:02 texmath 14:35:02 tgui 14:35:02 thefuck 14:35:02 tin 14:35:02 tinyxml 14:35:02 tinyxml2 14:35:02 todoman 14:35:02 tox 14:35:02 trafficserver 14:35:02 translate-toolkit 14:35:02 trash-cli 14:35:02 ttyd 14:35:02 twarc 14:35:02 twine-pypi 14:35:02 txr 14:35:02 txt2tags 14:35:02 u-boot-tools 14:35:02 urh 14:35:02 v2ray-plugin 14:35:02 v8 14:35:02 vala 14:35:02 vale 14:35:02 valgrind 14:35:02 vault 14:35:02 velero 14:35:02 verilator 14:35:02 vips 14:35:02 visp 14:35:02 vowpal-wabbit 14:35:02 vsts-cli 14:35:02 vtk 14:35:02 vttest 14:35:02 vulkan-headers 14:35:02 wabt 14:35:02 wal2json 14:35:02 wapm 14:35:02 wartremover 14:35:02 wasmer 14:35:02 watson 14:35:02 wcslib 14:35:02 webdis 14:35:02 websocat 14:35:02 webtorrent-cli 14:35:02 wp-cli-completion 14:35:02 wpscan 14:35:02 wxmaxima 14:35:02 x3270 14:35:02 xclip 14:35:02 xdotool 14:35:02 xmlto 14:35:02 xmrig 14:35:02 xonsh 14:35:02 xtensor 14:35:02 xxhash 14:35:02 yamllint 14:35:02 yapf 14:35:02 yarn 14:35:02 ydcv 14:35:02 ykman 14:35:02 you-get 14:35:02 youtube-dl 14:35:02 z 14:35:02 zabbix-cli 14:35:02 zeek 14:35:02 zimg 14:35:02 znc 14:35:02 zola 14:35:02 ==> Deleted Formulae 14:35:02 camlp4 14:35:02 ck 14:35:02 djmount 14:35:02 field3d 14:35:02 gmediaserver 14:35:02 14:35:03 ==> Fetching homebrew/master 14:35:06 From https://github.com/Homebrew/homebrew-core 14:35:06 * branch master -> FETCH_HEAD 14:35:06 257bb3e8cb..d3eb71b109 master -> homebrew/master 14:35:06 ==> minikube exists in homebrew/master 14:35:07 Previous HEAD position was a99a5e418f... Merge branch homebrew/master into linuxbrew/master 14:35:07 HEAD is now at d3eb71b109... composer 1.10.1 (#51704) 14:35:07 ==> replace "v1.8.1" with "v1.8.2" 14:35:07 ==> replace "cbda04cf6bbe65e987ae52bb393c10099ab62014" with "eb13446e786c9ef70cb0a9f85a633194e62396a1" 14:35:09 Fetching gem metadata from https://rubygems.org/......... 14:35:10 Using concurrent-ruby 1.1.6 14:35:10 Using minitest 5.14.0 14:35:10 Using thread_safe 0.3.6 14:35:10 Using zeitwerk 2.3.0 14:35:10 Using ast 2.4.0 14:35:10 Using bundler 1.17.2 14:35:10 Using connection_pool 2.2.2 14:35:10 Using json 2.3.0 14:35:10 Using docile 1.3.2 14:35:10 Using simplecov-html 0.10.2 14:35:10 Using sync 0.5.0 14:35:10 Using thor 1.0.1 14:35:10 Using diff-lcs 1.3 14:35:10 Using unf_ext 0.0.7.6 14:35:10 Using hpricot 0.8.6 14:35:10 Using jaro_winkler 1.5.4 14:35:10 Using mime-types-data 3.2019.1009 14:35:10 Using net-http-digest_auth 1.4.1 14:35:10 Using mini_portile2 2.4.0 14:35:10 Using ntlm-http 0.1.1 14:35:10 Using webrobots 0.1.2 14:35:10 Using mustache 1.1.1 14:35:10 Using parallel 1.19.1 14:35:10 Using plist 3.5.0 14:35:10 Using rainbow 3.0.0 14:35:10 Using rdiscount 2.2.0.1 14:35:10 Using rexml 3.2.4 14:35:10 Using rspec-support 3.9.2 14:35:10 Using ruby-progressbar 1.10.1 14:35:10 Using unicode-display_width 1.6.1 14:35:10 Using ruby-macho 2.2.0 14:35:10 Using i18n 1.8.2 14:35:10 Using tzinfo 1.2.6 14:35:10 Using parser 2.7.0.4 14:35:10 Using tins 1.24.1 14:35:10 Using net-http-persistent 3.1.0 14:35:10 Using unf 0.1.4 14:35:10 Using mime-types 3.3.1 14:35:10 Using simplecov 0.16.1 14:35:10 Using ronn 0.7.3 14:35:10 Using rspec-core 3.9.1 14:35:10 Using rspec-expectations 3.9.0 14:35:10 Using rspec-mocks 3.9.1 14:35:10 Using activesupport 6.0.2.1 14:35:10 Using parallel_tests 2.31.0 14:35:10 Using rubocop 0.80.1 14:35:10 Using term-ansicolor 1.7.1 14:35:10 Fetching nokogiri 1.10.9 14:35:10 Using coveralls 0.8.23 14:35:10 Using domain_name 0.5.20190701 14:35:10 Using rspec 3.9.0 14:35:10 Using rspec-its 1.3.0 14:35:10 Using rspec-retry 0.6.2 14:35:10 Using rubocop-performance 1.5.2 14:35:10 Using rubocop-rspec 1.38.1 14:35:10 Using http-cookie 1.0.3 14:35:10 Using rspec-wait 0.0.9 14:35:10 Installing nokogiri 1.10.9 with native extensions 14:35:12 Gem::Ext::BuildError: ERROR: Failed to build gem native extension. 14:35:12 14:35:12 current directory: 14:35:12 /home/linuxbrew/.linuxbrew/Homebrew/Library/Homebrew/vendor/bundle/ruby/2.6.0/gems/nokogiri-1.10.9/ext/nokogiri 14:35:12 /home/linuxbrew/.linuxbrew/Homebrew/Library/Homebrew/vendor/portable-ruby/2.6.3/bin/ruby 14:35:12 -I 14:35:12 /home/linuxbrew/.linuxbrew/Homebrew/Library/Homebrew/vendor/portable-ruby/2.6.3/lib/ruby/2.6.0 14:35:12 -r ./siteconf20200313-30573-18g5kc4.rb extconf.rb 14:35:12 checking if the C compiler accepts ... yes 14:35:12 Building nokogiri using packaged libraries. 14:35:12 Using mini_portile version 2.4.0 14:35:12 checking for gzdopen() in -lz... no 14:35:12 zlib is missing; necessary for building libxml2 14:35:12 *** extconf.rb failed *** 14:35:12 Could not create Makefile due to some reason, probably lack of necessary 14:35:12 libraries and/or headers. Check the mkmf.log file for more details. You may 14:35:12 need configuration options. 14:35:12 14:35:12 Provided configuration options: 14:35:12 --with-opt-dir 14:35:12 --with-opt-include 14:35:12 --without-opt-include=${opt-dir}/include 14:35:12 --with-opt-lib 14:35:12 --without-opt-lib=${opt-dir}/lib 14:35:12 --with-make-prog 14:35:12 --without-make-prog 14:35:12 --srcdir=. 14:35:12 --curdir 14:35:12 --ruby=/home/linuxbrew/.linuxbrew/Homebrew/Library/Homebrew/vendor/portable-ruby/2.6.3/bin/$(RUBY_BASE_NAME) 14:35:12 --help 14:35:12 --clean 14:35:12 --use-system-libraries 14:35:12 --enable-static 14:35:12 --disable-static 14:35:12 --with-zlib-dir 14:35:12 --without-zlib-dir 14:35:12 --with-zlib-include 14:35:12 --without-zlib-include=${zlib-dir}/include 14:35:12 --with-zlib-lib 14:35:12 --without-zlib-lib=${zlib-dir}/lib 14:35:12 --enable-cross-build 14:35:12 --disable-cross-build 14:35:12 14:35:12 To see why this extension failed to compile, please check the mkmf.log which can 14:35:12 be found here: 14:35:12 14:35:12 /home/linuxbrew/.linuxbrew/Homebrew/Library/Homebrew/vendor/bundle/ruby/2.6.0/extensions/x86_64-linux/2.6.0-static/nokogiri-1.10.9/mkmf.log 14:35:12 14:35:12 extconf failed, exit code 1 14:35:12 14:35:12 Gem files will remain installed in 14:35:12 /home/linuxbrew/.linuxbrew/Homebrew/Library/Homebrew/vendor/bundle/ruby/2.6.0/gems/nokogiri-1.10.9 14:35:12 for inspection. 14:35:12 Results logged to 14:35:12 /home/linuxbrew/.linuxbrew/Homebrew/Library/Homebrew/vendor/bundle/ruby/2.6.0/extensions/x86_64-linux/2.6.0-static/nokogiri-1.10.9/gem_make.out 14:35:12 14:35:12 An error occurred while installing nokogiri (1.10.9), and Bundler cannot 14:35:12 continue. 14:35:12 Make sure that `gem install nokogiri -v '1.10.9' --source 14:35:12 'https://rubygems.org/'` succeeds before bundling. 14:35:12 14:35:12 In Gemfile: 14:35:12 mechanize was resolved to 2.7.6, which depends on 14:35:12 nokogiri 14:35:12 Error: failed to run `/home/linuxbrew/.linuxbrew/Homebrew/Library/Homebrew/vendor/portable-ruby/2.6.3/bin/bundle install`! 14:35:12 Error: `brew audit` failed! </pre> </details>
1.0
Release process: run brew bump-formula-pr in docker - we need to run this command in a docker so it doesnt die of gem dependency ``` brew bump-formula-pr \ --strict minikube \ --revision="${revision}" \ --message="This PR was automatically created by minikube release scripts. Contact @tstromberg with any questions." \ --no-browse \ --tag="${TAG}" \ && status=0 || status=$? ``` https://github.com/kubernetes/minikube/blob/eb13446e786c9ef70cb0a9f85a633194e62396a1/hack/jenkins/release_update_brew.sh#L48 here is the current gem dep error from out latest 1.8.2 release <details> <pre> 14:33:14 + brew bump-formula-pr --strict minikube --revision=eb13446e786c9ef70cb0a9f85a633194e62396a1 '--message=This PR was automatically created by minikube release scripts. Contact @tstromberg with any questions.' --no-browse --tag=v1.8.2 14:33:18 Updating Homebrew... 14:33:22 ==> Auto-updated Homebrew! 14:33:22 Updated Homebrew from 78140cef4 to db121c7b1. 14:35:02 Updated 1 tap (homebrew/core). 14:35:02 ==> New Formulae 14:35:02 adoptopenjdk 14:35:02 adoptopenjdk@11 14:35:02 alsa-lib 14:35:02 apt 14:35:02 archiver 14:35:02 at-spi2-atk 14:35:02 at-spi2-core 14:35:02 attr 14:35:02 avahi 14:35:02 aws-vault 14:35:02 bsdmainutils 14:35:02 cpio 14:35:02 device-mapper 14:35:02 elfutils 14:35:02 gcc@9 14:35:02 git-trim 14:35:02 glibc 14:35:02 hdf5@1.10 14:35:02 hdt 14:35:02 isl@0.18 14:35:02 jdk 14:35:02 jdk@7 14:35:02 jdk@8 14:35:02 ladspa-sdk 14:35:02 libaio 14:35:02 libbsd 14:35:02 libcap 14:35:02 libcap-ng 14:35:02 libfuse 14:35:02 libmnl 14:35:02 libnetfilter-queue 14:35:02 libnfnetlink 14:35:02 libnsl 14:35:02 libprelude 14:35:02 libsasl2 14:35:02 libseccomp 14:35:02 libtirpc 14:35:02 linux-headers 14:35:02 linux-pam 14:35:02 lm-sensors 14:35:02 minimal-python 14:35:02 mockolo 14:35:02 net-tools 14:35:02 newrelic-cli 14:35:02 onedrive 14:35:02 openjdk@10 14:35:02 openjdk@9 14:35:02 procps 14:35:02 python-dbus 14:35:02 python@2 14:35:02 qp 14:35:02 singularity 14:35:02 strace 14:35:02 sysstat 14:35:02 systemd 14:35:02 texlive 14:35:02 tlx 14:35:02 xclogparser 14:35:02 ==> Updated Formulae 14:35:02 abseil 14:35:02 advancemenu 14:35:02 adwaita-icon-theme 14:35:02 afuse 14:35:02 agda 14:35:02 algol68g 14:35:02 angular-cli 14:35:02 ansible 14:35:02 ansible-lint 14:35:02 apache-drill 14:35:02 apibuilder-cli 14:35:02 apng2gif 14:35:02 apr-util 14:35:02 arangodb 14:35:02 argon2 14:35:02 aria2 14:35:02 armadillo 14:35:02 arping 14:35:02 artifactory 14:35:02 asciidoc 14:35:02 asciinema 14:35:02 astrometry-net 14:35:02 atdtool 14:35:02 autoconf 14:35:02 aws-cdk 14:35:02 aws-elasticbeanstalk 14:35:02 aws-shell 14:35:02 awscli 14:35:02 awscli@1 14:35:02 awscurl 14:35:02 awslogs 14:35:02 awsume 14:35:02 azure-cli 14:35:02 azure-storage-cpp 14:35:02 bacula-fd 14:35:02 balena-cli 14:35:02 bandcamp-dl 14:35:02 bazaar 14:35:02 bde 14:35:02 benthos 14:35:02 biosig 14:35:02 bison 14:35:02 bitcoin 14:35:02 black 14:35:02 blazegraph 14:35:02 bonnie++ 14:35:02 buildifier 14:35:02 bundletool 14:35:02 bzt 14:35:02 caffe 14:35:02 camlp5 14:35:02 cargo-completion 14:35:02 cartridge-cli 14:35:02 ceres-solver 14:35:02 certbot 14:35:02 cfn-lint 14:35:02 cfr-decompiler 14:35:02 cgns 14:35:02 cgrep 14:35:02 chakra 14:35:02 chapel 14:35:02 charm-tools 14:35:02 cheat 14:35:02 cli53 14:35:02 clojure 14:35:02 clojure-lsp 14:35:02 clutter 14:35:02 cobalt 14:35:02 cocoapods 14:35:02 codec2 14:35:02 codemod 14:35:02 codespell 14:35:02 cogl 14:35:02 comby 14:35:02 composer 14:35:02 conan 14:35:02 confluent-platform 14:35:02 conjure-up 14:35:02 cookiecutter 14:35:02 cpl 14:35:02 cppman 14:35:02 cquery 14:35:02 csound 14:35:02 csvkit 14:35:02 csvtomd 14:35:02 curl 14:35:02 curl-openssl 14:35:02 cvs 14:35:02 dartsim 14:35:02 davmail 14:35:02 db-vcs 14:35:02 dependency-check 14:35:02 derby 14:35:02 detekt 14:35:02 dialog 14:35:02 diceware 14:35:02 diff-pdf 14:35:02 diffoscope 14:35:02 dmd 14:35:02 docbook-xsl 14:35:02 docbook2x 14:35:02 docfx 14:35:02 docker 14:35:02 docker-completion 14:35:02 docker-compose 14:35:02 docker-slim 14:35:02 docker-squash 14:35:02 dosbox-x 14:35:02 double-conversion 14:35:02 dovecot 14:35:02 dtc 14:35:02 dub 14:35:02 dune 14:35:02 duplicity 14:35:02 dust 14:35:02 dxpy 14:35:02 dynare 14:35:02 e2fsprogs 14:35:02 easyrpg-player 14:35:02 ecasound 14:35:02 eccodes 14:35:02 eg-examples 14:35:02 ejdb 14:35:02 elasticsearch 14:35:02 embree 14:35:02 emscripten 14:35:02 ensmallen 14:35:02 esptool 14:35:02 etsh 14:35:02 evince 14:35:02 exim 14:35:02 expect 14:35:02 exploitdb 14:35:02 ext2fuse 14:35:02 ezstream 14:35:02 fairymax 14:35:02 faiss 14:35:02 fantom 14:35:02 fastlane 14:35:02 faudio 14:35:02 fcl 14:35:02 fdroidserver 14:35:02 file-roller 14:35:02 findomain 14:35:02 flann 14:35:02 flatbuffers 14:35:02 flintrock 14:35:02 fluent-bit 14:35:02 flyway 14:35:02 fn 14:35:02 folly 14:35:02 foma 14:35:02 fontconfig 14:35:02 fonttools 14:35:02 fq 14:35:02 frpc 14:35:02 frps 14:35:02 frugal 14:35:02 fselect 14:35:02 fzf 14:35:02 gambit 14:35:02 gambit-scheme 14:35:02 gammaray 14:35:02 gandi.cli 14:35:02 gatsby-cli 14:35:02 gcc 14:35:02 gcc@8 14:35:02 gdal 14:35:02 gdcm 14:35:02 geeqie 14:35:02 geocode-glib 14:35:02 geos 14:35:02 gerbil-scheme 14:35:02 gettext 14:35:02 gimme-aws-creds 14:35:02 git 14:35:02 git-annex 14:35:02 git-plus 14:35:02 git-review 14:35:02 git-revise 14:35:02 git-standup 14:35:02 gitmoji 14:35:02 gitup 14:35:02 gjs 14:35:02 glances 14:35:02 glib 14:35:02 glog 14:35:02 gmsh 14:35:02 gmt 14:35:02 gmt@5 14:35:02 gobject-introspection 14:35:02 google-sparsehash 14:35:02 gpgme 14:35:02 grin 14:35:02 grin-wallet 14:35:02 gromacs 14:35:02 groovy 14:35:02 groovysdk 14:35:02 gsettings-desktop-schemas 14:35:02 gti 14:35:02 gtkmm 14:35:02 gtksourceview4 14:35:02 hdf5 14:35:02 helm 14:35:02 hexyl 14:35:02 homeassistant-cli 14:35:02 hostess 14:35:02 howdoi 14:35:02 htop 14:35:02 httpie 14:35:02 hugo 14:35:02 hyperkit 14:35:02 ibex 14:35:02 ilmbase 14:35:02 imagemagick 14:35:02 imagemagick@6 14:35:02 imgproxy 14:35:02 inadyn 14:35:02 inlets 14:35:02 instalooter 14:35:02 internetarchive 14:35:02 intltool 14:35:02 io 14:35:02 ios-sim 14:35:02 ipmiutil 14:35:02 ipython 14:35:02 isl 14:35:02 itk 14:35:02 jbig2dec 14:35:02 jed 14:35:02 jenkins 14:35:02 jenkins-job-builder 14:35:02 jenkins-lts 14:35:02 jhead 14:35:02 jhipster 14:35:02 jinja2-cli 14:35:02 joplin 14:35:02 jrnl 14:35:02 jruby 14:35:02 juise 14:35:02 juju 14:35:02 juju-wait 14:35:02 jupyterlab 14:35:02 kallisto 14:35:02 keepkey-agent 14:35:02 khard 14:35:02 kibana 14:35:02 krew 14:35:02 ktoblzcheck 14:35:02 kubeprod 14:35:02 kubernetes-cli 14:35:02 kubeseal 14:35:02 kumo 14:35:02 lablgtk 14:35:02 lammps 14:35:02 ldc 14:35:02 legit 14:35:02 lgogdownloader 14:35:02 libbi 14:35:02 libcouchbase 14:35:02 libdazzle 14:35:02 libewf 14:35:02 libexosip 14:35:02 libfabric 14:35:02 libgetdata 14:35:02 libgosu 14:35:02 libgweather 14:35:02 libhttpseverywhere 14:35:02 libical 14:35:02 libmatio 14:35:02 libowfat 14:35:02 libpeas 14:35:02 libproxy 14:35:02 libpsl 14:35:02 libqalculate 14:35:02 librealsense 14:35:02 libsecret 14:35:02 libserdes 14:35:02 libsoup 14:35:02 libssh 14:35:02 libsvm 14:35:02 libtcod 14:35:02 libtool 14:35:02 libtorrent-rasterbar 14:35:02 libupnp 14:35:02 libwebsockets 14:35:02 libxml++ 14:35:02 libzdb 14:35:02 liquibase 14:35:02 literate-git 14:35:02 llvm 14:35:02 llvm@6 14:35:02 llvm@7 14:35:02 llvm@8 14:35:02 lolcat 14:35:02 lua 14:35:02 lua@5.1 14:35:02 luaradio 14:35:02 luarocks 14:35:02 luit 14:35:02 mackup 14:35:02 magic-wormhole 14:35:02 mapserver 14:35:02 mbedtls 14:35:02 mdds 14:35:02 mdv 14:35:02 memcached 14:35:02 menhir 14:35:02 mercurial 14:35:02 mercury 14:35:02 mesa 14:35:02 meson-internal 14:35:02 metabase 14:35:02 micronaut 14:35:02 minikube 14:35:02 minio 14:35:02 minizinc 14:35:02 minizip2 14:35:02 mitmproxy 14:35:02 mkdocs 14:35:02 mkl-dnn 14:35:02 mkvtoolnix 14:35:02 mlpack 14:35:02 molecule 14:35:02 mongo-cxx-driver 14:35:02 mongo-orchestration 14:35:02 mosh 14:35:02 mosquitto 14:35:02 mpd 14:35:02 mruby 14:35:02 mujs 14:35:02 mycli 14:35:02 mypy 14:35:02 mysql-connector-c++ 14:35:02 n 14:35:02 nagios-plugins 14:35:02 nbdime 14:35:02 nco 14:35:02 ncview 14:35:02 neofetch 14:35:02 neovim 14:35:02 netcdf 14:35:02 nethack 14:35:02 ninja 14:35:02 nng 14:35:02 node-sass 14:35:02 nodeenv 14:35:02 nss 14:35:02 nudoku 14:35:02 nushell 14:35:02 oauth2l 14:35:02 ocaml 14:35:02 ocaml-findlib 14:35:02 ocaml-num 14:35:02 ocamlbuild 14:35:02 ocamlsdl 14:35:02 ocrmypdf 14:35:02 octave 14:35:02 ode 14:35:02 odpi 14:35:02 onnxruntime 14:35:02 open-image-denoise 14:35:02 open-mpi 14:35:02 openexr 14:35:02 openssl@1.1 14:35:02 pastel 14:35:02 pcl 14:35:02 pdal 14:35:02 pgcli 14:35:02 pipenv 14:35:02 pipx 14:35:02 pkg-config 14:35:02 platformio 14:35:02 plotutils 14:35:02 pmd 14:35:02 pnpm 14:35:02 points2grid 14:35:02 pre-commit 14:35:02 procs 14:35:02 prodigal 14:35:02 proselint 14:35:02 protoc-gen-go 14:35:02 psc-package 14:35:02 pssh 14:35:02 pulseaudio 14:35:02 pulumi 14:35:02 pushpin 14:35:02 pwntools 14:35:02 pyenv 14:35:02 pygitup 14:35:02 pygments 14:35:02 pygobject3 14:35:02 pyinvoke 14:35:02 pylint 14:35:02 python 14:35:02 python@3.8 14:35:02 pyvim 14:35:02 qalculate-gtk 14:35:02 qt 14:35:02 rabbitmq 14:35:02 radamsa 14:35:02 radare2 14:35:02 rakudo-star 14:35:02 re-flex 14:35:02 remarshal 14:35:02 rex 14:35:02 rst-lint 14:35:02 rtv 14:35:02 rubberband 14:35:02 rust 14:35:02 s3cmd 14:35:02 s3fs 14:35:02 s6 14:35:02 sbjson 14:35:02 sceptre 14:35:02 scons 14:35:02 scour 14:35:02 sdcc 14:35:02 sdedit 14:35:02 sdl2 14:35:02 serf 14:35:02 serverless 14:35:02 sfk 14:35:02 shogun 14:35:02 simg2img 14:35:02 simple-tiles 14:35:02 sjk 14:35:02 skinny 14:35:02 sn0int 14:35:02 sonar-scanner 14:35:02 sonarqube 14:35:02 sphinx-doc 14:35:02 spice-protocol 14:35:02 spidermonkey 14:35:02 spigot 14:35:02 spirv-cross 14:35:02 spirv-tools 14:35:02 sqliteodbc 14:35:02 sqlparse 14:35:02 ssdb 14:35:02 ssh-audit 14:35:02 sshuttle 14:35:02 statik 14:35:02 stdman 14:35:02 stellar-core 14:35:02 stgit 14:35:02 stolon 14:35:02 streamlink 14:35:02 stress-ng 14:35:02 stubby 14:35:02 subversion 14:35:02 suite-sparse 14:35:02 sundials 14:35:02 supervisor 14:35:02 svtplay-dl 14:35:02 swagger-codegen 14:35:02 swift-protobuf 14:35:02 swift-sh 14:35:02 swiftformat 14:35:02 switch-lan-play 14:35:02 talisman 14:35:02 tasksh 14:35:02 tbox 14:35:02 tektoncd-cli 14:35:02 terragrunt 14:35:02 terrahelp 14:35:02 terrahub 14:35:02 texinfo 14:35:02 texmath 14:35:02 tgui 14:35:02 thefuck 14:35:02 tin 14:35:02 tinyxml 14:35:02 tinyxml2 14:35:02 todoman 14:35:02 tox 14:35:02 trafficserver 14:35:02 translate-toolkit 14:35:02 trash-cli 14:35:02 ttyd 14:35:02 twarc 14:35:02 twine-pypi 14:35:02 txr 14:35:02 txt2tags 14:35:02 u-boot-tools 14:35:02 urh 14:35:02 v2ray-plugin 14:35:02 v8 14:35:02 vala 14:35:02 vale 14:35:02 valgrind 14:35:02 vault 14:35:02 velero 14:35:02 verilator 14:35:02 vips 14:35:02 visp 14:35:02 vowpal-wabbit 14:35:02 vsts-cli 14:35:02 vtk 14:35:02 vttest 14:35:02 vulkan-headers 14:35:02 wabt 14:35:02 wal2json 14:35:02 wapm 14:35:02 wartremover 14:35:02 wasmer 14:35:02 watson 14:35:02 wcslib 14:35:02 webdis 14:35:02 websocat 14:35:02 webtorrent-cli 14:35:02 wp-cli-completion 14:35:02 wpscan 14:35:02 wxmaxima 14:35:02 x3270 14:35:02 xclip 14:35:02 xdotool 14:35:02 xmlto 14:35:02 xmrig 14:35:02 xonsh 14:35:02 xtensor 14:35:02 xxhash 14:35:02 yamllint 14:35:02 yapf 14:35:02 yarn 14:35:02 ydcv 14:35:02 ykman 14:35:02 you-get 14:35:02 youtube-dl 14:35:02 z 14:35:02 zabbix-cli 14:35:02 zeek 14:35:02 zimg 14:35:02 znc 14:35:02 zola 14:35:02 ==> Deleted Formulae 14:35:02 camlp4 14:35:02 ck 14:35:02 djmount 14:35:02 field3d 14:35:02 gmediaserver 14:35:02 14:35:03 ==> Fetching homebrew/master 14:35:06 From https://github.com/Homebrew/homebrew-core 14:35:06 * branch master -> FETCH_HEAD 14:35:06 257bb3e8cb..d3eb71b109 master -> homebrew/master 14:35:06 ==> minikube exists in homebrew/master 14:35:07 Previous HEAD position was a99a5e418f... Merge branch homebrew/master into linuxbrew/master 14:35:07 HEAD is now at d3eb71b109... composer 1.10.1 (#51704) 14:35:07 ==> replace "v1.8.1" with "v1.8.2" 14:35:07 ==> replace "cbda04cf6bbe65e987ae52bb393c10099ab62014" with "eb13446e786c9ef70cb0a9f85a633194e62396a1" 14:35:09 Fetching gem metadata from https://rubygems.org/......... 14:35:10 Using concurrent-ruby 1.1.6 14:35:10 Using minitest 5.14.0 14:35:10 Using thread_safe 0.3.6 14:35:10 Using zeitwerk 2.3.0 14:35:10 Using ast 2.4.0 14:35:10 Using bundler 1.17.2 14:35:10 Using connection_pool 2.2.2 14:35:10 Using json 2.3.0 14:35:10 Using docile 1.3.2 14:35:10 Using simplecov-html 0.10.2 14:35:10 Using sync 0.5.0 14:35:10 Using thor 1.0.1 14:35:10 Using diff-lcs 1.3 14:35:10 Using unf_ext 0.0.7.6 14:35:10 Using hpricot 0.8.6 14:35:10 Using jaro_winkler 1.5.4 14:35:10 Using mime-types-data 3.2019.1009 14:35:10 Using net-http-digest_auth 1.4.1 14:35:10 Using mini_portile2 2.4.0 14:35:10 Using ntlm-http 0.1.1 14:35:10 Using webrobots 0.1.2 14:35:10 Using mustache 1.1.1 14:35:10 Using parallel 1.19.1 14:35:10 Using plist 3.5.0 14:35:10 Using rainbow 3.0.0 14:35:10 Using rdiscount 2.2.0.1 14:35:10 Using rexml 3.2.4 14:35:10 Using rspec-support 3.9.2 14:35:10 Using ruby-progressbar 1.10.1 14:35:10 Using unicode-display_width 1.6.1 14:35:10 Using ruby-macho 2.2.0 14:35:10 Using i18n 1.8.2 14:35:10 Using tzinfo 1.2.6 14:35:10 Using parser 2.7.0.4 14:35:10 Using tins 1.24.1 14:35:10 Using net-http-persistent 3.1.0 14:35:10 Using unf 0.1.4 14:35:10 Using mime-types 3.3.1 14:35:10 Using simplecov 0.16.1 14:35:10 Using ronn 0.7.3 14:35:10 Using rspec-core 3.9.1 14:35:10 Using rspec-expectations 3.9.0 14:35:10 Using rspec-mocks 3.9.1 14:35:10 Using activesupport 6.0.2.1 14:35:10 Using parallel_tests 2.31.0 14:35:10 Using rubocop 0.80.1 14:35:10 Using term-ansicolor 1.7.1 14:35:10 Fetching nokogiri 1.10.9 14:35:10 Using coveralls 0.8.23 14:35:10 Using domain_name 0.5.20190701 14:35:10 Using rspec 3.9.0 14:35:10 Using rspec-its 1.3.0 14:35:10 Using rspec-retry 0.6.2 14:35:10 Using rubocop-performance 1.5.2 14:35:10 Using rubocop-rspec 1.38.1 14:35:10 Using http-cookie 1.0.3 14:35:10 Using rspec-wait 0.0.9 14:35:10 Installing nokogiri 1.10.9 with native extensions 14:35:12 Gem::Ext::BuildError: ERROR: Failed to build gem native extension. 14:35:12 14:35:12 current directory: 14:35:12 /home/linuxbrew/.linuxbrew/Homebrew/Library/Homebrew/vendor/bundle/ruby/2.6.0/gems/nokogiri-1.10.9/ext/nokogiri 14:35:12 /home/linuxbrew/.linuxbrew/Homebrew/Library/Homebrew/vendor/portable-ruby/2.6.3/bin/ruby 14:35:12 -I 14:35:12 /home/linuxbrew/.linuxbrew/Homebrew/Library/Homebrew/vendor/portable-ruby/2.6.3/lib/ruby/2.6.0 14:35:12 -r ./siteconf20200313-30573-18g5kc4.rb extconf.rb 14:35:12 checking if the C compiler accepts ... yes 14:35:12 Building nokogiri using packaged libraries. 14:35:12 Using mini_portile version 2.4.0 14:35:12 checking for gzdopen() in -lz... no 14:35:12 zlib is missing; necessary for building libxml2 14:35:12 *** extconf.rb failed *** 14:35:12 Could not create Makefile due to some reason, probably lack of necessary 14:35:12 libraries and/or headers. Check the mkmf.log file for more details. You may 14:35:12 need configuration options. 14:35:12 14:35:12 Provided configuration options: 14:35:12 --with-opt-dir 14:35:12 --with-opt-include 14:35:12 --without-opt-include=${opt-dir}/include 14:35:12 --with-opt-lib 14:35:12 --without-opt-lib=${opt-dir}/lib 14:35:12 --with-make-prog 14:35:12 --without-make-prog 14:35:12 --srcdir=. 14:35:12 --curdir 14:35:12 --ruby=/home/linuxbrew/.linuxbrew/Homebrew/Library/Homebrew/vendor/portable-ruby/2.6.3/bin/$(RUBY_BASE_NAME) 14:35:12 --help 14:35:12 --clean 14:35:12 --use-system-libraries 14:35:12 --enable-static 14:35:12 --disable-static 14:35:12 --with-zlib-dir 14:35:12 --without-zlib-dir 14:35:12 --with-zlib-include 14:35:12 --without-zlib-include=${zlib-dir}/include 14:35:12 --with-zlib-lib 14:35:12 --without-zlib-lib=${zlib-dir}/lib 14:35:12 --enable-cross-build 14:35:12 --disable-cross-build 14:35:12 14:35:12 To see why this extension failed to compile, please check the mkmf.log which can 14:35:12 be found here: 14:35:12 14:35:12 /home/linuxbrew/.linuxbrew/Homebrew/Library/Homebrew/vendor/bundle/ruby/2.6.0/extensions/x86_64-linux/2.6.0-static/nokogiri-1.10.9/mkmf.log 14:35:12 14:35:12 extconf failed, exit code 1 14:35:12 14:35:12 Gem files will remain installed in 14:35:12 /home/linuxbrew/.linuxbrew/Homebrew/Library/Homebrew/vendor/bundle/ruby/2.6.0/gems/nokogiri-1.10.9 14:35:12 for inspection. 14:35:12 Results logged to 14:35:12 /home/linuxbrew/.linuxbrew/Homebrew/Library/Homebrew/vendor/bundle/ruby/2.6.0/extensions/x86_64-linux/2.6.0-static/nokogiri-1.10.9/gem_make.out 14:35:12 14:35:12 An error occurred while installing nokogiri (1.10.9), and Bundler cannot 14:35:12 continue. 14:35:12 Make sure that `gem install nokogiri -v '1.10.9' --source 14:35:12 'https://rubygems.org/'` succeeds before bundling. 14:35:12 14:35:12 In Gemfile: 14:35:12 mechanize was resolved to 2.7.6, which depends on 14:35:12 nokogiri 14:35:12 Error: failed to run `/home/linuxbrew/.linuxbrew/Homebrew/Library/Homebrew/vendor/portable-ruby/2.6.3/bin/bundle install`! 14:35:12 Error: `brew audit` failed! </pre> </details>
process
release process run brew bump formula pr in docker we need to run this command in a docker so it doesnt die of gem dependency brew bump formula pr strict minikube revision revision message this pr was automatically created by minikube release scripts contact tstromberg with any questions no browse tag tag status status here is the current gem dep error from out latest release brew bump formula pr strict minikube revision message this pr was automatically created by minikube release scripts contact tstromberg with any questions no browse tag updating homebrew auto updated homebrew updated homebrew from to updated tap homebrew core new formulae adoptopenjdk adoptopenjdk alsa lib apt archiver at atk at core attr avahi aws vault bsdmainutils cpio device mapper elfutils gcc git trim glibc hdt isl jdk jdk jdk ladspa sdk libaio libbsd libcap libcap ng libfuse libmnl libnetfilter queue libnfnetlink libnsl libprelude libseccomp libtirpc linux headers linux pam lm sensors minimal python mockolo net tools newrelic cli onedrive openjdk openjdk procps python dbus python qp singularity strace sysstat systemd texlive tlx xclogparser updated formulae abseil advancemenu adwaita icon theme afuse agda angular cli ansible ansible lint apache drill apibuilder cli apr util arangodb armadillo arping artifactory asciidoc asciinema astrometry net atdtool autoconf aws cdk aws elasticbeanstalk aws shell awscli awscli awscurl awslogs awsume azure cli azure storage cpp bacula fd balena cli bandcamp dl bazaar bde benthos biosig bison bitcoin black blazegraph bonnie buildifier bundletool bzt caffe cargo completion cartridge cli ceres solver certbot cfn lint cfr decompiler cgns cgrep chakra chapel charm tools cheat clojure clojure lsp clutter cobalt cocoapods codemod codespell cogl comby composer conan confluent platform conjure up cookiecutter cpl cppman cquery csound csvkit csvtomd curl curl openssl cvs dartsim davmail db vcs dependency check derby detekt dialog diceware diff pdf diffoscope dmd docbook xsl docfx docker docker completion docker compose docker slim docker squash dosbox x double conversion dovecot dtc dub dune duplicity dust dxpy dynare easyrpg player ecasound eccodes eg examples ejdb elasticsearch embree emscripten ensmallen esptool etsh evince exim expect exploitdb ezstream fairymax faiss fantom fastlane faudio fcl fdroidserver file roller findomain flann flatbuffers flintrock fluent bit flyway fn folly foma fontconfig fonttools fq frpc frps frugal fselect fzf gambit gambit scheme gammaray gandi cli gatsby cli gcc gcc gdal gdcm geeqie geocode glib geos gerbil scheme gettext gimme aws creds git git annex git plus git review git revise git standup gitmoji gitup gjs glances glib glog gmsh gmt gmt gobject introspection google sparsehash gpgme grin grin wallet gromacs groovy groovysdk gsettings desktop schemas gti gtkmm helm hexyl homeassistant cli hostess howdoi htop httpie hugo hyperkit ibex ilmbase imagemagick imagemagick imgproxy inadyn inlets instalooter internetarchive intltool io ios sim ipmiutil ipython isl itk jed jenkins jenkins job builder jenkins lts jhead jhipster cli joplin jrnl jruby juise juju juju wait jupyterlab kallisto keepkey agent khard kibana krew ktoblzcheck kubeprod kubernetes cli kubeseal kumo lablgtk lammps ldc legit lgogdownloader libbi libcouchbase libdazzle libewf libexosip libfabric libgetdata libgosu libgweather libhttpseverywhere libical libmatio libowfat libpeas libproxy libpsl libqalculate librealsense libsecret libserdes libsoup libssh libsvm libtcod libtool libtorrent rasterbar libupnp libwebsockets libxml libzdb liquibase literate git llvm llvm llvm llvm lolcat lua lua luaradio luarocks luit mackup magic wormhole mapserver mbedtls mdds mdv memcached menhir mercurial mercury mesa meson internal metabase micronaut minikube minio minizinc mitmproxy mkdocs mkl dnn mkvtoolnix mlpack molecule mongo cxx driver mongo orchestration mosh mosquitto mpd mruby mujs mycli mypy mysql connector c n nagios plugins nbdime nco ncview neofetch neovim netcdf nethack ninja nng node sass nodeenv nss nudoku nushell ocaml ocaml findlib ocaml num ocamlbuild ocamlsdl ocrmypdf octave ode odpi onnxruntime open image denoise open mpi openexr openssl pastel pcl pdal pgcli pipenv pipx pkg config platformio plotutils pmd pnpm pre commit procs prodigal proselint protoc gen go psc package pssh pulseaudio pulumi pushpin pwntools pyenv pygitup pygments pyinvoke pylint python python pyvim qalculate gtk qt rabbitmq radamsa rakudo star re flex remarshal rex rst lint rtv rubberband rust sbjson sceptre scons scour sdcc sdedit serf serverless sfk shogun simple tiles sjk skinny sonar scanner sonarqube sphinx doc spice protocol spidermonkey spigot spirv cross spirv tools sqliteodbc sqlparse ssdb ssh audit sshuttle statik stdman stellar core stgit stolon streamlink stress ng stubby subversion suite sparse sundials supervisor svtplay dl swagger codegen swift protobuf swift sh swiftformat switch lan play talisman tasksh tbox tektoncd cli terragrunt terrahelp terrahub texinfo texmath tgui thefuck tin tinyxml todoman tox trafficserver translate toolkit trash cli ttyd twarc twine pypi txr u boot tools urh plugin vala vale valgrind vault velero verilator vips visp vowpal wabbit vsts cli vtk vttest vulkan headers wabt wapm wartremover wasmer watson wcslib webdis websocat webtorrent cli wp cli completion wpscan wxmaxima xclip xdotool xmlto xmrig xonsh xtensor xxhash yamllint yapf yarn ydcv ykman you get youtube dl z zabbix cli zeek zimg znc zola deleted formulae ck djmount gmediaserver fetching homebrew master from branch master fetch head master homebrew master minikube exists in homebrew master previous head position was merge branch homebrew master into linuxbrew master head is now at composer replace with replace with fetching gem metadata from using concurrent ruby using minitest using thread safe using zeitwerk using ast using bundler using connection pool using json using docile using simplecov html using sync using thor using diff lcs using unf ext using hpricot using jaro winkler using mime types data using net http digest auth using mini using ntlm http using webrobots using mustache using parallel using plist using rainbow using rdiscount using rexml using rspec support using ruby progressbar using unicode display width using ruby macho using using tzinfo using parser using tins using net http persistent using unf using mime types using simplecov using ronn using rspec core using rspec expectations using rspec mocks using activesupport using parallel tests using rubocop using term ansicolor fetching nokogiri using coveralls using domain name using rspec using rspec its using rspec retry using rubocop performance using rubocop rspec using http cookie using rspec wait installing nokogiri with native extensions gem ext builderror error failed to build gem native extension current directory home linuxbrew linuxbrew homebrew library homebrew vendor bundle ruby gems nokogiri ext nokogiri home linuxbrew linuxbrew homebrew library homebrew vendor portable ruby bin ruby i home linuxbrew linuxbrew homebrew library homebrew vendor portable ruby lib ruby r rb extconf rb checking if the c compiler accepts yes building nokogiri using packaged libraries using mini portile version checking for gzdopen in lz no zlib is missing necessary for building extconf rb failed could not create makefile due to some reason probably lack of necessary libraries and or headers check the mkmf log file for more details you may need configuration options provided configuration options with opt dir with opt include without opt include opt dir include with opt lib without opt lib opt dir lib with make prog without make prog srcdir curdir ruby home linuxbrew linuxbrew homebrew library homebrew vendor portable ruby bin ruby base name help clean use system libraries enable static disable static with zlib dir without zlib dir with zlib include without zlib include zlib dir include with zlib lib without zlib lib zlib dir lib enable cross build disable cross build to see why this extension failed to compile please check the mkmf log which can be found here home linuxbrew linuxbrew homebrew library homebrew vendor bundle ruby extensions linux static nokogiri mkmf log extconf failed exit code gem files will remain installed in home linuxbrew linuxbrew homebrew library homebrew vendor bundle ruby gems nokogiri for inspection results logged to home linuxbrew linuxbrew homebrew library homebrew vendor bundle ruby extensions linux static nokogiri gem make out an error occurred while installing nokogiri and bundler cannot continue make sure that gem install nokogiri v source succeeds before bundling in gemfile mechanize was resolved to which depends on nokogiri error failed to run home linuxbrew linuxbrew homebrew library homebrew vendor portable ruby bin bundle install error brew audit failed
1
70,877
18,289,823,649
IssuesEvent
2021-10-05 14:11:27
google/or-tools
https://api.github.com/repos/google/or-tools
closed
Fedora 34 build: `CMake Error at cmake/FindSCIP.cmake:68 (set_property): set_property not given a PROPERTY <name> argument.`
Bug Build: CMake Solver: SCIP
**What version of OR-Tools and what language are you using?** Version: stable Language: C++ **Which solver are you using (e.g. CP-SAT, Routing Solver, GLOP, BOP, Gurobi)** SCIP **What operating system (Linux, Windows, ...) and version?** Fedora 34, cmake version 3.20.5 **What did you do?** Steps to reproduce the behavior: I tried to build the `stable` branch like so: ``` git clone https://github.com/google/or-tools.git && cd or-tools mkdir build && cd build && SCIP_ROOT=/usr/local/ cmake .. ``` **What did you expect to see** A successful build **What did you see instead?** The build fails with the following error messages: <pre> -- SCIP_ROOT: /usr/local/ -- TARGET_PROPERTY: CMake Error at cmake/FindSCIP.cmake:68 (set_property): <b>set_property not given a PROPERTY <name> argument. </b> Call Stack (most recent call first): cmake/deps.cmake:53 (find_package) CMakeLists.txt:246 (include) </pre> Note: I posted this issue as a problem about linking the pthreads library before. I edited it now, because I think the issue is in fact a different one - namely the Error above.
1.0
Fedora 34 build: `CMake Error at cmake/FindSCIP.cmake:68 (set_property): set_property not given a PROPERTY <name> argument.` - **What version of OR-Tools and what language are you using?** Version: stable Language: C++ **Which solver are you using (e.g. CP-SAT, Routing Solver, GLOP, BOP, Gurobi)** SCIP **What operating system (Linux, Windows, ...) and version?** Fedora 34, cmake version 3.20.5 **What did you do?** Steps to reproduce the behavior: I tried to build the `stable` branch like so: ``` git clone https://github.com/google/or-tools.git && cd or-tools mkdir build && cd build && SCIP_ROOT=/usr/local/ cmake .. ``` **What did you expect to see** A successful build **What did you see instead?** The build fails with the following error messages: <pre> -- SCIP_ROOT: /usr/local/ -- TARGET_PROPERTY: CMake Error at cmake/FindSCIP.cmake:68 (set_property): <b>set_property not given a PROPERTY <name> argument. </b> Call Stack (most recent call first): cmake/deps.cmake:53 (find_package) CMakeLists.txt:246 (include) </pre> Note: I posted this issue as a problem about linking the pthreads library before. I edited it now, because I think the issue is in fact a different one - namely the Error above.
non_process
fedora build cmake error at cmake findscip cmake set property set property not given a property argument what version of or tools and what language are you using version stable language c which solver are you using e g cp sat routing solver glop bop gurobi scip what operating system linux windows and version fedora cmake version what did you do steps to reproduce the behavior i tried to build the stable branch like so git clone cd or tools mkdir build cd build scip root usr local cmake what did you expect to see a successful build what did you see instead the build fails with the following error messages scip root usr local target property cmake error at cmake findscip cmake set property set property not given a property argument call stack most recent call first cmake deps cmake find package cmakelists txt include note i posted this issue as a problem about linking the pthreads library before i edited it now because i think the issue is in fact a different one namely the error above
0
273,754
29,831,080,086
IssuesEvent
2023-06-18 09:28:42
RG4421/ampere-centos-kernel
https://api.github.com/repos/RG4421/ampere-centos-kernel
closed
CVE-2019-11599 (High) detected in linuxv5.2 - autoclosed
Mend: dependency security vulnerability
## CVE-2019-11599 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv5.2</b></p></summary> <p> <p>Linux kernel source tree</p> <p>Library home page: <a href=https://github.com/torvalds/linux.git>https://github.com/torvalds/linux.git</a></p> <p>Found in base branch: <b>amp-centos-8.0-kernel</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/proc/task_mmu.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/proc/task_mmu.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> The coredump implementation in the Linux kernel before 5.0.10 does not use locking or other mechanisms to prevent vma layout or vma flags changes while it runs, which allows local users to obtain sensitive information, cause a denial of service, or possibly have unspecified other impact by triggering a race condition with mmget_not_zero or get_task_mm calls. This is related to fs/userfaultfd.c, mm/mmap.c, fs/proc/task_mmu.c, and drivers/infiniband/core/uverbs_main.c. <p>Publish Date: 2019-04-29 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-11599>CVE-2019-11599</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.0</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11599">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11599</a></p> <p>Release Date: 2019-04-29</p> <p>Fix Resolution: v5.1-rc6</p> </p> </details> <p></p>
True
CVE-2019-11599 (High) detected in linuxv5.2 - autoclosed - ## CVE-2019-11599 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv5.2</b></p></summary> <p> <p>Linux kernel source tree</p> <p>Library home page: <a href=https://github.com/torvalds/linux.git>https://github.com/torvalds/linux.git</a></p> <p>Found in base branch: <b>amp-centos-8.0-kernel</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/proc/task_mmu.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/proc/task_mmu.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> The coredump implementation in the Linux kernel before 5.0.10 does not use locking or other mechanisms to prevent vma layout or vma flags changes while it runs, which allows local users to obtain sensitive information, cause a denial of service, or possibly have unspecified other impact by triggering a race condition with mmget_not_zero or get_task_mm calls. This is related to fs/userfaultfd.c, mm/mmap.c, fs/proc/task_mmu.c, and drivers/infiniband/core/uverbs_main.c. <p>Publish Date: 2019-04-29 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-11599>CVE-2019-11599</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.0</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11599">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11599</a></p> <p>Release Date: 2019-04-29</p> <p>Fix Resolution: v5.1-rc6</p> </p> </details> <p></p>
non_process
cve high detected in autoclosed cve high severity vulnerability vulnerable library linux kernel source tree library home page a href found in base branch amp centos kernel vulnerable source files fs proc task mmu c fs proc task mmu c vulnerability details the coredump implementation in the linux kernel before does not use locking or other mechanisms to prevent vma layout or vma flags changes while it runs which allows local users to obtain sensitive information cause a denial of service or possibly have unspecified other impact by triggering a race condition with mmget not zero or get task mm calls this is related to fs userfaultfd c mm mmap c fs proc task mmu c and drivers infiniband core uverbs main c publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution
0
20,255
26,874,022,498
IssuesEvent
2023-02-04 20:31:17
SatDump/SatDump
https://api.github.com/repos/SatDump/SatDump
opened
L1b product quality flags
enhancement Processing
We need quality flags to make any L2 products, or trust the generated products for research purposes. - [ ] AVHRR - [ ] AMSU-A - [ ] MHS
1.0
L1b product quality flags - We need quality flags to make any L2 products, or trust the generated products for research purposes. - [ ] AVHRR - [ ] AMSU-A - [ ] MHS
process
product quality flags we need quality flags to make any products or trust the generated products for research purposes avhrr amsu a mhs
1
21,923
30,446,460,024
IssuesEvent
2023-07-15 18:31:24
h4sh5/pypi-auto-scanner
https://api.github.com/repos/h4sh5/pypi-auto-scanner
opened
pyutils 0.0.1b18 has 2 GuardDog issues
guarddog typosquatting silent-process-execution
https://pypi.org/project/pyutils https://inspector.pypi.io/project/pyutils ```{ "dependency": "pyutils", "version": "0.0.1b18", "result": { "issues": 2, "errors": {}, "results": { "typosquatting": "This package closely ressembles the following package names, and might be a typosquatting attempt: pytils, python-utils", "silent-process-execution": [ { "location": "pyutils/exec_utils.py/pyutils/exec_utils.py:214", "code": " subproc = subprocess.Popen(\n args,\n stdin=subprocess.DEVNULL,\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n )", "message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null" } ] }, "path": "/tmp/tmpooezcz_t/pyutils" } }```
1.0
pyutils 0.0.1b18 has 2 GuardDog issues - https://pypi.org/project/pyutils https://inspector.pypi.io/project/pyutils ```{ "dependency": "pyutils", "version": "0.0.1b18", "result": { "issues": 2, "errors": {}, "results": { "typosquatting": "This package closely ressembles the following package names, and might be a typosquatting attempt: pytils, python-utils", "silent-process-execution": [ { "location": "pyutils/exec_utils.py/pyutils/exec_utils.py:214", "code": " subproc = subprocess.Popen(\n args,\n stdin=subprocess.DEVNULL,\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n )", "message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null" } ] }, "path": "/tmp/tmpooezcz_t/pyutils" } }```
process
pyutils has guarddog issues dependency pyutils version result issues errors results typosquatting this package closely ressembles the following package names and might be a typosquatting attempt pytils python utils silent process execution location pyutils exec utils py pyutils exec utils py code subproc subprocess popen n args n stdin subprocess devnull n stdout subprocess devnull n stderr subprocess devnull n message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null path tmp tmpooezcz t pyutils
1
1,701
4,349,710,064
IssuesEvent
2016-07-30 19:07:13
pwittchen/ReactiveNetwork
https://api.github.com/repos/pwittchen/ReactiveNetwork
closed
Release 0.5.1
release process
**Initial release notes**: - bumped RxJava to v. 1.1.8 **Things to do**: - [x] bump library version to 0.5.1 - [x] upload Archives to Maven Central Repository - [x] close and release artifact on Nexus - [x] update `CHANGELOG.md` after Maven Sync - [x] bump library version in `README.md` after Maven Sync - [x] create new GitHub release
1.0
Release 0.5.1 - **Initial release notes**: - bumped RxJava to v. 1.1.8 **Things to do**: - [x] bump library version to 0.5.1 - [x] upload Archives to Maven Central Repository - [x] close and release artifact on Nexus - [x] update `CHANGELOG.md` after Maven Sync - [x] bump library version in `README.md` after Maven Sync - [x] create new GitHub release
process
release initial release notes bumped rxjava to v things to do bump library version to upload archives to maven central repository close and release artifact on nexus update changelog md after maven sync bump library version in readme md after maven sync create new github release
1
13,251
15,722,798,243
IssuesEvent
2021-03-29 06:28:10
Ultimate-Hosts-Blacklist/whitelist
https://api.github.com/repos/Ultimate-Hosts-Blacklist/whitelist
closed
[FALSE-POSITIVE?] fossbytes.com
whitelisting process
**Domains or links** - fossbytes.com **More Information** ``` Exact match for fossbytes.com found in: - https://raw.githubusercontent.com/Ultimate-Hosts-Blacklist/Ultimate.Hosts.Blacklist/master/domains/domains0.list ``` **Have you requested removal from other sources?** Nope. **Additional context** N/A, but happy to provide if necessary
1.0
[FALSE-POSITIVE?] fossbytes.com - **Domains or links** - fossbytes.com **More Information** ``` Exact match for fossbytes.com found in: - https://raw.githubusercontent.com/Ultimate-Hosts-Blacklist/Ultimate.Hosts.Blacklist/master/domains/domains0.list ``` **Have you requested removal from other sources?** Nope. **Additional context** N/A, but happy to provide if necessary
process
fossbytes com domains or links fossbytes com more information exact match for fossbytes com found in have you requested removal from other sources nope additional context n a but happy to provide if necessary
1
9,894
12,891,758,134
IssuesEvent
2020-07-13 18:19:21
google/ground-android
https://api.github.com/repos/google/ground-android
opened
[UXR] Research open questions during pilot
priority: p2 type: process
- [ ] Is context of view/add/edit observation clear (i.e. which layer and which point is selected), or do we need breadcrumbs, or some other affordance in the header, etc.? @sergeydolgov1 please add other open questions above.
1.0
[UXR] Research open questions during pilot - - [ ] Is context of view/add/edit observation clear (i.e. which layer and which point is selected), or do we need breadcrumbs, or some other affordance in the header, etc.? @sergeydolgov1 please add other open questions above.
process
research open questions during pilot is context of view add edit observation clear i e which layer and which point is selected or do we need breadcrumbs or some other affordance in the header etc please add other open questions above
1