Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
4
112
repo_url
stringlengths
33
141
action
stringclasses
3 values
title
stringlengths
1
1.02k
labels
stringlengths
4
1.54k
body
stringlengths
1
262k
index
stringclasses
17 values
text_combine
stringlengths
95
262k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
19,762
26,135,599,237
IssuesEvent
2022-12-29 11:40:54
firebase/firebase-cpp-sdk
https://api.github.com/repos/firebase/firebase-cpp-sdk
closed
[C++] Nightly Integration Testing Report for Firestore
type: process nightly-testing
<hidden value="integration-test-status-comment"></hidden> ### [build against repo] Integration test with FLAKINESS (succeeded after retry) Requested by @sunmou99 on commit b07793ae015b4a69f2ec68e1c8f46206f9fac0c7 Last updated: Wed Dec 28 10:56 PST 2022 **[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/3793209772)** | Failures | Configs | |----------|---------| | firestore | [TEST] [FLAKINESS] [Android] [1/3 os: macos] [1/4 android_device: android_latest]<details><summary>(1 failed tests)</summary>&nbsp;&nbsp;NumericTransformsTest.DoubleIncrementWithExistingDouble</details> | Add flaky tests to **[go/fpl-cpp-flake-tracker](http://go/fpl-cpp-flake-tracker)** <hidden value="integration-test-status-comment"></hidden> *** ### βœ…&nbsp; [build against SDK] Integration test succeeded! Requested by @firebase-workflow-trigger[bot] on commit b07793ae015b4a69f2ec68e1c8f46206f9fac0c7 Last updated: Wed Dec 28 10:56 PST 2022 **[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/3793839941)** <hidden value="integration-test-status-comment"></hidden> *** ### βœ…&nbsp; [build against tip] Integration test succeeded! Requested by @sunmou99 on commit b07793ae015b4a69f2ec68e1c8f46206f9fac0c7 Last updated: Wed Dec 28 10:55 PST 2022 **[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/3793510600)**
1.0
[C++] Nightly Integration Testing Report for Firestore - <hidden value="integration-test-status-comment"></hidden> ### [build against repo] Integration test with FLAKINESS (succeeded after retry) Requested by @sunmou99 on commit b07793ae015b4a69f2ec68e1c8f46206f9fac0c7 Last updated: Wed Dec 28 10:56 PST 2022 **[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/3793209772)** | Failures | Configs | |----------|---------| | firestore | [TEST] [FLAKINESS] [Android] [1/3 os: macos] [1/4 android_device: android_latest]<details><summary>(1 failed tests)</summary>&nbsp;&nbsp;NumericTransformsTest.DoubleIncrementWithExistingDouble</details> | Add flaky tests to **[go/fpl-cpp-flake-tracker](http://go/fpl-cpp-flake-tracker)** <hidden value="integration-test-status-comment"></hidden> *** ### βœ…&nbsp; [build against SDK] Integration test succeeded! Requested by @firebase-workflow-trigger[bot] on commit b07793ae015b4a69f2ec68e1c8f46206f9fac0c7 Last updated: Wed Dec 28 10:56 PST 2022 **[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/3793839941)** <hidden value="integration-test-status-comment"></hidden> *** ### βœ…&nbsp; [build against tip] Integration test succeeded! Requested by @sunmou99 on commit b07793ae015b4a69f2ec68e1c8f46206f9fac0c7 Last updated: Wed Dec 28 10:55 PST 2022 **[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/3793510600)**
non_test
nightly integration testing report for firestore integration test with flakiness succeeded after retry requested by on commit last updated wed dec pst failures configs firestore failed tests nbsp nbsp numerictransformstest doubleincrementwithexistingdouble add flaky tests to βœ… nbsp integration test succeeded requested by firebase workflow trigger on commit last updated wed dec pst βœ… nbsp integration test succeeded requested by on commit last updated wed dec pst
0
388,476
26,767,984,371
IssuesEvent
2023-01-31 12:06:30
readthedocs/readthedocs.org
https://api.github.com/repos/readthedocs/readthedocs.org
opened
Docs: explain how the" standard build outputs" work
Needed: documentation Accepted
After implementing #9888 we have improved how our build process works. We need to expand the documentation to reflect these changes. There are good examples in https://github.com/readthedocs/readthedocs.org/issues/1939#issuecomment-1410230381 and other issues. Also, we have to update our "Downloadable documentation" page: https://docs.readthedocs.io/en/stable/downloadable-documentation.html
1.0
Docs: explain how the" standard build outputs" work - After implementing #9888 we have improved how our build process works. We need to expand the documentation to reflect these changes. There are good examples in https://github.com/readthedocs/readthedocs.org/issues/1939#issuecomment-1410230381 and other issues. Also, we have to update our "Downloadable documentation" page: https://docs.readthedocs.io/en/stable/downloadable-documentation.html
non_test
docs explain how the standard build outputs work after implementing we have improved how our build process works we need to expand the documentation to reflect these changes there are good examples in and other issues also we have to update our downloadable documentation page
0
248,718
7,935,293,824
IssuesEvent
2018-07-09 04:04:15
architecture-building-systems/CityEnergyAnalyst
https://api.github.com/repos/architecture-building-systems/CityEnergyAnalyst
closed
Maybe a BUG: electricity demand for datacenter might be too high
Priority 1 not a bug but...
The 500 W/m2 power for `SERVERRROM` in the archetypes DB might be too high. It causes the BAU 16 in the Zug case study to overheat already in winter. Heating is not needed at all in that building. See also the comments in #1189 ![image](https://user-images.githubusercontent.com/17002288/38483631-3e59fd02-3c06-11e8-9ac7-06f12eaa9712.png) ![image](https://user-images.githubusercontent.com/17002288/38483697-74d28926-3c06-11e8-8042-259af6d87f6d.png) @JIMENOFONSECA can I assign this to you? Or does @martin-mosteiro have any measured values from ETH?
1.0
Maybe a BUG: electricity demand for datacenter might be too high - The 500 W/m2 power for `SERVERRROM` in the archetypes DB might be too high. It causes the BAU 16 in the Zug case study to overheat already in winter. Heating is not needed at all in that building. See also the comments in #1189 ![image](https://user-images.githubusercontent.com/17002288/38483631-3e59fd02-3c06-11e8-9ac7-06f12eaa9712.png) ![image](https://user-images.githubusercontent.com/17002288/38483697-74d28926-3c06-11e8-8042-259af6d87f6d.png) @JIMENOFONSECA can I assign this to you? Or does @martin-mosteiro have any measured values from ETH?
non_test
maybe a bug electricity demand for datacenter might be too high the w power for serverrrom in the archetypes db might be too high it causes the bau in the zug case study to overheat already in winter heating is not needed at all in that building see also the comments in jimenofonseca can i assign this to you or does martin mosteiro have any measured values from eth
0
209,867
16,064,173,549
IssuesEvent
2021-04-23 16:25:26
godotengine/godot
https://api.github.com/repos/godotengine/godot
closed
Software crash when merging one Kinematicbody into another scene
bug crash needs testing topic:editor
**Godot version:** stable 3.2.3 **OS/device including version:** Windows 10 Professional Device name DESKTOP-T3MH083 Prozessor AMD Ryzen 5 2600 Six-Core Processor 3.40 GHz Installed RAM 16.0 GB Device ID CE860F4C-F151-429D-B0F8-52E69AF2E314 Product ID 00330-50000-00000-AAOEM System type 64-bit operating system, x64-based processor Pen and touch input No pen or touch input is available for this display. **Issue description:** When I wanted to add a 3D Kinematicbody with some subnodes to another scene via the "merge from scene" function, Godot suddenly crashed. This has happened many times lately.
1.0
Software crash when merging one Kinematicbody into another scene - **Godot version:** stable 3.2.3 **OS/device including version:** Windows 10 Professional Device name DESKTOP-T3MH083 Prozessor AMD Ryzen 5 2600 Six-Core Processor 3.40 GHz Installed RAM 16.0 GB Device ID CE860F4C-F151-429D-B0F8-52E69AF2E314 Product ID 00330-50000-00000-AAOEM System type 64-bit operating system, x64-based processor Pen and touch input No pen or touch input is available for this display. **Issue description:** When I wanted to add a 3D Kinematicbody with some subnodes to another scene via the "merge from scene" function, Godot suddenly crashed. This has happened many times lately.
test
software crash when merging one kinematicbody into another scene godot version stable os device including version windows professional device name desktop prozessor amd ryzen six core processor ghz installed ram gb device id product id aaoem system type bit operating system based processor pen and touch input no pen or touch input is available for this display issue description when i wanted to add a kinematicbody with some subnodes to another scene via the merge from scene function godot suddenly crashed this has happened many times lately
1
21,917
11,425,095,422
IssuesEvent
2020-02-03 19:07:38
tensorflow/tensorflow
https://api.github.com/repos/tensorflow/tensorflow
closed
Training loop freezes and then stops when try to run Cycle-GAN on multi-GPU system on Google GCP using 'tf.distribute.MirroredStrategy()' .
TF 2.0 comp:dist-strat type:performance
<em>Please make sure that this is a bug. As per our [GitHub Policy](https://github.com/tensorflow/tensorflow/blob/master/ISSUES.md), we only address code/doc bugs, performance issues, feature requests and build/installation issues on GitHub. tag:bug_template</em> **System information** - Have I written custom code (as opposed to using a stock example script provided in TensorFlow):Yes - OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Linux Ubuntu 18.04 - Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device:NaN - TensorFlow installed from (source or binary): source - TensorFlow version (use command below): 2.0.0 - Python version:3.6 - Bazel version (if compiling from source): NaN - GCC/Compiler version (if compiling from source): NaN - CUDA/cuDNN version: 10 - GPU model and memory: 4 Nvidia Tesla K80's of 11GB each You can collect some of this information using our environment capture [script](https://github.com/tensorflow/tensorflow/tree/master/tools/tf_env_collect.sh) You can also obtain the TensorFlow version with: 1. TF 1.0: `python -c "import tensorflow as tf; print(tf.GIT_VERSION, tf.VERSION)"` 2. TF 2.0: `python -c "import tensorflow as tf; print(tf.version.GIT_VERSION, tf.version.VERSION)"` **Describe the current behavior** Training loop stops after few steps in the first epoch itself. **Describe the expected behavior** Training loop should run smoothly. **Other info / logs** The link below contains a zip file in which the python scripts which I used to train this model on Multi GPU using 'tf.distribute.MirroredStrategy()' is stored. [The python scripts can be found here](https://drive.google.com/file/d/17TBNfs1h0Lnolq5ZaomqE9pTzdoGjEWb/view?usp=sharing) > Running 'download.py' will download the dataset that I used. > 'model.py' contains code for the architecture of the generator and the discriminator that I used for training this Cycle-GAN > 'utils.py' contains code that I used for preprocessing the images > 'main.py' contains the code in which I used 'tf.distribute.MirroredStrategy()' for multi-GPU training but it stops after few steps in the first epoch.
True
Training loop freezes and then stops when try to run Cycle-GAN on multi-GPU system on Google GCP using 'tf.distribute.MirroredStrategy()' . - <em>Please make sure that this is a bug. As per our [GitHub Policy](https://github.com/tensorflow/tensorflow/blob/master/ISSUES.md), we only address code/doc bugs, performance issues, feature requests and build/installation issues on GitHub. tag:bug_template</em> **System information** - Have I written custom code (as opposed to using a stock example script provided in TensorFlow):Yes - OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Linux Ubuntu 18.04 - Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device:NaN - TensorFlow installed from (source or binary): source - TensorFlow version (use command below): 2.0.0 - Python version:3.6 - Bazel version (if compiling from source): NaN - GCC/Compiler version (if compiling from source): NaN - CUDA/cuDNN version: 10 - GPU model and memory: 4 Nvidia Tesla K80's of 11GB each You can collect some of this information using our environment capture [script](https://github.com/tensorflow/tensorflow/tree/master/tools/tf_env_collect.sh) You can also obtain the TensorFlow version with: 1. TF 1.0: `python -c "import tensorflow as tf; print(tf.GIT_VERSION, tf.VERSION)"` 2. TF 2.0: `python -c "import tensorflow as tf; print(tf.version.GIT_VERSION, tf.version.VERSION)"` **Describe the current behavior** Training loop stops after few steps in the first epoch itself. **Describe the expected behavior** Training loop should run smoothly. **Other info / logs** The link below contains a zip file in which the python scripts which I used to train this model on Multi GPU using 'tf.distribute.MirroredStrategy()' is stored. [The python scripts can be found here](https://drive.google.com/file/d/17TBNfs1h0Lnolq5ZaomqE9pTzdoGjEWb/view?usp=sharing) > Running 'download.py' will download the dataset that I used. > 'model.py' contains code for the architecture of the generator and the discriminator that I used for training this Cycle-GAN > 'utils.py' contains code that I used for preprocessing the images > 'main.py' contains the code in which I used 'tf.distribute.MirroredStrategy()' for multi-GPU training but it stops after few steps in the first epoch.
non_test
training loop freezes and then stops when try to run cycle gan on multi gpu system on google gcp using tf distribute mirroredstrategy please make sure that this is a bug as per our we only address code doc bugs performance issues feature requests and build installation issues on github tag bug template system information have i written custom code as opposed to using a stock example script provided in tensorflow yes os platform and distribution e g linux ubuntu linux ubuntu mobile device e g iphone pixel samsung galaxy if the issue happens on mobile device nan tensorflow installed from source or binary source tensorflow version use command below python version bazel version if compiling from source nan gcc compiler version if compiling from source nan cuda cudnn version gpu model and memory nvidia tesla s of each you can collect some of this information using our environment capture you can also obtain the tensorflow version with tf python c import tensorflow as tf print tf git version tf version tf python c import tensorflow as tf print tf version git version tf version version describe the current behavior training loop stops after few steps in the first epoch itself describe the expected behavior training loop should run smoothly other info logs the link below contains a zip file in which the python scripts which i used to train this model on multi gpu using tf distribute mirroredstrategy is stored running download py will download the dataset that i used model py contains code for the architecture of the generator and the discriminator that i used for training this cycle gan utils py contains code that i used for preprocessing the images main py contains the code in which i used tf distribute mirroredstrategy for multi gpu training but it stops after few steps in the first epoch
0
205,494
23,340,552,684
IssuesEvent
2022-08-09 13:43:22
elastic/kibana
https://api.github.com/repos/elastic/kibana
closed
[Security Solution]Add HTTP header toggle remains enabled with zero header key value pair
bug impact:medium Team:Threat Hunting Team: SecuritySolution Team:Threat Hunting:Explore
**Describe the bug** `Add HTTP header` toggle remains enabled with zero header key value pair **Build Details:** ``` Version:8.4.0 BC1 Commit:58f7eaf0f8dc3c43cbfcd393e587f155e97b3d0d Build:54999 ``` **Steps** - Go to Stack Management - Select Webhook-Case Management connector - Fill in below details - connector name - enable authentication and fill in the correct auth - enable HTTP header toggle and add one header key value pair - content-type:application/json - now remove this by clicking on minus icon - Observed that Add HTTP header toggle remains enabled with zero key value pair and also user can navigate to next step of connector form - Suggestion: - Either restrict user to have at least one key value pair when toggle is enabled and disable the minus icon for that first key value field - Turn Off the `Add HTTP Header` toggle as soon as user deletes all the header key value pair **Scre https://user-images.githubusercontent.com/59917825/182149124-9d7c5e3b-e60b-4c20-ae58-daa476928897.mp4 en-Cast**
True
[Security Solution]Add HTTP header toggle remains enabled with zero header key value pair - **Describe the bug** `Add HTTP header` toggle remains enabled with zero header key value pair **Build Details:** ``` Version:8.4.0 BC1 Commit:58f7eaf0f8dc3c43cbfcd393e587f155e97b3d0d Build:54999 ``` **Steps** - Go to Stack Management - Select Webhook-Case Management connector - Fill in below details - connector name - enable authentication and fill in the correct auth - enable HTTP header toggle and add one header key value pair - content-type:application/json - now remove this by clicking on minus icon - Observed that Add HTTP header toggle remains enabled with zero key value pair and also user can navigate to next step of connector form - Suggestion: - Either restrict user to have at least one key value pair when toggle is enabled and disable the minus icon for that first key value field - Turn Off the `Add HTTP Header` toggle as soon as user deletes all the header key value pair **Scre https://user-images.githubusercontent.com/59917825/182149124-9d7c5e3b-e60b-4c20-ae58-daa476928897.mp4 en-Cast**
non_test
add http header toggle remains enabled with zero header key value pair describe the bug add http header toggle remains enabled with zero header key value pair build details version commit build steps go to stack management select webhook case management connector fill in below details connector name enable authentication and fill in the correct auth enable http header toggle and add one header key value pair content type application json now remove this by clicking on minus icon observed that add http header toggle remains enabled with zero key value pair and also user can navigate to next step of connector form suggestion either restrict user to have at least one key value pair when toggle is enabled and disable the minus icon for that first key value field turn off the add http header toggle as soon as user deletes all the header key value pair scre en cast
0
126,758
26,909,537,172
IssuesEvent
2023-02-06 22:09:18
MetaMask/design-tokens
https://api.github.com/repos/MetaMask/design-tokens
opened
[Ext] Insight Report: Tooltip
code design-system
### **Description** Fill out the `Tooltip` insight report from your findings from the audit The insight report will be part of our decision making framework and is intend to: - Document all findings from the component audit - Confirm as many component details as possible - Mitigate component inconsistencies across Figma, Mobile and Extension Include your thoughts on component name, description, api and any comments or topic to discuss relating to the component or it's make up. We will review the audit and insight report in our Wednesday technical sync to finalize the details of the component for all platforms. ### **Technical Details** The insight report should include the following for the component - name - description - variants/props - requirements (optional) - discussion/questions (optional) ### **Acceptance Criteria** - name, description variants/props is filled out and matches with other platforms where possible ### **References** - [FigJam](https://www.figma.com/file/KP10I7OHiuUsGZ53xAgZTj/Popover-Audit?node-id=0%3A1&t=u3rGtzERyA5zkSPu-1) - Read exercised `#05 Identify Existing Paradigms in Design and Code` and `#06 IdentifyEmergingandInteresting Paradigms in Design and Code` in the Design System in 90 Days workbook
1.0
[Ext] Insight Report: Tooltip - ### **Description** Fill out the `Tooltip` insight report from your findings from the audit The insight report will be part of our decision making framework and is intend to: - Document all findings from the component audit - Confirm as many component details as possible - Mitigate component inconsistencies across Figma, Mobile and Extension Include your thoughts on component name, description, api and any comments or topic to discuss relating to the component or it's make up. We will review the audit and insight report in our Wednesday technical sync to finalize the details of the component for all platforms. ### **Technical Details** The insight report should include the following for the component - name - description - variants/props - requirements (optional) - discussion/questions (optional) ### **Acceptance Criteria** - name, description variants/props is filled out and matches with other platforms where possible ### **References** - [FigJam](https://www.figma.com/file/KP10I7OHiuUsGZ53xAgZTj/Popover-Audit?node-id=0%3A1&t=u3rGtzERyA5zkSPu-1) - Read exercised `#05 Identify Existing Paradigms in Design and Code` and `#06 IdentifyEmergingandInteresting Paradigms in Design and Code` in the Design System in 90 Days workbook
non_test
insight report tooltip description fill out the tooltip insight report from your findings from the audit the insight report will be part of our decision making framework and is intend to document all findings from the component audit confirm as many component details as possible mitigate component inconsistencies across figma mobile and extension include your thoughts on component name description api and any comments or topic to discuss relating to the component or it s make up we will review the audit and insight report in our wednesday technical sync to finalize the details of the component for all platforms technical details the insight report should include the following for the component name description variants props requirements optional discussion questions optional acceptance criteria name description variants props is filled out and matches with other platforms where possible references read exercised identify existing paradigms in design and code and identifyemergingandinteresting paradigms in design and code in the design system in days workbook
0
820,802
30,789,669,914
IssuesEvent
2023-07-31 15:20:01
RobotLocomotion/drake
https://api.github.com/repos/RobotLocomotion/drake
closed
System::Clone has incorrect SystemIDs for the ports
type: bug priority: medium component: system framework
### What happened? Initially reported by @AlexandreAmice ... the following test ``` GTEST_TEST(PendulumPlantTest, Clone) { const PendulumPlant<double> plant; auto clone = plant.Clone(); auto context = clone->CreateDefaultContext(); clone->get_input_port(0).FixValue(context.get(), Vector1d::Zero()); } ``` fails with ``` C++ exception with description "InputPort: The Context given as an argument was not created for this InputPort[0] (tau) of System ::_ (PendulumPlant<double>)" thrown in the test body. ``` Digging around in lldb (with a breakpoint on the last line), I see ``` (lldb) p plant.get_system_id() (drake::systems::internal::SystemId) $5 = (value_ = 6) (lldb) p clone->get_system_id() (drake::systems::internal::SystemId) $6 = (value_ = 9) (lldb) p plant.CreateDefaultContext()->get_system_id() (drake::systems::internal::SystemId) $7 = (value_ = 6) (lldb) p clone.CreateDefaultContext()->get_system_id() (drake::systems::internal::SystemId) $8 = (value_ = 9) (lldb) p clone->CreateDefaultContext()->get_system_id() (drake::systems::internal::SystemId) $11 = (value_ = 9) (lldb) p clone->get_input_port(0).owning_system_id_ (const drake::systems::internal::SystemId) $12 = (value_ = 8) ``` Stepping into the code, I believe that the problem probably happens [here](https://github.com/RobotLocomotion/drake/blob/8b25bb8cd378802e1b819d66b4fba724d4b2fcc9/systems/framework/system.cc#L42).... it looks like the system's ID gets reset, but the port's IDs do not. ### Version _No response_ ### What operating system are you using? _No response_ ### What installation option are you using? _No response_ ### Relevant log output _No response_
1.0
System::Clone has incorrect SystemIDs for the ports - ### What happened? Initially reported by @AlexandreAmice ... the following test ``` GTEST_TEST(PendulumPlantTest, Clone) { const PendulumPlant<double> plant; auto clone = plant.Clone(); auto context = clone->CreateDefaultContext(); clone->get_input_port(0).FixValue(context.get(), Vector1d::Zero()); } ``` fails with ``` C++ exception with description "InputPort: The Context given as an argument was not created for this InputPort[0] (tau) of System ::_ (PendulumPlant<double>)" thrown in the test body. ``` Digging around in lldb (with a breakpoint on the last line), I see ``` (lldb) p plant.get_system_id() (drake::systems::internal::SystemId) $5 = (value_ = 6) (lldb) p clone->get_system_id() (drake::systems::internal::SystemId) $6 = (value_ = 9) (lldb) p plant.CreateDefaultContext()->get_system_id() (drake::systems::internal::SystemId) $7 = (value_ = 6) (lldb) p clone.CreateDefaultContext()->get_system_id() (drake::systems::internal::SystemId) $8 = (value_ = 9) (lldb) p clone->CreateDefaultContext()->get_system_id() (drake::systems::internal::SystemId) $11 = (value_ = 9) (lldb) p clone->get_input_port(0).owning_system_id_ (const drake::systems::internal::SystemId) $12 = (value_ = 8) ``` Stepping into the code, I believe that the problem probably happens [here](https://github.com/RobotLocomotion/drake/blob/8b25bb8cd378802e1b819d66b4fba724d4b2fcc9/systems/framework/system.cc#L42).... it looks like the system's ID gets reset, but the port's IDs do not. ### Version _No response_ ### What operating system are you using? _No response_ ### What installation option are you using? _No response_ ### Relevant log output _No response_
non_test
system clone has incorrect systemids for the ports what happened initially reported by alexandreamice the following test gtest test pendulumplanttest clone const pendulumplant plant auto clone plant clone auto context clone createdefaultcontext clone get input port fixvalue context get zero fails with c exception with description inputport the context given as an argument was not created for this inputport tau of system pendulumplant thrown in the test body digging around in lldb with a breakpoint on the last line i see lldb p plant get system id drake systems internal systemid value lldb p clone get system id drake systems internal systemid value lldb p plant createdefaultcontext get system id drake systems internal systemid value lldb p clone createdefaultcontext get system id drake systems internal systemid value lldb p clone createdefaultcontext get system id drake systems internal systemid value lldb p clone get input port owning system id const drake systems internal systemid value stepping into the code i believe that the problem probably happens it looks like the system s id gets reset but the port s ids do not version no response what operating system are you using no response what installation option are you using no response relevant log output no response
0
306,892
9,412,570,838
IssuesEvent
2019-04-10 04:42:40
metabase/metabase
https://api.github.com/repos/metabase/metabase
closed
Metabase hangs with unknown reason, CPU usage full, saying "The database(H2) has been closed"
Bug Priority/P1 Running Metabase
- Your databases: (e.x. MariaDB) - Metabase version: (e.x. 0.31.2) - Metabase hosting environment: (e.x. Debian 9) - Metabase internal database: (e.x. H2) - *Repeatable steps to reproduce the issue* After long time run, the metabase hangs and CPU is full (400% at a 4 core cpu) `kill` doesn't work and I have to do `kill -9` to force terminating. Sorry I forgot to do `jstack`. ``` 02-21 08:02:32 DEBUG sync.util :: STARTING: step 'sync-fks' for mysql Database 2 'xxxxxxx' 02-21 08:43:12 DEBUG metabase.middleware :: GET /api/user/current 200 (29 mins) (2 DB calls). Jetty threads: 8/50 (34 busy, 2 idle, 0 queued) 02-21 08:43:27 ERROR jdbcjobstore.JobStoreTX :: Couldn't rollback jdbc connection. The database has been closed [90098-197] org.h2.jdbc.JdbcSQLException: The database has been closed [90098-197] at org.h2.message.DbException.getJdbcSQLException(DbException.java:357) at org.h2.message.DbException.get(DbException.java:179) at org.h2.message.DbException.get(DbException.java:155) at org.h2.message.DbException.get(DbException.java:144) at org.h2.engine.Session.getTransaction(Session.java:1686) at org.h2.engine.Session.getStatementSavepoint(Session.java:1696) at org.h2.engine.Session.setSavepoint(Session.java:859) at org.h2.command.Command.executeUpdate(Command.java:255) at org.h2.jdbc.JdbcConnection.rollbackInternal(JdbcConnection.java:1558) at org.h2.jdbc.JdbcConnection.rollback(JdbcConnection.java:518) at com.mchange.v2.c3p0.impl.NewProxyConnection.rollback(NewProxyConnection.java:1033) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.quartz.impl.jdbcjobstore.AttributeRestoringConnectionInvocationHandler.invoke(AttributeRestoringConnectionInvocationHandler.java:73) at com.sun.proxy.$Proxy12.rollback(Unknown Source) at org.quartz.impl.jdbcjobstore.JobStoreSupport.rollbackConnection(JobStoreSupport.java:3639) at org.quartz.impl.jdbcjobstore.JobStoreSupport.doCheckin(JobStoreSupport.java:3264) at org.quartz.impl.jdbcjobstore.JobStoreSupport$ClusterManager.manage(JobStoreSupport.java:3857) at org.quartz.impl.jdbcjobstore.JobStoreSupport$ClusterManager.run(JobStoreSupport.java:3894) 02-21 08:43:31 WARN server.HttpChannel :: /api/collection/root org.h2.jdbc.JdbcSQLException: The database has been closed [90098-197] [665/1998] at org.h2.message.DbException.getJdbcSQLException(DbException.java:357) at org.h2.message.DbException.get(DbException.java:179) at org.h2.message.DbException.get(DbException.java:155) at org.h2.message.DbException.get(DbException.java:144) at org.h2.engine.Database.checkPowerOff(Database.java:536) at org.h2.command.Command.executeQuery(Command.java:228) at org.h2.jdbc.JdbcPreparedStatement.executeQuery(JdbcPreparedStatement.java:114) at com.mchange.v2.c3p0.impl.NewProxyPreparedStatement.executeQuery(NewProxyPreparedStatement.java:353) at clojure.java.jdbc$execute_query_with_params.invokeStatic(jdbc.clj:1002) at clojure.java.jdbc$execute_query_with_params.invoke(jdbc.clj:996) at clojure.java.jdbc$db_query_with_resultset_STAR_.invokeStatic(jdbc.clj:1025) at clojure.java.jdbc$db_query_with_resultset_STAR_.invoke(jdbc.clj:1005) at clojure.java.jdbc$query.invokeStatic(jdbc.clj:1099) at clojure.java.jdbc$query.invoke(jdbc.clj:1056) at toucan.db$query.invokeStatic(db.clj:275) at toucan.db$query.doInvoke(db.clj:271) at clojure.lang.RestFn.invoke(RestFn.java:410) at toucan.db$simple_select.invokeStatic(db.clj:379) at toucan.db$simple_select.invoke(db.clj:368) at toucan.db$simple_select_one.invokeStatic(db.clj:405) at toucan.db$simple_select_one.invoke(db.clj:394) at toucan.db$select_one.invokeStatic(db.clj:606) at toucan.db$select_one.doInvoke(db.clj:599) at clojure.lang.RestFn.invoke(RestFn.java:516) at metabase.middleware$session_with_id.invokeStatic(middleware.clj:73) at metabase.middleware$session_with_id.invoke(middleware.clj:70) at metabase.middleware$current_user_info_for_session.invokeStatic(middleware.clj:94) at metabase.middleware$current_user_info_for_session.invoke(middleware.clj:90) at metabase.middleware$add_current_user_info.invokeStatic(middleware.clj:100) at metabase.middleware$add_current_user_info.invoke(middleware.clj:99) at clojure.core$comp$fn__5529.invoke(core.clj:2561) at clojure.core$comp$fn__5529.invoke(core.clj:2561) at clojure.core$comp$fn__5529.invoke(core.clj:2561) at metabase.middleware$maybe_set_site_url$fn__56239.invoke(middleware.clj:290) at puppetlabs.i18n.core$locale_negotiator$fn__124.invoke(core.clj:357) at ring.middleware.cookies$wrap_cookies$fn__62401.invoke(cookies.clj:175) at ring.middleware.session$wrap_session$fn__62658.invoke(session.clj:108) at metabase.middleware$add_content_type$fn__56232.invoke(middleware.clj:262) at ring.middleware.gzip$wrap_gzip$fn__62432.invoke(gzip.clj:65) at ring.adapter.jetty$proxy_handler$fn__62260.invoke(jetty.clj:25) at ring.adapter.jetty.proxy$org.eclipse.jetty.server.handler.AbstractHandler$ff19274a.handle(Unknown Source) at org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:132) [623/1998] at org.eclipse.jetty.server.Server.handle(Server.java:531) at org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:352) at org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:260) at org.eclipse.jetty.io.AbstractConnection$ReadCallback.succeeded(AbstractConnection.java:281) at org.eclipse.jetty.io.FillInterest.fillable(FillInterest.java:102) at org.eclipse.jetty.io.ChannelEndPoint$2.run(ChannelEndPoint.java:118) at org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:762) at org.eclipse.jetty.util.thread.QueuedThreadPool$2.run(QueuedThreadPool.java:680) at java.lang.Thread.run(Thread.java:748) 02-21 08:43:41 ERROR jdbcjobstore.JobStoreTX :: Couldn't rollback jdbc connection. The database has been closed [90098-197] org.h2.jdbc.JdbcSQLException: The database has been closed [90098-197] at org.h2.message.DbException.getJdbcSQLException(DbException.java:357) at org.h2.message.DbException.get(DbException.java:179) at org.h2.message.DbException.get(DbException.java:155) at org.h2.message.DbException.get(DbException.java:144) at org.h2.engine.Session.getTransaction(Session.java:1686) at org.h2.engine.Session.getStatementSavepoint(Session.java:1696) at org.h2.engine.Session.setSavepoint(Session.java:859) at org.h2.command.Command.executeUpdate(Command.java:255) at org.h2.jdbc.JdbcConnection.rollbackInternal(JdbcConnection.java:1558) at org.h2.jdbc.JdbcConnection.rollback(JdbcConnection.java:518) at com.mchange.v2.c3p0.impl.NewProxyConnection.rollback(NewProxyConnection.java:1033) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.quartz.impl.jdbcjobstore.AttributeRestoringConnectionInvocationHandler.invoke(AttributeRestoringConnectionInvocationHandler.java:73) at com.sun.proxy.$Proxy12.rollback(Unknown Source) at org.quartz.impl.jdbcjobstore.JobStoreSupport.rollbackConnection(JobStoreSupport.java:3639) at org.quartz.impl.jdbcjobstore.JobStoreSupport.doRecoverMisfires(JobStoreSupport.java:3183) at org.quartz.impl.jdbcjobstore.JobStoreSupport$MisfireHandler.manage(JobStoreSupport.java:3934) at org.quartz.impl.jdbcjobstore.JobStoreSupport$MisfireHandler.run(JobStoreSupport.java:3955) 02-21 08:43:41 WARN server.HttpChannel :: /api/dashboard/1 org.h2.jdbc.JdbcSQLException: The database has been closed [90098-197] at org.h2.message.DbException.getJdbcSQLException(DbException.java:357) at org.h2.message.DbException.get(DbException.java:179) at org.h2.message.DbException.get(DbException.java:155) at org.h2.message.DbException.get(DbException.java:144) at org.h2.engine.Database.checkPowerOff(Database.java:536) at org.h2.command.Command.executeQuery(Command.java:228) at org.h2.jdbc.JdbcPreparedStatement.executeQuery(JdbcPreparedStatement.java:114) at com.mchange.v2.c3p0.impl.NewProxyPreparedStatement.executeQuery(NewProxyPreparedStatement.java:353) at clojure.java.jdbc$execute_query_with_params.invokeStatic(jdbc.clj:1002) at clojure.java.jdbc$execute_query_with_params.invoke(jdbc.clj:996) at clojure.java.jdbc$db_query_with_resultset_STAR_.invokeStatic(jdbc.clj:1025) [578/1998] at clojure.java.jdbc$db_query_with_resultset_STAR_.invoke(jdbc.clj:1005) at clojure.java.jdbc$query.invokeStatic(jdbc.clj:1099) at clojure.java.jdbc$query.invoke(jdbc.clj:1056) at toucan.db$query.invokeStatic(db.clj:275) at toucan.db$query.doInvoke(db.clj:271) at clojure.lang.RestFn.invoke(RestFn.java:410) at toucan.db$simple_select.invokeStatic(db.clj:379) at toucan.db$simple_select.invoke(db.clj:368) at toucan.db$simple_select_one.invokeStatic(db.clj:405) at toucan.db$simple_select_one.invoke(db.clj:394) at toucan.db$select_one.invokeStatic(db.clj:606) at toucan.db$select_one.doInvoke(db.clj:599) at clojure.lang.RestFn.invoke(RestFn.java:516) at metabase.middleware$session_with_id.invokeStatic(middleware.clj:73) at metabase.middleware$session_with_id.invoke(middleware.clj:70) at metabase.middleware$current_user_info_for_session.invokeStatic(middleware.clj:94) at metabase.middleware$current_user_info_for_session.invoke(middleware.clj:90) at metabase.middleware$add_current_user_info.invokeStatic(middleware.clj:100) at metabase.middleware$add_current_user_info.invoke(middleware.clj:99) at clojure.core$comp$fn__5529.invoke(core.clj:2561) at clojure.core$comp$fn__5529.invoke(core.clj:2561) at clojure.core$comp$fn__5529.invoke(core.clj:2561) at metabase.middleware$maybe_set_site_url$fn__56239.invoke(middleware.clj:290) at puppetlabs.i18n.core$locale_negotiator$fn__124.invoke(core.clj:357) at ring.middleware.cookies$wrap_cookies$fn__62401.invoke(cookies.clj:175) at ring.middleware.session$wrap_session$fn__62658.invoke(session.clj:108) at metabase.middleware$add_content_type$fn__56232.invoke(middleware.clj:262) at ring.middleware.gzip$wrap_gzip$fn__62432.invoke(gzip.clj:65) at ring.adapter.jetty$proxy_handler$fn__62260.invoke(jetty.clj:25) at ring.adapter.jetty.proxy$org.eclipse.jetty.server.handler.AbstractHandler$ff19274a.handle(Unknown Source) at org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:132) at org.eclipse.jetty.server.Server.handle(Server.java:531) at org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:352) at org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:260) at org.eclipse.jetty.io.AbstractConnection$ReadCallback.succeeded(AbstractConnection.java:281) at org.eclipse.jetty.io.FillInterest.fillable(FillInterest.java:102) at org.eclipse.jetty.io.ChannelEndPoint$2.run(ChannelEndPoint.java:118) at org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:762) at org.eclipse.jetty.util.thread.QueuedThreadPool$2.run(QueuedThreadPool.java:680) at java.lang.Thread.run(Thread.java:748) ```
1.0
Metabase hangs with unknown reason, CPU usage full, saying "The database(H2) has been closed" - - Your databases: (e.x. MariaDB) - Metabase version: (e.x. 0.31.2) - Metabase hosting environment: (e.x. Debian 9) - Metabase internal database: (e.x. H2) - *Repeatable steps to reproduce the issue* After long time run, the metabase hangs and CPU is full (400% at a 4 core cpu) `kill` doesn't work and I have to do `kill -9` to force terminating. Sorry I forgot to do `jstack`. ``` 02-21 08:02:32 DEBUG sync.util :: STARTING: step 'sync-fks' for mysql Database 2 'xxxxxxx' 02-21 08:43:12 DEBUG metabase.middleware :: GET /api/user/current 200 (29 mins) (2 DB calls). Jetty threads: 8/50 (34 busy, 2 idle, 0 queued) 02-21 08:43:27 ERROR jdbcjobstore.JobStoreTX :: Couldn't rollback jdbc connection. The database has been closed [90098-197] org.h2.jdbc.JdbcSQLException: The database has been closed [90098-197] at org.h2.message.DbException.getJdbcSQLException(DbException.java:357) at org.h2.message.DbException.get(DbException.java:179) at org.h2.message.DbException.get(DbException.java:155) at org.h2.message.DbException.get(DbException.java:144) at org.h2.engine.Session.getTransaction(Session.java:1686) at org.h2.engine.Session.getStatementSavepoint(Session.java:1696) at org.h2.engine.Session.setSavepoint(Session.java:859) at org.h2.command.Command.executeUpdate(Command.java:255) at org.h2.jdbc.JdbcConnection.rollbackInternal(JdbcConnection.java:1558) at org.h2.jdbc.JdbcConnection.rollback(JdbcConnection.java:518) at com.mchange.v2.c3p0.impl.NewProxyConnection.rollback(NewProxyConnection.java:1033) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.quartz.impl.jdbcjobstore.AttributeRestoringConnectionInvocationHandler.invoke(AttributeRestoringConnectionInvocationHandler.java:73) at com.sun.proxy.$Proxy12.rollback(Unknown Source) at org.quartz.impl.jdbcjobstore.JobStoreSupport.rollbackConnection(JobStoreSupport.java:3639) at org.quartz.impl.jdbcjobstore.JobStoreSupport.doCheckin(JobStoreSupport.java:3264) at org.quartz.impl.jdbcjobstore.JobStoreSupport$ClusterManager.manage(JobStoreSupport.java:3857) at org.quartz.impl.jdbcjobstore.JobStoreSupport$ClusterManager.run(JobStoreSupport.java:3894) 02-21 08:43:31 WARN server.HttpChannel :: /api/collection/root org.h2.jdbc.JdbcSQLException: The database has been closed [90098-197] [665/1998] at org.h2.message.DbException.getJdbcSQLException(DbException.java:357) at org.h2.message.DbException.get(DbException.java:179) at org.h2.message.DbException.get(DbException.java:155) at org.h2.message.DbException.get(DbException.java:144) at org.h2.engine.Database.checkPowerOff(Database.java:536) at org.h2.command.Command.executeQuery(Command.java:228) at org.h2.jdbc.JdbcPreparedStatement.executeQuery(JdbcPreparedStatement.java:114) at com.mchange.v2.c3p0.impl.NewProxyPreparedStatement.executeQuery(NewProxyPreparedStatement.java:353) at clojure.java.jdbc$execute_query_with_params.invokeStatic(jdbc.clj:1002) at clojure.java.jdbc$execute_query_with_params.invoke(jdbc.clj:996) at clojure.java.jdbc$db_query_with_resultset_STAR_.invokeStatic(jdbc.clj:1025) at clojure.java.jdbc$db_query_with_resultset_STAR_.invoke(jdbc.clj:1005) at clojure.java.jdbc$query.invokeStatic(jdbc.clj:1099) at clojure.java.jdbc$query.invoke(jdbc.clj:1056) at toucan.db$query.invokeStatic(db.clj:275) at toucan.db$query.doInvoke(db.clj:271) at clojure.lang.RestFn.invoke(RestFn.java:410) at toucan.db$simple_select.invokeStatic(db.clj:379) at toucan.db$simple_select.invoke(db.clj:368) at toucan.db$simple_select_one.invokeStatic(db.clj:405) at toucan.db$simple_select_one.invoke(db.clj:394) at toucan.db$select_one.invokeStatic(db.clj:606) at toucan.db$select_one.doInvoke(db.clj:599) at clojure.lang.RestFn.invoke(RestFn.java:516) at metabase.middleware$session_with_id.invokeStatic(middleware.clj:73) at metabase.middleware$session_with_id.invoke(middleware.clj:70) at metabase.middleware$current_user_info_for_session.invokeStatic(middleware.clj:94) at metabase.middleware$current_user_info_for_session.invoke(middleware.clj:90) at metabase.middleware$add_current_user_info.invokeStatic(middleware.clj:100) at metabase.middleware$add_current_user_info.invoke(middleware.clj:99) at clojure.core$comp$fn__5529.invoke(core.clj:2561) at clojure.core$comp$fn__5529.invoke(core.clj:2561) at clojure.core$comp$fn__5529.invoke(core.clj:2561) at metabase.middleware$maybe_set_site_url$fn__56239.invoke(middleware.clj:290) at puppetlabs.i18n.core$locale_negotiator$fn__124.invoke(core.clj:357) at ring.middleware.cookies$wrap_cookies$fn__62401.invoke(cookies.clj:175) at ring.middleware.session$wrap_session$fn__62658.invoke(session.clj:108) at metabase.middleware$add_content_type$fn__56232.invoke(middleware.clj:262) at ring.middleware.gzip$wrap_gzip$fn__62432.invoke(gzip.clj:65) at ring.adapter.jetty$proxy_handler$fn__62260.invoke(jetty.clj:25) at ring.adapter.jetty.proxy$org.eclipse.jetty.server.handler.AbstractHandler$ff19274a.handle(Unknown Source) at org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:132) [623/1998] at org.eclipse.jetty.server.Server.handle(Server.java:531) at org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:352) at org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:260) at org.eclipse.jetty.io.AbstractConnection$ReadCallback.succeeded(AbstractConnection.java:281) at org.eclipse.jetty.io.FillInterest.fillable(FillInterest.java:102) at org.eclipse.jetty.io.ChannelEndPoint$2.run(ChannelEndPoint.java:118) at org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:762) at org.eclipse.jetty.util.thread.QueuedThreadPool$2.run(QueuedThreadPool.java:680) at java.lang.Thread.run(Thread.java:748) 02-21 08:43:41 ERROR jdbcjobstore.JobStoreTX :: Couldn't rollback jdbc connection. The database has been closed [90098-197] org.h2.jdbc.JdbcSQLException: The database has been closed [90098-197] at org.h2.message.DbException.getJdbcSQLException(DbException.java:357) at org.h2.message.DbException.get(DbException.java:179) at org.h2.message.DbException.get(DbException.java:155) at org.h2.message.DbException.get(DbException.java:144) at org.h2.engine.Session.getTransaction(Session.java:1686) at org.h2.engine.Session.getStatementSavepoint(Session.java:1696) at org.h2.engine.Session.setSavepoint(Session.java:859) at org.h2.command.Command.executeUpdate(Command.java:255) at org.h2.jdbc.JdbcConnection.rollbackInternal(JdbcConnection.java:1558) at org.h2.jdbc.JdbcConnection.rollback(JdbcConnection.java:518) at com.mchange.v2.c3p0.impl.NewProxyConnection.rollback(NewProxyConnection.java:1033) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.quartz.impl.jdbcjobstore.AttributeRestoringConnectionInvocationHandler.invoke(AttributeRestoringConnectionInvocationHandler.java:73) at com.sun.proxy.$Proxy12.rollback(Unknown Source) at org.quartz.impl.jdbcjobstore.JobStoreSupport.rollbackConnection(JobStoreSupport.java:3639) at org.quartz.impl.jdbcjobstore.JobStoreSupport.doRecoverMisfires(JobStoreSupport.java:3183) at org.quartz.impl.jdbcjobstore.JobStoreSupport$MisfireHandler.manage(JobStoreSupport.java:3934) at org.quartz.impl.jdbcjobstore.JobStoreSupport$MisfireHandler.run(JobStoreSupport.java:3955) 02-21 08:43:41 WARN server.HttpChannel :: /api/dashboard/1 org.h2.jdbc.JdbcSQLException: The database has been closed [90098-197] at org.h2.message.DbException.getJdbcSQLException(DbException.java:357) at org.h2.message.DbException.get(DbException.java:179) at org.h2.message.DbException.get(DbException.java:155) at org.h2.message.DbException.get(DbException.java:144) at org.h2.engine.Database.checkPowerOff(Database.java:536) at org.h2.command.Command.executeQuery(Command.java:228) at org.h2.jdbc.JdbcPreparedStatement.executeQuery(JdbcPreparedStatement.java:114) at com.mchange.v2.c3p0.impl.NewProxyPreparedStatement.executeQuery(NewProxyPreparedStatement.java:353) at clojure.java.jdbc$execute_query_with_params.invokeStatic(jdbc.clj:1002) at clojure.java.jdbc$execute_query_with_params.invoke(jdbc.clj:996) at clojure.java.jdbc$db_query_with_resultset_STAR_.invokeStatic(jdbc.clj:1025) [578/1998] at clojure.java.jdbc$db_query_with_resultset_STAR_.invoke(jdbc.clj:1005) at clojure.java.jdbc$query.invokeStatic(jdbc.clj:1099) at clojure.java.jdbc$query.invoke(jdbc.clj:1056) at toucan.db$query.invokeStatic(db.clj:275) at toucan.db$query.doInvoke(db.clj:271) at clojure.lang.RestFn.invoke(RestFn.java:410) at toucan.db$simple_select.invokeStatic(db.clj:379) at toucan.db$simple_select.invoke(db.clj:368) at toucan.db$simple_select_one.invokeStatic(db.clj:405) at toucan.db$simple_select_one.invoke(db.clj:394) at toucan.db$select_one.invokeStatic(db.clj:606) at toucan.db$select_one.doInvoke(db.clj:599) at clojure.lang.RestFn.invoke(RestFn.java:516) at metabase.middleware$session_with_id.invokeStatic(middleware.clj:73) at metabase.middleware$session_with_id.invoke(middleware.clj:70) at metabase.middleware$current_user_info_for_session.invokeStatic(middleware.clj:94) at metabase.middleware$current_user_info_for_session.invoke(middleware.clj:90) at metabase.middleware$add_current_user_info.invokeStatic(middleware.clj:100) at metabase.middleware$add_current_user_info.invoke(middleware.clj:99) at clojure.core$comp$fn__5529.invoke(core.clj:2561) at clojure.core$comp$fn__5529.invoke(core.clj:2561) at clojure.core$comp$fn__5529.invoke(core.clj:2561) at metabase.middleware$maybe_set_site_url$fn__56239.invoke(middleware.clj:290) at puppetlabs.i18n.core$locale_negotiator$fn__124.invoke(core.clj:357) at ring.middleware.cookies$wrap_cookies$fn__62401.invoke(cookies.clj:175) at ring.middleware.session$wrap_session$fn__62658.invoke(session.clj:108) at metabase.middleware$add_content_type$fn__56232.invoke(middleware.clj:262) at ring.middleware.gzip$wrap_gzip$fn__62432.invoke(gzip.clj:65) at ring.adapter.jetty$proxy_handler$fn__62260.invoke(jetty.clj:25) at ring.adapter.jetty.proxy$org.eclipse.jetty.server.handler.AbstractHandler$ff19274a.handle(Unknown Source) at org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:132) at org.eclipse.jetty.server.Server.handle(Server.java:531) at org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:352) at org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:260) at org.eclipse.jetty.io.AbstractConnection$ReadCallback.succeeded(AbstractConnection.java:281) at org.eclipse.jetty.io.FillInterest.fillable(FillInterest.java:102) at org.eclipse.jetty.io.ChannelEndPoint$2.run(ChannelEndPoint.java:118) at org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:762) at org.eclipse.jetty.util.thread.QueuedThreadPool$2.run(QueuedThreadPool.java:680) at java.lang.Thread.run(Thread.java:748) ```
non_test
metabase hangs with unknown reason cpu usage full saying the database has been closed your databases e x mariadb metabase version e x metabase hosting environment e x debian metabase internal database e x repeatable steps to reproduce the issue after long time run the metabase hangs and cpu is full at a core cpu kill doesn t work and i have to do kill to force terminating sorry i forgot to do jstack debug sync util starting step sync fks for mysql database xxxxxxx debug metabase middleware get api user current mins db calls jetty threads busy idle queued error jdbcjobstore jobstoretx couldn t rollback jdbc connection the database has been closed org jdbc jdbcsqlexception the database has been closed at org message dbexception getjdbcsqlexception dbexception java at org message dbexception get dbexception java at org message dbexception get dbexception java at org message dbexception get dbexception java at org engine session gettransaction session java at org engine session getstatementsavepoint session java at org engine session setsavepoint session java at org command command executeupdate command java at org jdbc jdbcconnection rollbackinternal jdbcconnection java at org jdbc jdbcconnection rollback jdbcconnection java at com mchange impl newproxyconnection rollback newproxyconnection java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org quartz impl jdbcjobstore attributerestoringconnectioninvocationhandler invoke attributerestoringconnectioninvocationhandler java at com sun proxy rollback unknown source at org quartz impl jdbcjobstore jobstoresupport rollbackconnection jobstoresupport java at org quartz impl jdbcjobstore jobstoresupport docheckin jobstoresupport java at org quartz impl jdbcjobstore jobstoresupport clustermanager manage jobstoresupport java at org quartz impl jdbcjobstore jobstoresupport clustermanager run jobstoresupport java warn server httpchannel api collection root org jdbc jdbcsqlexception the database has been closed at org message dbexception getjdbcsqlexception dbexception java at org message dbexception get dbexception java at org message dbexception get dbexception java at org message dbexception get dbexception java at org engine database checkpoweroff database java at org command command executequery command java at org jdbc jdbcpreparedstatement executequery jdbcpreparedstatement java at com mchange impl newproxypreparedstatement executequery newproxypreparedstatement java at clojure java jdbc execute query with params invokestatic jdbc clj at clojure java jdbc execute query with params invoke jdbc clj at clojure java jdbc db query with resultset star invokestatic jdbc clj at clojure java jdbc db query with resultset star invoke jdbc clj at clojure java jdbc query invokestatic jdbc clj at clojure java jdbc query invoke jdbc clj at toucan db query invokestatic db clj at toucan db query doinvoke db clj at clojure lang restfn invoke restfn java at toucan db simple select invokestatic db clj at toucan db simple select invoke db clj at toucan db simple select one invokestatic db clj at toucan db simple select one invoke db clj at toucan db select one invokestatic db clj at toucan db select one doinvoke db clj at clojure lang restfn invoke restfn java at metabase middleware session with id invokestatic middleware clj at metabase middleware session with id invoke middleware clj at metabase middleware current user info for session invokestatic middleware clj at metabase middleware current user info for session invoke middleware clj at metabase middleware add current user info invokestatic middleware clj at metabase middleware add current user info invoke middleware clj at clojure core comp fn invoke core clj at clojure core comp fn invoke core clj at clojure core comp fn invoke core clj at metabase middleware maybe set site url fn invoke middleware clj at puppetlabs core locale negotiator fn invoke core clj at ring middleware cookies wrap cookies fn invoke cookies clj at ring middleware session wrap session fn invoke session clj at metabase middleware add content type fn invoke middleware clj at ring middleware gzip wrap gzip fn invoke gzip clj at ring adapter jetty proxy handler fn invoke jetty clj at ring adapter jetty proxy org eclipse jetty server handler abstracthandler handle unknown source at org eclipse jetty server handler handlerwrapper handle handlerwrapper java at org eclipse jetty server server handle server java at org eclipse jetty server httpchannel handle httpchannel java at org eclipse jetty server httpconnection onfillable httpconnection java at org eclipse jetty io abstractconnection readcallback succeeded abstractconnection java at org eclipse jetty io fillinterest fillable fillinterest java at org eclipse jetty io channelendpoint run channelendpoint java at org eclipse jetty util thread queuedthreadpool runjob queuedthreadpool java at org eclipse jetty util thread queuedthreadpool run queuedthreadpool java at java lang thread run thread java error jdbcjobstore jobstoretx couldn t rollback jdbc connection the database has been closed org jdbc jdbcsqlexception the database has been closed at org message dbexception getjdbcsqlexception dbexception java at org message dbexception get dbexception java at org message dbexception get dbexception java at org message dbexception get dbexception java at org engine session gettransaction session java at org engine session getstatementsavepoint session java at org engine session setsavepoint session java at org command command executeupdate command java at org jdbc jdbcconnection rollbackinternal jdbcconnection java at org jdbc jdbcconnection rollback jdbcconnection java at com mchange impl newproxyconnection rollback newproxyconnection java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org quartz impl jdbcjobstore attributerestoringconnectioninvocationhandler invoke attributerestoringconnectioninvocationhandler java at com sun proxy rollback unknown source at org quartz impl jdbcjobstore jobstoresupport rollbackconnection jobstoresupport java at org quartz impl jdbcjobstore jobstoresupport dorecovermisfires jobstoresupport java at org quartz impl jdbcjobstore jobstoresupport misfirehandler manage jobstoresupport java at org quartz impl jdbcjobstore jobstoresupport misfirehandler run jobstoresupport java warn server httpchannel api dashboard org jdbc jdbcsqlexception the database has been closed at org message dbexception getjdbcsqlexception dbexception java at org message dbexception get dbexception java at org message dbexception get dbexception java at org message dbexception get dbexception java at org engine database checkpoweroff database java at org command command executequery command java at org jdbc jdbcpreparedstatement executequery jdbcpreparedstatement java at com mchange impl newproxypreparedstatement executequery newproxypreparedstatement java at clojure java jdbc execute query with params invokestatic jdbc clj at clojure java jdbc execute query with params invoke jdbc clj at clojure java jdbc db query with resultset star invokestatic jdbc clj at clojure java jdbc db query with resultset star invoke jdbc clj at clojure java jdbc query invokestatic jdbc clj at clojure java jdbc query invoke jdbc clj at toucan db query invokestatic db clj at toucan db query doinvoke db clj at clojure lang restfn invoke restfn java at toucan db simple select invokestatic db clj at toucan db simple select invoke db clj at toucan db simple select one invokestatic db clj at toucan db simple select one invoke db clj at toucan db select one invokestatic db clj at toucan db select one doinvoke db clj at clojure lang restfn invoke restfn java at metabase middleware session with id invokestatic middleware clj at metabase middleware session with id invoke middleware clj at metabase middleware current user info for session invokestatic middleware clj at metabase middleware current user info for session invoke middleware clj at metabase middleware add current user info invokestatic middleware clj at metabase middleware add current user info invoke middleware clj at clojure core comp fn invoke core clj at clojure core comp fn invoke core clj at clojure core comp fn invoke core clj at metabase middleware maybe set site url fn invoke middleware clj at puppetlabs core locale negotiator fn invoke core clj at ring middleware cookies wrap cookies fn invoke cookies clj at ring middleware session wrap session fn invoke session clj at metabase middleware add content type fn invoke middleware clj at ring middleware gzip wrap gzip fn invoke gzip clj at ring adapter jetty proxy handler fn invoke jetty clj at ring adapter jetty proxy org eclipse jetty server handler abstracthandler handle unknown source at org eclipse jetty server handler handlerwrapper handle handlerwrapper java at org eclipse jetty server server handle server java at org eclipse jetty server httpchannel handle httpchannel java at org eclipse jetty server httpconnection onfillable httpconnection java at org eclipse jetty io abstractconnection readcallback succeeded abstractconnection java at org eclipse jetty io fillinterest fillable fillinterest java at org eclipse jetty io channelendpoint run channelendpoint java at org eclipse jetty util thread queuedthreadpool runjob queuedthreadpool java at org eclipse jetty util thread queuedthreadpool run queuedthreadpool java at java lang thread run thread java
0
114,281
14,544,513,273
IssuesEvent
2020-12-15 18:17:22
cloudfour/cloudfour.com-patterns
https://api.github.com/repos/cloudfour/cloudfour.com-patterns
opened
Button group pattern(s)
size:2 🎨 design
We could use a layout object for aligning a group of related buttons. This can be seen in our post-article notification area: <img width="793" alt="Screen Shot 2020-12-15 at 10 15 08 AM" src="https://user-images.githubusercontent.com/69633/102255222-9fc30a80-3ebe-11eb-8695-ed47d764ede9.png"> And when replying to comments: <img width="321" alt="Screen Shot 2020-12-15 at 10 15 59 AM" src="https://user-images.githubusercontent.com/69633/102255242-a782af00-3ebe-11eb-9de5-9608bc3428b6.png">
1.0
Button group pattern(s) - We could use a layout object for aligning a group of related buttons. This can be seen in our post-article notification area: <img width="793" alt="Screen Shot 2020-12-15 at 10 15 08 AM" src="https://user-images.githubusercontent.com/69633/102255222-9fc30a80-3ebe-11eb-8695-ed47d764ede9.png"> And when replying to comments: <img width="321" alt="Screen Shot 2020-12-15 at 10 15 59 AM" src="https://user-images.githubusercontent.com/69633/102255242-a782af00-3ebe-11eb-9de5-9608bc3428b6.png">
non_test
button group pattern s we could use a layout object for aligning a group of related buttons this can be seen in our post article notification area img width alt screen shot at am src and when replying to comments img width alt screen shot at am src
0
134,840
12,627,960,990
IssuesEvent
2020-06-15 00:22:57
tamasfe/taplo
https://api.github.com/repos/tamasfe/taplo
closed
Documentation
documentation
Proper documentation must be written before releasing 1.0.0. This issue is just simply a reminder and an entry in the checklist of the milestone.
1.0
Documentation - Proper documentation must be written before releasing 1.0.0. This issue is just simply a reminder and an entry in the checklist of the milestone.
non_test
documentation proper documentation must be written before releasing this issue is just simply a reminder and an entry in the checklist of the milestone
0
248,198
21,002,517,685
IssuesEvent
2022-03-29 18:54:57
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
roachtest: psycopg failed
C-test-failure O-robot O-roachtest branch-master release-blocker
roachtest.psycopg [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=4718770&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=4718770&tab=artifacts#/psycopg) on master @ [327f886758e64973e9e6ed221688622b6e1bde69](https://github.com/cockroachdb/cockroach/commits/327f886758e64973e9e6ed221688622b6e1bde69): ``` The test failed on branch=master, cloud=gce: test artifacts and logs in: /artifacts/psycopg/run_1 orm_helpers.go:190,orm_helpers.go:116,psycopg.go:144,psycopg.go:156,test_runner.go:875: Tests run on Cockroach v22.1.0-alpha.4-449-g327f886758 Tests run against psycopg 2_8_6 453 Total Tests Run 451 tests passed 2 tests failed 319 tests skipped 3 tests ignored 0 tests passed unexpectedly 1 test failed unexpectedly 0 tests expected failed but skipped 0 tests expected failed but not run --- --- FAIL: tests.test_async.AsyncTests.test_error (unexpected) For a full summary look at the psycopg artifacts An updated blocklist (psycopgBlockList22_1) is available in the artifacts' psycopg log ``` <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/sql-experience <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*psycopg.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
2.0
roachtest: psycopg failed - roachtest.psycopg [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=4718770&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=4718770&tab=artifacts#/psycopg) on master @ [327f886758e64973e9e6ed221688622b6e1bde69](https://github.com/cockroachdb/cockroach/commits/327f886758e64973e9e6ed221688622b6e1bde69): ``` The test failed on branch=master, cloud=gce: test artifacts and logs in: /artifacts/psycopg/run_1 orm_helpers.go:190,orm_helpers.go:116,psycopg.go:144,psycopg.go:156,test_runner.go:875: Tests run on Cockroach v22.1.0-alpha.4-449-g327f886758 Tests run against psycopg 2_8_6 453 Total Tests Run 451 tests passed 2 tests failed 319 tests skipped 3 tests ignored 0 tests passed unexpectedly 1 test failed unexpectedly 0 tests expected failed but skipped 0 tests expected failed but not run --- --- FAIL: tests.test_async.AsyncTests.test_error (unexpected) For a full summary look at the psycopg artifacts An updated blocklist (psycopgBlockList22_1) is available in the artifacts' psycopg log ``` <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/sql-experience <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*psycopg.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
test
roachtest psycopg failed roachtest psycopg with on master the test failed on branch master cloud gce test artifacts and logs in artifacts psycopg run orm helpers go orm helpers go psycopg go psycopg go test runner go tests run on cockroach alpha tests run against psycopg total tests run tests passed tests failed tests skipped tests ignored tests passed unexpectedly test failed unexpectedly tests expected failed but skipped tests expected failed but not run fail tests test async asynctests test error unexpected for a full summary look at the psycopg artifacts an updated blocklist is available in the artifacts psycopg log help see see cc cockroachdb sql experience
1
327,303
28,052,073,572
IssuesEvent
2023-03-29 06:46:10
ALTA-LapakUMKM-Group-2/LapakUMKM-APITesting
https://api.github.com/repos/ALTA-LapakUMKM-Group-2/LapakUMKM-APITesting
closed
[User-A026] POST Update Photo Profile With Valid Key
Manual Api Testing
**Given** Post update photo profile with valid json data **When** User tick on body key **Then** Send post update data **And** API Should be 200 OK
1.0
[User-A026] POST Update Photo Profile With Valid Key - **Given** Post update photo profile with valid json data **When** User tick on body key **Then** Send post update data **And** API Should be 200 OK
test
post update photo profile with valid key given post update photo profile with valid json data when user tick on body key then send post update data and api should be ok
1
262,898
23,019,001,009
IssuesEvent
2022-07-22 01:48:59
sigp/lighthouse
https://api.github.com/repos/sigp/lighthouse
closed
Add transactions to EE integration tests
test improvement bellatrix
## Description In https://github.com/sigp/lighthouse/pull/3157 we removed finalized payloads from the database in favour of reconstructing them from the execution layer. Our reconstruction depends on the correctness of the `ethers` library's transaction encoding, and it would be good to ensure this is correct in CI. To do this we need to add some transactions to the execution blocks in our EE integration tests. This will involve creating and signing transactions, preferably of several different types (EIP-1559 and legacy), publishing them to the EE, ensuring they get included in payloads and that these payloads can be reconstructed. Deeper testing could involve fuzzing, or live network testing (e.g. trying to reconstruct every block on Kiln)
1.0
Add transactions to EE integration tests - ## Description In https://github.com/sigp/lighthouse/pull/3157 we removed finalized payloads from the database in favour of reconstructing them from the execution layer. Our reconstruction depends on the correctness of the `ethers` library's transaction encoding, and it would be good to ensure this is correct in CI. To do this we need to add some transactions to the execution blocks in our EE integration tests. This will involve creating and signing transactions, preferably of several different types (EIP-1559 and legacy), publishing them to the EE, ensuring they get included in payloads and that these payloads can be reconstructed. Deeper testing could involve fuzzing, or live network testing (e.g. trying to reconstruct every block on Kiln)
test
add transactions to ee integration tests description in we removed finalized payloads from the database in favour of reconstructing them from the execution layer our reconstruction depends on the correctness of the ethers library s transaction encoding and it would be good to ensure this is correct in ci to do this we need to add some transactions to the execution blocks in our ee integration tests this will involve creating and signing transactions preferably of several different types eip and legacy publishing them to the ee ensuring they get included in payloads and that these payloads can be reconstructed deeper testing could involve fuzzing or live network testing e g trying to reconstruct every block on kiln
1
188,828
14,476,420,350
IssuesEvent
2020-12-10 04:05:02
openshift/odo
https://api.github.com/repos/openshift/odo
closed
Namespaces creation in parallel are not handled properly by the test script
area/testing kind/bug lifecycle/rotten
/kind bug <!-- Welcome! - We kindly ask you to: 1. Fill out the issue template below 2. Use the Google group if you have a question rather than a bug or feature request. The group is at: https://groups.google.com/forum/#!forum/odo-users Thanks for understanding, and for contributing to the project! --> ## What versions of software are you using? **Operating System:** All supported **Output of `odo version`:** master ## How did you run odo exactly? Running test on travis CI. ## Actual behavior ``` Creating a new project: bivsrltawh Running odo with args [odo project create bivsrltawh -w -v4] [odo] β€’ Waiting for project to come up ... [odo] I0425 14:55:05.512561 9447 occlient.go:531] Status of creation of project bivsrltawh is Active [odo] I0425 14:55:05.512642 9447 occlient.go:536] Project bivsrltawh now exists [odo] I0425 14:55:05.528385 9447 occlient.go:571] Status of creation of service account &ServiceAccount{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:default,GenerateName:,Namespace:bivsrltawh,SelfLink:/api/v1/namespaces/bivsrltawh/serviceaccounts/default,UID:9ad08575-62da-48b5-af5f-bf30f409df1a,ResourceVersion:22723,Generation:0,CreationTimestamp:2020-04-25 14:55:04 +0000 UTC,DeletionTimestamp:<nil>,DeletionGracePeriodSeconds:nil,Labels:map[string]string{},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Secrets:[{ default-token-w5gvn } { default-dockercfg-w62w8 }],ImagePullSecrets:[{default-dockercfg-w62w8}],AutomountServiceAccountToken:nil,} is ready [odo] βœ“ Waiting for project to come up [2s] [odo] βœ“ Project 'bivsrltawh' is ready for use [odo] βœ“ New project created and now using project: bivsrltawh [odo] I0425 14:55:05.546426 9447 odo.go:80] Could not get the latest release information in time. Never mind, exiting gracefully :) Running odo with args [odo create nodejs --context /tmp/220268031 --project bivsrltawh jjpzrc --ref master --git https://github.com/openshift/nodejs-ex --port 8080,8000] [odo] Validation [odo] β€’ Validating component ... [odo] βœ“ Validating component [4ms] [odo] [odo] Please use `odo push` command to create the component with source deployed Running odo with args [odo push --context /tmp/220268031] [odo] βœ— projectrequests.project.openshift.io is forbidden: User "system:serviceaccount:ci-op-jkg9fswi:default" cannot create projectrequests.project.openshift.io at the cluster scope: no RBAC policy matched Deleting project: bivsrltawh Running odo with args [odo project delete bivsrltawh -f] [odo] βœ— The project bivsrltawh does not exist. Please check the list of projects using `odo project list` β€’ Failure [2.313 seconds] odo url command tests /go/src/github.com/openshift/odo/tests/integration/cmd_url_test.go:15 Listing urls /go/src/github.com/openshift/odo/tests/integration/cmd_url_test.go:41 should list appropriate URLs and push message [It] /go/src/github.com/openshift/odo/tests/integration/cmd_url_test.go:42 No future change is possible. Bailing out early after 0.132s. Running odo with args [odo push --context /tmp/220268031] Expected <int>: 1 to match exit code: <int>: 0 ``` On 4 test node we are observing this failure more prominently. ## Expected behavior It should create the namespace successfully and do ```odo push``` ## Any logs, error output, etc? Logs : https://prow.svc.ci.openshift.org/view/gcs/origin-ci-test/pr-logs/pull/openshift_odo/2965/pull-ci-openshift-odo-master-v4.4-integration-e2e-benchmark/431#1:build-log.txt%3A429
1.0
Namespaces creation in parallel are not handled properly by the test script - /kind bug <!-- Welcome! - We kindly ask you to: 1. Fill out the issue template below 2. Use the Google group if you have a question rather than a bug or feature request. The group is at: https://groups.google.com/forum/#!forum/odo-users Thanks for understanding, and for contributing to the project! --> ## What versions of software are you using? **Operating System:** All supported **Output of `odo version`:** master ## How did you run odo exactly? Running test on travis CI. ## Actual behavior ``` Creating a new project: bivsrltawh Running odo with args [odo project create bivsrltawh -w -v4] [odo] β€’ Waiting for project to come up ... [odo] I0425 14:55:05.512561 9447 occlient.go:531] Status of creation of project bivsrltawh is Active [odo] I0425 14:55:05.512642 9447 occlient.go:536] Project bivsrltawh now exists [odo] I0425 14:55:05.528385 9447 occlient.go:571] Status of creation of service account &ServiceAccount{ObjectMeta:k8s_io_apimachinery_pkg_apis_meta_v1.ObjectMeta{Name:default,GenerateName:,Namespace:bivsrltawh,SelfLink:/api/v1/namespaces/bivsrltawh/serviceaccounts/default,UID:9ad08575-62da-48b5-af5f-bf30f409df1a,ResourceVersion:22723,Generation:0,CreationTimestamp:2020-04-25 14:55:04 +0000 UTC,DeletionTimestamp:<nil>,DeletionGracePeriodSeconds:nil,Labels:map[string]string{},Annotations:map[string]string{},OwnerReferences:[],Finalizers:[],ClusterName:,Initializers:nil,},Secrets:[{ default-token-w5gvn } { default-dockercfg-w62w8 }],ImagePullSecrets:[{default-dockercfg-w62w8}],AutomountServiceAccountToken:nil,} is ready [odo] βœ“ Waiting for project to come up [2s] [odo] βœ“ Project 'bivsrltawh' is ready for use [odo] βœ“ New project created and now using project: bivsrltawh [odo] I0425 14:55:05.546426 9447 odo.go:80] Could not get the latest release information in time. Never mind, exiting gracefully :) Running odo with args [odo create nodejs --context /tmp/220268031 --project bivsrltawh jjpzrc --ref master --git https://github.com/openshift/nodejs-ex --port 8080,8000] [odo] Validation [odo] β€’ Validating component ... [odo] βœ“ Validating component [4ms] [odo] [odo] Please use `odo push` command to create the component with source deployed Running odo with args [odo push --context /tmp/220268031] [odo] βœ— projectrequests.project.openshift.io is forbidden: User "system:serviceaccount:ci-op-jkg9fswi:default" cannot create projectrequests.project.openshift.io at the cluster scope: no RBAC policy matched Deleting project: bivsrltawh Running odo with args [odo project delete bivsrltawh -f] [odo] βœ— The project bivsrltawh does not exist. Please check the list of projects using `odo project list` β€’ Failure [2.313 seconds] odo url command tests /go/src/github.com/openshift/odo/tests/integration/cmd_url_test.go:15 Listing urls /go/src/github.com/openshift/odo/tests/integration/cmd_url_test.go:41 should list appropriate URLs and push message [It] /go/src/github.com/openshift/odo/tests/integration/cmd_url_test.go:42 No future change is possible. Bailing out early after 0.132s. Running odo with args [odo push --context /tmp/220268031] Expected <int>: 1 to match exit code: <int>: 0 ``` On 4 test node we are observing this failure more prominently. ## Expected behavior It should create the namespace successfully and do ```odo push``` ## Any logs, error output, etc? Logs : https://prow.svc.ci.openshift.org/view/gcs/origin-ci-test/pr-logs/pull/openshift_odo/2965/pull-ci-openshift-odo-master-v4.4-integration-e2e-benchmark/431#1:build-log.txt%3A429
test
namespaces creation in parallel are not handled properly by the test script kind bug welcome we kindly ask you to fill out the issue template below use the google group if you have a question rather than a bug or feature request the group is at thanks for understanding and for contributing to the project what versions of software are you using operating system all supported output of odo version master how did you run odo exactly running test on travis ci actual behavior creating a new project bivsrltawh running odo with args β€’ waiting for project to come up occlient go status of creation of project bivsrltawh is active occlient go project bivsrltawh now exists occlient go status of creation of service account serviceaccount objectmeta io apimachinery pkg apis meta objectmeta name default generatename namespace bivsrltawh selflink api namespaces bivsrltawh serviceaccounts default uid resourceversion generation creationtimestamp utc deletiontimestamp deletiongraceperiodseconds nil labels map string annotations map string ownerreferences finalizers clustername initializers nil secrets imagepullsecrets automountserviceaccounttoken nil is ready βœ“ waiting for project to come up βœ“ project bivsrltawh is ready for use βœ“ new project created and now using project bivsrltawh odo go could not get the latest release information in time never mind exiting gracefully running odo with args validation β€’ validating component βœ“ validating component please use odo push command to create the component with source deployed running odo with args βœ— projectrequests project openshift io is forbidden user system serviceaccount ci op default cannot create projectrequests project openshift io at the cluster scope no rbac policy matched deleting project bivsrltawh running odo with args βœ— the project bivsrltawh does not exist please check the list of projects using odo project list β€’ failure odo url command tests go src github com openshift odo tests integration cmd url test go listing urls go src github com openshift odo tests integration cmd url test go should list appropriate urls and push message go src github com openshift odo tests integration cmd url test go no future change is possible bailing out early after running odo with args expected to match exit code on test node we are observing this failure more prominently expected behavior it should create the namespace successfully and do odo push any logs error output etc logs
1
78,546
27,584,301,109
IssuesEvent
2023-03-08 18:27:55
idaholab/malamute
https://api.github.com/repos/idaholab/malamute
closed
MALAMUTE issue templates are not compatible with GitHub
c: infrastructure p: normal t: defect
## Bug Description Given the transition to GitHub from GitLab, the issue templates are no longer compatible. ## Steps to Reproduce Try to make an issue - no templates are available. ## Impact Prevents easy usage of an issue template, in violation of SQA.
1.0
MALAMUTE issue templates are not compatible with GitHub - ## Bug Description Given the transition to GitHub from GitLab, the issue templates are no longer compatible. ## Steps to Reproduce Try to make an issue - no templates are available. ## Impact Prevents easy usage of an issue template, in violation of SQA.
non_test
malamute issue templates are not compatible with github bug description given the transition to github from gitlab the issue templates are no longer compatible steps to reproduce try to make an issue no templates are available impact prevents easy usage of an issue template in violation of sqa
0
3,396
4,288,418,671
IssuesEvent
2016-07-17 12:55:04
php-coder/mystamps
https://api.github.com/repos/php-coder/mystamps
closed
Update Spring Boot to 1.3.5.RELEASE
area/infrastructure in progress
- Update `pom.xml` - [x] Use new version (also update the comments) - Read the change logs - Spring Boot - [x] https://spring.io/blog/2015/12/18/spring-boot-1-3-1-and-1-2-8-available-now - [x] https://github.com/spring-projects/spring-boot/issues?q=milestone%3A1.3.1+is%3Aclosed - [x] https://spring.io/blog/2016/01/22/spring-boot-1-3-2-released - [x] https://github.com/spring-projects/spring-boot/issues?q=milestone%3A1.3.2+is%3Aclosed - [x] https://spring.io/blog/2016/02/26/spring-boot-1-3-3-and-1-4-0-m1-available-now - [x] https://github.com/spring-projects/spring-boot/issues?q=milestone%3A1.3.3 - [x] https://spring.io/blog/2016/05/06/spring-boot-1-3-4-released - [x] https://github.com/spring-projects/spring-boot/issues?q=milestone%3A1.3.4+is%3Aclosed - [x] https://spring.io/blog/2016/05/10/spring-boot-1-3-5-released - [x] https://github.com/spring-projects/spring-boot/issues?q=milestone%3A1.3.5+is%3Aclosed - `spring` (4.2.3.RELEASE -> 4.2.6.RELEASE) - [x] https://spring.io/blog/2015/12/17/spring-framework-4-2-4-4-1-9-released - [x] https://jira.spring.io/jira/secure/ReleaseNote.jspa?projectId=10000&version=15353 - [x] https://spring.io/blog/2016/02/25/spring-framework-4-2-5-available-now - [x] https://jira.spring.io/jira/secure/ReleaseNote.jspa?projectId=10000&version=15439 - [x] https://spring.io/blog/2016/05/06/spring-framework-4-3-rc2-4-2-6-and-3-2-17-available-now - [x] https://jira.spring.io/secure/ReleaseNote.jspa?projectId=10000&version=15460 - `spring-security` (4.0.3 -> 4.0.4) - [x] https://spring.io/blog/2016/02/26/spring-security-4-0-4-released - [x] https://github.com/spring-projects/spring-security/issues?utf8=%E2%9C%93&q=milestone%3A4.0.4 - `liquibase` (3.4.1 -> 3.4.2) - [x] www.liquibase.org/2015/11/liquibase-3-4-2-released.html - `jetty` (9.2.11.v20150529 -> 9.2.16.v20160414 because of `java.lang.UnsupportedOperationException: Use setIncludedMimeTypes or setExcludedMimeTypes instead`) - [x] https://dev.eclipse.org/mhonarc/lists/jetty-announce/msg00081.html - [x] https://dev.eclipse.org/mhonarc/lists/jetty-announce/msg00084.html - [x] https://dev.eclipse.org/mhonarc/lists/jetty-announce/msg00088.html - [x] https://github.com/eclipse/jetty.project/releases/tag/jetty-9.2.16.v20160414 - Migrate - [x] Compare dependencies - [x] Try to use `server.server-header` (Spring Boot GH 4461) - [x] Try `server.server-name` property (Spring Boot GH 4504) - [x] Configure dispatcher types (Spring Boot GH 4505) - [x] Remove `install: true` workaround (Spring Boot GH 4922) (resolution: I don't see the benefits) - [x] Try to use `AuthenticationAuditListener` (resolution: need more investigation and depends on #86, so we'll back to it later) - [x] Check `*.conf` file permissions on server (Spring Boot GH 4935) (resolution: so hard permissions (`0400`) on config file will make our auto-releases harder and I don't see any problem that someone could read the options of JVM)
1.0
Update Spring Boot to 1.3.5.RELEASE - - Update `pom.xml` - [x] Use new version (also update the comments) - Read the change logs - Spring Boot - [x] https://spring.io/blog/2015/12/18/spring-boot-1-3-1-and-1-2-8-available-now - [x] https://github.com/spring-projects/spring-boot/issues?q=milestone%3A1.3.1+is%3Aclosed - [x] https://spring.io/blog/2016/01/22/spring-boot-1-3-2-released - [x] https://github.com/spring-projects/spring-boot/issues?q=milestone%3A1.3.2+is%3Aclosed - [x] https://spring.io/blog/2016/02/26/spring-boot-1-3-3-and-1-4-0-m1-available-now - [x] https://github.com/spring-projects/spring-boot/issues?q=milestone%3A1.3.3 - [x] https://spring.io/blog/2016/05/06/spring-boot-1-3-4-released - [x] https://github.com/spring-projects/spring-boot/issues?q=milestone%3A1.3.4+is%3Aclosed - [x] https://spring.io/blog/2016/05/10/spring-boot-1-3-5-released - [x] https://github.com/spring-projects/spring-boot/issues?q=milestone%3A1.3.5+is%3Aclosed - `spring` (4.2.3.RELEASE -> 4.2.6.RELEASE) - [x] https://spring.io/blog/2015/12/17/spring-framework-4-2-4-4-1-9-released - [x] https://jira.spring.io/jira/secure/ReleaseNote.jspa?projectId=10000&version=15353 - [x] https://spring.io/blog/2016/02/25/spring-framework-4-2-5-available-now - [x] https://jira.spring.io/jira/secure/ReleaseNote.jspa?projectId=10000&version=15439 - [x] https://spring.io/blog/2016/05/06/spring-framework-4-3-rc2-4-2-6-and-3-2-17-available-now - [x] https://jira.spring.io/secure/ReleaseNote.jspa?projectId=10000&version=15460 - `spring-security` (4.0.3 -> 4.0.4) - [x] https://spring.io/blog/2016/02/26/spring-security-4-0-4-released - [x] https://github.com/spring-projects/spring-security/issues?utf8=%E2%9C%93&q=milestone%3A4.0.4 - `liquibase` (3.4.1 -> 3.4.2) - [x] www.liquibase.org/2015/11/liquibase-3-4-2-released.html - `jetty` (9.2.11.v20150529 -> 9.2.16.v20160414 because of `java.lang.UnsupportedOperationException: Use setIncludedMimeTypes or setExcludedMimeTypes instead`) - [x] https://dev.eclipse.org/mhonarc/lists/jetty-announce/msg00081.html - [x] https://dev.eclipse.org/mhonarc/lists/jetty-announce/msg00084.html - [x] https://dev.eclipse.org/mhonarc/lists/jetty-announce/msg00088.html - [x] https://github.com/eclipse/jetty.project/releases/tag/jetty-9.2.16.v20160414 - Migrate - [x] Compare dependencies - [x] Try to use `server.server-header` (Spring Boot GH 4461) - [x] Try `server.server-name` property (Spring Boot GH 4504) - [x] Configure dispatcher types (Spring Boot GH 4505) - [x] Remove `install: true` workaround (Spring Boot GH 4922) (resolution: I don't see the benefits) - [x] Try to use `AuthenticationAuditListener` (resolution: need more investigation and depends on #86, so we'll back to it later) - [x] Check `*.conf` file permissions on server (Spring Boot GH 4935) (resolution: so hard permissions (`0400`) on config file will make our auto-releases harder and I don't see any problem that someone could read the options of JVM)
non_test
update spring boot to release update pom xml use new version also update the comments read the change logs spring boot spring release release spring security liquibase jetty because of java lang unsupportedoperationexception use setincludedmimetypes or setexcludedmimetypes instead migrate compare dependencies try to use server server header spring boot gh try server server name property spring boot gh configure dispatcher types spring boot gh remove install true workaround spring boot gh resolution i don t see the benefits try to use authenticationauditlistener resolution need more investigation and depends on so we ll back to it later check conf file permissions on server spring boot gh resolution so hard permissions on config file will make our auto releases harder and i don t see any problem that someone could read the options of jvm
0
140,059
21,004,376,012
IssuesEvent
2022-03-29 20:49:38
WatershedXiaolan/Xiaolan-s-Gitbook
https://api.github.com/repos/WatershedXiaolan/Xiaolan-s-Gitbook
opened
Long-Polling vs WebSockets vs Server-Sent Events
system design
They are communication protocols between a client like a web browser and a web server. **Ajax Polling[#](https://www.educative.io/courses/grokking-the-system-design-interview/gx7wZzWn5Vj#Ajax-Polling)** The basic idea is that the client repeatedly polls (or requests) a server for data. If no data is available, an empty response is returned. <img width="675" alt="Screen Shot 2022-03-29 at 3 47 54 PM" src="https://user-images.githubusercontent.com/5552371/160704427-bb6fad23-38c2-4519-8b5f-3027c16896f4.png"> **HTTP Long-Polling[#](https://www.educative.io/courses/grokking-the-system-design-interview/gx7wZzWn5Vj#HTTP-Long-Polling)** With Long-Polling, the client requests information from the server exactly as in normal polling, but with the expectation that the server may not respond immediately. Each Long-Poll request has a timeout. The client has to reconnect periodically after the connection is closed due to timeouts. <img width="625" alt="Screen Shot 2022-03-29 at 3 47 59 PM" src="https://user-images.githubusercontent.com/5552371/160704448-77c0f40b-bbf4-4229-97d8-171e4245f38a.png"> **WebSockets[#](https://www.educative.io/courses/grokking-the-system-design-interview/gx7wZzWn5Vj#WebSockets)** - It provides a persistent connection between a client and a server that both parties can use to start sending data at any time. - The client establishes a WebSocket connection through a process known as the WebSocket handshake. 、 - If the process succeeds, then the server and client can exchange data in both directions at any time. - The WebSocket protocol enables the communication between a client and a server with lower overheads, facilitating real-time data transfer from and to the server. - **a two-way (bi-directional) ongoing conversation can take place between a client and a server.** <img width="632" alt="Screen Shot 2022-03-29 at 3 48 04 PM" src="https://user-images.githubusercontent.com/5552371/160704476-1c9c91f1-c4dc-4fda-a67d-129673bd5b29.png"> **Server-Sent Events (SSEs)[#](https://www.educative.io/courses/grokking-the-system-design-interview/gx7wZzWn5Vj#Server-Sent-Events-(SSEs))** - Under SSEs the client establishes a persistent and long-term connection with the server. The server uses this connection to send data to a client. - If the client wants to send data to the server, it would require the use of another technology/protocol to do so. - SSEs are best when we need real-time traffic from the server to the client or if the server is generating data in a loop and will be sending multiple events to the client. - <img width="647" alt="Screen Shot 2022-03-29 at 3 48 08 PM" src="https://user-images.githubusercontent.com/5552371/160704497-2d545150-3afc-4119-88ff-4b39f3b7c6e3.png">
1.0
Long-Polling vs WebSockets vs Server-Sent Events - They are communication protocols between a client like a web browser and a web server. **Ajax Polling[#](https://www.educative.io/courses/grokking-the-system-design-interview/gx7wZzWn5Vj#Ajax-Polling)** The basic idea is that the client repeatedly polls (or requests) a server for data. If no data is available, an empty response is returned. <img width="675" alt="Screen Shot 2022-03-29 at 3 47 54 PM" src="https://user-images.githubusercontent.com/5552371/160704427-bb6fad23-38c2-4519-8b5f-3027c16896f4.png"> **HTTP Long-Polling[#](https://www.educative.io/courses/grokking-the-system-design-interview/gx7wZzWn5Vj#HTTP-Long-Polling)** With Long-Polling, the client requests information from the server exactly as in normal polling, but with the expectation that the server may not respond immediately. Each Long-Poll request has a timeout. The client has to reconnect periodically after the connection is closed due to timeouts. <img width="625" alt="Screen Shot 2022-03-29 at 3 47 59 PM" src="https://user-images.githubusercontent.com/5552371/160704448-77c0f40b-bbf4-4229-97d8-171e4245f38a.png"> **WebSockets[#](https://www.educative.io/courses/grokking-the-system-design-interview/gx7wZzWn5Vj#WebSockets)** - It provides a persistent connection between a client and a server that both parties can use to start sending data at any time. - The client establishes a WebSocket connection through a process known as the WebSocket handshake. 、 - If the process succeeds, then the server and client can exchange data in both directions at any time. - The WebSocket protocol enables the communication between a client and a server with lower overheads, facilitating real-time data transfer from and to the server. - **a two-way (bi-directional) ongoing conversation can take place between a client and a server.** <img width="632" alt="Screen Shot 2022-03-29 at 3 48 04 PM" src="https://user-images.githubusercontent.com/5552371/160704476-1c9c91f1-c4dc-4fda-a67d-129673bd5b29.png"> **Server-Sent Events (SSEs)[#](https://www.educative.io/courses/grokking-the-system-design-interview/gx7wZzWn5Vj#Server-Sent-Events-(SSEs))** - Under SSEs the client establishes a persistent and long-term connection with the server. The server uses this connection to send data to a client. - If the client wants to send data to the server, it would require the use of another technology/protocol to do so. - SSEs are best when we need real-time traffic from the server to the client or if the server is generating data in a loop and will be sending multiple events to the client. - <img width="647" alt="Screen Shot 2022-03-29 at 3 48 08 PM" src="https://user-images.githubusercontent.com/5552371/160704497-2d545150-3afc-4119-88ff-4b39f3b7c6e3.png">
non_test
long polling vs websockets vs server sent events they are communication protocols between a client like a web browser and a web server ajax polling the basic idea is that the client repeatedly polls or requests a server for data if no data is available an empty response is returned img width alt screen shot at pm src http long polling with long polling the client requests information from the server exactly as in normal polling but with the expectation that the server may not respond immediately each long poll request has a timeout the client has to reconnect periodically after the connection is closed due to timeouts img width alt screen shot at pm src websockets it provides a persistent connection between a client and a server that both parties can use to start sending data at any time the client establishes a websocket connection through a process known as the websocket handshake 、 if the process succeeds then the server and client can exchange data in both directions at any time the websocket protocol enables the communication between a client and a server with lower overheads facilitating real time data transfer from and to the server a two way bi directional ongoing conversation can take place between a client and a server img width alt screen shot at pm src server sent events sses under sses the client establishes a persistent and long term connection with the server the server uses this connection to send data to a client if the client wants to send data to the server it would require the use of another technology protocol to do so sses are best when we need real time traffic from the server to the client or if the server is generating data in a loop and will be sending multiple events to the client img width alt screen shot at pm src
0
11,070
3,171,971,194
IssuesEvent
2015-09-23 03:16:15
balderdashy/sails-mysql
https://api.github.com/repos/balderdashy/sails-mysql
closed
Table names like "schemas" need to be backticked
needs review needs testing
```` module.exports = { adapter : 'mysql-default', migrate : 'safe', tableName : '`schemas`', attributes: { /* e.g. nickname: 'string' */ } }; ````
1.0
Table names like "schemas" need to be backticked - ```` module.exports = { adapter : 'mysql-default', migrate : 'safe', tableName : '`schemas`', attributes: { /* e.g. nickname: 'string' */ } }; ````
test
table names like schemas need to be backticked module exports adapter mysql default migrate safe tablename schemas attributes e g nickname string
1
15,466
27,227,695,398
IssuesEvent
2023-02-21 10:52:27
renovatebot/renovate
https://api.github.com/repos/renovatebot/renovate
opened
Lazy onboarding PR refresh
type:feature priority-2-high status:requirements performance
### What would you like Renovate to be able to do? Avoid refreshing onboarding PRs every time there's a commit to the default branch. Onboarding PRs sometime remain open for a long time and users don't need it refreshed non-stop. ### If you have any ideas on how this should be implemented, please tell us here. For platforms which support checkboxes, only refresh onboarding PRs when either: - a commit is made to the onboarding branch, or - users tick a (new) checkbox in the onboarding PR which says something like "Refresh this PR". If the PR isn't checked or the branch not modified, skip cloning and any other work. We might need to do some brainstorming on whether the first point requires repositoryCache to be enabled or not. One alternative is to embed the branch commit SHA into the PR body ### Is this a feature you are interested in implementing yourself? No
1.0
Lazy onboarding PR refresh - ### What would you like Renovate to be able to do? Avoid refreshing onboarding PRs every time there's a commit to the default branch. Onboarding PRs sometime remain open for a long time and users don't need it refreshed non-stop. ### If you have any ideas on how this should be implemented, please tell us here. For platforms which support checkboxes, only refresh onboarding PRs when either: - a commit is made to the onboarding branch, or - users tick a (new) checkbox in the onboarding PR which says something like "Refresh this PR". If the PR isn't checked or the branch not modified, skip cloning and any other work. We might need to do some brainstorming on whether the first point requires repositoryCache to be enabled or not. One alternative is to embed the branch commit SHA into the PR body ### Is this a feature you are interested in implementing yourself? No
non_test
lazy onboarding pr refresh what would you like renovate to be able to do avoid refreshing onboarding prs every time there s a commit to the default branch onboarding prs sometime remain open for a long time and users don t need it refreshed non stop if you have any ideas on how this should be implemented please tell us here for platforms which support checkboxes only refresh onboarding prs when either a commit is made to the onboarding branch or users tick a new checkbox in the onboarding pr which says something like refresh this pr if the pr isn t checked or the branch not modified skip cloning and any other work we might need to do some brainstorming on whether the first point requires repositorycache to be enabled or not one alternative is to embed the branch commit sha into the pr body is this a feature you are interested in implementing yourself no
0
56,012
14,078,373,087
IssuesEvent
2020-11-04 13:28:57
themagicalmammal/android_kernel_samsung_j7elte
https://api.github.com/repos/themagicalmammal/android_kernel_samsung_j7elte
opened
CVE-2013-7268 (Medium) detected in linuxlinux-3.13.5
security vulnerability
## CVE-2013-7268 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-3.13.5</b></p></summary> <p> <p>Apache Software Foundation (ASF)</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v3.0/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v3.0/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/themagicalmammal/android_kernel_samsung_j7elte/commit/adc86a86e0ac98007fd3af905bc71e9f29c1502c">adc86a86e0ac98007fd3af905bc71e9f29c1502c</a></p> <p>Found in base branch: <b>xsentinel-1.7-experimental</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (0)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The ipx_recvmsg function in net/ipx/af_ipx.c in the Linux kernel before 3.12.4 updates a certain length value without ensuring that an associated data structure has been initialized, which allows local users to obtain sensitive information from kernel memory via a (1) recvfrom, (2) recvmmsg, or (3) recvmsg system call. <p>Publish Date: 2014-01-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2013-7268>CVE-2013-7268</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>4.9</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2013-7268">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2013-7268</a></p> <p>Release Date: 2014-01-06</p> <p>Fix Resolution: v3.13-rc1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2013-7268 (Medium) detected in linuxlinux-3.13.5 - ## CVE-2013-7268 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-3.13.5</b></p></summary> <p> <p>Apache Software Foundation (ASF)</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v3.0/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v3.0/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/themagicalmammal/android_kernel_samsung_j7elte/commit/adc86a86e0ac98007fd3af905bc71e9f29c1502c">adc86a86e0ac98007fd3af905bc71e9f29c1502c</a></p> <p>Found in base branch: <b>xsentinel-1.7-experimental</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (0)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The ipx_recvmsg function in net/ipx/af_ipx.c in the Linux kernel before 3.12.4 updates a certain length value without ensuring that an associated data structure has been initialized, which allows local users to obtain sensitive information from kernel memory via a (1) recvfrom, (2) recvmmsg, or (3) recvmsg system call. <p>Publish Date: 2014-01-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2013-7268>CVE-2013-7268</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>4.9</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2013-7268">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2013-7268</a></p> <p>Release Date: 2014-01-06</p> <p>Fix Resolution: v3.13-rc1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve medium detected in linuxlinux cve medium severity vulnerability vulnerable library linuxlinux apache software foundation asf library home page a href found in head commit a href found in base branch xsentinel experimental vulnerable source files vulnerability details the ipx recvmsg function in net ipx af ipx c in the linux kernel before updates a certain length value without ensuring that an associated data structure has been initialized which allows local users to obtain sensitive information from kernel memory via a recvfrom recvmmsg or recvmsg system call publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
262,444
22,841,261,639
IssuesEvent
2022-07-12 22:13:46
mapbox/mapbox-gl-js
https://api.github.com/repos/mapbox/mapbox-gl-js
closed
`dynamic-filter/symbols/line/distance-nofilter` render test is flaky
testing :100:
`render-tests/dynamic-filter/symbols/line/distance-nofilter` failed once on a completely unrelated change. We should fix it
1.0
`dynamic-filter/symbols/line/distance-nofilter` render test is flaky - `render-tests/dynamic-filter/symbols/line/distance-nofilter` failed once on a completely unrelated change. We should fix it
test
dynamic filter symbols line distance nofilter render test is flaky render tests dynamic filter symbols line distance nofilter failed once on a completely unrelated change we should fix it
1
66,634
3,256,614,748
IssuesEvent
2015-10-20 14:32:38
ox-it/ords
https://api.github.com/repos/ox-it/ords
closed
Saved datasets can't be re-run since ODBC code was amended
auto-migrated David Paine Priority-Critical
Attempting to run a saved dynamic query generated the following error: Unable to run query. Possible causes include malformed SQL, or a database error. Please check your query and try again. If you think you have encountered a problem with the ORDS system, please contact the help desk by emailing ords@it.ox.ac.uk Message was: org.postgresql.util.PSQLException: ERROR: relation "country" does not exist Position: 40 Attempting to run a saved static query results in a stack trace error: HTTP Status 500 - Unable to compile class for JSP: An error occurred at line: 67 in the jsp file: /queryResult.jsp The constructor DBUtils(String, String, OrdsUser) is undefined 64: Project project = DBGateway.getProject(database.getDatabaseProjectId()); 65: String cleanProjectName = StringEscapeUtils.escapeHtml4(project.getName()); 66: OrdsPhysicalDatabase opd = DBGatewayProjectNode.getPhysicalDatabaseByPhysicalDatabaseId(Integer.parseInt(cleanPhysicalId)); 67: DBUtils dbUtils = new DBUtils(DBGateway.getDbServerByProjectId(cleanProjectId), opd.getDbConsumedName(), user); 68: Collection<DataRow> tableList = ORDSDatabaseUtils.getTableListForMainDatabase(opd.getLogicalDatabaseId(), user); 69: boolean advanced = true; 70: if ((request.getParameter("qtype") == null) || (request.getParameter("qtype").equals("simple"))) { Stacktrace: type Exception report message Unable to compile class for JSP: An error occurred at line: 67 in the jsp file: /queryResult.jsp The constructor DBUtils(String, String, OrdsUser) is undefined 64: Project project = DBGateway.getProject(database.getDatabaseProjectId()); 65: String cleanProjectName = StringEscapeUtils.escapeHtml4(project.getName()); 66: OrdsPhysicalDatabase opd = DBGatewayProjectNode.getPhysicalDatabaseByPhysicalDatabaseId(Integer.parseInt(cleanPhysicalId)); 67: DBUtils dbUtils = new DBUtils(DBGateway.getDbServerByProjectId(cleanProjectId), opd.getDbConsumedName(), user); 68: Collection<DataRow> tableList = ORDSDatabaseUtils.getTableListForMainDatabase(opd.getLogicalDatabaseId(), user); 69: boolean advanced = true; 70: if ((request.getParameter("qtype") == null) || (request.getParameter("qtype").equals("simple"))) { Stacktrace: description The server encountered an internal error (Unable to compile class for JSP: An error occurred at line: 67 in the jsp file: /queryResult.jsp The constructor DBUtils(String, String, OrdsUser) is undefined 64: Project project = DBGateway.getProject(database.getDatabaseProjectId()); 65: String cleanProjectName = StringEscapeUtils.escapeHtml4(project.getName()); 66: OrdsPhysicalDatabase opd = DBGatewayProjectNode.getPhysicalDatabaseByPhysicalDatabaseId(Integer.parseInt(cleanPhysicalId)); 67: DBUtils dbUtils = new DBUtils(DBGateway.getDbServerByProjectId(cleanProjectId), opd.getDbConsumedName(), user); 68: Collection<DataRow> tableList = ORDSDatabaseUtils.getTableListForMainDatabase(opd.getLogicalDatabaseId(), user); 69: boolean advanced = true; 70: if ((request.getParameter("qtype") == null) || (request.getParameter("qtype").equals("simple"))) { Stacktrace:) that prevented it from fulfilling this request. exception org.apache.jasper.JasperException: Unable to compile class for JSP: An error occurred at line: 67 in the jsp file: /queryResult.jsp The constructor DBUtils(String, String, OrdsUser) is undefined 64: Project project = DBGateway.getProject(database.getDatabaseProjectId()); 65: String cleanProjectName = StringEscapeUtils.escapeHtml4(project.getName()); 66: OrdsPhysicalDatabase opd = DBGatewayProjectNode.getPhysicalDatabaseByPhysicalDatabaseId(Integer.parseInt(cleanPhysicalId)); 67: DBUtils dbUtils = new DBUtils(DBGateway.getDbServerByProjectId(cleanProjectId), opd.getDbConsumedName(), user); 68: Collection<DataRow> tableList = ORDSDatabaseUtils.getTableListForMainDatabase(opd.getLogicalDatabaseId(), user); 69: boolean advanced = true; 70: if ((request.getParameter("qtype") == null) || (request.getParameter("qtype").equals("simple"))) { Stacktrace: org.apache.jasper.compiler.DefaultErrorHandler.javacError(DefaultErrorHandler.java:102) org.apache.jasper.compiler.ErrorDispatcher.javacError(ErrorDispatcher.java:331) org.apache.jasper.compiler.JDTCompiler.generateClass(JDTCompiler.java:469) org.apache.jasper.compiler.Compiler.compile(Compiler.java:378) org.apache.jasper.compiler.Compiler.compile(Compiler.java:353) org.apache.jasper.compiler.Compiler.compile(Compiler.java:340) org.apache.jasper.JspCompilationContext.compile(JspCompilationContext.java:646) org.apache.jasper.servlet.JspServletWrapper.service(JspServletWrapper.java:357) org.apache.jasper.servlet.JspServlet.serviceJspFile(JspServlet.java:390) org.apache.jasper.servlet.JspServlet.service(JspServlet.java:334) javax.servlet.http.HttpServlet.service(HttpServlet.java:722) note The full stack trace of the root cause is available in the Apache Tomcat/7.0.28 logs. Apache Tomcat/7.0.28
1.0
Saved datasets can't be re-run since ODBC code was amended - Attempting to run a saved dynamic query generated the following error: Unable to run query. Possible causes include malformed SQL, or a database error. Please check your query and try again. If you think you have encountered a problem with the ORDS system, please contact the help desk by emailing ords@it.ox.ac.uk Message was: org.postgresql.util.PSQLException: ERROR: relation "country" does not exist Position: 40 Attempting to run a saved static query results in a stack trace error: HTTP Status 500 - Unable to compile class for JSP: An error occurred at line: 67 in the jsp file: /queryResult.jsp The constructor DBUtils(String, String, OrdsUser) is undefined 64: Project project = DBGateway.getProject(database.getDatabaseProjectId()); 65: String cleanProjectName = StringEscapeUtils.escapeHtml4(project.getName()); 66: OrdsPhysicalDatabase opd = DBGatewayProjectNode.getPhysicalDatabaseByPhysicalDatabaseId(Integer.parseInt(cleanPhysicalId)); 67: DBUtils dbUtils = new DBUtils(DBGateway.getDbServerByProjectId(cleanProjectId), opd.getDbConsumedName(), user); 68: Collection<DataRow> tableList = ORDSDatabaseUtils.getTableListForMainDatabase(opd.getLogicalDatabaseId(), user); 69: boolean advanced = true; 70: if ((request.getParameter("qtype") == null) || (request.getParameter("qtype").equals("simple"))) { Stacktrace: type Exception report message Unable to compile class for JSP: An error occurred at line: 67 in the jsp file: /queryResult.jsp The constructor DBUtils(String, String, OrdsUser) is undefined 64: Project project = DBGateway.getProject(database.getDatabaseProjectId()); 65: String cleanProjectName = StringEscapeUtils.escapeHtml4(project.getName()); 66: OrdsPhysicalDatabase opd = DBGatewayProjectNode.getPhysicalDatabaseByPhysicalDatabaseId(Integer.parseInt(cleanPhysicalId)); 67: DBUtils dbUtils = new DBUtils(DBGateway.getDbServerByProjectId(cleanProjectId), opd.getDbConsumedName(), user); 68: Collection<DataRow> tableList = ORDSDatabaseUtils.getTableListForMainDatabase(opd.getLogicalDatabaseId(), user); 69: boolean advanced = true; 70: if ((request.getParameter("qtype") == null) || (request.getParameter("qtype").equals("simple"))) { Stacktrace: description The server encountered an internal error (Unable to compile class for JSP: An error occurred at line: 67 in the jsp file: /queryResult.jsp The constructor DBUtils(String, String, OrdsUser) is undefined 64: Project project = DBGateway.getProject(database.getDatabaseProjectId()); 65: String cleanProjectName = StringEscapeUtils.escapeHtml4(project.getName()); 66: OrdsPhysicalDatabase opd = DBGatewayProjectNode.getPhysicalDatabaseByPhysicalDatabaseId(Integer.parseInt(cleanPhysicalId)); 67: DBUtils dbUtils = new DBUtils(DBGateway.getDbServerByProjectId(cleanProjectId), opd.getDbConsumedName(), user); 68: Collection<DataRow> tableList = ORDSDatabaseUtils.getTableListForMainDatabase(opd.getLogicalDatabaseId(), user); 69: boolean advanced = true; 70: if ((request.getParameter("qtype") == null) || (request.getParameter("qtype").equals("simple"))) { Stacktrace:) that prevented it from fulfilling this request. exception org.apache.jasper.JasperException: Unable to compile class for JSP: An error occurred at line: 67 in the jsp file: /queryResult.jsp The constructor DBUtils(String, String, OrdsUser) is undefined 64: Project project = DBGateway.getProject(database.getDatabaseProjectId()); 65: String cleanProjectName = StringEscapeUtils.escapeHtml4(project.getName()); 66: OrdsPhysicalDatabase opd = DBGatewayProjectNode.getPhysicalDatabaseByPhysicalDatabaseId(Integer.parseInt(cleanPhysicalId)); 67: DBUtils dbUtils = new DBUtils(DBGateway.getDbServerByProjectId(cleanProjectId), opd.getDbConsumedName(), user); 68: Collection<DataRow> tableList = ORDSDatabaseUtils.getTableListForMainDatabase(opd.getLogicalDatabaseId(), user); 69: boolean advanced = true; 70: if ((request.getParameter("qtype") == null) || (request.getParameter("qtype").equals("simple"))) { Stacktrace: org.apache.jasper.compiler.DefaultErrorHandler.javacError(DefaultErrorHandler.java:102) org.apache.jasper.compiler.ErrorDispatcher.javacError(ErrorDispatcher.java:331) org.apache.jasper.compiler.JDTCompiler.generateClass(JDTCompiler.java:469) org.apache.jasper.compiler.Compiler.compile(Compiler.java:378) org.apache.jasper.compiler.Compiler.compile(Compiler.java:353) org.apache.jasper.compiler.Compiler.compile(Compiler.java:340) org.apache.jasper.JspCompilationContext.compile(JspCompilationContext.java:646) org.apache.jasper.servlet.JspServletWrapper.service(JspServletWrapper.java:357) org.apache.jasper.servlet.JspServlet.serviceJspFile(JspServlet.java:390) org.apache.jasper.servlet.JspServlet.service(JspServlet.java:334) javax.servlet.http.HttpServlet.service(HttpServlet.java:722) note The full stack trace of the root cause is available in the Apache Tomcat/7.0.28 logs. Apache Tomcat/7.0.28
non_test
saved datasets can t be re run since odbc code was amended attempting to run a saved dynamic query generated the following error unable to run query possible causes include malformed sql or a database error please check your query and try again if you think you have encountered a problem with the ords system please contact the help desk by emailing ords it ox ac uk message was org postgresql util psqlexception error relation country does not exist position attempting to run a saved static query results in a stack trace error http status unable to compile class for jsp an error occurred at line in the jsp file queryresult jsp the constructor dbutils string string ordsuser is undefined project project dbgateway getproject database getdatabaseprojectid string cleanprojectname stringescapeutils project getname ordsphysicaldatabase opd dbgatewayprojectnode getphysicaldatabasebyphysicaldatabaseid integer parseint cleanphysicalid dbutils dbutils new dbutils dbgateway getdbserverbyprojectid cleanprojectid opd getdbconsumedname user collection tablelist ordsdatabaseutils gettablelistformaindatabase opd getlogicaldatabaseid user boolean advanced true if request getparameter qtype null request getparameter qtype equals simple stacktrace type exception report message unable to compile class for jsp an error occurred at line in the jsp file queryresult jsp the constructor dbutils string string ordsuser is undefined project project dbgateway getproject database getdatabaseprojectid string cleanprojectname stringescapeutils project getname ordsphysicaldatabase opd dbgatewayprojectnode getphysicaldatabasebyphysicaldatabaseid integer parseint cleanphysicalid dbutils dbutils new dbutils dbgateway getdbserverbyprojectid cleanprojectid opd getdbconsumedname user collection tablelist ordsdatabaseutils gettablelistformaindatabase opd getlogicaldatabaseid user boolean advanced true if request getparameter qtype null request getparameter qtype equals simple stacktrace description the server encountered an internal error unable to compile class for jsp an error occurred at line in the jsp file queryresult jsp the constructor dbutils string string ordsuser is undefined project project dbgateway getproject database getdatabaseprojectid string cleanprojectname stringescapeutils project getname ordsphysicaldatabase opd dbgatewayprojectnode getphysicaldatabasebyphysicaldatabaseid integer parseint cleanphysicalid dbutils dbutils new dbutils dbgateway getdbserverbyprojectid cleanprojectid opd getdbconsumedname user collection tablelist ordsdatabaseutils gettablelistformaindatabase opd getlogicaldatabaseid user boolean advanced true if request getparameter qtype null request getparameter qtype equals simple stacktrace that prevented it from fulfilling this request exception org apache jasper jasperexception unable to compile class for jsp an error occurred at line in the jsp file queryresult jsp the constructor dbutils string string ordsuser is undefined project project dbgateway getproject database getdatabaseprojectid string cleanprojectname stringescapeutils project getname ordsphysicaldatabase opd dbgatewayprojectnode getphysicaldatabasebyphysicaldatabaseid integer parseint cleanphysicalid dbutils dbutils new dbutils dbgateway getdbserverbyprojectid cleanprojectid opd getdbconsumedname user collection tablelist ordsdatabaseutils gettablelistformaindatabase opd getlogicaldatabaseid user boolean advanced true if request getparameter qtype null request getparameter qtype equals simple stacktrace org apache jasper compiler defaulterrorhandler javacerror defaulterrorhandler java org apache jasper compiler errordispatcher javacerror errordispatcher java org apache jasper compiler jdtcompiler generateclass jdtcompiler java org apache jasper compiler compiler compile compiler java org apache jasper compiler compiler compile compiler java org apache jasper compiler compiler compile compiler java org apache jasper jspcompilationcontext compile jspcompilationcontext java org apache jasper servlet jspservletwrapper service jspservletwrapper java org apache jasper servlet jspservlet servicejspfile jspservlet java org apache jasper servlet jspservlet service jspservlet java javax servlet http httpservlet service httpservlet java note the full stack trace of the root cause is available in the apache tomcat logs apache tomcat
0
85,801
10,684,584,740
IssuesEvent
2019-10-22 10:48:31
korsmakolnikov/hauer
https://api.github.com/repos/korsmakolnikov/hauer
closed
Design command line commands to manipulate endpoints
Design help wanted question
Let's discuss how command line commands should be. How we want them to work. This is related with #1 since different approach are required if we build the router at compile time. Commands should recompile the project after write the routes config file in that case. This task should start from analysis of loopback command line utilities.
1.0
Design command line commands to manipulate endpoints - Let's discuss how command line commands should be. How we want them to work. This is related with #1 since different approach are required if we build the router at compile time. Commands should recompile the project after write the routes config file in that case. This task should start from analysis of loopback command line utilities.
non_test
design command line commands to manipulate endpoints let s discuss how command line commands should be how we want them to work this is related with since different approach are required if we build the router at compile time commands should recompile the project after write the routes config file in that case this task should start from analysis of loopback command line utilities
0
258,578
22,328,713,913
IssuesEvent
2022-06-14 12:56:34
elastic/elasticsearch
https://api.github.com/repos/elastic/elasticsearch
closed
[CI] FrozenSearchableSnapshotsIntegTests testCreateAndRestorePartialSearchableSnapshot failing
:Distributed/Snapshot/Restore >test-failure Team:Distributed
**Build scan:** https://gradle-enterprise.elastic.co/s/nzkgnqet7vwt4/tests/:x-pack:plugin:searchable-snapshots:internalClusterTest/org.elasticsearch.xpack.searchablesnapshots.FrozenSearchableSnapshotsIntegTests/testCreateAndRestorePartialSearchableSnapshot **Reproduction line:** `./gradlew ':x-pack:plugin:searchable-snapshots:internalClusterTest' --tests "org.elasticsearch.xpack.searchablesnapshots.FrozenSearchableSnapshotsIntegTests.testCreateAndRestorePartialSearchableSnapshot" -Dtests.seed=AC36A259FD46B46A -Dtests.locale=no -Dtests.timezone=Hongkong -Druntime.java=18` **Applicable branches:** 8.2 **Reproduces locally?:** Didn't try **Failure history:** https://gradle-enterprise.elastic.co/scans/tests?tests.container=org.elasticsearch.xpack.searchablesnapshots.FrozenSearchableSnapshotsIntegTests&tests.test=testCreateAndRestorePartialSearchableSnapshot **Failure excerpt:** ``` com.carrotsearch.randomizedtesting.UncaughtExceptionError: Captured an uncaught exception in thread: Thread[id=44, name=elasticsearch[node_s0][masterService#updateTask][T#1], state=RUNNABLE, group=TGRP-FrozenSearchableSnapshotsIntegTests] at __randomizedtesting.SeedInfo.seed([AC36A259FD46B46A:D8B16388C0219A88]:0) Caused by: java.lang.AssertionError: (No message provided) at __randomizedtesting.SeedInfo.seed([AC36A259FD46B46A]:0) at org.elasticsearch.cluster.routing.allocation.IndexMetadataUpdater.applyChanges(IndexMetadataUpdater.java:134) at org.elasticsearch.cluster.routing.allocation.RoutingAllocation.updateMetadataWithRoutingChanges(RoutingAllocation.java:274) at org.elasticsearch.cluster.routing.allocation.AllocationService.buildResultAndLogHealthChange(AllocationService.java:150) at org.elasticsearch.cluster.routing.allocation.AllocationService.reroute(AllocationService.java:489) at org.elasticsearch.cluster.routing.BatchedRerouteService$1.execute(BatchedRerouteService.java:116) at org.elasticsearch.cluster.ClusterStateTaskExecutor$1.execute(ClusterStateTaskExecutor.java:95) at org.elasticsearch.cluster.service.MasterService.innerExecuteTasks(MasterService.java:901) at org.elasticsearch.cluster.service.MasterService.executeTasks(MasterService.java:871) at org.elasticsearch.cluster.service.MasterService.runTasks(MasterService.java:253) at org.elasticsearch.cluster.service.MasterService$Batcher.run(MasterService.java:153) at org.elasticsearch.cluster.service.TaskBatcher.runIfNotProcessed(TaskBatcher.java:114) at org.elasticsearch.cluster.service.TaskBatcher$BatchedTask.run(TaskBatcher.java:170) at org.elasticsearch.common.util.concurrent.ThreadContext$ContextPreservingRunnable.run(ThreadContext.java:714) at org.elasticsearch.common.util.concurrent.PrioritizedEsThreadPoolExecutor$TieBreakingPrioritizedRunnable.runAndClean(PrioritizedEsThreadPoolExecutor.java:260) at org.elasticsearch.common.util.concurrent.PrioritizedEsThreadPoolExecutor$TieBreakingPrioritizedRunnable.run(PrioritizedEsThreadPoolExecutor.java:223) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) at java.lang.Thread.run(Thread.java:833) ```
1.0
[CI] FrozenSearchableSnapshotsIntegTests testCreateAndRestorePartialSearchableSnapshot failing - **Build scan:** https://gradle-enterprise.elastic.co/s/nzkgnqet7vwt4/tests/:x-pack:plugin:searchable-snapshots:internalClusterTest/org.elasticsearch.xpack.searchablesnapshots.FrozenSearchableSnapshotsIntegTests/testCreateAndRestorePartialSearchableSnapshot **Reproduction line:** `./gradlew ':x-pack:plugin:searchable-snapshots:internalClusterTest' --tests "org.elasticsearch.xpack.searchablesnapshots.FrozenSearchableSnapshotsIntegTests.testCreateAndRestorePartialSearchableSnapshot" -Dtests.seed=AC36A259FD46B46A -Dtests.locale=no -Dtests.timezone=Hongkong -Druntime.java=18` **Applicable branches:** 8.2 **Reproduces locally?:** Didn't try **Failure history:** https://gradle-enterprise.elastic.co/scans/tests?tests.container=org.elasticsearch.xpack.searchablesnapshots.FrozenSearchableSnapshotsIntegTests&tests.test=testCreateAndRestorePartialSearchableSnapshot **Failure excerpt:** ``` com.carrotsearch.randomizedtesting.UncaughtExceptionError: Captured an uncaught exception in thread: Thread[id=44, name=elasticsearch[node_s0][masterService#updateTask][T#1], state=RUNNABLE, group=TGRP-FrozenSearchableSnapshotsIntegTests] at __randomizedtesting.SeedInfo.seed([AC36A259FD46B46A:D8B16388C0219A88]:0) Caused by: java.lang.AssertionError: (No message provided) at __randomizedtesting.SeedInfo.seed([AC36A259FD46B46A]:0) at org.elasticsearch.cluster.routing.allocation.IndexMetadataUpdater.applyChanges(IndexMetadataUpdater.java:134) at org.elasticsearch.cluster.routing.allocation.RoutingAllocation.updateMetadataWithRoutingChanges(RoutingAllocation.java:274) at org.elasticsearch.cluster.routing.allocation.AllocationService.buildResultAndLogHealthChange(AllocationService.java:150) at org.elasticsearch.cluster.routing.allocation.AllocationService.reroute(AllocationService.java:489) at org.elasticsearch.cluster.routing.BatchedRerouteService$1.execute(BatchedRerouteService.java:116) at org.elasticsearch.cluster.ClusterStateTaskExecutor$1.execute(ClusterStateTaskExecutor.java:95) at org.elasticsearch.cluster.service.MasterService.innerExecuteTasks(MasterService.java:901) at org.elasticsearch.cluster.service.MasterService.executeTasks(MasterService.java:871) at org.elasticsearch.cluster.service.MasterService.runTasks(MasterService.java:253) at org.elasticsearch.cluster.service.MasterService$Batcher.run(MasterService.java:153) at org.elasticsearch.cluster.service.TaskBatcher.runIfNotProcessed(TaskBatcher.java:114) at org.elasticsearch.cluster.service.TaskBatcher$BatchedTask.run(TaskBatcher.java:170) at org.elasticsearch.common.util.concurrent.ThreadContext$ContextPreservingRunnable.run(ThreadContext.java:714) at org.elasticsearch.common.util.concurrent.PrioritizedEsThreadPoolExecutor$TieBreakingPrioritizedRunnable.runAndClean(PrioritizedEsThreadPoolExecutor.java:260) at org.elasticsearch.common.util.concurrent.PrioritizedEsThreadPoolExecutor$TieBreakingPrioritizedRunnable.run(PrioritizedEsThreadPoolExecutor.java:223) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) at java.lang.Thread.run(Thread.java:833) ```
test
frozensearchablesnapshotsintegtests testcreateandrestorepartialsearchablesnapshot failing build scan reproduction line gradlew x pack plugin searchable snapshots internalclustertest tests org elasticsearch xpack searchablesnapshots frozensearchablesnapshotsintegtests testcreateandrestorepartialsearchablesnapshot dtests seed dtests locale no dtests timezone hongkong druntime java applicable branches reproduces locally didn t try failure history failure excerpt com carrotsearch randomizedtesting uncaughtexceptionerror captured an uncaught exception in thread thread state runnable group tgrp frozensearchablesnapshotsintegtests at randomizedtesting seedinfo seed caused by java lang assertionerror no message provided at randomizedtesting seedinfo seed at org elasticsearch cluster routing allocation indexmetadataupdater applychanges indexmetadataupdater java at org elasticsearch cluster routing allocation routingallocation updatemetadatawithroutingchanges routingallocation java at org elasticsearch cluster routing allocation allocationservice buildresultandloghealthchange allocationservice java at org elasticsearch cluster routing allocation allocationservice reroute allocationservice java at org elasticsearch cluster routing batchedrerouteservice execute batchedrerouteservice java at org elasticsearch cluster clusterstatetaskexecutor execute clusterstatetaskexecutor java at org elasticsearch cluster service masterservice innerexecutetasks masterservice java at org elasticsearch cluster service masterservice executetasks masterservice java at org elasticsearch cluster service masterservice runtasks masterservice java at org elasticsearch cluster service masterservice batcher run masterservice java at org elasticsearch cluster service taskbatcher runifnotprocessed taskbatcher java at org elasticsearch cluster service taskbatcher batchedtask run taskbatcher java at org elasticsearch common util concurrent threadcontext contextpreservingrunnable run threadcontext java at org elasticsearch common util concurrent prioritizedesthreadpoolexecutor tiebreakingprioritizedrunnable runandclean prioritizedesthreadpoolexecutor java at org elasticsearch common util concurrent prioritizedesthreadpoolexecutor tiebreakingprioritizedrunnable run prioritizedesthreadpoolexecutor java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java
1
228,208
25,169,552,938
IssuesEvent
2022-11-11 01:05:58
ghuangsnl/spring-boot
https://api.github.com/repos/ghuangsnl/spring-boot
opened
CVE-2022-37866 (High) detected in ivy-2.5.0.jar
security vulnerability
## CVE-2022-37866 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ivy-2.5.0.jar</b></p></summary> <p>The Apache Software Foundation provides support for the Apache community of open-source software projects. The Apache projects are characterized by a collaborative, consensus based development process, an open and pragmatic software license, and a desire to create high quality software that leads the way in its field. We consider ourselves not simply a group of projects sharing a server, but rather a community of developers and users.</p> <p>Library home page: <a href="http://ant.apache.org/ivy/">http://ant.apache.org/ivy/</a></p> <p>Path to vulnerable library: /spring-boot-project/spring-boot-tools/spring-boot-antlib/build.gradle</p> <p> Dependency Hierarchy: - :x: **ivy-2.5.0.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ghuangsnl/spring-boot/commit/275c27d9dd5c88d8db426ebfb734d89d3f8e7412">275c27d9dd5c88d8db426ebfb734d89d3f8e7412</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> When Apache Ivy downloads artifacts from a repository it stores them in the local file system based on a user-supplied "pattern" that may include placeholders for artifacts coordinates like the organisation, module or version. If said coordinates contain "../" sequences - which are valid characters for Ivy coordinates in general - it is possible the artifacts are stored outside of Ivy's local cache or repository or can overwrite different artifacts inside of the local cache. In order to exploit this vulnerability an attacker needs collaboration by the remote repository as Ivy will issue http requests containing ".." sequences and a "normal" repository will not interpret them as part of the artifact coordinates. Users of Apache Ivy 2.0.0 to 2.5.1 should upgrade to Ivy 2.5.1. <p>Publish Date: 2022-11-07 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-37866>CVE-2022-37866</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://lists.apache.org/thread/htxbr8oc464hxrgroftnz3my70whk93b">https://lists.apache.org/thread/htxbr8oc464hxrgroftnz3my70whk93b</a></p> <p>Release Date: 2022-11-07</p> <p>Fix Resolution: org.apache.ivy:ivy:2.5.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-37866 (High) detected in ivy-2.5.0.jar - ## CVE-2022-37866 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ivy-2.5.0.jar</b></p></summary> <p>The Apache Software Foundation provides support for the Apache community of open-source software projects. The Apache projects are characterized by a collaborative, consensus based development process, an open and pragmatic software license, and a desire to create high quality software that leads the way in its field. We consider ourselves not simply a group of projects sharing a server, but rather a community of developers and users.</p> <p>Library home page: <a href="http://ant.apache.org/ivy/">http://ant.apache.org/ivy/</a></p> <p>Path to vulnerable library: /spring-boot-project/spring-boot-tools/spring-boot-antlib/build.gradle</p> <p> Dependency Hierarchy: - :x: **ivy-2.5.0.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ghuangsnl/spring-boot/commit/275c27d9dd5c88d8db426ebfb734d89d3f8e7412">275c27d9dd5c88d8db426ebfb734d89d3f8e7412</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> When Apache Ivy downloads artifacts from a repository it stores them in the local file system based on a user-supplied "pattern" that may include placeholders for artifacts coordinates like the organisation, module or version. If said coordinates contain "../" sequences - which are valid characters for Ivy coordinates in general - it is possible the artifacts are stored outside of Ivy's local cache or repository or can overwrite different artifacts inside of the local cache. In order to exploit this vulnerability an attacker needs collaboration by the remote repository as Ivy will issue http requests containing ".." sequences and a "normal" repository will not interpret them as part of the artifact coordinates. Users of Apache Ivy 2.0.0 to 2.5.1 should upgrade to Ivy 2.5.1. <p>Publish Date: 2022-11-07 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-37866>CVE-2022-37866</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://lists.apache.org/thread/htxbr8oc464hxrgroftnz3my70whk93b">https://lists.apache.org/thread/htxbr8oc464hxrgroftnz3my70whk93b</a></p> <p>Release Date: 2022-11-07</p> <p>Fix Resolution: org.apache.ivy:ivy:2.5.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in ivy jar cve high severity vulnerability vulnerable library ivy jar the apache software foundation provides support for the apache community of open source software projects the apache projects are characterized by a collaborative consensus based development process an open and pragmatic software license and a desire to create high quality software that leads the way in its field we consider ourselves not simply a group of projects sharing a server but rather a community of developers and users library home page a href path to vulnerable library spring boot project spring boot tools spring boot antlib build gradle dependency hierarchy x ivy jar vulnerable library found in head commit a href vulnerability details when apache ivy downloads artifacts from a repository it stores them in the local file system based on a user supplied pattern that may include placeholders for artifacts coordinates like the organisation module or version if said coordinates contain sequences which are valid characters for ivy coordinates in general it is possible the artifacts are stored outside of ivy s local cache or repository or can overwrite different artifacts inside of the local cache in order to exploit this vulnerability an attacker needs collaboration by the remote repository as ivy will issue http requests containing sequences and a normal repository will not interpret them as part of the artifact coordinates users of apache ivy to should upgrade to ivy publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache ivy ivy step up your open source security game with mend
0
20,491
4,553,780,995
IssuesEvent
2016-09-13 06:55:16
Quick/Nimble
https://api.github.com/repos/Quick/Nimble
closed
Examples using `evaluate` in README need update
bug documentation needed
The following example, from [this page](https://github.com/Quick/Nimble#writing-your-own-matchers) on the README fails with a compile-time error, since `actionExpression.evaluate` can throw but is not caught. Full error: "Call can throw, but it is not marked with 'try' and the error is not handled" ```swift public func equals<T: Equatable>(expectedValue: T?) -> MatcherFunc<T?> { return MatcherFunc { actionExpression, failureMessage in failureMessage.postfixMessage = "equal <\(expectedValue)>" return actionExpression.evaluate() == expectedValue } } ```
1.0
Examples using `evaluate` in README need update - The following example, from [this page](https://github.com/Quick/Nimble#writing-your-own-matchers) on the README fails with a compile-time error, since `actionExpression.evaluate` can throw but is not caught. Full error: "Call can throw, but it is not marked with 'try' and the error is not handled" ```swift public func equals<T: Equatable>(expectedValue: T?) -> MatcherFunc<T?> { return MatcherFunc { actionExpression, failureMessage in failureMessage.postfixMessage = "equal <\(expectedValue)>" return actionExpression.evaluate() == expectedValue } } ```
non_test
examples using evaluate in readme need update the following example from on the readme fails with a compile time error since actionexpression evaluate can throw but is not caught full error call can throw but it is not marked with try and the error is not handled swift public func equals expectedvalue t matcherfunc return matcherfunc actionexpression failuremessage in failuremessage postfixmessage equal return actionexpression evaluate expectedvalue
0
181,849
21,664,454,322
IssuesEvent
2022-05-07 01:24:11
venkateshreddypala/post-it-a4
https://api.github.com/repos/venkateshreddypala/post-it-a4
closed
WS-2015-0024 (High) detected in uglify-js-2.3.6.tgz - autoclosed
security vulnerability
## WS-2015-0024 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>uglify-js-2.3.6.tgz</b></p></summary> <p>JavaScript parser, mangler/compressor and beautifier toolkit</p> <p>Library home page: <a href="https://registry.npmjs.org/uglify-js/-/uglify-js-2.3.6.tgz">https://registry.npmjs.org/uglify-js/-/uglify-js-2.3.6.tgz</a></p> <p>Path to dependency file: /post-it-a4/package.json</p> <p>Path to vulnerable library: /tmp/git/post-it-a4/node_modules/handlebars/node_modules/uglify-js/package.json</p> <p> Dependency Hierarchy: - cli-1.1.1.tgz (Root Library) - postcss-url-5.1.2.tgz - directory-encoder-0.7.2.tgz - handlebars-1.3.0.tgz - :x: **uglify-js-2.3.6.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://api.github.com/repos/venkateshreddypala/post-it-a4/commits/c71d001cc56ebfa721446ecdffd026c4e7337310">c71d001cc56ebfa721446ecdffd026c4e7337310</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> UglifyJS versions 2.4.23 and earlier are affected by a vulnerability which allows a specially crafted Javascript file to have altered functionality after minification. <p>Publish Date: 2015-08-24 <p>URL: <a href=https://github.com/mishoo/UglifyJS2/issues/751>WS-2015-0024</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>8.3</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/mishoo/UglifyJS2/commit/905b6011784ca60d41919ac1a499962b7c1d4b02">https://github.com/mishoo/UglifyJS2/commit/905b6011784ca60d41919ac1a499962b7c1d4b02</a></p> <p>Release Date: 2017-01-31</p> <p>Fix Resolution: v2.4.24</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2015-0024 (High) detected in uglify-js-2.3.6.tgz - autoclosed - ## WS-2015-0024 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>uglify-js-2.3.6.tgz</b></p></summary> <p>JavaScript parser, mangler/compressor and beautifier toolkit</p> <p>Library home page: <a href="https://registry.npmjs.org/uglify-js/-/uglify-js-2.3.6.tgz">https://registry.npmjs.org/uglify-js/-/uglify-js-2.3.6.tgz</a></p> <p>Path to dependency file: /post-it-a4/package.json</p> <p>Path to vulnerable library: /tmp/git/post-it-a4/node_modules/handlebars/node_modules/uglify-js/package.json</p> <p> Dependency Hierarchy: - cli-1.1.1.tgz (Root Library) - postcss-url-5.1.2.tgz - directory-encoder-0.7.2.tgz - handlebars-1.3.0.tgz - :x: **uglify-js-2.3.6.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://api.github.com/repos/venkateshreddypala/post-it-a4/commits/c71d001cc56ebfa721446ecdffd026c4e7337310">c71d001cc56ebfa721446ecdffd026c4e7337310</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> UglifyJS versions 2.4.23 and earlier are affected by a vulnerability which allows a specially crafted Javascript file to have altered functionality after minification. <p>Publish Date: 2015-08-24 <p>URL: <a href=https://github.com/mishoo/UglifyJS2/issues/751>WS-2015-0024</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>8.3</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/mishoo/UglifyJS2/commit/905b6011784ca60d41919ac1a499962b7c1d4b02">https://github.com/mishoo/UglifyJS2/commit/905b6011784ca60d41919ac1a499962b7c1d4b02</a></p> <p>Release Date: 2017-01-31</p> <p>Fix Resolution: v2.4.24</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
ws high detected in uglify js tgz autoclosed ws high severity vulnerability vulnerable library uglify js tgz javascript parser mangler compressor and beautifier toolkit library home page a href path to dependency file post it package json path to vulnerable library tmp git post it node modules handlebars node modules uglify js package json dependency hierarchy cli tgz root library postcss url tgz directory encoder tgz handlebars tgz x uglify js tgz vulnerable library found in head commit a href vulnerability details uglifyjs versions and earlier are affected by a vulnerability which allows a specially crafted javascript file to have altered functionality after minification publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
66,235
20,086,240,408
IssuesEvent
2022-02-05 02:14:31
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
opened
Clicking on the jump to unreads button gives a "Failed to load timeline position" and dumps me in a blank timeline
T-Defect A-Timeline A-Read-Marker
### Steps to reproduce 1. Visit a room 1. Click the "Jump to first unread message." arrow ![](https://user-images.githubusercontent.com/558581/152602295-4dee9b11-2e0c-4782-bb1c-10e42a004c08.png) 1. `Failed to load timeline position` error modal is thrown up > Failed to load timeline position > > Tried to load a specific point in this room's timeline, but was unable to find it. ![](https://user-images.githubusercontent.com/558581/152602305-7068332b-156b-46e7-a782-2395171786f0.png) 1. After exiting out of the modal, the timeline is blank and no way to scroll 1. Have to switch rooms to get the timeline again --- Reproduction/rage-shake logs: https://github.com/matrix-org/element-web-rageshakes/issues/10392 I was in the `!ltpmbfSHehPTwnLDNS:jki.re` room and clicked on the "Jump to first unread message." arrow and got the `Failed to load timeline position` error modal. With the following error in the devtools console. Relevant code for the error: [`src/components/structures/TimelinePanel.tsx#L1219-L1244`](https://github.com/matrix-org/matrix-react-sdk/blob/7f3f18604417d9567c109201673e60a5f24a30ce/src/components/structures/TimelinePanel.tsx#L1219-L1244) ``` Error loading timeline panel at $iR-8DJRGNhmIpDFDHmRy70gUTurU77eCba-Jt5YIXbo: Error: getEventTimeline result didn't include requested event ``` I'm not sure what room the `$iR-8DJRGNhmIpDFDHmRy70gUTurU77eCba-Jt5YIXbo` event pertains to. Is there a way to look it up and tell which room it actually belongs to? ### Outcome #### What did you expect? Jump to first unread button jumps me to whatever the unreads are for that room. Or not be shown if it was for another room and this room doesn't have any unread. #### What happened instead? Error modal is shown for a situation that probably shouldn't occur in the first place. ### Operating system Windows 10 ### Browser information Chrome 97.0.4692.99 ### URL for webapp https://develop.element.io/ ### Application version Element version: 64242a004eb7-react-78e78292cb62-js-b07457726bf5 Olm version: 3.2.8 ### Homeserver matrix.org ### Will you send logs? Yes
1.0
Clicking on the jump to unreads button gives a "Failed to load timeline position" and dumps me in a blank timeline - ### Steps to reproduce 1. Visit a room 1. Click the "Jump to first unread message." arrow ![](https://user-images.githubusercontent.com/558581/152602295-4dee9b11-2e0c-4782-bb1c-10e42a004c08.png) 1. `Failed to load timeline position` error modal is thrown up > Failed to load timeline position > > Tried to load a specific point in this room's timeline, but was unable to find it. ![](https://user-images.githubusercontent.com/558581/152602305-7068332b-156b-46e7-a782-2395171786f0.png) 1. After exiting out of the modal, the timeline is blank and no way to scroll 1. Have to switch rooms to get the timeline again --- Reproduction/rage-shake logs: https://github.com/matrix-org/element-web-rageshakes/issues/10392 I was in the `!ltpmbfSHehPTwnLDNS:jki.re` room and clicked on the "Jump to first unread message." arrow and got the `Failed to load timeline position` error modal. With the following error in the devtools console. Relevant code for the error: [`src/components/structures/TimelinePanel.tsx#L1219-L1244`](https://github.com/matrix-org/matrix-react-sdk/blob/7f3f18604417d9567c109201673e60a5f24a30ce/src/components/structures/TimelinePanel.tsx#L1219-L1244) ``` Error loading timeline panel at $iR-8DJRGNhmIpDFDHmRy70gUTurU77eCba-Jt5YIXbo: Error: getEventTimeline result didn't include requested event ``` I'm not sure what room the `$iR-8DJRGNhmIpDFDHmRy70gUTurU77eCba-Jt5YIXbo` event pertains to. Is there a way to look it up and tell which room it actually belongs to? ### Outcome #### What did you expect? Jump to first unread button jumps me to whatever the unreads are for that room. Or not be shown if it was for another room and this room doesn't have any unread. #### What happened instead? Error modal is shown for a situation that probably shouldn't occur in the first place. ### Operating system Windows 10 ### Browser information Chrome 97.0.4692.99 ### URL for webapp https://develop.element.io/ ### Application version Element version: 64242a004eb7-react-78e78292cb62-js-b07457726bf5 Olm version: 3.2.8 ### Homeserver matrix.org ### Will you send logs? Yes
non_test
clicking on the jump to unreads button gives a failed to load timeline position and dumps me in a blank timeline steps to reproduce visit a room click the jump to first unread message arrow failed to load timeline position error modal is thrown up failed to load timeline position tried to load a specific point in this room s timeline but was unable to find it after exiting out of the modal the timeline is blank and no way to scroll have to switch rooms to get the timeline again reproduction rage shake logs i was in the ltpmbfshehptwnldns jki re room and clicked on the jump to first unread message arrow and got the failed to load timeline position error modal with the following error in the devtools console relevant code for the error error loading timeline panel at ir error geteventtimeline result didn t include requested event i m not sure what room the ir event pertains to is there a way to look it up and tell which room it actually belongs to outcome what did you expect jump to first unread button jumps me to whatever the unreads are for that room or not be shown if it was for another room and this room doesn t have any unread what happened instead error modal is shown for a situation that probably shouldn t occur in the first place operating system windows browser information chrome url for webapp application version element version react js olm version homeserver matrix org will you send logs yes
0
32,649
2,757,202,628
IssuesEvent
2015-04-27 13:42:49
geometalab/osmaxx
https://api.github.com/repos/geometalab/osmaxx
closed
Use native UUID-Fieldtype with django 1.8
priority:high type:refactoring
Starting with Django 1.8 there will be a native UUIDField in [Django](https://docs.djangoproject.com/en/1.8/ref/models/fields/#uuidfield) Depends on #80
1.0
Use native UUID-Fieldtype with django 1.8 - Starting with Django 1.8 there will be a native UUIDField in [Django](https://docs.djangoproject.com/en/1.8/ref/models/fields/#uuidfield) Depends on #80
non_test
use native uuid fieldtype with django starting with django there will be a native uuidfield in depends on
0
378,100
11,196,166,400
IssuesEvent
2020-01-03 09:18:49
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
closed
[0.9.0 staging-1293] Civics: Laws for craft
Fixed Medium Priority
Need to cjeck crafting trigger. 1. I have a law like this ![bandicam 2019-12-13 15-53-10-239](https://user-images.githubusercontent.com/27898520/70795907-c7225b80-1dc2-11ea-806b-cfae8620fa13.jpg) I suppose it should prevent charred food and charred papaya orders. But I still can craft them. ![bandicam 2019-12-13 15-54-22-794](https://user-images.githubusercontent.com/27898520/70795979-ef11bf00-1dc2-11ea-8797-8c5580bdfa74.jpg) ![bandicam 2019-12-13 15-57-10-323](https://user-images.githubusercontent.com/27898520/70795993-f46f0980-1dc2-11ea-8a8c-40407d1ac3a8.jpg) 2. Why do we have two fields: matching item and Creating Product? Also Creating Product won't show me all the products (eg. no Charred Animal).
1.0
[0.9.0 staging-1293] Civics: Laws for craft - Need to cjeck crafting trigger. 1. I have a law like this ![bandicam 2019-12-13 15-53-10-239](https://user-images.githubusercontent.com/27898520/70795907-c7225b80-1dc2-11ea-806b-cfae8620fa13.jpg) I suppose it should prevent charred food and charred papaya orders. But I still can craft them. ![bandicam 2019-12-13 15-54-22-794](https://user-images.githubusercontent.com/27898520/70795979-ef11bf00-1dc2-11ea-8797-8c5580bdfa74.jpg) ![bandicam 2019-12-13 15-57-10-323](https://user-images.githubusercontent.com/27898520/70795993-f46f0980-1dc2-11ea-8a8c-40407d1ac3a8.jpg) 2. Why do we have two fields: matching item and Creating Product? Also Creating Product won't show me all the products (eg. no Charred Animal).
non_test
civics laws for craft need to cjeck crafting trigger i have a law like this i suppose it should prevent charred food and charred papaya orders but i still can craft them why do we have two fields matching item and creating product also creating product won t show me all the products eg no charred animal
0
48,391
5,956,308,349
IssuesEvent
2017-05-28 15:36:37
mapbox/carto
https://api.github.com/repos/mapbox/carto
closed
Nested conditions break using same key
needs-test
This breaks quite messily: ``` #ncn_label[zoom>=11][zoom<=14][reflen<=8][type='bicycle'] { [network='ncn'],[network='rcn'] { shield-name: "[ref]"; shield-face-name: @sans_bold; shield-file: url(img/shield-motorway-1.png); [network='ncn'] { [reflen=8] { shield-file: url(img/shield-motorway-8.png); } } [network='rcn'] { [reflen=8] { shield-file: url(img/shield-trunk-8.png); } } } } ``` with repeated errors of "labels.mss:377:16 Property shield-face-name required for defining shield styles." If the nesting is unrolled slightly, then it's happy: ``` #ncn_label[zoom>=11][zoom<=14][reflen<=8][type='bicycle'] { [network='ncn'],[network='rcn'] { shield-name: "[ref]"; shield-face-name: @sans_bold; shield-file: url(img/shield-motorway-1.png); } [network='ncn'] { [reflen=8] { shield-file: url(img/shield-motorway-8.png); } } [network='rcn'] { [reflen=8] { shield-file: url(img/shield-trunk-8.png); } } } ``` (Code truncated for clarity but it should still illustrate the point, I hope!)
1.0
Nested conditions break using same key - This breaks quite messily: ``` #ncn_label[zoom>=11][zoom<=14][reflen<=8][type='bicycle'] { [network='ncn'],[network='rcn'] { shield-name: "[ref]"; shield-face-name: @sans_bold; shield-file: url(img/shield-motorway-1.png); [network='ncn'] { [reflen=8] { shield-file: url(img/shield-motorway-8.png); } } [network='rcn'] { [reflen=8] { shield-file: url(img/shield-trunk-8.png); } } } } ``` with repeated errors of "labels.mss:377:16 Property shield-face-name required for defining shield styles." If the nesting is unrolled slightly, then it's happy: ``` #ncn_label[zoom>=11][zoom<=14][reflen<=8][type='bicycle'] { [network='ncn'],[network='rcn'] { shield-name: "[ref]"; shield-face-name: @sans_bold; shield-file: url(img/shield-motorway-1.png); } [network='ncn'] { [reflen=8] { shield-file: url(img/shield-motorway-8.png); } } [network='rcn'] { [reflen=8] { shield-file: url(img/shield-trunk-8.png); } } } ``` (Code truncated for clarity but it should still illustrate the point, I hope!)
test
nested conditions break using same key this breaks quite messily ncn label shield name shield face name sans bold shield file url img shield motorway png shield file url img shield motorway png shield file url img shield trunk png with repeated errors of labels mss property shield face name required for defining shield styles if the nesting is unrolled slightly then it s happy ncn label shield name shield face name sans bold shield file url img shield motorway png shield file url img shield motorway png shield file url img shield trunk png code truncated for clarity but it should still illustrate the point i hope
1
244,151
20,612,572,416
IssuesEvent
2022-03-07 10:05:11
keymanapp/keyman
https://api.github.com/repos/keymanapp/keyman
opened
test(android): Regression Test B15S4
android/ tests
# Keyman for Android Acceptance Test Procedures # 1. These test procedures are to be run before moving from alpha to beta, or beta to stable, or before PRs are merged into stable branches. 2. Copy these checklists of tests into a new issue (for tier transitions) or a comment on the PR. 3. When annotating results, use `>` at the start of a new line under the checkbox to note the result. (For an example, see #3973). ## Gather Assets for Testing ## - [ ] Physical Android phone (record device's Android -- version 5.0 Lollipop or later) - [ ] External Bluetooth or USB QWERTY keyboard - [ ] External Bluetooth or USB AZERTY keyboard (European AZERTY layout with 102nd key) ## SUITE_SETUP: Setup Steps ## - **TEST_UNINSTALL**: Uninstall previous version of Keyman on the device (don't forget to reset user settings for the app; in Android Pie: Settings>Storage>Other apps>Keyman>Clear Data) - **TEST_UNINSTALL**: Install test build APK - **TEST_ACCEPT_PERMISSION**: Accept all the Android permission requests for storage (applies to Android 6.0 Marshmallow onward) # User Testing ## **SUITE_INSPECT: Visual Inspection of master branch** - **TEST_CI_WEB_MASTER**: Verify latest CI Android build of master is successful on https://status.keyman.com/ - **TEST_VERIFY_HISTORY**: Verify `/history.md` contains all the current changes ## Various versions of Androids UI for these tests assume default English locale * GROUP_ANDROID_5 * GROUP_ANDROID_8 * GROUP_ANDROID_11 ## SUITE_MENU_FUNCTIONALITY: Menu Functionality Tests ## <details> <summary>click to expand</summary> - **TEST_GET_STARTED:** This tests the "Get Started" menu from a fresh installation <details> <summary>Expand to see how to do and verify this</summary> 1. When the app loads, verify the "Get Started" menu is visible 2. From the "Get Started" menu, touch "Add a keyboard for your language" 3. Verify "Install Keyboard or Dictionary" menu appears 4. From the "Install Keyboard or Dictionary" menu, click the back arrow to return to "Get Started" 5. From the "Get Started" menu, touch "Enable Keyman as system-wide keyboard" 6. Verify Android system menus appear for controlling virtual keyboard 7. Scroll down and enable Keyman as a system keyboard. 8. On the "Attention" dialog, read and hit "OK" to dismiss the dialogs. 9. Touch the back button to return to "Get Started" 10. Verify "Enable Keyman as system-wide keyboard" now has a ticked checkbox. 11. From the "Get Started" menu, touch "Set Keyman as default keyboard" 12. On the "Change keyboard" dialog, select Keyman 13. Verify on the "Get Started" menu that "Set Keyman as default keyboard" now has a ticked checkbox. 14. From the "Get Started" menu, touch "More info" 15. Verify the Info page appears with a version string at the bottom of the page 16. Touch the back button (arrow) to return to "Get Started" 17. From the "Get Started" menu, untick the last option 'Show "Get Started" on startup' 18. Close the "Get Started" menu and exit the app. 19. From the Android device, start Keyman for Android 20. Verify the "Get Started" menu does not appear. </details> - **TEST_INSTALL_KD:** This menu is accessed via **Get Started** menu or the Settings button-->"Install Keyboard or Dictionary" <details> <summary>Expand to see how to do and verify this</summary> 1. On the device, enable internet connectivity 2. Touch "Install from keyman.com" 3. Verify a new keyboard can be downloaded and selected 4. On the device, disable internet connectivity 5. From "Install Keyboard or Dictionary", touch "Install from keyman.com" 6. Verify device displays a message "Cannot connect to Keyman server!" 7. On the device, re-enable internet connectivity </details> - **TEST_KEYMAN_SETTTINGS:** <details> <summary>Expand to see how to do and verify this</summary> 1. From "Settings" --> "Installed Languages" 2. Select "English" --> English Settings 3. Select keyboard settings menu 4. Verify Info appears with keyboard version, help link, and QR code 5. If keyboard update is available, click on version and verify update 6. If help link is available, click on link 7. Verify keyboard help documentation appears 8. Return to English Settings menu 9. Select Dictionary --> English dictionary (MTNT) 10. Verify the Dictionary info appears 11. Return to English Settings menu 12. Turn on/off 'Enable Predictions' and validate each 13. Turn on/off 'Enable Corrections' and validate each. For example, when Corrections are enabled and context "tooo" should still suggest "too" 14. Validate that when both predictions and corrections are off, banner is not visible 15. Add/Remove dictionaries - validate 16. If multiple dictionaries are available, test swapping between them </details> - **TEST_SHARE:** <details> <summary>Expand to see how to do and verify this</summary> - Type and verify text can be shared to external app > Failed - no external apps shown on Android 11 #6231 </details> - **TEST_KEYMAN_BROWSER:** Keyman Browser ### <details> <summary>Expand to see how to do and verify this</summary> 1. On default page, click on text field and set system keyboard to Keyman 2. Verify can type with Keyman as system keyboard 3. Close and reopen Keyman app </details> - **TEST_TEXT_SIZE:** <details> <summary>Expand to see how to do and verify this</summary> - Verify text can be rescaled from Text Size 16 to 72 </details> - **TEST_CLEAR_TEXT:** <details> <summary>Expand to see how to do and verify this</summary> - Verify text can be cleared </details> - **TEST_INFO:** <details> <summary>Expand to see how to do and verify this</summary> 1. From the device, disable internet access 2. From the Keyman app menu, select "Info" 3. Verify Keyman for Android version appears at the bottom of the screen 4. Verify Keyman for Android help documentation is displayed. 5. From the device, enable internet access </details> - **TEST_INSTALL_UPDATES:** This menu option only appears when a language resource (keyboard or lexical-model) update is available <details> <summary>Expand to see how to do and verify this</summary> 1. Click on "Install Updates" 2. Accept "Download" on the dialogue 3. Verify resource is updated after background download complete > Not tested because updates not available </details> - **TEST_CHANGE_DISPLAY_LANG:** This tests changing the display language for the App and assumes starting with English locale <details> <summary>Expand to see how to do and verify this</summary> 1. From "Settings" --> "Display Language" 2. Select "Khmer" 3. From "Settings" --> Verify menu selections in Khmer 4. From the overflow menu --> clear text 5. Verify Keyman text field has Khmer prompt (for "Start typing here") 6. From "Settings" --> "Change Display Language" 7. If the string is translated, the first row should be "Default Locale" in Khmer 8. Select "English" 9. Verify the app reloads in English </details> - **TEST_ADJUST_K_HEIGHT:** This menu option allows the user to adjust the OSK height for the current orientation (portrait or landscape). The customized height is saved as a preference <details> <summary>Expand to see how to do and verify this</summary> 1. Have the device in portrait orientation 2. From "Settings" --> "Adjust keyboard height" 3. On the menu, drag the OSK height to change the keyboard height 4. Click back 5. Verify the OSK is refreshed to the selected height 6. From "Settings" --> "Adjust keyboard height" 7. Click the "reset to defaults" button and then click back 8. Verify the OSK reverts to the original height 9. Rotate the device to landscape orientation 10. From "Settings" --> "Adjust keyboard height" 11. On the menu, drag the OSK height to change the keyboard height 12. Click back 13. Verify the OSK is refreshed to the selected height 14. From "Settings" --> "Adjust keyboard height" 15. Click the "reset to defaults" button and then click back 16. Verify the OSK reverts to the original height 17. Rotate the device back to portrait orientation </details> - **TEST_SPACEBAR_CAPTION:** This menu option allows the user to adjust the label displayed on the keyboard spacebar <details> <summary>Expand to see how to do and verify this</summary> 1. Select the default SIL EuroLatin keyboard 2. From "Settings" --> Spacebar caption 3. On the menu, select "Language" 4. Click back until the OSK is displayed 5. Verify the spacebar label is "English" 6. From "Settings" --> Spacebar caption 7. On the menu, select "Keyboard" 8. Click back until the OSK is displayed 9. Verify the spacebar label is "EuroLatin (SIL)" 10. From "Settings" --> Spacebar caption 11. On the menu, select "Blank" 12. Click back until the OSK is displayed 13. Verify the spacebar label is blank 14. From "Settings" --> Spacebar caption 15. On the menu, select "Language+Keyboard" 16. Click back until the OSK is displayed 17. Verify the spacebar label is "English - EuroLatin (SIL)" </details> </details> ---- ## SUITE_KEYBOARD_FUNCTIONALITY: Keyboard Tests <details> <summary>click to expand</summary> ### In-app Keyboards These tests are run in the context of typing in the Keyman app - **TEST_INAPP_LATIN_KEYBOARD:** English(EuroLatin SIL) In portrait orientation, verify OSK is visible and fills the width the bottom of the screen <details> <summary>Expand to see how to do and verify this</summary> 1. Verify long-press `q` key works 2. Verify long-press `p` key works 3. Verify uppercase layer can be selected via `SHIFT` 4. Verify number layer can be selected via `123` 5. Verify long-press `1` key works 6. Verify long-press `0` key works 7. Verify backspace, space, and enter keys work 8. Verify *Khmer* -> *Khmer Angkor* keyboard can be added via Settings menu 9. Repeat in landscape orientation > Landscape orientation failed - The popup key on certain EuroLatin keys (such as "q", "y", "p") are truncated. Workaround is adjusting the keyboard height taller #6232 </details> - **TEST_INAPP_NON_LATIN_KEYBOARD:** Non-Latin script Keyboard <details> <summary>Expand to see how to do and verify this</summary> 1. In the app, add/select a random non-Latin script keyboard 2. Verify OSK is visible 3. Test several keys and modifiers and verify Latin characters are not appearing. </details> ### System Keyboards These tests are run with Keyman selected as a system keyboard, and in the context of typing in an external app (like Google Keep) - **TEST_SYSTEM_LATIN_KEYBOARD:** English (EuroLatin SIL) In portrait orientation, verify OSK is visible and fills the width the bottom of the screen <details> <summary>Expand to see how to do and verify this</summary> 1. Verify long-press `q` key works 2. Verify long-press `p` key works 3. Verify uppercase layer can be selected via `SHIFT` 4. Verify number layer can be selected via `123` 5. Verify long-press `1` key works 6. Verify long-press `0` key works 7. Verify backspace, space, and enter keys work 8. Verify can change to random non-Latin script keyboard via globe button 9. Repeat in landscape orientation </details> - **TEST_KEYBOARD_PICKER:** Keyboard Picker menu to switch input method This tests the Keyboard Picker menu which lists other system input methods at the bottom of the menu. These appear only in Keyman system keyboard <details> <summary>Expand to see how to do and verify this</summary> 1. With only 1 Keyman keyboard installed and selected as a system keyboard, long-press and release the the globe button 2. Verify the bottom of the Keyboards picker menu lists other available input methods 3. Click on one of the other input methods 4. Verify the Keyman system keyboard switches to the selected input method 5. On the device's Settings, select Keyman as the input method 6. On the Keyman keyboard, short-press and release the globe button 7. Verify the previous input method is selected </details> - **TEST_EXT_BLUETOOTH_KEYBOARD:** External (Bluetooth) keyboard This test is run with an external keyboard (USB or bluetooth) connected to the device <details> <summary>Expand to see how to do and verify this</summary> 1. Start with English(EuroLatin SIL) as the selected keyboard 2. On the device's Settings, show the input method (OSK) 3. Verify can type with numbers, letters, and punctuation 4. Verify `caps lock` has no effect when typing 5. Verify long-press on a character repeats that character 6. Verify `tab` toggles to next field 7. Verify `enter` adds a newline 8. Verify `backspace` removes a character 9. Verify <kbd>ctrl</kbd>+<kbd>tab</kbd> brings up the keyboard picker menu 10. Use <kbd>tab</kbd> or arrow keys to navigate the keyboard picker menu > Failed - navigating to other input methods selected the entire group #6233 11. Select a non-Latin keyboard 12. Resume typing and verify new script appears </details> - **TEST_EXT_AZERTY_KEYBOARD:** External European AZERTY keyboard This test is run with an external European AZERTY keyboard paired to the device. It should contain the 102nd key `<>` on the bottom row next to the left-shift. <details> <summary>Expand to see how to do and verify this</summary> 1. Start Keyman and install the French Basic keyboard 2. From the physical keyboard, type on the top letter row (azerty) 3. Verify `azerty` appears 4. Type on the 102nd key 5. Verify `<` appears 6. Type shift-102nd key 7. Verify `>` appears </details> </details> ---- ## SUITE_KB_DISTRIBUTION: Ad-hoc Keyboard Distribution Tests <details> <summary>Expand to see how to do and verify this</summary> - **TEST_KMP_DISTRIBUTION** KMP distribution <details> <summary>Expand to see how to do and verify this</summary> 1. Reset app settings (in Android Pie: Settings>Storage>Other apps>Keyman>Clear Storage) 2. Verify kmp [distribution](https://help.keyman.com/developer/current-version/guides/distribute/) successfully installs<br/> Example: [Khmer Angkor](https://downloads.keyman.com/keyboards/khmer_angkor/1.0.2/khmer_angkor.kmp) 3. Verify kmp successfully installs from Settings --> Install Keyboard or Dictionary --> Install from local file </details> </details> ---- ## What's New Tests - [ ] Refer to the new changes in `/HISTORY.md` and verify functionality
1.0
test(android): Regression Test B15S4 - # Keyman for Android Acceptance Test Procedures # 1. These test procedures are to be run before moving from alpha to beta, or beta to stable, or before PRs are merged into stable branches. 2. Copy these checklists of tests into a new issue (for tier transitions) or a comment on the PR. 3. When annotating results, use `>` at the start of a new line under the checkbox to note the result. (For an example, see #3973). ## Gather Assets for Testing ## - [ ] Physical Android phone (record device's Android -- version 5.0 Lollipop or later) - [ ] External Bluetooth or USB QWERTY keyboard - [ ] External Bluetooth or USB AZERTY keyboard (European AZERTY layout with 102nd key) ## SUITE_SETUP: Setup Steps ## - **TEST_UNINSTALL**: Uninstall previous version of Keyman on the device (don't forget to reset user settings for the app; in Android Pie: Settings>Storage>Other apps>Keyman>Clear Data) - **TEST_UNINSTALL**: Install test build APK - **TEST_ACCEPT_PERMISSION**: Accept all the Android permission requests for storage (applies to Android 6.0 Marshmallow onward) # User Testing ## **SUITE_INSPECT: Visual Inspection of master branch** - **TEST_CI_WEB_MASTER**: Verify latest CI Android build of master is successful on https://status.keyman.com/ - **TEST_VERIFY_HISTORY**: Verify `/history.md` contains all the current changes ## Various versions of Androids UI for these tests assume default English locale * GROUP_ANDROID_5 * GROUP_ANDROID_8 * GROUP_ANDROID_11 ## SUITE_MENU_FUNCTIONALITY: Menu Functionality Tests ## <details> <summary>click to expand</summary> - **TEST_GET_STARTED:** This tests the "Get Started" menu from a fresh installation <details> <summary>Expand to see how to do and verify this</summary> 1. When the app loads, verify the "Get Started" menu is visible 2. From the "Get Started" menu, touch "Add a keyboard for your language" 3. Verify "Install Keyboard or Dictionary" menu appears 4. From the "Install Keyboard or Dictionary" menu, click the back arrow to return to "Get Started" 5. From the "Get Started" menu, touch "Enable Keyman as system-wide keyboard" 6. Verify Android system menus appear for controlling virtual keyboard 7. Scroll down and enable Keyman as a system keyboard. 8. On the "Attention" dialog, read and hit "OK" to dismiss the dialogs. 9. Touch the back button to return to "Get Started" 10. Verify "Enable Keyman as system-wide keyboard" now has a ticked checkbox. 11. From the "Get Started" menu, touch "Set Keyman as default keyboard" 12. On the "Change keyboard" dialog, select Keyman 13. Verify on the "Get Started" menu that "Set Keyman as default keyboard" now has a ticked checkbox. 14. From the "Get Started" menu, touch "More info" 15. Verify the Info page appears with a version string at the bottom of the page 16. Touch the back button (arrow) to return to "Get Started" 17. From the "Get Started" menu, untick the last option 'Show "Get Started" on startup' 18. Close the "Get Started" menu and exit the app. 19. From the Android device, start Keyman for Android 20. Verify the "Get Started" menu does not appear. </details> - **TEST_INSTALL_KD:** This menu is accessed via **Get Started** menu or the Settings button-->"Install Keyboard or Dictionary" <details> <summary>Expand to see how to do and verify this</summary> 1. On the device, enable internet connectivity 2. Touch "Install from keyman.com" 3. Verify a new keyboard can be downloaded and selected 4. On the device, disable internet connectivity 5. From "Install Keyboard or Dictionary", touch "Install from keyman.com" 6. Verify device displays a message "Cannot connect to Keyman server!" 7. On the device, re-enable internet connectivity </details> - **TEST_KEYMAN_SETTTINGS:** <details> <summary>Expand to see how to do and verify this</summary> 1. From "Settings" --> "Installed Languages" 2. Select "English" --> English Settings 3. Select keyboard settings menu 4. Verify Info appears with keyboard version, help link, and QR code 5. If keyboard update is available, click on version and verify update 6. If help link is available, click on link 7. Verify keyboard help documentation appears 8. Return to English Settings menu 9. Select Dictionary --> English dictionary (MTNT) 10. Verify the Dictionary info appears 11. Return to English Settings menu 12. Turn on/off 'Enable Predictions' and validate each 13. Turn on/off 'Enable Corrections' and validate each. For example, when Corrections are enabled and context "tooo" should still suggest "too" 14. Validate that when both predictions and corrections are off, banner is not visible 15. Add/Remove dictionaries - validate 16. If multiple dictionaries are available, test swapping between them </details> - **TEST_SHARE:** <details> <summary>Expand to see how to do and verify this</summary> - Type and verify text can be shared to external app > Failed - no external apps shown on Android 11 #6231 </details> - **TEST_KEYMAN_BROWSER:** Keyman Browser ### <details> <summary>Expand to see how to do and verify this</summary> 1. On default page, click on text field and set system keyboard to Keyman 2. Verify can type with Keyman as system keyboard 3. Close and reopen Keyman app </details> - **TEST_TEXT_SIZE:** <details> <summary>Expand to see how to do and verify this</summary> - Verify text can be rescaled from Text Size 16 to 72 </details> - **TEST_CLEAR_TEXT:** <details> <summary>Expand to see how to do and verify this</summary> - Verify text can be cleared </details> - **TEST_INFO:** <details> <summary>Expand to see how to do and verify this</summary> 1. From the device, disable internet access 2. From the Keyman app menu, select "Info" 3. Verify Keyman for Android version appears at the bottom of the screen 4. Verify Keyman for Android help documentation is displayed. 5. From the device, enable internet access </details> - **TEST_INSTALL_UPDATES:** This menu option only appears when a language resource (keyboard or lexical-model) update is available <details> <summary>Expand to see how to do and verify this</summary> 1. Click on "Install Updates" 2. Accept "Download" on the dialogue 3. Verify resource is updated after background download complete > Not tested because updates not available </details> - **TEST_CHANGE_DISPLAY_LANG:** This tests changing the display language for the App and assumes starting with English locale <details> <summary>Expand to see how to do and verify this</summary> 1. From "Settings" --> "Display Language" 2. Select "Khmer" 3. From "Settings" --> Verify menu selections in Khmer 4. From the overflow menu --> clear text 5. Verify Keyman text field has Khmer prompt (for "Start typing here") 6. From "Settings" --> "Change Display Language" 7. If the string is translated, the first row should be "Default Locale" in Khmer 8. Select "English" 9. Verify the app reloads in English </details> - **TEST_ADJUST_K_HEIGHT:** This menu option allows the user to adjust the OSK height for the current orientation (portrait or landscape). The customized height is saved as a preference <details> <summary>Expand to see how to do and verify this</summary> 1. Have the device in portrait orientation 2. From "Settings" --> "Adjust keyboard height" 3. On the menu, drag the OSK height to change the keyboard height 4. Click back 5. Verify the OSK is refreshed to the selected height 6. From "Settings" --> "Adjust keyboard height" 7. Click the "reset to defaults" button and then click back 8. Verify the OSK reverts to the original height 9. Rotate the device to landscape orientation 10. From "Settings" --> "Adjust keyboard height" 11. On the menu, drag the OSK height to change the keyboard height 12. Click back 13. Verify the OSK is refreshed to the selected height 14. From "Settings" --> "Adjust keyboard height" 15. Click the "reset to defaults" button and then click back 16. Verify the OSK reverts to the original height 17. Rotate the device back to portrait orientation </details> - **TEST_SPACEBAR_CAPTION:** This menu option allows the user to adjust the label displayed on the keyboard spacebar <details> <summary>Expand to see how to do and verify this</summary> 1. Select the default SIL EuroLatin keyboard 2. From "Settings" --> Spacebar caption 3. On the menu, select "Language" 4. Click back until the OSK is displayed 5. Verify the spacebar label is "English" 6. From "Settings" --> Spacebar caption 7. On the menu, select "Keyboard" 8. Click back until the OSK is displayed 9. Verify the spacebar label is "EuroLatin (SIL)" 10. From "Settings" --> Spacebar caption 11. On the menu, select "Blank" 12. Click back until the OSK is displayed 13. Verify the spacebar label is blank 14. From "Settings" --> Spacebar caption 15. On the menu, select "Language+Keyboard" 16. Click back until the OSK is displayed 17. Verify the spacebar label is "English - EuroLatin (SIL)" </details> </details> ---- ## SUITE_KEYBOARD_FUNCTIONALITY: Keyboard Tests <details> <summary>click to expand</summary> ### In-app Keyboards These tests are run in the context of typing in the Keyman app - **TEST_INAPP_LATIN_KEYBOARD:** English(EuroLatin SIL) In portrait orientation, verify OSK is visible and fills the width the bottom of the screen <details> <summary>Expand to see how to do and verify this</summary> 1. Verify long-press `q` key works 2. Verify long-press `p` key works 3. Verify uppercase layer can be selected via `SHIFT` 4. Verify number layer can be selected via `123` 5. Verify long-press `1` key works 6. Verify long-press `0` key works 7. Verify backspace, space, and enter keys work 8. Verify *Khmer* -> *Khmer Angkor* keyboard can be added via Settings menu 9. Repeat in landscape orientation > Landscape orientation failed - The popup key on certain EuroLatin keys (such as "q", "y", "p") are truncated. Workaround is adjusting the keyboard height taller #6232 </details> - **TEST_INAPP_NON_LATIN_KEYBOARD:** Non-Latin script Keyboard <details> <summary>Expand to see how to do and verify this</summary> 1. In the app, add/select a random non-Latin script keyboard 2. Verify OSK is visible 3. Test several keys and modifiers and verify Latin characters are not appearing. </details> ### System Keyboards These tests are run with Keyman selected as a system keyboard, and in the context of typing in an external app (like Google Keep) - **TEST_SYSTEM_LATIN_KEYBOARD:** English (EuroLatin SIL) In portrait orientation, verify OSK is visible and fills the width the bottom of the screen <details> <summary>Expand to see how to do and verify this</summary> 1. Verify long-press `q` key works 2. Verify long-press `p` key works 3. Verify uppercase layer can be selected via `SHIFT` 4. Verify number layer can be selected via `123` 5. Verify long-press `1` key works 6. Verify long-press `0` key works 7. Verify backspace, space, and enter keys work 8. Verify can change to random non-Latin script keyboard via globe button 9. Repeat in landscape orientation </details> - **TEST_KEYBOARD_PICKER:** Keyboard Picker menu to switch input method This tests the Keyboard Picker menu which lists other system input methods at the bottom of the menu. These appear only in Keyman system keyboard <details> <summary>Expand to see how to do and verify this</summary> 1. With only 1 Keyman keyboard installed and selected as a system keyboard, long-press and release the the globe button 2. Verify the bottom of the Keyboards picker menu lists other available input methods 3. Click on one of the other input methods 4. Verify the Keyman system keyboard switches to the selected input method 5. On the device's Settings, select Keyman as the input method 6. On the Keyman keyboard, short-press and release the globe button 7. Verify the previous input method is selected </details> - **TEST_EXT_BLUETOOTH_KEYBOARD:** External (Bluetooth) keyboard This test is run with an external keyboard (USB or bluetooth) connected to the device <details> <summary>Expand to see how to do and verify this</summary> 1. Start with English(EuroLatin SIL) as the selected keyboard 2. On the device's Settings, show the input method (OSK) 3. Verify can type with numbers, letters, and punctuation 4. Verify `caps lock` has no effect when typing 5. Verify long-press on a character repeats that character 6. Verify `tab` toggles to next field 7. Verify `enter` adds a newline 8. Verify `backspace` removes a character 9. Verify <kbd>ctrl</kbd>+<kbd>tab</kbd> brings up the keyboard picker menu 10. Use <kbd>tab</kbd> or arrow keys to navigate the keyboard picker menu > Failed - navigating to other input methods selected the entire group #6233 11. Select a non-Latin keyboard 12. Resume typing and verify new script appears </details> - **TEST_EXT_AZERTY_KEYBOARD:** External European AZERTY keyboard This test is run with an external European AZERTY keyboard paired to the device. It should contain the 102nd key `<>` on the bottom row next to the left-shift. <details> <summary>Expand to see how to do and verify this</summary> 1. Start Keyman and install the French Basic keyboard 2. From the physical keyboard, type on the top letter row (azerty) 3. Verify `azerty` appears 4. Type on the 102nd key 5. Verify `<` appears 6. Type shift-102nd key 7. Verify `>` appears </details> </details> ---- ## SUITE_KB_DISTRIBUTION: Ad-hoc Keyboard Distribution Tests <details> <summary>Expand to see how to do and verify this</summary> - **TEST_KMP_DISTRIBUTION** KMP distribution <details> <summary>Expand to see how to do and verify this</summary> 1. Reset app settings (in Android Pie: Settings>Storage>Other apps>Keyman>Clear Storage) 2. Verify kmp [distribution](https://help.keyman.com/developer/current-version/guides/distribute/) successfully installs<br/> Example: [Khmer Angkor](https://downloads.keyman.com/keyboards/khmer_angkor/1.0.2/khmer_angkor.kmp) 3. Verify kmp successfully installs from Settings --> Install Keyboard or Dictionary --> Install from local file </details> </details> ---- ## What's New Tests - [ ] Refer to the new changes in `/HISTORY.md` and verify functionality
test
test android regression test keyman for android acceptance test procedures these test procedures are to be run before moving from alpha to beta or beta to stable or before prs are merged into stable branches copy these checklists of tests into a new issue for tier transitions or a comment on the pr when annotating results use at the start of a new line under the checkbox to note the result for an example see gather assets for testing physical android phone record device s android version lollipop or later external bluetooth or usb qwerty keyboard external bluetooth or usb azerty keyboard european azerty layout with key suite setup setup steps test uninstall uninstall previous version of keyman on the device don t forget to reset user settings for the app in android pie settings storage other apps keyman clear data test uninstall install test build apk test accept permission accept all the android permission requests for storage applies to android marshmallow onward user testing suite inspect visual inspection of master branch test ci web master verify latest ci android build of master is successful on test verify history verify history md contains all the current changes various versions of androids ui for these tests assume default english locale group android group android group android suite menu functionality menu functionality tests click to expand test get started this tests the get started menu from a fresh installation expand to see how to do and verify this when the app loads verify the get started menu is visible from the get started menu touch add a keyboard for your language verify install keyboard or dictionary menu appears from the install keyboard or dictionary menu click the back arrow to return to get started from the get started menu touch enable keyman as system wide keyboard verify android system menus appear for controlling virtual keyboard scroll down and enable keyman as a system keyboard on the attention dialog read and hit ok to dismiss the dialogs touch the back button to return to get started verify enable keyman as system wide keyboard now has a ticked checkbox from the get started menu touch set keyman as default keyboard on the change keyboard dialog select keyman verify on the get started menu that set keyman as default keyboard now has a ticked checkbox from the get started menu touch more info verify the info page appears with a version string at the bottom of the page touch the back button arrow to return to get started from the get started menu untick the last option show get started on startup close the get started menu and exit the app from the android device start keyman for android verify the get started menu does not appear test install kd this menu is accessed via get started menu or the settings button install keyboard or dictionary expand to see how to do and verify this on the device enable internet connectivity touch install from keyman com verify a new keyboard can be downloaded and selected on the device disable internet connectivity from install keyboard or dictionary touch install from keyman com verify device displays a message cannot connect to keyman server on the device re enable internet connectivity test keyman setttings expand to see how to do and verify this from settings installed languages select english english settings select keyboard settings menu verify info appears with keyboard version help link and qr code if keyboard update is available click on version and verify update if help link is available click on link verify keyboard help documentation appears return to english settings menu select dictionary english dictionary mtnt verify the dictionary info appears return to english settings menu turn on off enable predictions and validate each turn on off enable corrections and validate each for example when corrections are enabled and context tooo should still suggest too validate that when both predictions and corrections are off banner is not visible add remove dictionaries validate if multiple dictionaries are available test swapping between them test share expand to see how to do and verify this type and verify text can be shared to external app failed no external apps shown on android test keyman browser keyman browser expand to see how to do and verify this on default page click on text field and set system keyboard to keyman verify can type with keyman as system keyboard close and reopen keyman app test text size expand to see how to do and verify this verify text can be rescaled from text size to test clear text expand to see how to do and verify this verify text can be cleared test info expand to see how to do and verify this from the device disable internet access from the keyman app menu select info verify keyman for android version appears at the bottom of the screen verify keyman for android help documentation is displayed from the device enable internet access test install updates this menu option only appears when a language resource keyboard or lexical model update is available expand to see how to do and verify this click on install updates accept download on the dialogue verify resource is updated after background download complete not tested because updates not available test change display lang this tests changing the display language for the app and assumes starting with english locale expand to see how to do and verify this from settings display language select khmer from settings verify menu selections in khmer from the overflow menu clear text verify keyman text field has khmer prompt for start typing here from settings change display language if the string is translated the first row should be default locale in khmer select english verify the app reloads in english test adjust k height this menu option allows the user to adjust the osk height for the current orientation portrait or landscape the customized height is saved as a preference expand to see how to do and verify this have the device in portrait orientation from settings adjust keyboard height on the menu drag the osk height to change the keyboard height click back verify the osk is refreshed to the selected height from settings adjust keyboard height click the reset to defaults button and then click back verify the osk reverts to the original height rotate the device to landscape orientation from settings adjust keyboard height on the menu drag the osk height to change the keyboard height click back verify the osk is refreshed to the selected height from settings adjust keyboard height click the reset to defaults button and then click back verify the osk reverts to the original height rotate the device back to portrait orientation test spacebar caption this menu option allows the user to adjust the label displayed on the keyboard spacebar expand to see how to do and verify this select the default sil eurolatin keyboard from settings spacebar caption on the menu select language click back until the osk is displayed verify the spacebar label is english from settings spacebar caption on the menu select keyboard click back until the osk is displayed verify the spacebar label is eurolatin sil from settings spacebar caption on the menu select blank click back until the osk is displayed verify the spacebar label is blank from settings spacebar caption on the menu select language keyboard click back until the osk is displayed verify the spacebar label is english eurolatin sil suite keyboard functionality keyboard tests click to expand in app keyboards these tests are run in the context of typing in the keyman app test inapp latin keyboard english eurolatin sil in portrait orientation verify osk is visible and fills the width the bottom of the screen expand to see how to do and verify this verify long press q key works verify long press p key works verify uppercase layer can be selected via shift verify number layer can be selected via verify long press key works verify long press key works verify backspace space and enter keys work verify khmer khmer angkor keyboard can be added via settings menu repeat in landscape orientation landscape orientation failed the popup key on certain eurolatin keys such as q y p are truncated workaround is adjusting the keyboard height taller test inapp non latin keyboard non latin script keyboard expand to see how to do and verify this in the app add select a random non latin script keyboard verify osk is visible test several keys and modifiers and verify latin characters are not appearing system keyboards these tests are run with keyman selected as a system keyboard and in the context of typing in an external app like google keep test system latin keyboard english eurolatin sil in portrait orientation verify osk is visible and fills the width the bottom of the screen expand to see how to do and verify this verify long press q key works verify long press p key works verify uppercase layer can be selected via shift verify number layer can be selected via verify long press key works verify long press key works verify backspace space and enter keys work verify can change to random non latin script keyboard via globe button repeat in landscape orientation test keyboard picker keyboard picker menu to switch input method this tests the keyboard picker menu which lists other system input methods at the bottom of the menu these appear only in keyman system keyboard expand to see how to do and verify this with only keyman keyboard installed and selected as a system keyboard long press and release the the globe button verify the bottom of the keyboards picker menu lists other available input methods click on one of the other input methods verify the keyman system keyboard switches to the selected input method on the device s settings select keyman as the input method on the keyman keyboard short press and release the globe button verify the previous input method is selected test ext bluetooth keyboard external bluetooth keyboard this test is run with an external keyboard usb or bluetooth connected to the device expand to see how to do and verify this start with english eurolatin sil as the selected keyboard on the device s settings show the input method osk verify can type with numbers letters and punctuation verify caps lock has no effect when typing verify long press on a character repeats that character verify tab toggles to next field verify enter adds a newline verify backspace removes a character verify ctrl tab brings up the keyboard picker menu use tab or arrow keys to navigate the keyboard picker menu failed navigating to other input methods selected the entire group select a non latin keyboard resume typing and verify new script appears test ext azerty keyboard external european azerty keyboard this test is run with an external european azerty keyboard paired to the device it should contain the key on the bottom row next to the left shift expand to see how to do and verify this start keyman and install the french basic keyboard from the physical keyboard type on the top letter row azerty verify azerty appears type on the key verify appears type shift key verify appears suite kb distribution ad hoc keyboard distribution tests expand to see how to do and verify this test kmp distribution kmp distribution expand to see how to do and verify this reset app settings in android pie settings storage other apps keyman clear storage verify kmp successfully installs example verify kmp successfully installs from settings install keyboard or dictionary install from local file what s new tests refer to the new changes in history md and verify functionality
1
233,385
18,981,726,949
IssuesEvent
2021-11-21 01:39:22
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
opened
[Failing test][sig-api-machinery] aks-engine-windows-containerd-master
kind/failing-test
### Which jobs are failing? aks-engine-windows-containerd-master ### Which tests are failing? - ci-kubernetes-e2e-aks-engine-azure-master-windows-containerd.OverallChanges - Kubernetes e2e suite.[sig-api-machinery] Aggregator Should be able to support the 1.17 Sample API Server using the current Aggregator [Conformance]Changes - kubetest.TestChanges ### Since when has it been failing? 11/20/2021 03:27 CET ### Testgrid link https://testgrid.k8s.io/sig-release-master-informing#aks-engine-windows-containerd-master ### Reason for failure (if possible) [see run](https://prow.k8s.io/view/gs/kubernetes-jenkins/logs/ci-kubernetes-e2e-aks-engine-azure-master-windows-containerd/1462200742394728448) Kubernetes e2e suite: [sig-api-machinery] Aggregator Should be able to support the 1.17 Sample API Server using the current Aggregator [Conformance] ``` error waiting for deployment "sample-apiserver-deployment" ``` ```bash Failure [303.376 seconds] [sig-api-machinery] Aggregator Should be able to support the 1.17 Sample API Server using the current Aggregator [Conformance] [It] [Fail] [sig-api-machinery] Aggregator [It] Should be able to support the 1.17 Sample API Server using the current Aggregator [Conformance] Ran 305 of 7040 Specs in 2001.704 seconds FAIL! -- 304 Passed | 1 Failed | 0 Pending | 6735 Skipped ``` ### Anything else we need to know? /cc @jsturtevant ### Relevant SIG(s) /sig api-machinery
1.0
[Failing test][sig-api-machinery] aks-engine-windows-containerd-master - ### Which jobs are failing? aks-engine-windows-containerd-master ### Which tests are failing? - ci-kubernetes-e2e-aks-engine-azure-master-windows-containerd.OverallChanges - Kubernetes e2e suite.[sig-api-machinery] Aggregator Should be able to support the 1.17 Sample API Server using the current Aggregator [Conformance]Changes - kubetest.TestChanges ### Since when has it been failing? 11/20/2021 03:27 CET ### Testgrid link https://testgrid.k8s.io/sig-release-master-informing#aks-engine-windows-containerd-master ### Reason for failure (if possible) [see run](https://prow.k8s.io/view/gs/kubernetes-jenkins/logs/ci-kubernetes-e2e-aks-engine-azure-master-windows-containerd/1462200742394728448) Kubernetes e2e suite: [sig-api-machinery] Aggregator Should be able to support the 1.17 Sample API Server using the current Aggregator [Conformance] ``` error waiting for deployment "sample-apiserver-deployment" ``` ```bash Failure [303.376 seconds] [sig-api-machinery] Aggregator Should be able to support the 1.17 Sample API Server using the current Aggregator [Conformance] [It] [Fail] [sig-api-machinery] Aggregator [It] Should be able to support the 1.17 Sample API Server using the current Aggregator [Conformance] Ran 305 of 7040 Specs in 2001.704 seconds FAIL! -- 304 Passed | 1 Failed | 0 Pending | 6735 Skipped ``` ### Anything else we need to know? /cc @jsturtevant ### Relevant SIG(s) /sig api-machinery
test
aks engine windows containerd master which jobs are failing aks engine windows containerd master which tests are failing ci kubernetes aks engine azure master windows containerd overallchanges kubernetes suite aggregator should be able to support the sample api server using the current aggregator changes kubetest testchanges since when has it been failing cet testgrid link reason for failure if possible kubernetes suite aggregator should be able to support the sample api server using the current aggregator error waiting for deployment sample apiserver deployment bash failure aggregator should be able to support the sample api server using the current aggregator aggregator should be able to support the sample api server using the current aggregator ran of specs in seconds fail passed failed pending skipped anything else we need to know cc jsturtevant relevant sig s sig api machinery
1
642,288
20,883,495,155
IssuesEvent
2022-03-23 00:40:05
epicmaxco/vuestic-ui
https://api.github.com/repos/epicmaxco/vuestic-ui
opened
Composables refactoring
DISCUSSION refactoring LOW PRIORITY
Let's spend some time on composables improvements. - [ ] Having tests and some readmes for complex composables. - [ ] Move service composables to composables directory - [ ] Move component specific composables to components folder if it used ones. (useSelections, useAlign e.g.) - [ ] Rename hooks directories to composables. - [ ] Make tools that extracts prop types from useComposableProps instead of making extra interfaces (I made it already somewhere...) - [ ] useSyncProp vs useStateful? This two makes the same thing in different ways. - [ ] Move composables outside from mixins directory. - [ ] Make composables as part of vuestic-ui, maybe we have some useful composables for end-user (not sure if we have...). - [ ] Remove old mixin composables... Like useLoading or useColor.
1.0
Composables refactoring - Let's spend some time on composables improvements. - [ ] Having tests and some readmes for complex composables. - [ ] Move service composables to composables directory - [ ] Move component specific composables to components folder if it used ones. (useSelections, useAlign e.g.) - [ ] Rename hooks directories to composables. - [ ] Make tools that extracts prop types from useComposableProps instead of making extra interfaces (I made it already somewhere...) - [ ] useSyncProp vs useStateful? This two makes the same thing in different ways. - [ ] Move composables outside from mixins directory. - [ ] Make composables as part of vuestic-ui, maybe we have some useful composables for end-user (not sure if we have...). - [ ] Remove old mixin composables... Like useLoading or useColor.
non_test
composables refactoring let s spend some time on composables improvements having tests and some readmes for complex composables move service composables to composables directory move component specific composables to components folder if it used ones useselections usealign e g rename hooks directories to composables make tools that extracts prop types from usecomposableprops instead of making extra interfaces i made it already somewhere usesyncprop vs usestateful this two makes the same thing in different ways move composables outside from mixins directory make composables as part of vuestic ui maybe we have some useful composables for end user not sure if we have remove old mixin composables like useloading or usecolor
0
203,294
15,360,993,446
IssuesEvent
2021-03-01 17:35:08
abhiTronix/raspberry-pi-cross-compilers
https://api.github.com/repos/abhiTronix/raspberry-pi-cross-compilers
closed
Passing --sysroot option to compiler does not seem to work ?
Question :grey_question: Ready to Test :man_scientist:
Hello, from a x64 linux host, cross-compiling for a 32-bit Pi 4: (/opt/ossia-sdk-rpi/pi/sysroot/ contains the rsync'd /usr/lib and /usr/include from my pi) ```bash echo 'int main() { return 0; }' > foo.cpp /opt/ossia-sdk-rpi/cross-pi-gcc-10.2.0-2/bin/arm-linux-gnueabihf-g++ \ --sysroot=/opt/ossia-sdk-rpi/pi/sysroot/ \ foo.cpp ``` gives me ```bash /opt/ossia-sdk-rpi/cross-pi-gcc-10.2.0-2/bin/../lib/gcc/arm-linux-gnueabihf/10.2.0/../../../../arm-linux-gnueabihf/bin/ld: cannot find crt1.o: Aucun fichier ou dossier de ce type /opt/ossia-sdk-rpi/cross-pi-gcc-10.2.0-2/bin/../lib/gcc/arm-linux-gnueabihf/10.2.0/../../../../arm-linux-gnueabihf/bin/ld: cannot find crti.o: Aucun fichier ou dossier de ce type /opt/ossia-sdk-rpi/cross-pi-gcc-10.2.0-2/bin/../lib/gcc/arm-linux-gnueabihf/10.2.0/../../../../arm-linux-gnueabihf/bin/ld: cannot find -lm collect2: error: ld returned 1 exit status ``` what am I missing ? I tried adding a ton of link paths but that did not help: ```bash -L/opt/ossia-sdk-rpi/cross-pi-gcc-10.2.0-2/lib/ -L/opt/ossia-sdk-rpi/cross-pi-gcc-10.2.0-2/arm-linux-gnueabihf/lib/ -L/opt/ossia-sdk-rpi/cross-pi-gcc-10.2.0-2/arm-linux-gnueabihf/libc/usr/lib/ -L/opt/ossia-sdk-rpi/cross-pi-gcc-10.2.0-2/arm-linux-gnueabihf/libc/lib/ -L/opt/ossia-sdk-rpi/pi/sysroot/usr/lib/arm-linux-gnueabihf ```
1.0
Passing --sysroot option to compiler does not seem to work ? - Hello, from a x64 linux host, cross-compiling for a 32-bit Pi 4: (/opt/ossia-sdk-rpi/pi/sysroot/ contains the rsync'd /usr/lib and /usr/include from my pi) ```bash echo 'int main() { return 0; }' > foo.cpp /opt/ossia-sdk-rpi/cross-pi-gcc-10.2.0-2/bin/arm-linux-gnueabihf-g++ \ --sysroot=/opt/ossia-sdk-rpi/pi/sysroot/ \ foo.cpp ``` gives me ```bash /opt/ossia-sdk-rpi/cross-pi-gcc-10.2.0-2/bin/../lib/gcc/arm-linux-gnueabihf/10.2.0/../../../../arm-linux-gnueabihf/bin/ld: cannot find crt1.o: Aucun fichier ou dossier de ce type /opt/ossia-sdk-rpi/cross-pi-gcc-10.2.0-2/bin/../lib/gcc/arm-linux-gnueabihf/10.2.0/../../../../arm-linux-gnueabihf/bin/ld: cannot find crti.o: Aucun fichier ou dossier de ce type /opt/ossia-sdk-rpi/cross-pi-gcc-10.2.0-2/bin/../lib/gcc/arm-linux-gnueabihf/10.2.0/../../../../arm-linux-gnueabihf/bin/ld: cannot find -lm collect2: error: ld returned 1 exit status ``` what am I missing ? I tried adding a ton of link paths but that did not help: ```bash -L/opt/ossia-sdk-rpi/cross-pi-gcc-10.2.0-2/lib/ -L/opt/ossia-sdk-rpi/cross-pi-gcc-10.2.0-2/arm-linux-gnueabihf/lib/ -L/opt/ossia-sdk-rpi/cross-pi-gcc-10.2.0-2/arm-linux-gnueabihf/libc/usr/lib/ -L/opt/ossia-sdk-rpi/cross-pi-gcc-10.2.0-2/arm-linux-gnueabihf/libc/lib/ -L/opt/ossia-sdk-rpi/pi/sysroot/usr/lib/arm-linux-gnueabihf ```
test
passing sysroot option to compiler does not seem to work hello from a linux host cross compiling for a bit pi opt ossia sdk rpi pi sysroot contains the rsync d usr lib and usr include from my pi bash echo int main return foo cpp opt ossia sdk rpi cross pi gcc bin arm linux gnueabihf g sysroot opt ossia sdk rpi pi sysroot foo cpp gives me bash opt ossia sdk rpi cross pi gcc bin lib gcc arm linux gnueabihf arm linux gnueabihf bin ld cannot find o aucun fichier ou dossier de ce type opt ossia sdk rpi cross pi gcc bin lib gcc arm linux gnueabihf arm linux gnueabihf bin ld cannot find crti o aucun fichier ou dossier de ce type opt ossia sdk rpi cross pi gcc bin lib gcc arm linux gnueabihf arm linux gnueabihf bin ld cannot find lm error ld returned exit status what am i missing i tried adding a ton of link paths but that did not help bash l opt ossia sdk rpi cross pi gcc lib l opt ossia sdk rpi cross pi gcc arm linux gnueabihf lib l opt ossia sdk rpi cross pi gcc arm linux gnueabihf libc usr lib l opt ossia sdk rpi cross pi gcc arm linux gnueabihf libc lib l opt ossia sdk rpi pi sysroot usr lib arm linux gnueabihf
1
184,104
21,784,795,325
IssuesEvent
2022-05-14 01:20:48
szb512/cypress
https://api.github.com/repos/szb512/cypress
opened
CVE-2022-1650 (High) detected in eventsource-1.0.7.tgz
security vulnerability
## CVE-2022-1650 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>eventsource-1.0.7.tgz</b></p></summary> <p>W3C compliant EventSource client for Node.js and browser (polyfill)</p> <p>Library home page: <a href="https://registry.npmjs.org/eventsource/-/eventsource-1.0.7.tgz">https://registry.npmjs.org/eventsource/-/eventsource-1.0.7.tgz</a></p> <p>Path to dependency file: /cypress/packages/server/package.json</p> <p>Path to vulnerable library: /packages/server/node_modules/eventsource/package.json</p> <p> Dependency Hierarchy: - :x: **eventsource-1.0.7.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/szb512/cypress/commit/29dcad339d37f2169e5a640bf8d0d1438f7c18c2">29dcad339d37f2169e5a640bf8d0d1438f7c18c2</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Exposure of Sensitive Information to an Unauthorized Actor in GitHub repository eventsource/eventsource prior to v2.0.2. <p>Publish Date: 2022-05-12 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-1650>CVE-2022-1650</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://huntr.dev/bounties/dc9e467f-be5d-4945-867d-1044d27e9b8e/">https://huntr.dev/bounties/dc9e467f-be5d-4945-867d-1044d27e9b8e/</a></p> <p>Release Date: 2022-05-12</p> <p>Fix Resolution: eventsource - 2.0.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-1650 (High) detected in eventsource-1.0.7.tgz - ## CVE-2022-1650 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>eventsource-1.0.7.tgz</b></p></summary> <p>W3C compliant EventSource client for Node.js and browser (polyfill)</p> <p>Library home page: <a href="https://registry.npmjs.org/eventsource/-/eventsource-1.0.7.tgz">https://registry.npmjs.org/eventsource/-/eventsource-1.0.7.tgz</a></p> <p>Path to dependency file: /cypress/packages/server/package.json</p> <p>Path to vulnerable library: /packages/server/node_modules/eventsource/package.json</p> <p> Dependency Hierarchy: - :x: **eventsource-1.0.7.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/szb512/cypress/commit/29dcad339d37f2169e5a640bf8d0d1438f7c18c2">29dcad339d37f2169e5a640bf8d0d1438f7c18c2</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Exposure of Sensitive Information to an Unauthorized Actor in GitHub repository eventsource/eventsource prior to v2.0.2. <p>Publish Date: 2022-05-12 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-1650>CVE-2022-1650</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://huntr.dev/bounties/dc9e467f-be5d-4945-867d-1044d27e9b8e/">https://huntr.dev/bounties/dc9e467f-be5d-4945-867d-1044d27e9b8e/</a></p> <p>Release Date: 2022-05-12</p> <p>Fix Resolution: eventsource - 2.0.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in eventsource tgz cve high severity vulnerability vulnerable library eventsource tgz compliant eventsource client for node js and browser polyfill library home page a href path to dependency file cypress packages server package json path to vulnerable library packages server node modules eventsource package json dependency hierarchy x eventsource tgz vulnerable library found in head commit a href vulnerability details exposure of sensitive information to an unauthorized actor in github repository eventsource eventsource prior to publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution eventsource step up your open source security game with whitesource
0
211,248
16,191,933,420
IssuesEvent
2021-05-04 09:42:27
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachtest: schemachange/database-version-upgrade failed
C-test-failure O-roachtest O-robot branch-master release-blocker
roachtest.schemachange/database-version-upgrade [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2944550&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=2944550&tab=artifacts#/schemachange/database-version-upgrade) on master @ [684e753c15f3fc58df79b6ea70e7b6715eae4835](https://github.com/cockroachdb/cockroach/commits/684e753c15f3fc58df79b6ea70e7b6715eae4835): ``` The test failed on branch=master, cloud=gce: test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/schemachange/database-version-upgrade/run_1 schema_change_database_version_upgrade.go:51,schema_change_database_version_upgrade.go:38,test_runner.go:777: prev version not set for version: 21.2 (1) attached stack trace -- stack trace: | main.PredecessorVersion | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/test_runner.go:1210 | main.runSchemaChangeDatabaseVersionUpgrade | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/schema_change_database_version_upgrade.go:49 | main.registerSchemaChangeDatabaseVersionUpgrade.func1 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/schema_change_database_version_upgrade.go:38 | main.(*testRunner).runTest.func2 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/test_runner.go:777 | runtime.goexit | /usr/local/go/src/runtime/asm_amd64.s:1374 Wraps: (2) prev version not set for version: 21.2 Error types: (1) *withstack.withStack (2) *errutil.leafError ``` <details><summary>Reproduce</summary> <p> <p>To reproduce, try: ```bash # From https://go.crdb.dev/p/roachstress, perhaps edited lightly. caffeinate ./roachstress.sh schemachange/database-version-upgrade ``` </p> </p> </details> /cc @cockroachdb/sql-schema <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*schemachange/database-version-upgrade.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
2.0
roachtest: schemachange/database-version-upgrade failed - roachtest.schemachange/database-version-upgrade [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2944550&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=2944550&tab=artifacts#/schemachange/database-version-upgrade) on master @ [684e753c15f3fc58df79b6ea70e7b6715eae4835](https://github.com/cockroachdb/cockroach/commits/684e753c15f3fc58df79b6ea70e7b6715eae4835): ``` The test failed on branch=master, cloud=gce: test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/schemachange/database-version-upgrade/run_1 schema_change_database_version_upgrade.go:51,schema_change_database_version_upgrade.go:38,test_runner.go:777: prev version not set for version: 21.2 (1) attached stack trace -- stack trace: | main.PredecessorVersion | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/test_runner.go:1210 | main.runSchemaChangeDatabaseVersionUpgrade | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/schema_change_database_version_upgrade.go:49 | main.registerSchemaChangeDatabaseVersionUpgrade.func1 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/schema_change_database_version_upgrade.go:38 | main.(*testRunner).runTest.func2 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/test_runner.go:777 | runtime.goexit | /usr/local/go/src/runtime/asm_amd64.s:1374 Wraps: (2) prev version not set for version: 21.2 Error types: (1) *withstack.withStack (2) *errutil.leafError ``` <details><summary>Reproduce</summary> <p> <p>To reproduce, try: ```bash # From https://go.crdb.dev/p/roachstress, perhaps edited lightly. caffeinate ./roachstress.sh schemachange/database-version-upgrade ``` </p> </p> </details> /cc @cockroachdb/sql-schema <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*schemachange/database-version-upgrade.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
test
roachtest schemachange database version upgrade failed roachtest schemachange database version upgrade with on master the test failed on branch master cloud gce test artifacts and logs in home agent work go src github com cockroachdb cockroach artifacts schemachange database version upgrade run schema change database version upgrade go schema change database version upgrade go test runner go prev version not set for version attached stack trace stack trace main predecessorversion home agent work go src github com cockroachdb cockroach pkg cmd roachtest test runner go main runschemachangedatabaseversionupgrade home agent work go src github com cockroachdb cockroach pkg cmd roachtest schema change database version upgrade go main registerschemachangedatabaseversionupgrade home agent work go src github com cockroachdb cockroach pkg cmd roachtest schema change database version upgrade go main testrunner runtest home agent work go src github com cockroachdb cockroach pkg cmd roachtest test runner go runtime goexit usr local go src runtime asm s wraps prev version not set for version error types withstack withstack errutil leaferror reproduce to reproduce try bash from perhaps edited lightly caffeinate roachstress sh schemachange database version upgrade cc cockroachdb sql schema
1
298,138
25,791,951,595
IssuesEvent
2022-12-10 06:32:18
EddieHubCommunity/LinkFree
https://api.github.com/repos/EddieHubCommunity/LinkFree
closed
New Testimonial
testimonial
### Name eddiejaoude ### Title Updated testimonial for EJ ### Description testing update
1.0
New Testimonial - ### Name eddiejaoude ### Title Updated testimonial for EJ ### Description testing update
test
new testimonial name eddiejaoude title updated testimonial for ej description testing update
1
27,625
5,377,469,505
IssuesEvent
2017-02-23 12:12:02
gap-system/gap
https://api.github.com/repos/gap-system/gap
closed
minor glitch in reference manual
documentation newcomer-friendly
The Index in the Reference manual has an entry "PCore" immediately after "OnTuplesTuples".
1.0
minor glitch in reference manual - The Index in the Reference manual has an entry "PCore" immediately after "OnTuplesTuples".
non_test
minor glitch in reference manual the index in the reference manual has an entry pcore immediately after ontuplestuples
0
40,099
10,451,900,156
IssuesEvent
2019-09-19 13:44:47
CleverRaven/Cataclysm-DDA
https://api.github.com/repos/CleverRaven/Cataclysm-DDA
closed
[Tools] Better pre commit support for linting, specifically win/vs builds
<Enhancement / Feature> Code: Astyle / Optimization / Static Analysis Code: Build OS: Windows [JSON] stale
I think I've now run into astyle or json linting issues dozens of times.... This is wasting a hell of a lot of people's time. I think this needs better linting support, especially if developing on win/vs2015+. My understanding is that the cmake build process catches all those errors, vs builds *do not*. So.... "_Need better tools_". Feel free to assign this to me. I'm heavily leaning to pre-commit hooks which simply barf on linting issues. A few powershell scripts shouldn't be too much of a hassle.
1.0
[Tools] Better pre commit support for linting, specifically win/vs builds - I think I've now run into astyle or json linting issues dozens of times.... This is wasting a hell of a lot of people's time. I think this needs better linting support, especially if developing on win/vs2015+. My understanding is that the cmake build process catches all those errors, vs builds *do not*. So.... "_Need better tools_". Feel free to assign this to me. I'm heavily leaning to pre-commit hooks which simply barf on linting issues. A few powershell scripts shouldn't be too much of a hassle.
non_test
better pre commit support for linting specifically win vs builds i think i ve now run into astyle or json linting issues dozens of times this is wasting a hell of a lot of people s time i think this needs better linting support especially if developing on win my understanding is that the cmake build process catches all those errors vs builds do not so need better tools feel free to assign this to me i m heavily leaning to pre commit hooks which simply barf on linting issues a few powershell scripts shouldn t be too much of a hassle
0
8,586
2,611,531,725
IssuesEvent
2015-02-27 06:03:16
chrsmith/hedgewars
https://api.github.com/repos/chrsmith/hedgewars
closed
debian ubuntu hedgewars dependencies
auto-migrated Priority-Medium Type-Defect
``` Hi, I'm trying to look at the dependencies you listed here [1] and I found that in debian/ubuntu we have some more build-deps fp-compiler bzip2 fp-units-gfx libghc-stm-dev fp-units-misc Are them really needed? could we safely drop them? thanks G. [1] http://code.google.com/p/hedgewars/wiki/BuildingOnLinux ``` Original issue reported on code.google.com by `costamag...@gmail.com` on 23 May 2013 at 12:17
1.0
debian ubuntu hedgewars dependencies - ``` Hi, I'm trying to look at the dependencies you listed here [1] and I found that in debian/ubuntu we have some more build-deps fp-compiler bzip2 fp-units-gfx libghc-stm-dev fp-units-misc Are them really needed? could we safely drop them? thanks G. [1] http://code.google.com/p/hedgewars/wiki/BuildingOnLinux ``` Original issue reported on code.google.com by `costamag...@gmail.com` on 23 May 2013 at 12:17
non_test
debian ubuntu hedgewars dependencies hi i m trying to look at the dependencies you listed here and i found that in debian ubuntu we have some more build deps fp compiler fp units gfx libghc stm dev fp units misc are them really needed could we safely drop them thanks g original issue reported on code google com by costamag gmail com on may at
0
7,800
2,934,164,973
IssuesEvent
2015-06-30 06:30:52
e-government-ua/i
https://api.github.com/repos/e-government-ua/i
closed
Ошибка Π² словС "РСєстраційні Π΄Π°Π½Ρ–" Π² мСню "ΠŸΠΎΡΠ»ΡƒΠ³ΠΈ"
bug test
1. ΠžΡ‚ΠΊΡ€Ρ‹Ρ‚ΡŒ мСню "ΠŸΠΎΡΠ»ΡƒΠ³ΠΈ". 2. Π’ Ρ€Π°Π·Π΄Π΅Π»Π°Ρ… "Громадянам" ΠΈ "БізнСсу" Π½Π°ΠΉΡ‚ΠΈ ΠΏΠΎΠ΄Ρ€Π°Π·Π΄Π΅Π» "РСєстраційни Π΄Π°Π½Ρ–". ![data](https://cloud.githubusercontent.com/assets/12966279/8397488/7532f1b4-1dd5-11e5-8efd-d847ae30ac72.jpg) ![data1](https://cloud.githubusercontent.com/assets/12966279/8397489/75339042-1dd5-11e5-9add-78803f9574ef.jpg) ЀактичСский Ρ€Π΅Π·ΡƒΠ»ΡŒΡ‚Π°Ρ‚: Ошибка Π² словС "РСєстраційни Π΄Π°Π½Ρ–" ΠžΠΆΠΈΠ΄Π°Π΅ΠΌΡ‹ΠΉ Ρ€Π΅Π·ΡƒΠ»ΡŒΡ‚Π°Ρ‚: Π‘Π»ΠΎΠ²ΠΎ "РСєстраційні Π΄Π°Π½Ρ–" Π±Π΅Π· ошибки
1.0
Ошибка Π² словС "РСєстраційні Π΄Π°Π½Ρ–" Π² мСню "ΠŸΠΎΡΠ»ΡƒΠ³ΠΈ" - 1. ΠžΡ‚ΠΊΡ€Ρ‹Ρ‚ΡŒ мСню "ΠŸΠΎΡΠ»ΡƒΠ³ΠΈ". 2. Π’ Ρ€Π°Π·Π΄Π΅Π»Π°Ρ… "Громадянам" ΠΈ "БізнСсу" Π½Π°ΠΉΡ‚ΠΈ ΠΏΠΎΠ΄Ρ€Π°Π·Π΄Π΅Π» "РСєстраційни Π΄Π°Π½Ρ–". ![data](https://cloud.githubusercontent.com/assets/12966279/8397488/7532f1b4-1dd5-11e5-8efd-d847ae30ac72.jpg) ![data1](https://cloud.githubusercontent.com/assets/12966279/8397489/75339042-1dd5-11e5-9add-78803f9574ef.jpg) ЀактичСский Ρ€Π΅Π·ΡƒΠ»ΡŒΡ‚Π°Ρ‚: Ошибка Π² словС "РСєстраційни Π΄Π°Π½Ρ–" ΠžΠΆΠΈΠ΄Π°Π΅ΠΌΡ‹ΠΉ Ρ€Π΅Π·ΡƒΠ»ΡŒΡ‚Π°Ρ‚: Π‘Π»ΠΎΠ²ΠΎ "РСєстраційні Π΄Π°Π½Ρ–" Π±Π΅Π· ошибки
test
ошибка Π² словС рСєстраційні Π΄Π°Π½Ρ– Π² мСню послуги ΠΎΡ‚ΠΊΡ€Ρ‹Ρ‚ΡŒ мСню послуги Π² Ρ€Π°Π·Π΄Π΅Π»Π°Ρ… громадянам ΠΈ бізнСсу Π½Π°ΠΉΡ‚ΠΈ ΠΏΠΎΠ΄Ρ€Π°Π·Π΄Π΅Π» рСєстраційни Π΄Π°Π½Ρ– фактичСский Ρ€Π΅Π·ΡƒΠ»ΡŒΡ‚Π°Ρ‚ ошибка Π² словС рСєстраційни Π΄Π°Π½Ρ– ΠΎΠΆΠΈΠ΄Π°Π΅ΠΌΡ‹ΠΉ Ρ€Π΅Π·ΡƒΠ»ΡŒΡ‚Π°Ρ‚ слово рСєстраційні Π΄Π°Π½Ρ– Π±Π΅Π· ошибки
1
415,505
28,039,788,406
IssuesEvent
2023-03-28 17:35:29
hashicorp/terraform-provider-aws
https://api.github.com/repos/hashicorp/terraform-provider-aws
opened
[Docs]: Route53 Record import is misleading
documentation needs-triage
### Documentation Link https://registry.terraform.io/providers/hashicorp/aws/latest/docs/resources/route53_record#import ### Description The documentation for importing a Route53 record shows an example of the following: ``` terraform import aws_route53_record.myrecord Z4KAPRWWNC7JR_dev.example.com_NS` ``` After some experimentation, we realized that (at least for A records), you need to use just the record name and not the FQDN in the import. If you use the FQDN, then Terraform later tries to replace the record. I think the example should use just `dev` and not the FQDN: ``` $ terraform import aws_route53_record.myrecord Z4KAPRWWNC7JR_dev_NS ``` ### References _No response_ ### Would you like to implement a fix? None
1.0
[Docs]: Route53 Record import is misleading - ### Documentation Link https://registry.terraform.io/providers/hashicorp/aws/latest/docs/resources/route53_record#import ### Description The documentation for importing a Route53 record shows an example of the following: ``` terraform import aws_route53_record.myrecord Z4KAPRWWNC7JR_dev.example.com_NS` ``` After some experimentation, we realized that (at least for A records), you need to use just the record name and not the FQDN in the import. If you use the FQDN, then Terraform later tries to replace the record. I think the example should use just `dev` and not the FQDN: ``` $ terraform import aws_route53_record.myrecord Z4KAPRWWNC7JR_dev_NS ``` ### References _No response_ ### Would you like to implement a fix? None
non_test
record import is misleading documentation link description the documentation for importing a record shows an example of the following terraform import aws record myrecord dev example com ns after some experimentation we realized that at least for a records you need to use just the record name and not the fqdn in the import if you use the fqdn then terraform later tries to replace the record i think the example should use just dev and not the fqdn terraform import aws record myrecord dev ns references no response would you like to implement a fix none
0
171,341
13,228,975,539
IssuesEvent
2020-08-18 07:20:03
NKCR-INPROVE/evidence.periodik
https://api.github.com/repos/NKCR-INPROVE/evidence.periodik
closed
Kde se bere mutace a proč není ve fasetÑch?
ready for test
![image](https://user-images.githubusercontent.com/811537/85961598-c4c4cf00-b9ab-11ea-842d-09f99bafe462.png) KdyΕΎ kliknu na hlavnΓ­ strΓ‘nce na LidovΓ© noviny - VydΓ‘nΓ­ Brno, tak se mi zobrazΓ­ seznam se zΓ‘hlavΓ­m Metatitul: LidovΓ© noviny - VydΓ‘nΓ­ Brno Mutace: Brno. Kde se bere info o mutaci, proč je v zΓ‘hlavΓ­ a nenΓ­ ve fasetΔ›?
1.0
Kde se bere mutace a proč nenΓ­ ve fasetΓ‘ch? - ![image](https://user-images.githubusercontent.com/811537/85961598-c4c4cf00-b9ab-11ea-842d-09f99bafe462.png) KdyΕΎ kliknu na hlavnΓ­ strΓ‘nce na LidovΓ© noviny - VydΓ‘nΓ­ Brno, tak se mi zobrazΓ­ seznam se zΓ‘hlavΓ­m Metatitul: LidovΓ© noviny - VydΓ‘nΓ­ Brno Mutace: Brno. Kde se bere info o mutaci, proč je v zΓ‘hlavΓ­ a nenΓ­ ve fasetΔ›?
test
kde se bere mutace a proč nenΓ­ ve fasetΓ‘ch kdyΕΎ kliknu na hlavnΓ­ strΓ‘nce na lidovΓ© noviny vydΓ‘nΓ­ brno tak se mi zobrazΓ­ seznam se zΓ‘hlavΓ­m metatitul lidovΓ© noviny vydΓ‘nΓ­ brno mutace brno kde se bere info o mutaci proč je v zΓ‘hlavΓ­ a nenΓ­ ve fasetΔ›
1
234,018
19,090,508,121
IssuesEvent
2021-11-29 11:32:42
spring-projects/spring-framework
https://api.github.com/repos/spring-projects/spring-framework
closed
Remove dependency on Commons FileUpload in ContentRequestMatchers
status: waiting-for-triage in: test
In Spring 5.3, `multipartData()` methods were added to `org.springframework.test.web.client.match.ContentRequestMatchers`, which are useful for testing a multipart request body. As a slight limitation, to use this method with Spring 5.3, we need to add dependency on Commons-Fileupload in test scope. The dependency was removed from Spring 6.0 in conjunction with #27423. It would great if this change to `ContentRequestMatchers` would be backported to Spring 5.3.x.
1.0
Remove dependency on Commons FileUpload in ContentRequestMatchers - In Spring 5.3, `multipartData()` methods were added to `org.springframework.test.web.client.match.ContentRequestMatchers`, which are useful for testing a multipart request body. As a slight limitation, to use this method with Spring 5.3, we need to add dependency on Commons-Fileupload in test scope. The dependency was removed from Spring 6.0 in conjunction with #27423. It would great if this change to `ContentRequestMatchers` would be backported to Spring 5.3.x.
test
remove dependency on commons fileupload in contentrequestmatchers in spring multipartdata methods were added to org springframework test web client match contentrequestmatchers which are useful for testing a multipart request body as a slight limitation to use this method with spring we need to add dependency on commons fileupload in test scope the dependency was removed from spring in conjunction with it would great if this change to contentrequestmatchers would be backported to spring x
1
566,966
16,835,512,020
IssuesEvent
2021-06-18 11:29:21
metabase/metabase
https://api.github.com/repos/metabase/metabase
closed
Unable to share sub-collections from Personal Collection
.Reproduced Administration/Permissions Organization/Collections Priority:P3 Type:Bug
### Bugs If you're filing an issue about a bug please include as much information as you can including the following. - Your browser and the version: Chrome 68.0.3440.106 - Your operating system: OS X 10 - Your databases: Redshift - Metabase version: 0.30.1 - Metabase hosting environment: ECS/Docker - Metabase internal database: Postgres ## Repro Steps: 1. Create a sub-collection in your Personal Collection 2. Navigate into the sub-collection, hit Edit the permissions for this collection - at this point no one has access to this collection: ![image](https://user-images.githubusercontent.com/40185566/44677427-bcd02880-a9ea-11e8-93bf-226b18175b9f.png) 3. Make a change granting another group view access: ![image](https://user-images.githubusercontent.com/40185566/44677438-c5c0fa00-a9ea-11e8-8c0a-64dae8c28760.png) 4. **Issue 1:** Get message about All Users already having greater access? Doesn't make much sense... ![image](https://user-images.githubusercontent.com/40185566/44677469-d5404300-a9ea-11e8-93e8-5d69d12a7a95.png) 5. **Issue 2:** Hit Revoke access, then Save, get this message: ![image](https://user-images.githubusercontent.com/40185566/44677497-f012b780-a9ea-11e8-8ab9-a915b59131a8.png) 6. Logs show: ```PUT /api/collection/graph 500 (71 ms) (7 DB calls). {:message "You cannot edit permissions for a Personal Collection or its descendants.", :type java.lang.Exception, :stacktrace ["models.permissions$check_not_personal_collection_or_descendant.invokeStatic(permissions.clj:474)" "models.permissions$check_not_personal_collection_or_descendant.invoke(permissions.clj:461)" "models.permissions$revoke_collection_permissions_BANG_.invokeStatic(permissions.clj:479)" "models.permissions$revoke_collection_permissions_BANG_.invoke(permissions.clj:476)" "models.collection$fn__27294$update_collection_permissions_BANG___27299$fn__27300.invoke(collection.clj:878)" "models.collection$fn__27294$update_collection_permissions_BANG___27299.invoke(collection.clj:870)" "models.collection$fn__27319$update_group_permissions_BANG___27324$fn__27325.invoke(collection.clj:887)" "models.collection$fn__27319$update_group_permissions_BANG___27324.invoke(collection.clj:884)" "models.collection$fn__27360$update_graph_BANG___27369$fn__27370$fn__27374.invoke(collection.clj:914)" "models.collection$fn__27360$update_graph_BANG___27369$fn__27370.invoke(collection.clj:912)" "models.collection$fn__27360$update_graph_BANG___27369.invoke(collection.clj:902)" "api.collection$fn__46474.invokeStatic(collection.clj:304)" "api.collection$fn__46474.invoke(collection.clj:299)" "middleware$enforce_authentication$fn__55250.invoke(middleware.clj:116)" "api.routes$fn__55392.invokeStatic(routes.clj:62)" "api.routes$fn__55392.invoke(routes.clj:62)" "routes$fn__55477$fn__55478.doInvoke(routes.clj:108)" "routes$fn__55477.invokeStatic(routes.clj:103)" "routes$fn__55477.invoke(routes.clj:103)" "middleware$catch_api_exceptions$fn__55379.invoke(middleware.clj:424)" "middleware$log_api_call$fn__55356$fn__55358.invoke(middleware.clj:351)" "middleware$log_api_call$fn__55356.invoke(middleware.clj:350)" "middleware$add_security_headers$fn__55304.invoke(middleware.clj:253)" "core$wrap_streamed_json_response$fn__60666.invoke(core.clj:67)" "middleware$bind_current_user$fn__55255.invoke(middleware.clj:140)" "middleware$maybe_set_site_url$fn__55308.invoke(middleware.clj:277)"]} ``` I'm really hoping the intention is that we're able to share sub-collections from our Personal Collection with different groups (assuming that's why the option exists in the UI). There's already been a lot of confusion about this after we upgraded this morning, right now everyone is just saving everything in the 'Migrated Questions' collections to get around this.
1.0
Unable to share sub-collections from Personal Collection - ### Bugs If you're filing an issue about a bug please include as much information as you can including the following. - Your browser and the version: Chrome 68.0.3440.106 - Your operating system: OS X 10 - Your databases: Redshift - Metabase version: 0.30.1 - Metabase hosting environment: ECS/Docker - Metabase internal database: Postgres ## Repro Steps: 1. Create a sub-collection in your Personal Collection 2. Navigate into the sub-collection, hit Edit the permissions for this collection - at this point no one has access to this collection: ![image](https://user-images.githubusercontent.com/40185566/44677427-bcd02880-a9ea-11e8-93bf-226b18175b9f.png) 3. Make a change granting another group view access: ![image](https://user-images.githubusercontent.com/40185566/44677438-c5c0fa00-a9ea-11e8-8c0a-64dae8c28760.png) 4. **Issue 1:** Get message about All Users already having greater access? Doesn't make much sense... ![image](https://user-images.githubusercontent.com/40185566/44677469-d5404300-a9ea-11e8-93e8-5d69d12a7a95.png) 5. **Issue 2:** Hit Revoke access, then Save, get this message: ![image](https://user-images.githubusercontent.com/40185566/44677497-f012b780-a9ea-11e8-8ab9-a915b59131a8.png) 6. Logs show: ```PUT /api/collection/graph 500 (71 ms) (7 DB calls). {:message "You cannot edit permissions for a Personal Collection or its descendants.", :type java.lang.Exception, :stacktrace ["models.permissions$check_not_personal_collection_or_descendant.invokeStatic(permissions.clj:474)" "models.permissions$check_not_personal_collection_or_descendant.invoke(permissions.clj:461)" "models.permissions$revoke_collection_permissions_BANG_.invokeStatic(permissions.clj:479)" "models.permissions$revoke_collection_permissions_BANG_.invoke(permissions.clj:476)" "models.collection$fn__27294$update_collection_permissions_BANG___27299$fn__27300.invoke(collection.clj:878)" "models.collection$fn__27294$update_collection_permissions_BANG___27299.invoke(collection.clj:870)" "models.collection$fn__27319$update_group_permissions_BANG___27324$fn__27325.invoke(collection.clj:887)" "models.collection$fn__27319$update_group_permissions_BANG___27324.invoke(collection.clj:884)" "models.collection$fn__27360$update_graph_BANG___27369$fn__27370$fn__27374.invoke(collection.clj:914)" "models.collection$fn__27360$update_graph_BANG___27369$fn__27370.invoke(collection.clj:912)" "models.collection$fn__27360$update_graph_BANG___27369.invoke(collection.clj:902)" "api.collection$fn__46474.invokeStatic(collection.clj:304)" "api.collection$fn__46474.invoke(collection.clj:299)" "middleware$enforce_authentication$fn__55250.invoke(middleware.clj:116)" "api.routes$fn__55392.invokeStatic(routes.clj:62)" "api.routes$fn__55392.invoke(routes.clj:62)" "routes$fn__55477$fn__55478.doInvoke(routes.clj:108)" "routes$fn__55477.invokeStatic(routes.clj:103)" "routes$fn__55477.invoke(routes.clj:103)" "middleware$catch_api_exceptions$fn__55379.invoke(middleware.clj:424)" "middleware$log_api_call$fn__55356$fn__55358.invoke(middleware.clj:351)" "middleware$log_api_call$fn__55356.invoke(middleware.clj:350)" "middleware$add_security_headers$fn__55304.invoke(middleware.clj:253)" "core$wrap_streamed_json_response$fn__60666.invoke(core.clj:67)" "middleware$bind_current_user$fn__55255.invoke(middleware.clj:140)" "middleware$maybe_set_site_url$fn__55308.invoke(middleware.clj:277)"]} ``` I'm really hoping the intention is that we're able to share sub-collections from our Personal Collection with different groups (assuming that's why the option exists in the UI). There's already been a lot of confusion about this after we upgraded this morning, right now everyone is just saving everything in the 'Migrated Questions' collections to get around this.
non_test
unable to share sub collections from personal collection bugs if you re filing an issue about a bug please include as much information as you can including the following your browser and the version chrome your operating system os x your databases redshift metabase version metabase hosting environment ecs docker metabase internal database postgres repro steps create a sub collection in your personal collection navigate into the sub collection hit edit the permissions for this collection at this point no one has access to this collection make a change granting another group view access issue get message about all users already having greater access doesn t make much sense issue hit revoke access then save get this message logs show put api collection graph ms db calls message you cannot edit permissions for a personal collection or its descendants type java lang exception stacktrace models permissions check not personal collection or descendant invokestatic permissions clj models permissions check not personal collection or descendant invoke permissions clj models permissions revoke collection permissions bang invokestatic permissions clj models permissions revoke collection permissions bang invoke permissions clj models collection fn update collection permissions bang fn invoke collection clj models collection fn update collection permissions bang invoke collection clj models collection fn update group permissions bang fn invoke collection clj models collection fn update group permissions bang invoke collection clj models collection fn update graph bang fn fn invoke collection clj models collection fn update graph bang fn invoke collection clj models collection fn update graph bang invoke collection clj api collection fn invokestatic collection clj api collection fn invoke collection clj middleware enforce authentication fn invoke middleware clj api routes fn invokestatic routes clj api routes fn invoke routes clj routes fn fn doinvoke routes clj routes fn invokestatic routes clj routes fn invoke routes clj middleware catch api exceptions fn invoke middleware clj middleware log api call fn fn invoke middleware clj middleware log api call fn invoke middleware clj middleware add security headers fn invoke middleware clj core wrap streamed json response fn invoke core clj middleware bind current user fn invoke middleware clj middleware maybe set site url fn invoke middleware clj i m really hoping the intention is that we re able to share sub collections from our personal collection with different groups assuming that s why the option exists in the ui there s already been a lot of confusion about this after we upgraded this morning right now everyone is just saving everything in the migrated questions collections to get around this
0
88,909
8,180,506,998
IssuesEvent
2018-08-28 19:41:46
cmu-db/terrier
https://api.github.com/repos/cmu-db/terrier
opened
New scenarios for large_transaction_test (w/ GC)
beginner tests
We have large scale tests that generate randomized workloads to run against the storage layer. The general framework is in place, and it has helped us find many bugs, but we have discovered that certain bugs only appear under certain configurations (e.g. high abort rates, large number of read-only transactions). Currently we only have one test scenario for each of the test cases, with numbers that largely don't mean anything. The goal would be to come up with a variety of meaningful configurations that will help us smoke out different classes of bugs. The tests themselves are under `test/storage/large_garbage_collector_test.cpp` and `test/transaction/large_transaction_test.cpp`. The small framework we use for them can be found in `test/util/transaction_test_util.cpp` and `test/include/util/transaction_test_util.h`. The tunable parameters exposed are documented and the existing test cases should serve as good examples. Come up with a configuration, describe the scenario it models, and if you are curious, verify that it can find bugs by injecting the kind of bug you think this would lead to into the transaction system.
1.0
New scenarios for large_transaction_test (w/ GC) - We have large scale tests that generate randomized workloads to run against the storage layer. The general framework is in place, and it has helped us find many bugs, but we have discovered that certain bugs only appear under certain configurations (e.g. high abort rates, large number of read-only transactions). Currently we only have one test scenario for each of the test cases, with numbers that largely don't mean anything. The goal would be to come up with a variety of meaningful configurations that will help us smoke out different classes of bugs. The tests themselves are under `test/storage/large_garbage_collector_test.cpp` and `test/transaction/large_transaction_test.cpp`. The small framework we use for them can be found in `test/util/transaction_test_util.cpp` and `test/include/util/transaction_test_util.h`. The tunable parameters exposed are documented and the existing test cases should serve as good examples. Come up with a configuration, describe the scenario it models, and if you are curious, verify that it can find bugs by injecting the kind of bug you think this would lead to into the transaction system.
test
new scenarios for large transaction test w gc we have large scale tests that generate randomized workloads to run against the storage layer the general framework is in place and it has helped us find many bugs but we have discovered that certain bugs only appear under certain configurations e g high abort rates large number of read only transactions currently we only have one test scenario for each of the test cases with numbers that largely don t mean anything the goal would be to come up with a variety of meaningful configurations that will help us smoke out different classes of bugs the tests themselves are under test storage large garbage collector test cpp and test transaction large transaction test cpp the small framework we use for them can be found in test util transaction test util cpp and test include util transaction test util h the tunable parameters exposed are documented and the existing test cases should serve as good examples come up with a configuration describe the scenario it models and if you are curious verify that it can find bugs by injecting the kind of bug you think this would lead to into the transaction system
1
140,661
11,355,044,113
IssuesEvent
2020-01-24 19:04:58
pandas-dev/pandas
https://api.github.com/repos/pandas-dev/pandas
closed
BUG: Calling DataFrame.stack on an out-of-order column MultiIndex leads to swapped values
Needs Tests good first issue
Please run the code below. Notice how the column values are swapped to the wrong labels. This is due to `stack()` failing to preserve the order in the MultiIndex. #### Code Sample, a copy-pastable example if possible ```python import numpy as np import pandas as pd values = np.arange(5) data = np.vstack([['b{}'.format(x) for x in values], # b0, b1, .. ['a{}'.format(x) for x in values]]) # a0, a1, .. df = pd.DataFrame(data.T, columns=['b', 'a']) df.columns.name = 'first' # Call pd.concat to get the 2-level MultiIndex *unsorted* columns. # The bug seems to happen when having one of these unsorted MultiIndexes. second_level_dict = {'x': df} multi_level_df = pd.concat(second_level_dict, axis=1) multi_level_df.columns.names = ['second', 'first'] # Sort the columns, i.e. [a, b] instead of [b, a]. sorted_cols_df = multi_level_df.reindex(sorted(multi_level_df.columns), axis=1) print('Before the restack:') print(sorted_cols_df) # Stack and unstack, should be the same. # This is what causes the bug. sorted_cols_df.stack() also exposes the problem restacked = sorted_cols_df.stack(['first', 'second']).unstack(['first', 'second']) print() print('Restacked:') print(restacked) print('(Notice the swapped column values)') ``` #### Output ``` $ python pandas_bug.py Before the restack: second x first a b 0 a0 b0 1 a1 b1 2 a2 b2 3 a3 b3 4 a4 b4 Restacked: first a b second x x 0 b0 a0 <-- notice the swapped values 1 b1 a1 2 b2 a2 3 b3 a3 4 b4 a4 ``` #### Output of ``pd.show_versions()`` I've reproduced this on both 0.21 and 0.20. <details> In [2]: pd.show_versions() INSTALLED VERSIONS ------------------ commit: None python: 3.6.3.final.0 python-bits: 64 OS: Linux OS-release: 4.4.0-97-generic machine: x86_64 processor: x86_64 byteorder: little LC_ALL: None LANG: en_US.UTF-8 LOCALE: en_US.UTF-8 pandas: 0.21.0 pytest: 3.2.1 pip: 9.0.1 setuptools: 36.6.0 Cython: 0.26.1 numpy: 1.13.3 scipy: 0.19.1 pyarrow: None xarray: None IPython: 6.2.1 sphinx: 1.6.3 patsy: 0.4.1 dateutil: 2.6.1 pytz: 2017.3 blosc: None bottleneck: 1.2.1 tables: 3.4.2 numexpr: 2.6.2 feather: None matplotlib: 2.1.0 openpyxl: 2.4.8 xlrd: 1.1.0 xlwt: 1.3.0 xlsxwriter: 1.0.2 lxml: 4.1.0 bs4: 4.6.0 html5lib: 0.999999999 sqlalchemy: 1.1.13 pymysql: None psycopg2: None jinja2: 2.9.6 s3fs: None fastparquet: None pandas_gbq: None pandas_datareader: None </details>
1.0
BUG: Calling DataFrame.stack on an out-of-order column MultiIndex leads to swapped values - Please run the code below. Notice how the column values are swapped to the wrong labels. This is due to `stack()` failing to preserve the order in the MultiIndex. #### Code Sample, a copy-pastable example if possible ```python import numpy as np import pandas as pd values = np.arange(5) data = np.vstack([['b{}'.format(x) for x in values], # b0, b1, .. ['a{}'.format(x) for x in values]]) # a0, a1, .. df = pd.DataFrame(data.T, columns=['b', 'a']) df.columns.name = 'first' # Call pd.concat to get the 2-level MultiIndex *unsorted* columns. # The bug seems to happen when having one of these unsorted MultiIndexes. second_level_dict = {'x': df} multi_level_df = pd.concat(second_level_dict, axis=1) multi_level_df.columns.names = ['second', 'first'] # Sort the columns, i.e. [a, b] instead of [b, a]. sorted_cols_df = multi_level_df.reindex(sorted(multi_level_df.columns), axis=1) print('Before the restack:') print(sorted_cols_df) # Stack and unstack, should be the same. # This is what causes the bug. sorted_cols_df.stack() also exposes the problem restacked = sorted_cols_df.stack(['first', 'second']).unstack(['first', 'second']) print() print('Restacked:') print(restacked) print('(Notice the swapped column values)') ``` #### Output ``` $ python pandas_bug.py Before the restack: second x first a b 0 a0 b0 1 a1 b1 2 a2 b2 3 a3 b3 4 a4 b4 Restacked: first a b second x x 0 b0 a0 <-- notice the swapped values 1 b1 a1 2 b2 a2 3 b3 a3 4 b4 a4 ``` #### Output of ``pd.show_versions()`` I've reproduced this on both 0.21 and 0.20. <details> In [2]: pd.show_versions() INSTALLED VERSIONS ------------------ commit: None python: 3.6.3.final.0 python-bits: 64 OS: Linux OS-release: 4.4.0-97-generic machine: x86_64 processor: x86_64 byteorder: little LC_ALL: None LANG: en_US.UTF-8 LOCALE: en_US.UTF-8 pandas: 0.21.0 pytest: 3.2.1 pip: 9.0.1 setuptools: 36.6.0 Cython: 0.26.1 numpy: 1.13.3 scipy: 0.19.1 pyarrow: None xarray: None IPython: 6.2.1 sphinx: 1.6.3 patsy: 0.4.1 dateutil: 2.6.1 pytz: 2017.3 blosc: None bottleneck: 1.2.1 tables: 3.4.2 numexpr: 2.6.2 feather: None matplotlib: 2.1.0 openpyxl: 2.4.8 xlrd: 1.1.0 xlwt: 1.3.0 xlsxwriter: 1.0.2 lxml: 4.1.0 bs4: 4.6.0 html5lib: 0.999999999 sqlalchemy: 1.1.13 pymysql: None psycopg2: None jinja2: 2.9.6 s3fs: None fastparquet: None pandas_gbq: None pandas_datareader: None </details>
test
bug calling dataframe stack on an out of order column multiindex leads to swapped values please run the code below notice how the column values are swapped to the wrong labels this is due to stack failing to preserve the order in the multiindex code sample a copy pastable example if possible python import numpy as np import pandas as pd values np arange data np vstack df pd dataframe data t columns df columns name first call pd concat to get the level multiindex unsorted columns the bug seems to happen when having one of these unsorted multiindexes second level dict x df multi level df pd concat second level dict axis multi level df columns names sort the columns i e instead of sorted cols df multi level df reindex sorted multi level df columns axis print before the restack print sorted cols df stack and unstack should be the same this is what causes the bug sorted cols df stack also exposes the problem restacked sorted cols df stack unstack print print restacked print restacked print notice the swapped column values output python pandas bug py before the restack second x first a b restacked first a b second x x notice the swapped values output of pd show versions i ve reproduced this on both and in pd show versions installed versions commit none python final python bits os linux os release generic machine processor byteorder little lc all none lang en us utf locale en us utf pandas pytest pip setuptools cython numpy scipy pyarrow none xarray none ipython sphinx patsy dateutil pytz blosc none bottleneck tables numexpr feather none matplotlib openpyxl xlrd xlwt xlsxwriter lxml sqlalchemy pymysql none none none fastparquet none pandas gbq none pandas datareader none
1
69,513
3,304,917,315
IssuesEvent
2015-11-04 00:29:55
Sonarr/Sonarr
https://api.github.com/repos/Sonarr/Sonarr
closed
Updater fails to enumerate processes on FreeBSD
bug priority:medium
``` 15-7-21 10:48:38.7|Debug|ProcessProvider|Finding process with Id:60550 15-7-21 10:48:38.7|Fatal|UpdateApp|An error has occurred while applying update package. System.NotSupportedException: This system does not support EnumProcesses at (wrapper managed-to-native) System.Diagnostics.Process:GetProcesses_internal () at System.Diagnostics.Process.GetProcesses () [0x00000] in /usr/ports/lang/mono/work/mono-4.0.1/mcs/class/System/System.Diagnostics/Process.cs:834 at NzbDrone.Common.Processes.ProcessProvider.GetProcessById (Int32 id) [0x00023] in m:\BuildAgent\work\6c3239faf2b92630\src\NzbDrone.Common\Processes\ProcessProvider.cs:73 at NzbDrone.Common.Processes.ProcessProvider.Exists (Int32 processId) [0x00000] in m:\BuildAgent\work\6c3239faf2b92630\src\NzbDrone.Common\Processes\ProcessProvider.cs:56 at NzbDrone.Update.UpdateEngine.InstallUpdateService.Verify (System.String targetFolder, Int32 processId) [0x0005c] in m:\BuildAgent\work\6c3239faf2b92630\src\NzbDrone.Update\UpdateEngine\InstallUpdateService.cs:67 at NzbDrone.Update.UpdateEngine.InstallUpdateService.Start (System.String installationFolder, Int32 processId) [0x00000] in m:\BuildAgent\work\6c3239faf2b92630\src\NzbDrone.Update\UpdateEngine\InstallUpdateService.cs:79 at NzbDrone.Update.UpdateApp.Start (System.String[] args) [0x00020] in m:\BuildAgent\work\6c3239faf2b92630\src\NzbDrone.Update\UpdateApp.cs:59 at NzbDrone.Update.UpdateApp.Main (System.String[] args) [0x00042] in m:\BuildAgent\work\6c3239faf2b92630\src\NzbDrone.Update\UpdateApp.cs:43 ```
1.0
Updater fails to enumerate processes on FreeBSD - ``` 15-7-21 10:48:38.7|Debug|ProcessProvider|Finding process with Id:60550 15-7-21 10:48:38.7|Fatal|UpdateApp|An error has occurred while applying update package. System.NotSupportedException: This system does not support EnumProcesses at (wrapper managed-to-native) System.Diagnostics.Process:GetProcesses_internal () at System.Diagnostics.Process.GetProcesses () [0x00000] in /usr/ports/lang/mono/work/mono-4.0.1/mcs/class/System/System.Diagnostics/Process.cs:834 at NzbDrone.Common.Processes.ProcessProvider.GetProcessById (Int32 id) [0x00023] in m:\BuildAgent\work\6c3239faf2b92630\src\NzbDrone.Common\Processes\ProcessProvider.cs:73 at NzbDrone.Common.Processes.ProcessProvider.Exists (Int32 processId) [0x00000] in m:\BuildAgent\work\6c3239faf2b92630\src\NzbDrone.Common\Processes\ProcessProvider.cs:56 at NzbDrone.Update.UpdateEngine.InstallUpdateService.Verify (System.String targetFolder, Int32 processId) [0x0005c] in m:\BuildAgent\work\6c3239faf2b92630\src\NzbDrone.Update\UpdateEngine\InstallUpdateService.cs:67 at NzbDrone.Update.UpdateEngine.InstallUpdateService.Start (System.String installationFolder, Int32 processId) [0x00000] in m:\BuildAgent\work\6c3239faf2b92630\src\NzbDrone.Update\UpdateEngine\InstallUpdateService.cs:79 at NzbDrone.Update.UpdateApp.Start (System.String[] args) [0x00020] in m:\BuildAgent\work\6c3239faf2b92630\src\NzbDrone.Update\UpdateApp.cs:59 at NzbDrone.Update.UpdateApp.Main (System.String[] args) [0x00042] in m:\BuildAgent\work\6c3239faf2b92630\src\NzbDrone.Update\UpdateApp.cs:43 ```
non_test
updater fails to enumerate processes on freebsd debug processprovider finding process with id fatal updateapp an error has occurred while applying update package system notsupportedexception this system does not support enumprocesses at wrapper managed to native system diagnostics process getprocesses internal at system diagnostics process getprocesses in usr ports lang mono work mono mcs class system system diagnostics process cs at nzbdrone common processes processprovider getprocessbyid id in m buildagent work src nzbdrone common processes processprovider cs at nzbdrone common processes processprovider exists processid in m buildagent work src nzbdrone common processes processprovider cs at nzbdrone update updateengine installupdateservice verify system string targetfolder processid in m buildagent work src nzbdrone update updateengine installupdateservice cs at nzbdrone update updateengine installupdateservice start system string installationfolder processid in m buildagent work src nzbdrone update updateengine installupdateservice cs at nzbdrone update updateapp start system string args in m buildagent work src nzbdrone update updateapp cs at nzbdrone update updateapp main system string args in m buildagent work src nzbdrone update updateapp cs
0
327,082
28,042,048,735
IssuesEvent
2023-03-28 19:19:47
aws/eks-anywhere
https://api.github.com/repos/aws/eks-anywhere
opened
Add RHEL Tinkerbell tests
area/testing
RHEL only supports running in BIOS mode. Add support for switching BIOS mode on test framework to enable testing rhel
1.0
Add RHEL Tinkerbell tests - RHEL only supports running in BIOS mode. Add support for switching BIOS mode on test framework to enable testing rhel
test
add rhel tinkerbell tests rhel only supports running in bios mode add support for switching bios mode on test framework to enable testing rhel
1
382,402
26,497,463,452
IssuesEvent
2023-01-18 07:28:15
zauberzeug/nicegui
https://api.github.com/repos/zauberzeug/nicegui
closed
Serving app behind reverse proxy subpath
documentation
When trying to serve a nicegui app behind Nginx reverse proxy, I get an Error message "Connection lost. Trying to reconnect... Basic Nginx configuration ``` ... location /nicegui/ { proxy_http_version 1.1; proxy_set_header Connection ""; proxy_pass http://127.0.0.1:5556/; } ... ``` main.py from the Basic Example ``` from nicegui import ui ui.label('Hello NiceGUI!') ui.run(host="127.0.0.1", port=5556, show=False) ``` Any advice would be appreciated.
1.0
Serving app behind reverse proxy subpath - When trying to serve a nicegui app behind Nginx reverse proxy, I get an Error message "Connection lost. Trying to reconnect... Basic Nginx configuration ``` ... location /nicegui/ { proxy_http_version 1.1; proxy_set_header Connection ""; proxy_pass http://127.0.0.1:5556/; } ... ``` main.py from the Basic Example ``` from nicegui import ui ui.label('Hello NiceGUI!') ui.run(host="127.0.0.1", port=5556, show=False) ``` Any advice would be appreciated.
non_test
serving app behind reverse proxy subpath when trying to serve a nicegui app behind nginx reverse proxy i get an error message connection lost trying to reconnect basic nginx configuration location nicegui proxy http version proxy set header connection proxy pass main py from the basic example from nicegui import ui ui label hello nicegui ui run host port show false any advice would be appreciated
0
67,519
14,879,932,060
IssuesEvent
2021-01-20 08:28:40
loggly/cloudwatch-metrics-to-loggly
https://api.github.com/repos/loggly/cloudwatch-metrics-to-loggly
opened
CVE-2020-11022 (Medium) detected in jquery-1.7.2.min.js
security vulnerability
## CVE-2020-11022 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.7.2.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.2/jquery.min.js</a></p> <p>Path to dependency file: cloudwatch-metrics-to-loggly/node_modules/jmespath/index.html</p> <p>Path to vulnerable library: cloudwatch-metrics-to-loggly/node_modules/jmespath/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.7.2.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://api.github.com/repos/loggly/cloudwatch-metrics-to-loggly/commits/0e5d12923faa2f8d052a0c407b6f4f21b2a06fd1">0e5d12923faa2f8d052a0c407b6f4f21b2a06fd1</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022>CVE-2020-11022</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/">https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jQuery - 3.5.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.7.2","isTransitiveDependency":false,"dependencyTree":"jquery:1.7.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jQuery - 3.5.0"}],"vulnerabilityIdentifier":"CVE-2020-11022","vulnerabilityDetails":"In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery\u0027s DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-11022 (Medium) detected in jquery-1.7.2.min.js - ## CVE-2020-11022 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.7.2.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.2/jquery.min.js</a></p> <p>Path to dependency file: cloudwatch-metrics-to-loggly/node_modules/jmespath/index.html</p> <p>Path to vulnerable library: cloudwatch-metrics-to-loggly/node_modules/jmespath/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.7.2.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://api.github.com/repos/loggly/cloudwatch-metrics-to-loggly/commits/0e5d12923faa2f8d052a0c407b6f4f21b2a06fd1">0e5d12923faa2f8d052a0c407b6f4f21b2a06fd1</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022>CVE-2020-11022</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/">https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jQuery - 3.5.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.7.2","isTransitiveDependency":false,"dependencyTree":"jquery:1.7.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jQuery - 3.5.0"}],"vulnerabilityIdentifier":"CVE-2020-11022","vulnerabilityDetails":"In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery\u0027s DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
non_test
cve medium detected in jquery min js cve medium severity vulnerability vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file cloudwatch metrics to loggly node modules jmespath index html path to vulnerable library cloudwatch metrics to loggly node modules jmespath index html dependency hierarchy x jquery min js vulnerable library found in head commit a href found in base branch master vulnerability details in jquery versions greater than or equal to and before passing html from untrusted sources even after sanitizing it to one of jquery s dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails in jquery versions greater than or equal to and before passing html from untrusted sources even after sanitizing it to one of jquery dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery vulnerabilityurl
0
396,935
11,716,000,704
IssuesEvent
2020-03-09 14:59:40
craftercms/craftercms
https://api.github.com/repos/craftercms/craftercms
closed
[studio] S3 datasources with only slash in path should upload to root of s3 folder
bug priority: medium
## Describe the bug Currently having only `/` in the path for any S3 datasources cause the upload and upload from browse to save the files in `//` ## To Reproduce Steps to reproduce the behavior: 1. Add a browse / upload S3 datasource and set the Repository Path to `/` 2. Upload a file to S3 3. The file in Studio is reference as `/remote-assets/s3/s3-default//download.html` notice the double `/` ## Expected behavior When Repository Path is set to `/` Studio should save the files in the root of S3 ## Screenshots File uploaded in S3 is saved in `/` folder in the root of S3 ![imagen](https://user-images.githubusercontent.com/6722074/75591395-3e23d380-5a45-11ea-9828-a640eb8e8270.png) ## Logs N/A ## Specs ### Version `3.1.6-SNAPSHOT` ### OS Any ### Browser Any ## Additional context Related to https://github.com/craftercms/craftercms/issues/3680
1.0
[studio] S3 datasources with only slash in path should upload to root of s3 folder - ## Describe the bug Currently having only `/` in the path for any S3 datasources cause the upload and upload from browse to save the files in `//` ## To Reproduce Steps to reproduce the behavior: 1. Add a browse / upload S3 datasource and set the Repository Path to `/` 2. Upload a file to S3 3. The file in Studio is reference as `/remote-assets/s3/s3-default//download.html` notice the double `/` ## Expected behavior When Repository Path is set to `/` Studio should save the files in the root of S3 ## Screenshots File uploaded in S3 is saved in `/` folder in the root of S3 ![imagen](https://user-images.githubusercontent.com/6722074/75591395-3e23d380-5a45-11ea-9828-a640eb8e8270.png) ## Logs N/A ## Specs ### Version `3.1.6-SNAPSHOT` ### OS Any ### Browser Any ## Additional context Related to https://github.com/craftercms/craftercms/issues/3680
non_test
datasources with only slash in path should upload to root of folder describe the bug currently having only in the path for any datasources cause the upload and upload from browse to save the files in to reproduce steps to reproduce the behavior add a browse upload datasource and set the repository path to upload a file to the file in studio is reference as remote assets default download html notice the double expected behavior when repository path is set to studio should save the files in the root of screenshots file uploaded in is saved in folder in the root of logs n a specs version snapshot os any browser any additional context related to
0
2,810
5,738,545,505
IssuesEvent
2017-04-23 05:28:07
SIMEXP/niak
https://api.github.com/repos/SIMEXP/niak
closed
Automatic removing of failed QC subjects when grabbing preprocessed data
enhancement preprocessing quality control
Add an option in niak_grab_fmri_preprocess that grab only qc passed subject if it find a qc_report.csv file.
1.0
Automatic removing of failed QC subjects when grabbing preprocessed data - Add an option in niak_grab_fmri_preprocess that grab only qc passed subject if it find a qc_report.csv file.
non_test
automatic removing of failed qc subjects when grabbing preprocessed data add an option in niak grab fmri preprocess that grab only qc passed subject if it find a qc report csv file
0
628,187
19,977,612,997
IssuesEvent
2022-01-29 10:54:01
slynch8/10x
https://api.github.com/repos/slynch8/10x
closed
Syntax highlighter keeps blinking when typing
bug check Priority 3
Sometimes, when I'm typing, sytanx highlighter keeps refreshing and blinking, like rapidly turning on and off. The GIF doesn't show clearly, as it happens very fast, but you can notice two or three blinks. ![Animation4](https://user-images.githubusercontent.com/5081921/136603126-d06d14f9-34d9-47cc-9789-e5444c491be1.gif)
1.0
Syntax highlighter keeps blinking when typing - Sometimes, when I'm typing, sytanx highlighter keeps refreshing and blinking, like rapidly turning on and off. The GIF doesn't show clearly, as it happens very fast, but you can notice two or three blinks. ![Animation4](https://user-images.githubusercontent.com/5081921/136603126-d06d14f9-34d9-47cc-9789-e5444c491be1.gif)
non_test
syntax highlighter keeps blinking when typing sometimes when i m typing sytanx highlighter keeps refreshing and blinking like rapidly turning on and off the gif doesn t show clearly as it happens very fast but you can notice two or three blinks
0
38,382
5,184,282,323
IssuesEvent
2017-01-20 05:15:27
gravityview/GravityView
https://api.github.com/repos/gravityview/GravityView
closed
Helper_Functions_Test errors and GravityView_API_Test failure
Status: Needs Testing
``` Time: 34.26 seconds, Memory: 68.00MB There were 2 errors: 1) GravityView_Helper_Functions_Test::test_gv_empty Undefined variable: not_defined /tests/unit-tests/helper-functions_Test.php:21 2) GravityView_Helper_Functions_Test::test_gravityview_number_format A non well formed numeric value encountered /src/wp-includes/functions.php:219 /tests/unit-tests/helper-functions_Test.php:155 -- There was 1 failure: 1) GravityView_API_Test::test_gravityview_get_current_views The $current_views array didn't have a value set at $post->ID key Failed asserting that false is true. /tests/unit-tests/GravityView_API_Test.php:356 ERRORS! Tests: 83, Assertions: 1216, Errors: 2, Failures: 1. ``` Against 1e56843 using PHPUnit 5.7, WordPress 4.7 and GravityForms HEAD at 976fd3ea (~Dec 23 master), on custom-built PHP 7.0.13 (cli) (built: Nov 18 2016 00:35:50) ( NTS ), stock 10.0.28-MariaDB, and Ubuntu 16.04.1 LTS (Linux 4.4.1). Let me know if there's any other information I can provide.
1.0
Helper_Functions_Test errors and GravityView_API_Test failure - ``` Time: 34.26 seconds, Memory: 68.00MB There were 2 errors: 1) GravityView_Helper_Functions_Test::test_gv_empty Undefined variable: not_defined /tests/unit-tests/helper-functions_Test.php:21 2) GravityView_Helper_Functions_Test::test_gravityview_number_format A non well formed numeric value encountered /src/wp-includes/functions.php:219 /tests/unit-tests/helper-functions_Test.php:155 -- There was 1 failure: 1) GravityView_API_Test::test_gravityview_get_current_views The $current_views array didn't have a value set at $post->ID key Failed asserting that false is true. /tests/unit-tests/GravityView_API_Test.php:356 ERRORS! Tests: 83, Assertions: 1216, Errors: 2, Failures: 1. ``` Against 1e56843 using PHPUnit 5.7, WordPress 4.7 and GravityForms HEAD at 976fd3ea (~Dec 23 master), on custom-built PHP 7.0.13 (cli) (built: Nov 18 2016 00:35:50) ( NTS ), stock 10.0.28-MariaDB, and Ubuntu 16.04.1 LTS (Linux 4.4.1). Let me know if there's any other information I can provide.
test
helper functions test errors and gravityview api test failure time seconds memory there were errors gravityview helper functions test test gv empty undefined variable not defined tests unit tests helper functions test php gravityview helper functions test test gravityview number format a non well formed numeric value encountered src wp includes functions php tests unit tests helper functions test php there was failure gravityview api test test gravityview get current views the current views array didn t have a value set at post id key failed asserting that false is true tests unit tests gravityview api test php errors tests assertions errors failures against using phpunit wordpress and gravityforms head at dec master on custom built php cli built nov nts stock mariadb and ubuntu lts linux let me know if there s any other information i can provide
1
252,219
19,007,880,689
IssuesEvent
2021-11-23 04:10:39
girlscript/winter-of-contributing
https://api.github.com/repos/girlscript/winter-of-contributing
closed
[FRONTEND WITH FRAMEWORK]: Angular(8.4) -> Build the application
documentation GWOC21 Angular Frontend Dev React/Angular/Vue
### Description - How to build an angular application using CLI command? - Explain with syntax/example - Reference :- https://angular.io/cli/build (Week 8 Issue) Note:- 1. Don't copy paste, we would be merging after reviewing the PR. 2. Changes should be made inside "Frontend_Web_Development_React_Angular_Vue/Angular/Week8" directory 3. Tasks would be assigned on First Come First Serve Basis ### Domain Frontend Dev React/Angular/Vue ### Type of Contribution Documentation ### Code of Conduct - [X] I follow [Contributing Guidelines](https://github.com/girlscript/winter-of-contributing/blob/main/.github/CONTRIBUTING.md) & [Code of conduct](https://github.com/girlscript/winter-of-contributing/blob/main/.github/CODE_OF_CONDUCT.md) of this project.
1.0
[FRONTEND WITH FRAMEWORK]: Angular(8.4) -> Build the application - ### Description - How to build an angular application using CLI command? - Explain with syntax/example - Reference :- https://angular.io/cli/build (Week 8 Issue) Note:- 1. Don't copy paste, we would be merging after reviewing the PR. 2. Changes should be made inside "Frontend_Web_Development_React_Angular_Vue/Angular/Week8" directory 3. Tasks would be assigned on First Come First Serve Basis ### Domain Frontend Dev React/Angular/Vue ### Type of Contribution Documentation ### Code of Conduct - [X] I follow [Contributing Guidelines](https://github.com/girlscript/winter-of-contributing/blob/main/.github/CONTRIBUTING.md) & [Code of conduct](https://github.com/girlscript/winter-of-contributing/blob/main/.github/CODE_OF_CONDUCT.md) of this project.
non_test
angular build the application description how to build an angular application using cli command explain with syntax example reference week issue note don t copy paste we would be merging after reviewing the pr changes should be made inside frontend web development react angular vue angular directory tasks would be assigned on first come first serve basis domain frontend dev react angular vue type of contribution documentation code of conduct i follow of this project
0
69,561
7,153,469,622
IssuesEvent
2018-01-26 01:56:31
GeoDaCenter/geoda
https://api.github.com/repos/GeoDaCenter/geoda
closed
Crash: Merge Table
to be tested
From: Details: os: 1-10-10 vs: 1-12-1-59 /Applications/GeoDa.app/Contents/MacOS/../Resources/logger.txt GdaFrame::ShowOpenDatasourceDlg() ConnectDatasourceDlg::InitSamplePanel() Check auto update: AutoUpdate::CheckUpdate() AutoUpdate::GetCheckList() AutoUpdate::ReadUrlContent() Entering ConnectDatasourceDlg::OnOkClick ConnectDatasourceDlg::CreateDataSource() Entering ConnectDatasourceDlg::SaveRecentDataSource Exiting ConnectDatasourceDlg::SaveRecentDataSource Exiting ConnectDatasourceDlg::OnOkClick Entering Project::Project (new project) Project::CommonProjectInit() Entering Project::InitFromOgrLayer Datasource name: Project::GetSpatialReference() Exiting Project::Project Click GdaFrame::InitWithProject() Open TableFrame. Open MapFrame. MapCanvas::MapCanvas() MapCanvas::ChangeMapType() MapCanvas::VarInfoAttributeChange() MapCanvas::CreateAndUpdateCategories() In TableFrame::OnActivate Open MergeTableDlg. Entering MergeTableDlg::OnOpenClick() Entering ConnectDatasourceDlg::OnOkClick ConnectDatasourceDlg::CreateDataSource() Entering ConnectDatasourceDlg::SaveRecentDataSource Exiting ConnectDatasourceDlg::SaveRecentDataSource In TableFrame::OnActivate Exiting ConnectDatasourceDlg::OnOkClick
1.0
Crash: Merge Table - From: Details: os: 1-10-10 vs: 1-12-1-59 /Applications/GeoDa.app/Contents/MacOS/../Resources/logger.txt GdaFrame::ShowOpenDatasourceDlg() ConnectDatasourceDlg::InitSamplePanel() Check auto update: AutoUpdate::CheckUpdate() AutoUpdate::GetCheckList() AutoUpdate::ReadUrlContent() Entering ConnectDatasourceDlg::OnOkClick ConnectDatasourceDlg::CreateDataSource() Entering ConnectDatasourceDlg::SaveRecentDataSource Exiting ConnectDatasourceDlg::SaveRecentDataSource Exiting ConnectDatasourceDlg::OnOkClick Entering Project::Project (new project) Project::CommonProjectInit() Entering Project::InitFromOgrLayer Datasource name: Project::GetSpatialReference() Exiting Project::Project Click GdaFrame::InitWithProject() Open TableFrame. Open MapFrame. MapCanvas::MapCanvas() MapCanvas::ChangeMapType() MapCanvas::VarInfoAttributeChange() MapCanvas::CreateAndUpdateCategories() In TableFrame::OnActivate Open MergeTableDlg. Entering MergeTableDlg::OnOpenClick() Entering ConnectDatasourceDlg::OnOkClick ConnectDatasourceDlg::CreateDataSource() Entering ConnectDatasourceDlg::SaveRecentDataSource Exiting ConnectDatasourceDlg::SaveRecentDataSource In TableFrame::OnActivate Exiting ConnectDatasourceDlg::OnOkClick
test
crash merge table from details os vs applications geoda app contents macos resources logger txt gdaframe showopendatasourcedlg connectdatasourcedlg initsamplepanel check auto update autoupdate checkupdate autoupdate getchecklist autoupdate readurlcontent entering connectdatasourcedlg onokclick connectdatasourcedlg createdatasource entering connectdatasourcedlg saverecentdatasource exiting connectdatasourcedlg saverecentdatasource exiting connectdatasourcedlg onokclick entering project project new project project commonprojectinit entering project initfromogrlayer datasource name project getspatialreference exiting project project click gdaframe initwithproject open tableframe open mapframe mapcanvas mapcanvas mapcanvas changemaptype mapcanvas varinfoattributechange mapcanvas createandupdatecategories in tableframe onactivate open mergetabledlg entering mergetabledlg onopenclick entering connectdatasourcedlg onokclick connectdatasourcedlg createdatasource entering connectdatasourcedlg saverecentdatasource exiting connectdatasourcedlg saverecentdatasource in tableframe onactivate exiting connectdatasourcedlg onokclick
1
194,759
22,262,305,070
IssuesEvent
2022-06-10 02:25:14
bazelbuild/bazel-buildfarm
https://api.github.com/repos/bazelbuild/bazel-buildfarm
closed
several critical cve in the image: bazelbuild/buildfarm-worker:171ac6fe
security
can see more details by run this command: trivy image bazelbuild/buildfarm-worker:171ac6fe
True
several critical cve in the image: bazelbuild/buildfarm-worker:171ac6fe - can see more details by run this command: trivy image bazelbuild/buildfarm-worker:171ac6fe
non_test
several critical cve in the image bazelbuild buildfarm worker can see more details by run this command trivy image bazelbuild buildfarm worker
0
203,068
15,343,301,559
IssuesEvent
2021-02-27 19:42:54
hasii2011/PyUt
https://api.github.com/repos/hasii2011/PyUt
closed
Move Command Tests
UnitTests
Since commands package is now a sub-package of the history manager; need to move unit tests to matching package
1.0
Move Command Tests - Since commands package is now a sub-package of the history manager; need to move unit tests to matching package
test
move command tests since commands package is now a sub package of the history manager need to move unit tests to matching package
1
147,133
13,201,213,846
IssuesEvent
2020-08-14 09:42:20
tektoncd/website
https://api.github.com/repos/tektoncd/website
reopened
Getting started guides for project
help wanted kind/documentation lifecycle/stale
We need better "Getting started guide" for the tekton projects. - They should be on the `docs/` folder of each of those project (like `docs/getting-started/…`) - Repositories to tackle - [`tektoncd/pipeline`](https://github.com/tektoncd/pipeline) - [`tektoncd/triggers`](https://github.com/tektoncd/triggers) β€” the [existing one](https://github.com/tektoncd/triggers/tree/master/docs/getting-started) needs enhancement - [`tektoncd/cli`](https://github.com/tektoncd/cli) - [`tektoncd/dashboard`](https://github.com/tektoncd/dashboard) /kind docs
1.0
Getting started guides for project - We need better "Getting started guide" for the tekton projects. - They should be on the `docs/` folder of each of those project (like `docs/getting-started/…`) - Repositories to tackle - [`tektoncd/pipeline`](https://github.com/tektoncd/pipeline) - [`tektoncd/triggers`](https://github.com/tektoncd/triggers) β€” the [existing one](https://github.com/tektoncd/triggers/tree/master/docs/getting-started) needs enhancement - [`tektoncd/cli`](https://github.com/tektoncd/cli) - [`tektoncd/dashboard`](https://github.com/tektoncd/dashboard) /kind docs
non_test
getting started guides for project we need better getting started guide for the tekton projects they should be on the docs folder of each of those project like docs getting started … repositories to tackle β€” the needs enhancement kind docs
0
470,121
13,531,426,191
IssuesEvent
2020-09-15 21:37:29
kubernetes/kubeadm
https://api.github.com/repos/kubernetes/kubeadm
closed
Regression - etcd datadir permissions not set on etcd grow
area/etcd area/security kind/bug kind/regression priority/backlog
## Is this a BUG REPORT or FEATURE REQUEST? BUG REPORT ## Versions **kubeadm version** (use `kubeadm version`): 1.14+ **Environment**: N/A ## What happened? With the release of etcd 3.4.10, the datadir permissions now need to be 0700 or etcd won't start. There was an issue (#1308) where perms were set on join before starting the etcd container as a security control, overriding the default behavior of creating a non-existant directory mode 0755. However, in a [cleanup](https://github.com/kubernetes/kubernetes/commit/6bbed9fef00d9789abf041f3961c9543e7a4ff45#diff-0960dc0bb7c3e7113a0daa027856b8f9), that necessary os.mkdirall was removed. This was transparently ignored for several releases since etcd didn't complain, but with etcd-io/etcd#11798 (in 3.4.10), the new etcd cluster on the second node does not start. I'm pretty sure this will break anyone on k8s 1.14 or newer who upgrades to etcd 3.4.10 or newer without first fixing the /var/lib/etcd perms. ## What you expected to happen? /var/lib/etcd (or whatever the var is set to) should be set to 0700. :) ## How to reproduce it (as minimally and precisely as possible)? Join a second master node, then `ls -ld /var/lib/etcd` on the node. With an etcd 3.4.10 or newer runtime ## Anything else we need to know? It's worth explicitly noting that the first control plane node added works fine. It's just the second and subsequent nodes which were handled in a separate location in the code which exhibit the problem.
1.0
Regression - etcd datadir permissions not set on etcd grow - ## Is this a BUG REPORT or FEATURE REQUEST? BUG REPORT ## Versions **kubeadm version** (use `kubeadm version`): 1.14+ **Environment**: N/A ## What happened? With the release of etcd 3.4.10, the datadir permissions now need to be 0700 or etcd won't start. There was an issue (#1308) where perms were set on join before starting the etcd container as a security control, overriding the default behavior of creating a non-existant directory mode 0755. However, in a [cleanup](https://github.com/kubernetes/kubernetes/commit/6bbed9fef00d9789abf041f3961c9543e7a4ff45#diff-0960dc0bb7c3e7113a0daa027856b8f9), that necessary os.mkdirall was removed. This was transparently ignored for several releases since etcd didn't complain, but with etcd-io/etcd#11798 (in 3.4.10), the new etcd cluster on the second node does not start. I'm pretty sure this will break anyone on k8s 1.14 or newer who upgrades to etcd 3.4.10 or newer without first fixing the /var/lib/etcd perms. ## What you expected to happen? /var/lib/etcd (or whatever the var is set to) should be set to 0700. :) ## How to reproduce it (as minimally and precisely as possible)? Join a second master node, then `ls -ld /var/lib/etcd` on the node. With an etcd 3.4.10 or newer runtime ## Anything else we need to know? It's worth explicitly noting that the first control plane node added works fine. It's just the second and subsequent nodes which were handled in a separate location in the code which exhibit the problem.
non_test
regression etcd datadir permissions not set on etcd grow is this a bug report or feature request bug report versions kubeadm version use kubeadm version environment n a what happened with the release of etcd the datadir permissions now need to be or etcd won t start there was an issue where perms were set on join before starting the etcd container as a security control overriding the default behavior of creating a non existant directory mode however in a that necessary os mkdirall was removed this was transparently ignored for several releases since etcd didn t complain but with etcd io etcd in the new etcd cluster on the second node does not start i m pretty sure this will break anyone on or newer who upgrades to etcd or newer without first fixing the var lib etcd perms what you expected to happen var lib etcd or whatever the var is set to should be set to how to reproduce it as minimally and precisely as possible join a second master node then ls ld var lib etcd on the node with an etcd or newer runtime anything else we need to know it s worth explicitly noting that the first control plane node added works fine it s just the second and subsequent nodes which were handled in a separate location in the code which exhibit the problem
0
135,700
18,717,799,426
IssuesEvent
2021-11-03 08:13:06
shaimael/IdentityServer4
https://api.github.com/repos/shaimael/IdentityServer4
closed
CVE-2021-23337 (High) detected in lodash-1.0.2.tgz - autoclosed
security vulnerability
## CVE-2021-23337 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-1.0.2.tgz</b></p></summary> <p>A utility library delivering consistency, customization, performance, and extras.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-1.0.2.tgz">https://registry.npmjs.org/lodash/-/lodash-1.0.2.tgz</a></p> <p>Path to dependency file: IdentityServer4/samples/Clients/old/MvcImplicit/package.json</p> <p>Path to vulnerable library: IdentityServer4/samples/Clients/old/MvcImplicit/node_modules/lodash/package.json,IdentityServer4/samples/Clients/old/MvcImplicitJwtRequest/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - gulp-3.8.11.tgz (Root Library) - vinyl-fs-0.3.14.tgz - glob-watcher-0.0.6.tgz - gaze-0.5.2.tgz - globule-0.1.0.tgz - :x: **lodash-1.0.2.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/shaimael/IdentityServer4/commit/1a971ceee12750a348ada2520e1769e6c763fb5f">1a971ceee12750a348ada2520e1769e6c763fb5f</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Lodash versions prior to 4.17.21 are vulnerable to Command Injection via the template function. <p>Publish Date: 2021-02-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23337>CVE-2021-23337</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.2</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c">https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c</a></p> <p>Release Date: 2021-02-15</p> <p>Fix Resolution: lodash - 4.17.21</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"1.0.2","packageFilePaths":["/samples/Clients/old/MvcImplicit/package.json","/samples/Clients/old/MvcImplicitJwtRequest/package.json"],"isTransitiveDependency":true,"dependencyTree":"gulp:3.8.11;vinyl-fs:0.3.14;glob-watcher:0.0.6;gaze:0.5.2;globule:0.1.0;lodash:1.0.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"lodash - 4.17.21"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2021-23337","vulnerabilityDetails":"Lodash versions prior to 4.17.21 are vulnerable to Command Injection via the template function.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23337","cvss3Severity":"high","cvss3Score":"7.2","cvss3Metrics":{"A":"High","AC":"Low","PR":"High","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2021-23337 (High) detected in lodash-1.0.2.tgz - autoclosed - ## CVE-2021-23337 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-1.0.2.tgz</b></p></summary> <p>A utility library delivering consistency, customization, performance, and extras.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-1.0.2.tgz">https://registry.npmjs.org/lodash/-/lodash-1.0.2.tgz</a></p> <p>Path to dependency file: IdentityServer4/samples/Clients/old/MvcImplicit/package.json</p> <p>Path to vulnerable library: IdentityServer4/samples/Clients/old/MvcImplicit/node_modules/lodash/package.json,IdentityServer4/samples/Clients/old/MvcImplicitJwtRequest/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - gulp-3.8.11.tgz (Root Library) - vinyl-fs-0.3.14.tgz - glob-watcher-0.0.6.tgz - gaze-0.5.2.tgz - globule-0.1.0.tgz - :x: **lodash-1.0.2.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/shaimael/IdentityServer4/commit/1a971ceee12750a348ada2520e1769e6c763fb5f">1a971ceee12750a348ada2520e1769e6c763fb5f</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Lodash versions prior to 4.17.21 are vulnerable to Command Injection via the template function. <p>Publish Date: 2021-02-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23337>CVE-2021-23337</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.2</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c">https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c</a></p> <p>Release Date: 2021-02-15</p> <p>Fix Resolution: lodash - 4.17.21</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"1.0.2","packageFilePaths":["/samples/Clients/old/MvcImplicit/package.json","/samples/Clients/old/MvcImplicitJwtRequest/package.json"],"isTransitiveDependency":true,"dependencyTree":"gulp:3.8.11;vinyl-fs:0.3.14;glob-watcher:0.0.6;gaze:0.5.2;globule:0.1.0;lodash:1.0.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"lodash - 4.17.21"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2021-23337","vulnerabilityDetails":"Lodash versions prior to 4.17.21 are vulnerable to Command Injection via the template function.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23337","cvss3Severity":"high","cvss3Score":"7.2","cvss3Metrics":{"A":"High","AC":"Low","PR":"High","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_test
cve high detected in lodash tgz autoclosed cve high severity vulnerability vulnerable library lodash tgz a utility library delivering consistency customization performance and extras library home page a href path to dependency file samples clients old mvcimplicit package json path to vulnerable library samples clients old mvcimplicit node modules lodash package json samples clients old mvcimplicitjwtrequest node modules lodash package json dependency hierarchy gulp tgz root library vinyl fs tgz glob watcher tgz gaze tgz globule tgz x lodash tgz vulnerable library found in head commit a href found in base branch main vulnerability details lodash versions prior to are vulnerable to command injection via the template function publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution lodash isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree gulp vinyl fs glob watcher gaze globule lodash isminimumfixversionavailable true minimumfixversion lodash basebranches vulnerabilityidentifier cve vulnerabilitydetails lodash versions prior to are vulnerable to command injection via the template function vulnerabilityurl
0
135,135
12,675,804,826
IssuesEvent
2020-06-19 02:57:04
Azure/azure-cli
https://api.github.com/repos/Azure/azure-cli
closed
Clarification of --scopes parameter for `az ad sp create-for-rbac --skip-assignment ...`
Documentation Graph OKR3.4 Candidate
Could you explain the meaning of the `--scopes` parameter for the command `az ad sp create-for-rbac --skip-assignment ...`. The help says: > --scopes : Space-separated list of scopes the service principal's role assignment applies to. Defaults to the root of the current subscription. As I have no assignment due to `--skip-assignment`, there is no assignment yet... What does `--scopes` do? Does it limit the scope where a role can be assigned to the service principal? ``` az role assignment create --assignee $ASSIGNEE --scope $SCOPE --role $ROLE ``` Or does it just do nothing in the `--skip-assignment` case? I'd be glad if you could clarify this and add some more information to the documentation regarding `--scopes`. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: e99d25ff-03e2-e015-37f4-a29b59ec50d4 * Version Independent ID: 1a8e88ed-72c7-fbe7-b0ad-68f23b3946cd * Content: [az ad sp](https://docs.microsoft.com/en-us/cli/azure/ad/sp?view=azure-cli-latest#az-ad-sp-create-for-rbac) * Content Source: [src/azure-cli/azure/cli/command_modules/role/_help.py](https://github.com/Azure/azure-cli/blob/dev/src/azure-cli/azure/cli/command_modules/role/_help.py) * Service: **active-directory** * GitHub Login: @rloutlaw * Microsoft Alias: **routlaw**
1.0
Clarification of --scopes parameter for `az ad sp create-for-rbac --skip-assignment ...` - Could you explain the meaning of the `--scopes` parameter for the command `az ad sp create-for-rbac --skip-assignment ...`. The help says: > --scopes : Space-separated list of scopes the service principal's role assignment applies to. Defaults to the root of the current subscription. As I have no assignment due to `--skip-assignment`, there is no assignment yet... What does `--scopes` do? Does it limit the scope where a role can be assigned to the service principal? ``` az role assignment create --assignee $ASSIGNEE --scope $SCOPE --role $ROLE ``` Or does it just do nothing in the `--skip-assignment` case? I'd be glad if you could clarify this and add some more information to the documentation regarding `--scopes`. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: e99d25ff-03e2-e015-37f4-a29b59ec50d4 * Version Independent ID: 1a8e88ed-72c7-fbe7-b0ad-68f23b3946cd * Content: [az ad sp](https://docs.microsoft.com/en-us/cli/azure/ad/sp?view=azure-cli-latest#az-ad-sp-create-for-rbac) * Content Source: [src/azure-cli/azure/cli/command_modules/role/_help.py](https://github.com/Azure/azure-cli/blob/dev/src/azure-cli/azure/cli/command_modules/role/_help.py) * Service: **active-directory** * GitHub Login: @rloutlaw * Microsoft Alias: **routlaw**
non_test
clarification of scopes parameter for az ad sp create for rbac skip assignment could you explain the meaning of the scopes parameter for the command az ad sp create for rbac skip assignment the help says scopes space separated list of scopes the service principal s role assignment applies to defaults to the root of the current subscription as i have no assignment due to skip assignment there is no assignment yet what does scopes do does it limit the scope where a role can be assigned to the service principal az role assignment create assignee assignee scope scope role role or does it just do nothing in the skip assignment case i d be glad if you could clarify this and add some more information to the documentation regarding scopes document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service active directory github login rloutlaw microsoft alias routlaw
0
236,966
7,754,225,812
IssuesEvent
2018-05-31 05:34:53
metabase/metabase
https://api.github.com/repos/metabase/metabase
closed
Cannot Use Chinese?
Bug Priority/P2 i18n
I am trying to input Chinese as title of Dashboard, Questions or etc. But after it's saved, it shows me some question mark. ```???``` Multilingual are not supported yet? Here is the database I'm using for Metabase: * MySQL 5.7.17 * Default collation: utf8_unicode_ci I also tried to update the Dashboard's title directly in MySQL with Chinese, then it can show Chinese correctly in Metabase interface. But if I change it again in Metabase interface, I still get ```???```. I feel when Metabase interface send the characters to MySQL, it does not use UTF8?
1.0
Cannot Use Chinese? - I am trying to input Chinese as title of Dashboard, Questions or etc. But after it's saved, it shows me some question mark. ```???``` Multilingual are not supported yet? Here is the database I'm using for Metabase: * MySQL 5.7.17 * Default collation: utf8_unicode_ci I also tried to update the Dashboard's title directly in MySQL with Chinese, then it can show Chinese correctly in Metabase interface. But if I change it again in Metabase interface, I still get ```???```. I feel when Metabase interface send the characters to MySQL, it does not use UTF8?
non_test
cannot use chinese i am trying to input chinese as title of dashboard questions or etc but after it s saved it shows me some question mark multilingual are not supported yet here is the database i m using for metabase mysql default collation unicode ci i also tried to update the dashboard s title directly in mysql with chinese then it can show chinese correctly in metabase interface but if i change it again in metabase interface i still get i feel when metabase interface send the characters to mysql it does not use
0
187,853
14,433,173,798
IssuesEvent
2020-12-07 04:05:44
kalexmills/github-vet-tests-dec2020
https://api.github.com/repos/kalexmills/github-vet-tests-dec2020
closed
lightningnetwork/lnd: nursery_store_test.go; 3 LoC
fresh test tiny
Found a possible issue in [lightningnetwork/lnd](https://www.github.com/lightningnetwork/lnd) at [nursery_store_test.go](https://github.com/lightningnetwork/lnd/blob/125dbbf0daf913ff40bfe5f17ff5eefc2871399b/nursery_store_test.go#L133-L135) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call which takes a reference to htlcOutput at line 134 may start a goroutine [Click here to see the code in its original context.](https://github.com/lightningnetwork/lnd/blob/125dbbf0daf913ff40bfe5f17ff5eefc2871399b/nursery_store_test.go#L133-L135) <details> <summary>Click here to show the 3 line(s) of Go which triggered the analyzer.</summary> ```go for _, htlcOutput := range test.htlcOutputs { assertCribAtExpiryHeight(t, ns, &htlcOutput) } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 125dbbf0daf913ff40bfe5f17ff5eefc2871399b
1.0
lightningnetwork/lnd: nursery_store_test.go; 3 LoC - Found a possible issue in [lightningnetwork/lnd](https://www.github.com/lightningnetwork/lnd) at [nursery_store_test.go](https://github.com/lightningnetwork/lnd/blob/125dbbf0daf913ff40bfe5f17ff5eefc2871399b/nursery_store_test.go#L133-L135) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call which takes a reference to htlcOutput at line 134 may start a goroutine [Click here to see the code in its original context.](https://github.com/lightningnetwork/lnd/blob/125dbbf0daf913ff40bfe5f17ff5eefc2871399b/nursery_store_test.go#L133-L135) <details> <summary>Click here to show the 3 line(s) of Go which triggered the analyzer.</summary> ```go for _, htlcOutput := range test.htlcOutputs { assertCribAtExpiryHeight(t, ns, &htlcOutput) } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 125dbbf0daf913ff40bfe5f17ff5eefc2871399b
test
lightningnetwork lnd nursery store test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message function call which takes a reference to htlcoutput at line may start a goroutine click here to show the line s of go which triggered the analyzer go for htlcoutput range test htlcoutputs assertcribatexpiryheight t ns htlcoutput leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
1
377,970
26,277,489,616
IssuesEvent
2023-01-07 00:32:04
UnBArqDsw2022-2/2022.2_G4_IdotPet
https://api.github.com/repos/UnBArqDsw2022-2/2022.2_G4_IdotPet
closed
GoFs - Comportamentais
documentation
# DescriΓ§Γ£o: Essa issue tem como objetivo realizar a elaboraΓ§Γ£o dos GoFs Comportamentais. # Tarefas: - [ ] Definir quais sΓ£o os GoFs Comportamentais - [ ] Elaborar a base dos GoFs Comportamentais # CritΓ©rio de aceitaΓ§Γ£o: - [ ] GoFs Comportamentais definidos - [ ] GoFs Comportamentais elaborados
1.0
GoFs - Comportamentais - # DescriΓ§Γ£o: Essa issue tem como objetivo realizar a elaboraΓ§Γ£o dos GoFs Comportamentais. # Tarefas: - [ ] Definir quais sΓ£o os GoFs Comportamentais - [ ] Elaborar a base dos GoFs Comportamentais # CritΓ©rio de aceitaΓ§Γ£o: - [ ] GoFs Comportamentais definidos - [ ] GoFs Comportamentais elaborados
non_test
gofs comportamentais descriΓ§Γ£o essa issue tem como objetivo realizar a elaboraΓ§Γ£o dos gofs comportamentais tarefas definir quais sΓ£o os gofs comportamentais elaborar a base dos gofs comportamentais critΓ©rio de aceitaΓ§Γ£o gofs comportamentais definidos gofs comportamentais elaborados
0
149,588
11,906,232,264
IssuesEvent
2020-03-30 20:00:02
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachtest: psycopg failed
C-test-failure O-roachtest O-robot branch-master release-blocker
[(roachtest).psycopg failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=1776248&tab=buildLog) on [master@14094e3d5ea5f548dfada7bdb6e0e1158f53e168](https://github.com/cockroachdb/cockroach/commits/14094e3d5ea5f548dfada7bdb6e0e1158f53e168): ``` The test failed on branch=master, cloud=gce: test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/20200301-1776248/psycopg/run_1 orm_helpers.go:214,orm_helpers.go:144,psycopg.go:121,psycopg.go:134,test_runner.go:741: Tests run on Cockroach v20.1.0-beta.1-505-g14094e3 Tests run against psycopg 2_8_4 672 Total Tests Run 466 tests passed 206 tests failed 96 tests skipped 2 tests ignored 1 test passed unexpectedly 0 tests failed unexpectedly 0 tests expected failed but skipped 0 tests expected failed but not run --- --- PASS: tests.test_async.AsyncTests.test_flush_on_write - https://github.com/cockroachdb/cockroach/issues/44709 (unexpected) For a full summary look at the psycopg artifacts An updated blacklist (psycopgBlackList20_1) is available in the artifacts' psycopg log ``` <details><summary>More</summary><p> Artifacts: [/psycopg](https://teamcity.cockroachdb.com/viewLog.html?buildId=1776248&tab=artifacts#/psycopg) [See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Apsycopg.%2A&sort=title&restgroup=false&display=lastcommented+project) <sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
2.0
roachtest: psycopg failed - [(roachtest).psycopg failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=1776248&tab=buildLog) on [master@14094e3d5ea5f548dfada7bdb6e0e1158f53e168](https://github.com/cockroachdb/cockroach/commits/14094e3d5ea5f548dfada7bdb6e0e1158f53e168): ``` The test failed on branch=master, cloud=gce: test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/20200301-1776248/psycopg/run_1 orm_helpers.go:214,orm_helpers.go:144,psycopg.go:121,psycopg.go:134,test_runner.go:741: Tests run on Cockroach v20.1.0-beta.1-505-g14094e3 Tests run against psycopg 2_8_4 672 Total Tests Run 466 tests passed 206 tests failed 96 tests skipped 2 tests ignored 1 test passed unexpectedly 0 tests failed unexpectedly 0 tests expected failed but skipped 0 tests expected failed but not run --- --- PASS: tests.test_async.AsyncTests.test_flush_on_write - https://github.com/cockroachdb/cockroach/issues/44709 (unexpected) For a full summary look at the psycopg artifacts An updated blacklist (psycopgBlackList20_1) is available in the artifacts' psycopg log ``` <details><summary>More</summary><p> Artifacts: [/psycopg](https://teamcity.cockroachdb.com/viewLog.html?buildId=1776248&tab=artifacts#/psycopg) [See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Apsycopg.%2A&sort=title&restgroup=false&display=lastcommented+project) <sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
test
roachtest psycopg failed on the test failed on branch master cloud gce test artifacts and logs in home agent work go src github com cockroachdb cockroach artifacts psycopg run orm helpers go orm helpers go psycopg go psycopg go test runner go tests run on cockroach beta tests run against psycopg total tests run tests passed tests failed tests skipped tests ignored test passed unexpectedly tests failed unexpectedly tests expected failed but skipped tests expected failed but not run pass tests test async asynctests test flush on write unexpected for a full summary look at the psycopg artifacts an updated blacklist is available in the artifacts psycopg log more artifacts powered by
1
415,596
12,131,803,883
IssuesEvent
2020-04-23 05:49:54
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
closed
Chat feedback
Priority: Medium Status: Fixed Status: Reopen Week Task
![image](https://user-images.githubusercontent.com/3536496/77038078-30cf8880-69f6-11ea-848f-bd12fbe59fbb.png) - [x] Clicking on the election should take you to the web page for the election (like the 'vote' button does in the government page). The icon on the right is clickable but doesnt do anything - the icon on any notification should do the action that makes sense, in this case go to view the election in the webpage. - [x] For elections that add/modify/remove something, clicking on the election UILink or the notification button should show you that thing. Right now it shows a detail page of the election - this should never be shown (what's worse, the user can edit the election right in that ui) ![image](https://user-images.githubusercontent.com/3536496/77038236-9885d380-69f6-11ea-9a5a-a7b9380bdd3a.png) - [x] Clicking Active took me to a page where I can edit the Active demographic. It should open a read-only viewer of it. Same for eveyrthing else that can be clicked like this.
1.0
Chat feedback - ![image](https://user-images.githubusercontent.com/3536496/77038078-30cf8880-69f6-11ea-848f-bd12fbe59fbb.png) - [x] Clicking on the election should take you to the web page for the election (like the 'vote' button does in the government page). The icon on the right is clickable but doesnt do anything - the icon on any notification should do the action that makes sense, in this case go to view the election in the webpage. - [x] For elections that add/modify/remove something, clicking on the election UILink or the notification button should show you that thing. Right now it shows a detail page of the election - this should never be shown (what's worse, the user can edit the election right in that ui) ![image](https://user-images.githubusercontent.com/3536496/77038236-9885d380-69f6-11ea-9a5a-a7b9380bdd3a.png) - [x] Clicking Active took me to a page where I can edit the Active demographic. It should open a read-only viewer of it. Same for eveyrthing else that can be clicked like this.
non_test
chat feedback clicking on the election should take you to the web page for the election like the vote button does in the government page the icon on the right is clickable but doesnt do anything the icon on any notification should do the action that makes sense in this case go to view the election in the webpage for elections that add modify remove something clicking on the election uilink or the notification button should show you that thing right now it shows a detail page of the election this should never be shown what s worse the user can edit the election right in that ui clicking active took me to a page where i can edit the active demographic it should open a read only viewer of it same for eveyrthing else that can be clicked like this
0
520,808
15,094,096,325
IssuesEvent
2021-02-07 04:20:40
TravelMapping/DataProcessing
https://api.github.com/repos/TravelMapping/DataProcessing
opened
"editing" DUPLICATE_LABEL entries
datacheck enhancement low priority
Each **`Route`** already has a **`duplicate_labels`** set used for writing warnings to userlogs. Making these into a hash table is a simple modification in both C++ & Python. If a label is included in a file more than twice, we can look up the existing datacheck and modify its info. We could keep a list, for example, of... * bounding box for the HB * all line numbers in the .wpt file * first & last line number in the .wpt file * whatever might be most useful
1.0
"editing" DUPLICATE_LABEL entries - Each **`Route`** already has a **`duplicate_labels`** set used for writing warnings to userlogs. Making these into a hash table is a simple modification in both C++ & Python. If a label is included in a file more than twice, we can look up the existing datacheck and modify its info. We could keep a list, for example, of... * bounding box for the HB * all line numbers in the .wpt file * first & last line number in the .wpt file * whatever might be most useful
non_test
editing duplicate label entries each route already has a duplicate labels set used for writing warnings to userlogs making these into a hash table is a simple modification in both c python if a label is included in a file more than twice we can look up the existing datacheck and modify its info we could keep a list for example of bounding box for the hb all line numbers in the wpt file first last line number in the wpt file whatever might be most useful
0
184,797
14,289,965,901
IssuesEvent
2020-11-23 20:06:49
github-vet/rangeclosure-findings
https://api.github.com/repos/github-vet/rangeclosure-findings
closed
liambarkley/jenkins: jenkins_home/tools/org.jenkinsci.plugins.golang.GolangInstallation/go_1.8.3/src/os/exec/exec_test.go; 20 LoC
fresh small test
Found a possible issue in [liambarkley/jenkins](https://www.github.com/liambarkley/jenkins) at [jenkins_home/tools/org.jenkinsci.plugins.golang.GolangInstallation/go_1.8.3/src/os/exec/exec_test.go](https://github.com/liambarkley/jenkins/blob/fe753e9b5089c2c8f088b5549490c1f4c5a8e9a6/jenkins_home/tools/org.jenkinsci.plugins.golang.GolangInstallation/go_1.8.3/src/os/exec/exec_test.go#L788-L807) The below snippet of Go code triggered static analysis which searches for goroutines and/or defer statements which capture loop variables. [Click here to see the code in its original context.](https://github.com/liambarkley/jenkins/blob/fe753e9b5089c2c8f088b5549490c1f4c5a8e9a6/jenkins_home/tools/org.jenkinsci.plugins.golang.GolangInstallation/go_1.8.3/src/os/exec/exec_test.go#L788-L807) <details> <summary>Click here to show the 20 line(s) of Go which triggered the analyzer.</summary> ```go for i, r := range pipes { ch := make(chan string, 1) go func(c chan string) { buf := make([]byte, 10) n, err := r.Read(buf) if err != nil { fmt.Fprintf(os.Stderr, "Child: read error: %v on pipe %d\n", err, i) os.Exit(1) } c <- string(buf[:n]) close(c) }(ch) select { case m := <-ch: response = response + m case <-time.After(5 * time.Second): fmt.Fprintf(os.Stderr, "Child: Timeout reading from pipe: %d\n", i) os.Exit(1) } } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: fe753e9b5089c2c8f088b5549490c1f4c5a8e9a6
1.0
liambarkley/jenkins: jenkins_home/tools/org.jenkinsci.plugins.golang.GolangInstallation/go_1.8.3/src/os/exec/exec_test.go; 20 LoC - Found a possible issue in [liambarkley/jenkins](https://www.github.com/liambarkley/jenkins) at [jenkins_home/tools/org.jenkinsci.plugins.golang.GolangInstallation/go_1.8.3/src/os/exec/exec_test.go](https://github.com/liambarkley/jenkins/blob/fe753e9b5089c2c8f088b5549490c1f4c5a8e9a6/jenkins_home/tools/org.jenkinsci.plugins.golang.GolangInstallation/go_1.8.3/src/os/exec/exec_test.go#L788-L807) The below snippet of Go code triggered static analysis which searches for goroutines and/or defer statements which capture loop variables. [Click here to see the code in its original context.](https://github.com/liambarkley/jenkins/blob/fe753e9b5089c2c8f088b5549490c1f4c5a8e9a6/jenkins_home/tools/org.jenkinsci.plugins.golang.GolangInstallation/go_1.8.3/src/os/exec/exec_test.go#L788-L807) <details> <summary>Click here to show the 20 line(s) of Go which triggered the analyzer.</summary> ```go for i, r := range pipes { ch := make(chan string, 1) go func(c chan string) { buf := make([]byte, 10) n, err := r.Read(buf) if err != nil { fmt.Fprintf(os.Stderr, "Child: read error: %v on pipe %d\n", err, i) os.Exit(1) } c <- string(buf[:n]) close(c) }(ch) select { case m := <-ch: response = response + m case <-time.After(5 * time.Second): fmt.Fprintf(os.Stderr, "Child: Timeout reading from pipe: %d\n", i) os.Exit(1) } } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: fe753e9b5089c2c8f088b5549490c1f4c5a8e9a6
test
liambarkley jenkins jenkins home tools org jenkinsci plugins golang golanginstallation go src os exec exec test go loc found a possible issue in at the below snippet of go code triggered static analysis which searches for goroutines and or defer statements which capture loop variables click here to show the line s of go which triggered the analyzer go for i r range pipes ch make chan string go func c chan string buf make byte n err r read buf if err nil fmt fprintf os stderr child read error v on pipe d n err i os exit c string buf close c ch select case m ch response response m case time after time second fmt fprintf os stderr child timeout reading from pipe d n i os exit leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
1
73,232
9,654,059,079
IssuesEvent
2019-05-19 10:54:38
cornellius-gp/gpytorch
https://api.github.com/repos/cornellius-gp/gpytorch
opened
Relationship between your LCM in Multi-output GP and LMC/ICM/SLFM
documentation
Hi everyone, I am a bit confused the LCM. However, I am familiar this paper: Kernels for Vector-Valued Functions: a Review. https://arxiv.org/pdf/1106.6251.pdf I can't quite tell whether the LCM kernel is the LMC kernel in the paper above. Could you please tell me what is the relationship between LCM and LMC? Thanks in advance.
1.0
Relationship between your LCM in Multi-output GP and LMC/ICM/SLFM - Hi everyone, I am a bit confused the LCM. However, I am familiar this paper: Kernels for Vector-Valued Functions: a Review. https://arxiv.org/pdf/1106.6251.pdf I can't quite tell whether the LCM kernel is the LMC kernel in the paper above. Could you please tell me what is the relationship between LCM and LMC? Thanks in advance.
non_test
relationship between your lcm in multi output gp and lmc icm slfm hi everyone i am a bit confused the lcm however i am familiar this paper kernels for vector valued functions a review i can t quite tell whether the lcm kernel is the lmc kernel in the paper above could you please tell me what is the relationship between lcm and lmc thanks in advance
0
343,865
30,696,765,051
IssuesEvent
2023-07-26 19:15:13
jdi-testing/jdi-light
https://api.github.com/repos/jdi-testing/jdi-light
closed
Angular rework: Slider page is 'empty' on new site
🐞 bug Angular TestSite
On new testsite https://jdi-testing.github.io/jdi-light/angular-page/#/ page Slider is empty. ![image](https://github.com/jdi-testing/jdi-light/assets/55189259/cde79b51-8ddb-42dc-9cda-1d53feb72754)
1.0
Angular rework: Slider page is 'empty' on new site - On new testsite https://jdi-testing.github.io/jdi-light/angular-page/#/ page Slider is empty. ![image](https://github.com/jdi-testing/jdi-light/assets/55189259/cde79b51-8ddb-42dc-9cda-1d53feb72754)
test
angular rework slider page is empty on new site on new testsite page slider is empty
1
186,505
14,394,741,323
IssuesEvent
2020-12-03 02:01:16
github-vet/rangeclosure-findings
https://api.github.com/repos/github-vet/rangeclosure-findings
closed
wongma7/efs-provisioner: vendor/k8s.io/kubernetes/pkg/cloudprovider/providers/azure/azure_test.go; 46 LoC
fresh small test vendored
Found a possible issue in [wongma7/efs-provisioner](https://www.github.com/wongma7/efs-provisioner) at [vendor/k8s.io/kubernetes/pkg/cloudprovider/providers/azure/azure_test.go](https://github.com/wongma7/efs-provisioner/blob/30dd5311bb03da485386707c351579ff5d580722/vendor/k8s.io/kubernetes/pkg/cloudprovider/providers/azure/azure_test.go#L327-L372) The below snippet of Go code triggered static analysis which searches for goroutines and/or defer statements which capture loop variables. [Click here to see the code in its original context.](https://github.com/wongma7/efs-provisioner/blob/30dd5311bb03da485386707c351579ff5d580722/vendor/k8s.io/kubernetes/pkg/cloudprovider/providers/azure/azure_test.go#L327-L372) <details> <summary>Click here to show the 46 line(s) of Go which triggered the analyzer.</summary> ```go for _, svc := range services { for _, wantedRule := range svc.Spec.Ports { expectedRuleCount++ wantedRuleName := getRuleName(&svc, wantedRule) foundRule := false for _, actualRule := range *loadBalancer.LoadBalancingRules { if strings.EqualFold(*actualRule.Name, wantedRuleName) && *actualRule.FrontendPort == wantedRule.Port && *actualRule.BackendPort == wantedRule.Port { foundRule = true break } } if !foundRule { t.Errorf("Expected load balancer rule but didn't find it: %q", wantedRuleName) } foundProbe := false if serviceapi.NeedsHealthCheck(&svc) { path, port := serviceapi.GetServiceHealthCheckPathPort(&svc) for _, actualProbe := range *loadBalancer.Probes { if strings.EqualFold(*actualProbe.Name, wantedRuleName) && *actualProbe.Port == port && *actualProbe.RequestPath == path && actualProbe.Protocol == network.ProbeProtocolHTTP { foundProbe = true break } } } else { for _, actualProbe := range *loadBalancer.Probes { if strings.EqualFold(*actualProbe.Name, wantedRuleName) && *actualProbe.Port == wantedRule.NodePort { foundProbe = true break } } } if !foundProbe { for _, actualProbe := range *loadBalancer.Probes { t.Logf("Probe: %s %d", *actualProbe.Name, *actualProbe.Port) } t.Errorf("Expected loadbalancer probe but didn't find it: %q", wantedRuleName) } } } ``` Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call which takes a reference to svc at line 345 may start a goroutine </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 30dd5311bb03da485386707c351579ff5d580722
1.0
wongma7/efs-provisioner: vendor/k8s.io/kubernetes/pkg/cloudprovider/providers/azure/azure_test.go; 46 LoC - Found a possible issue in [wongma7/efs-provisioner](https://www.github.com/wongma7/efs-provisioner) at [vendor/k8s.io/kubernetes/pkg/cloudprovider/providers/azure/azure_test.go](https://github.com/wongma7/efs-provisioner/blob/30dd5311bb03da485386707c351579ff5d580722/vendor/k8s.io/kubernetes/pkg/cloudprovider/providers/azure/azure_test.go#L327-L372) The below snippet of Go code triggered static analysis which searches for goroutines and/or defer statements which capture loop variables. [Click here to see the code in its original context.](https://github.com/wongma7/efs-provisioner/blob/30dd5311bb03da485386707c351579ff5d580722/vendor/k8s.io/kubernetes/pkg/cloudprovider/providers/azure/azure_test.go#L327-L372) <details> <summary>Click here to show the 46 line(s) of Go which triggered the analyzer.</summary> ```go for _, svc := range services { for _, wantedRule := range svc.Spec.Ports { expectedRuleCount++ wantedRuleName := getRuleName(&svc, wantedRule) foundRule := false for _, actualRule := range *loadBalancer.LoadBalancingRules { if strings.EqualFold(*actualRule.Name, wantedRuleName) && *actualRule.FrontendPort == wantedRule.Port && *actualRule.BackendPort == wantedRule.Port { foundRule = true break } } if !foundRule { t.Errorf("Expected load balancer rule but didn't find it: %q", wantedRuleName) } foundProbe := false if serviceapi.NeedsHealthCheck(&svc) { path, port := serviceapi.GetServiceHealthCheckPathPort(&svc) for _, actualProbe := range *loadBalancer.Probes { if strings.EqualFold(*actualProbe.Name, wantedRuleName) && *actualProbe.Port == port && *actualProbe.RequestPath == path && actualProbe.Protocol == network.ProbeProtocolHTTP { foundProbe = true break } } } else { for _, actualProbe := range *loadBalancer.Probes { if strings.EqualFold(*actualProbe.Name, wantedRuleName) && *actualProbe.Port == wantedRule.NodePort { foundProbe = true break } } } if !foundProbe { for _, actualProbe := range *loadBalancer.Probes { t.Logf("Probe: %s %d", *actualProbe.Name, *actualProbe.Port) } t.Errorf("Expected loadbalancer probe but didn't find it: %q", wantedRuleName) } } } ``` Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call which takes a reference to svc at line 345 may start a goroutine </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 30dd5311bb03da485386707c351579ff5d580722
test
efs provisioner vendor io kubernetes pkg cloudprovider providers azure azure test go loc found a possible issue in at the below snippet of go code triggered static analysis which searches for goroutines and or defer statements which capture loop variables click here to show the line s of go which triggered the analyzer go for svc range services for wantedrule range svc spec ports expectedrulecount wantedrulename getrulename svc wantedrule foundrule false for actualrule range loadbalancer loadbalancingrules if strings equalfold actualrule name wantedrulename actualrule frontendport wantedrule port actualrule backendport wantedrule port foundrule true break if foundrule t errorf expected load balancer rule but didn t find it q wantedrulename foundprobe false if serviceapi needshealthcheck svc path port serviceapi getservicehealthcheckpathport svc for actualprobe range loadbalancer probes if strings equalfold actualprobe name wantedrulename actualprobe port port actualprobe requestpath path actualprobe protocol network probeprotocolhttp foundprobe true break else for actualprobe range loadbalancer probes if strings equalfold actualprobe name wantedrulename actualprobe port wantedrule nodeport foundprobe true break if foundprobe for actualprobe range loadbalancer probes t logf probe s d actualprobe name actualprobe port t errorf expected loadbalancer probe but didn t find it q wantedrulename below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message function call which takes a reference to svc at line may start a goroutine leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
1
293,052
25,264,355,887
IssuesEvent
2022-11-16 02:37:00
optuna/optuna
https://api.github.com/repos/optuna/optuna
closed
Gather stub objective functions
test
### Motivation This is part of #2974 The Optuna's tests have many stub objective functions. We can reduce coding cost to write tests by organizing them. ### Description By gathering the stub objective functions to `optuna.testing`, we can reuse prewritten objective functions and simplify the tests. ### Alternatives (optional) _No response_ ### Additional context (optional) _No response_
1.0
Gather stub objective functions - ### Motivation This is part of #2974 The Optuna's tests have many stub objective functions. We can reduce coding cost to write tests by organizing them. ### Description By gathering the stub objective functions to `optuna.testing`, we can reuse prewritten objective functions and simplify the tests. ### Alternatives (optional) _No response_ ### Additional context (optional) _No response_
test
gather stub objective functions motivation this is part of the optuna s tests have many stub objective functions we can reduce coding cost to write tests by organizing them description by gathering the stub objective functions to optuna testing we can reuse prewritten objective functions and simplify the tests alternatives optional no response additional context optional no response
1
765,148
26,835,122,468
IssuesEvent
2023-02-02 18:52:03
iterative/vscode-dvc
https://api.github.com/repos/iterative/vscode-dvc
closed
Test and release LSP
priority-p1 triage πŸ“¦ product A: lsp
- [ ] Regarding the possibility of the crashes coming from parsing the files, this can be done in the usual integration test in the server. An idea to do it is to send the same files a lot of times - [ ] General smoke tests must be added to make sure the server is loading properly with the extension.
1.0
Test and release LSP - - [ ] Regarding the possibility of the crashes coming from parsing the files, this can be done in the usual integration test in the server. An idea to do it is to send the same files a lot of times - [ ] General smoke tests must be added to make sure the server is loading properly with the extension.
non_test
test and release lsp regarding the possibility of the crashes coming from parsing the files this can be done in the usual integration test in the server an idea to do it is to send the same files a lot of times general smoke tests must be added to make sure the server is loading properly with the extension
0
137,903
11,167,241,432
IssuesEvent
2019-12-27 16:21:44
paritytech/polkassembly
https://api.github.com/repos/paritytech/polkassembly
closed
Add tests for mutations.
i-4-tests s-auth-server
Add tests for mutations: - [x] changeEmail.ts - [x] changeName.ts - [x] changePassword.ts - [x] login.ts - [x] logout.ts - [ ] requestResetPassword.ts - [x] resetPassword.ts - [x] signup.ts - [x] verifyEmail.ts
1.0
Add tests for mutations. - Add tests for mutations: - [x] changeEmail.ts - [x] changeName.ts - [x] changePassword.ts - [x] login.ts - [x] logout.ts - [ ] requestResetPassword.ts - [x] resetPassword.ts - [x] signup.ts - [x] verifyEmail.ts
test
add tests for mutations add tests for mutations changeemail ts changename ts changepassword ts login ts logout ts requestresetpassword ts resetpassword ts signup ts verifyemail ts
1
128,222
5,051,010,173
IssuesEvent
2016-12-20 20:31:15
aayaffe/SailingRaceCourseManager
https://api.github.com/repos/aayaffe/SailingRaceCourseManager
closed
When pressing an info window, with no GPS available the app crashes
Priority: Critical Type: Bug
Tries to set a line with one point missing (The own location)
1.0
When pressing an info window, with no GPS available the app crashes - Tries to set a line with one point missing (The own location)
non_test
when pressing an info window with no gps available the app crashes tries to set a line with one point missing the own location
0
239,944
19,982,209,364
IssuesEvent
2022-01-30 03:54:00
Airbenders-490/AirbendersFE
https://api.github.com/repos/Airbenders-490/AirbendersFE
opened
[STUD-167] As a student, I want to be able to retrieve messages for specific chat rooms so I can read my messages.
acceptance test RELEASE 2
**Trigger:** - The user chooses to view a list of the chat rooms they are currently a part of. **Acceptance Criteria:** - The user can read the messages for a chosen chat room. **Conditions of Success:** - The user can choose a chat room to open - The user can view the messages in the chat room they’ve opened **User Acceptance Flow:** - User is on the list of the chat rooms - User clicks on a chat room - User is able to view the conversation
1.0
[STUD-167] As a student, I want to be able to retrieve messages for specific chat rooms so I can read my messages. - **Trigger:** - The user chooses to view a list of the chat rooms they are currently a part of. **Acceptance Criteria:** - The user can read the messages for a chosen chat room. **Conditions of Success:** - The user can choose a chat room to open - The user can view the messages in the chat room they’ve opened **User Acceptance Flow:** - User is on the list of the chat rooms - User clicks on a chat room - User is able to view the conversation
test
as a student i want to be able to retrieve messages for specific chat rooms so i can read my messages trigger the user chooses to view a list of the chat rooms they are currently a part of acceptance criteria the user can read the messages for a chosen chat room conditions of success the user can choose a chat room to open the user can view the messages in the chat room they’ve opened user acceptance flow user is on the list of the chat rooms user clicks on a chat room user is able to view the conversation
1
142,743
21,878,922,386
IssuesEvent
2022-05-19 12:48:58
ballerina-platform/ballerina-spec
https://api.github.com/repos/ballerina-platform/ballerina-spec
opened
Supporting GraphQL Interfaces in Ballerina
design/implementability design/usability graphql
**Description:** We are to implement GraphQL interface support in Ballerina. As per the last discussion, we agreed to implement the GraphQL interfaces using the distinct service types. Ideally, we thought the following is the way to define a GraphQL interface in Ballerina. ```ballerina type Person distinct service object { resource function get name() returns string; }; distinct service class Student { *Person; resource function get name() returns string { return "Student"; } } ``` But since the [resource functions are not considered as a part of the service type](https://github.com/ballerina-platform/ballerina-spec/issues/669) yet, we are thinking about the following approach to define interfaces. ```ballerina distinct service class Person { resource function get name() returns string { return ""; } } distinct service class Student { resource function get name() returns string { return "Student"; } } ``` We can validate the interface implementation using a compiler plugin to check whether the implementations contain the same resource function signatures as the interface type. But the issue with this approach is that the interface type can have instances of its own, which violates the interface concept. @jclark WDYT? Related issues: #559 #669
2.0
Supporting GraphQL Interfaces in Ballerina - **Description:** We are to implement GraphQL interface support in Ballerina. As per the last discussion, we agreed to implement the GraphQL interfaces using the distinct service types. Ideally, we thought the following is the way to define a GraphQL interface in Ballerina. ```ballerina type Person distinct service object { resource function get name() returns string; }; distinct service class Student { *Person; resource function get name() returns string { return "Student"; } } ``` But since the [resource functions are not considered as a part of the service type](https://github.com/ballerina-platform/ballerina-spec/issues/669) yet, we are thinking about the following approach to define interfaces. ```ballerina distinct service class Person { resource function get name() returns string { return ""; } } distinct service class Student { resource function get name() returns string { return "Student"; } } ``` We can validate the interface implementation using a compiler plugin to check whether the implementations contain the same resource function signatures as the interface type. But the issue with this approach is that the interface type can have instances of its own, which violates the interface concept. @jclark WDYT? Related issues: #559 #669
non_test
supporting graphql interfaces in ballerina description we are to implement graphql interface support in ballerina as per the last discussion we agreed to implement the graphql interfaces using the distinct service types ideally we thought the following is the way to define a graphql interface in ballerina ballerina type person distinct service object resource function get name returns string distinct service class student person resource function get name returns string return student but since the yet we are thinking about the following approach to define interfaces ballerina distinct service class person resource function get name returns string return distinct service class student resource function get name returns string return student we can validate the interface implementation using a compiler plugin to check whether the implementations contain the same resource function signatures as the interface type but the issue with this approach is that the interface type can have instances of its own which violates the interface concept jclark wdyt related issues
0
260,399
22,617,316,829
IssuesEvent
2022-06-30 00:20:03
Uuvana-Studios/longvinter-windows-client
https://api.github.com/repos/Uuvana-Studios/longvinter-windows-client
opened
[US-WEST] Uuvana 1 is a buggy mess.
Bug Not Tested
**Describe the bug** All assets load in way too late, to the point that you could be walking and a house with a fence loads onto you and you become stuck. You can even harvest other's gated and protected plants because their fences don't load in for awhile. This happens even with a low ping, my ping is 19 and the issue persists. **To Reproduce** Steps to reproduce the behavior: 1. Go to server browser 2. Click on [US-WEST] Uuvana 1 3. Realize that the asset loading times are absurdly high, to the point that the game is not playable. 4. Leave **Expected behavior** Assets load into the world before you see them. **Screenshots** Upon initial arrival of this random base (I am able to enter and harvest): ![image](https://user-images.githubusercontent.com/108439168/176566743-87119411-7dfa-45c5-a13b-7b16d3155d16.png) After 1 minute of waiting without moving: ![image](https://user-images.githubusercontent.com/108439168/176566891-5e61b2e0-9312-4b1c-a2fc-b9ddd273cc33.png) **Desktop (please complete the following information):** - OS: Windows 10 - Game Version 1.07b - Steam Version (most recent of July 29th, 2022)
1.0
[US-WEST] Uuvana 1 is a buggy mess. - **Describe the bug** All assets load in way too late, to the point that you could be walking and a house with a fence loads onto you and you become stuck. You can even harvest other's gated and protected plants because their fences don't load in for awhile. This happens even with a low ping, my ping is 19 and the issue persists. **To Reproduce** Steps to reproduce the behavior: 1. Go to server browser 2. Click on [US-WEST] Uuvana 1 3. Realize that the asset loading times are absurdly high, to the point that the game is not playable. 4. Leave **Expected behavior** Assets load into the world before you see them. **Screenshots** Upon initial arrival of this random base (I am able to enter and harvest): ![image](https://user-images.githubusercontent.com/108439168/176566743-87119411-7dfa-45c5-a13b-7b16d3155d16.png) After 1 minute of waiting without moving: ![image](https://user-images.githubusercontent.com/108439168/176566891-5e61b2e0-9312-4b1c-a2fc-b9ddd273cc33.png) **Desktop (please complete the following information):** - OS: Windows 10 - Game Version 1.07b - Steam Version (most recent of July 29th, 2022)
test
uuvana is a buggy mess describe the bug all assets load in way too late to the point that you could be walking and a house with a fence loads onto you and you become stuck you can even harvest other s gated and protected plants because their fences don t load in for awhile this happens even with a low ping my ping is and the issue persists to reproduce steps to reproduce the behavior go to server browser click on uuvana realize that the asset loading times are absurdly high to the point that the game is not playable leave expected behavior assets load into the world before you see them screenshots upon initial arrival of this random base i am able to enter and harvest after minute of waiting without moving desktop please complete the following information os windows game version steam version most recent of july
1
170,588
13,194,027,360
IssuesEvent
2020-08-13 16:07:41
Thy-Vipe/BeastsOfBermuda-issues
https://api.github.com/repos/Thy-Vipe/BeastsOfBermuda-issues
closed
[Major Bug] Dark sphere around your creature have loop holes
Fixed! Gameplay Major Bug Potential fix tester-team
_Originally written by **Anatze | 76561198302304278**_ Game Version: 1.1.982 *===== System Specs ===== CPU Brand: Intel(R) Core(TM) i7-8750H CPU @ 2.20GHz Vendor: GenuineIntel GPU Brand: Intel(R) UHD Graphics 630 GPU Driver Info: Unknown Num CPU Cores: 6 ===================* Context: **all creatures in a burrow** Map: Rival_Shores When your creature stand in a burrow, you can change your resolution like "FOV 120" in console and then you can see outside of your black sphere
1.0
[Major Bug] Dark sphere around your creature have loop holes - _Originally written by **Anatze | 76561198302304278**_ Game Version: 1.1.982 *===== System Specs ===== CPU Brand: Intel(R) Core(TM) i7-8750H CPU @ 2.20GHz Vendor: GenuineIntel GPU Brand: Intel(R) UHD Graphics 630 GPU Driver Info: Unknown Num CPU Cores: 6 ===================* Context: **all creatures in a burrow** Map: Rival_Shores When your creature stand in a burrow, you can change your resolution like "FOV 120" in console and then you can see outside of your black sphere
test
dark sphere around your creature have loop holes originally written by anatze game version system specs cpu brand intel r core tm cpu vendor genuineintel gpu brand intel r uhd graphics gpu driver info unknown num cpu cores context all creatures in a burrow map rival shores when your creature stand in a burrow you can change your resolution like fov in console and then you can see outside of your black sphere
1
101,631
12,699,509,712
IssuesEvent
2020-06-22 14:57:42
prisma/studio
https://api.github.com/repos/prisma/studio
closed
Modal Editing behavior
kind/improvement process/needs-design topic: design
Modal editing behavior in the databrowser can lead to some unintuitive behavior (see #185) - Should we have modal editing behavior - What's the prior art here?
2.0
Modal Editing behavior - Modal editing behavior in the databrowser can lead to some unintuitive behavior (see #185) - Should we have modal editing behavior - What's the prior art here?
non_test
modal editing behavior modal editing behavior in the databrowser can lead to some unintuitive behavior see should we have modal editing behavior what s the prior art here
0
266,450
23,238,975,880
IssuesEvent
2022-08-03 14:07:17
apache/skywalking
https://api.github.com/repos/apache/skywalking
closed
[Feature] Remove the primary index in a block
feature test database
### Search before asking - [X] I had searched in the [issues](https://github.com/apache/skywalking/issues?q=is%3Aissue) and found no similar feature requirement. ### Description The primary index is an iteration helper to scan the series based on a time range. The scanner could use the main store which organizes the data over time. The underlying TSet should provide an iterator interface to support the scanning operation. ### Use case _No response_ ### Related issues _No response_ ### Are you willing to submit a PR? - [X] Yes I am willing to submit a PR! ### Code of Conduct - [X] I agree to follow this project's [Code of Conduct](https://www.apache.org/foundation/policies/conduct)
1.0
[Feature] Remove the primary index in a block - ### Search before asking - [X] I had searched in the [issues](https://github.com/apache/skywalking/issues?q=is%3Aissue) and found no similar feature requirement. ### Description The primary index is an iteration helper to scan the series based on a time range. The scanner could use the main store which organizes the data over time. The underlying TSet should provide an iterator interface to support the scanning operation. ### Use case _No response_ ### Related issues _No response_ ### Are you willing to submit a PR? - [X] Yes I am willing to submit a PR! ### Code of Conduct - [X] I agree to follow this project's [Code of Conduct](https://www.apache.org/foundation/policies/conduct)
test
remove the primary index in a block search before asking i had searched in the and found no similar feature requirement description the primary index is an iteration helper to scan the series based on a time range the scanner could use the main store which organizes the data over time the underlying tset should provide an iterator interface to support the scanning operation use case no response related issues no response are you willing to submit a pr yes i am willing to submit a pr code of conduct i agree to follow this project s
1
279,850
24,258,045,358
IssuesEvent
2022-09-27 19:38:46
OpenLiberty/openliberty.io
https://api.github.com/repos/OpenLiberty/openliberty.io
closed
Write Cypress tests for runtime release ordering
test
Test the ordering for runtime releases on the starter page using Cypress tests (1-2 tests) Ordering was changed in Issue #2573 , these changes require Cypress testing
1.0
Write Cypress tests for runtime release ordering - Test the ordering for runtime releases on the starter page using Cypress tests (1-2 tests) Ordering was changed in Issue #2573 , these changes require Cypress testing
test
write cypress tests for runtime release ordering test the ordering for runtime releases on the starter page using cypress tests tests ordering was changed in issue these changes require cypress testing
1
152,948
24,042,214,415
IssuesEvent
2022-09-16 03:44:25
woocommerce/pinterest-for-woocommerce
https://api.github.com/repos/woocommerce/pinterest-for-woocommerce
closed
Store stays connected to Pinterest account after disconnecting. Prevents connecting with a different Pinterest account
type: bug priority: low needs design
### Steps to reproduce: <!-- Describe the steps to reproduce the behavior.--> 1. Create a new store 2. Disconnect the store 3. connect it to another store ### Expected behavior: Should allow users to connect to different pinterest account ### Actual behavior: Prevents connecting with a different Pinterest account ### Additional details: <!--Any additional details you think might be helpful.--> <!--Ticket numbers/links, plugin versions, system statuses etc.--> ![1Cl8CMFtdVB_DXBkQBHyzFzyTlvYObS5438wgF0CA](https://user-images.githubusercontent.com/1510328/153259671-f3196ade-cb11-4eeb-9b1b-e25bdbd81c65.png)
1.0
Store stays connected to Pinterest account after disconnecting. Prevents connecting with a different Pinterest account - ### Steps to reproduce: <!-- Describe the steps to reproduce the behavior.--> 1. Create a new store 2. Disconnect the store 3. connect it to another store ### Expected behavior: Should allow users to connect to different pinterest account ### Actual behavior: Prevents connecting with a different Pinterest account ### Additional details: <!--Any additional details you think might be helpful.--> <!--Ticket numbers/links, plugin versions, system statuses etc.--> ![1Cl8CMFtdVB_DXBkQBHyzFzyTlvYObS5438wgF0CA](https://user-images.githubusercontent.com/1510328/153259671-f3196ade-cb11-4eeb-9b1b-e25bdbd81c65.png)
non_test
store stays connected to pinterest account after disconnecting prevents connecting with a different pinterest account steps to reproduce create a new store disconnect the store connect it to another store expected behavior should allow users to connect to different pinterest account actual behavior prevents connecting with a different pinterest account additional details
0
59,229
6,641,589,556
IssuesEvent
2017-09-27 02:10:58
pydata/xarray
https://api.github.com/repos/pydata/xarray
closed
Some tests still check pandas version
pandas-like testing
Although we updated pandas minimum version to 0.18 in #1530 , we still check its version like `pd.__version__ >= '0.17'` in some tests. (I forgot to remove these statements in #1530 .)
1.0
Some tests still check pandas version - Although we updated pandas minimum version to 0.18 in #1530 , we still check its version like `pd.__version__ >= '0.17'` in some tests. (I forgot to remove these statements in #1530 .)
test
some tests still check pandas version although we updated pandas minimum version to in we still check its version like pd version in some tests i forgot to remove these statements in
1
151,490
12,041,526,202
IssuesEvent
2020-04-14 08:59:34
WoWManiaUK/Redemption
https://api.github.com/repos/WoWManiaUK/Redemption
closed
[NPC] Longrunner Proudhoof does not respawn after killed by NPC in quest We Strike
Fix - Tester Confirmed
**Links:** https://www.wow-mania.com/armory/?npc=25335 **What is Happening:** If Proudhoof is killed during quest "We Strike!" then he does not respawn at the caravan at all (even with a GM-forced .respawn) ![image](https://user-images.githubusercontent.com/46089380/78717159-18c59780-7918-11ea-9a65-cf52e52e4ab5.png) **What Should happen:** According to armory, Proudhoof should have a 2 min respawn.
1.0
[NPC] Longrunner Proudhoof does not respawn after killed by NPC in quest We Strike - **Links:** https://www.wow-mania.com/armory/?npc=25335 **What is Happening:** If Proudhoof is killed during quest "We Strike!" then he does not respawn at the caravan at all (even with a GM-forced .respawn) ![image](https://user-images.githubusercontent.com/46089380/78717159-18c59780-7918-11ea-9a65-cf52e52e4ab5.png) **What Should happen:** According to armory, Proudhoof should have a 2 min respawn.
test
longrunner proudhoof does not respawn after killed by npc in quest we strike links what is happening if proudhoof is killed during quest we strike then he does not respawn at the caravan at all even with a gm forced respawn what should happen according to armory proudhoof should have a min respawn
1
204,357
15,439,172,014
IssuesEvent
2021-03-07 23:12:13
trevorNgo/Measure2.0
https://api.github.com/repos/trevorNgo/Measure2.0
opened
CS4ZP6 Tester Feedback: "FORCE CHECK-IN" button on the "Archive Year Term" page does not do anything
tester
**Description:** Clicking the **FORCE CHECK-IN** button under the **Archive Year Term** page when logged in as an **Admin** does not do anything. **OS:** Windows 10 Enterprise **Browser:** Chrome Version 89.0.4389.82 **Reproduction steps:** * Sign in as an **Admin**. * Click on the **Archive Year Term** button. * Click on the **FORCE CHECK-IN** button. **Expected result:** The **FORCE CHECK-IN** button should do something **Actual result:** Nothing happens when the button is clicked ![image](https://user-images.githubusercontent.com/32877366/110258417-8935c800-7f70-11eb-8558-2aa4deb9fe86.png)
1.0
CS4ZP6 Tester Feedback: "FORCE CHECK-IN" button on the "Archive Year Term" page does not do anything - **Description:** Clicking the **FORCE CHECK-IN** button under the **Archive Year Term** page when logged in as an **Admin** does not do anything. **OS:** Windows 10 Enterprise **Browser:** Chrome Version 89.0.4389.82 **Reproduction steps:** * Sign in as an **Admin**. * Click on the **Archive Year Term** button. * Click on the **FORCE CHECK-IN** button. **Expected result:** The **FORCE CHECK-IN** button should do something **Actual result:** Nothing happens when the button is clicked ![image](https://user-images.githubusercontent.com/32877366/110258417-8935c800-7f70-11eb-8558-2aa4deb9fe86.png)
test
tester feedback force check in button on the archive year term page does not do anything description clicking the force check in button under the archive year term page when logged in as an admin does not do anything os windows enterprise browser chrome version reproduction steps sign in as an admin click on the archive year term button click on the force check in button expected result the force check in button should do something actual result nothing happens when the button is clicked
1
288,214
24,890,404,465
IssuesEvent
2022-10-28 11:29:12
vishal-testgh20221021/testgh
https://api.github.com/repos/vishal-testgh20221021/testgh
opened
Test Failed - t1 - Step 2
testcollab ttc
h3. Test Case Details *Test case title*: t1 *Test plan title*: Test Plan 28 October 2022 *Steps*: |S.No.|Step|Expected Result|Status|Comment| |-|-|-|-|-| |1|s1|r1|Fail| | |*2*|s2 |r2|Fail| |
1.0
Test Failed - t1 - Step 2 - h3. Test Case Details *Test case title*: t1 *Test plan title*: Test Plan 28 October 2022 *Steps*: |S.No.|Step|Expected Result|Status|Comment| |-|-|-|-|-| |1|s1|r1|Fail| | |*2*|s2 |r2|Fail| |
test
test failed step test case details test case title test plan title test plan october steps s no step expected result status comment fail fail
1
23,243
4,004,253,011
IssuesEvent
2016-05-12 06:14:03
golang/go
https://api.github.com/repos/golang/go
opened
os/signal: TestReset flake
Testing
Darwin flake: https://build.golang.org/log/60153f56055e11452f86d300cd83adb10a2456ab ``` --- FAIL: TestReset (0.10s) signal_test.go:152: unexpected signal window size changes FAIL FAIL os/signal 4.121s ``` Paging @ianlancetaylor.
1.0
os/signal: TestReset flake - Darwin flake: https://build.golang.org/log/60153f56055e11452f86d300cd83adb10a2456ab ``` --- FAIL: TestReset (0.10s) signal_test.go:152: unexpected signal window size changes FAIL FAIL os/signal 4.121s ``` Paging @ianlancetaylor.
test
os signal testreset flake darwin flake fail testreset signal test go unexpected signal window size changes fail fail os signal paging ianlancetaylor
1
179,513
13,884,607,849
IssuesEvent
2020-10-18 16:49:51
nicorithner/rails_engine
https://api.github.com/repos/nicorithner/rails_engine
closed
Transactions model tests and table
Tests database
**transactions** id, invoice_id, credit_card_number, credit_card_expiration_date, result, created_at, updated_at
1.0
Transactions model tests and table - **transactions** id, invoice_id, credit_card_number, credit_card_expiration_date, result, created_at, updated_at
test
transactions model tests and table transactions id invoice id credit card number credit card expiration date result created at updated at
1
196,255
14,851,526,755
IssuesEvent
2021-01-18 07:04:22
thoth-station/thoth-application
https://api.github.com/repos/thoth-station/thoth-application
closed
User system:serviceaccount:thoth-test-core:management-api cannot create resource jobs in API group batch in the namespace thoth-test-core
bug deployment_name/moc deployment_name/ocp-stage deployment_name/ocp-test human_intervention_required priority/important-soon sig/devops
**Describe the bug** User system:serviceaccount:thoth-test-core:management-api cannot create resource jobs in API group batch in the namespace thoth-test-core **To Reproduce** Steps to reproduce the behavior: 1. Use `/graph/initialize` in management-API in test envioronment **Expected behavior** A clear and concise description of what you expected to happen. **Screenshots** ``` {"name": "thoth.management_api.openapi_server", "levelname": "ERROR", "module": "app", "lineno": 1891, "funcname": "log_exception", "created": 1610642726.0111866, "asctime": "2021-01-14 16:45:26,011", "msecs": 11.186599731445312, "relative_created": 58228.846073150635, "process": 23, "message": "Exception on /api/v1/graph/initialize [PUT]\nTraceback (most recent call last):\n File \"/opt/app-root/lib64/python3.8/site-packages/openshift/dynamic/client.py\", line 42, in inner\n resp = func(self, *args, **kwargs)\n File \"/opt/app-root/lib64/python3.8/site-packages/openshift/dynamic/client.py\", line 235, in request\n return self.client.call_api(\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py\", line 340, in call_api\n return self.__call_api(resource_path, method,\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py\", line 172, in __call_api\n response_data = self.request(\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py\", line 382, in request\n return self.rest_client.POST(url,\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/rest.py\", line 272, in POST\n return self.request(\"POST\", url,\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/rest.py\", line 231, in request\n raise ApiException(http_resp=r)\nkubernetes.client.rest.ApiException: (403)\nReason: Forbidden\nHTTP response headers: HTTPHeaderDict({'Audit-Id': '49583870-4187-49f8-9539-aef793ac349e', 'Cache-Control': 'no-cache, private', 'Content-Type': 'application/json', 'X-Content-Type-Options': 'nosniff', 'Date': 'Thu, 14 Jan 2021 16:45:25 GMT', 'Content-Length': '334'})\nHTTP response body: b'{\"kind\":\"Status\",\"apiVersion\":\"v1\",\"metadata\":{},\"status\":\"Failure\",\"message\":\"jobs.batch is forbidden: User \\\\\"system:serviceaccount:thoth-test-core:management-api\\\\\" cannot create resource \\\\\"jobs\\\\\" in API group \\\\\"batch\\\\\" in the namespace \\\\\"thoth-test-core\\\\\"\",\"reason\":\"Forbidden\",\"details\":{\"group\":\"batch\",\"kind\":\"jobs\"},\"code\":403}\\n'\n\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File \"/opt/app-root/lib64/python3.8/site-packages/flask/app.py\", line 2447, in wsgi_app\n response = self.full_dispatch_request()\n File \"/opt/app-root/lib64/python3.8/site-packages/flask/app.py\", line 1952, in full_dispatch_request\n rv = self.handle_user_exception(e)\n File \"/opt/app-root/lib64/python3.8/site-packages/flask/app.py\", line 1821, in handle_user_exception\n reraise(exc_type, exc_value, tb)\n File \"/opt/app-root/lib64/python3.8/site-packages/flask/_compat.py\", line 39, in reraise\n raise value\n File \"/opt/app-root/lib64/python3.8/site-packages/flask/app.py\", line 1950, in full_dispatch_request\n rv = self.dispatch_request()\n File \"/opt/app-root/lib64/python3.8/site-packages/flask/app.py\", line 1936, in dispatch_request\n return self.view_functions[rule.endpoint](**req.view_args)\n File \"/opt/app-root/lib64/python3.8/site-packages/connexion/decorators/decorator.py\", line 48, in wrapper\n response = function(request)\n File \"/opt/app-root/lib64/python3.8/site-packages/connexion/decorators/uri_parsing.py\", line 144, in wrapper\n response = function(request)\n File \"/opt/app-root/lib64/python3.8/site-packages/connexion/decorators/validation.py\", line 384, in wrapper\n return function(request)\n File \"/opt/app-root/lib64/python3.8/site-packages/connexion/decorators/parameter.py\", line 121, in wrapper\n return function(**kwargs)\n File \"/opt/app-root/src/thoth/management_api/api_v1.py\", line 348, in initialize_schema\n job_id = _OPENSHIFT.schedule_graph_schema_update()\n File \"/opt/app-root/lib64/python3.8/site-packages/thoth/common/openshift.py\", line 791, in schedule_graph_schema_update\n response = self.ocp_client.resources.get(\n File \"/opt/app-root/lib64/python3.8/site-packages/openshift/dynamic/client.py\", line 101, in create\n return self.request('post', path, body=body, **kwargs)\n File \"/opt/app-root/lib64/python3.8/site-packages/openshift/dynamic/client.py\", line 44, in inner\n raise api_exception(e)\nopenshift.dynamic.exceptions.ForbiddenError: 403\nReason: Forbidden\nHTTP response headers: HTTPHeaderDict({'Audit-Id': '49583870-4187-49f8-9539-aef793ac349e', 'Cache-Control': 'no-cache, private', 'Content-Type': 'application/json', 'X-Content-Type-Options': 'nosniff', 'Date': 'Thu, 14 Jan 2021 16:45:25 GMT', 'Content-Length': '334'})\nHTTP response body: b'{\"kind\":\"Status\",\"apiVersion\":\"v1\",\"metadata\":{},\"status\":\"Failure\",\"message\":\"jobs.batch is forbidden: User \\\\\"system:serviceaccount:thoth-test-core:management-api\\\\\" cannot create resource \\\\\"jobs\\\\\" in API group \\\\\"batch\\\\\" in the namespace \\\\\"thoth-test-core\\\\\"\",\"reason\":\"Forbidden\",\"details\":{\"group\":\"batch\",\"kind\":\"jobs\"},\"code\":403}\\n'\nOriginal traceback: \n File \"/opt/app-root/lib64/python3.8/site-packages/openshift/dynamic/client.py\", line 42, in inner\n resp = func(self, *args, **kwargs)\n\n File \"/opt/app-root/lib64/python3.8/site-packages/openshift/dynamic/client.py\", line 235, in request\n return self.client.call_api(\n\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py\", line 340, in call_api\n return self.__call_api(resource_path, method,\n\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py\", line 172, in __call_api\n response_data = self.request(\n\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py\", line 382, in request\n return self.rest_client.POST(url,\n\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/rest.py\", line 272, in POST\n return self.request(\"POST\", url,\n\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/rest.py\", line 231, in request\n raise ApiException(http_resp=r)\n"} {"name": "sentry_sdk.errors", "levelname": "INFO", "module": "scope", "lineno": 345, "funcname": "_drop", "created": 1610642726.0191312, "asctime": "2021-01-14 16:45:26,019", "msecs": 19.131183624267578, "relative_created": 58236.79065704346, "process": 23, "message": "event processor (<function DedupeIntegration.setup_once.<locals>.processor at 0x7f4e8280b3a0>) dropped event ({'level': 'error', 'exception': {'values': [{'module': 'kubernetes.client.rest', 'type': 'ApiException', 'value': '(403)\\nReason: Forbidden\\nHTTP response headers: HTTPHeaderDict({\\'Audit-Id\\': \\'49583870-4187-49f8-9539-aef793ac349e\\', \\'Cache-Control\\': \\'no-cache, private\\', \\'Content-Type\\': \\'application/json\\', \\'X-Content-Type-Options\\': \\'nosniff\\', \\'Date\\': \\'Thu, 14 Jan 2021 16:45:25 GMT\\', \\'Content-Length\\': \\'334\\'})\\nHTTP response body: b\\'{\"kind\":\"Status\",\"apiVersion\":\"v1\",\"metadata\":{},\"status\":\"Failure\",\"message\":\"jobs.batch is forbidden: User \\\\\\\\\"system:serviceaccount:thoth-test-core:management-api\\\\\\\\\" cannot create resource \\\\\\\\\"jobs\\\\\\\\\" in API group \\\\\\\\\"batch\\\\\\\\\" in the namespace \\\\\\\\\"thoth-test-core\\\\\\\\\"\",\"reason\":\"Forbidden\",\"details\":{\"group\":\"batch\",\"kind\":\"jobs\"},\"code\":403}\\\\n\\'\\n', 'mechanism': {'type': 'logging', 'handled': True}, 'stacktrace': {'frames': [{'filename': 'openshift/dynamic/client.py', 'abs_path': '/opt/app-root/lib64/python3.8/site-packages/openshift/dynamic/client.py', 'function': 'inner', 'module': 'openshift.dynamic.client', 'lineno': 42, 'pre_context': [' \"\"\" Handles parsing response structure and translating API Exceptions \"\"\"', ' def inner(self, *args, **kwargs):', \" serialize_response = kwargs.pop('serialize', True)\", \" serializer = kwargs.pop('serializer', ResourceInstance)\", ' try:'], 'context_line': ' resp = func(self, *args, **kwargs)', 'post_context': [' except ApiException as e:', ' raise api_exception(e)', ' if serialize_response:', ' try:', ' if six.PY2:'], 'vars': {'self': <openshift.dynamic.client.DynamicClient object at 0x7f4e826c1c70>, 'args': ('post', '/apis/batch/v1/namespaces/thoth-test-core/jobs'), 'kwargs': {'body': {'apiVersion': 'batch/v1', 'kind': 'Job', 'metadata': {'labels': {'app': 'thoth', 'component': 'graph-schema-update-job', 'mark': 'cleanup'}, 'name': 'graph-update-schema-7f56a52c400a58e992985a33c2bce4d5'}, 'spec': {'backoffLimit': 7, 'template': {'metadata': {'labels': {'app': 'thoth', 'component': 'graph-schema-update-job', 'mark': 'cleanup'}}, 'spec': {'containers': [{'env': [{'name': 'THOTH_WORKFLOW_TASK', 'value': 'graph_schema_update'}, {'name': 'THOTH_DEPLOYMENT_NAME', 'valueFrom': {'configMapKeyRef': {'key': 'deployment-name', 'name': 'thoth'}}}, {'name': 'KNOWLEDGE_GRAPH_HOST', 'valueFrom': {'configMapKeyRef': {'key': 'postgresql-host', 'name': 'thoth'}}}, {'name': 'KNOWLEDGE_GRAPH_PORT', 'value': '5432'}, {'name': 'KNOWLEDGE_GRAPH_SSL_DISABLED', 'value': '1'}, {'name': 'KNOWLEDGE_GRAPH_USER', 'valueFrom': {'secretKeyRef': {'key': 'database-user', 'name': 'postgresql'}}}, {'name': 'KNOWLEDGE_GRAPH_PASSWORD', 'valueFrom': {'secretKeyRef': {'key': 'database-password', 'name': 'postgresql'}}}, {'name': 'KNOWLEDGE_GRAPH_DATABASE', 'valueFrom': {'secretKeyRef': {'key': 'database-name', 'name': 'postgresql'}}}, {'name': 'THOTH_CEPH_BUCKET', 'valueFrom': {'configMapKeyRef': {'key': 'bucket-name', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_BUCKET_PREFIX', 'valueFrom': {'configMapKeyRef': {'key': 'bucket-prefix', 'name': 'ceph'}}}, {'name': 'THOTH_S3_ENDPOINT_URL', 'valueFrom': {'configMapKeyRef': {'key': 'host', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_KEY_ID', 'valueFrom': {'secretKeyRef': {'key': 'key-id', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_SECRET_KEY', 'valueFrom': {'secretKeyRef': {'key': 'secret-key', 'name': 'ceph'}}}], 'image': 'workflow-helpers', 'name': 'graph-schema-update-job', 'resources': {'limits': {'cpu': '1', 'memory': '512Mi'}, 'requests': {'cpu': '1', 'memory': '256Mi'}}}], 'restartPolicy': 'OnFailure'}}}}}, 'serialize_response': True, 'serializer': <class 'openshift.dynamic.resource.ResourceInstance'>, 'func': <function DynamicClient.request at 0x7f4e86f494c0>}}, {'filename': 'openshift/dynamic/client.py', 'abs_path': '/opt/app-root/lib64/python3.8/site-packages/openshift/dynamic/client.py', 'function': 'request', 'module': 'openshift.dynamic.client', 'lineno': 235, 'pre_context': [\" header_params['Content-Type'] = self.client.select_header_content_type(['*/*'])\", '', ' # Authentication setting', \" auth_settings = ['BearerToken']\", ''], 'context_line': ' return self.client.call_api(', 'post_context': [' path,', ' method.upper(),', ' path_params,', ' query_params,', ' header_params,'], 'vars': {'self': <openshift.dynamic.client.DynamicClient object at 0x7f4e826c1c70>, 'method': 'post', 'path': '/apis/batch/v1/namespaces/thoth-test-core/jobs', 'body': {'apiVersion': 'batch/v1', 'kind': 'Job', 'metadata': {'labels': {'app': 'thoth', 'component': 'graph-schema-update-job', 'mark': 'cleanup'}, 'name': 'graph-update-schema-7f56a52c400a58e992985a33c2bce4d5'}, 'spec': {'backoffLimit': 7, 'template': {'metadata': {'labels': {'app': 'thoth', 'component': 'graph-schema-update-job', 'mark': 'cleanup'}}, 'spec': {'containers': [{'env': [{'name': 'THOTH_WORKFLOW_TASK', 'value': 'graph_schema_update'}, {'name': 'THOTH_DEPLOYMENT_NAME', 'valueFrom': {'configMapKeyRef': {'key': 'deployment-name', 'name': 'thoth'}}}, {'name': 'KNOWLEDGE_GRAPH_HOST', 'valueFrom': {'configMapKeyRef': {'key': 'postgresql-host', 'name': 'thoth'}}}, {'name': 'KNOWLEDGE_GRAPH_PORT', 'value': '5432'}, {'name': 'KNOWLEDGE_GRAPH_SSL_DISABLED', 'value': '1'}, {'name': 'KNOWLEDGE_GRAPH_USER', 'valueFrom': {'secretKeyRef': {'key': 'database-user', 'name': 'postgresql'}}}, {'name': 'KNOWLEDGE_GRAPH_PASSWORD', 'valueFrom': {'secretKeyRef': {'key': 'database-password', 'name': 'postgresql'}}}, {'name': 'KNOWLEDGE_GRAPH_DATABASE', 'valueFrom': {'secretKeyRef': {'key': 'database-name', 'name': 'postgresql'}}}, {'name': 'THOTH_CEPH_BUCKET', 'valueFrom': {'configMapKeyRef': {'key': 'bucket-name', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_BUCKET_PREFIX', 'valueFrom': {'configMapKeyRef': {'key': 'bucket-prefix', 'name': 'ceph'}}}, {'name': 'THOTH_S3_ENDPOINT_URL', 'valueFrom': {'configMapKeyRef': {'key': 'host', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_KEY_ID', 'valueFrom': {'secretKeyRef': {'key': 'key-id', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_SECRET_KEY', 'valueFrom': {'secretKeyRef': {'key': 'secret-key', 'name': 'ceph'}}}], 'image': 'workflow-helpers', 'name': 'graph-schema-update-job', 'resources': {'limits': {'cpu': '1', 'memory': '512Mi'}, 'requests': {'cpu': '1', 'memory': '256Mi'}}}], 'restartPolicy': 'OnFailure'}}}}, 'params': {}, 'path_params': {}, 'query_params': [], 'header_params': {'Accept': 'application/json', 'Content-Type': 'application/json', 'User-Agent': 'OpenAPI-Generator/11.0.0/python'}, 'form_params': [], 'local_var_files': {}, 'auth_settings': ['BearerToken']}}, {'filename': 'kubernetes/client/api_client.py', 'abs_path': '/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py', 'function': 'call_api', 'module': 'kubernetes.client.api_client', 'lineno': 340, 'pre_context': [' The method will return the request thread.', ' If parameter async_req is False or missing,', ' then the method will return the response directly.', ' \"\"\"', ' if not async_req:'], 'context_line': ' return self.__call_api(resource_path, method,', 'post_context': [' path_params, query_params, header_params,', ' body, post_params, files,', ' response_type, auth_settings,', ' _return_http_data_only, collection_formats,', ' _preload_content, _request_timeout)'], 'vars': {'self': <kubernetes.client.api_client.ApiClient object at 0x7f4e826c1730>, 'resource_path': '/apis/batch/v1/namespaces/thoth-test-core/jobs', 'method': 'POST', 'path_params': {}, 'query_params': [], 'header_params': {'Accept': 'application/json', 'Content-Type': 'application/json', 'User-Agent': 'OpenAPI-Generator/11.0.0/python'}, 'body': {'apiVersion': 'batch/v1', 'kind': 'Job', 'metadata': {'labels': {'app': 'thoth', 'component': 'graph-schema-update-job', 'mark': 'cleanup'}, 'name': 'graph-update-schema-7f56a52c400a58e992985a33c2bce4d5'}, 'spec': {'backoffLimit': 7, 'template': {'metadata': {'labels': {'app': 'thoth', 'component': 'graph-schema-update-job', 'mark': 'cleanup'}}, 'spec': {'containers': [{'env': [{'name': 'THOTH_WORKFLOW_TASK', 'value': 'graph_schema_update'}, {'name': 'THOTH_DEPLOYMENT_NAME', 'valueFrom': {'configMapKeyRef': {'key': 'deployment-name', 'name': 'thoth'}}}, {'name': 'KNOWLEDGE_GRAPH_HOST', 'valueFrom': {'configMapKeyRef': {'key': 'postgresql-host', 'name': 'thoth'}}}, {'name': 'KNOWLEDGE_GRAPH_PORT', 'value': '5432'}, {'name': 'KNOWLEDGE_GRAPH_SSL_DISABLED', 'value': '1'}, {'name': 'KNOWLEDGE_GRAPH_USER', 'valueFrom': {'secretKeyRef': {'key': 'database-user', 'name': 'postgresql'}}}, {'name': 'KNOWLEDGE_GRAPH_PASSWORD', 'valueFrom': {'secretKeyRef': {'key': 'database-password', 'name': 'postgresql'}}}, {'name': 'KNOWLEDGE_GRAPH_DATABASE', 'valueFrom': {'secretKeyRef': {'key': 'database-name', 'name': 'postgresql'}}}, {'name': 'THOTH_CEPH_BUCKET', 'valueFrom': {'configMapKeyRef': {'key': 'bucket-name', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_BUCKET_PREFIX', 'valueFrom': {'configMapKeyRef': {'key': 'bucket-prefix', 'name': 'ceph'}}}, {'name': 'THOTH_S3_ENDPOINT_URL', 'valueFrom': {'configMapKeyRef': {'key': 'host', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_KEY_ID', 'valueFrom': {'secretKeyRef': {'key': 'key-id', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_SECRET_KEY', 'valueFrom': {'secretKeyRef': {'key': 'secret-key', 'name': 'ceph'}}}], 'image': 'workflow-helpers', 'name': 'graph-schema-update-job', 'resources': {'limits': {'cpu': '1', 'memory': '512Mi'}, 'requests': {'cpu': '1', 'memory': '256Mi'}}}], 'restartPolicy': 'OnFailure'}}}}, 'post_params': [], 'files': {}, 'response_type': None, 'auth_settings': ['BearerToken'], 'async_req': None, '_return_http_data_only': True, 'collection_formats': None, '_preload_content': False, '_request_timeout': None}}, {'filename': 'kubernetes/client/api_client.py', 'abs_path': '/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py', 'function': '__call_api', 'module': 'kubernetes.client.api_client', 'lineno': 172, 'pre_context': ['', ' # request url', ' url = self.configuration.host + resource_path', '', ' # perform request and return response'], 'context_line': ' response_data = self.request(', 'post_context': [' method, url, query_params=query_params, headers=header_params,', ' post_params=post_params, body=body,', ' _preload_content=_preload_content,', ' _request_timeout=_request_timeout)', ''], 'vars': {'self': <kubernetes.client.api_client.ApiClient object at 0x7f4e826c1730>, 'resource_path': '/apis/batch/v1/namespaces/thoth-test-core/jobs', 'method': 'POST', 'path_params': {}, 'query_params': [], 'header_params': {'Accept': 'application/json', 'Content-Type': 'application/json', 'User-Agent': 'OpenAPI-Generator/11.0.0/python', 'authorization': 'bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IlhaVlc3YnFrZ1BaSllHSkpoNnVLenpTQlE1X3lrcEJzVndHWUdJclVDVEUifQ.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJ0aG90aC10ZXN0LWNvcmUiLCJrdWJlcm5ldGVzLmlvL3NlcnZpY2VhY2NvdW50L3NlY3JldC5uYW1lIjoibWFuYWdlbWVudC1hcGktdG9rZW4tNXRqcDkiLCJrdWJlcm5ldGVzLmlvL3NlcnZpY2VhY2NvdW50L3NlcnZpY2UtYWNjb3VudC5uYW1lIjoibWFuYWdlbWVudC1hcGkiLCJrdWJlcm5ldGVzLmlvL3NlcnZpY2VhY2NvdW50L3NlcnZpY2UtYWNjb3VudC51aWQiOiIyMTZlODJlZi05N2RhLTRiNGYtODk4Mi02ZDQzOWEwNjRmNjIiLCJzdWIiOiJzeXN0ZW06c2VydmljZWFjY291bnQ6dGhvdGgtdGVzdC1jb3JlOm1hbmFnZW1lbnQtYXBpIn0.KLDhPJFSnfyXXY5tV5788iPFbYUQ48TXZW2GEi6N-8QZ8qVkcXmubhlvlO8_P93Me9-CS8iAgsUtRH8RYBUOLssZ_R6ioZaOH3Qe7TQhMwbgu6cqFp0wYMRJnvY-RrXHLZBRjUrstgKmEeE6UsclRxy88TRGpp2XYjlTSgW-XHYUj_EMqTePkQXrQZ8MM7EwQiAvwRv653dq3mSmNt4iwF5ewN7C6lPm1mUpw2Kg-kn1ojcP_G1d9gr1u41FUnPD_BDyWfa0Y2fS1fmR33QL7whdmvC26t6qANbHyMlz-jzorH9gfvgdTzhEjXE9-FHu54zj63y6GJM1kJy0vGGKfw'}, 'body': {'apiVersion': 'batch/v1', 'kind': 'Job', 'metadata': {'labels': {'app': 'thoth', 'component': 'graph-schema-update-job', 'mark': 'cleanup'}, 'name': 'graph-update-schema-7f56a52c400a58e992985a33c2bce4d5'}, 'spec': {'backoffLimit': 7, 'template': {'metadata': {'labels': {'app': 'thoth', 'component': 'graph-schema-update-job', 'mark': 'cleanup'}}, 'spec': {'containers': [{'env': [{'name': 'THOTH_WORKFLOW_TASK', 'value': 'graph_schema_update'}, {'name': 'THOTH_DEPLOYMENT_NAME', 'valueFrom': {'configMapKeyRef': {'key': 'deployment-name', 'name': 'thoth'}}}, {'name': 'KNOWLEDGE_GRAPH_HOST', 'valueFrom': {'configMapKeyRef': {'key': 'postgresql-host', 'name': 'thoth'}}}, {'name': 'KNOWLEDGE_GRAPH_PORT', 'value': '5432'}, {'name': 'KNOWLEDGE_GRAPH_SSL_DISABLED', 'value': '1'}, {'name': 'KNOWLEDGE_GRAPH_USER', 'valueFrom': {'secretKeyRef': {'key': 'database-user', 'name': 'postgresql'}}}, {'name': 'KNOWLEDGE_GRAPH_PASSWORD', 'valueFrom': {'secretKeyRef': {'key': 'database-password', 'name': 'postgresql'}}}, {'name': 'KNOWLEDGE_GRAPH_DATABASE', 'valueFrom': {'secretKeyRef': {'key': 'database-name', 'name': 'postgresql'}}}, {'name': 'THOTH_CEPH_BUCKET', 'valueFrom': {'configMapKeyRef': {'key': 'bucket-name', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_BUCKET_PREFIX', 'valueFrom': {'configMapKeyRef': {'key': 'bucket-prefix', 'name': 'ceph'}}}, {'name': 'THOTH_S3_ENDPOINT_URL', 'valueFrom': {'configMapKeyRef': {'key': 'host', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_KEY_ID', 'valueFrom': {'secretKeyRef': {'key': 'key-id', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_SECRET_KEY', 'valueFrom': {'secretKeyRef': {'key': 'secret-key', 'name': 'ceph'}}}], 'image': 'workflow-helpers', 'name': 'graph-schema-update-job', 'resources': {'limits': {'cpu': '1', 'memory': '512Mi'}, 'requests': {'cpu': '1', 'memory': '256Mi'}}}], 'restartPolicy': 'OnFailure'}}}}, 'post_params': [], 'files': {}, 'response_type': None, 'auth_settings': ['BearerToken'], '_return_http_data_only': True, 'collection_formats': None, '_preload_content': False, '_request_timeout': None, 'config': <kubernetes.client.configuration.Configuration object at 0x7f4e826c1940>, 'url': 'https://172.30.0.1:443/apis/batch/v1/namespaces/thoth-test-core/jobs'}}, {'filename': 'kubernetes/client/api_client.py', 'abs_path': '/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py', 'function': 'request', 'module': 'kubernetes.client.api_client', 'lineno': 382, 'pre_context': [' post_params=post_params,', ' _preload_content=_preload_content,', ' _request_timeout=_request_timeout,', ' body=body)', ' elif method == \"POST\":'], 'context_line': ' return self.rest_client.POST(url,', 'post_context': [' query_params=query_params,', ' headers=headers,', ' post_params=post_params,', ' _preload_content=_preload_content,', ' _request_timeout=_request_timeout,'], 'vars': {'self': <kubernetes.client.api_client.ApiClient object at 0x7f4e826c1730>, 'method': 'POST', 'url': 'https://172.30.0.1:443/apis/batch/v1/namespaces/thoth-test-core/jobs', 'query_params': [], 'headers': {'Accept': 'application/json', 'Content-Type': 'application/json', 'User-Agent': 'OpenAPI-Generator/11.0.0/python', 'authorization': 'bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IlhaVlc3YnFrZ1BaSllHSkpoNnVLenpTQlE1X3lrcEJzVndHWUdJclVDVEUifQ.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJ0aG90aC10ZXN0LWNvcmUiLCJrdWJlcm5ldGVzLmlvL3NlcnZpY2VhY2NvdW50L3NlY3JldC5uYW1lIjoibWFuYWdlbWVudC1hcGktdG9rZW4tNXRqcDkiLCJrdWJlcm5ldGVzLmlvL3NlcnZpY2VhY2NvdW50L3NlcnZpY2UtYWNjb3VudC5uYW1lIjoibWFuYWdlbWVudC1hcGkiLCJrdWJlcm5ldGVzLmlvL3NlcnZpY2VhY2NvdW50L3NlcnZpY2UtYWNjb3VudC51aWQiOiIyMTZlODJlZi05N2RhLTRiNGYtODk4Mi02ZDQzOWEwNjRmNjIiLCJzdWIiOiJzeXN0ZW06c2VydmljZWFjY291bnQ6dGhvdGgtdGVzdC1jb3JlOm1hbmFnZW1lbnQtYXBpIn0.KLDhPJFSnfyXXY5tV5788iPFbYUQ48TXZW2GEi6N-8QZ8qVkcXmubhlvlO8_P93Me9-CS8iAgsUtRH8RYBUOLssZ_R6ioZaOH3Qe7TQhMwbgu6cqFp0wYMRJnvY-RrXHLZBRjUrstgKmEeE6UsclRxy88TRGpp2XYjlTSgW-XHYUj_EMqTePkQXrQZ8MM7EwQiAvwRv653dq3mSmNt4iwF5ewN7C6lPm1mUpw2Kg-kn1ojcP_G1d9gr1u41FUnPD_BDyWfa0Y2fS1fmR33QL7whdmvC26t6qANbHyMlz-jzorH9gfvgdTzhEjXE9-FHu54zj63y6GJM1kJy0vGGKfw'}, 'post_params': [], 'body': {'apiVersion': 'batch/v1', 'kind': 'Job', 'metadata': {'labels': {'app': 'thoth', 'component': 'graph-schema-update-job', 'mark': 'cleanup'}, 'name': 'graph-update-schema-7f56a52c400a58e992985a33c2bce4d5'}, 'spec': {'backoffLimit': 7, 'template': {'metadata': {'labels': {'app': 'thoth', 'component': 'graph-schema-update-job', 'mark': 'cleanup'}}, 'spec': {'containers': [{'env': [{'name': 'THOTH_WORKFLOW_TASK', 'value': 'graph_schema_update'}, {'name': 'THOTH_DEPLOYMENT_NAME', 'valueFrom': {'configMapKeyRef': {'key': 'deployment-name', 'name': 'thoth'}}}, {'name': 'KNOWLEDGE_GRAPH_HOST', 'valueFrom': {'configMapKeyRef': {'key': 'postgresql-host', 'name': 'thoth'}}}, {'name': 'KNOWLEDGE_GRAPH_PORT', 'value': '5432'}, {'name': 'KNOWLEDGE_GRAPH_SSL_DISABLED', 'value': '1'}, {'name': 'KNOWLEDGE_GRAPH_USER', 'valueFrom': {'secretKeyRef': {'key': 'database-user', 'name': 'postgresql'}}}, {'name': 'KNOWLEDGE_GRAPH_PASSWORD', 'valueFrom': {'secretKeyRef': {'key': 'database-password', 'name': 'postgresql'}}}, {'name': 'KNOWLEDGE_GRAPH_DATABASE', 'valueFrom': {'secretKeyRef': {'key': 'database-name', 'name': 'postgresql'}}}, {'name': 'THOTH_CEPH_BUCKET', 'valueFrom': {'configMapKeyRef': {'key': 'bucket-name', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_BUCKET_PREFIX', 'valueFrom': {'configMapKeyRef': {'key': 'bucket-prefix', 'name': 'ceph'}}}, {'name': 'THOTH_S3_ENDPOINT_URL', 'valueFrom': {'configMapKeyRef': {'key': 'host', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_KEY_ID', 'valueFrom': {'secretKeyRef': {'key': 'key-id', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_SECRET_KEY', 'valueFrom': {'secretKeyRef': {'key': 'secret-key', 'name': 'ceph'}}}], 'image': 'workflow-helpers', 'name': 'graph-schema-update-job', 'resources': {'limits': {'cpu': '1', 'memory': '512Mi'}, 'requests': {'cpu': '1', 'memory': '256Mi'}}}], 'restartPolicy': 'OnFailure'}}}}, '_preload_content': False, '_request_timeout': None}}, {'filename': 'kubernetes/client/rest.py', 'abs_path': '/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/rest.py', 'function': 'POST', 'module': 'kubernetes.client.rest', 'lineno': 272, 'pre_context': [' _request_timeout=_request_timeout,', ' body=body)', '', ' def POST(self, url, headers=None, query_params=None, post_params=None,', ' body=None, _preload_content=True, _request_timeout=None):'], 'context_line': ' return self.request(\"POST\", url,', 'post_context': [' headers=headers,', ' query_params=query_params,', ' post_params=post_params,', ' _preload_content=_preload_content,', ' _request_timeout=_request_timeout,'], 'vars': {'self': <kubernetes.client.rest.RESTClientObject object at 0x7f4e826c17c0>, 'url': 'https://172.30.0.1:443/apis/batch/v1/namespaces/thoth-test-core/jobs', 'headers': {'Accept': 'application/json', 'Content-Type': 'application/json', 'User-Agent': 'OpenAPI-Generator/11.0.0/python', 'authorization': 'bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IlhaVlc3YnFrZ1BaSllHSkpoNnVLenpTQlE1X3lrcEJzVndHWUdJclVDVEUifQ.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJ0aG90aC10ZXN0LWNvcmUiLCJrdWJlcm5ldGVzLmlvL3NlcnZpY2VhY2NvdW50L3NlY3JldC5uYW1lIjoibWFuYWdlbWVudC1hcGktdG9rZW4tNXRqcDkiLCJrdWJlcm5ldGVzLmlvL3NlcnZpY2VhY2NvdW50L3NlcnZpY2UtYWNjb3VudC5uYW1lIjoibWFuYWdlbWVudC1hcGkiLCJrdWJlcm5ldGVzLmlvL3NlcnZpY2VhY2NvdW50L3NlcnZpY2UtYWNjb3VudC51aWQiOiIyMTZlODJlZi05N2RhLTRiNGYtODk4Mi02ZDQzOWEwNjRmNjIiLCJzdWIiOiJzeXN0ZW06c2VydmljZWFjY291bnQ6dGhvdGgtdGVzdC1jb3JlOm1hbmFnZW1lbnQtYXBpIn0.KLDhPJFSnfyXXY5tV5788iPFbYUQ48TXZW2GEi6N-8QZ8qVkcXmubhlvlO8_P93Me9-CS8iAgsUtRH8RYBUOLssZ_R6ioZaOH3Qe7TQhMwbgu6cqFp0wYMRJnvY-RrXHLZBRjUrstgKmEeE6UsclRxy88TRGpp2XYjlTSgW-XHYUj_EMqTePkQXrQZ8MM7EwQiAvwRv653dq3mSmNt4iwF5ewN7C6lPm1mUpw2Kg-kn1ojcP_G1d9gr1u41FUnPD_BDyWfa0Y2fS1fmR33QL7whdmvC26t6qANbHyMlz-jzorH9gfvgdTzhEjXE9-FHu54zj63y6GJM1kJy0vGGKfw'}, 'query_params': [], 'post_params': [], 'body': {'apiVersion': 'batch/v1', 'kind': 'Job', 'metadata': {'labels': {'app': 'thoth', 'component': 'graph-schema-update-job', 'mark': 'cleanup'}, 'name': 'graph-update-schema-7f56a52c400a58e992985a33c2bce4d5'}, 'spec': {'backoffLimit': 7, 'template': {'metadata': {'labels': {'app': 'thoth', 'component': 'graph-schema-update-job', 'mark': 'cleanup'}}, 'spec': {'containers': [{'env': [{'name': 'THOTH_WORKFLOW_TASK', 'value': 'graph_schema_update'}, {'name': 'THOTH_DEPLOYMENT_NAME', 'valueFrom': {'configMapKeyRef': {'key': 'deployment-name', 'name': 'thoth'}}}, {'name': 'KNOWLEDGE_GRAPH_HOST', 'valueFrom': {'configMapKeyRef': {'key': 'postgresql-host', 'name': 'thoth'}}}, {'name': 'KNOWLEDGE_GRAPH_PORT', 'value': '5432'}, {'name': 'KNOWLEDGE_GRAPH_SSL_DISABLED', 'value': '1'}, {'name': 'KNOWLEDGE_GRAPH_USER', 'valueFrom': {'secretKeyRef': {'key': 'database-user', 'name': 'postgresql'}}}, {'name': 'KNOWLEDGE_GRAPH_PASSWORD', 'valueFrom': {'secretKeyRef': {'key': 'database-password', 'name': 'postgresql'}}}, {'name': 'KNOWLEDGE_GRAPH_DATABASE', 'valueFrom': {'secretKeyRef': {'key': 'database-name', 'name': 'postgresql'}}}, {'name': 'THOTH_CEPH_BUCKET', 'valueFrom': {'configMapKeyRef': {'key': 'bucket-name', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_BUCKET_PREFIX', 'valueFrom': {'configMapKeyRef': {'key': 'bucket-prefix', 'name': 'ceph'}}}, {'name': 'THOTH_S3_ENDPOINT_URL', 'valueFrom': {'configMapKeyRef': {'key': 'host', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_KEY_ID', 'valueFrom': {'secretKeyRef': {'key': 'key-id', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_SECRET_KEY', 'valueFrom': {'secretKeyRef': {'key': 'secret-key', 'name': 'ceph'}}}], 'image': 'workflow-helpers', 'name': 'graph-schema-update-job', 'resources': {'limits': {'cpu': '1', 'memory': '512Mi'}, 'requests': {'cpu': '1', 'memory': '256Mi'}}}], 'restartPolicy': 'OnFailure'}}}}, '_preload_content': False, '_request_timeout': None}}, {'filename': 'kubernetes/client/rest.py', 'abs_path': '/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/rest.py', 'function': 'request', 'module': 'kubernetes.client.rest', 'lineno': 231, 'pre_context': ['', ' # log response body', ' logger.debug(\"response body: %s\", r.data)', '', ' if not 200 <= r.status <= 299:'], 'context_line': ' raise ApiException(http_resp=r)', 'post_context': ['', ' return r', '', ' def GET(self, url, headers=None, query_params=None, _preload_content=True,', ' _request_timeout=None):'], 'vars': {'self': <kubernetes.client.rest.RESTClientObject object at 0x7f4e826c17c0>, 'method': 'POST', 'url': 'https://172.30.0.1:443/apis/batch/v1/namespaces/thoth-test-core/jobs', 'query_params': [], 'headers': {'Accept': 'application/json', 'Content-Type': 'application/json', 'User-Agent': 'OpenAPI-Generator/11.0.0/python', 'authorization': 'bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IlhaVlc3YnFrZ1BaSllHSkpoNnVLenpTQlE1X3lrcEJzVndHWUdJclVDVEUifQ.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJ0aG90aC10ZXN0LWNvcmUiLCJrdWJlcm5ldGVzLmlvL3NlcnZpY2VhY2NvdW50L3NlY3JldC5uYW1lIjoibWFuYWdlbWVudC1hcGktdG9rZW4tNXRqcDkiLCJrdWJlcm5ldGVzLmlvL3NlcnZpY2VhY2NvdW50L3NlcnZpY2UtYWNjb3VudC5uYW1lIjoibWFuYWdlbWVudC1hcGkiLCJrdWJlcm5ldGVzLmlvL3NlcnZpY2VhY2NvdW50L3NlcnZpY2UtYWNjb3VudC51aWQiOiIyMTZlODJlZi05N2RhLTRiNGYtODk4Mi02ZDQzOWEwNjRmNjIiLCJzdWIiOiJzeXN0ZW06c2VydmljZWFjY291bnQ6dGhvdGgtdGVzdC1jb3JlOm1hbmFnZW1lbnQtYXBpIn0.KLDhPJFSnfyXXY5tV5788iPFbYUQ48TXZW2GEi6N-8QZ8qVkcXmubhlvlO8_P93Me9-CS8iAgsUtRH8RYBUOLssZ_R6ioZaOH3Qe7TQhMwbgu6cqFp0wYMRJnvY-RrXHLZBRjUrstgKmEeE6UsclRxy88TRGpp2XYjlTSgW-XHYUj_EMqTePkQXrQZ8MM7EwQiAvwRv653dq3mSmNt4iwF5ewN7C6lPm1mUpw2Kg-kn1ojcP_G1d9gr1u41FUnPD_BDyWfa0Y2fS1fmR33QL7whdmvC26t6qANbHyMlz-jzorH9gfvgdTzhEjXE9-FHu54zj63y6GJM1kJy0vGGKfw'}, 'body': {'apiVersion': 'batch/v1', 'kind': 'Job', 'metadata': {'labels': {'app': 'thoth', 'component': 'graph-schema-update-job', 'mark': 'cleanup'}, 'name': 'graph-update-schema-7f56a52c400a58e992985a33c2bce4d5'}, 'spec': {'backoffLimit': 7, 'template': {'metadata': {'labels': {'app': 'thoth', 'component': 'graph-schema-update-job', 'mark': 'cleanup'}}, 'spec': {'containers': [{'env': [{'name': 'THOTH_WORKFLOW_TASK', 'value': 'graph_schema_update'}, {'name': 'THOTH_DEPLOYMENT_NAME', 'valueFrom': {'configMapKeyRef': {'key': 'deployment-name', 'name': 'thoth'}}}, {'name': 'KNOWLEDGE_GRAPH_HOST', 'valueFrom': {'configMapKeyRef': {'key': 'postgresql-host', 'name': 'thoth'}}}, {'name': 'KNOWLEDGE_GRAPH_PORT', 'value': '5432'}, {'name': 'KNOWLEDGE_GRAPH_SSL_DISABLED', 'value': '1'}, {'name': 'KNOWLEDGE_GRAPH_USER', 'valueFrom': {'secretKeyRef': {'key': 'database-user', 'name': 'postgresql'}}}, {'name': 'KNOWLEDGE_GRAPH_PASSWORD', 'valueFrom': {'secretKeyRef': {'key': 'database-password', 'name': 'postgresql'}}}, {'name': 'KNOWLEDGE_GRAPH_DATABASE', 'valueFrom': {'secretKeyRef': {'key': 'database-name', 'name': 'postgresql'}}}, {'name': 'THOTH_CEPH_BUCKET', 'valueFrom': {'configMapKeyRef': {'key': 'bucket-name', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_BUCKET_PREFIX', 'valueFrom': {'configMapKeyRef': {'key': 'bucket-prefix', 'name': 'ceph'}}}, {'name': 'THOTH_S3_ENDPOINT_URL', 'valueFrom': {'configMapKeyRef': {'key': 'host', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_KEY_ID', 'valueFrom': {'secretKeyRef': {'key': 'key-id', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_SECRET_KEY', 'valueFrom': {'secretKeyRef': {'key': 'secret-key', 'name': 'ceph'}}}], 'image': 'workflow-helpers', 'name': 'graph-schema-update-job', 'resources': {'limits': {'cpu': '1', 'memory': '512Mi'}, 'requests': {'cpu': '1', 'memory': '256Mi'}}}], 'restartPolicy': 'OnFailure'}}}}, 'post_params': {}, '_preload_content': False, '_request_timeout': None, 'timeout': None, 'request_body': '{\"apiVersion\": \"batch/v1\", \"kind\": \"Job\", \"metadata\": {\"labels\": {\"app\": \"thoth\", \"component\": \"graph-schema-update-job\", \"mark\": \"cleanup\"}, \"name\": \"graph-update-schema-7f56a52c400a58e992985a33c2bce4d5\"}, \"spec\": {\"backoffLimit\": 7, \"template\": {\"metadata\": {\"labels\": {\"app\": \"thoth\", \"component\": \"graph-schema-update-job\", \"mark\": \"cleanup\"}}, \"spec\": {\"containers\": [{\"env\": [{\"name\": \"THOTH_WORKFLOW_TASK\", \"value\": \"graph_schema_update\"}, {\"name\": \"THOTH_DEPLOYMENT_NAME\", \"valueFrom\": {\"configMapKeyRef\": {\"key\": \"deployment-name\", \"name\": \"thoth\"}}}, {\"name\": \"KNOWLEDGE_GRAPH_HOST\", \"valueFrom\": {\"configMapKeyRef\": {\"key\": \"postgresql-host\", \"name\": \"thoth\"}}}, {\"name\": \"KNOWLEDGE_GRAPH_PORT\", \"value\": \"5432\"}, {\"name\": \"KNOWLEDGE_GRAPH_SSL_DISABLED\", \"value\": \"1\"}, {\"name\": \"KNOWLEDGE_GRAPH_USER\", \"valueFrom\": {\"secretKeyRef\": {\"key\": \"database-user\", \"name\": \"postgresql\"}}}, {\"name\": \"KNOWLEDGE_GRAPH_PASSWORD\", \"valueFrom\": {\"secretKeyRef\": {\"key\": \"database-password\", \"name\": \"postgresql\"}}}, {\"name\": \"KNOWLEDGE_GRAPH_DATABASE\", \"valueFrom\": {\"secretKeyRef\": {\"key\": \"database-name\", \"name\": \"postgresql\"}}}, {\"name\": \"THOTH_CEPH_BUCKET\", \"valueFrom\": {\"configMapKeyRef\": {\"key\": \"bucket-name\", \"name\": \"ceph\"}}}, {\"name\": \"THOTH_CEPH_BUCKET_PREFIX\", \"valueFrom\": {\"configMapKeyRef\": {\"key\": \"bucket-prefix\", \"name\": \"ceph\"}}}, {\"name\": \"THOTH_S3_ENDPOINT_URL\", \"valueFrom\": {\"configMapKeyRef\": {\"key\": \"host\", \"name\": \"ceph\"}}}, {\"name\": \"THOTH_CEPH_KEY_ID\", \"valueFrom\": {\"secretKeyRef\": {\"key\": \"key-id\", \"name\": \"ceph\"}}}, {\"name\": \"THOTH_CEPH_SECRET_KEY\", \"valueFrom\": {\"secretKeyRef\": {\"key\": \"secret-key\", \"name\": \"ceph\"}}}], \"image\": \"workflow-helpers\", \"name\": \"graph-schema-update-job\", \"resources\": {\"limits\": {\"cpu\": \"1\", \"memory\": \"512Mi\"}, \"requests\": {\"cpu\": \"1\", \"memory\": \"256Mi\"}}}], \"restartPolicy\": \"OnFailure\"}}}}', 'r': <urllib3.response.HTTPResponse object at 0x7f4e80044eb0>}}]}}, {'module': 'openshift.dynamic.exceptions', 'type': 'ForbiddenError', 'value': '403\\nReason: Forbidden\\nHTTP response headers: HTTPHeaderDict({\\'Audit-Id\\': \\'49583870-4187-49f8-9539-aef793ac349e\\', \\'Cache-Control\\': \\'no-cache, private\\', \\'Content-Type\\': \\'application/json\\', \\'X-Content-Type-Options\\': \\'nosniff\\', \\'Date\\': \\'Thu, 14 Jan 2021 16:45:25 GMT\\', \\'Content-Length\\': \\'334\\'})\\nHTTP response body: b\\'{\"kind\":\"Status\",\"apiVersion\":\"v1\",\"metadata\":{},\"status\":\"Failure\",\"message\":\"jobs.batch is forbidden: User \\\\\\\\\"system:serviceaccount:thoth-test-core:management-api\\\\\\\\\" cannot create resource \\\\\\\\\"jobs\\\\\\\\\" in API group \\\\\\\\\"batch\\\\\\\\\" in the namespace \\\\\\\\\"thoth-test-core\\\\\\\\\"\",\"reason\":\"Forbidden\",\"details\":{\"group\":\"batch\",\"kind\":\"jobs\"},\"code\":403}\\\\n\\'\\nOriginal traceback: \\n File \"/opt/app-root/lib64/python3.8/site-packages/openshift/dynamic/client.py\", line 42, in inner\\n resp = func(self, *args, **kwargs)\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/openshift/dynamic/client.py\", line 235, in request\\n return self.client.call_api(\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py\", line 340, in call_api\\n return self.__call_api(resource_path, method,\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py\", line 172, in __call_api\\n response_data = self.request(\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py\", line 382, in request\\n return self.rest_client.POST(url,\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/rest.py\", line 272, in POST\\n return self.request(\"POST\", url,\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/rest.py\", line 231, in request\\n raise ApiException(http_resp=r)\\n', 'mechanism': {'type': 'logging', 'handled': True}, 'stacktrace': {'frames': [{'filename': 'flask/app.py', 'abs_path': '/opt/app-root/lib64/python3.8/site-packages/flask/app.py', 'function': 'wsgi_app', 'module': 'flask.app', 'lineno': 2447, 'pre_context': [' ctx = self.request_context(environ)', ' error = None', ' try:', ' try:', ' ctx.push()'], 'context_line': ' response = self.full_dispatch_request()', 'post_context': [' except Exception as e:', ' error = e', ' response = self.handle_exception(e)', ' except: # noqa: B001', ' error = sys.exc_info()[1]'], 'vars': {'self': <Flask 'thoth.management_api.openapi_server'>, 'environ': {'wsgi.errors': <gunicorn.http.wsgi.WSGIErrorsWrapper object at 0x7f4e800311c0>, 'wsgi.version': (1, 0), 'wsgi.multithread': False, 'wsgi.multiprocess': False, 'wsgi.run_once': False, 'wsgi.file_wrapper': <class 'gunicorn.http.wsgi.FileWrapper'>, 'wsgi.input_terminated': True, 'SERVER_SOFTWARE': 'gunicorn/20.0.4', 'wsgi.input': <gunicorn.http.body.Body object at 0x7f4e800315e0>, 'gunicorn.socket': <socket.socket fd=16, family=AddressFamily.AF_INET, type=SocketKind.SOCK_STREAM, proto=0, laddr=('172.21.12.117', 8080), raddr=('172.20.8.1', 39828)>, 'REQUEST_METHOD': 'PUT', 'QUERY_STRING': 'secret=secret101', 'RAW_URI': '/api/v1/graph/initialize?secret=secret101', 'SERVER_PROTOCOL': 'HTTP/1.1', 'CONTENT_LENGTH': '0', 'HTTP_ACCEPT': '*/*', 'HTTP_USER_AGENT': 'Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/87.0.4280.141 Safari/537.36', 'HTTP_ORIGIN': 'https://management.test.thoth-station.ninja', 'HTTP_SEC_FETCH_SITE': 'same-origin', 'HTTP_SEC_FETCH_MODE': 'cors', 'HTTP_SEC_FETCH_DEST': 'empty', 'HTTP_REFERER': 'https://management.test.thoth-station.ninja/api/v1/ui/', 'HTTP_ACCEPT_ENCODING': 'gzip, deflate, br', 'HTTP_ACCEPT_LANGUAGE': 'it-IT,it;q=0.9,en-US;q=0.8,en;q=0.7,es;q=0.6', 'HTTP_HOST': 'management.test.thoth-station.ninja', 'HTTP_X_FORWARDED_HOST': 'management.test.thoth-station.ninja', 'HTTP_X_FORWARDED_PORT': '443', 'HTTP_X_FORWARDED_PROTO': 'https', 'HTTP_FORWARDED': 'for=10.36.112.242;host=management.test.thoth-station.ninja;proto=https', 'HTTP_X_FORWARDED_FOR': '10.36.112.242', 'HTTP_CONNECTION': 'close', 'wsgi.url_scheme': 'http', 'REMOTE_ADDR': '172.20.8.1', 'REMOTE_PORT': '39828', 'SERVER_NAME': '0.0.0.0', 'SERVER_PORT': '8080', 'PATH_INFO': '/api/v1/graph/initialize', 'SCRIPT_NAME': '', 'werkzeug.request': <Request 'http://management.test.thoth-station.ninja/api/v1/graph/initialize?secret=secret101' [PUT]>}, 'start_response': functools.partial(<function _sentry_start_response at 0x7f4e8293ea60>, <bound method Response.start_response of <gunicorn.http.wsgi.Response object at 0x7f4e80031190>>, <Transaction(name='/api/v1.thoth_management_api_api_v1_initialize_schema', op='http.server', trace_id='df56345840f44f009335865766b30f2f', span_id='a032e586b4eb3cf8', parent_span_id=None, sampled=True)>), 'ctx': <RequestContext 'http://management.test.thoth-station.ninja/api/v1/graph/initialize?secret=secret101' [PUT] of thoth.management_api.openapi_server>, 'error': ForbiddenError(ApiException(), ' File \"/opt/app-root/lib64/python3.8/site-packages/openshift/dynamic/client.py\", line 42, in inner\\n resp = func(self, *args, **kwargs)\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/openshift/dynamic/client.py\", line 235, in request\\n return self.client.call_api(\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py\", line 340, in call_api\\n return self.__call_api(resource_path, method,\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py\", line 172, in __call_api\\n response_data = self.request(\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py\", line 382, in request\\n return self.rest_client.POST(url,\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/rest.py\", line 272, in POST\\n return self.request(\"POST\", url,\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/rest.py\", line 231, in request\\n raise ApiException(http_resp=r)\\n'), 'e': ForbiddenError(ApiException(), ' File \"/opt/app-root/lib64/python3.8/site-packages/openshift/dynamic/client.py\", line 42, in inner\\n resp = func(self, *args, **kwargs)\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/openshift/dynamic/client.py\", line 235, in request\\n return self.client.call_api(\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py\", line 340, in call_api\\n return self.__call_api(resource_path, method,\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py\", line 172, in __call_api\\n response_data = self.request(\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py\", line 382, in request\\n return self.rest_client.POST(url,\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/rest.py\", line 272, in POST\\n return self.request(\"POST\", url,\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/rest.py\", line 231, in request\\n raise ApiException(http_resp=r)\\n')}}, {'filename': 'flask/app.py', 'abs_path': '/opt/app-root/lib64/python3.8/site-packages/flask/app.py', 'function': 'full_dispatch_request', 'module': 'flask.app', 'lineno': 1952, 'pre_context': [' request_started.send(self)', ' rv = self.preprocess_request()', ' if rv is None:', ' rv = self.dispatch_request()', ' except Exception as e:'], 'context_line': ' rv = self.handle_user_exception(e)', 'post_context': [' return self.finalize_request(rv)', '', ' def finalize_request(self, rv, from_error_handler=False):', ' \"\"\"Given the return value from a view function this finalizes', ' the request by converting it into a response and invoking the'], 'vars': {'self': <Flask 'thoth.management_api.openapi_server'>, 'rv': None}}, {'filename': 'flask/app.py', 'abs_path': '/opt/app-root/lib64/python3.8/site-packages/flask/app.py', 'function': 'handle_user_exception', 'module': 'flask.app', 'lineno': 1821, 'pre_context': [' return self.handle_http_exception(e)', '', ' handler = self._find_error_handler(e)', '', ' if handler is None:'], 'context_line': ' reraise(exc_type, exc_value, tb)', 'post_context': [' return handler(e)', '', ' def handle_exception(self, e):', ' \"\"\"Handle an exception that did not have an error handler', ' associated with it, or that was raised from an error handler.'], 'vars': {'self': <Flask 'thoth.management_api.openapi_server'>, 'e': ForbiddenError(ApiException(), ' File \"/opt/app-root/lib64/python3.8/site-packages/openshift/dynamic/client.py\", line 42, in inner\\n resp = func(self, *args, **kwargs)\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/openshift/dynamic/client.py\", line 235, in request\\n return self.client.call_api(\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py\", line 340, in call_api\\n return self.__call_api(resource_path, method,\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py\", line 172, in __call_api\\n response_data = self.request(\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py\", line 382, in request\\n return self.rest_client.POST(url,\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/rest.py\", line 272, in POST\\n return self.request(\"POST\", url,\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/rest.py\", line 231, in request\\n raise ApiException(http_resp=r)\\n'), 'exc_type': <class 'openshift.dynamic.exceptions.ForbiddenError'>, 'exc_value': ForbiddenError(ApiException(), ' File \"/opt/app-root/lib64/python3.8/site-packages/openshift/dynamic/client.py\", line 42, in inner\\n resp = func(self, *args, **kwargs)\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/openshift/dynamic/client.py\", line 235, in request\\n return self.client.call_api(\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py\", line 340, in call_api\\n return self.__call_api(resource_path, method,\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py\", line 172, in __call_api\\n response_data = self.request(\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py\", line 382, in request\\n return self.rest_client.POST(url,\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/rest.py\", line 272, in POST\\n return self.request(\"POST\", url,\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/rest.py\", line 231, in request\\n raise ApiException(http_resp=r)\\n'), 'tb': <traceback object at 0x7f4e7a70fd00>, 'handler': None}}, {'filename': 'flask/_compat.py', 'abs_path': '/opt/app-root/lib64/python3.8/site-packages/flask/_compat.py', 'function': 'reraise', 'module': 'flask._compat', 'lineno': 39, 'pre_context': [' import collections.abc as collections_abc', '', ' def reraise(tp, value, tb=None):', ' if value.__traceback__ is not tb:', ' raise value.with_traceback(tb)'], 'context_line': ' raise value', 'post_context': ['', ' implements_to_string = _identity', '', 'else:', ' iterkeys = lambda d: d.iterkeys()'], 'vars': {'tp': <class 'openshift.dynamic.exceptions.ForbiddenError'>, 'value': ForbiddenError(ApiException(), ' File \"/opt/app-root/lib64/python3.8/site-packages/openshift/dynamic/client.py\", line 42, in inner\\n resp = func(self, *args, **kwargs)\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/openshift/dynamic/client.py\", line 235, in request\\n return self.client.call_api(\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py\", line 340, in call_api\\n return self.__call_api(resource_path, method,\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py\", line 172, in __call_api\\n response_data = self.request(\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py\", line 382, in request\\n return self.rest_client.POST(url,\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/rest.py\", line 272, in POST\\n return self.request(\"POST\", url,\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/rest.py\", line 231, in request\\n raise ApiException(http_resp=r)\\n'), 'tb': <traceback object at 0x7f4e7a70fd00>}}, {'filename': 'flask/app.py', 'abs_path': '/opt/app-root/lib64/python3.8/site-packages/flask/app.py', 'function': 'full_dispatch_request', 'module': 'flask.app', 'lineno': 1950, 'pre_context': [' self.try_trigger_before_first_request_functions()', ' try:', ' request_started.send(self)', ' rv = self.preprocess_request()', ' if rv is None:'], 'context_line': ' rv = self.dispatch_request()', 'post_context': [' except Exception as e:', ' rv = self.handle_user_exception(e)', ' return self.finalize_request(rv)', '', ' def finalize_request(self, rv, from_error_handler=False):'], 'vars': {'self': <Flask 'thoth.management_api.openapi_server'>, 'rv': None}}, {'filename': 'flask/app.py', 'abs_path': '/opt/app-root/lib64/python3.8/site-packages/flask/app.py', 'function': 'dispatch_request', 'module': 'flask.app', 'lineno': 1936, 'pre_context': [' getattr(rule, \"provide_automatic_options\", False)', ' and req.method == \"OPTIONS\"', ' ):', ' return self.make_default_options_response()', ' # otherwise dispatch to the handler for that endpoint'], 'context_line': ' return self.view_functions[rule.endpoint](**req.view_args)', 'post_context': ['', ' def full_dispatch_request(self):', ' \"\"\"Dispatches the request and on top of that performs request', ' pre and postprocessing as well as HTTP exception catching and', ' error handling.'], 'vars': {'self': <Flask 'thoth.management_api.openapi_server'>, 'req': <Request 'http://management.test.thoth-station.ninja/api/v1/graph/initialize?secret=secret101' [PUT]>, 'rule': <Rule '/api/v1/graph/initialize' (OPTIONS, PUT) -> /api/v1.thoth_management_api_api_v1_initialize_schema>}}, {'filename': 'connexion/decorators/decorator.py', 'abs_path': '/opt/app-root/lib64/python3.8/site-packages/connexion/decorators/decorator.py', 'function': 'wrapper', 'module': 'connexion.decorators.decorator', 'lineno': 48, 'pre_context': ['', ' else: # pragma: 3 no cover', ' @functools.wraps(function)', ' def wrapper(*args, **kwargs):', ' request = self.api.get_request(*args, **kwargs)'], 'context_line': ' response = function(request)', 'post_context': [' return self.api.get_response(response, self.mimetype, request)', '', ' return wrapper'], 'vars': {'args': (), 'kwargs': {}, 'request': <connexion.lifecycle.ConnexionRequest object at 0x7f4e800aa820>, 'function': <function initialize_schema at 0x7f4e82802ee0>, 'self': <BaseDecorator>}}, {'filename': 'connexion/decorators/uri_parsing.py', 'abs_path': '/opt/app-root/lib64/python3.8/site-packages/connexion/decorators/uri_parsing.py', 'function': 'wrapper', 'module': 'connexion.decorators.uri_parsing', 'lineno': 144, 'pre_context': [' form = coerce_dict(request.form)', '', ' request.query = self.resolve_query(query)', ' request.path_params = self.resolve_path(path_params)', ' request.form = self.resolve_form(form)'], 'context_line': ' response = function(request)', 'post_context': [' return response', '', ' return wrapper', '', ''], 'vars': {'request': <connexion.lifecycle.ConnexionRequest object at 0x7f4e800aa820>, 'coerce_dict': <function AbstractURIParser.__call__.<locals>.wrapper.<locals>.coerce_dict at 0x7f4e8004bd30>, 'query': {'secret': ['secret101']}, 'path_params': {}, 'form': {}, 'function': <function initialize_schema at 0x7f4e82802e50>, 'self': <OpenAPIURIParser>}}, {'filename': 'connexion/decorators/validation.py', 'abs_path': '/opt/app-root/lib64/python3.8/site-packages/connexion/decorators/validation.py', 'function': 'wrapper', 'module': 'connexion.decorators.validation', 'lineno': 384, 'pre_context': [\" for param in self.parameters.get('formData', []):\", ' error = self.validate_formdata_parameter(param[\"name\"], param, request)', ' if error:', ' raise BadRequestProblem(detail=error)', ''], 'context_line': ' return function(request)', 'post_context': ['', ' return wrapper'], 'vars': {'request': <connexion.lifecycle.ConnexionRequest object at 0x7f4e800aa820>, 'query_errors': set(), 'formdata_errors': set(), 'param': {'name': 'secret', 'in': 'query', 'required': True, 'description': 'A secret to authorize operation.', 'schema': {'type': 'string'}}, 'error': None, 'function': <function initialize_schema at 0x7f4e82802dc0>, 'self': <connexion.decorators.validation.ParameterValidator object at 0x7f4e827fc340>}}, {'filename': 'connexion/decorators/parameter.py', 'abs_path': '/opt/app-root/lib64/python3.8/site-packages/connexion/decorators/parameter.py', 'function': 'wrapper', 'module': 'connexion.decorators.parameter', 'lineno': 121, 'pre_context': ['', ' # attempt to provide the request context to the function', ' if pass_context_arg_name and (has_kwargs or pass_context_arg_name in arguments):', ' kwargs[pass_context_arg_name] = request.context', ''], 'context_line': ' return function(**kwargs)', 'post_context': ['', ' return wrapper'], 'vars': {'request': <connexion.lifecycle.ConnexionRequest object at 0x7f4e800aa820>, 'kwargs': {'secret': 'secret101'}, 'request_body': None, 'query': {'secret': 'secret101'}, 'arguments': ['secret'], 'consumes': ['application/json'], 'function': <function initialize_schema at 0x7f4e82694d30>, 'has_kwargs': False, 'operation': <connexion.operations.openapi.OpenAPIOperation object at 0x7f4e827fc130>, 'pass_context_arg_name': None, 'pythonic_params': False, 'sanitize': <function parameter_to_arg.<locals>.sanitized at 0x7f4e82802ca0>}}, {'filename': '/opt/app-root/src/thoth/management_api/api_v1.py', 'abs_path': '/opt/app-root/src/thoth/management_api/api_v1.py', 'function': 'initialize_schema', 'module': 'thoth.management_api.api_v1', 'lineno': 348, 'pre_context': ['def initialize_schema(secret: str):', ' \"\"\"Initialize/update schema in graph database (async).\"\"\"', ' if secret != Configuration.THOTH_MANAGEMENT_API_TOKEN:', ' return {\"error\": \"Wrong secret provided\"}, 401', ''], 'context_line': ' job_id = _OPENSHIFT.schedule_graph_schema_update()', 'post_context': [' return {\"job_id\": job_id}, 201', '', '', 'def schedule_solver_unsolvable(secret: str, solver_name: str) -> tuple:', ' \"\"\"Schedule solving of unsolvable packages for the given solver.\"\"\"'], 'vars': {'secret': 'secret101'}}, {'filename': '/opt/app-root/lib64/python3.8/site-packages/thoth/common/openshift.py', 'abs_path': '/opt/app-root/lib64/python3.8/site-packages/thoth/common/openshift.py', 'function': 'schedule_graph_schema_update', 'module': 'thoth.common.openshift', 'lineno': 791, 'pre_context': ['', ' template = self.oc_process(namespace, template)', '', ' graph_update_schema_template = template[\"objects\"][0]', ''], 'context_line': ' response = self.ocp_client.resources.get(', 'post_context': [' api_version=graph_update_schema_template[\"apiVersion\"],', ' kind=graph_update_schema_template[\"kind\"],', ' ).create(body=graph_update_schema_template, namespace=namespace)', '', ' _LOGGER.debug(\"OpenShift response for creating a pod: %r\", response.to_dict())'], 'vars': {'self': <thoth.common.openshift.OpenShift object at 0x7f4e826e3310>, 'job_id': 'graph-update-schema-7f56a52c400a58e992985a33c2bce4d5', 'namespace': 'thoth-test-core', 'template': {'kind': 'Template', 'apiVersion': 'template.openshift.io/v1', 'metadata': {'name': 'graph-schema-update-job', 'namespace': 'thoth-test-core', 'selfLink': '/apis/template.openshift.io/v1/namespaces/thoth-test-core/processedtemplates/graph-schema-update-job', 'uid': '1c4d59d8-dd65-4f1b-82f9-c2505cc3c6b2', 'resourceVersion': '391642130', 'creationTimestamp': '2021-01-14T16:43:05Z', 'labels': {'template': 'graph-schema-update-job'}, 'annotations': {'description': 'Thoth: Graph Schema Update Job for updating schema in database', 'kubectl.kubernetes.io/last-applied-configuration': '{\"apiVersion\":\"template.openshift.io/v1\",\"kind\":\"Template\",\"metadata\":{\"annotations\":{\"description\":\"Thoth: Graph Schema Update Job for updating schema in database\",\"openshift.io/display-name\":\"Thoth: Graph Schema Update Job Multiple Documents\",\"tags\":\"thoth,ai-stacks,aistacks,graph-schema-update-job\",\"template.openshift.io/documentation-url\":\"https://github.com/Thoth-Station/\",\"template.openshift.io/long-description\":\"This template defines resources needed to deploy Thoth\\'s Graph Schema Update Job on OpenShift.\\\\n\",\"template.openshift.io/provider-display-name\":\"Red Hat, Inc.\"},\"labels\":{\"template\":\"graph-schema-update-job\"},\"name\":\"graph-schema-update-job\",\"namespace\":\"thoth-test-core\"},\"objects\":[{\"apiVersion\":\"batch/v1\",\"kind\":\"Job\",\"metadata\":{\"labels\":{\"app\":\"thoth\",\"component\":\"graph-schema-update-job\",\"mark\":\"cleanup\"},\"name\":\"${THOTH_SCHEMA_UPDATE_JOB_ID}\"},\"spec\":{\"backoffLimit\":7,\"template\":{\"metadata\":{\"labels\":{\"app\":\"thoth\",\"component\":\"graph-schema-update-job\",\"mark\":\"cleanup\"}},\"spec\":{\"containers\":[{\"env\":[{\"name\":\"THOTH_WORKFLOW_TASK\",\"value\":\"graph_schema_update\"},{\"name\":\"THOTH_DEPLOYMENT_NAME\",\"valueFrom\":{\"configMapKeyRef\":{\"key\":\"deployment-name\",\"name\":\"thoth\"}}},{\"name\":\"KNOWLEDGE_GRAPH_HOST\",\"valueFrom\":{\"configMapKeyRef\":{\"key\":\"postgresql-host\",\"name\":\"thoth\"}}},{\"name\":\"KNOWLEDGE_GRAPH_PORT\",\"value\":\"5432\"},{\"name\":\"KNOWLEDGE_GRAPH_SSL_DISABLED\",\"value\":\"1\"},{\"name\":\"KNOWLEDGE_GRAPH_USER\",\"valueFrom\":{\"secretKeyRef\":{\"key\":\"database-user\",\"name\":\"postgresql\"}}},{\"name\":\"KNOWLEDGE_GRAPH_PASSWORD\",\"valueFrom\":{\"secretKeyRef\":{\"key\":\"database-password\",\"name\":\"postgresql\"}}},{\"name\":\"KNOWLEDGE_GRAPH_DATABASE\",\"valueFrom\":{\"secretKeyRef\":{\"key\":\"database-name\",\"name\":\"postgresql\"}}},{\"name\":\"THOTH_CEPH_BUCKET\",\"valueFrom\":{\"configMapKeyRef\":{\"key\":\"bucket-name\",\"name\":\"ceph\"}}},{\"name\":\"THOTH_CEPH_BUCKET_PREFIX\",\"valueFrom\":{\"configMapKeyRef\":{\"key\":\"bucket-prefix\",\"name\":\"ceph\"}}},{\"name\":\"THOTH_S3_ENDPOINT_URL\",\"valueFrom\":{\"configMapKeyRef\":{\"key\":\"host\",\"name\":\"ceph\"}}},{\"name\":\"THOTH_CEPH_KEY_ID\",\"valueFrom\":{\"secretKeyRef\":{\"key\":\"key-id\",\"name\":\"ceph\"}}},{\"name\":\"THOTH_CEPH_SECRET_KEY\",\"valueFrom\":{\"secretKeyRef\":{\"key\":\"secret-key\",\"name\":\"ceph\"}}}],\"image\":\"workflow-helpers\",\"name\":\"graph-schema-update-job\",\"resources\":{\"limits\":{\"cpu\":\"1\",\"memory\":\"512Mi\"},\"requests\":{\"cpu\":\"1\",\"memory\":\"256Mi\"}}}],\"restartPolicy\":\"OnFailure\"}}}}],\"parameters\":[{\"description\":\"A unique identifier of job\",\"displayName\":\"A unique identifier of job\",\"name\":\"THOTH_SCHEMA_UPDATE_JOB_ID\",\"required\":true}]}\\n', 'openshift.io/display-name': 'Thoth: Graph Schema Update Job Multiple Documents', 'tags': 'thoth,ai-stacks,aistacks,graph-schema-update-job', 'template.openshift.io/documentation-url': 'https://github.com/Thoth-Station/', 'template.openshift.io/long-description': \"This template defines resources needed to deploy Thoth's Graph Schema Update Job on OpenShift.\\n\", 'template.openshift.io/provider-display-name': 'Red Hat, Inc.'}, 'managedFields': [{'manager': 'python-requests', 'operation': 'Update', 'apiVersion': 'template.openshift.io/v1', 'time': '2021-01-14T16:45:25Z', 'fieldsType': 'FieldsV1', 'fieldsV1': {'f:metadata': {'f:annotations': {'.': {}, 'f:description': {}, 'f:kubectl.kubernetes.io/last-applied-configuration': {}, 'f:openshift.io/display-name': {}, 'f:tags': {}, 'f:template.openshift.io/documentation-url': {}, 'f:template.openshift.io/long-description': {}, 'f:template.openshift.io/provider-display-name': {}}, 'f:labels': {'.': {}, 'f:template': {}}}, 'f:objects': {}, 'f:parameters': {}}}]}, 'objects': [{'apiVersion': 'batch/v1', 'kind': 'Job', 'metadata': {'labels': {'app': 'thoth', 'component': 'graph-schema-update-job', 'mark': 'cleanup'}, 'name': 'graph-update-schema-7f56a52c400a58e992985a33c2bce4d5'}, 'spec': {'backoffLimit': 7, 'template': {'metadata': {'labels': {'app': 'thoth', 'component': 'graph-schema-update-job', 'mark': 'cleanup'}}, 'spec': {'containers': [{'env': [{'name': 'THOTH_WORKFLOW_TASK', 'value': 'graph_schema_update'}, {'name': 'THOTH_DEPLOYMENT_NAME', 'valueFrom': {'configMapKeyRef': {'key': 'deployment-name', 'name': 'thoth'}}}, {'name': 'KNOWLEDGE_GRAPH_HOST', 'valueFrom': {'configMapKeyRef': {'key': 'postgresql-host', 'name': 'thoth'}}}, {'name': 'KNOWLEDGE_GRAPH_PORT', 'value': '5432'}, {'name': 'KNOWLEDGE_GRAPH_SSL_DISABLED', 'value': '1'}, {'name': 'KNOWLEDGE_GRAPH_USER', 'valueFrom': {'secretKeyRef': {'key': 'database-user', 'name': 'postgresql'}}}, {'name': 'KNOWLEDGE_GRAPH_PASSWORD', 'valueFrom': {'secretKeyRef': {'key': 'database-password', 'name': 'postgresql'}}}, {'name': 'KNOWLEDGE_GRAPH_DATABASE', 'valueFrom': {'secretKeyRef': {'key': 'database-name', 'name': 'postgresql'}}}, {'name': 'THOTH_CEPH_BUCKET', 'valueFrom': {'configMapKeyRef': {'key': 'bucket-name', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_BUCKET_PREFIX', 'valueFrom': {'configMapKeyRef': {'key': 'bucket-prefix', 'name': 'ceph'}}}, {'name': 'THOTH_S3_ENDPOINT_URL', 'valueFrom': {'configMapKeyRef': {'key': 'host', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_KEY_ID', 'valueFrom': {'secretKeyRef': {'key': 'key-id', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_SECRET_KEY', 'valueFrom': {'secretKeyRef': {'key': 'secret-key', 'name': 'ceph'}}}], 'image': 'workflow-helpers', 'name': 'graph-schema-update-job', 'resources': {'limits': {'cpu': '1', 'memory': '512Mi'}, 'requests': {'cpu': '1', 'memory': '256Mi'}}}], 'restartPolicy': 'OnFailure'}}}}], 'parameters': [{'name': 'THOTH_SCHEMA_UPDATE_JOB_ID', 'displayName': 'A unique identifier of job', 'description': 'A unique identifier of job', 'value': 'graph-update-schema-7f56a52c400a58e992985a33c2bce4d5', 'required': True}]}, 'graph_update_schema_template': {'apiVersion': 'batch/v1', 'kind': 'Job', 'metadata': {'labels': {'app': 'thoth', 'component': 'graph-schema-update-job', 'mark': 'cleanup'}, 'name': 'graph-update-schema-7f56a52c400a58e992985a33c2bce4d5'}, 'spec': {'backoffLimit': 7, 'template': {'metadata': {'labels': {'app': 'thoth', 'component': 'graph-schema-update-job', 'mark': 'cleanup'}}, 'spec': {'containers': [{'env': [{'name': 'THOTH_WORKFLOW_TASK', 'value': 'graph_schema_update'}, {'name': 'THOTH_DEPLOYMENT_NAME', 'valueFrom': {'configMapKeyRef': {'key': 'deployment-name', 'name': 'thoth'}}}, {'name': 'KNOWLEDGE_GRAPH_HOST', 'valueFrom': {'configMapKeyRef': {'key': 'postgresql-host', 'name': 'thoth'}}}, {'name': 'KNOWLEDGE_GRAPH_PORT', 'value': '5432'}, {'name': 'KNOWLEDGE_GRAPH_SSL_DISABLED', 'value': '1'}, {'name': 'KNOWLEDGE_GRAPH_USER', 'valueFrom': {'secretKeyRef': {'key': 'database-user', 'name': 'postgresql'}}}, {'name': 'KNOWLEDGE_GRAPH_PASSWORD', 'valueFrom': {'secretKeyRef': {'key': 'database-password', 'name': 'postgresql'}}}, {'name': 'KNOWLEDGE_GRAPH_DATABASE', 'valueFrom': {'secretKeyRef': {'key': 'database-name', 'name': 'postgresql'}}}, {'name': 'THOTH_CEPH_BUCKET', 'valueFrom': {'configMapKeyRef': {'key': 'bucket-name', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_BUCKET_PREFIX', 'valueFrom': {'configMapKeyRef': {'key': 'bucket-prefix', 'name': 'ceph'}}}, {'name': 'THOTH_S3_ENDPOINT_URL', 'valueFrom': {'configMapKeyRef': {'key': 'host', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_KEY_ID', 'valueFrom': {'secretKeyRef': {'key': 'key-id', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_SECRET_KEY', 'valueFrom': {'secretKeyRef': {'key': 'secret-key', 'name': 'ceph'}}}], 'image': 'workflow-helpers', 'name': 'graph-schema-update-job', 'resources': {'limits': {'cpu': '1', 'memory': '512Mi'}, 'requests': {'cpu': '1', 'memory': '256Mi'}}}], 'restartPolicy': 'OnFailure'}}}}}}, {'filename': 'openshift/dynamic/client.py', 'abs_path': '/opt/app-root/lib64/python3.8/site-packages/openshift/dynamic/client.py', 'function': 'create', 'module': 'openshift.dynamic.client', 'lineno': 101, 'pre_context': [' def create(self, resource, body=None, namespace=None, **kwargs):', ' body = self.serialize_body(body)', ' if resource.namespaced:', ' namespace = self.ensure_namespace(resource, namespace, body)', ' path = resource.path(namespace=namespace)'], 'context_line': \" return self.request('post', path, body=body, **kwargs)\", 'post_context': ['', ' def delete(self, resource, name=None, namespace=None, label_selector=None, field_selector=None, **kwargs):', ' if not (name or label_selector or field_selector):', ' raise ValueError(\"At least one of name|label_selector|field_selector is required\")', ' if resource.namespaced and not (label_selector or field_selector or namespace):'], 'vars': {'self': <openshift.dynamic.client.DynamicClient object at 0x7f4e826c1c70>, 'resource': <Resource(batch/v1/jobs)>, 'body': {'apiVersion': 'batch/v1', 'kind': 'Job', 'metadata': {'labels': {'app': 'thoth', 'component': 'graph-schema-update-job', 'mark': 'cleanup'}, 'name': 'graph-update-schema-7f56a52c400a58e992985a33c2bce4d5'}, 'spec': {'backoffLimit': 7, 'template': {'metadata': {'labels': {'app': 'thoth', 'component': 'graph-schema-update-job', 'mark': 'cleanup'}}, 'spec': {'containers': [{'env': [{'name': 'THOTH_WORKFLOW_TASK', 'value': 'graph_schema_update'}, {'name': 'THOTH_DEPLOYMENT_NAME', 'valueFrom': {'configMapKeyRef': {'key': 'deployment-name', 'name': 'thoth'}}}, {'name': 'KNOWLEDGE_GRAPH_HOST', 'valueFrom': {'configMapKeyRef': {'key': 'postgresql-host', 'name': 'thoth'}}}, {'name': 'KNOWLEDGE_GRAPH_PORT', 'value': '5432'}, {'name': 'KNOWLEDGE_GRAPH_SSL_DISABLED', 'value': '1'}, {'name': 'KNOWLEDGE_GRAPH_USER', 'valueFrom': {'secretKeyRef': {'key': 'database-user', 'name': 'postgresql'}}}, {'name': 'KNOWLEDGE_GRAPH_PASSWORD', 'valueFrom': {'secretKeyRef': {'key': 'database-password', 'name': 'postgresql'}}}, {'name': 'KNOWLEDGE_GRAPH_DATABASE', 'valueFrom': {'secretKeyRef': {'key': 'database-name', 'name': 'postgresql'}}}, {'name': 'THOTH_CEPH_BUCKET', 'valueFrom': {'configMapKeyRef': {'key': 'bucket-name', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_BUCKET_PREFIX', 'valueFrom': {'configMapKeyRef': {'key': 'bucket-prefix', 'name': 'ceph'}}}, {'name': 'THOTH_S3_ENDPOINT_URL', 'valueFrom': {'configMapKeyRef': {'key': 'host', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_KEY_ID', 'valueFrom': {'secretKeyRef': {'key': 'key-id', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_SECRET_KEY', 'valueFrom': {'secretKeyRef': {'key': 'secret-key', 'name': 'ceph'}}}], 'image': 'workflow-helpers', 'name': 'graph-schema-update-job', 'resources': {'limits': {'cpu': '1', 'memory': '512Mi'}, 'requests': {'cpu': '1', 'memory': '256Mi'}}}], 'restartPolicy': 'OnFailure'}}}}, 'namespace': 'thoth-test-core', 'kwargs': {}, 'path': '/apis/batch/v1/namespaces/thoth-test-core/jobs'}}, {'filename': 'openshift/dynamic/client.py', 'abs_path': '/opt/app-root/lib64/python3.8/site-packages/openshift/dynamic/client.py', 'function': 'inner', 'module': 'openshift.dynamic.client', 'lineno': 44, 'pre_context': [\" serialize_response = kwargs.pop('serialize', True)\", \" serializer = kwargs.pop('serializer', ResourceInstance)\", ' try:', ' resp = func(self, *args, **kwargs)', ' except ApiException as e:'], 'context_line': ' raise api_exception(e)', 'post_context': [' if serialize_response:', ' try:', ' if six.PY2:', ' return serializer(self, json.loads(resp.data))', \" return serializer(self, json.loads(resp.data.decode('utf8')))\"], 'vars': {'self': <openshift.dynamic.client.DynamicClient object at 0x7f4e826c1c70>, 'args': ('post', '/apis/batch/v1/namespaces/thoth-test-core/jobs'), 'kwargs': {'body': {'apiVersion': 'batch/v1', 'kind': 'Job', 'metadata': {'labels': {'app': 'thoth', 'component': 'graph-schema-update-job', 'mark': 'cleanup'}, 'name': 'graph-update-schema-7f56a52c400a58e992985a33c2bce4d5'}, 'spec': {'backoffLimit': 7, 'template': {'metadata': {'labels': {'app': 'thoth', 'component': 'graph-schema-update-job', 'mark': 'cleanup'}}, 'spec': {'containers': [{'env': [{'name': 'THOTH_WORKFLOW_TASK', 'value': 'graph_schema_update'}, {'name': 'THOTH_DEPLOYMENT_NAME', 'valueFrom': {'configMapKeyRef': {'key': 'deployment-name', 'name': 'thoth'}}}, {'name': 'KNOWLEDGE_GRAPH_HOST', 'valueFrom': {'configMapKeyRef': {'key': 'postgresql-host', 'name': 'thoth'}}}, {'name': 'KNOWLEDGE_GRAPH_PORT', 'value': '5432'}, {'name': 'KNOWLEDGE_GRAPH_SSL_DISABLED', 'value': '1'}, {'name': 'KNOWLEDGE_GRAPH_USER', 'valueFrom': {'secretKeyRef': {'key': 'database-user', 'name': 'postgresql'}}}, {'name': 'KNOWLEDGE_GRAPH_PASSWORD', 'valueFrom': {'secretKeyRef': {'key': 'database-password', 'name': 'postgresql'}}}, {'name': 'KNOWLEDGE_GRAPH_DATABASE', 'valueFrom': {'secretKeyRef': {'key': 'database-name', 'name': 'postgresql'}}}, {'name': 'THOTH_CEPH_BUCKET', 'valueFrom': {'configMapKeyRef': {'key': 'bucket-name', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_BUCKET_PREFIX', 'valueFrom': {'configMapKeyRef': {'key': 'bucket-prefix', 'name': 'ceph'}}}, {'name': 'THOTH_S3_ENDPOINT_URL', 'valueFrom': {'configMapKeyRef': {'key': 'host', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_KEY_ID', 'valueFrom': {'secretKeyRef': {'key': 'key-id', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_SECRET_KEY', 'valueFrom': {'secretKeyRef': {'key': 'secret-key', 'name': 'ceph'}}}], 'image': 'workflow-helpers', 'name': 'graph-schema-update-job', 'resources': {'limits': {'cpu': '1', 'memory': '512Mi'}, 'requests': {'cpu': '1', 'memory': '256Mi'}}}], 'restartPolicy': 'OnFailure'}}}}}, 'serialize_response': True, 'serializer': <class 'openshift.dynamic.resource.ResourceInstance'>, 'func': <function DynamicClient.request at 0x7f4e86f494c0>}}]}}]}, 'logger': 'thoth.management_api.openapi_server', 'logentry': {'message': 'Exception on /api/v1/graph/initialize [PUT]', 'params': ()}, 'extra': {'asctime': '2021-01-14 16:45:26,011'}, 'event_id': '92d2f914f6214dbbb420e91f4e7399fd', 'timestamp': datetime.datetime(2021, 1, 14, 16, 45, 26, 19047), 'breadcrumbs': {'values': [{'type': 'http', 'category': 'httplib', 'data': {'method': 'GET', 'url': 'https://172.30.0.1/apis/template.openshift.io/v1', 'status_code': 200, 'reason': 'OK'}, 'timestamp': datetime.datetime(2021, 1, 14, 16, 45, 25, 718394)}, {'type': 'http', 'category': 'httplib', 'data': {'method': 'GET', 'url': 'https://172.30.0.1/apis/template.openshift.io/v1/namespaces/thoth-test-core/templates?labelSelector=template%3Dgraph-schema-update-job', 'status_code': 200, 'reason': 'OK'}, 'timestamp': datetime.datetime(2021, 1, 14, 16, 45, 25, 733828)}, {'type': 'http', 'category': 'httplib', 'data': {'method': 'POST', 'url': 'https://172.30.0.1/apis/template.openshift.io/v1/namespaces/thoth-test-core/processedtemplates', 'status_code': 201, 'reason': 'Created'}, 'timestamp': datetime.datetime(2021, 1, 14, 16, 45, 25, 761455)}, {'type': 'http', 'category': 'httplib', 'data': {'method': 'GET', 'url': 'https://172.30.0.1/apis/batch/v1', 'status_code': 200, 'reason': 'OK'}, 'timestamp': datetime.datetime(2021, 1, 14, 16, 45, 25, 766532)}, {'type': 'http', 'category': 'httplib', 'data': {'method': 'POST', 'url': 'https://172.30.0.1/apis/batch/v1/namespaces/thoth-test-core/jobs', 'status_code': 403, 'reason': 'Forbidden'}, 'timestamp': datetime.datetime(2021, 1, 14, 16, 45, 25, 773598)}]}, 'transaction': '/api/v1.thoth_management_api_api_v1_initialize_schema', 'contexts': {'trace': {'trace_id': 'df56345840f44f009335865766b30f2f', 'span_id': 'a032e586b4eb3cf8', 'parent_span_id': None, 'op': 'http.server', 'description': None}, 'runtime': {'name': 'CPython', 'version': '3.8.3', 'build': '3.8.3 (default, Aug 18 2020, 08:56:04) \\n[GCC 8.3.1 20191121 (Red Hat 8.3.1-5)]'}}})"} {"remote": "172.20.8.1", "date": "[14/Jan/2021:16:45:26 +0000]", "status": "500", "response_length": 177, "referer": "https://management.test.thoth-station.ninja/api/v1/ui/", "user_agent": "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/87.0.4280.141 Safari/537.36", "request_method": "PUT", "url_path": "/api/v1/graph/initialize", "protocol": "HTTP/1.1", "request_time": 0} ``` **Additional context** Add any other context about the problem here.
1.0
User system:serviceaccount:thoth-test-core:management-api cannot create resource jobs in API group batch in the namespace thoth-test-core - **Describe the bug** User system:serviceaccount:thoth-test-core:management-api cannot create resource jobs in API group batch in the namespace thoth-test-core **To Reproduce** Steps to reproduce the behavior: 1. Use `/graph/initialize` in management-API in test envioronment **Expected behavior** A clear and concise description of what you expected to happen. **Screenshots** ``` {"name": "thoth.management_api.openapi_server", "levelname": "ERROR", "module": "app", "lineno": 1891, "funcname": "log_exception", "created": 1610642726.0111866, "asctime": "2021-01-14 16:45:26,011", "msecs": 11.186599731445312, "relative_created": 58228.846073150635, "process": 23, "message": "Exception on /api/v1/graph/initialize [PUT]\nTraceback (most recent call last):\n File \"/opt/app-root/lib64/python3.8/site-packages/openshift/dynamic/client.py\", line 42, in inner\n resp = func(self, *args, **kwargs)\n File \"/opt/app-root/lib64/python3.8/site-packages/openshift/dynamic/client.py\", line 235, in request\n return self.client.call_api(\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py\", line 340, in call_api\n return self.__call_api(resource_path, method,\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py\", line 172, in __call_api\n response_data = self.request(\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py\", line 382, in request\n return self.rest_client.POST(url,\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/rest.py\", line 272, in POST\n return self.request(\"POST\", url,\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/rest.py\", line 231, in request\n raise ApiException(http_resp=r)\nkubernetes.client.rest.ApiException: (403)\nReason: Forbidden\nHTTP response headers: HTTPHeaderDict({'Audit-Id': '49583870-4187-49f8-9539-aef793ac349e', 'Cache-Control': 'no-cache, private', 'Content-Type': 'application/json', 'X-Content-Type-Options': 'nosniff', 'Date': 'Thu, 14 Jan 2021 16:45:25 GMT', 'Content-Length': '334'})\nHTTP response body: b'{\"kind\":\"Status\",\"apiVersion\":\"v1\",\"metadata\":{},\"status\":\"Failure\",\"message\":\"jobs.batch is forbidden: User \\\\\"system:serviceaccount:thoth-test-core:management-api\\\\\" cannot create resource \\\\\"jobs\\\\\" in API group \\\\\"batch\\\\\" in the namespace \\\\\"thoth-test-core\\\\\"\",\"reason\":\"Forbidden\",\"details\":{\"group\":\"batch\",\"kind\":\"jobs\"},\"code\":403}\\n'\n\n\nDuring handling of the above exception, another exception occurred:\n\nTraceback (most recent call last):\n File \"/opt/app-root/lib64/python3.8/site-packages/flask/app.py\", line 2447, in wsgi_app\n response = self.full_dispatch_request()\n File \"/opt/app-root/lib64/python3.8/site-packages/flask/app.py\", line 1952, in full_dispatch_request\n rv = self.handle_user_exception(e)\n File \"/opt/app-root/lib64/python3.8/site-packages/flask/app.py\", line 1821, in handle_user_exception\n reraise(exc_type, exc_value, tb)\n File \"/opt/app-root/lib64/python3.8/site-packages/flask/_compat.py\", line 39, in reraise\n raise value\n File \"/opt/app-root/lib64/python3.8/site-packages/flask/app.py\", line 1950, in full_dispatch_request\n rv = self.dispatch_request()\n File \"/opt/app-root/lib64/python3.8/site-packages/flask/app.py\", line 1936, in dispatch_request\n return self.view_functions[rule.endpoint](**req.view_args)\n File \"/opt/app-root/lib64/python3.8/site-packages/connexion/decorators/decorator.py\", line 48, in wrapper\n response = function(request)\n File \"/opt/app-root/lib64/python3.8/site-packages/connexion/decorators/uri_parsing.py\", line 144, in wrapper\n response = function(request)\n File \"/opt/app-root/lib64/python3.8/site-packages/connexion/decorators/validation.py\", line 384, in wrapper\n return function(request)\n File \"/opt/app-root/lib64/python3.8/site-packages/connexion/decorators/parameter.py\", line 121, in wrapper\n return function(**kwargs)\n File \"/opt/app-root/src/thoth/management_api/api_v1.py\", line 348, in initialize_schema\n job_id = _OPENSHIFT.schedule_graph_schema_update()\n File \"/opt/app-root/lib64/python3.8/site-packages/thoth/common/openshift.py\", line 791, in schedule_graph_schema_update\n response = self.ocp_client.resources.get(\n File \"/opt/app-root/lib64/python3.8/site-packages/openshift/dynamic/client.py\", line 101, in create\n return self.request('post', path, body=body, **kwargs)\n File \"/opt/app-root/lib64/python3.8/site-packages/openshift/dynamic/client.py\", line 44, in inner\n raise api_exception(e)\nopenshift.dynamic.exceptions.ForbiddenError: 403\nReason: Forbidden\nHTTP response headers: HTTPHeaderDict({'Audit-Id': '49583870-4187-49f8-9539-aef793ac349e', 'Cache-Control': 'no-cache, private', 'Content-Type': 'application/json', 'X-Content-Type-Options': 'nosniff', 'Date': 'Thu, 14 Jan 2021 16:45:25 GMT', 'Content-Length': '334'})\nHTTP response body: b'{\"kind\":\"Status\",\"apiVersion\":\"v1\",\"metadata\":{},\"status\":\"Failure\",\"message\":\"jobs.batch is forbidden: User \\\\\"system:serviceaccount:thoth-test-core:management-api\\\\\" cannot create resource \\\\\"jobs\\\\\" in API group \\\\\"batch\\\\\" in the namespace \\\\\"thoth-test-core\\\\\"\",\"reason\":\"Forbidden\",\"details\":{\"group\":\"batch\",\"kind\":\"jobs\"},\"code\":403}\\n'\nOriginal traceback: \n File \"/opt/app-root/lib64/python3.8/site-packages/openshift/dynamic/client.py\", line 42, in inner\n resp = func(self, *args, **kwargs)\n\n File \"/opt/app-root/lib64/python3.8/site-packages/openshift/dynamic/client.py\", line 235, in request\n return self.client.call_api(\n\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py\", line 340, in call_api\n return self.__call_api(resource_path, method,\n\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py\", line 172, in __call_api\n response_data = self.request(\n\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py\", line 382, in request\n return self.rest_client.POST(url,\n\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/rest.py\", line 272, in POST\n return self.request(\"POST\", url,\n\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/rest.py\", line 231, in request\n raise ApiException(http_resp=r)\n"} {"name": "sentry_sdk.errors", "levelname": "INFO", "module": "scope", "lineno": 345, "funcname": "_drop", "created": 1610642726.0191312, "asctime": "2021-01-14 16:45:26,019", "msecs": 19.131183624267578, "relative_created": 58236.79065704346, "process": 23, "message": "event processor (<function DedupeIntegration.setup_once.<locals>.processor at 0x7f4e8280b3a0>) dropped event ({'level': 'error', 'exception': {'values': [{'module': 'kubernetes.client.rest', 'type': 'ApiException', 'value': '(403)\\nReason: Forbidden\\nHTTP response headers: HTTPHeaderDict({\\'Audit-Id\\': \\'49583870-4187-49f8-9539-aef793ac349e\\', \\'Cache-Control\\': \\'no-cache, private\\', \\'Content-Type\\': \\'application/json\\', \\'X-Content-Type-Options\\': \\'nosniff\\', \\'Date\\': \\'Thu, 14 Jan 2021 16:45:25 GMT\\', \\'Content-Length\\': \\'334\\'})\\nHTTP response body: b\\'{\"kind\":\"Status\",\"apiVersion\":\"v1\",\"metadata\":{},\"status\":\"Failure\",\"message\":\"jobs.batch is forbidden: User \\\\\\\\\"system:serviceaccount:thoth-test-core:management-api\\\\\\\\\" cannot create resource \\\\\\\\\"jobs\\\\\\\\\" in API group \\\\\\\\\"batch\\\\\\\\\" in the namespace \\\\\\\\\"thoth-test-core\\\\\\\\\"\",\"reason\":\"Forbidden\",\"details\":{\"group\":\"batch\",\"kind\":\"jobs\"},\"code\":403}\\\\n\\'\\n', 'mechanism': {'type': 'logging', 'handled': True}, 'stacktrace': {'frames': [{'filename': 'openshift/dynamic/client.py', 'abs_path': '/opt/app-root/lib64/python3.8/site-packages/openshift/dynamic/client.py', 'function': 'inner', 'module': 'openshift.dynamic.client', 'lineno': 42, 'pre_context': [' \"\"\" Handles parsing response structure and translating API Exceptions \"\"\"', ' def inner(self, *args, **kwargs):', \" serialize_response = kwargs.pop('serialize', True)\", \" serializer = kwargs.pop('serializer', ResourceInstance)\", ' try:'], 'context_line': ' resp = func(self, *args, **kwargs)', 'post_context': [' except ApiException as e:', ' raise api_exception(e)', ' if serialize_response:', ' try:', ' if six.PY2:'], 'vars': {'self': <openshift.dynamic.client.DynamicClient object at 0x7f4e826c1c70>, 'args': ('post', '/apis/batch/v1/namespaces/thoth-test-core/jobs'), 'kwargs': {'body': {'apiVersion': 'batch/v1', 'kind': 'Job', 'metadata': {'labels': {'app': 'thoth', 'component': 'graph-schema-update-job', 'mark': 'cleanup'}, 'name': 'graph-update-schema-7f56a52c400a58e992985a33c2bce4d5'}, 'spec': {'backoffLimit': 7, 'template': {'metadata': {'labels': {'app': 'thoth', 'component': 'graph-schema-update-job', 'mark': 'cleanup'}}, 'spec': {'containers': [{'env': [{'name': 'THOTH_WORKFLOW_TASK', 'value': 'graph_schema_update'}, {'name': 'THOTH_DEPLOYMENT_NAME', 'valueFrom': {'configMapKeyRef': {'key': 'deployment-name', 'name': 'thoth'}}}, {'name': 'KNOWLEDGE_GRAPH_HOST', 'valueFrom': {'configMapKeyRef': {'key': 'postgresql-host', 'name': 'thoth'}}}, {'name': 'KNOWLEDGE_GRAPH_PORT', 'value': '5432'}, {'name': 'KNOWLEDGE_GRAPH_SSL_DISABLED', 'value': '1'}, {'name': 'KNOWLEDGE_GRAPH_USER', 'valueFrom': {'secretKeyRef': {'key': 'database-user', 'name': 'postgresql'}}}, {'name': 'KNOWLEDGE_GRAPH_PASSWORD', 'valueFrom': {'secretKeyRef': {'key': 'database-password', 'name': 'postgresql'}}}, {'name': 'KNOWLEDGE_GRAPH_DATABASE', 'valueFrom': {'secretKeyRef': {'key': 'database-name', 'name': 'postgresql'}}}, {'name': 'THOTH_CEPH_BUCKET', 'valueFrom': {'configMapKeyRef': {'key': 'bucket-name', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_BUCKET_PREFIX', 'valueFrom': {'configMapKeyRef': {'key': 'bucket-prefix', 'name': 'ceph'}}}, {'name': 'THOTH_S3_ENDPOINT_URL', 'valueFrom': {'configMapKeyRef': {'key': 'host', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_KEY_ID', 'valueFrom': {'secretKeyRef': {'key': 'key-id', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_SECRET_KEY', 'valueFrom': {'secretKeyRef': {'key': 'secret-key', 'name': 'ceph'}}}], 'image': 'workflow-helpers', 'name': 'graph-schema-update-job', 'resources': {'limits': {'cpu': '1', 'memory': '512Mi'}, 'requests': {'cpu': '1', 'memory': '256Mi'}}}], 'restartPolicy': 'OnFailure'}}}}}, 'serialize_response': True, 'serializer': <class 'openshift.dynamic.resource.ResourceInstance'>, 'func': <function DynamicClient.request at 0x7f4e86f494c0>}}, {'filename': 'openshift/dynamic/client.py', 'abs_path': '/opt/app-root/lib64/python3.8/site-packages/openshift/dynamic/client.py', 'function': 'request', 'module': 'openshift.dynamic.client', 'lineno': 235, 'pre_context': [\" header_params['Content-Type'] = self.client.select_header_content_type(['*/*'])\", '', ' # Authentication setting', \" auth_settings = ['BearerToken']\", ''], 'context_line': ' return self.client.call_api(', 'post_context': [' path,', ' method.upper(),', ' path_params,', ' query_params,', ' header_params,'], 'vars': {'self': <openshift.dynamic.client.DynamicClient object at 0x7f4e826c1c70>, 'method': 'post', 'path': '/apis/batch/v1/namespaces/thoth-test-core/jobs', 'body': {'apiVersion': 'batch/v1', 'kind': 'Job', 'metadata': {'labels': {'app': 'thoth', 'component': 'graph-schema-update-job', 'mark': 'cleanup'}, 'name': 'graph-update-schema-7f56a52c400a58e992985a33c2bce4d5'}, 'spec': {'backoffLimit': 7, 'template': {'metadata': {'labels': {'app': 'thoth', 'component': 'graph-schema-update-job', 'mark': 'cleanup'}}, 'spec': {'containers': [{'env': [{'name': 'THOTH_WORKFLOW_TASK', 'value': 'graph_schema_update'}, {'name': 'THOTH_DEPLOYMENT_NAME', 'valueFrom': {'configMapKeyRef': {'key': 'deployment-name', 'name': 'thoth'}}}, {'name': 'KNOWLEDGE_GRAPH_HOST', 'valueFrom': {'configMapKeyRef': {'key': 'postgresql-host', 'name': 'thoth'}}}, {'name': 'KNOWLEDGE_GRAPH_PORT', 'value': '5432'}, {'name': 'KNOWLEDGE_GRAPH_SSL_DISABLED', 'value': '1'}, {'name': 'KNOWLEDGE_GRAPH_USER', 'valueFrom': {'secretKeyRef': {'key': 'database-user', 'name': 'postgresql'}}}, {'name': 'KNOWLEDGE_GRAPH_PASSWORD', 'valueFrom': {'secretKeyRef': {'key': 'database-password', 'name': 'postgresql'}}}, {'name': 'KNOWLEDGE_GRAPH_DATABASE', 'valueFrom': {'secretKeyRef': {'key': 'database-name', 'name': 'postgresql'}}}, {'name': 'THOTH_CEPH_BUCKET', 'valueFrom': {'configMapKeyRef': {'key': 'bucket-name', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_BUCKET_PREFIX', 'valueFrom': {'configMapKeyRef': {'key': 'bucket-prefix', 'name': 'ceph'}}}, {'name': 'THOTH_S3_ENDPOINT_URL', 'valueFrom': {'configMapKeyRef': {'key': 'host', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_KEY_ID', 'valueFrom': {'secretKeyRef': {'key': 'key-id', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_SECRET_KEY', 'valueFrom': {'secretKeyRef': {'key': 'secret-key', 'name': 'ceph'}}}], 'image': 'workflow-helpers', 'name': 'graph-schema-update-job', 'resources': {'limits': {'cpu': '1', 'memory': '512Mi'}, 'requests': {'cpu': '1', 'memory': '256Mi'}}}], 'restartPolicy': 'OnFailure'}}}}, 'params': {}, 'path_params': {}, 'query_params': [], 'header_params': {'Accept': 'application/json', 'Content-Type': 'application/json', 'User-Agent': 'OpenAPI-Generator/11.0.0/python'}, 'form_params': [], 'local_var_files': {}, 'auth_settings': ['BearerToken']}}, {'filename': 'kubernetes/client/api_client.py', 'abs_path': '/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py', 'function': 'call_api', 'module': 'kubernetes.client.api_client', 'lineno': 340, 'pre_context': [' The method will return the request thread.', ' If parameter async_req is False or missing,', ' then the method will return the response directly.', ' \"\"\"', ' if not async_req:'], 'context_line': ' return self.__call_api(resource_path, method,', 'post_context': [' path_params, query_params, header_params,', ' body, post_params, files,', ' response_type, auth_settings,', ' _return_http_data_only, collection_formats,', ' _preload_content, _request_timeout)'], 'vars': {'self': <kubernetes.client.api_client.ApiClient object at 0x7f4e826c1730>, 'resource_path': '/apis/batch/v1/namespaces/thoth-test-core/jobs', 'method': 'POST', 'path_params': {}, 'query_params': [], 'header_params': {'Accept': 'application/json', 'Content-Type': 'application/json', 'User-Agent': 'OpenAPI-Generator/11.0.0/python'}, 'body': {'apiVersion': 'batch/v1', 'kind': 'Job', 'metadata': {'labels': {'app': 'thoth', 'component': 'graph-schema-update-job', 'mark': 'cleanup'}, 'name': 'graph-update-schema-7f56a52c400a58e992985a33c2bce4d5'}, 'spec': {'backoffLimit': 7, 'template': {'metadata': {'labels': {'app': 'thoth', 'component': 'graph-schema-update-job', 'mark': 'cleanup'}}, 'spec': {'containers': [{'env': [{'name': 'THOTH_WORKFLOW_TASK', 'value': 'graph_schema_update'}, {'name': 'THOTH_DEPLOYMENT_NAME', 'valueFrom': {'configMapKeyRef': {'key': 'deployment-name', 'name': 'thoth'}}}, {'name': 'KNOWLEDGE_GRAPH_HOST', 'valueFrom': {'configMapKeyRef': {'key': 'postgresql-host', 'name': 'thoth'}}}, {'name': 'KNOWLEDGE_GRAPH_PORT', 'value': '5432'}, {'name': 'KNOWLEDGE_GRAPH_SSL_DISABLED', 'value': '1'}, {'name': 'KNOWLEDGE_GRAPH_USER', 'valueFrom': {'secretKeyRef': {'key': 'database-user', 'name': 'postgresql'}}}, {'name': 'KNOWLEDGE_GRAPH_PASSWORD', 'valueFrom': {'secretKeyRef': {'key': 'database-password', 'name': 'postgresql'}}}, {'name': 'KNOWLEDGE_GRAPH_DATABASE', 'valueFrom': {'secretKeyRef': {'key': 'database-name', 'name': 'postgresql'}}}, {'name': 'THOTH_CEPH_BUCKET', 'valueFrom': {'configMapKeyRef': {'key': 'bucket-name', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_BUCKET_PREFIX', 'valueFrom': {'configMapKeyRef': {'key': 'bucket-prefix', 'name': 'ceph'}}}, {'name': 'THOTH_S3_ENDPOINT_URL', 'valueFrom': {'configMapKeyRef': {'key': 'host', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_KEY_ID', 'valueFrom': {'secretKeyRef': {'key': 'key-id', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_SECRET_KEY', 'valueFrom': {'secretKeyRef': {'key': 'secret-key', 'name': 'ceph'}}}], 'image': 'workflow-helpers', 'name': 'graph-schema-update-job', 'resources': {'limits': {'cpu': '1', 'memory': '512Mi'}, 'requests': {'cpu': '1', 'memory': '256Mi'}}}], 'restartPolicy': 'OnFailure'}}}}, 'post_params': [], 'files': {}, 'response_type': None, 'auth_settings': ['BearerToken'], 'async_req': None, '_return_http_data_only': True, 'collection_formats': None, '_preload_content': False, '_request_timeout': None}}, {'filename': 'kubernetes/client/api_client.py', 'abs_path': '/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py', 'function': '__call_api', 'module': 'kubernetes.client.api_client', 'lineno': 172, 'pre_context': ['', ' # request url', ' url = self.configuration.host + resource_path', '', ' # perform request and return response'], 'context_line': ' response_data = self.request(', 'post_context': [' method, url, query_params=query_params, headers=header_params,', ' post_params=post_params, body=body,', ' _preload_content=_preload_content,', ' _request_timeout=_request_timeout)', ''], 'vars': {'self': <kubernetes.client.api_client.ApiClient object at 0x7f4e826c1730>, 'resource_path': '/apis/batch/v1/namespaces/thoth-test-core/jobs', 'method': 'POST', 'path_params': {}, 'query_params': [], 'header_params': {'Accept': 'application/json', 'Content-Type': 'application/json', 'User-Agent': 'OpenAPI-Generator/11.0.0/python', 'authorization': 'bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IlhaVlc3YnFrZ1BaSllHSkpoNnVLenpTQlE1X3lrcEJzVndHWUdJclVDVEUifQ.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJ0aG90aC10ZXN0LWNvcmUiLCJrdWJlcm5ldGVzLmlvL3NlcnZpY2VhY2NvdW50L3NlY3JldC5uYW1lIjoibWFuYWdlbWVudC1hcGktdG9rZW4tNXRqcDkiLCJrdWJlcm5ldGVzLmlvL3NlcnZpY2VhY2NvdW50L3NlcnZpY2UtYWNjb3VudC5uYW1lIjoibWFuYWdlbWVudC1hcGkiLCJrdWJlcm5ldGVzLmlvL3NlcnZpY2VhY2NvdW50L3NlcnZpY2UtYWNjb3VudC51aWQiOiIyMTZlODJlZi05N2RhLTRiNGYtODk4Mi02ZDQzOWEwNjRmNjIiLCJzdWIiOiJzeXN0ZW06c2VydmljZWFjY291bnQ6dGhvdGgtdGVzdC1jb3JlOm1hbmFnZW1lbnQtYXBpIn0.KLDhPJFSnfyXXY5tV5788iPFbYUQ48TXZW2GEi6N-8QZ8qVkcXmubhlvlO8_P93Me9-CS8iAgsUtRH8RYBUOLssZ_R6ioZaOH3Qe7TQhMwbgu6cqFp0wYMRJnvY-RrXHLZBRjUrstgKmEeE6UsclRxy88TRGpp2XYjlTSgW-XHYUj_EMqTePkQXrQZ8MM7EwQiAvwRv653dq3mSmNt4iwF5ewN7C6lPm1mUpw2Kg-kn1ojcP_G1d9gr1u41FUnPD_BDyWfa0Y2fS1fmR33QL7whdmvC26t6qANbHyMlz-jzorH9gfvgdTzhEjXE9-FHu54zj63y6GJM1kJy0vGGKfw'}, 'body': {'apiVersion': 'batch/v1', 'kind': 'Job', 'metadata': {'labels': {'app': 'thoth', 'component': 'graph-schema-update-job', 'mark': 'cleanup'}, 'name': 'graph-update-schema-7f56a52c400a58e992985a33c2bce4d5'}, 'spec': {'backoffLimit': 7, 'template': {'metadata': {'labels': {'app': 'thoth', 'component': 'graph-schema-update-job', 'mark': 'cleanup'}}, 'spec': {'containers': [{'env': [{'name': 'THOTH_WORKFLOW_TASK', 'value': 'graph_schema_update'}, {'name': 'THOTH_DEPLOYMENT_NAME', 'valueFrom': {'configMapKeyRef': {'key': 'deployment-name', 'name': 'thoth'}}}, {'name': 'KNOWLEDGE_GRAPH_HOST', 'valueFrom': {'configMapKeyRef': {'key': 'postgresql-host', 'name': 'thoth'}}}, {'name': 'KNOWLEDGE_GRAPH_PORT', 'value': '5432'}, {'name': 'KNOWLEDGE_GRAPH_SSL_DISABLED', 'value': '1'}, {'name': 'KNOWLEDGE_GRAPH_USER', 'valueFrom': {'secretKeyRef': {'key': 'database-user', 'name': 'postgresql'}}}, {'name': 'KNOWLEDGE_GRAPH_PASSWORD', 'valueFrom': {'secretKeyRef': {'key': 'database-password', 'name': 'postgresql'}}}, {'name': 'KNOWLEDGE_GRAPH_DATABASE', 'valueFrom': {'secretKeyRef': {'key': 'database-name', 'name': 'postgresql'}}}, {'name': 'THOTH_CEPH_BUCKET', 'valueFrom': {'configMapKeyRef': {'key': 'bucket-name', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_BUCKET_PREFIX', 'valueFrom': {'configMapKeyRef': {'key': 'bucket-prefix', 'name': 'ceph'}}}, {'name': 'THOTH_S3_ENDPOINT_URL', 'valueFrom': {'configMapKeyRef': {'key': 'host', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_KEY_ID', 'valueFrom': {'secretKeyRef': {'key': 'key-id', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_SECRET_KEY', 'valueFrom': {'secretKeyRef': {'key': 'secret-key', 'name': 'ceph'}}}], 'image': 'workflow-helpers', 'name': 'graph-schema-update-job', 'resources': {'limits': {'cpu': '1', 'memory': '512Mi'}, 'requests': {'cpu': '1', 'memory': '256Mi'}}}], 'restartPolicy': 'OnFailure'}}}}, 'post_params': [], 'files': {}, 'response_type': None, 'auth_settings': ['BearerToken'], '_return_http_data_only': True, 'collection_formats': None, '_preload_content': False, '_request_timeout': None, 'config': <kubernetes.client.configuration.Configuration object at 0x7f4e826c1940>, 'url': 'https://172.30.0.1:443/apis/batch/v1/namespaces/thoth-test-core/jobs'}}, {'filename': 'kubernetes/client/api_client.py', 'abs_path': '/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py', 'function': 'request', 'module': 'kubernetes.client.api_client', 'lineno': 382, 'pre_context': [' post_params=post_params,', ' _preload_content=_preload_content,', ' _request_timeout=_request_timeout,', ' body=body)', ' elif method == \"POST\":'], 'context_line': ' return self.rest_client.POST(url,', 'post_context': [' query_params=query_params,', ' headers=headers,', ' post_params=post_params,', ' _preload_content=_preload_content,', ' _request_timeout=_request_timeout,'], 'vars': {'self': <kubernetes.client.api_client.ApiClient object at 0x7f4e826c1730>, 'method': 'POST', 'url': 'https://172.30.0.1:443/apis/batch/v1/namespaces/thoth-test-core/jobs', 'query_params': [], 'headers': {'Accept': 'application/json', 'Content-Type': 'application/json', 'User-Agent': 'OpenAPI-Generator/11.0.0/python', 'authorization': 'bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IlhaVlc3YnFrZ1BaSllHSkpoNnVLenpTQlE1X3lrcEJzVndHWUdJclVDVEUifQ.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJ0aG90aC10ZXN0LWNvcmUiLCJrdWJlcm5ldGVzLmlvL3NlcnZpY2VhY2NvdW50L3NlY3JldC5uYW1lIjoibWFuYWdlbWVudC1hcGktdG9rZW4tNXRqcDkiLCJrdWJlcm5ldGVzLmlvL3NlcnZpY2VhY2NvdW50L3NlcnZpY2UtYWNjb3VudC5uYW1lIjoibWFuYWdlbWVudC1hcGkiLCJrdWJlcm5ldGVzLmlvL3NlcnZpY2VhY2NvdW50L3NlcnZpY2UtYWNjb3VudC51aWQiOiIyMTZlODJlZi05N2RhLTRiNGYtODk4Mi02ZDQzOWEwNjRmNjIiLCJzdWIiOiJzeXN0ZW06c2VydmljZWFjY291bnQ6dGhvdGgtdGVzdC1jb3JlOm1hbmFnZW1lbnQtYXBpIn0.KLDhPJFSnfyXXY5tV5788iPFbYUQ48TXZW2GEi6N-8QZ8qVkcXmubhlvlO8_P93Me9-CS8iAgsUtRH8RYBUOLssZ_R6ioZaOH3Qe7TQhMwbgu6cqFp0wYMRJnvY-RrXHLZBRjUrstgKmEeE6UsclRxy88TRGpp2XYjlTSgW-XHYUj_EMqTePkQXrQZ8MM7EwQiAvwRv653dq3mSmNt4iwF5ewN7C6lPm1mUpw2Kg-kn1ojcP_G1d9gr1u41FUnPD_BDyWfa0Y2fS1fmR33QL7whdmvC26t6qANbHyMlz-jzorH9gfvgdTzhEjXE9-FHu54zj63y6GJM1kJy0vGGKfw'}, 'post_params': [], 'body': {'apiVersion': 'batch/v1', 'kind': 'Job', 'metadata': {'labels': {'app': 'thoth', 'component': 'graph-schema-update-job', 'mark': 'cleanup'}, 'name': 'graph-update-schema-7f56a52c400a58e992985a33c2bce4d5'}, 'spec': {'backoffLimit': 7, 'template': {'metadata': {'labels': {'app': 'thoth', 'component': 'graph-schema-update-job', 'mark': 'cleanup'}}, 'spec': {'containers': [{'env': [{'name': 'THOTH_WORKFLOW_TASK', 'value': 'graph_schema_update'}, {'name': 'THOTH_DEPLOYMENT_NAME', 'valueFrom': {'configMapKeyRef': {'key': 'deployment-name', 'name': 'thoth'}}}, {'name': 'KNOWLEDGE_GRAPH_HOST', 'valueFrom': {'configMapKeyRef': {'key': 'postgresql-host', 'name': 'thoth'}}}, {'name': 'KNOWLEDGE_GRAPH_PORT', 'value': '5432'}, {'name': 'KNOWLEDGE_GRAPH_SSL_DISABLED', 'value': '1'}, {'name': 'KNOWLEDGE_GRAPH_USER', 'valueFrom': {'secretKeyRef': {'key': 'database-user', 'name': 'postgresql'}}}, {'name': 'KNOWLEDGE_GRAPH_PASSWORD', 'valueFrom': {'secretKeyRef': {'key': 'database-password', 'name': 'postgresql'}}}, {'name': 'KNOWLEDGE_GRAPH_DATABASE', 'valueFrom': {'secretKeyRef': {'key': 'database-name', 'name': 'postgresql'}}}, {'name': 'THOTH_CEPH_BUCKET', 'valueFrom': {'configMapKeyRef': {'key': 'bucket-name', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_BUCKET_PREFIX', 'valueFrom': {'configMapKeyRef': {'key': 'bucket-prefix', 'name': 'ceph'}}}, {'name': 'THOTH_S3_ENDPOINT_URL', 'valueFrom': {'configMapKeyRef': {'key': 'host', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_KEY_ID', 'valueFrom': {'secretKeyRef': {'key': 'key-id', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_SECRET_KEY', 'valueFrom': {'secretKeyRef': {'key': 'secret-key', 'name': 'ceph'}}}], 'image': 'workflow-helpers', 'name': 'graph-schema-update-job', 'resources': {'limits': {'cpu': '1', 'memory': '512Mi'}, 'requests': {'cpu': '1', 'memory': '256Mi'}}}], 'restartPolicy': 'OnFailure'}}}}, '_preload_content': False, '_request_timeout': None}}, {'filename': 'kubernetes/client/rest.py', 'abs_path': '/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/rest.py', 'function': 'POST', 'module': 'kubernetes.client.rest', 'lineno': 272, 'pre_context': [' _request_timeout=_request_timeout,', ' body=body)', '', ' def POST(self, url, headers=None, query_params=None, post_params=None,', ' body=None, _preload_content=True, _request_timeout=None):'], 'context_line': ' return self.request(\"POST\", url,', 'post_context': [' headers=headers,', ' query_params=query_params,', ' post_params=post_params,', ' _preload_content=_preload_content,', ' _request_timeout=_request_timeout,'], 'vars': {'self': <kubernetes.client.rest.RESTClientObject object at 0x7f4e826c17c0>, 'url': 'https://172.30.0.1:443/apis/batch/v1/namespaces/thoth-test-core/jobs', 'headers': {'Accept': 'application/json', 'Content-Type': 'application/json', 'User-Agent': 'OpenAPI-Generator/11.0.0/python', 'authorization': 'bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IlhaVlc3YnFrZ1BaSllHSkpoNnVLenpTQlE1X3lrcEJzVndHWUdJclVDVEUifQ.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJ0aG90aC10ZXN0LWNvcmUiLCJrdWJlcm5ldGVzLmlvL3NlcnZpY2VhY2NvdW50L3NlY3JldC5uYW1lIjoibWFuYWdlbWVudC1hcGktdG9rZW4tNXRqcDkiLCJrdWJlcm5ldGVzLmlvL3NlcnZpY2VhY2NvdW50L3NlcnZpY2UtYWNjb3VudC5uYW1lIjoibWFuYWdlbWVudC1hcGkiLCJrdWJlcm5ldGVzLmlvL3NlcnZpY2VhY2NvdW50L3NlcnZpY2UtYWNjb3VudC51aWQiOiIyMTZlODJlZi05N2RhLTRiNGYtODk4Mi02ZDQzOWEwNjRmNjIiLCJzdWIiOiJzeXN0ZW06c2VydmljZWFjY291bnQ6dGhvdGgtdGVzdC1jb3JlOm1hbmFnZW1lbnQtYXBpIn0.KLDhPJFSnfyXXY5tV5788iPFbYUQ48TXZW2GEi6N-8QZ8qVkcXmubhlvlO8_P93Me9-CS8iAgsUtRH8RYBUOLssZ_R6ioZaOH3Qe7TQhMwbgu6cqFp0wYMRJnvY-RrXHLZBRjUrstgKmEeE6UsclRxy88TRGpp2XYjlTSgW-XHYUj_EMqTePkQXrQZ8MM7EwQiAvwRv653dq3mSmNt4iwF5ewN7C6lPm1mUpw2Kg-kn1ojcP_G1d9gr1u41FUnPD_BDyWfa0Y2fS1fmR33QL7whdmvC26t6qANbHyMlz-jzorH9gfvgdTzhEjXE9-FHu54zj63y6GJM1kJy0vGGKfw'}, 'query_params': [], 'post_params': [], 'body': {'apiVersion': 'batch/v1', 'kind': 'Job', 'metadata': {'labels': {'app': 'thoth', 'component': 'graph-schema-update-job', 'mark': 'cleanup'}, 'name': 'graph-update-schema-7f56a52c400a58e992985a33c2bce4d5'}, 'spec': {'backoffLimit': 7, 'template': {'metadata': {'labels': {'app': 'thoth', 'component': 'graph-schema-update-job', 'mark': 'cleanup'}}, 'spec': {'containers': [{'env': [{'name': 'THOTH_WORKFLOW_TASK', 'value': 'graph_schema_update'}, {'name': 'THOTH_DEPLOYMENT_NAME', 'valueFrom': {'configMapKeyRef': {'key': 'deployment-name', 'name': 'thoth'}}}, {'name': 'KNOWLEDGE_GRAPH_HOST', 'valueFrom': {'configMapKeyRef': {'key': 'postgresql-host', 'name': 'thoth'}}}, {'name': 'KNOWLEDGE_GRAPH_PORT', 'value': '5432'}, {'name': 'KNOWLEDGE_GRAPH_SSL_DISABLED', 'value': '1'}, {'name': 'KNOWLEDGE_GRAPH_USER', 'valueFrom': {'secretKeyRef': {'key': 'database-user', 'name': 'postgresql'}}}, {'name': 'KNOWLEDGE_GRAPH_PASSWORD', 'valueFrom': {'secretKeyRef': {'key': 'database-password', 'name': 'postgresql'}}}, {'name': 'KNOWLEDGE_GRAPH_DATABASE', 'valueFrom': {'secretKeyRef': {'key': 'database-name', 'name': 'postgresql'}}}, {'name': 'THOTH_CEPH_BUCKET', 'valueFrom': {'configMapKeyRef': {'key': 'bucket-name', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_BUCKET_PREFIX', 'valueFrom': {'configMapKeyRef': {'key': 'bucket-prefix', 'name': 'ceph'}}}, {'name': 'THOTH_S3_ENDPOINT_URL', 'valueFrom': {'configMapKeyRef': {'key': 'host', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_KEY_ID', 'valueFrom': {'secretKeyRef': {'key': 'key-id', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_SECRET_KEY', 'valueFrom': {'secretKeyRef': {'key': 'secret-key', 'name': 'ceph'}}}], 'image': 'workflow-helpers', 'name': 'graph-schema-update-job', 'resources': {'limits': {'cpu': '1', 'memory': '512Mi'}, 'requests': {'cpu': '1', 'memory': '256Mi'}}}], 'restartPolicy': 'OnFailure'}}}}, '_preload_content': False, '_request_timeout': None}}, {'filename': 'kubernetes/client/rest.py', 'abs_path': '/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/rest.py', 'function': 'request', 'module': 'kubernetes.client.rest', 'lineno': 231, 'pre_context': ['', ' # log response body', ' logger.debug(\"response body: %s\", r.data)', '', ' if not 200 <= r.status <= 299:'], 'context_line': ' raise ApiException(http_resp=r)', 'post_context': ['', ' return r', '', ' def GET(self, url, headers=None, query_params=None, _preload_content=True,', ' _request_timeout=None):'], 'vars': {'self': <kubernetes.client.rest.RESTClientObject object at 0x7f4e826c17c0>, 'method': 'POST', 'url': 'https://172.30.0.1:443/apis/batch/v1/namespaces/thoth-test-core/jobs', 'query_params': [], 'headers': {'Accept': 'application/json', 'Content-Type': 'application/json', 'User-Agent': 'OpenAPI-Generator/11.0.0/python', 'authorization': 'bearer eyJhbGciOiJSUzI1NiIsImtpZCI6IlhaVlc3YnFrZ1BaSllHSkpoNnVLenpTQlE1X3lrcEJzVndHWUdJclVDVEUifQ.eyJpc3MiOiJrdWJlcm5ldGVzL3NlcnZpY2VhY2NvdW50Iiwia3ViZXJuZXRlcy5pby9zZXJ2aWNlYWNjb3VudC9uYW1lc3BhY2UiOiJ0aG90aC10ZXN0LWNvcmUiLCJrdWJlcm5ldGVzLmlvL3NlcnZpY2VhY2NvdW50L3NlY3JldC5uYW1lIjoibWFuYWdlbWVudC1hcGktdG9rZW4tNXRqcDkiLCJrdWJlcm5ldGVzLmlvL3NlcnZpY2VhY2NvdW50L3NlcnZpY2UtYWNjb3VudC5uYW1lIjoibWFuYWdlbWVudC1hcGkiLCJrdWJlcm5ldGVzLmlvL3NlcnZpY2VhY2NvdW50L3NlcnZpY2UtYWNjb3VudC51aWQiOiIyMTZlODJlZi05N2RhLTRiNGYtODk4Mi02ZDQzOWEwNjRmNjIiLCJzdWIiOiJzeXN0ZW06c2VydmljZWFjY291bnQ6dGhvdGgtdGVzdC1jb3JlOm1hbmFnZW1lbnQtYXBpIn0.KLDhPJFSnfyXXY5tV5788iPFbYUQ48TXZW2GEi6N-8QZ8qVkcXmubhlvlO8_P93Me9-CS8iAgsUtRH8RYBUOLssZ_R6ioZaOH3Qe7TQhMwbgu6cqFp0wYMRJnvY-RrXHLZBRjUrstgKmEeE6UsclRxy88TRGpp2XYjlTSgW-XHYUj_EMqTePkQXrQZ8MM7EwQiAvwRv653dq3mSmNt4iwF5ewN7C6lPm1mUpw2Kg-kn1ojcP_G1d9gr1u41FUnPD_BDyWfa0Y2fS1fmR33QL7whdmvC26t6qANbHyMlz-jzorH9gfvgdTzhEjXE9-FHu54zj63y6GJM1kJy0vGGKfw'}, 'body': {'apiVersion': 'batch/v1', 'kind': 'Job', 'metadata': {'labels': {'app': 'thoth', 'component': 'graph-schema-update-job', 'mark': 'cleanup'}, 'name': 'graph-update-schema-7f56a52c400a58e992985a33c2bce4d5'}, 'spec': {'backoffLimit': 7, 'template': {'metadata': {'labels': {'app': 'thoth', 'component': 'graph-schema-update-job', 'mark': 'cleanup'}}, 'spec': {'containers': [{'env': [{'name': 'THOTH_WORKFLOW_TASK', 'value': 'graph_schema_update'}, {'name': 'THOTH_DEPLOYMENT_NAME', 'valueFrom': {'configMapKeyRef': {'key': 'deployment-name', 'name': 'thoth'}}}, {'name': 'KNOWLEDGE_GRAPH_HOST', 'valueFrom': {'configMapKeyRef': {'key': 'postgresql-host', 'name': 'thoth'}}}, {'name': 'KNOWLEDGE_GRAPH_PORT', 'value': '5432'}, {'name': 'KNOWLEDGE_GRAPH_SSL_DISABLED', 'value': '1'}, {'name': 'KNOWLEDGE_GRAPH_USER', 'valueFrom': {'secretKeyRef': {'key': 'database-user', 'name': 'postgresql'}}}, {'name': 'KNOWLEDGE_GRAPH_PASSWORD', 'valueFrom': {'secretKeyRef': {'key': 'database-password', 'name': 'postgresql'}}}, {'name': 'KNOWLEDGE_GRAPH_DATABASE', 'valueFrom': {'secretKeyRef': {'key': 'database-name', 'name': 'postgresql'}}}, {'name': 'THOTH_CEPH_BUCKET', 'valueFrom': {'configMapKeyRef': {'key': 'bucket-name', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_BUCKET_PREFIX', 'valueFrom': {'configMapKeyRef': {'key': 'bucket-prefix', 'name': 'ceph'}}}, {'name': 'THOTH_S3_ENDPOINT_URL', 'valueFrom': {'configMapKeyRef': {'key': 'host', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_KEY_ID', 'valueFrom': {'secretKeyRef': {'key': 'key-id', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_SECRET_KEY', 'valueFrom': {'secretKeyRef': {'key': 'secret-key', 'name': 'ceph'}}}], 'image': 'workflow-helpers', 'name': 'graph-schema-update-job', 'resources': {'limits': {'cpu': '1', 'memory': '512Mi'}, 'requests': {'cpu': '1', 'memory': '256Mi'}}}], 'restartPolicy': 'OnFailure'}}}}, 'post_params': {}, '_preload_content': False, '_request_timeout': None, 'timeout': None, 'request_body': '{\"apiVersion\": \"batch/v1\", \"kind\": \"Job\", \"metadata\": {\"labels\": {\"app\": \"thoth\", \"component\": \"graph-schema-update-job\", \"mark\": \"cleanup\"}, \"name\": \"graph-update-schema-7f56a52c400a58e992985a33c2bce4d5\"}, \"spec\": {\"backoffLimit\": 7, \"template\": {\"metadata\": {\"labels\": {\"app\": \"thoth\", \"component\": \"graph-schema-update-job\", \"mark\": \"cleanup\"}}, \"spec\": {\"containers\": [{\"env\": [{\"name\": \"THOTH_WORKFLOW_TASK\", \"value\": \"graph_schema_update\"}, {\"name\": \"THOTH_DEPLOYMENT_NAME\", \"valueFrom\": {\"configMapKeyRef\": {\"key\": \"deployment-name\", \"name\": \"thoth\"}}}, {\"name\": \"KNOWLEDGE_GRAPH_HOST\", \"valueFrom\": {\"configMapKeyRef\": {\"key\": \"postgresql-host\", \"name\": \"thoth\"}}}, {\"name\": \"KNOWLEDGE_GRAPH_PORT\", \"value\": \"5432\"}, {\"name\": \"KNOWLEDGE_GRAPH_SSL_DISABLED\", \"value\": \"1\"}, {\"name\": \"KNOWLEDGE_GRAPH_USER\", \"valueFrom\": {\"secretKeyRef\": {\"key\": \"database-user\", \"name\": \"postgresql\"}}}, {\"name\": \"KNOWLEDGE_GRAPH_PASSWORD\", \"valueFrom\": {\"secretKeyRef\": {\"key\": \"database-password\", \"name\": \"postgresql\"}}}, {\"name\": \"KNOWLEDGE_GRAPH_DATABASE\", \"valueFrom\": {\"secretKeyRef\": {\"key\": \"database-name\", \"name\": \"postgresql\"}}}, {\"name\": \"THOTH_CEPH_BUCKET\", \"valueFrom\": {\"configMapKeyRef\": {\"key\": \"bucket-name\", \"name\": \"ceph\"}}}, {\"name\": \"THOTH_CEPH_BUCKET_PREFIX\", \"valueFrom\": {\"configMapKeyRef\": {\"key\": \"bucket-prefix\", \"name\": \"ceph\"}}}, {\"name\": \"THOTH_S3_ENDPOINT_URL\", \"valueFrom\": {\"configMapKeyRef\": {\"key\": \"host\", \"name\": \"ceph\"}}}, {\"name\": \"THOTH_CEPH_KEY_ID\", \"valueFrom\": {\"secretKeyRef\": {\"key\": \"key-id\", \"name\": \"ceph\"}}}, {\"name\": \"THOTH_CEPH_SECRET_KEY\", \"valueFrom\": {\"secretKeyRef\": {\"key\": \"secret-key\", \"name\": \"ceph\"}}}], \"image\": \"workflow-helpers\", \"name\": \"graph-schema-update-job\", \"resources\": {\"limits\": {\"cpu\": \"1\", \"memory\": \"512Mi\"}, \"requests\": {\"cpu\": \"1\", \"memory\": \"256Mi\"}}}], \"restartPolicy\": \"OnFailure\"}}}}', 'r': <urllib3.response.HTTPResponse object at 0x7f4e80044eb0>}}]}}, {'module': 'openshift.dynamic.exceptions', 'type': 'ForbiddenError', 'value': '403\\nReason: Forbidden\\nHTTP response headers: HTTPHeaderDict({\\'Audit-Id\\': \\'49583870-4187-49f8-9539-aef793ac349e\\', \\'Cache-Control\\': \\'no-cache, private\\', \\'Content-Type\\': \\'application/json\\', \\'X-Content-Type-Options\\': \\'nosniff\\', \\'Date\\': \\'Thu, 14 Jan 2021 16:45:25 GMT\\', \\'Content-Length\\': \\'334\\'})\\nHTTP response body: b\\'{\"kind\":\"Status\",\"apiVersion\":\"v1\",\"metadata\":{},\"status\":\"Failure\",\"message\":\"jobs.batch is forbidden: User \\\\\\\\\"system:serviceaccount:thoth-test-core:management-api\\\\\\\\\" cannot create resource \\\\\\\\\"jobs\\\\\\\\\" in API group \\\\\\\\\"batch\\\\\\\\\" in the namespace \\\\\\\\\"thoth-test-core\\\\\\\\\"\",\"reason\":\"Forbidden\",\"details\":{\"group\":\"batch\",\"kind\":\"jobs\"},\"code\":403}\\\\n\\'\\nOriginal traceback: \\n File \"/opt/app-root/lib64/python3.8/site-packages/openshift/dynamic/client.py\", line 42, in inner\\n resp = func(self, *args, **kwargs)\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/openshift/dynamic/client.py\", line 235, in request\\n return self.client.call_api(\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py\", line 340, in call_api\\n return self.__call_api(resource_path, method,\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py\", line 172, in __call_api\\n response_data = self.request(\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py\", line 382, in request\\n return self.rest_client.POST(url,\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/rest.py\", line 272, in POST\\n return self.request(\"POST\", url,\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/rest.py\", line 231, in request\\n raise ApiException(http_resp=r)\\n', 'mechanism': {'type': 'logging', 'handled': True}, 'stacktrace': {'frames': [{'filename': 'flask/app.py', 'abs_path': '/opt/app-root/lib64/python3.8/site-packages/flask/app.py', 'function': 'wsgi_app', 'module': 'flask.app', 'lineno': 2447, 'pre_context': [' ctx = self.request_context(environ)', ' error = None', ' try:', ' try:', ' ctx.push()'], 'context_line': ' response = self.full_dispatch_request()', 'post_context': [' except Exception as e:', ' error = e', ' response = self.handle_exception(e)', ' except: # noqa: B001', ' error = sys.exc_info()[1]'], 'vars': {'self': <Flask 'thoth.management_api.openapi_server'>, 'environ': {'wsgi.errors': <gunicorn.http.wsgi.WSGIErrorsWrapper object at 0x7f4e800311c0>, 'wsgi.version': (1, 0), 'wsgi.multithread': False, 'wsgi.multiprocess': False, 'wsgi.run_once': False, 'wsgi.file_wrapper': <class 'gunicorn.http.wsgi.FileWrapper'>, 'wsgi.input_terminated': True, 'SERVER_SOFTWARE': 'gunicorn/20.0.4', 'wsgi.input': <gunicorn.http.body.Body object at 0x7f4e800315e0>, 'gunicorn.socket': <socket.socket fd=16, family=AddressFamily.AF_INET, type=SocketKind.SOCK_STREAM, proto=0, laddr=('172.21.12.117', 8080), raddr=('172.20.8.1', 39828)>, 'REQUEST_METHOD': 'PUT', 'QUERY_STRING': 'secret=secret101', 'RAW_URI': '/api/v1/graph/initialize?secret=secret101', 'SERVER_PROTOCOL': 'HTTP/1.1', 'CONTENT_LENGTH': '0', 'HTTP_ACCEPT': '*/*', 'HTTP_USER_AGENT': 'Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/87.0.4280.141 Safari/537.36', 'HTTP_ORIGIN': 'https://management.test.thoth-station.ninja', 'HTTP_SEC_FETCH_SITE': 'same-origin', 'HTTP_SEC_FETCH_MODE': 'cors', 'HTTP_SEC_FETCH_DEST': 'empty', 'HTTP_REFERER': 'https://management.test.thoth-station.ninja/api/v1/ui/', 'HTTP_ACCEPT_ENCODING': 'gzip, deflate, br', 'HTTP_ACCEPT_LANGUAGE': 'it-IT,it;q=0.9,en-US;q=0.8,en;q=0.7,es;q=0.6', 'HTTP_HOST': 'management.test.thoth-station.ninja', 'HTTP_X_FORWARDED_HOST': 'management.test.thoth-station.ninja', 'HTTP_X_FORWARDED_PORT': '443', 'HTTP_X_FORWARDED_PROTO': 'https', 'HTTP_FORWARDED': 'for=10.36.112.242;host=management.test.thoth-station.ninja;proto=https', 'HTTP_X_FORWARDED_FOR': '10.36.112.242', 'HTTP_CONNECTION': 'close', 'wsgi.url_scheme': 'http', 'REMOTE_ADDR': '172.20.8.1', 'REMOTE_PORT': '39828', 'SERVER_NAME': '0.0.0.0', 'SERVER_PORT': '8080', 'PATH_INFO': '/api/v1/graph/initialize', 'SCRIPT_NAME': '', 'werkzeug.request': <Request 'http://management.test.thoth-station.ninja/api/v1/graph/initialize?secret=secret101' [PUT]>}, 'start_response': functools.partial(<function _sentry_start_response at 0x7f4e8293ea60>, <bound method Response.start_response of <gunicorn.http.wsgi.Response object at 0x7f4e80031190>>, <Transaction(name='/api/v1.thoth_management_api_api_v1_initialize_schema', op='http.server', trace_id='df56345840f44f009335865766b30f2f', span_id='a032e586b4eb3cf8', parent_span_id=None, sampled=True)>), 'ctx': <RequestContext 'http://management.test.thoth-station.ninja/api/v1/graph/initialize?secret=secret101' [PUT] of thoth.management_api.openapi_server>, 'error': ForbiddenError(ApiException(), ' File \"/opt/app-root/lib64/python3.8/site-packages/openshift/dynamic/client.py\", line 42, in inner\\n resp = func(self, *args, **kwargs)\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/openshift/dynamic/client.py\", line 235, in request\\n return self.client.call_api(\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py\", line 340, in call_api\\n return self.__call_api(resource_path, method,\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py\", line 172, in __call_api\\n response_data = self.request(\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py\", line 382, in request\\n return self.rest_client.POST(url,\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/rest.py\", line 272, in POST\\n return self.request(\"POST\", url,\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/rest.py\", line 231, in request\\n raise ApiException(http_resp=r)\\n'), 'e': ForbiddenError(ApiException(), ' File \"/opt/app-root/lib64/python3.8/site-packages/openshift/dynamic/client.py\", line 42, in inner\\n resp = func(self, *args, **kwargs)\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/openshift/dynamic/client.py\", line 235, in request\\n return self.client.call_api(\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py\", line 340, in call_api\\n return self.__call_api(resource_path, method,\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py\", line 172, in __call_api\\n response_data = self.request(\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py\", line 382, in request\\n return self.rest_client.POST(url,\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/rest.py\", line 272, in POST\\n return self.request(\"POST\", url,\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/rest.py\", line 231, in request\\n raise ApiException(http_resp=r)\\n')}}, {'filename': 'flask/app.py', 'abs_path': '/opt/app-root/lib64/python3.8/site-packages/flask/app.py', 'function': 'full_dispatch_request', 'module': 'flask.app', 'lineno': 1952, 'pre_context': [' request_started.send(self)', ' rv = self.preprocess_request()', ' if rv is None:', ' rv = self.dispatch_request()', ' except Exception as e:'], 'context_line': ' rv = self.handle_user_exception(e)', 'post_context': [' return self.finalize_request(rv)', '', ' def finalize_request(self, rv, from_error_handler=False):', ' \"\"\"Given the return value from a view function this finalizes', ' the request by converting it into a response and invoking the'], 'vars': {'self': <Flask 'thoth.management_api.openapi_server'>, 'rv': None}}, {'filename': 'flask/app.py', 'abs_path': '/opt/app-root/lib64/python3.8/site-packages/flask/app.py', 'function': 'handle_user_exception', 'module': 'flask.app', 'lineno': 1821, 'pre_context': [' return self.handle_http_exception(e)', '', ' handler = self._find_error_handler(e)', '', ' if handler is None:'], 'context_line': ' reraise(exc_type, exc_value, tb)', 'post_context': [' return handler(e)', '', ' def handle_exception(self, e):', ' \"\"\"Handle an exception that did not have an error handler', ' associated with it, or that was raised from an error handler.'], 'vars': {'self': <Flask 'thoth.management_api.openapi_server'>, 'e': ForbiddenError(ApiException(), ' File \"/opt/app-root/lib64/python3.8/site-packages/openshift/dynamic/client.py\", line 42, in inner\\n resp = func(self, *args, **kwargs)\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/openshift/dynamic/client.py\", line 235, in request\\n return self.client.call_api(\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py\", line 340, in call_api\\n return self.__call_api(resource_path, method,\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py\", line 172, in __call_api\\n response_data = self.request(\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py\", line 382, in request\\n return self.rest_client.POST(url,\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/rest.py\", line 272, in POST\\n return self.request(\"POST\", url,\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/rest.py\", line 231, in request\\n raise ApiException(http_resp=r)\\n'), 'exc_type': <class 'openshift.dynamic.exceptions.ForbiddenError'>, 'exc_value': ForbiddenError(ApiException(), ' File \"/opt/app-root/lib64/python3.8/site-packages/openshift/dynamic/client.py\", line 42, in inner\\n resp = func(self, *args, **kwargs)\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/openshift/dynamic/client.py\", line 235, in request\\n return self.client.call_api(\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py\", line 340, in call_api\\n return self.__call_api(resource_path, method,\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py\", line 172, in __call_api\\n response_data = self.request(\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py\", line 382, in request\\n return self.rest_client.POST(url,\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/rest.py\", line 272, in POST\\n return self.request(\"POST\", url,\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/rest.py\", line 231, in request\\n raise ApiException(http_resp=r)\\n'), 'tb': <traceback object at 0x7f4e7a70fd00>, 'handler': None}}, {'filename': 'flask/_compat.py', 'abs_path': '/opt/app-root/lib64/python3.8/site-packages/flask/_compat.py', 'function': 'reraise', 'module': 'flask._compat', 'lineno': 39, 'pre_context': [' import collections.abc as collections_abc', '', ' def reraise(tp, value, tb=None):', ' if value.__traceback__ is not tb:', ' raise value.with_traceback(tb)'], 'context_line': ' raise value', 'post_context': ['', ' implements_to_string = _identity', '', 'else:', ' iterkeys = lambda d: d.iterkeys()'], 'vars': {'tp': <class 'openshift.dynamic.exceptions.ForbiddenError'>, 'value': ForbiddenError(ApiException(), ' File \"/opt/app-root/lib64/python3.8/site-packages/openshift/dynamic/client.py\", line 42, in inner\\n resp = func(self, *args, **kwargs)\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/openshift/dynamic/client.py\", line 235, in request\\n return self.client.call_api(\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py\", line 340, in call_api\\n return self.__call_api(resource_path, method,\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py\", line 172, in __call_api\\n response_data = self.request(\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/api_client.py\", line 382, in request\\n return self.rest_client.POST(url,\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/rest.py\", line 272, in POST\\n return self.request(\"POST\", url,\\n\\n File \"/opt/app-root/lib64/python3.8/site-packages/kubernetes/client/rest.py\", line 231, in request\\n raise ApiException(http_resp=r)\\n'), 'tb': <traceback object at 0x7f4e7a70fd00>}}, {'filename': 'flask/app.py', 'abs_path': '/opt/app-root/lib64/python3.8/site-packages/flask/app.py', 'function': 'full_dispatch_request', 'module': 'flask.app', 'lineno': 1950, 'pre_context': [' self.try_trigger_before_first_request_functions()', ' try:', ' request_started.send(self)', ' rv = self.preprocess_request()', ' if rv is None:'], 'context_line': ' rv = self.dispatch_request()', 'post_context': [' except Exception as e:', ' rv = self.handle_user_exception(e)', ' return self.finalize_request(rv)', '', ' def finalize_request(self, rv, from_error_handler=False):'], 'vars': {'self': <Flask 'thoth.management_api.openapi_server'>, 'rv': None}}, {'filename': 'flask/app.py', 'abs_path': '/opt/app-root/lib64/python3.8/site-packages/flask/app.py', 'function': 'dispatch_request', 'module': 'flask.app', 'lineno': 1936, 'pre_context': [' getattr(rule, \"provide_automatic_options\", False)', ' and req.method == \"OPTIONS\"', ' ):', ' return self.make_default_options_response()', ' # otherwise dispatch to the handler for that endpoint'], 'context_line': ' return self.view_functions[rule.endpoint](**req.view_args)', 'post_context': ['', ' def full_dispatch_request(self):', ' \"\"\"Dispatches the request and on top of that performs request', ' pre and postprocessing as well as HTTP exception catching and', ' error handling.'], 'vars': {'self': <Flask 'thoth.management_api.openapi_server'>, 'req': <Request 'http://management.test.thoth-station.ninja/api/v1/graph/initialize?secret=secret101' [PUT]>, 'rule': <Rule '/api/v1/graph/initialize' (OPTIONS, PUT) -> /api/v1.thoth_management_api_api_v1_initialize_schema>}}, {'filename': 'connexion/decorators/decorator.py', 'abs_path': '/opt/app-root/lib64/python3.8/site-packages/connexion/decorators/decorator.py', 'function': 'wrapper', 'module': 'connexion.decorators.decorator', 'lineno': 48, 'pre_context': ['', ' else: # pragma: 3 no cover', ' @functools.wraps(function)', ' def wrapper(*args, **kwargs):', ' request = self.api.get_request(*args, **kwargs)'], 'context_line': ' response = function(request)', 'post_context': [' return self.api.get_response(response, self.mimetype, request)', '', ' return wrapper'], 'vars': {'args': (), 'kwargs': {}, 'request': <connexion.lifecycle.ConnexionRequest object at 0x7f4e800aa820>, 'function': <function initialize_schema at 0x7f4e82802ee0>, 'self': <BaseDecorator>}}, {'filename': 'connexion/decorators/uri_parsing.py', 'abs_path': '/opt/app-root/lib64/python3.8/site-packages/connexion/decorators/uri_parsing.py', 'function': 'wrapper', 'module': 'connexion.decorators.uri_parsing', 'lineno': 144, 'pre_context': [' form = coerce_dict(request.form)', '', ' request.query = self.resolve_query(query)', ' request.path_params = self.resolve_path(path_params)', ' request.form = self.resolve_form(form)'], 'context_line': ' response = function(request)', 'post_context': [' return response', '', ' return wrapper', '', ''], 'vars': {'request': <connexion.lifecycle.ConnexionRequest object at 0x7f4e800aa820>, 'coerce_dict': <function AbstractURIParser.__call__.<locals>.wrapper.<locals>.coerce_dict at 0x7f4e8004bd30>, 'query': {'secret': ['secret101']}, 'path_params': {}, 'form': {}, 'function': <function initialize_schema at 0x7f4e82802e50>, 'self': <OpenAPIURIParser>}}, {'filename': 'connexion/decorators/validation.py', 'abs_path': '/opt/app-root/lib64/python3.8/site-packages/connexion/decorators/validation.py', 'function': 'wrapper', 'module': 'connexion.decorators.validation', 'lineno': 384, 'pre_context': [\" for param in self.parameters.get('formData', []):\", ' error = self.validate_formdata_parameter(param[\"name\"], param, request)', ' if error:', ' raise BadRequestProblem(detail=error)', ''], 'context_line': ' return function(request)', 'post_context': ['', ' return wrapper'], 'vars': {'request': <connexion.lifecycle.ConnexionRequest object at 0x7f4e800aa820>, 'query_errors': set(), 'formdata_errors': set(), 'param': {'name': 'secret', 'in': 'query', 'required': True, 'description': 'A secret to authorize operation.', 'schema': {'type': 'string'}}, 'error': None, 'function': <function initialize_schema at 0x7f4e82802dc0>, 'self': <connexion.decorators.validation.ParameterValidator object at 0x7f4e827fc340>}}, {'filename': 'connexion/decorators/parameter.py', 'abs_path': '/opt/app-root/lib64/python3.8/site-packages/connexion/decorators/parameter.py', 'function': 'wrapper', 'module': 'connexion.decorators.parameter', 'lineno': 121, 'pre_context': ['', ' # attempt to provide the request context to the function', ' if pass_context_arg_name and (has_kwargs or pass_context_arg_name in arguments):', ' kwargs[pass_context_arg_name] = request.context', ''], 'context_line': ' return function(**kwargs)', 'post_context': ['', ' return wrapper'], 'vars': {'request': <connexion.lifecycle.ConnexionRequest object at 0x7f4e800aa820>, 'kwargs': {'secret': 'secret101'}, 'request_body': None, 'query': {'secret': 'secret101'}, 'arguments': ['secret'], 'consumes': ['application/json'], 'function': <function initialize_schema at 0x7f4e82694d30>, 'has_kwargs': False, 'operation': <connexion.operations.openapi.OpenAPIOperation object at 0x7f4e827fc130>, 'pass_context_arg_name': None, 'pythonic_params': False, 'sanitize': <function parameter_to_arg.<locals>.sanitized at 0x7f4e82802ca0>}}, {'filename': '/opt/app-root/src/thoth/management_api/api_v1.py', 'abs_path': '/opt/app-root/src/thoth/management_api/api_v1.py', 'function': 'initialize_schema', 'module': 'thoth.management_api.api_v1', 'lineno': 348, 'pre_context': ['def initialize_schema(secret: str):', ' \"\"\"Initialize/update schema in graph database (async).\"\"\"', ' if secret != Configuration.THOTH_MANAGEMENT_API_TOKEN:', ' return {\"error\": \"Wrong secret provided\"}, 401', ''], 'context_line': ' job_id = _OPENSHIFT.schedule_graph_schema_update()', 'post_context': [' return {\"job_id\": job_id}, 201', '', '', 'def schedule_solver_unsolvable(secret: str, solver_name: str) -> tuple:', ' \"\"\"Schedule solving of unsolvable packages for the given solver.\"\"\"'], 'vars': {'secret': 'secret101'}}, {'filename': '/opt/app-root/lib64/python3.8/site-packages/thoth/common/openshift.py', 'abs_path': '/opt/app-root/lib64/python3.8/site-packages/thoth/common/openshift.py', 'function': 'schedule_graph_schema_update', 'module': 'thoth.common.openshift', 'lineno': 791, 'pre_context': ['', ' template = self.oc_process(namespace, template)', '', ' graph_update_schema_template = template[\"objects\"][0]', ''], 'context_line': ' response = self.ocp_client.resources.get(', 'post_context': [' api_version=graph_update_schema_template[\"apiVersion\"],', ' kind=graph_update_schema_template[\"kind\"],', ' ).create(body=graph_update_schema_template, namespace=namespace)', '', ' _LOGGER.debug(\"OpenShift response for creating a pod: %r\", response.to_dict())'], 'vars': {'self': <thoth.common.openshift.OpenShift object at 0x7f4e826e3310>, 'job_id': 'graph-update-schema-7f56a52c400a58e992985a33c2bce4d5', 'namespace': 'thoth-test-core', 'template': {'kind': 'Template', 'apiVersion': 'template.openshift.io/v1', 'metadata': {'name': 'graph-schema-update-job', 'namespace': 'thoth-test-core', 'selfLink': '/apis/template.openshift.io/v1/namespaces/thoth-test-core/processedtemplates/graph-schema-update-job', 'uid': '1c4d59d8-dd65-4f1b-82f9-c2505cc3c6b2', 'resourceVersion': '391642130', 'creationTimestamp': '2021-01-14T16:43:05Z', 'labels': {'template': 'graph-schema-update-job'}, 'annotations': {'description': 'Thoth: Graph Schema Update Job for updating schema in database', 'kubectl.kubernetes.io/last-applied-configuration': '{\"apiVersion\":\"template.openshift.io/v1\",\"kind\":\"Template\",\"metadata\":{\"annotations\":{\"description\":\"Thoth: Graph Schema Update Job for updating schema in database\",\"openshift.io/display-name\":\"Thoth: Graph Schema Update Job Multiple Documents\",\"tags\":\"thoth,ai-stacks,aistacks,graph-schema-update-job\",\"template.openshift.io/documentation-url\":\"https://github.com/Thoth-Station/\",\"template.openshift.io/long-description\":\"This template defines resources needed to deploy Thoth\\'s Graph Schema Update Job on OpenShift.\\\\n\",\"template.openshift.io/provider-display-name\":\"Red Hat, Inc.\"},\"labels\":{\"template\":\"graph-schema-update-job\"},\"name\":\"graph-schema-update-job\",\"namespace\":\"thoth-test-core\"},\"objects\":[{\"apiVersion\":\"batch/v1\",\"kind\":\"Job\",\"metadata\":{\"labels\":{\"app\":\"thoth\",\"component\":\"graph-schema-update-job\",\"mark\":\"cleanup\"},\"name\":\"${THOTH_SCHEMA_UPDATE_JOB_ID}\"},\"spec\":{\"backoffLimit\":7,\"template\":{\"metadata\":{\"labels\":{\"app\":\"thoth\",\"component\":\"graph-schema-update-job\",\"mark\":\"cleanup\"}},\"spec\":{\"containers\":[{\"env\":[{\"name\":\"THOTH_WORKFLOW_TASK\",\"value\":\"graph_schema_update\"},{\"name\":\"THOTH_DEPLOYMENT_NAME\",\"valueFrom\":{\"configMapKeyRef\":{\"key\":\"deployment-name\",\"name\":\"thoth\"}}},{\"name\":\"KNOWLEDGE_GRAPH_HOST\",\"valueFrom\":{\"configMapKeyRef\":{\"key\":\"postgresql-host\",\"name\":\"thoth\"}}},{\"name\":\"KNOWLEDGE_GRAPH_PORT\",\"value\":\"5432\"},{\"name\":\"KNOWLEDGE_GRAPH_SSL_DISABLED\",\"value\":\"1\"},{\"name\":\"KNOWLEDGE_GRAPH_USER\",\"valueFrom\":{\"secretKeyRef\":{\"key\":\"database-user\",\"name\":\"postgresql\"}}},{\"name\":\"KNOWLEDGE_GRAPH_PASSWORD\",\"valueFrom\":{\"secretKeyRef\":{\"key\":\"database-password\",\"name\":\"postgresql\"}}},{\"name\":\"KNOWLEDGE_GRAPH_DATABASE\",\"valueFrom\":{\"secretKeyRef\":{\"key\":\"database-name\",\"name\":\"postgresql\"}}},{\"name\":\"THOTH_CEPH_BUCKET\",\"valueFrom\":{\"configMapKeyRef\":{\"key\":\"bucket-name\",\"name\":\"ceph\"}}},{\"name\":\"THOTH_CEPH_BUCKET_PREFIX\",\"valueFrom\":{\"configMapKeyRef\":{\"key\":\"bucket-prefix\",\"name\":\"ceph\"}}},{\"name\":\"THOTH_S3_ENDPOINT_URL\",\"valueFrom\":{\"configMapKeyRef\":{\"key\":\"host\",\"name\":\"ceph\"}}},{\"name\":\"THOTH_CEPH_KEY_ID\",\"valueFrom\":{\"secretKeyRef\":{\"key\":\"key-id\",\"name\":\"ceph\"}}},{\"name\":\"THOTH_CEPH_SECRET_KEY\",\"valueFrom\":{\"secretKeyRef\":{\"key\":\"secret-key\",\"name\":\"ceph\"}}}],\"image\":\"workflow-helpers\",\"name\":\"graph-schema-update-job\",\"resources\":{\"limits\":{\"cpu\":\"1\",\"memory\":\"512Mi\"},\"requests\":{\"cpu\":\"1\",\"memory\":\"256Mi\"}}}],\"restartPolicy\":\"OnFailure\"}}}}],\"parameters\":[{\"description\":\"A unique identifier of job\",\"displayName\":\"A unique identifier of job\",\"name\":\"THOTH_SCHEMA_UPDATE_JOB_ID\",\"required\":true}]}\\n', 'openshift.io/display-name': 'Thoth: Graph Schema Update Job Multiple Documents', 'tags': 'thoth,ai-stacks,aistacks,graph-schema-update-job', 'template.openshift.io/documentation-url': 'https://github.com/Thoth-Station/', 'template.openshift.io/long-description': \"This template defines resources needed to deploy Thoth's Graph Schema Update Job on OpenShift.\\n\", 'template.openshift.io/provider-display-name': 'Red Hat, Inc.'}, 'managedFields': [{'manager': 'python-requests', 'operation': 'Update', 'apiVersion': 'template.openshift.io/v1', 'time': '2021-01-14T16:45:25Z', 'fieldsType': 'FieldsV1', 'fieldsV1': {'f:metadata': {'f:annotations': {'.': {}, 'f:description': {}, 'f:kubectl.kubernetes.io/last-applied-configuration': {}, 'f:openshift.io/display-name': {}, 'f:tags': {}, 'f:template.openshift.io/documentation-url': {}, 'f:template.openshift.io/long-description': {}, 'f:template.openshift.io/provider-display-name': {}}, 'f:labels': {'.': {}, 'f:template': {}}}, 'f:objects': {}, 'f:parameters': {}}}]}, 'objects': [{'apiVersion': 'batch/v1', 'kind': 'Job', 'metadata': {'labels': {'app': 'thoth', 'component': 'graph-schema-update-job', 'mark': 'cleanup'}, 'name': 'graph-update-schema-7f56a52c400a58e992985a33c2bce4d5'}, 'spec': {'backoffLimit': 7, 'template': {'metadata': {'labels': {'app': 'thoth', 'component': 'graph-schema-update-job', 'mark': 'cleanup'}}, 'spec': {'containers': [{'env': [{'name': 'THOTH_WORKFLOW_TASK', 'value': 'graph_schema_update'}, {'name': 'THOTH_DEPLOYMENT_NAME', 'valueFrom': {'configMapKeyRef': {'key': 'deployment-name', 'name': 'thoth'}}}, {'name': 'KNOWLEDGE_GRAPH_HOST', 'valueFrom': {'configMapKeyRef': {'key': 'postgresql-host', 'name': 'thoth'}}}, {'name': 'KNOWLEDGE_GRAPH_PORT', 'value': '5432'}, {'name': 'KNOWLEDGE_GRAPH_SSL_DISABLED', 'value': '1'}, {'name': 'KNOWLEDGE_GRAPH_USER', 'valueFrom': {'secretKeyRef': {'key': 'database-user', 'name': 'postgresql'}}}, {'name': 'KNOWLEDGE_GRAPH_PASSWORD', 'valueFrom': {'secretKeyRef': {'key': 'database-password', 'name': 'postgresql'}}}, {'name': 'KNOWLEDGE_GRAPH_DATABASE', 'valueFrom': {'secretKeyRef': {'key': 'database-name', 'name': 'postgresql'}}}, {'name': 'THOTH_CEPH_BUCKET', 'valueFrom': {'configMapKeyRef': {'key': 'bucket-name', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_BUCKET_PREFIX', 'valueFrom': {'configMapKeyRef': {'key': 'bucket-prefix', 'name': 'ceph'}}}, {'name': 'THOTH_S3_ENDPOINT_URL', 'valueFrom': {'configMapKeyRef': {'key': 'host', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_KEY_ID', 'valueFrom': {'secretKeyRef': {'key': 'key-id', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_SECRET_KEY', 'valueFrom': {'secretKeyRef': {'key': 'secret-key', 'name': 'ceph'}}}], 'image': 'workflow-helpers', 'name': 'graph-schema-update-job', 'resources': {'limits': {'cpu': '1', 'memory': '512Mi'}, 'requests': {'cpu': '1', 'memory': '256Mi'}}}], 'restartPolicy': 'OnFailure'}}}}], 'parameters': [{'name': 'THOTH_SCHEMA_UPDATE_JOB_ID', 'displayName': 'A unique identifier of job', 'description': 'A unique identifier of job', 'value': 'graph-update-schema-7f56a52c400a58e992985a33c2bce4d5', 'required': True}]}, 'graph_update_schema_template': {'apiVersion': 'batch/v1', 'kind': 'Job', 'metadata': {'labels': {'app': 'thoth', 'component': 'graph-schema-update-job', 'mark': 'cleanup'}, 'name': 'graph-update-schema-7f56a52c400a58e992985a33c2bce4d5'}, 'spec': {'backoffLimit': 7, 'template': {'metadata': {'labels': {'app': 'thoth', 'component': 'graph-schema-update-job', 'mark': 'cleanup'}}, 'spec': {'containers': [{'env': [{'name': 'THOTH_WORKFLOW_TASK', 'value': 'graph_schema_update'}, {'name': 'THOTH_DEPLOYMENT_NAME', 'valueFrom': {'configMapKeyRef': {'key': 'deployment-name', 'name': 'thoth'}}}, {'name': 'KNOWLEDGE_GRAPH_HOST', 'valueFrom': {'configMapKeyRef': {'key': 'postgresql-host', 'name': 'thoth'}}}, {'name': 'KNOWLEDGE_GRAPH_PORT', 'value': '5432'}, {'name': 'KNOWLEDGE_GRAPH_SSL_DISABLED', 'value': '1'}, {'name': 'KNOWLEDGE_GRAPH_USER', 'valueFrom': {'secretKeyRef': {'key': 'database-user', 'name': 'postgresql'}}}, {'name': 'KNOWLEDGE_GRAPH_PASSWORD', 'valueFrom': {'secretKeyRef': {'key': 'database-password', 'name': 'postgresql'}}}, {'name': 'KNOWLEDGE_GRAPH_DATABASE', 'valueFrom': {'secretKeyRef': {'key': 'database-name', 'name': 'postgresql'}}}, {'name': 'THOTH_CEPH_BUCKET', 'valueFrom': {'configMapKeyRef': {'key': 'bucket-name', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_BUCKET_PREFIX', 'valueFrom': {'configMapKeyRef': {'key': 'bucket-prefix', 'name': 'ceph'}}}, {'name': 'THOTH_S3_ENDPOINT_URL', 'valueFrom': {'configMapKeyRef': {'key': 'host', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_KEY_ID', 'valueFrom': {'secretKeyRef': {'key': 'key-id', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_SECRET_KEY', 'valueFrom': {'secretKeyRef': {'key': 'secret-key', 'name': 'ceph'}}}], 'image': 'workflow-helpers', 'name': 'graph-schema-update-job', 'resources': {'limits': {'cpu': '1', 'memory': '512Mi'}, 'requests': {'cpu': '1', 'memory': '256Mi'}}}], 'restartPolicy': 'OnFailure'}}}}}}, {'filename': 'openshift/dynamic/client.py', 'abs_path': '/opt/app-root/lib64/python3.8/site-packages/openshift/dynamic/client.py', 'function': 'create', 'module': 'openshift.dynamic.client', 'lineno': 101, 'pre_context': [' def create(self, resource, body=None, namespace=None, **kwargs):', ' body = self.serialize_body(body)', ' if resource.namespaced:', ' namespace = self.ensure_namespace(resource, namespace, body)', ' path = resource.path(namespace=namespace)'], 'context_line': \" return self.request('post', path, body=body, **kwargs)\", 'post_context': ['', ' def delete(self, resource, name=None, namespace=None, label_selector=None, field_selector=None, **kwargs):', ' if not (name or label_selector or field_selector):', ' raise ValueError(\"At least one of name|label_selector|field_selector is required\")', ' if resource.namespaced and not (label_selector or field_selector or namespace):'], 'vars': {'self': <openshift.dynamic.client.DynamicClient object at 0x7f4e826c1c70>, 'resource': <Resource(batch/v1/jobs)>, 'body': {'apiVersion': 'batch/v1', 'kind': 'Job', 'metadata': {'labels': {'app': 'thoth', 'component': 'graph-schema-update-job', 'mark': 'cleanup'}, 'name': 'graph-update-schema-7f56a52c400a58e992985a33c2bce4d5'}, 'spec': {'backoffLimit': 7, 'template': {'metadata': {'labels': {'app': 'thoth', 'component': 'graph-schema-update-job', 'mark': 'cleanup'}}, 'spec': {'containers': [{'env': [{'name': 'THOTH_WORKFLOW_TASK', 'value': 'graph_schema_update'}, {'name': 'THOTH_DEPLOYMENT_NAME', 'valueFrom': {'configMapKeyRef': {'key': 'deployment-name', 'name': 'thoth'}}}, {'name': 'KNOWLEDGE_GRAPH_HOST', 'valueFrom': {'configMapKeyRef': {'key': 'postgresql-host', 'name': 'thoth'}}}, {'name': 'KNOWLEDGE_GRAPH_PORT', 'value': '5432'}, {'name': 'KNOWLEDGE_GRAPH_SSL_DISABLED', 'value': '1'}, {'name': 'KNOWLEDGE_GRAPH_USER', 'valueFrom': {'secretKeyRef': {'key': 'database-user', 'name': 'postgresql'}}}, {'name': 'KNOWLEDGE_GRAPH_PASSWORD', 'valueFrom': {'secretKeyRef': {'key': 'database-password', 'name': 'postgresql'}}}, {'name': 'KNOWLEDGE_GRAPH_DATABASE', 'valueFrom': {'secretKeyRef': {'key': 'database-name', 'name': 'postgresql'}}}, {'name': 'THOTH_CEPH_BUCKET', 'valueFrom': {'configMapKeyRef': {'key': 'bucket-name', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_BUCKET_PREFIX', 'valueFrom': {'configMapKeyRef': {'key': 'bucket-prefix', 'name': 'ceph'}}}, {'name': 'THOTH_S3_ENDPOINT_URL', 'valueFrom': {'configMapKeyRef': {'key': 'host', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_KEY_ID', 'valueFrom': {'secretKeyRef': {'key': 'key-id', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_SECRET_KEY', 'valueFrom': {'secretKeyRef': {'key': 'secret-key', 'name': 'ceph'}}}], 'image': 'workflow-helpers', 'name': 'graph-schema-update-job', 'resources': {'limits': {'cpu': '1', 'memory': '512Mi'}, 'requests': {'cpu': '1', 'memory': '256Mi'}}}], 'restartPolicy': 'OnFailure'}}}}, 'namespace': 'thoth-test-core', 'kwargs': {}, 'path': '/apis/batch/v1/namespaces/thoth-test-core/jobs'}}, {'filename': 'openshift/dynamic/client.py', 'abs_path': '/opt/app-root/lib64/python3.8/site-packages/openshift/dynamic/client.py', 'function': 'inner', 'module': 'openshift.dynamic.client', 'lineno': 44, 'pre_context': [\" serialize_response = kwargs.pop('serialize', True)\", \" serializer = kwargs.pop('serializer', ResourceInstance)\", ' try:', ' resp = func(self, *args, **kwargs)', ' except ApiException as e:'], 'context_line': ' raise api_exception(e)', 'post_context': [' if serialize_response:', ' try:', ' if six.PY2:', ' return serializer(self, json.loads(resp.data))', \" return serializer(self, json.loads(resp.data.decode('utf8')))\"], 'vars': {'self': <openshift.dynamic.client.DynamicClient object at 0x7f4e826c1c70>, 'args': ('post', '/apis/batch/v1/namespaces/thoth-test-core/jobs'), 'kwargs': {'body': {'apiVersion': 'batch/v1', 'kind': 'Job', 'metadata': {'labels': {'app': 'thoth', 'component': 'graph-schema-update-job', 'mark': 'cleanup'}, 'name': 'graph-update-schema-7f56a52c400a58e992985a33c2bce4d5'}, 'spec': {'backoffLimit': 7, 'template': {'metadata': {'labels': {'app': 'thoth', 'component': 'graph-schema-update-job', 'mark': 'cleanup'}}, 'spec': {'containers': [{'env': [{'name': 'THOTH_WORKFLOW_TASK', 'value': 'graph_schema_update'}, {'name': 'THOTH_DEPLOYMENT_NAME', 'valueFrom': {'configMapKeyRef': {'key': 'deployment-name', 'name': 'thoth'}}}, {'name': 'KNOWLEDGE_GRAPH_HOST', 'valueFrom': {'configMapKeyRef': {'key': 'postgresql-host', 'name': 'thoth'}}}, {'name': 'KNOWLEDGE_GRAPH_PORT', 'value': '5432'}, {'name': 'KNOWLEDGE_GRAPH_SSL_DISABLED', 'value': '1'}, {'name': 'KNOWLEDGE_GRAPH_USER', 'valueFrom': {'secretKeyRef': {'key': 'database-user', 'name': 'postgresql'}}}, {'name': 'KNOWLEDGE_GRAPH_PASSWORD', 'valueFrom': {'secretKeyRef': {'key': 'database-password', 'name': 'postgresql'}}}, {'name': 'KNOWLEDGE_GRAPH_DATABASE', 'valueFrom': {'secretKeyRef': {'key': 'database-name', 'name': 'postgresql'}}}, {'name': 'THOTH_CEPH_BUCKET', 'valueFrom': {'configMapKeyRef': {'key': 'bucket-name', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_BUCKET_PREFIX', 'valueFrom': {'configMapKeyRef': {'key': 'bucket-prefix', 'name': 'ceph'}}}, {'name': 'THOTH_S3_ENDPOINT_URL', 'valueFrom': {'configMapKeyRef': {'key': 'host', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_KEY_ID', 'valueFrom': {'secretKeyRef': {'key': 'key-id', 'name': 'ceph'}}}, {'name': 'THOTH_CEPH_SECRET_KEY', 'valueFrom': {'secretKeyRef': {'key': 'secret-key', 'name': 'ceph'}}}], 'image': 'workflow-helpers', 'name': 'graph-schema-update-job', 'resources': {'limits': {'cpu': '1', 'memory': '512Mi'}, 'requests': {'cpu': '1', 'memory': '256Mi'}}}], 'restartPolicy': 'OnFailure'}}}}}, 'serialize_response': True, 'serializer': <class 'openshift.dynamic.resource.ResourceInstance'>, 'func': <function DynamicClient.request at 0x7f4e86f494c0>}}]}}]}, 'logger': 'thoth.management_api.openapi_server', 'logentry': {'message': 'Exception on /api/v1/graph/initialize [PUT]', 'params': ()}, 'extra': {'asctime': '2021-01-14 16:45:26,011'}, 'event_id': '92d2f914f6214dbbb420e91f4e7399fd', 'timestamp': datetime.datetime(2021, 1, 14, 16, 45, 26, 19047), 'breadcrumbs': {'values': [{'type': 'http', 'category': 'httplib', 'data': {'method': 'GET', 'url': 'https://172.30.0.1/apis/template.openshift.io/v1', 'status_code': 200, 'reason': 'OK'}, 'timestamp': datetime.datetime(2021, 1, 14, 16, 45, 25, 718394)}, {'type': 'http', 'category': 'httplib', 'data': {'method': 'GET', 'url': 'https://172.30.0.1/apis/template.openshift.io/v1/namespaces/thoth-test-core/templates?labelSelector=template%3Dgraph-schema-update-job', 'status_code': 200, 'reason': 'OK'}, 'timestamp': datetime.datetime(2021, 1, 14, 16, 45, 25, 733828)}, {'type': 'http', 'category': 'httplib', 'data': {'method': 'POST', 'url': 'https://172.30.0.1/apis/template.openshift.io/v1/namespaces/thoth-test-core/processedtemplates', 'status_code': 201, 'reason': 'Created'}, 'timestamp': datetime.datetime(2021, 1, 14, 16, 45, 25, 761455)}, {'type': 'http', 'category': 'httplib', 'data': {'method': 'GET', 'url': 'https://172.30.0.1/apis/batch/v1', 'status_code': 200, 'reason': 'OK'}, 'timestamp': datetime.datetime(2021, 1, 14, 16, 45, 25, 766532)}, {'type': 'http', 'category': 'httplib', 'data': {'method': 'POST', 'url': 'https://172.30.0.1/apis/batch/v1/namespaces/thoth-test-core/jobs', 'status_code': 403, 'reason': 'Forbidden'}, 'timestamp': datetime.datetime(2021, 1, 14, 16, 45, 25, 773598)}]}, 'transaction': '/api/v1.thoth_management_api_api_v1_initialize_schema', 'contexts': {'trace': {'trace_id': 'df56345840f44f009335865766b30f2f', 'span_id': 'a032e586b4eb3cf8', 'parent_span_id': None, 'op': 'http.server', 'description': None}, 'runtime': {'name': 'CPython', 'version': '3.8.3', 'build': '3.8.3 (default, Aug 18 2020, 08:56:04) \\n[GCC 8.3.1 20191121 (Red Hat 8.3.1-5)]'}}})"} {"remote": "172.20.8.1", "date": "[14/Jan/2021:16:45:26 +0000]", "status": "500", "response_length": 177, "referer": "https://management.test.thoth-station.ninja/api/v1/ui/", "user_agent": "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/87.0.4280.141 Safari/537.36", "request_method": "PUT", "url_path": "/api/v1/graph/initialize", "protocol": "HTTP/1.1", "request_time": 0} ``` **Additional context** Add any other context about the problem here.
test
user system serviceaccount thoth test core management api cannot create resource jobs in api group batch in the namespace thoth test core describe the bug user system serviceaccount thoth test core management api cannot create resource jobs in api group batch in the namespace thoth test core to reproduce steps to reproduce the behavior use graph initialize in management api in test envioronment expected behavior a clear and concise description of what you expected to happen screenshots name thoth management api openapi server levelname error module app lineno funcname log exception created asctime msecs relative created process message exception on api graph initialize ntraceback most recent call last n file opt app root site packages openshift dynamic client py line in inner n resp func self args kwargs n file opt app root site packages openshift dynamic client py line in request n return self client call api n file opt app root site packages kubernetes client api client py line in call api n return self call api resource path method n file opt app root site packages kubernetes client api client py line in call api n response data self request n file opt app root site packages kubernetes client api client py line in request n return self rest client post url n file opt app root site packages kubernetes client rest py line in post n return self request post url n file opt app root site packages kubernetes client rest py line in request n raise apiexception http resp r nkubernetes client rest apiexception nreason forbidden nhttp response headers httpheaderdict audit id cache control no cache private content type application json x content type options nosniff date thu jan gmt content length nhttp response body b kind status apiversion metadata status failure message jobs batch is forbidden user system serviceaccount thoth test core management api cannot create resource jobs in api group batch in the namespace thoth test core reason forbidden details group batch kind jobs code n n n nduring handling of the above exception another exception occurred n ntraceback most recent call last n file opt app root site packages flask app py line in wsgi app n response self full dispatch request n file opt app root site packages flask app py line in full dispatch request n rv self handle user exception e n file opt app root site packages flask app py line in handle user exception n reraise exc type exc value tb n file opt app root site packages flask compat py line in reraise n raise value n file opt app root site packages flask app py line in full dispatch request n rv self dispatch request n file opt app root site packages flask app py line in dispatch request n return self view functions req view args n file opt app root site packages connexion decorators decorator py line in wrapper n response function request n file opt app root site packages connexion decorators uri parsing py line in wrapper n response function request n file opt app root site packages connexion decorators validation py line in wrapper n return function request n file opt app root site packages connexion decorators parameter py line in wrapper n return function kwargs n file opt app root src thoth management api api py line in initialize schema n job id openshift schedule graph schema update n file opt app root site packages thoth common openshift py line in schedule graph schema update n response self ocp client resources get n file opt app root site packages openshift dynamic client py line in create n return self request post path body body kwargs n file opt app root site packages openshift dynamic client py line in inner n raise api exception e nopenshift dynamic exceptions forbiddenerror nreason forbidden nhttp response headers httpheaderdict audit id cache control no cache private content type application json x content type options nosniff date thu jan gmt content length nhttp response body b kind status apiversion metadata status failure message jobs batch is forbidden user system serviceaccount thoth test core management api cannot create resource jobs in api group batch in the namespace thoth test core reason forbidden details group batch kind jobs code n noriginal traceback n file opt app root site packages openshift dynamic client py line in inner n resp func self args kwargs n n file opt app root site packages openshift dynamic client py line in request n return self client call api n n file opt app root site packages kubernetes client api client py line in call api n return self call api resource path method n n file opt app root site packages kubernetes client api client py line in call api n response data self request n n file opt app root site packages kubernetes client api client py line in request n return self rest client post url n n file opt app root site packages kubernetes client rest py line in post n return self request post url n n file opt app root site packages kubernetes client rest py line in request n raise apiexception http resp r n name sentry sdk errors levelname info module scope lineno funcname drop created asctime msecs relative created process message event processor processor at dropped event level error exception values context line resp func self args kwargs post context vars self args post apis batch namespaces thoth test core jobs kwargs body apiversion batch kind job metadata labels app thoth component graph schema update job mark cleanup name graph update schema spec backofflimit template metadata labels app thoth component graph schema update job mark cleanup spec containers image workflow helpers name graph schema update job resources limits cpu memory requests cpu memory restartpolicy onfailure serialize response true serializer func filename openshift dynamic client py abs path opt app root site packages openshift dynamic client py function request module openshift dynamic client lineno pre context self client select header content type authentication setting auth settings context line return self client call api post context vars self method post path apis batch namespaces thoth test core jobs body apiversion batch kind job metadata labels app thoth component graph schema update job mark cleanup name graph update schema spec backofflimit template metadata labels app thoth component graph schema update job mark cleanup spec containers image workflow helpers name graph schema update job resources limits cpu memory requests cpu memory restartpolicy onfailure params path params query params header params accept application json content type application json user agent openapi generator python form params local var files auth settings filename kubernetes client api client py abs path opt app root site packages kubernetes client api client py function call api module kubernetes client api client lineno pre context context line return self call api resource path method post context vars self resource path apis batch namespaces thoth test core jobs method post path params query params header params accept application json content type application json user agent openapi generator python body apiversion batch kind job metadata labels app thoth component graph schema update job mark cleanup name graph update schema spec backofflimit template metadata labels app thoth component graph schema update job mark cleanup spec containers image workflow helpers name graph schema update job resources limits cpu memory requests cpu memory restartpolicy onfailure post params files response type none auth settings async req none return http data only true collection formats none preload content false request timeout none filename kubernetes client api client py abs path opt app root site packages kubernetes client api client py function call api module kubernetes client api client lineno pre context context line response data self request post context vars self resource path apis batch namespaces thoth test core jobs method post path params query params header params accept application json content type application json user agent openapi generator python authorization bearer xhyuj body apiversion batch kind job metadata labels app thoth component graph schema update job mark cleanup name graph update schema spec backofflimit template metadata labels app thoth component graph schema update job mark cleanup spec containers image workflow helpers name graph schema update job resources limits cpu memory requests cpu memory restartpolicy onfailure post params files response type none auth settings return http data only true collection formats none preload content false request timeout none config url filename kubernetes client api client py abs path opt app root site packages kubernetes client api client py function request module kubernetes client api client lineno pre context context line return self rest client post url post context vars self method post url query params headers accept application json content type application json user agent openapi generator python authorization bearer xhyuj post params body apiversion batch kind job metadata labels app thoth component graph schema update job mark cleanup name graph update schema spec backofflimit template metadata labels app thoth component graph schema update job mark cleanup spec containers image workflow helpers name graph schema update job resources limits cpu memory requests cpu memory restartpolicy onfailure preload content false request timeout none filename kubernetes client rest py abs path opt app root site packages kubernetes client rest py function post module kubernetes client rest lineno pre context context line return self request post url post context vars self url headers accept application json content type application json user agent openapi generator python authorization bearer xhyuj query params post params body apiversion batch kind job metadata labels app thoth component graph schema update job mark cleanup name graph update schema spec backofflimit template metadata labels app thoth component graph schema update job mark cleanup spec containers image workflow helpers name graph schema update job resources limits cpu memory requests cpu memory restartpolicy onfailure preload content false request timeout none filename kubernetes client rest py abs path opt app root site packages kubernetes client rest py function request module kubernetes client rest lineno pre context context line raise apiexception http resp r post context vars self method post url query params headers accept application json content type application json user agent openapi generator python authorization bearer xhyuj body apiversion batch kind job metadata labels app thoth component graph schema update job mark cleanup name graph update schema spec backofflimit template metadata labels app thoth component graph schema update job mark cleanup spec containers image workflow helpers name graph schema update job resources limits cpu memory requests cpu memory restartpolicy onfailure post params preload content false request timeout none timeout none request body apiversion batch kind job metadata labels app thoth component graph schema update job mark cleanup name graph update schema spec backofflimit template metadata labels app thoth component graph schema update job mark cleanup spec containers image workflow helpers name graph schema update job resources limits cpu memory requests cpu memory restartpolicy onfailure r module openshift dynamic exceptions type forbiddenerror value nreason forbidden nhttp response headers httpheaderdict audit id cache control no cache private content type application json x content type options nosniff date thu jan gmt content length nhttp response body b kind status apiversion metadata status failure message jobs batch is forbidden user system serviceaccount thoth test core management api cannot create resource jobs in api group batch in the namespace thoth test core reason forbidden details group batch kind jobs code n noriginal traceback n file opt app root site packages openshift dynamic client py line in inner n resp func self args kwargs n n file opt app root site packages openshift dynamic client py line in request n return self client call api n n file opt app root site packages kubernetes client api client py line in call api n return self call api resource path method n n file opt app root site packages kubernetes client api client py line in call api n response data self request n n file opt app root site packages kubernetes client api client py line in request n return self rest client post url n n file opt app root site packages kubernetes client rest py line in post n return self request post url n n file opt app root site packages kubernetes client rest py line in request n raise apiexception http resp r n mechanism type logging handled true stacktrace frames context line response self full dispatch request post context vars self environ wsgi errors wsgi version wsgi multithread false wsgi multiprocess false wsgi run once false wsgi file wrapper wsgi input terminated true server software gunicorn wsgi input gunicorn socket request method put query string secret raw uri api graph initialize secret server protocol http content length http accept http user agent mozilla linux applewebkit khtml like gecko chrome safari http origin http sec fetch site same origin http sec fetch mode cors http sec fetch dest empty http referer http accept encoding gzip deflate br http accept language it it it q en us q en q es q http host management test thoth station ninja http x forwarded host management test thoth station ninja http x forwarded port http x forwarded proto https http forwarded for host management test thoth station ninja proto https http x forwarded for http connection close wsgi url scheme http remote addr remote port server name server port path info api graph initialize script name werkzeug request start response functools partial ctx error forbiddenerror apiexception file opt app root site packages openshift dynamic client py line in inner n resp func self args kwargs n n file opt app root site packages openshift dynamic client py line in request n return self client call api n n file opt app root site packages kubernetes client api client py line in call api n return self call api resource path method n n file opt app root site packages kubernetes client api client py line in call api n response data self request n n file opt app root site packages kubernetes client api client py line in request n return self rest client post url n n file opt app root site packages kubernetes client rest py line in post n return self request post url n n file opt app root site packages kubernetes client rest py line in request n raise apiexception http resp r n e forbiddenerror apiexception file opt app root site packages openshift dynamic client py line in inner n resp func self args kwargs n n file opt app root site packages openshift dynamic client py line in request n return self client call api n n file opt app root site packages kubernetes client api client py line in call api n return self call api resource path method n n file opt app root site packages kubernetes client api client py line in call api n response data self request n n file opt app root site packages kubernetes client api client py line in request n return self rest client post url n n file opt app root site packages kubernetes client rest py line in post n return self request post url n n file opt app root site packages kubernetes client rest py line in request n raise apiexception http resp r n filename flask app py abs path opt app root site packages flask app py function full dispatch request module flask app lineno pre context context line rv self handle user exception e post context vars self rv none filename flask app py abs path opt app root site packages flask app py function handle user exception module flask app lineno pre context context line reraise exc type exc value tb post context vars self e forbiddenerror apiexception file opt app root site packages openshift dynamic client py line in inner n resp func self args kwargs n n file opt app root site packages openshift dynamic client py line in request n return self client call api n n file opt app root site packages kubernetes client api client py line in call api n return self call api resource path method n n file opt app root site packages kubernetes client api client py line in call api n response data self request n n file opt app root site packages kubernetes client api client py line in request n return self rest client post url n n file opt app root site packages kubernetes client rest py line in post n return self request post url n n file opt app root site packages kubernetes client rest py line in request n raise apiexception http resp r n exc type exc value forbiddenerror apiexception file opt app root site packages openshift dynamic client py line in inner n resp func self args kwargs n n file opt app root site packages openshift dynamic client py line in request n return self client call api n n file opt app root site packages kubernetes client api client py line in call api n return self call api resource path method n n file opt app root site packages kubernetes client api client py line in call api n response data self request n n file opt app root site packages kubernetes client api client py line in request n return self rest client post url n n file opt app root site packages kubernetes client rest py line in post n return self request post url n n file opt app root site packages kubernetes client rest py line in request n raise apiexception http resp r n tb handler none filename flask compat py abs path opt app root site packages flask compat py function reraise module flask compat lineno pre context context line raise value post context vars tp value forbiddenerror apiexception file opt app root site packages openshift dynamic client py line in inner n resp func self args kwargs n n file opt app root site packages openshift dynamic client py line in request n return self client call api n n file opt app root site packages kubernetes client api client py line in call api n return self call api resource path method n n file opt app root site packages kubernetes client api client py line in call api n response data self request n n file opt app root site packages kubernetes client api client py line in request n return self rest client post url n n file opt app root site packages kubernetes client rest py line in post n return self request post url n n file opt app root site packages kubernetes client rest py line in request n raise apiexception http resp r n tb filename flask app py abs path opt app root site packages flask app py function full dispatch request module flask app lineno pre context context line rv self dispatch request post context vars self rv none filename flask app py abs path opt app root site packages flask app py function dispatch request module flask app lineno pre context context line return self view functions req view args post context vars self req rule api thoth management api api initialize schema filename connexion decorators decorator py abs path opt app root site packages connexion decorators decorator py function wrapper module connexion decorators decorator lineno pre context context line response function request post context vars args kwargs request function self filename connexion decorators uri parsing py abs path opt app root site packages connexion decorators uri parsing py function wrapper module connexion decorators uri parsing lineno pre context context line response function request post context vars request coerce dict wrapper coerce dict at query secret path params form function self filename connexion decorators validation py abs path opt app root site packages connexion decorators validation py function wrapper module connexion decorators validation lineno pre context error self validate formdata parameter param param request if error raise badrequestproblem detail error context line return function request post context vars request query errors set formdata errors set param name secret in query required true description a secret to authorize operation schema type string error none function self filename connexion decorators parameter py abs path opt app root site packages connexion decorators parameter py function wrapper module connexion decorators parameter lineno pre context request context context line return function kwargs post context vars request kwargs secret request body none query secret arguments consumes function has kwargs false operation pass context arg name none pythonic params false sanitize sanitized at filename opt app root src thoth management api api py abs path opt app root src thoth management api api py function initialize schema module thoth management api api lineno pre context context line job id openshift schedule graph schema update post context vars secret filename opt app root site packages thoth common openshift py abs path opt app root site packages thoth common openshift py function schedule graph schema update module thoth common openshift lineno pre context context line response self ocp client resources get post context kind graph update schema template create body graph update schema template namespace namespace logger debug openshift response for creating a pod r response to dict vars self job id graph update schema namespace thoth test core template kind template apiversion template openshift io metadata name graph schema update job namespace thoth test core selflink apis template openshift io namespaces thoth test core processedtemplates graph schema update job uid resourceversion creationtimestamp labels template graph schema update job annotations description thoth graph schema update job for updating schema in database kubectl kubernetes io last applied configuration apiversion template openshift io kind template metadata annotations description thoth graph schema update job for updating schema in database openshift io display name thoth graph schema update job multiple documents tags thoth ai stacks aistacks graph schema update job template openshift io documentation url template defines resources needed to deploy thoth s graph schema update job on openshift n template openshift io provider display name red hat inc labels template graph schema update job name graph schema update job namespace thoth test core objects image workflow helpers name graph schema update job resources limits cpu memory requests cpu memory restartpolicy onfailure parameters n openshift io display name thoth graph schema update job multiple documents tags thoth ai stacks aistacks graph schema update job template openshift io documentation url template openshift io long description this template defines resources needed to deploy thoth s graph schema update job on openshift n template openshift io provider display name red hat inc managedfields objects image workflow helpers name graph schema update job resources limits cpu memory requests cpu memory restartpolicy onfailure parameters graph update schema template apiversion batch kind job metadata labels app thoth component graph schema update job mark cleanup name graph update schema spec backofflimit template metadata labels app thoth component graph schema update job mark cleanup spec containers image workflow helpers name graph schema update job resources limits cpu memory requests cpu memory restartpolicy onfailure filename openshift dynamic client py abs path opt app root site packages openshift dynamic client py function create module openshift dynamic client lineno pre context context line return self request post path body body kwargs post context vars self resource body apiversion batch kind job metadata labels app thoth component graph schema update job mark cleanup name graph update schema spec backofflimit template metadata labels app thoth component graph schema update job mark cleanup spec containers image workflow helpers name graph schema update job resources limits cpu memory requests cpu memory restartpolicy onfailure namespace thoth test core kwargs path apis batch namespaces thoth test core jobs filename openshift dynamic client py abs path opt app root site packages openshift dynamic client py function inner module openshift dynamic client lineno pre context context line raise api exception e post context vars self args post apis batch namespaces thoth test core jobs kwargs body apiversion batch kind job metadata labels app thoth component graph schema update job mark cleanup name graph update schema spec backofflimit template metadata labels app thoth component graph schema update job mark cleanup spec containers image workflow helpers name graph schema update job resources limits cpu memory requests cpu memory restartpolicy onfailure serialize response true serializer func logger thoth management api openapi server logentry message exception on api graph initialize params extra asctime event id timestamp datetime datetime breadcrumbs values transaction api thoth management api api initialize schema contexts trace trace id span id parent span id none op http server description none runtime name cpython version build default aug n remote date status response length referer user agent mozilla linux applewebkit khtml like gecko chrome safari request method put url path api graph initialize protocol http request time additional context add any other context about the problem here
1
133,414
29,172,918,924
IssuesEvent
2023-05-19 04:44:48
EddieHubCommunity/LinkFree
https://api.github.com/repos/EddieHubCommunity/LinkFree
closed
Found two user profiles whose GitHub profiles does not exist
πŸ’» aspect: code πŸ›  goal: fix 🟧 priority: high πŸ”’ points: 5
### Description 1) `Skyhero-admin.json` 2) `aaquibdev.json` this error occurred due to this `upstream image response failed for https://github.com/Skyhero-admin.png 404` ### Screenshots this caused the random user section to generate only 4 user profiles, the 5th threw error ![Screenshot from 2023-01-09 00-40-14](https://user-images.githubusercontent.com/95094057/211214416-29c8c1e6-db39-4acf-926d-584dc6fc148c.png) ### Additional information There must be more accounts like this, it might be negligible for a large number of users (this happened on my local dev server, with smaller user list)
1.0
Found two user profiles whose GitHub profiles does not exist - ### Description 1) `Skyhero-admin.json` 2) `aaquibdev.json` this error occurred due to this `upstream image response failed for https://github.com/Skyhero-admin.png 404` ### Screenshots this caused the random user section to generate only 4 user profiles, the 5th threw error ![Screenshot from 2023-01-09 00-40-14](https://user-images.githubusercontent.com/95094057/211214416-29c8c1e6-db39-4acf-926d-584dc6fc148c.png) ### Additional information There must be more accounts like this, it might be negligible for a large number of users (this happened on my local dev server, with smaller user list)
non_test
found two user profiles whose github profiles does not exist description skyhero admin json aaquibdev json this error occurred due to this upstream image response failed for screenshots this caused the random user section to generate only user profiles the threw error additional information there must be more accounts like this it might be negligible for a large number of users this happened on my local dev server with smaller user list
0
240,433
20,030,389,705
IssuesEvent
2022-02-02 04:38:27
flutter/flutter
https://api.github.com/repos/flutter/flutter
opened
Symbolize stack traces dumped by Fuchsia on the FEMU target
a: tests dependency: fuchsia
Fuchsia+FEMU is a nontrivial test dependency to setup and maintain for developers, and so at least having symbolized stack traces when segfaults happen on FEMU would save many hours of debugging time. For example, mysterious FEMU sefgaults blocked https://github.com/flutter/engine/pull/29391 for days.
1.0
Symbolize stack traces dumped by Fuchsia on the FEMU target - Fuchsia+FEMU is a nontrivial test dependency to setup and maintain for developers, and so at least having symbolized stack traces when segfaults happen on FEMU would save many hours of debugging time. For example, mysterious FEMU sefgaults blocked https://github.com/flutter/engine/pull/29391 for days.
test
symbolize stack traces dumped by fuchsia on the femu target fuchsia femu is a nontrivial test dependency to setup and maintain for developers and so at least having symbolized stack traces when segfaults happen on femu would save many hours of debugging time for example mysterious femu sefgaults blocked for days
1
301,309
26,033,773,634
IssuesEvent
2022-12-22 01:15:38
sergeirocks100/GearStation_Next
https://api.github.com/repos/sergeirocks100/GearStation_Next
opened
Flaky test nuke_cinematic: Cannot modify null.dir.
πŸ€– Flaky Test Report
<!-- This issue can be renamed, but do not change the next comment! --> <!-- title: Flaky test nuke_cinematic: Cannot modify null.dir. --> Flaky tests were detected in [this test run](https://github.com/sergeirocks100/GearStation_Next/actions/runs/3753931985/attempts/1). This means that there was a failure that was cleared when the tests were simply restarted. Failures: ``` nuke_cinematic: [00:45:05] Runtime in bot.dm,1063: Cannot modify null.dir. proc name: set path (/mob/living/simple_animal/bot/proc/set_path) src: Officer Beepsky (/mob/living/simple_animal/bot/secbot/beepsky/officer) src.loc: the floor (172,116,2) (/turf/open/floor/iron) call stack: Officer Beepsky (/mob/living/simple_animal/bot/secbot/beepsky/officer): set path(/list (/list)) Officer Beepsky (/mob/living/simple_animal/bot/secbot/beepsky/officer): calc path(null) Officer Beepsky (/mob/living/simple_animal/bot/secbot/beepsky/officer): target patrol() world: ImmediateInvokeAsync(Officer Beepsky (/mob/living/simple_animal/bot/secbot/beepsky/officer), /mob/living/simple_animal/bot/... (/mob/living/simple_animal/bot/proc/target_patrol)) at bot.dm:1063 ```
1.0
Flaky test nuke_cinematic: Cannot modify null.dir. - <!-- This issue can be renamed, but do not change the next comment! --> <!-- title: Flaky test nuke_cinematic: Cannot modify null.dir. --> Flaky tests were detected in [this test run](https://github.com/sergeirocks100/GearStation_Next/actions/runs/3753931985/attempts/1). This means that there was a failure that was cleared when the tests were simply restarted. Failures: ``` nuke_cinematic: [00:45:05] Runtime in bot.dm,1063: Cannot modify null.dir. proc name: set path (/mob/living/simple_animal/bot/proc/set_path) src: Officer Beepsky (/mob/living/simple_animal/bot/secbot/beepsky/officer) src.loc: the floor (172,116,2) (/turf/open/floor/iron) call stack: Officer Beepsky (/mob/living/simple_animal/bot/secbot/beepsky/officer): set path(/list (/list)) Officer Beepsky (/mob/living/simple_animal/bot/secbot/beepsky/officer): calc path(null) Officer Beepsky (/mob/living/simple_animal/bot/secbot/beepsky/officer): target patrol() world: ImmediateInvokeAsync(Officer Beepsky (/mob/living/simple_animal/bot/secbot/beepsky/officer), /mob/living/simple_animal/bot/... (/mob/living/simple_animal/bot/proc/target_patrol)) at bot.dm:1063 ```
test
flaky test nuke cinematic cannot modify null dir flaky tests were detected in this means that there was a failure that was cleared when the tests were simply restarted failures nuke cinematic runtime in bot dm cannot modify null dir proc name set path mob living simple animal bot proc set path src officer beepsky mob living simple animal bot secbot beepsky officer src loc the floor turf open floor iron call stack officer beepsky mob living simple animal bot secbot beepsky officer set path list list officer beepsky mob living simple animal bot secbot beepsky officer calc path null officer beepsky mob living simple animal bot secbot beepsky officer target patrol world immediateinvokeasync officer beepsky mob living simple animal bot secbot beepsky officer mob living simple animal bot mob living simple animal bot proc target patrol at bot dm
1
300,908
9,213,758,278
IssuesEvent
2019-03-10 14:42:37
StratusNetwork/issues
https://api.github.com/repos/StratusNetwork/issues
closed
Glina lacks Item Repair of Chain Chestplates and Iron Axes
Priority: Medium Scope: Map Type: Bug
Glina doesn't have item repair for it's axes or chain chestplates. You might also wanna item remove water buckets/buckets/wood/glass so they don't stack up in the inventory too.
1.0
Glina lacks Item Repair of Chain Chestplates and Iron Axes - Glina doesn't have item repair for it's axes or chain chestplates. You might also wanna item remove water buckets/buckets/wood/glass so they don't stack up in the inventory too.
non_test
glina lacks item repair of chain chestplates and iron axes glina doesn t have item repair for it s axes or chain chestplates you might also wanna item remove water buckets buckets wood glass so they don t stack up in the inventory too
0
46,882
5,832,561,399
IssuesEvent
2017-05-08 22:09:47
loconomics/loconomics
https://api.github.com/repos/loconomics/loconomics
closed
Add additional landing pages
C: Content F: Onboarding P2 T4: Ready for Testing
- [x] /welcome/healing-arts-professionals - Massage Therapist 106 - Reiki Practitioner 283 - Nutritionist 141 - Hypnotherapist 146 - Rolfer 318 - Life Coach 186 - [x] /welcome/licensed-professionals - Massage Therapist 106 - Psychotherapist 193 - Chiropractor 170 - Acupuncturist 185 - Esthetician 179 - [x] /welcome/local-service-professionals - Photographer 90 - Tutor 32 - Art Tutor 296 - [x] /welcome/home-care-professionals - Cleaning Professional 14 - Gardener 22 - Handyman 82 - Electrician 34
1.0
Add additional landing pages - - [x] /welcome/healing-arts-professionals - Massage Therapist 106 - Reiki Practitioner 283 - Nutritionist 141 - Hypnotherapist 146 - Rolfer 318 - Life Coach 186 - [x] /welcome/licensed-professionals - Massage Therapist 106 - Psychotherapist 193 - Chiropractor 170 - Acupuncturist 185 - Esthetician 179 - [x] /welcome/local-service-professionals - Photographer 90 - Tutor 32 - Art Tutor 296 - [x] /welcome/home-care-professionals - Cleaning Professional 14 - Gardener 22 - Handyman 82 - Electrician 34
test
add additional landing pages welcome healing arts professionals massage therapist reiki practitioner nutritionist hypnotherapist rolfer life coach welcome licensed professionals massage therapist psychotherapist chiropractor acupuncturist esthetician welcome local service professionals photographer tutor art tutor welcome home care professionals cleaning professional gardener handyman electrician
1
178,362
13,776,003,266
IssuesEvent
2020-10-08 08:51:06
kinvolk/lokomotive
https://api.github.com/repos/kinvolk/lokomotive
closed
In TestSelfHostedKubeletLabels test also delete kubelet pod
area/testing bug
Delete the kubelet pod apart from deleting the node object. I think sometimes when we just delete the node object the register sequence never starts. And that's why we are seeing flaky tests. I am not, sure if it is upto the kubelet to identify if it is deregistered and register again. So solving the problem who has the responsibility to fix the kubelet is out of scope of this issue.
1.0
In TestSelfHostedKubeletLabels test also delete kubelet pod - Delete the kubelet pod apart from deleting the node object. I think sometimes when we just delete the node object the register sequence never starts. And that's why we are seeing flaky tests. I am not, sure if it is upto the kubelet to identify if it is deregistered and register again. So solving the problem who has the responsibility to fix the kubelet is out of scope of this issue.
test
in testselfhostedkubeletlabels test also delete kubelet pod delete the kubelet pod apart from deleting the node object i think sometimes when we just delete the node object the register sequence never starts and that s why we are seeing flaky tests i am not sure if it is upto the kubelet to identify if it is deregistered and register again so solving the problem who has the responsibility to fix the kubelet is out of scope of this issue
1