Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
4
112
repo_url
stringlengths
33
141
action
stringclasses
3 values
title
stringlengths
1
1.02k
labels
stringlengths
4
1.54k
body
stringlengths
1
262k
index
stringclasses
17 values
text_combine
stringlengths
95
262k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
83,838
7,881,890,907
IssuesEvent
2018-06-26 20:37:27
nasa-gibs/worldview
https://api.github.com/repos/nasa-gibs/worldview
closed
Bring in start/end dates from GC docs into layer descriptions
testing
Ben took an initial look at this: WV-1797 Here is a list of changes discussed on 11/14/17 - [x] Change process_temporal_value to build an array of all dates and add to wv.json. i.e. this: <Value>2011-09-30/2011-11-25/P1D</Value> <Value>2011-11-27/2011-11-27/P1D</Value> <Value>2011-11-29/2012-01-22/P1D</Value> Becomes: September 30th, 2011 - November 25th, 2011 November 27th, 2011 - November 27th, 2011 Novemeber 29th, 2011 - January 22nd, 2012 - [x] Display date range at top - [x] Format date to be '24 November 2015' style - [x] Add indicator to top date range to tell user there are gaps in that range. - [x] Display exact dates at bottom in a list. - [x] Remove titles from markdown files - [x] Add date range to the layer list hover state when no data is available: i.e. this -> "No data on selected date for this layer" to this -> "Data available between 27 May 2001 - 3 March 2008" ----------------------------------------------------------------------- Phase 2: Link the bottom dates to the timeline
1.0
Bring in start/end dates from GC docs into layer descriptions - Ben took an initial look at this: WV-1797 Here is a list of changes discussed on 11/14/17 - [x] Change process_temporal_value to build an array of all dates and add to wv.json. i.e. this: <Value>2011-09-30/2011-11-25/P1D</Value> <Value>2011-11-27/2011-11-27/P1D</Value> <Value>2011-11-29/2012-01-22/P1D</Value> Becomes: September 30th, 2011 - November 25th, 2011 November 27th, 2011 - November 27th, 2011 Novemeber 29th, 2011 - January 22nd, 2012 - [x] Display date range at top - [x] Format date to be '24 November 2015' style - [x] Add indicator to top date range to tell user there are gaps in that range. - [x] Display exact dates at bottom in a list. - [x] Remove titles from markdown files - [x] Add date range to the layer list hover state when no data is available: i.e. this -> "No data on selected date for this layer" to this -> "Data available between 27 May 2001 - 3 March 2008" ----------------------------------------------------------------------- Phase 2: Link the bottom dates to the timeline
test
bring in start end dates from gc docs into layer descriptions ben took an initial look at this wv here is a list of changes discussed on change process temporal value to build an array of all dates and add to wv json i e this becomes september november november november novemeber january display date range at top format date to be november style add indicator to top date range to tell user there are gaps in that range display exact dates at bottom in a list remove titles from markdown files add date range to the layer list hover state when no data is available i e this no data on selected date for this layer to this data available between may march phase link the bottom dates to the timeline
1
127,426
10,469,446,701
IssuesEvent
2019-09-22 20:45:53
RPTools/maptool
https://api.github.com/repos/RPTools/maptool
closed
GM tokens FOW revealed to the players on auto reveal
bug claimed tested
Full history http://forums.rptools.net/viewtopic.php?f=3&p=269873#p269873 Maptool 1.4.1.8 a/ start a server with "auto reveal on move" option selected, connect a client to it b/ empty map (visible to player) with FOW enabled and one token assigned to server/GM ownership c/ hit CTRL+e CTRL+f and move the token around from the server FOW is revealed to the player.
1.0
GM tokens FOW revealed to the players on auto reveal - Full history http://forums.rptools.net/viewtopic.php?f=3&p=269873#p269873 Maptool 1.4.1.8 a/ start a server with "auto reveal on move" option selected, connect a client to it b/ empty map (visible to player) with FOW enabled and one token assigned to server/GM ownership c/ hit CTRL+e CTRL+f and move the token around from the server FOW is revealed to the player.
test
gm tokens fow revealed to the players on auto reveal full history maptool a start a server with auto reveal on move option selected connect a client to it b empty map visible to player with fow enabled and one token assigned to server gm ownership c hit ctrl e ctrl f and move the token around from the server fow is revealed to the player
1
311,523
26,795,870,338
IssuesEvent
2023-02-01 11:51:22
akademia-envelo-3/moovelo-front
https://api.github.com/repos/akademia-envelo-3/moovelo-front
closed
[Fe] [U] User - przegląda grupy
question User Frontend SP4 test-ok
### Story User może przegląda grupy ### Dodatkowe informacje - User przegląda grupy które są pobierane z endpointa /groups, wyświetlany jest tytuł, krótki opis, oraz przycisk dołączenia do grupy ### Link do Makiet: https://www.figma.com/file/VwoFt9OUGvu0j4IABdGZHs/Moovelo?node-id=613026%3A6275&t=jnOwK4MYswdWwiEY-0 ### Kryteria akceptacji: - [ ] Wyświetlenie listy grup
1.0
[Fe] [U] User - przegląda grupy - ### Story User może przegląda grupy ### Dodatkowe informacje - User przegląda grupy które są pobierane z endpointa /groups, wyświetlany jest tytuł, krótki opis, oraz przycisk dołączenia do grupy ### Link do Makiet: https://www.figma.com/file/VwoFt9OUGvu0j4IABdGZHs/Moovelo?node-id=613026%3A6275&t=jnOwK4MYswdWwiEY-0 ### Kryteria akceptacji: - [ ] Wyświetlenie listy grup
test
user przegląda grupy story user może przegląda grupy dodatkowe informacje user przegląda grupy które są pobierane z endpointa groups wyświetlany jest tytuł krótki opis oraz przycisk dołączenia do grupy link do makiet kryteria akceptacji wyświetlenie listy grup
1
275,437
23,915,751,058
IssuesEvent
2022-09-09 12:33:10
WordPress/gutenberg
https://api.github.com/repos/WordPress/gutenberg
closed
Navigation block: New menu list view inconsistent
[Type] Bug Needs Testing [Block] Navigation
### Description When I "create new menu" the links in the list view aren't updated: <img width="1440" alt="Screenshot 2022-08-01 at 14 35 40" src="https://user-images.githubusercontent.com/275961/182159970-003dafbc-3672-4491-abe9-42a16a0fb8a7.png"> <img width="1440" alt="Screenshot 2022-08-01 at 14 35 31" src="https://user-images.githubusercontent.com/275961/182159981-99eca27e-90aa-42cd-af2d-4d7c0b784f5d.png"> ### Step-by-step reproduction instructions In a block theme with a navigation block, - Open the list view so that you can see the list of items in the navigation block - Click on "Create new menu" in the nav block - Confirm that you still see the old nav items in the list view ### Screenshots, screen recording, code snippet _No response_ ### Environment info _No response_ ### Please confirm that you have searched existing issues in the repo. Yes ### Please confirm that you have tested with all plugins deactivated except Gutenberg. Yes
1.0
Navigation block: New menu list view inconsistent - ### Description When I "create new menu" the links in the list view aren't updated: <img width="1440" alt="Screenshot 2022-08-01 at 14 35 40" src="https://user-images.githubusercontent.com/275961/182159970-003dafbc-3672-4491-abe9-42a16a0fb8a7.png"> <img width="1440" alt="Screenshot 2022-08-01 at 14 35 31" src="https://user-images.githubusercontent.com/275961/182159981-99eca27e-90aa-42cd-af2d-4d7c0b784f5d.png"> ### Step-by-step reproduction instructions In a block theme with a navigation block, - Open the list view so that you can see the list of items in the navigation block - Click on "Create new menu" in the nav block - Confirm that you still see the old nav items in the list view ### Screenshots, screen recording, code snippet _No response_ ### Environment info _No response_ ### Please confirm that you have searched existing issues in the repo. Yes ### Please confirm that you have tested with all plugins deactivated except Gutenberg. Yes
test
navigation block new menu list view inconsistent description when i create new menu the links in the list view aren t updated img width alt screenshot at src img width alt screenshot at src step by step reproduction instructions in a block theme with a navigation block open the list view so that you can see the list of items in the navigation block click on create new menu in the nav block confirm that you still see the old nav items in the list view screenshots screen recording code snippet no response environment info no response please confirm that you have searched existing issues in the repo yes please confirm that you have tested with all plugins deactivated except gutenberg yes
1
444,145
12,807,021,989
IssuesEvent
2020-07-03 10:35:22
MEN-Mikro-Elektronik/13MD05-90
https://api.github.com/repos/MEN-Mikro-Elektronik/13MD05-90
opened
Global variable G_freeUsrBufList not protected for context switches / SMP
bug high priority
# Internal PR MAIN_PR006854 F75P IOP - F215 frequently lose frame in multi-thread receive mode # Background This problem was observed with the [MSCAN driver](https://github.com/MEN-Mikro-Elektronik/13Z015-06) but can probably be reproduce with any other driver. # Description of the problem The application is executed on multi-core CPU. Two different threads access two different devices (can_1 and can_2). In the application we observe frame lost (this is not suppose to happen because the two interfaces are connected on the same CAN bus and shall receive the same number of frames). # Workaround The user application shall protect all the mscan_xxx() api calls with semaphore. # Correction The correction applied in PR #203 fix the issue with our current test setup. # Work to do 1. Reproduce the problem and define a test setup for this problem (if needed I can provide some test SW for this) 2. Check and confirm if the correction in PR #203 is enough 3. Make some stress tests
1.0
Global variable G_freeUsrBufList not protected for context switches / SMP - # Internal PR MAIN_PR006854 F75P IOP - F215 frequently lose frame in multi-thread receive mode # Background This problem was observed with the [MSCAN driver](https://github.com/MEN-Mikro-Elektronik/13Z015-06) but can probably be reproduce with any other driver. # Description of the problem The application is executed on multi-core CPU. Two different threads access two different devices (can_1 and can_2). In the application we observe frame lost (this is not suppose to happen because the two interfaces are connected on the same CAN bus and shall receive the same number of frames). # Workaround The user application shall protect all the mscan_xxx() api calls with semaphore. # Correction The correction applied in PR #203 fix the issue with our current test setup. # Work to do 1. Reproduce the problem and define a test setup for this problem (if needed I can provide some test SW for this) 2. Check and confirm if the correction in PR #203 is enough 3. Make some stress tests
non_test
global variable g freeusrbuflist not protected for context switches smp internal pr main iop frequently lose frame in multi thread receive mode background this problem was observed with the but can probably be reproduce with any other driver description of the problem the application is executed on multi core cpu two different threads access two different devices can and can in the application we observe frame lost this is not suppose to happen because the two interfaces are connected on the same can bus and shall receive the same number of frames workaround the user application shall protect all the mscan xxx api calls with semaphore correction the correction applied in pr fix the issue with our current test setup work to do reproduce the problem and define a test setup for this problem if needed i can provide some test sw for this check and confirm if the correction in pr is enough make some stress tests
0
352,508
32,073,727,261
IssuesEvent
2023-09-25 09:38:11
input-output-hk/cardano-ledger
https://api.github.com/repos/input-output-hk/cardano-ledger
closed
Add tests to ensure Conway DCerts decoding/encoding is backwards-compatible
:detective: testing
The `EncCBOR` and `DecCBOR` instances for `ConwayDCert` should be backwards compatible with `ShelleyDCert`, meaning that it should be possible to decode a Shelley era certificate in Conway era (with the exception of MIR certificates). We should add tests to check that this is indeed the case.
1.0
Add tests to ensure Conway DCerts decoding/encoding is backwards-compatible - The `EncCBOR` and `DecCBOR` instances for `ConwayDCert` should be backwards compatible with `ShelleyDCert`, meaning that it should be possible to decode a Shelley era certificate in Conway era (with the exception of MIR certificates). We should add tests to check that this is indeed the case.
test
add tests to ensure conway dcerts decoding encoding is backwards compatible the enccbor and deccbor instances for conwaydcert should be backwards compatible with shelleydcert meaning that it should be possible to decode a shelley era certificate in conway era with the exception of mir certificates we should add tests to check that this is indeed the case
1
138,222
11,195,226,699
IssuesEvent
2020-01-03 05:22:09
jojapoppa/fedoragold-wallet-electron
https://api.github.com/repos/jojapoppa/fedoragold-wallet-electron
closed
Transaction history list issues
bug completed - in testing in next release
The transaction history list needs debugging and cleanup. This is mostly already completed, and just needs testing prior to next release.
1.0
Transaction history list issues - The transaction history list needs debugging and cleanup. This is mostly already completed, and just needs testing prior to next release.
test
transaction history list issues the transaction history list needs debugging and cleanup this is mostly already completed and just needs testing prior to next release
1
8,363
2,982,207,920
IssuesEvent
2015-07-17 09:28:26
owncloud/client
https://api.github.com/repos/owncloud/client
closed
[OS X] [Multi account] Fix integration with OSX-style settings dialog
approved by qa bug ReadyToTest sev2-high
* [x] Crash on account delete * [x] UI glitch on account add * [x] Possibly more wrong?
1.0
[OS X] [Multi account] Fix integration with OSX-style settings dialog - * [x] Crash on account delete * [x] UI glitch on account add * [x] Possibly more wrong?
test
fix integration with osx style settings dialog crash on account delete ui glitch on account add possibly more wrong
1
998
4,776,213,976
IssuesEvent
2016-10-27 13:08:10
SpartanRefactoring/Spartanizer
https://api.github.com/repos/SpartanRefactoring/Spartanizer
closed
Tutorial/Travis: sign-up
Architecture Folks GUI/IT Quality assurance Teaching
@orimarco lets write it together. * Step 1: enter [https://travis-ci.org/SpartanRefactoring/Spartanizer](https://travis-ci.org/SpartanRefactoring/Spartanizer) * Step 2: click on `Sign in with GitHub` * Step 3: Profit $$$.
1.0
Tutorial/Travis: sign-up - @orimarco lets write it together. * Step 1: enter [https://travis-ci.org/SpartanRefactoring/Spartanizer](https://travis-ci.org/SpartanRefactoring/Spartanizer) * Step 2: click on `Sign in with GitHub` * Step 3: Profit $$$.
non_test
tutorial travis sign up orimarco lets write it together step enter step click on sign in with github step profit
0
432,574
12,495,287,346
IssuesEvent
2020-06-01 12:55:41
hochschule-darmstadt/openartbrowser
https://api.github.com/repos/hochschule-darmstadt/openartbrowser
closed
Movement overview timeline on start page
User Interface feature high priority
**Is your feature request related to a problem? Please describe.** The user has no overview of the different movements an their chronological order. **Why do you want this feature? What goals should be achieved** Teach users the context of time and art movements. **Describe the solution you'd like** Adding an overview timeline containing (the most important) movements which are not a sub-movement to the start page. This information can be acquired by the wikidata dataset. An empty *is part of* field indicates that the movement is not a sub-movement itself. In the future it might be possible to add more movements (even sub-movements). But this depends on how this will blow up the start page. **Describe acceptance criteria** A timeline on the start page shows a few different important movements in chronological order. **Describe alternatives you've considered** None **What effects does your proposed solution have** We need to crawl more fields of the wikidata dataset and change the model. *This issue is closely coupled to issue #65*. Therefore the newly required data should be equal for both features. **Additional context** ![Timeline_Movement_overview](https://user-images.githubusercontent.com/28682931/69335000-2f669d00-0c5c-11ea-8fa7-e68475283d69.png)
1.0
Movement overview timeline on start page - **Is your feature request related to a problem? Please describe.** The user has no overview of the different movements an their chronological order. **Why do you want this feature? What goals should be achieved** Teach users the context of time and art movements. **Describe the solution you'd like** Adding an overview timeline containing (the most important) movements which are not a sub-movement to the start page. This information can be acquired by the wikidata dataset. An empty *is part of* field indicates that the movement is not a sub-movement itself. In the future it might be possible to add more movements (even sub-movements). But this depends on how this will blow up the start page. **Describe acceptance criteria** A timeline on the start page shows a few different important movements in chronological order. **Describe alternatives you've considered** None **What effects does your proposed solution have** We need to crawl more fields of the wikidata dataset and change the model. *This issue is closely coupled to issue #65*. Therefore the newly required data should be equal for both features. **Additional context** ![Timeline_Movement_overview](https://user-images.githubusercontent.com/28682931/69335000-2f669d00-0c5c-11ea-8fa7-e68475283d69.png)
non_test
movement overview timeline on start page is your feature request related to a problem please describe the user has no overview of the different movements an their chronological order why do you want this feature what goals should be achieved teach users the context of time and art movements describe the solution you d like adding an overview timeline containing the most important movements which are not a sub movement to the start page this information can be acquired by the wikidata dataset an empty is part of field indicates that the movement is not a sub movement itself in the future it might be possible to add more movements even sub movements but this depends on how this will blow up the start page describe acceptance criteria a timeline on the start page shows a few different important movements in chronological order describe alternatives you ve considered none what effects does your proposed solution have we need to crawl more fields of the wikidata dataset and change the model this issue is closely coupled to issue therefore the newly required data should be equal for both features additional context
0
281,731
24,415,182,586
IssuesEvent
2022-10-05 15:18:37
spring-projects/spring-framework
https://api.github.com/repos/spring-projects/spring-framework
opened
Simplify TestRuntimeHintsRegistrar API
in: test type: enhancement theme: aot
The `TestRuntimeHintsRegistrar` currently combines `MergedContextConfiguration` and test classes. However, it appears that only `spring-test` internals have a need for registering hints based on the `MergedContextConfiguration`. For example, Spring Boot's AOT testing support has not had such a need. In light of that, we should simplify the `TestRuntimeHintsRegistrar` API so that it only focuses on test classes, and we should move the hint registration code specific to `MergedContextConfiguration` to an internal mechanism.
1.0
Simplify TestRuntimeHintsRegistrar API - The `TestRuntimeHintsRegistrar` currently combines `MergedContextConfiguration` and test classes. However, it appears that only `spring-test` internals have a need for registering hints based on the `MergedContextConfiguration`. For example, Spring Boot's AOT testing support has not had such a need. In light of that, we should simplify the `TestRuntimeHintsRegistrar` API so that it only focuses on test classes, and we should move the hint registration code specific to `MergedContextConfiguration` to an internal mechanism.
test
simplify testruntimehintsregistrar api the testruntimehintsregistrar currently combines mergedcontextconfiguration and test classes however it appears that only spring test internals have a need for registering hints based on the mergedcontextconfiguration for example spring boot s aot testing support has not had such a need in light of that we should simplify the testruntimehintsregistrar api so that it only focuses on test classes and we should move the hint registration code specific to mergedcontextconfiguration to an internal mechanism
1
280,651
24,320,894,936
IssuesEvent
2022-09-30 10:37:25
modicio/modicio
https://api.github.com/repos/modicio/modicio
opened
Testbase Implementation
KPtest
Create a Unit- & Integration Testbase on top of #1 covering important functionality, focusing on assembly and dissassembly of unfolding hierarchies.
1.0
Testbase Implementation - Create a Unit- & Integration Testbase on top of #1 covering important functionality, focusing on assembly and dissassembly of unfolding hierarchies.
test
testbase implementation create a unit integration testbase on top of covering important functionality focusing on assembly and dissassembly of unfolding hierarchies
1
45,143
23,926,526,859
IssuesEvent
2022-09-10 00:02:33
pytorch/TensorRT
https://api.github.com/repos/pytorch/TensorRT
closed
❓ [Question] Why BERT Base is slower w/ Torch-TensorRT than native PyTorch?
question No Activity performance
## ❓ Question <!-- Your question --> I'm trying to optimize hugging face's BERT Base uncased model using Torch-TensorRT, the code works after disabling full compilation (`require_full_compilation=False`), and the avg latency is ~10ms on T4. However, it it slower than native PyTorch implementation (~6ms on T4). In contrast, running the same model with `trtexec` only takes ~4ms. So, for BERT Base, it's 2.5x slower than TensorRT. I wonder if this is expected? Here's the full code: ``` from transformers import BertModel, BertTokenizer, BertConfig import torch import time enc = BertTokenizer.from_pretrained("./bert-base-uncased") # Tokenizing input text text = "[CLS] Who was Jim Henson ? [SEP] Jim Henson was a puppeteer [SEP]" tokenized_text = enc.tokenize(text) # Masking one of the input tokens masked_index = 8 tokenized_text[masked_index] = '[MASK]' indexed_tokens = enc.convert_tokens_to_ids(tokenized_text) segments_ids = [0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1, 1] # Creating a dummy input tokens_tensor = torch.tensor([indexed_tokens]).to(torch.int32).cuda() segments_tensors = torch.tensor([segments_ids]).to(torch.int32).cuda() dummy_input = [tokens_tensor, segments_tensors] dummy_input_shapes = [list(v.size()) for v in dummy_input] # Initializing the model with the torchscript flag # Flag set to True even though it is not necessary as this model does not have an LM Head. config = BertConfig(vocab_size_or_config_json_file=32000, hidden_size=768, num_hidden_layers=12, num_attention_heads=12, intermediate_size=3072, torchscript=True) # Instantiating the model model = BertModel(config) # The model needs to be in evaluation mode model.eval() # If you are instantiating the model with `from_pretrained` you can also easily set the TorchScript flag model = BertModel.from_pretrained("./bert-base-uncased", torchscript=True) model = model.eval().cuda() # Creating the trace traced_model = torch.jit.trace(model, dummy_input) import torch_tensorrt compile_settings = { "require_full_compilation": False, "truncate_long_and_double": True, "torch_executed_ops": ["aten::Int"] } optimized_model = torch_tensorrt.compile(traced_model, inputs=dummy_input, **compile_settings) def benchmark(model, input): # Warming up for _ in range(10): model(*input) inference_count = 1000 # inference test start = time.time() for _ in range(inference_count): model(*input) end = time.time() print(f"use {(end-start)/inference_count*1000} ms each inference") print(f"{inference_count/(end-start)} step/s") print("before compile") benchmark(traced_model, dummy_input) print("after compile") benchmark(optimized_model, dummy_input) ``` So, my question is why it is slower than native PyTorch, and how do I fine-tune it? ## What you have already tried <!-- A clear and concise description of what you have already done. --> I've checked out the log from Torch-TensorRT, looks like the model is partitioned into 3 parts, separated by `at::Int` op, and looks like Int op is [hard to implement](https://github.com/NVIDIA/Torch-TensorRT/issues/513). Next, I profiled the inference process with Nsight System, here's the screenshot: ![CleanShot 2022-01-26 at 18 44 38](https://user-images.githubusercontent.com/552990/151149720-d707afcb-0fb0-467d-a468-b1b35eb9330a.png) It is expected to see 3 divided segments, however, there are 2 things that caught my attention: 1. Why segment 0 is slower than pure TensorRT? Is it due to over complicated conversion? 2. Why the `cudaMemcpyAsync` took so long? Shouldn't it only return the `last_hidden_state` tensor? ## Environment > Build information about Torch-TensorRT can be found by turning on debug messages - PyTorch Version (e.g., 1.0): 1.10 - CPU Architecture: - OS (e.g., Linux): Ubuntu 18.04 - How you installed PyTorch (`conda`, `pip`, `libtorch`, source): pip - Build command you used (if compiling from source): python setup.py develop - Are you using local sources or building from archives: local sources - Python version: 3.6.9 - CUDA version: 10.2 - GPU models and configuration: T4 - Any other relevant information: ## Additional context <!-- Add any other context about the problem here. -->
True
❓ [Question] Why BERT Base is slower w/ Torch-TensorRT than native PyTorch? - ## ❓ Question <!-- Your question --> I'm trying to optimize hugging face's BERT Base uncased model using Torch-TensorRT, the code works after disabling full compilation (`require_full_compilation=False`), and the avg latency is ~10ms on T4. However, it it slower than native PyTorch implementation (~6ms on T4). In contrast, running the same model with `trtexec` only takes ~4ms. So, for BERT Base, it's 2.5x slower than TensorRT. I wonder if this is expected? Here's the full code: ``` from transformers import BertModel, BertTokenizer, BertConfig import torch import time enc = BertTokenizer.from_pretrained("./bert-base-uncased") # Tokenizing input text text = "[CLS] Who was Jim Henson ? [SEP] Jim Henson was a puppeteer [SEP]" tokenized_text = enc.tokenize(text) # Masking one of the input tokens masked_index = 8 tokenized_text[masked_index] = '[MASK]' indexed_tokens = enc.convert_tokens_to_ids(tokenized_text) segments_ids = [0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1, 1] # Creating a dummy input tokens_tensor = torch.tensor([indexed_tokens]).to(torch.int32).cuda() segments_tensors = torch.tensor([segments_ids]).to(torch.int32).cuda() dummy_input = [tokens_tensor, segments_tensors] dummy_input_shapes = [list(v.size()) for v in dummy_input] # Initializing the model with the torchscript flag # Flag set to True even though it is not necessary as this model does not have an LM Head. config = BertConfig(vocab_size_or_config_json_file=32000, hidden_size=768, num_hidden_layers=12, num_attention_heads=12, intermediate_size=3072, torchscript=True) # Instantiating the model model = BertModel(config) # The model needs to be in evaluation mode model.eval() # If you are instantiating the model with `from_pretrained` you can also easily set the TorchScript flag model = BertModel.from_pretrained("./bert-base-uncased", torchscript=True) model = model.eval().cuda() # Creating the trace traced_model = torch.jit.trace(model, dummy_input) import torch_tensorrt compile_settings = { "require_full_compilation": False, "truncate_long_and_double": True, "torch_executed_ops": ["aten::Int"] } optimized_model = torch_tensorrt.compile(traced_model, inputs=dummy_input, **compile_settings) def benchmark(model, input): # Warming up for _ in range(10): model(*input) inference_count = 1000 # inference test start = time.time() for _ in range(inference_count): model(*input) end = time.time() print(f"use {(end-start)/inference_count*1000} ms each inference") print(f"{inference_count/(end-start)} step/s") print("before compile") benchmark(traced_model, dummy_input) print("after compile") benchmark(optimized_model, dummy_input) ``` So, my question is why it is slower than native PyTorch, and how do I fine-tune it? ## What you have already tried <!-- A clear and concise description of what you have already done. --> I've checked out the log from Torch-TensorRT, looks like the model is partitioned into 3 parts, separated by `at::Int` op, and looks like Int op is [hard to implement](https://github.com/NVIDIA/Torch-TensorRT/issues/513). Next, I profiled the inference process with Nsight System, here's the screenshot: ![CleanShot 2022-01-26 at 18 44 38](https://user-images.githubusercontent.com/552990/151149720-d707afcb-0fb0-467d-a468-b1b35eb9330a.png) It is expected to see 3 divided segments, however, there are 2 things that caught my attention: 1. Why segment 0 is slower than pure TensorRT? Is it due to over complicated conversion? 2. Why the `cudaMemcpyAsync` took so long? Shouldn't it only return the `last_hidden_state` tensor? ## Environment > Build information about Torch-TensorRT can be found by turning on debug messages - PyTorch Version (e.g., 1.0): 1.10 - CPU Architecture: - OS (e.g., Linux): Ubuntu 18.04 - How you installed PyTorch (`conda`, `pip`, `libtorch`, source): pip - Build command you used (if compiling from source): python setup.py develop - Are you using local sources or building from archives: local sources - Python version: 3.6.9 - CUDA version: 10.2 - GPU models and configuration: T4 - Any other relevant information: ## Additional context <!-- Add any other context about the problem here. -->
non_test
❓ why bert base is slower w torch tensorrt than native pytorch ❓ question i m trying to optimize hugging face s bert base uncased model using torch tensorrt the code works after disabling full compilation require full compilation false and the avg latency is on however it it slower than native pytorch implementation on in contrast running the same model with trtexec only takes so for bert base it s slower than tensorrt i wonder if this is expected here s the full code from transformers import bertmodel berttokenizer bertconfig import torch import time enc berttokenizer from pretrained bert base uncased tokenizing input text text who was jim henson jim henson was a puppeteer tokenized text enc tokenize text masking one of the input tokens masked index tokenized text indexed tokens enc convert tokens to ids tokenized text segments ids creating a dummy input tokens tensor torch tensor to torch cuda segments tensors torch tensor to torch cuda dummy input dummy input shapes initializing the model with the torchscript flag flag set to true even though it is not necessary as this model does not have an lm head config bertconfig vocab size or config json file hidden size num hidden layers num attention heads intermediate size torchscript true instantiating the model model bertmodel config the model needs to be in evaluation mode model eval if you are instantiating the model with from pretrained you can also easily set the torchscript flag model bertmodel from pretrained bert base uncased torchscript true model model eval cuda creating the trace traced model torch jit trace model dummy input import torch tensorrt compile settings require full compilation false truncate long and double true torch executed ops optimized model torch tensorrt compile traced model inputs dummy input compile settings def benchmark model input warming up for in range model input inference count inference test start time time for in range inference count model input end time time print f use end start inference count ms each inference print f inference count end start step s print before compile benchmark traced model dummy input print after compile benchmark optimized model dummy input so my question is why it is slower than native pytorch and how do i fine tune it what you have already tried i ve checked out the log from torch tensorrt looks like the model is partitioned into parts separated by at int op and looks like int op is next i profiled the inference process with nsight system here s the screenshot it is expected to see divided segments however there are things that caught my attention why segment is slower than pure tensorrt is it due to over complicated conversion why the cudamemcpyasync took so long shouldn t it only return the last hidden state tensor environment build information about torch tensorrt can be found by turning on debug messages pytorch version e g cpu architecture os e g linux ubuntu how you installed pytorch conda pip libtorch source pip build command you used if compiling from source python setup py develop are you using local sources or building from archives local sources python version cuda version gpu models and configuration any other relevant information additional context
0
231,044
7,622,968,114
IssuesEvent
2018-05-03 13:49:28
larray-project/larray
https://api.github.com/repos/larray-project/larray
closed
Session.summary should include non-LArray objects
difficulty: low enhancement priority: high work in progress
@gdementen should ``Session.summary`` also include Group objects?
1.0
Session.summary should include non-LArray objects - @gdementen should ``Session.summary`` also include Group objects?
non_test
session summary should include non larray objects gdementen should session summary also include group objects
0
7,105
10,441,062,259
IssuesEvent
2019-09-18 09:59:32
dan-solli/ottra
https://api.github.com/repos/dan-solli/ottra
closed
[User Story] Equipment
Requirement client server
As a user, I want to register actual objects so that I can keep track of where they are and for things-to-do to refer to them, so I know what I need to pick up to be able to complete a task. - If they are consumables, ie, there are only a certain number of AA batteries, I want the system to keep track of the amount and suggest when I should replenish them. - Equipment, some may have actions. I am thinking dish washer (run) or washing machine (run). Or vaccuum cleaner (change bag). - - An action may create a task to be completed, ie "Create task to change the bag" for a vaccuum cleaner. - - If I am to run a washing machine program, I want to be able to select the appropriate program because different programs may warrant different detergents, and they probably will run at variable lengths. For time management, this is important support. **Additional context** Add any other context or screenshots about the feature request here.
1.0
[User Story] Equipment - As a user, I want to register actual objects so that I can keep track of where they are and for things-to-do to refer to them, so I know what I need to pick up to be able to complete a task. - If they are consumables, ie, there are only a certain number of AA batteries, I want the system to keep track of the amount and suggest when I should replenish them. - Equipment, some may have actions. I am thinking dish washer (run) or washing machine (run). Or vaccuum cleaner (change bag). - - An action may create a task to be completed, ie "Create task to change the bag" for a vaccuum cleaner. - - If I am to run a washing machine program, I want to be able to select the appropriate program because different programs may warrant different detergents, and they probably will run at variable lengths. For time management, this is important support. **Additional context** Add any other context or screenshots about the feature request here.
non_test
equipment as a user i want to register actual objects so that i can keep track of where they are and for things to do to refer to them so i know what i need to pick up to be able to complete a task if they are consumables ie there are only a certain number of aa batteries i want the system to keep track of the amount and suggest when i should replenish them equipment some may have actions i am thinking dish washer run or washing machine run or vaccuum cleaner change bag an action may create a task to be completed ie create task to change the bag for a vaccuum cleaner if i am to run a washing machine program i want to be able to select the appropriate program because different programs may warrant different detergents and they probably will run at variable lengths for time management this is important support additional context add any other context or screenshots about the feature request here
0
170,662
13,196,932,686
IssuesEvent
2020-08-13 21:44:01
NVIDIA/spark-rapids
https://api.github.com/repos/NVIDIA/spark-rapids
closed
[FEA] Add some AQE-specific tests to the PySpark test suite
feature request test
**Is your feature request related to a problem? Please describe.** We currently do not have any PySpark tests that use Adaptive Query Execution (AQE). **Describe the solution you'd like** We should add some AQE tests to the PySpark test suite so that we catch regressions. **Describe alternatives you've considered** N/A **Additional context** N/A
1.0
[FEA] Add some AQE-specific tests to the PySpark test suite - **Is your feature request related to a problem? Please describe.** We currently do not have any PySpark tests that use Adaptive Query Execution (AQE). **Describe the solution you'd like** We should add some AQE tests to the PySpark test suite so that we catch regressions. **Describe alternatives you've considered** N/A **Additional context** N/A
test
add some aqe specific tests to the pyspark test suite is your feature request related to a problem please describe we currently do not have any pyspark tests that use adaptive query execution aqe describe the solution you d like we should add some aqe tests to the pyspark test suite so that we catch regressions describe alternatives you ve considered n a additional context n a
1
30,125
8,482,768,197
IssuesEvent
2018-10-25 19:30:51
bitshares/bitshares-ui
https://api.github.com/repos/bitshares/bitshares-ui
closed
Exchange getObject ["2.0.0"] timed out
[1c] Task [3] Bug [4d] Critical Priority [5b] Small [6] Staging Build
When loading the exchange for the first time you may have a request timed out for object "2.0.0" resulting in the exchange page not loading the rest of the section due to missing `feeStatus`. ![bild](https://user-images.githubusercontent.com/12114550/46918279-95e8a880-cfd0-11e8-926e-6caebe91d960.png) **Possible solution today** One way is to change node, which will send the data again and most likley have a success with it. The page will then load as expected.
1.0
Exchange getObject ["2.0.0"] timed out - When loading the exchange for the first time you may have a request timed out for object "2.0.0" resulting in the exchange page not loading the rest of the section due to missing `feeStatus`. ![bild](https://user-images.githubusercontent.com/12114550/46918279-95e8a880-cfd0-11e8-926e-6caebe91d960.png) **Possible solution today** One way is to change node, which will send the data again and most likley have a success with it. The page will then load as expected.
non_test
exchange getobject timed out when loading the exchange for the first time you may have a request timed out for object resulting in the exchange page not loading the rest of the section due to missing feestatus possible solution today one way is to change node which will send the data again and most likley have a success with it the page will then load as expected
0
116,353
24,902,791,788
IssuesEvent
2022-10-28 23:36:42
alefragnani/vscode-read-only-indicator
https://api.github.com/repos/alefragnani/vscode-read-only-indicator
closed
[FEATURE] - Be aware/evaluate how to integrate with the new read-only indicator on editor tabs
enhancement depends on vscode api adoption
<!-- Please search existing issues to avoid creating duplicates. --> <!-- Describe the feature you'd like. --> Keep looking how https://github.com/microsoft/vscode/issues/130526 will evolve. - Will it be just an indicator in editor tabs? - Will they provide an API?
1.0
[FEATURE] - Be aware/evaluate how to integrate with the new read-only indicator on editor tabs - <!-- Please search existing issues to avoid creating duplicates. --> <!-- Describe the feature you'd like. --> Keep looking how https://github.com/microsoft/vscode/issues/130526 will evolve. - Will it be just an indicator in editor tabs? - Will they provide an API?
non_test
be aware evaluate how to integrate with the new read only indicator on editor tabs keep looking how will evolve will it be just an indicator in editor tabs will they provide an api
0
1,000
3,287,541,117
IssuesEvent
2015-10-29 10:57:27
CartoDB/cartodb
https://api.github.com/repos/CartoDB/cartodb
closed
When importing through IMPORT API viz privacy is not set as dataset privacy is set on upload
0 - Backlog Data-services
According to the referenced commit: https://github.com/CartoDB/cartodb/pull/5022 "the privacy of the viz is the same as whatever the privacy param was set to." I can't see that behavior on the latest version. Setting a dataset's privacy does work but when create_vis is set to "true" on the import api the resulting visualization is always set as private regardless of the privacy setting specified through the IMPORT API call.
1.0
When importing through IMPORT API viz privacy is not set as dataset privacy is set on upload - According to the referenced commit: https://github.com/CartoDB/cartodb/pull/5022 "the privacy of the viz is the same as whatever the privacy param was set to." I can't see that behavior on the latest version. Setting a dataset's privacy does work but when create_vis is set to "true" on the import api the resulting visualization is always set as private regardless of the privacy setting specified through the IMPORT API call.
non_test
when importing through import api viz privacy is not set as dataset privacy is set on upload according to the referenced commit the privacy of the viz is the same as whatever the privacy param was set to i can t see that behavior on the latest version setting a dataset s privacy does work but when create vis is set to true on the import api the resulting visualization is always set as private regardless of the privacy setting specified through the import api call
0
54,504
6,393,284,018
IssuesEvent
2017-08-04 06:53:55
frappe/erpnext
https://api.github.com/repos/frappe/erpnext
closed
Tests related to Calculations
testing
- [x] Multi currency - [x] Multi UOM - [x] Discount on item - [x] Discount on total - [x] Tax - [x] Item wise Tax - [x] Change Price List - [x] Shipping Rule - [x] Serialized Item - [x] Batched Item
1.0
Tests related to Calculations - - [x] Multi currency - [x] Multi UOM - [x] Discount on item - [x] Discount on total - [x] Tax - [x] Item wise Tax - [x] Change Price List - [x] Shipping Rule - [x] Serialized Item - [x] Batched Item
test
tests related to calculations multi currency multi uom discount on item discount on total tax item wise tax change price list shipping rule serialized item batched item
1
368,101
10,866,193,077
IssuesEvent
2019-11-14 20:40:25
pbek/QOwnNotes
https://api.github.com/repos/pbek/QOwnNotes
closed
Highlighting of trailing spaces
enhancement priority-low
Trailing spaces will be highlighted in the note editor in the next release.
1.0
Highlighting of trailing spaces - Trailing spaces will be highlighted in the note editor in the next release.
non_test
highlighting of trailing spaces trailing spaces will be highlighted in the note editor in the next release
0
267,060
28,492,727,820
IssuesEvent
2023-04-18 12:22:36
nidhi7598/OPENSSL_1.0.2_G2.5_CVE-2023-0215
https://api.github.com/repos/nidhi7598/OPENSSL_1.0.2_G2.5_CVE-2023-0215
opened
CVE-2016-2842 (High) detected in opensslOpenSSL_1_0_2
Mend: dependency security vulnerability
## CVE-2016-2842 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>opensslOpenSSL_1_0_2</b></p></summary> <p> <p>TLS/SSL and crypto library</p> <p>Library home page: <a href=https://github.com/openssl/openssl.git>https://github.com/openssl/openssl.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/nidhi7598/OPENSSL_1.0.2_G2.5_CVE-2023-0215/commit/323924e277e1e2abec63ff28af8b21c235925c51">323924e277e1e2abec63ff28af8b21c235925c51</a></p> <p>Found in base branch: <b>main</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/crypto/bio/b_print.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> The doapr_outch function in crypto/bio/b_print.c in OpenSSL 1.0.1 before 1.0.1s and 1.0.2 before 1.0.2g does not verify that a certain memory allocation succeeds, which allows remote attackers to cause a denial of service (out-of-bounds write or memory consumption) or possibly have unspecified other impact via a long string, as demonstrated by a large amount of ASN.1 data, a different vulnerability than CVE-2016-0799. <p>Publish Date: 2016-03-03 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2016-2842>CVE-2016-2842</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2016-2842">https://nvd.nist.gov/vuln/detail/CVE-2016-2842</a></p> <p>Release Date: 2016-03-03</p> <p>Fix Resolution: 1.0.1s,1.0.2g</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2016-2842 (High) detected in opensslOpenSSL_1_0_2 - ## CVE-2016-2842 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>opensslOpenSSL_1_0_2</b></p></summary> <p> <p>TLS/SSL and crypto library</p> <p>Library home page: <a href=https://github.com/openssl/openssl.git>https://github.com/openssl/openssl.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/nidhi7598/OPENSSL_1.0.2_G2.5_CVE-2023-0215/commit/323924e277e1e2abec63ff28af8b21c235925c51">323924e277e1e2abec63ff28af8b21c235925c51</a></p> <p>Found in base branch: <b>main</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/crypto/bio/b_print.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> The doapr_outch function in crypto/bio/b_print.c in OpenSSL 1.0.1 before 1.0.1s and 1.0.2 before 1.0.2g does not verify that a certain memory allocation succeeds, which allows remote attackers to cause a denial of service (out-of-bounds write or memory consumption) or possibly have unspecified other impact via a long string, as demonstrated by a large amount of ASN.1 data, a different vulnerability than CVE-2016-0799. <p>Publish Date: 2016-03-03 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2016-2842>CVE-2016-2842</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2016-2842">https://nvd.nist.gov/vuln/detail/CVE-2016-2842</a></p> <p>Release Date: 2016-03-03</p> <p>Fix Resolution: 1.0.1s,1.0.2g</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in opensslopenssl cve high severity vulnerability vulnerable library opensslopenssl tls ssl and crypto library library home page a href found in head commit a href found in base branch main vulnerable source files crypto bio b print c vulnerability details the doapr outch function in crypto bio b print c in openssl before and before does not verify that a certain memory allocation succeeds which allows remote attackers to cause a denial of service out of bounds write or memory consumption or possibly have unspecified other impact via a long string as demonstrated by a large amount of asn data a different vulnerability than cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
2,488
2,691,825,413
IssuesEvent
2015-04-01 00:34:54
linkedin/dustjs
https://api.github.com/repos/linkedin/dustjs
closed
Tutorial or docs on asynchronous streaming interface (w/ Express)
documentation
### Use Case I'm working on an app using Node/Express and Dust. We're making a bunch of API calls in the server using Express, and then rendering the view with the data at the end. In doing so, however long it takes for the API calls to complete, that's how long it takes for the page to load. This isn't ideal, of course. Instead, we want to asynchronously pipe the data into the view from the server as it's loaded. ### The Problem I'm trying to figure out how to use the Dust streaming interface. I haven't found much information on how to implement the Dust streaming interface, specifically with a Node/Express framework. All of the docs I've seen only provide small code snippets of how bits and pieces work, but none provide the entire context. Could you guys create docs, or code samples, of how to implement this? Is there a 'Hello World' of the Dust streaming interface?
1.0
Tutorial or docs on asynchronous streaming interface (w/ Express) - ### Use Case I'm working on an app using Node/Express and Dust. We're making a bunch of API calls in the server using Express, and then rendering the view with the data at the end. In doing so, however long it takes for the API calls to complete, that's how long it takes for the page to load. This isn't ideal, of course. Instead, we want to asynchronously pipe the data into the view from the server as it's loaded. ### The Problem I'm trying to figure out how to use the Dust streaming interface. I haven't found much information on how to implement the Dust streaming interface, specifically with a Node/Express framework. All of the docs I've seen only provide small code snippets of how bits and pieces work, but none provide the entire context. Could you guys create docs, or code samples, of how to implement this? Is there a 'Hello World' of the Dust streaming interface?
non_test
tutorial or docs on asynchronous streaming interface w express use case i m working on an app using node express and dust we re making a bunch of api calls in the server using express and then rendering the view with the data at the end in doing so however long it takes for the api calls to complete that s how long it takes for the page to load this isn t ideal of course instead we want to asynchronously pipe the data into the view from the server as it s loaded the problem i m trying to figure out how to use the dust streaming interface i haven t found much information on how to implement the dust streaming interface specifically with a node express framework all of the docs i ve seen only provide small code snippets of how bits and pieces work but none provide the entire context could you guys create docs or code samples of how to implement this is there a hello world of the dust streaming interface
0
151,539
12,042,552,391
IssuesEvent
2020-04-14 10:46:46
WoWManiaUK/Redemption
https://api.github.com/repos/WoWManiaUK/Redemption
closed
Dahlia's Tears
Fix - Tester Confirmed
Starting off, I'd like to point out I am using the most current launcher for WoW Mania and my version is 2.6.0 I submitted a GM's Ticket and was aided by Divigrac who confirmed the issues and requested I submit a bug report. [Convo with Divigrac.txt](https://github.com/WoWManiaUK/Redemption/files/4472613/Convo.with.Divigrac.txt) **Links:** Quest: [Dahlia's Tears](https://www.wowhead.com/quest=13078/dahlias-tears) **What is Happening:** The Ruby Keepers are severely bugged. Their pats are insanely small or they are getting stuck in to the terrain. Bringing mobs to kill as per the questline is not causing the Keepers to aggro or trigger. We actually had to drag a mob to directly underneath its feet just for the dragon to aggro and attack the mob. However, instead of burning the mob as per the quest scripts, the Keepers simply attack them and the quest item, [Dahlia's Tears](https://www.wowhead.com/item=43084/dahlias-tears), finally spawns. **What Should happen:** There should be only a few of the Keepers flying about. Once a mob is brought underneath them, the Keeper is supposed to burn them with fire, and spawning the quest item.
1.0
Dahlia's Tears - Starting off, I'd like to point out I am using the most current launcher for WoW Mania and my version is 2.6.0 I submitted a GM's Ticket and was aided by Divigrac who confirmed the issues and requested I submit a bug report. [Convo with Divigrac.txt](https://github.com/WoWManiaUK/Redemption/files/4472613/Convo.with.Divigrac.txt) **Links:** Quest: [Dahlia's Tears](https://www.wowhead.com/quest=13078/dahlias-tears) **What is Happening:** The Ruby Keepers are severely bugged. Their pats are insanely small or they are getting stuck in to the terrain. Bringing mobs to kill as per the questline is not causing the Keepers to aggro or trigger. We actually had to drag a mob to directly underneath its feet just for the dragon to aggro and attack the mob. However, instead of burning the mob as per the quest scripts, the Keepers simply attack them and the quest item, [Dahlia's Tears](https://www.wowhead.com/item=43084/dahlias-tears), finally spawns. **What Should happen:** There should be only a few of the Keepers flying about. Once a mob is brought underneath them, the Keeper is supposed to burn them with fire, and spawning the quest item.
test
dahlia s tears starting off i d like to point out i am using the most current launcher for wow mania and my version is i submitted a gm s ticket and was aided by divigrac who confirmed the issues and requested i submit a bug report links quest what is happening the ruby keepers are severely bugged their pats are insanely small or they are getting stuck in to the terrain bringing mobs to kill as per the questline is not causing the keepers to aggro or trigger we actually had to drag a mob to directly underneath its feet just for the dragon to aggro and attack the mob however instead of burning the mob as per the quest scripts the keepers simply attack them and the quest item finally spawns what should happen there should be only a few of the keepers flying about once a mob is brought underneath them the keeper is supposed to burn them with fire and spawning the quest item
1
317,302
27,226,109,840
IssuesEvent
2023-02-21 09:49:52
KirilStrezikozin/BakeMaster-Blender-Addon
https://api.github.com/repos/KirilStrezikozin/BakeMaster-Blender-Addon
closed
REQUEST: Set up a texture set and select a map to sync all settings with
enhancement solved close on release needs testing
**This feature request is:** - [x] not a duplicate - [x] implemented **Is your feature request related to a problem? Please describe.** _atokim from BlenderMarket:_ Making resolution settings for texture sets and automatically applying them to baked maps is much faster and easier than doing it for every texture in every "baking job". It is logical to use the bake settings for each map for all texture sets automatically - there is no need to configure them individually each time - this is not a frequent case in my practice and in this case I will simply create a separate bake session. **Describe the solution you'd like to be implemented** Add functionality that will allow the user to set up a texture set and select a map to sync all settings with.
1.0
REQUEST: Set up a texture set and select a map to sync all settings with - **This feature request is:** - [x] not a duplicate - [x] implemented **Is your feature request related to a problem? Please describe.** _atokim from BlenderMarket:_ Making resolution settings for texture sets and automatically applying them to baked maps is much faster and easier than doing it for every texture in every "baking job". It is logical to use the bake settings for each map for all texture sets automatically - there is no need to configure them individually each time - this is not a frequent case in my practice and in this case I will simply create a separate bake session. **Describe the solution you'd like to be implemented** Add functionality that will allow the user to set up a texture set and select a map to sync all settings with.
test
request set up a texture set and select a map to sync all settings with this feature request is not a duplicate implemented is your feature request related to a problem please describe atokim from blendermarket making resolution settings for texture sets and automatically applying them to baked maps is much faster and easier than doing it for every texture in every baking job it is logical to use the bake settings for each map for all texture sets automatically there is no need to configure them individually each time this is not a frequent case in my practice and in this case i will simply create a separate bake session describe the solution you d like to be implemented add functionality that will allow the user to set up a texture set and select a map to sync all settings with
1
406,446
11,893,285,551
IssuesEvent
2020-03-29 10:49:01
bryntum/support
https://api.github.com/repos/bryntum/support
closed
Task incorrectly rendered after duration change
bug high-priority resolved
Go to advanced demo - Edit Install Apache task, set duration to 0 days - Edit the same again, set duration to 10 days - Task doesn't render properly and percent complete is incorrect https://www.bryntum.com/forum/viewtopic.php?f=52&t=13504 <img width="1178" alt="Снимок экрана 2020-03-18 в 16 49 41" src="https://user-images.githubusercontent.com/57486733/76968323-c3f6b880-6939-11ea-8a5d-511906443c82.png"> Or run in console: ``` gantt.taskStore.getById(11).setDuration(0) gantt.taskStore.getById(11).setDuration(10) ```
1.0
Task incorrectly rendered after duration change - Go to advanced demo - Edit Install Apache task, set duration to 0 days - Edit the same again, set duration to 10 days - Task doesn't render properly and percent complete is incorrect https://www.bryntum.com/forum/viewtopic.php?f=52&t=13504 <img width="1178" alt="Снимок экрана 2020-03-18 в 16 49 41" src="https://user-images.githubusercontent.com/57486733/76968323-c3f6b880-6939-11ea-8a5d-511906443c82.png"> Or run in console: ``` gantt.taskStore.getById(11).setDuration(0) gantt.taskStore.getById(11).setDuration(10) ```
non_test
task incorrectly rendered after duration change go to advanced demo edit install apache task set duration to days edit the same again set duration to days task doesn t render properly and percent complete is incorrect img width alt снимок экрана в src or run in console gantt taskstore getbyid setduration gantt taskstore getbyid setduration
0
11,476
3,204,653,720
IssuesEvent
2015-10-03 10:00:53
imixs/imixs-workflow
https://api.github.com/repos/imixs/imixs-workflow
closed
WorkflowSchedulerService - change timer object
bug testing
currently an error occurs when updating timer details
1.0
WorkflowSchedulerService - change timer object - currently an error occurs when updating timer details
test
workflowschedulerservice change timer object currently an error occurs when updating timer details
1
261,820
22,774,010,104
IssuesEvent
2022-07-08 12:51:11
MPMG-DCC-UFMG/F01
https://api.github.com/repos/MPMG-DCC-UFMG/F01
opened
Teste de generalizacao para a tag Informações institucionais - Link de acesso - Tabuleiro
generalization test development
DoD: Realizar o teste de Generalização do validador da tag Informações institucionais - Link de acesso para o Município de Tabuleiro.
1.0
Teste de generalizacao para a tag Informações institucionais - Link de acesso - Tabuleiro - DoD: Realizar o teste de Generalização do validador da tag Informações institucionais - Link de acesso para o Município de Tabuleiro.
test
teste de generalizacao para a tag informações institucionais link de acesso tabuleiro dod realizar o teste de generalização do validador da tag informações institucionais link de acesso para o município de tabuleiro
1
76,791
21,570,441,625
IssuesEvent
2022-05-02 07:29:08
Polymer/tools
https://api.github.com/repos/Polymer/tools
closed
Build process breaks with RxJs : this._r-- > 0
Package: build wontfix
### Description Our project uses RxJs. The command line `polymer build` seems to work ok. In the Gulp file, we are using this exactly currently: https://github.com/PolymerElements/generator-polymer-init-custom-build/blob/master/generators/app/gulpfile.js And we also tested with updated NPM modules all around `"polymer-build": "^0.7.0"` the rx.lite.min file contains a section of code that is this: `c.prototype.next = function (a) { this._r-- > 0 && (this._o.onNext(a), this._r <= 0 && this._o.onCompleted()) }` Either the `polymerBuild.PolymerProject(JSON).sources()` or the `mergeStream()` method is converting `this._r-- > 0` to `this._r==\x3e0&&` If I change `this._r-- > 0` to `(this._r-- > 0)` the build works. Please keep in mind that this is without any splitHtml or rejoinHtml methods or any other inline transformations other than sources(), dependencies() and mergeStream() ### Versions & Environment - polymer-build: 0.7.0 - node: v6.9.2 - Operating System: Mac 10.12.3 #### Steps to Reproduce `bower install rxjs` `<script src="../bower_components/rxjs/dist/rx.lite.min.js"></script>` Run gulp script from: https://github.com/PolymerElements/generator-polymer-init-custom-build/blob/master/generators/app/gulpfile.js #### Expected Results Working project with `this._r-- > 0` somewhere inside. #### Actual Results SyntaxError: Invalid or unexpected token: `this._r==\x3e0&&`
1.0
Build process breaks with RxJs : this._r-- > 0 - ### Description Our project uses RxJs. The command line `polymer build` seems to work ok. In the Gulp file, we are using this exactly currently: https://github.com/PolymerElements/generator-polymer-init-custom-build/blob/master/generators/app/gulpfile.js And we also tested with updated NPM modules all around `"polymer-build": "^0.7.0"` the rx.lite.min file contains a section of code that is this: `c.prototype.next = function (a) { this._r-- > 0 && (this._o.onNext(a), this._r <= 0 && this._o.onCompleted()) }` Either the `polymerBuild.PolymerProject(JSON).sources()` or the `mergeStream()` method is converting `this._r-- > 0` to `this._r==\x3e0&&` If I change `this._r-- > 0` to `(this._r-- > 0)` the build works. Please keep in mind that this is without any splitHtml or rejoinHtml methods or any other inline transformations other than sources(), dependencies() and mergeStream() ### Versions & Environment - polymer-build: 0.7.0 - node: v6.9.2 - Operating System: Mac 10.12.3 #### Steps to Reproduce `bower install rxjs` `<script src="../bower_components/rxjs/dist/rx.lite.min.js"></script>` Run gulp script from: https://github.com/PolymerElements/generator-polymer-init-custom-build/blob/master/generators/app/gulpfile.js #### Expected Results Working project with `this._r-- > 0` somewhere inside. #### Actual Results SyntaxError: Invalid or unexpected token: `this._r==\x3e0&&`
non_test
build process breaks with rxjs this r description our project uses rxjs the command line polymer build seems to work ok in the gulp file we are using this exactly currently and we also tested with updated npm modules all around polymer build the rx lite min file contains a section of code that is this c prototype next function a this r this o onnext a this r this o oncompleted either the polymerbuild polymerproject json sources or the mergestream method is converting this r to this r if i change this r to this r the build works please keep in mind that this is without any splithtml or rejoinhtml methods or any other inline transformations other than sources dependencies and mergestream versions environment polymer build node operating system mac steps to reproduce bower install rxjs run gulp script from expected results working project with this r somewhere inside actual results syntaxerror invalid or unexpected token this r
0
215,905
16,721,542,773
IssuesEvent
2021-06-10 07:56:55
SAPDocuments/Issues
https://api.github.com/repos/SAPDocuments/Issues
closed
Perform Actions On Cards To Support Your Business Case
High-Prio SCPTest-2105A SCPTest-mcard SCPTest-trial1 SCPTest-trial2 SCPTest-trial3
**Tutorials: https://developers.sap.com/tutorials/cp-mobile-cards-booster-starter.html** Steps: **Step 3: View supplier contact card on your mobile device Step 4: View sales order approval card on your mobile device** -------------------------- **Issue:** Contact Card and Sales Order Approval Card are auto subscribed, So please remove the steps to subscribe to these cards in Step 3 and Step 4. Instead add a note that if the cards are not visible, then we have to unsubscribe and subscribe to these cards again and then do a pull refresh for the cards to be visible. Best Regards, Priyanka
5.0
Perform Actions On Cards To Support Your Business Case - **Tutorials: https://developers.sap.com/tutorials/cp-mobile-cards-booster-starter.html** Steps: **Step 3: View supplier contact card on your mobile device Step 4: View sales order approval card on your mobile device** -------------------------- **Issue:** Contact Card and Sales Order Approval Card are auto subscribed, So please remove the steps to subscribe to these cards in Step 3 and Step 4. Instead add a note that if the cards are not visible, then we have to unsubscribe and subscribe to these cards again and then do a pull refresh for the cards to be visible. Best Regards, Priyanka
test
perform actions on cards to support your business case tutorials steps step view supplier contact card on your mobile device step view sales order approval card on your mobile device issue contact card and sales order approval card are auto subscribed so please remove the steps to subscribe to these cards in step and step instead add a note that if the cards are not visible then we have to unsubscribe and subscribe to these cards again and then do a pull refresh for the cards to be visible best regards priyanka
1
23,449
4,018,888,593
IssuesEvent
2016-05-16 12:54:24
CyclopsMC/EvilCraft
https://api.github.com/repos/CyclopsMC/EvilCraft
opened
Bare Cap/Brush/Rod without modifiers
alpha-testing enhancement
When shift-right clicking a bare cap without modifiers reads "Modifiers:" but doesn't list anything. It would be prettier if it read "no modifiers found" or something along that line
1.0
Bare Cap/Brush/Rod without modifiers - When shift-right clicking a bare cap without modifiers reads "Modifiers:" but doesn't list anything. It would be prettier if it read "no modifiers found" or something along that line
test
bare cap brush rod without modifiers when shift right clicking a bare cap without modifiers reads modifiers but doesn t list anything it would be prettier if it read no modifiers found or something along that line
1
276,396
23,990,991,284
IssuesEvent
2022-09-14 01:03:54
hvac/hvac
https://api.github.com/repos/hvac/hvac
closed
TestOIDC Failure in Vault 1.9.0
help wanted jwt/oidc test failures
Vault 1.9.0 introduced the following error in test_oidc.py. Creating an issue to track the resolution of this failure. ``` FAILED tests/integration_tests/api/auth_methods/test_oidc.py::TestOIDC::test_oidc_authorization_url_request_0_success - hvac.exceptions.InvalidRequest: cannot find key "oidc-test-key", on post https://localhost:8200/v1/identity/oidc/role/hvac-oidc-test ```
1.0
TestOIDC Failure in Vault 1.9.0 - Vault 1.9.0 introduced the following error in test_oidc.py. Creating an issue to track the resolution of this failure. ``` FAILED tests/integration_tests/api/auth_methods/test_oidc.py::TestOIDC::test_oidc_authorization_url_request_0_success - hvac.exceptions.InvalidRequest: cannot find key "oidc-test-key", on post https://localhost:8200/v1/identity/oidc/role/hvac-oidc-test ```
test
testoidc failure in vault vault introduced the following error in test oidc py creating an issue to track the resolution of this failure failed tests integration tests api auth methods test oidc py testoidc test oidc authorization url request success hvac exceptions invalidrequest cannot find key oidc test key on post
1
642,526
20,906,464,227
IssuesEvent
2022-03-24 03:08:18
o3de/o3de
https://api.github.com/repos/o3de/o3de
closed
[Material Editor] A crash happened when clicking a material file in Asset Browser within Editor to open it in the Material Editor.
kind/bug sig/graphics-audio priority/critical feature/graphics/tools
**Describe the bug** A clear and concise description of what the bug is. Try to isolate the issue to help the community to reproduce it easily and increase chances for a fast fix. **Steps to reproduce** Steps to reproduce the behavior: 1. Open O3DE.sln with Visual Studio 2019. 2. Set Editor as startup project. 3. Press F5 to start debug Editor. 4. When Editor is ready, click on 'Tools -> Material Editor' to open Material Editor. 5. Wait for a few seconds, then Material Editor is ready to work. 6. Browse files in Asset Browser within Editor, click a material file to view its content in Material Editor. 7. Check Process View in Visual Studio 2019, a second Material Editor process is running and failing to exit. 8. See error in Visual Studio 2019. **Expected behavior** The second Material Editor process could exit after its job is done. **Actual behavior** The second Material Editor process fails to exit normally. It crashed at end of `~AtomToolsApplication()` in AtomToolsApplication.cpp. **Assets required** NA **Screenshots/Video** NA **Found in Branch** development main stabilization/2111RTE **Desktop/Device (please complete the following information):** - Device: [PC] - OS: [Windows] - Version [10] - CPU [Intel I9-9900k] - GPU [NVidia RTX 3090] - Memory [32GB] **Additional context** Add any other context about the problem here.
1.0
[Material Editor] A crash happened when clicking a material file in Asset Browser within Editor to open it in the Material Editor. - **Describe the bug** A clear and concise description of what the bug is. Try to isolate the issue to help the community to reproduce it easily and increase chances for a fast fix. **Steps to reproduce** Steps to reproduce the behavior: 1. Open O3DE.sln with Visual Studio 2019. 2. Set Editor as startup project. 3. Press F5 to start debug Editor. 4. When Editor is ready, click on 'Tools -> Material Editor' to open Material Editor. 5. Wait for a few seconds, then Material Editor is ready to work. 6. Browse files in Asset Browser within Editor, click a material file to view its content in Material Editor. 7. Check Process View in Visual Studio 2019, a second Material Editor process is running and failing to exit. 8. See error in Visual Studio 2019. **Expected behavior** The second Material Editor process could exit after its job is done. **Actual behavior** The second Material Editor process fails to exit normally. It crashed at end of `~AtomToolsApplication()` in AtomToolsApplication.cpp. **Assets required** NA **Screenshots/Video** NA **Found in Branch** development main stabilization/2111RTE **Desktop/Device (please complete the following information):** - Device: [PC] - OS: [Windows] - Version [10] - CPU [Intel I9-9900k] - GPU [NVidia RTX 3090] - Memory [32GB] **Additional context** Add any other context about the problem here.
non_test
a crash happened when clicking a material file in asset browser within editor to open it in the material editor describe the bug a clear and concise description of what the bug is try to isolate the issue to help the community to reproduce it easily and increase chances for a fast fix steps to reproduce steps to reproduce the behavior open sln with visual studio set editor as startup project press to start debug editor when editor is ready click on tools material editor to open material editor wait for a few seconds then material editor is ready to work browse files in asset browser within editor click a material file to view its content in material editor check process view in visual studio a second material editor process is running and failing to exit see error in visual studio expected behavior the second material editor process could exit after its job is done actual behavior the second material editor process fails to exit normally it crashed at end of atomtoolsapplication in atomtoolsapplication cpp assets required na screenshots video na found in branch development main stabilization desktop device please complete the following information device os version cpu gpu memory additional context add any other context about the problem here
0
106,298
16,673,279,551
IssuesEvent
2021-06-07 13:31:14
VivekBuzruk/Hygieia
https://api.github.com/repos/VivekBuzruk/Hygieia
closed
WS-2019-0063 (High) detected in js-yaml-3.6.1.tgz - autoclosed
security vulnerability
## WS-2019-0063 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>js-yaml-3.6.1.tgz</b></p></summary> <p>YAML 1.2 parser and serializer</p> <p>Library home page: <a href="https://registry.npmjs.org/js-yaml/-/js-yaml-3.6.1.tgz">https://registry.npmjs.org/js-yaml/-/js-yaml-3.6.1.tgz</a></p> <p>Path to dependency file: Hygieia/UI/package.json</p> <p>Path to vulnerable library: Hygieia/UI/node_modules/js-yaml/package.json</p> <p> Dependency Hierarchy: - gulp-yaml-1.0.1.tgz (Root Library) - :x: **js-yaml-3.6.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/VivekBuzruk/Hygieia/commit/3c4f119e4343cf7fa276bb4756361b926902248e">3c4f119e4343cf7fa276bb4756361b926902248e</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Js-yaml prior to 3.13.1 are vulnerable to Code Injection. The load() function may execute arbitrary code injected through a malicious YAML file. <p>Publish Date: 2019-04-05 <p>URL: <a href=https://github.com/nodeca/js-yaml/pull/480>WS-2019-0063</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>8.0</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/813">https://www.npmjs.com/advisories/813</a></p> <p>Release Date: 2019-04-05</p> <p>Fix Resolution: js-yaml - 3.13.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2019-0063 (High) detected in js-yaml-3.6.1.tgz - autoclosed - ## WS-2019-0063 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>js-yaml-3.6.1.tgz</b></p></summary> <p>YAML 1.2 parser and serializer</p> <p>Library home page: <a href="https://registry.npmjs.org/js-yaml/-/js-yaml-3.6.1.tgz">https://registry.npmjs.org/js-yaml/-/js-yaml-3.6.1.tgz</a></p> <p>Path to dependency file: Hygieia/UI/package.json</p> <p>Path to vulnerable library: Hygieia/UI/node_modules/js-yaml/package.json</p> <p> Dependency Hierarchy: - gulp-yaml-1.0.1.tgz (Root Library) - :x: **js-yaml-3.6.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/VivekBuzruk/Hygieia/commit/3c4f119e4343cf7fa276bb4756361b926902248e">3c4f119e4343cf7fa276bb4756361b926902248e</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Js-yaml prior to 3.13.1 are vulnerable to Code Injection. The load() function may execute arbitrary code injected through a malicious YAML file. <p>Publish Date: 2019-04-05 <p>URL: <a href=https://github.com/nodeca/js-yaml/pull/480>WS-2019-0063</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>8.0</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/813">https://www.npmjs.com/advisories/813</a></p> <p>Release Date: 2019-04-05</p> <p>Fix Resolution: js-yaml - 3.13.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
ws high detected in js yaml tgz autoclosed ws high severity vulnerability vulnerable library js yaml tgz yaml parser and serializer library home page a href path to dependency file hygieia ui package json path to vulnerable library hygieia ui node modules js yaml package json dependency hierarchy gulp yaml tgz root library x js yaml tgz vulnerable library found in head commit a href found in base branch master vulnerability details js yaml prior to are vulnerable to code injection the load function may execute arbitrary code injected through a malicious yaml file publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution js yaml step up your open source security game with whitesource
0
204,660
15,526,590,098
IssuesEvent
2021-03-13 01:50:45
IntellectualSites/FastAsyncWorldEdit
https://api.github.com/repos/IntellectualSites/FastAsyncWorldEdit
opened
//br circle does not seem to follow player view for placement
Requires Testing
### Server Implementation Tuinity ### Server Version 1.16.5 ### Describe the bug `//br circle` doesn't seem to respect the direction the player is looking for circle placement. ### To Reproduce Type `//br circle` Click in various directions and they all seem to be ambiguous ### Expected behaviour Circle plane should be placed perpendicular to the player's direction of view ### Screenshots / Videos This was placed at an angle while directly facing the block: https://brdv.coffee/storage/uploads/4/GedIBEdXQjY0.png (verified 0 degree angle to block with F3 player facing indicator) Submitted by a builder on the server: https://imgur.com/a/F0lqzHe Circles seem to all be oriented in the same direction ### Error log (if applicable) _No response_ ### Fawe Debugpaste https://athion.net/ISPaster/paste/view/4f8ce65fc0a8467dbdfd2cec2b6933c6 ### Fawe Version 1.16-618;c8db3e0
1.0
//br circle does not seem to follow player view for placement - ### Server Implementation Tuinity ### Server Version 1.16.5 ### Describe the bug `//br circle` doesn't seem to respect the direction the player is looking for circle placement. ### To Reproduce Type `//br circle` Click in various directions and they all seem to be ambiguous ### Expected behaviour Circle plane should be placed perpendicular to the player's direction of view ### Screenshots / Videos This was placed at an angle while directly facing the block: https://brdv.coffee/storage/uploads/4/GedIBEdXQjY0.png (verified 0 degree angle to block with F3 player facing indicator) Submitted by a builder on the server: https://imgur.com/a/F0lqzHe Circles seem to all be oriented in the same direction ### Error log (if applicable) _No response_ ### Fawe Debugpaste https://athion.net/ISPaster/paste/view/4f8ce65fc0a8467dbdfd2cec2b6933c6 ### Fawe Version 1.16-618;c8db3e0
test
br circle does not seem to follow player view for placement server implementation tuinity server version describe the bug br circle doesn t seem to respect the direction the player is looking for circle placement to reproduce type br circle click in various directions and they all seem to be ambiguous expected behaviour circle plane should be placed perpendicular to the player s direction of view screenshots videos this was placed at an angle while directly facing the block verified degree angle to block with player facing indicator submitted by a builder on the server circles seem to all be oriented in the same direction error log if applicable no response fawe debugpaste fawe version
1
65,221
27,020,248,422
IssuesEvent
2023-02-11 00:31:55
devssa/onde-codar-em-salvador
https://api.github.com/repos/devssa/onde-codar-em-salvador
closed
[REMOTO - CLT] [BACKEND SR e PL] - Pessoa Desenvolvedora Back End nível Sênior e Pleno
BACK-END EFETIVO(CLT) PHP PYTHON PLENO SENIOR SOAP NODE.JS JSON XML REMOTO WEBSERVICES APIs HELP WANTED Stale
Estamos com oportunidades incríveis para o **time da MadeiraMadeira, pioneira e líder no segmento de Home Center**, atua com e-commerce, marketplace e também canais offline. A startup recebeu um investimento de US$190 milhões e se tornou o **16° unicórnio brasileiro**. Buscamos talentos que tenham propósitos, sejam proativos e tenham vontade de aprender e crescer exponencialmente. Para essa oportunidade são necessários conhecimentos em: -**Python, NodeJS ou PHP**; -**Desenvolvimento APIs**; -**Desenvolvimento de Integração webservice (JSON, XML/ SOAP)**; -Entre outros. Se você tem vontade de fazer parte de um time que oferece espaços e oportunidades reais de desenvolvimento pessoal e profissional, dá uma olhada no link a seguir: https://byintera.in/1z3 Um sorriso de orelha a orelha =D
1.0
[REMOTO - CLT] [BACKEND SR e PL] - Pessoa Desenvolvedora Back End nível Sênior e Pleno - Estamos com oportunidades incríveis para o **time da MadeiraMadeira, pioneira e líder no segmento de Home Center**, atua com e-commerce, marketplace e também canais offline. A startup recebeu um investimento de US$190 milhões e se tornou o **16° unicórnio brasileiro**. Buscamos talentos que tenham propósitos, sejam proativos e tenham vontade de aprender e crescer exponencialmente. Para essa oportunidade são necessários conhecimentos em: -**Python, NodeJS ou PHP**; -**Desenvolvimento APIs**; -**Desenvolvimento de Integração webservice (JSON, XML/ SOAP)**; -Entre outros. Se você tem vontade de fazer parte de um time que oferece espaços e oportunidades reais de desenvolvimento pessoal e profissional, dá uma olhada no link a seguir: https://byintera.in/1z3 Um sorriso de orelha a orelha =D
non_test
pessoa desenvolvedora back end nível sênior e pleno estamos com oportunidades incríveis para o time da madeiramadeira pioneira e líder no segmento de home center atua com e commerce marketplace e também canais offline a startup recebeu um investimento de us milhões e se tornou o ° unicórnio brasileiro buscamos talentos que tenham propósitos sejam proativos e tenham vontade de aprender e crescer exponencialmente para essa oportunidade são necessários conhecimentos em python nodejs ou php desenvolvimento apis desenvolvimento de integração webservice json xml soap entre outros se você tem vontade de fazer parte de um time que oferece espaços e oportunidades reais de desenvolvimento pessoal e profissional dá uma olhada no link a seguir um sorriso de orelha a orelha d
0
284,273
8,736,876,100
IssuesEvent
2018-12-11 20:50:14
aowen87/TicketTester
https://api.github.com/repos/aowen87/TicketTester
closed
Change cielo host profile to use rfta15.lanl.gov for the gateway.
bug likelihood medium priority reviewed severity medium
The old redfta cluster used as the gateway to lanl has been retired and replaced with the rfta cluster. The valid nodes are rfta13 rfta28. 15 seems like a good choice. -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. As such, not all information was able to be captured in the transition. Below is a complete record of the original redmine ticket. Ticket number: 1530 Status: Resolved Project: VisIt Tracker: Bug Priority: Immediate Subject: Change cielo host profile to use rfta15.lanl.gov for the gateway. Assigned to: Eric Brugger Category: Target version: 2.6.3 Author: Eric Brugger Start: 07/08/2013 Due date: % Done: 100 Estimated time: 1.0 Created: 07/08/2013 12:59 pm Updated: 07/08/2013 01:30 pm Likelihood: 3 - Occasional Severity: 3 - Major Irritation Found in version: 2.6.2 Impact: Expected Use: OS: All Support Group: Any Description: The old redfta cluster used as the gateway to lanl has been retired and replaced with the rfta cluster. The valid nodes are rfta13 rfta28. 15 seems like a good choice. Comments: Ladies and Gentlemen,I committed revisions 21309 and 21311 the 2.6 RC and trunk with thefollowing change:1) I changed the gateway host to rfta15.lanl.gov in the host profile for LANL's cielo system since the old gateway cluster was retired. This resolved #1530.M resources/help/en_US/relnotes2.6.3.htmlM resources/hosts/llnl_closed/host_lanl_closed_cielo.xmlEric
1.0
Change cielo host profile to use rfta15.lanl.gov for the gateway. - The old redfta cluster used as the gateway to lanl has been retired and replaced with the rfta cluster. The valid nodes are rfta13 rfta28. 15 seems like a good choice. -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. As such, not all information was able to be captured in the transition. Below is a complete record of the original redmine ticket. Ticket number: 1530 Status: Resolved Project: VisIt Tracker: Bug Priority: Immediate Subject: Change cielo host profile to use rfta15.lanl.gov for the gateway. Assigned to: Eric Brugger Category: Target version: 2.6.3 Author: Eric Brugger Start: 07/08/2013 Due date: % Done: 100 Estimated time: 1.0 Created: 07/08/2013 12:59 pm Updated: 07/08/2013 01:30 pm Likelihood: 3 - Occasional Severity: 3 - Major Irritation Found in version: 2.6.2 Impact: Expected Use: OS: All Support Group: Any Description: The old redfta cluster used as the gateway to lanl has been retired and replaced with the rfta cluster. The valid nodes are rfta13 rfta28. 15 seems like a good choice. Comments: Ladies and Gentlemen,I committed revisions 21309 and 21311 the 2.6 RC and trunk with thefollowing change:1) I changed the gateway host to rfta15.lanl.gov in the host profile for LANL's cielo system since the old gateway cluster was retired. This resolved #1530.M resources/help/en_US/relnotes2.6.3.htmlM resources/hosts/llnl_closed/host_lanl_closed_cielo.xmlEric
non_test
change cielo host profile to use lanl gov for the gateway the old redfta cluster used as the gateway to lanl has been retired and replaced with the rfta cluster the valid nodes are seems like a good choice redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker bug priority immediate subject change cielo host profile to use lanl gov for the gateway assigned to eric brugger category target version author eric brugger start due date done estimated time created pm updated pm likelihood occasional severity major irritation found in version impact expected use os all support group any description the old redfta cluster used as the gateway to lanl has been retired and replaced with the rfta cluster the valid nodes are seems like a good choice comments ladies and gentlemen i committed revisions and the rc and trunk with thefollowing change i changed the gateway host to lanl gov in the host profile for lanl s cielo system since the old gateway cluster was retired this resolved m resources help en us htmlm resources hosts llnl closed host lanl closed cielo xmleric
0
9,978
3,080,714,950
IssuesEvent
2015-08-22 01:02:58
rancher/rancher
https://api.github.com/repos/rancher/rancher
closed
Failed to access LB rules even though correct rules were programmed.
area/networking kind/bug status/to-test
Server version - V0.25.0-rc1 Create a LB service linked to a service -S1 with scale 3. I was able to access the Lb rules successfully. Scaled down the service to 1. After this , i am not able to access the LB rule: Get the following error: Sangeethas-MBP:testwgets sangeethahariharan1$ wget http://146.148.82.227:7777/name.html --2015-06-19 21:21:05-- http://146.148.82.227:7777/name.html Connecting to 146.148.82.227:7777... connected. HTTP request sent, awaiting response... 503 Service Unavailable 2015-06-19 21:21:05 ERROR 503: Service Unavailable. I see that the rules are being programmed correctly in /etc/haproxy/haproxy.cfg file , but still LB rules are not accessible. From the LB Agent , When I tried to ping to the contaier Ip address the LB rule resolves to , I am not able to ping. There seems to be problems with network connectivity between the LB agent and the container.
1.0
Failed to access LB rules even though correct rules were programmed. - Server version - V0.25.0-rc1 Create a LB service linked to a service -S1 with scale 3. I was able to access the Lb rules successfully. Scaled down the service to 1. After this , i am not able to access the LB rule: Get the following error: Sangeethas-MBP:testwgets sangeethahariharan1$ wget http://146.148.82.227:7777/name.html --2015-06-19 21:21:05-- http://146.148.82.227:7777/name.html Connecting to 146.148.82.227:7777... connected. HTTP request sent, awaiting response... 503 Service Unavailable 2015-06-19 21:21:05 ERROR 503: Service Unavailable. I see that the rules are being programmed correctly in /etc/haproxy/haproxy.cfg file , but still LB rules are not accessible. From the LB Agent , When I tried to ping to the contaier Ip address the LB rule resolves to , I am not able to ping. There seems to be problems with network connectivity between the LB agent and the container.
test
failed to access lb rules even though correct rules were programmed server version create a lb service linked to a service with scale i was able to access the lb rules successfully scaled down the service to after this i am not able to access the lb rule get the following error sangeethas mbp testwgets wget connecting to connected http request sent awaiting response service unavailable error service unavailable i see that the rules are being programmed correctly in etc haproxy haproxy cfg file but still lb rules are not accessible from the lb agent when i tried to ping to the contaier ip address the lb rule resolves to i am not able to ping there seems to be problems with network connectivity between the lb agent and the container
1
399,532
11,756,545,285
IssuesEvent
2020-03-13 11:49:07
airshipit/airshipctl
https://api.github.com/repos/airshipit/airshipctl
closed
Enhance ephemeral node iso-generation user and netconfig selection process
component/bootstrap enhancement priority/critical ready for review
**Problem description (if applicable)** Currently, the airshipctl iso-gen process makes a [call](https://github.com/airshipit/airshipctl/blob/9a47c9b423716836383c24c071df3ae840baeb32/pkg/bootstrap/isogen/command.go#L121) to the [cloud-init](https://github.com/airshipit/airshipctl/blob/9a47c9b423716836383c24c071df3ae840baeb32/pkg/bootstrap/cloudinit/cloud-init.go) pkg to generate the cloud-init for the ephemeral host. This looks up the appropriate user data and network data for the ephemeral host by finding a `Kind: Secret` in the document set with the appropriate annotation. This has several issues: - If more than one secret object exists with the ephemeral target annotation, this code as-is won't work as it won't know which to select. This is quite likely as the ephemeral manifest set starts to grow. - This may be incompatible with the current in-flight implementation allowing networkData for baremetal hosts to be defined. The following [pull request](https://github.com/metal3-io/baremetal-operator/pull/348/files#diff-3507adac177799b23dc0f7f5ad4ffc5fR74) in the BMO leverages a secret with a `networkData` key, however we use a `netplan` key the current implementation meaning we could not share/re-use this `Secret` between the BMO and airshipctl, resulting in a duplicate ephemeral host network configuration somewhere. **Proposed change** Originally, we wanted to drive ephemeral node selection by indicating a specific `BareMetalHost` as the ephemeral node which would allow code such as the cloud-init pkg to back into the appropriate networkData and any other host-specific information easily without duplicating that data for ephemeral iso generation. If one of the hosts had such an annotation, walking its `spec` back to the actual `Secret` containing the appropriate `networkData` for that host should be fairly straight forward and would not run into any conflict issues. This would also allow us a path to get that hosts BMC information for power control as well, avoiding the need to specify this separately in Airship configuration.
1.0
Enhance ephemeral node iso-generation user and netconfig selection process - **Problem description (if applicable)** Currently, the airshipctl iso-gen process makes a [call](https://github.com/airshipit/airshipctl/blob/9a47c9b423716836383c24c071df3ae840baeb32/pkg/bootstrap/isogen/command.go#L121) to the [cloud-init](https://github.com/airshipit/airshipctl/blob/9a47c9b423716836383c24c071df3ae840baeb32/pkg/bootstrap/cloudinit/cloud-init.go) pkg to generate the cloud-init for the ephemeral host. This looks up the appropriate user data and network data for the ephemeral host by finding a `Kind: Secret` in the document set with the appropriate annotation. This has several issues: - If more than one secret object exists with the ephemeral target annotation, this code as-is won't work as it won't know which to select. This is quite likely as the ephemeral manifest set starts to grow. - This may be incompatible with the current in-flight implementation allowing networkData for baremetal hosts to be defined. The following [pull request](https://github.com/metal3-io/baremetal-operator/pull/348/files#diff-3507adac177799b23dc0f7f5ad4ffc5fR74) in the BMO leverages a secret with a `networkData` key, however we use a `netplan` key the current implementation meaning we could not share/re-use this `Secret` between the BMO and airshipctl, resulting in a duplicate ephemeral host network configuration somewhere. **Proposed change** Originally, we wanted to drive ephemeral node selection by indicating a specific `BareMetalHost` as the ephemeral node which would allow code such as the cloud-init pkg to back into the appropriate networkData and any other host-specific information easily without duplicating that data for ephemeral iso generation. If one of the hosts had such an annotation, walking its `spec` back to the actual `Secret` containing the appropriate `networkData` for that host should be fairly straight forward and would not run into any conflict issues. This would also allow us a path to get that hosts BMC information for power control as well, avoiding the need to specify this separately in Airship configuration.
non_test
enhance ephemeral node iso generation user and netconfig selection process problem description if applicable currently the airshipctl iso gen process makes a to the pkg to generate the cloud init for the ephemeral host this looks up the appropriate user data and network data for the ephemeral host by finding a kind secret in the document set with the appropriate annotation this has several issues if more than one secret object exists with the ephemeral target annotation this code as is won t work as it won t know which to select this is quite likely as the ephemeral manifest set starts to grow this may be incompatible with the current in flight implementation allowing networkdata for baremetal hosts to be defined the following in the bmo leverages a secret with a networkdata key however we use a netplan key the current implementation meaning we could not share re use this secret between the bmo and airshipctl resulting in a duplicate ephemeral host network configuration somewhere proposed change originally we wanted to drive ephemeral node selection by indicating a specific baremetalhost as the ephemeral node which would allow code such as the cloud init pkg to back into the appropriate networkdata and any other host specific information easily without duplicating that data for ephemeral iso generation if one of the hosts had such an annotation walking its spec back to the actual secret containing the appropriate networkdata for that host should be fairly straight forward and would not run into any conflict issues this would also allow us a path to get that hosts bmc information for power control as well avoiding the need to specify this separately in airship configuration
0
69,993
7,167,285,443
IssuesEvent
2018-01-29 20:01:06
EasyCrossPlatformLib/EasyCrossPlatform_src
https://api.github.com/repos/EasyCrossPlatformLib/EasyCrossPlatform_src
opened
Bug Examination for 1.0 Release
help wanted test
I am planning to release the stable version of EasyCrossPlatform 1.0 Examination shall be needed to examine the stability of this library. I will need people to help me with it. I will start testing with raspberry pi(ARMv6), and Windows 10(x86, x64), Linux(Ubuntu 16.04LTS x64), MacOS(High Sierra x64). Anyone can do the linux x86 test for me?
1.0
Bug Examination for 1.0 Release - I am planning to release the stable version of EasyCrossPlatform 1.0 Examination shall be needed to examine the stability of this library. I will need people to help me with it. I will start testing with raspberry pi(ARMv6), and Windows 10(x86, x64), Linux(Ubuntu 16.04LTS x64), MacOS(High Sierra x64). Anyone can do the linux x86 test for me?
test
bug examination for release i am planning to release the stable version of easycrossplatform examination shall be needed to examine the stability of this library i will need people to help me with it i will start testing with raspberry pi and windows linux ubuntu macos high sierra anyone can do the linux test for me
1
67,827
7,064,796,078
IssuesEvent
2018-01-06 12:01:06
systemetric/sr-robot
https://api.github.com/repos/systemetric/sr-robot
closed
Support for logging images
enhancement needs testing
Shepherd should be able to pass an `--image-file` command-line argument that allows sr.robot to save its images to the filesystem – this will allow us to e.g. show images in the web UI, or to save them to a USB stick. So that we can potentially do both of these at once, sr.robot should be able to accept multiple places to save the image to.
1.0
Support for logging images - Shepherd should be able to pass an `--image-file` command-line argument that allows sr.robot to save its images to the filesystem – this will allow us to e.g. show images in the web UI, or to save them to a USB stick. So that we can potentially do both of these at once, sr.robot should be able to accept multiple places to save the image to.
test
support for logging images shepherd should be able to pass an image file command line argument that allows sr robot to save its images to the filesystem – this will allow us to e g show images in the web ui or to save them to a usb stick so that we can potentially do both of these at once sr robot should be able to accept multiple places to save the image to
1
600,837
18,359,680,759
IssuesEvent
2021-10-09 02:22:12
googleapis/google-cloud-go
https://api.github.com/repos/googleapis/google-cloud-go
closed
bigquery/storage/managedwriter: TestIntegration_ProtoNormalization failed
type: bug api: bigquery priority: p1 flakybot: issue
This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: 7ac8489d9344bf15a66691ab6220924a5b208066 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/c1d38e81-2b3c-459b-a3be-c3ec90801e49), [Sponge](http://sponge2/c1d38e81-2b3c-459b-a3be-c3ec90801e49) status: failed <details><summary>Test output</summary><br><pre> integration_test.go:617: error in response: rpc error: code = ResourceExhausted desc = Resource has been exhausted (e.g. check quota).</pre></details>
1.0
bigquery/storage/managedwriter: TestIntegration_ProtoNormalization failed - This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: 7ac8489d9344bf15a66691ab6220924a5b208066 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/c1d38e81-2b3c-459b-a3be-c3ec90801e49), [Sponge](http://sponge2/c1d38e81-2b3c-459b-a3be-c3ec90801e49) status: failed <details><summary>Test output</summary><br><pre> integration_test.go:617: error in response: rpc error: code = ResourceExhausted desc = Resource has been exhausted (e.g. check quota).</pre></details>
non_test
bigquery storage managedwriter testintegration protonormalization failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output integration test go error in response rpc error code resourceexhausted desc resource has been exhausted e g check quota
0
100,651
8,752,307,441
IssuesEvent
2018-12-14 02:20:17
actiontech/dble
https://api.github.com/repos/actiontech/dble
closed
error when join cross databases
from auto_test resolve
* **dble version: 5.6.29-dble-9.9.9.9-79933e5-20181211102340** * **preconditions :** lower_case_table_names =1 * **configs:** **schema.xml** ``` <schema dataNode="dn5" name="mytest" sqlMaxLimit="100"> <table dataNode="dn1,dn2,dn3,dn4" name="test" type="global"/> </schema> <schema name="DBTEST"> <table name="Test_Table" dataNode="dn1,dn2,dn3,dn4" rule="hash-four"/> </schema> ``` * **steps:** step1. execute in dble ``` mysql> use mytest; Database changed mysql> select t1.id from dbtest.test_table t1 left join test t2 on t1.id=t2.id; ERROR 4004 (42S22): column t1.id not found ``` * **expect result:** 1. query success * **real result:** 1.query failed * **supplements:** 1. auto_test position:lower_case_table_names.feature --line66-68 /label ~BUG
1.0
error when join cross databases - * **dble version: 5.6.29-dble-9.9.9.9-79933e5-20181211102340** * **preconditions :** lower_case_table_names =1 * **configs:** **schema.xml** ``` <schema dataNode="dn5" name="mytest" sqlMaxLimit="100"> <table dataNode="dn1,dn2,dn3,dn4" name="test" type="global"/> </schema> <schema name="DBTEST"> <table name="Test_Table" dataNode="dn1,dn2,dn3,dn4" rule="hash-four"/> </schema> ``` * **steps:** step1. execute in dble ``` mysql> use mytest; Database changed mysql> select t1.id from dbtest.test_table t1 left join test t2 on t1.id=t2.id; ERROR 4004 (42S22): column t1.id not found ``` * **expect result:** 1. query success * **real result:** 1.query failed * **supplements:** 1. auto_test position:lower_case_table_names.feature --line66-68 /label ~BUG
test
error when join cross databases dble version: dble preconditions : lower case table names configs: schema xml steps: execute in dble mysql use mytest database changed mysql select id from dbtest test table left join test on id id error column id not found expect result: query success real result: query failed supplements: auto test position lower case table names feature label bug
1
161,959
13,880,467,411
IssuesEvent
2020-10-17 18:54:48
spotify/backstage
https://api.github.com/repos/spotify/backstage
closed
Deployments: Provide standardised ways of deploying Backstage
documentation enhancement
<!--- Provide a general summary of the feature request in the Title above --> ## Feature Suggestion There's scant documentation on how to get Backstage deployed today. This was an issue which was influenced by my submission of https://github.com/spotify/backstage/pull/2452. We can share more details of our setup (including the our Terraform definitions, or how we deployed it on GitLab) -- let us know what else you'd like us to do. <!--- If you're looking for help, please see https://backstage.io/ ---> <!--- Tell us how we could improve your experience --> ## Possible Implementation <!--- Not obligatory, but ideas as to the implementation of the addition or change --> We (@grab) have built a deployment pipeline which extends concepts in the development image above - using the development image, install the dependencies, build the distribution, and then use a second stage to only include the final artifact of the first stage. This maximises reuse of an openly-available image (adding exactly the Backstage bits) and builds on top of the Dockerfiles which is already in the Backstage repository. Given the frontend and backend images, there's a Docker-compose file similar for the development environment (but does not include a database -- simple copy pastes from the Development Docker-compose file would also work) which starts both backend and frontend containers. This is where users of Backstage can decide how they want to integrate with their cloud. For us, we went with this approach: - In development/test environments, we use a similar set up as development, i.e. an ephemeral Postgres container, deployed onto a Kubernetes cluster. See https://github.com/spotify/backstage/issues/2178 for more details - ephemeral environments are there to prove that a feature/plugin on a branch works and we can see a deployment even before it has been merged to master. - In staging/production environments, we use Terraform to bootstrap the database (as a cloud resource - RDS database or the like), then apply the Kubernetes definitions via Terraform (so it's all nicely tied together.) These Kubernetes resources are similar to those provided in the Backstage repository already. ## Context <!--- What are you trying to accomplish? --> <!--- Providing context (e.g. links to configuration settings, stack trace or log data) --> <!--- helps us come up with a solution that is most useful in the real world --> These documents touch on the deployment aspects of Backstage, each with a slightly different slant: - https://backstage.io/docs/getting-started/running-backstage-locally - https://backstage.io/docs/getting-started/development-environment - https://backstage.io/docs/getting-started/deployment-other Furthermore, we have Dockerfiles, Docker-compose files, and Kubernetes definitions in the repository. These should really be unified to minimise confusion.
1.0
Deployments: Provide standardised ways of deploying Backstage - <!--- Provide a general summary of the feature request in the Title above --> ## Feature Suggestion There's scant documentation on how to get Backstage deployed today. This was an issue which was influenced by my submission of https://github.com/spotify/backstage/pull/2452. We can share more details of our setup (including the our Terraform definitions, or how we deployed it on GitLab) -- let us know what else you'd like us to do. <!--- If you're looking for help, please see https://backstage.io/ ---> <!--- Tell us how we could improve your experience --> ## Possible Implementation <!--- Not obligatory, but ideas as to the implementation of the addition or change --> We (@grab) have built a deployment pipeline which extends concepts in the development image above - using the development image, install the dependencies, build the distribution, and then use a second stage to only include the final artifact of the first stage. This maximises reuse of an openly-available image (adding exactly the Backstage bits) and builds on top of the Dockerfiles which is already in the Backstage repository. Given the frontend and backend images, there's a Docker-compose file similar for the development environment (but does not include a database -- simple copy pastes from the Development Docker-compose file would also work) which starts both backend and frontend containers. This is where users of Backstage can decide how they want to integrate with their cloud. For us, we went with this approach: - In development/test environments, we use a similar set up as development, i.e. an ephemeral Postgres container, deployed onto a Kubernetes cluster. See https://github.com/spotify/backstage/issues/2178 for more details - ephemeral environments are there to prove that a feature/plugin on a branch works and we can see a deployment even before it has been merged to master. - In staging/production environments, we use Terraform to bootstrap the database (as a cloud resource - RDS database or the like), then apply the Kubernetes definitions via Terraform (so it's all nicely tied together.) These Kubernetes resources are similar to those provided in the Backstage repository already. ## Context <!--- What are you trying to accomplish? --> <!--- Providing context (e.g. links to configuration settings, stack trace or log data) --> <!--- helps us come up with a solution that is most useful in the real world --> These documents touch on the deployment aspects of Backstage, each with a slightly different slant: - https://backstage.io/docs/getting-started/running-backstage-locally - https://backstage.io/docs/getting-started/development-environment - https://backstage.io/docs/getting-started/deployment-other Furthermore, we have Dockerfiles, Docker-compose files, and Kubernetes definitions in the repository. These should really be unified to minimise confusion.
non_test
deployments provide standardised ways of deploying backstage feature suggestion there s scant documentation on how to get backstage deployed today this was an issue which was influenced by my submission of we can share more details of our setup including the our terraform definitions or how we deployed it on gitlab let us know what else you d like us to do possible implementation we grab have built a deployment pipeline which extends concepts in the development image above using the development image install the dependencies build the distribution and then use a second stage to only include the final artifact of the first stage this maximises reuse of an openly available image adding exactly the backstage bits and builds on top of the dockerfiles which is already in the backstage repository given the frontend and backend images there s a docker compose file similar for the development environment but does not include a database simple copy pastes from the development docker compose file would also work which starts both backend and frontend containers this is where users of backstage can decide how they want to integrate with their cloud for us we went with this approach in development test environments we use a similar set up as development i e an ephemeral postgres container deployed onto a kubernetes cluster see for more details ephemeral environments are there to prove that a feature plugin on a branch works and we can see a deployment even before it has been merged to master in staging production environments we use terraform to bootstrap the database as a cloud resource rds database or the like then apply the kubernetes definitions via terraform so it s all nicely tied together these kubernetes resources are similar to those provided in the backstage repository already context these documents touch on the deployment aspects of backstage each with a slightly different slant furthermore we have dockerfiles docker compose files and kubernetes definitions in the repository these should really be unified to minimise confusion
0
20,969
16,374,726,799
IssuesEvent
2021-05-15 21:34:17
godotengine/godot
https://api.github.com/repos/godotengine/godot
closed
Restoring minimized Project Manager makes it bigger
bug topic:editor usability
**Godot version:** 5f1107a **OS/device including version:** Windows 10 **Issue description:** Uh, I had this happen once and can't reproduce anymore because of #37327, but when you minimize Project Manager and restore it, the window is bigger and gets bigger with each repeat of this action. Might be related to DPI scaling.
True
Restoring minimized Project Manager makes it bigger - **Godot version:** 5f1107a **OS/device including version:** Windows 10 **Issue description:** Uh, I had this happen once and can't reproduce anymore because of #37327, but when you minimize Project Manager and restore it, the window is bigger and gets bigger with each repeat of this action. Might be related to DPI scaling.
non_test
restoring minimized project manager makes it bigger godot version os device including version windows issue description uh i had this happen once and can t reproduce anymore because of but when you minimize project manager and restore it the window is bigger and gets bigger with each repeat of this action might be related to dpi scaling
0
310,449
26,717,030,622
IssuesEvent
2023-01-28 16:58:59
dotnetcore/BootstrapBlazor
https://api.github.com/repos/dotnetcore/BootstrapBlazor
closed
test: add unit test for cascader
test
### Which class is this unit test associated with? add unit test for cascader
1.0
test: add unit test for cascader - ### Which class is this unit test associated with? add unit test for cascader
test
test add unit test for cascader which class is this unit test associated with add unit test for cascader
1
12,205
2,685,464,128
IssuesEvent
2015-03-30 01:13:33
IssueMigrationTest/Test5
https://api.github.com/repos/IssueMigrationTest/Test5
closed
type( object ) is not supported.
auto-migrated Priority-Medium Type-Defect
**Issue by sears....@gmail.com** _17 Jan 2011 at 10:44 GMT_ _Originally opened on Google Code_ ---- ``` In python, often developers will use the type() function to verify the type of objects. Example: def make_negative(interger): if type(interger) != type(0): raise Exception('"interger" is not an interger!') else: return (-1)*interger On another note: I'd really like to help out with this project, but I am only proficient in python. If there is anything I can do to help out, leave a note here and I'll try my best to help out. Great project you've got here! ```
1.0
type( object ) is not supported. - **Issue by sears....@gmail.com** _17 Jan 2011 at 10:44 GMT_ _Originally opened on Google Code_ ---- ``` In python, often developers will use the type() function to verify the type of objects. Example: def make_negative(interger): if type(interger) != type(0): raise Exception('"interger" is not an interger!') else: return (-1)*interger On another note: I'd really like to help out with this project, but I am only proficient in python. If there is anything I can do to help out, leave a note here and I'll try my best to help out. Great project you've got here! ```
non_test
type object is not supported issue by sears gmail com jan at gmt originally opened on google code in python often developers will use the type function to verify the type of objects example def make negative interger if type interger type raise exception interger is not an interger else return interger on another note i d really like to help out with this project but i am only proficient in python if there is anything i can do to help out leave a note here and i ll try my best to help out great project you ve got here
0
448,449
12,950,759,104
IssuesEvent
2020-07-19 14:23:44
GiftForGood/website
https://api.github.com/repos/GiftForGood/website
closed
Mark Wishes as Completed for NPO
c.UserStory m.MVP priority.High
# User Story <!-- https://github.com/GiftForGood/website/issues?q=is%3Aissue+label%3Ac.UserStory --> ## Describe the user story in detail. As a NPO, I want to mark a wish as completed so that I will not get any interest in future.
1.0
Mark Wishes as Completed for NPO - # User Story <!-- https://github.com/GiftForGood/website/issues?q=is%3Aissue+label%3Ac.UserStory --> ## Describe the user story in detail. As a NPO, I want to mark a wish as completed so that I will not get any interest in future.
non_test
mark wishes as completed for npo user story describe the user story in detail as a npo i want to mark a wish as completed so that i will not get any interest in future
0
777,983
27,300,044,438
IssuesEvent
2023-02-24 00:37:49
KelvinTegelaar/CIPP
https://api.github.com/repos/KelvinTegelaar/CIPP
closed
[Feature Request]: Monitor for overprovisioned O365 licenses.
enhancement no-priority
### Description of the new feature - must be an in-depth explanation of the feature you want, reasoning why, and the added benefits for MSPs as a whole. Due to the new Microsoft NCE license model, we need a better way to monitor for overprovisioned licenses. It would be helpful to be alerted when the number of O365 consumed licenses is higher than provisioned licenses. ### PowerShell commands you would normally use to achieve above request #Connect to Microsoft Graph. Connect-Graph -Scopes Organization.Read.All #Read all assigned licenses for the SKU part that contain O365 in the part name. $Licenses = Get-MgSubscribedSku | where{$_.SkuPartNumber -like "*O365*"} | select {$_.PrepaidUnits.enabled}, ConsumedUnits, skupartnumber #Identify the total number of consumed and prepaid licenses. [int]$Consumed = $Licenses.ConsumedUnits [int]$Prepaid = $Licenses.{$_.PrepaidUnits.enabled} #Compare the number of consumed and prepaid licenses to identify overprovisioning. If($Consumed -gt $Prepaid){Write-Host -ForegroundColor Red "Licenses have been overprovisioned. Total Count $Consumed/$Prepaid"}
1.0
[Feature Request]: Monitor for overprovisioned O365 licenses. - ### Description of the new feature - must be an in-depth explanation of the feature you want, reasoning why, and the added benefits for MSPs as a whole. Due to the new Microsoft NCE license model, we need a better way to monitor for overprovisioned licenses. It would be helpful to be alerted when the number of O365 consumed licenses is higher than provisioned licenses. ### PowerShell commands you would normally use to achieve above request #Connect to Microsoft Graph. Connect-Graph -Scopes Organization.Read.All #Read all assigned licenses for the SKU part that contain O365 in the part name. $Licenses = Get-MgSubscribedSku | where{$_.SkuPartNumber -like "*O365*"} | select {$_.PrepaidUnits.enabled}, ConsumedUnits, skupartnumber #Identify the total number of consumed and prepaid licenses. [int]$Consumed = $Licenses.ConsumedUnits [int]$Prepaid = $Licenses.{$_.PrepaidUnits.enabled} #Compare the number of consumed and prepaid licenses to identify overprovisioning. If($Consumed -gt $Prepaid){Write-Host -ForegroundColor Red "Licenses have been overprovisioned. Total Count $Consumed/$Prepaid"}
non_test
monitor for overprovisioned licenses description of the new feature must be an in depth explanation of the feature you want reasoning why and the added benefits for msps as a whole due to the new microsoft nce license model we need a better way to monitor for overprovisioned licenses it would be helpful to be alerted when the number of consumed licenses is higher than provisioned licenses powershell commands you would normally use to achieve above request connect to microsoft graph connect graph scopes organization read all read all assigned licenses for the sku part that contain in the part name licenses get mgsubscribedsku where skupartnumber like select prepaidunits enabled consumedunits skupartnumber identify the total number of consumed and prepaid licenses consumed licenses consumedunits prepaid licenses prepaidunits enabled compare the number of consumed and prepaid licenses to identify overprovisioning if consumed gt prepaid write host foregroundcolor red licenses have been overprovisioned total count consumed prepaid
0
349,919
31,841,144,551
IssuesEvent
2023-09-14 16:23:53
submariner-io/subctl
https://api.github.com/repos/submariner-io/subctl
opened
subctl UTs on secret handling sometimes fail with mismatched data
bug testing
See https://github.com/submariner-io/subctl/actions/runs/6187928997/job/16798837452?pr=923 for example — there’s a race between the informer and the retrieval of the secret (from `EnsureTokenSecret`), so the secret being compared doesn’t necessarily have its `Data` field filled in, but might by the time it’s retrieved again for comparison.
1.0
subctl UTs on secret handling sometimes fail with mismatched data - See https://github.com/submariner-io/subctl/actions/runs/6187928997/job/16798837452?pr=923 for example — there’s a race between the informer and the retrieval of the secret (from `EnsureTokenSecret`), so the secret being compared doesn’t necessarily have its `Data` field filled in, but might by the time it’s retrieved again for comparison.
test
subctl uts on secret handling sometimes fail with mismatched data see for example — there’s a race between the informer and the retrieval of the secret from ensuretokensecret so the secret being compared doesn’t necessarily have its data field filled in but might by the time it’s retrieved again for comparison
1
334,261
29,829,412,349
IssuesEvent
2023-06-18 04:20:50
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
reopened
Fix elementwise.test_expm1
Sub Task Failing Test
| | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5048230116/jobs/9056152030" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/5048230116/jobs/9056152030" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5048230116/jobs/9056152030" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/5048230116/jobs/9056152030" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |paddle|<a href="null" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
1.0
Fix elementwise.test_expm1 - | | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5048230116/jobs/9056152030" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/5048230116/jobs/9056152030" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5048230116/jobs/9056152030" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/5048230116/jobs/9056152030" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |paddle|<a href="null" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
test
fix elementwise test tensorflow img src torch img src numpy img src jax img src paddle img src
1
321,823
27,558,446,635
IssuesEvent
2023-03-07 19:52:47
bcgov/itvr
https://api.github.com/repos/bcgov/itvr
closed
ITVR - Update Cancel application email
High Story Tested :)
**Title: Update the wording in the cancel application email** **Description:** the original wording for the cancel application email was based on household applications being cancelled by the spouse or automatically by the system. The household application has been turned off so this email format is now incorrect. There is now a cancel application feature in the console that the analyst can use, this might be done at the request of the applicant. As such we need a new email format for these cancelled applications. If household applications are ever enabled again we may need to distinguish these 2 different cancel scenarios. **Wireframe:** na **Purpose and benefit to user:** the current email is confusing and incorrect, the new email wording will make more sense to the applicant **Acceptance Criteria:** - [ ] Given I am an individual applicant, When my application is cancelled, then i get an email with the appropriate wording - [ ] Given I am a household applicant, When my application is cancelled, then I get an email with different wording **Development Checklist:** - [ ] Write the two cancellation emails - [ ] comment out the household email because currently household applications are not allowed - [ ] make sure that the individual's email sends when an idir user cancels the application **Notes:** new email wording Your application has been cancelled. If you haven’t yet received a rebate you are encouraged to apply again. If you think this was done in error or you have questions, please feel free to contact us at ZEVPrograms@gov.bc.ca
1.0
ITVR - Update Cancel application email - **Title: Update the wording in the cancel application email** **Description:** the original wording for the cancel application email was based on household applications being cancelled by the spouse or automatically by the system. The household application has been turned off so this email format is now incorrect. There is now a cancel application feature in the console that the analyst can use, this might be done at the request of the applicant. As such we need a new email format for these cancelled applications. If household applications are ever enabled again we may need to distinguish these 2 different cancel scenarios. **Wireframe:** na **Purpose and benefit to user:** the current email is confusing and incorrect, the new email wording will make more sense to the applicant **Acceptance Criteria:** - [ ] Given I am an individual applicant, When my application is cancelled, then i get an email with the appropriate wording - [ ] Given I am a household applicant, When my application is cancelled, then I get an email with different wording **Development Checklist:** - [ ] Write the two cancellation emails - [ ] comment out the household email because currently household applications are not allowed - [ ] make sure that the individual's email sends when an idir user cancels the application **Notes:** new email wording Your application has been cancelled. If you haven’t yet received a rebate you are encouraged to apply again. If you think this was done in error or you have questions, please feel free to contact us at ZEVPrograms@gov.bc.ca
test
itvr update cancel application email title update the wording in the cancel application email description the original wording for the cancel application email was based on household applications being cancelled by the spouse or automatically by the system the household application has been turned off so this email format is now incorrect there is now a cancel application feature in the console that the analyst can use this might be done at the request of the applicant as such we need a new email format for these cancelled applications if household applications are ever enabled again we may need to distinguish these different cancel scenarios wireframe na purpose and benefit to user the current email is confusing and incorrect the new email wording will make more sense to the applicant acceptance criteria given i am an individual applicant when my application is cancelled then i get an email with the appropriate wording given i am a household applicant when my application is cancelled then i get an email with different wording development checklist write the two cancellation emails comment out the household email because currently household applications are not allowed make sure that the individual s email sends when an idir user cancels the application notes new email wording your application has been cancelled if you haven’t yet received a rebate you are encouraged to apply again if you think this was done in error or you have questions please feel free to contact us at zevprograms gov bc ca
1
103,318
4,166,971,033
IssuesEvent
2016-06-20 07:30:31
TheScienceMuseum/collectionsonline
https://api.github.com/repos/TheScienceMuseum/collectionsonline
closed
Create simple API search route
in-progress priority-1 technical
* Connect to elastic search * Submit query * Retrieve results
1.0
Create simple API search route - * Connect to elastic search * Submit query * Retrieve results
non_test
create simple api search route connect to elastic search submit query retrieve results
0
124,865
10,327,013,240
IssuesEvent
2019-09-02 05:11:27
vgstation-coders/vgstation13
https://api.github.com/repos/vgstation-coders/vgstation13
closed
Fax timer
Bug / Fix Needs Moar Testing T-Thanks BYOND
(WEB REPORT BY: jeanvoerman REMOTE: 172.93.109.202:7777) # Revision b01534b42527c3ad3815baf98946d12a7c2e7fd6 # Description Transmitter arrays realigning. Please stand by for 84960.8 seconds. # Steps to Reproduce Try sending a fax I think? # What you Expected Sending a second fax # What Actually Happened 24 hours cooldown
1.0
Fax timer - (WEB REPORT BY: jeanvoerman REMOTE: 172.93.109.202:7777) # Revision b01534b42527c3ad3815baf98946d12a7c2e7fd6 # Description Transmitter arrays realigning. Please stand by for 84960.8 seconds. # Steps to Reproduce Try sending a fax I think? # What you Expected Sending a second fax # What Actually Happened 24 hours cooldown
test
fax timer web report by jeanvoerman remote revision description transmitter arrays realigning please stand by for seconds steps to reproduce try sending a fax i think what you expected sending a second fax what actually happened hours cooldown
1
233,236
18,956,111,945
IssuesEvent
2021-11-18 20:28:23
dotnet/msbuild
https://api.github.com/repos/dotnet/msbuild
opened
Add Regression Test for NugetSDKResolver Throwing
For consideration Area: Resolvers Area: Tests
Origin story: https://github.com/dotnet/msbuild/issues/7035 ### Context When we took the fix to throw when sdkresolvers throw, it revealed the nugetsdkresolver was throwing often. That _should have_ been caught by a regression test, so we should add one.
1.0
Add Regression Test for NugetSDKResolver Throwing - Origin story: https://github.com/dotnet/msbuild/issues/7035 ### Context When we took the fix to throw when sdkresolvers throw, it revealed the nugetsdkresolver was throwing often. That _should have_ been caught by a regression test, so we should add one.
test
add regression test for nugetsdkresolver throwing origin story context when we took the fix to throw when sdkresolvers throw it revealed the nugetsdkresolver was throwing often that should have been caught by a regression test so we should add one
1
140,183
18,895,243,106
IssuesEvent
2021-11-15 17:08:55
bgoonz/searchAwesome
https://api.github.com/repos/bgoonz/searchAwesome
closed
CVE-2018-19837 (Medium) detected in node-sass-4.11.0.tgz
security vulnerability
## CVE-2018-19837 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-sass-4.11.0.tgz</b></p></summary> <p>Wrapper around libsass</p> <p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.11.0.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.11.0.tgz</a></p> <p>Path to dependency file: searchAwesome/clones/awesome-stacks/package.json</p> <p>Path to vulnerable library: /clones/awesome-stacks/node_modules/node-sass/package.json</p> <p> Dependency Hierarchy: - :x: **node-sass-4.11.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/bgoonz/searchAwesome/commit/cb1b8421c464b43b24d4816929e575612a00cd49">cb1b8421c464b43b24d4816929e575612a00cd49</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In LibSass prior to 3.5.5, Sass::Eval::operator()(Sass::Binary_Expression*) inside eval.cpp allows attackers to cause a denial-of-service resulting from stack consumption via a crafted sass file, because of certain incorrect parsing of '%' as a modulo operator in parser.cpp. <p>Publish Date: 2018-12-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-19837>CVE-2018-19837</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/sass/libsass/releases/tag/3.5.5">https://github.com/sass/libsass/releases/tag/3.5.5</a></p> <p>Release Date: 2018-12-04</p> <p>Fix Resolution: libsass - 3.5.5;node-sass - 4.14.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-19837 (Medium) detected in node-sass-4.11.0.tgz - ## CVE-2018-19837 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-sass-4.11.0.tgz</b></p></summary> <p>Wrapper around libsass</p> <p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.11.0.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.11.0.tgz</a></p> <p>Path to dependency file: searchAwesome/clones/awesome-stacks/package.json</p> <p>Path to vulnerable library: /clones/awesome-stacks/node_modules/node-sass/package.json</p> <p> Dependency Hierarchy: - :x: **node-sass-4.11.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/bgoonz/searchAwesome/commit/cb1b8421c464b43b24d4816929e575612a00cd49">cb1b8421c464b43b24d4816929e575612a00cd49</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In LibSass prior to 3.5.5, Sass::Eval::operator()(Sass::Binary_Expression*) inside eval.cpp allows attackers to cause a denial-of-service resulting from stack consumption via a crafted sass file, because of certain incorrect parsing of '%' as a modulo operator in parser.cpp. <p>Publish Date: 2018-12-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-19837>CVE-2018-19837</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/sass/libsass/releases/tag/3.5.5">https://github.com/sass/libsass/releases/tag/3.5.5</a></p> <p>Release Date: 2018-12-04</p> <p>Fix Resolution: libsass - 3.5.5;node-sass - 4.14.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve medium detected in node sass tgz cve medium severity vulnerability vulnerable library node sass tgz wrapper around libsass library home page a href path to dependency file searchawesome clones awesome stacks package json path to vulnerable library clones awesome stacks node modules node sass package json dependency hierarchy x node sass tgz vulnerable library found in head commit a href found in base branch master vulnerability details in libsass prior to sass eval operator sass binary expression inside eval cpp allows attackers to cause a denial of service resulting from stack consumption via a crafted sass file because of certain incorrect parsing of as a modulo operator in parser cpp publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution libsass node sass step up your open source security game with whitesource
0
97,055
8,641,843,847
IssuesEvent
2018-11-24 22:19:47
jpbarraca/pai
https://api.github.com/repos/jpbarraca/pai
closed
Error when using arming in stay mode
bug mg testing required
When I arm the system in stay mode, I get the following error in the console: Currently I am still testing and familiarizing myself with everything. So it might very possibly be something I am doing wrong. It seem to still work, as the alarm is armed in stay mode. I would publish either arm_stay or STAY_ARM (depending if I have MQTT_HOMEBRIDGE_ENABLE as True or False). >topic=paradox/control/partitions/Area_1, payload=arm_stay >topic=paradox/control/partitions/Area_1, payload=STAY_ARM >topic=paradox/control/partitions/all, payload=arm_stay >topic=paradox/control/partitions/all, payload=STAY_ARM I experience the same when using the existing remote for the alarm. **Details**: Panel MG5050 version 5.33 build 2 IP150 Module (downgraded to <4 firmware) **Current user.conf**: >LOGGING_LEVEL_CONSOLE = logging.DEBUG >CONNECTION_TYPE = 'IP' >IP_CONNECTION_HOST = '1.1.1.1' >IP_CONNECTION_PASSWORD = b'my_password' >MQTT_ENABLE = True >MQTT_HOST = '1.1.1.2' >MQTT_USERNAME = None >MQTT_PASSWORD = None >PASSWORD = None >MQTT_HOMEBRIDGE_ENABLE = True _**(or False)**_ Test, disam the system, wait a few seconds and arm it again. Below is the console output with DEBUG switched on. **Console output**: REDACTED
1.0
Error when using arming in stay mode - When I arm the system in stay mode, I get the following error in the console: Currently I am still testing and familiarizing myself with everything. So it might very possibly be something I am doing wrong. It seem to still work, as the alarm is armed in stay mode. I would publish either arm_stay or STAY_ARM (depending if I have MQTT_HOMEBRIDGE_ENABLE as True or False). >topic=paradox/control/partitions/Area_1, payload=arm_stay >topic=paradox/control/partitions/Area_1, payload=STAY_ARM >topic=paradox/control/partitions/all, payload=arm_stay >topic=paradox/control/partitions/all, payload=STAY_ARM I experience the same when using the existing remote for the alarm. **Details**: Panel MG5050 version 5.33 build 2 IP150 Module (downgraded to <4 firmware) **Current user.conf**: >LOGGING_LEVEL_CONSOLE = logging.DEBUG >CONNECTION_TYPE = 'IP' >IP_CONNECTION_HOST = '1.1.1.1' >IP_CONNECTION_PASSWORD = b'my_password' >MQTT_ENABLE = True >MQTT_HOST = '1.1.1.2' >MQTT_USERNAME = None >MQTT_PASSWORD = None >PASSWORD = None >MQTT_HOMEBRIDGE_ENABLE = True _**(or False)**_ Test, disam the system, wait a few seconds and arm it again. Below is the console output with DEBUG switched on. **Console output**: REDACTED
test
error when using arming in stay mode when i arm the system in stay mode i get the following error in the console currently i am still testing and familiarizing myself with everything so it might very possibly be something i am doing wrong it seem to still work as the alarm is armed in stay mode i would publish either arm stay or stay arm depending if i have mqtt homebridge enable as true or false topic paradox control partitions area payload arm stay topic paradox control partitions area payload stay arm topic paradox control partitions all payload arm stay topic paradox control partitions all payload stay arm i experience the same when using the existing remote for the alarm details panel version build module downgraded to firmware current user conf logging level console logging debug connection type ip ip connection host ip connection password b my password mqtt enable true mqtt host mqtt username none mqtt password none password none mqtt homebridge enable true or false test disam the system wait a few seconds and arm it again below is the console output with debug switched on console output redacted
1
81,970
31,837,100,183
IssuesEvent
2023-09-14 14:08:05
SeleniumHQ/selenium
https://api.github.com/repos/SeleniumHQ/selenium
closed
[🐛 Bug]: Python selenium msedge does not work with headless mode disabled. Message: unknown error: msedge failed to start: crashed
R-awaiting answer I-defect needs-triaging
### What happened? I have a problem with a python selenium web scraper that uses microsoft edge as a webdriver. The msedgedriver works fine only when one of these is present: edge_options.add_argument("--user-data-dir=C:\\Users\\xxx\\Desktop\\Edge_user_data") edge_options.add_argument('--headless') In case i want to use both, the program crashes, i have tried: -giving msedge permissions -reseting msedge browser -duplicating the profile folder -many other solutions from stackoverflow of git, but nothing has worked ### How can we reproduce the issue? ```shell # Define Edge options edge_options = webdriver.EdgeOptions() edgepath = "C:\\Program Files (x86)\\Microsoft\\Edge\\Application\\msedge.exe" edge_options.setBinary = edgepath service = webdriver.EdgeService(service_args=['--log-level=SEVERE', '--disable-build-check']) # Using the user profile edge_options.add_argument("--user-data-dir=C:\\Users\\xxx\\Desktop\\Edge_user_data") # Options logger = logging.getLogger() logger.setLevel(logging.CRITICAL) #logging.basicConfig(level=logging.WARNING) edge_options.add_argument('--remote-debugging-port=0') edge_options.add_argument('--no-first-run') edge_options.add_argument('--no-default-browser-check') edge_options.add_argument('--headless') edge_options.add_argument('--log-level=3') edge_options.add_argument("--disable-logging") edge_options.add_argument('--start-maximized') edge_options.add_argument('--disable-infobars') edge_options.add_experimental_option('excludeSwitches', ['disable-popup-blocking']) driver = webdriver.Edge(options=edge_options, service=service) return driver ``` ### Relevant log output ```shell Traceback (most recent call last): File "C:\Users\xxx\Desktop\opus-test17.py", line 3414, in <module> slower_process(urls, result_file_path) File "C:\Users\xxx\Desktop\opus-test17.py", line 3338, in slower_process result = check_url(url) ^^^^^^^^^^^^^^ File "C:\Users\xxx\Desktop\opus-test17.py", line 42, in check_url result = 123_check(url, username) ^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Users\xxx\Desktop\opus-test17.py", line 534, in 123_check driver = initialize_driver5() ^^^^^^^^^^^^^^^^^^^^ File "C:\Users\xxx\Desktop\opus-test17.py", line 321, in initialize_driver5 driver = webdriver.Edge(options=edge_options, service=service) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Users\xxx\AppData\Local\Programs\Python\Python311\Lib\site-packages\selenium\webdriver\edge\webdriver.py", line 45, in __init__ super().__init__( File "C:\Users\xxx\AppData\Local\Programs\Python\Python311\Lib\site-packages\selenium\webdriver\chromium\webdriver.py", line 56, in __init__ super().__init__( File "C:\Users\xxx\AppData\Local\Programs\Python\Python311\Lib\site-packages\selenium\webdriver\remote\webdriver.py", line 206, in __init__ self.start_session(capabilities) File "C:\Users\xxx\AppData\Local\Programs\Python\Python311\Lib\site-packages\selenium\webdriver\remote\webdriver.py", line 290, in start_session response = self.execute(Command.NEW_SESSION, caps)["value"] ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Users\xxx\AppData\Local\Programs\Python\Python311\Lib\site-packages\selenium\webdriver\remote\webdriver.py", line 345, in execute self.error_handler.check_response(response) File "C:\Users\xxx\AppData\Local\Programs\Python\Python311\Lib\site-packages\selenium\webdriver\remote\errorhandler.py", line 229, in check_response raise exception_class(message, screen, stacktrace) selenium.common.exceptions.WebDriverException: Message: unknown error: msedge failed to start: crashed. (unknown error: DevToolsActivePort file doesn't exist) (The process started from msedge location C:\Program Files (x86)\Microsoft\Edge\Application\msedge.exe is no longer running, so msedgedriver is assuming that msedge has crashed.) Stacktrace: GetHandleVerifier [0x00007FF74D583DB2+61490] Microsoft::Applications::Events::EventProperty::~EventProperty [0x00007FF74D516002+740642] (No symbol) [0x00007FF74D2EB8AE] (No symbol) [0x00007FF74D31CBFE] (No symbol) [0x00007FF74D317F13] (No symbol) [0x00007FF74D35B7F8] (No symbol) [0x00007FF74D3536E3] (No symbol) [0x00007FF74D325EAA] (No symbol) [0x00007FF74D32518B] (No symbol) [0x00007FF74D326634] Microsoft::Applications::Events::ILogManager::DispatchEventBroadcast [0x00007FF74D748D69+1207369] (No symbol) [0x00007FF74D3A5304] Microsoft::Applications::Events::EventProperty::~EventProperty [0x00007FF74D4690F1+32273] Microsoft::Applications::Events::EventProperty::~EventProperty [0x00007FF74D4619E9+1801] Microsoft::Applications::Events::ILogManager::DispatchEventBroadcast [0x00007FF74D747944+1202212] Microsoft::Applications::Events::ILogConfiguration::operator* [0x00007FF74D51E998+19784] Microsoft::Applications::Events::ILogConfiguration::operator* [0x00007FF74D51AE54+4612] Microsoft::Applications::Events::ILogConfiguration::operator* [0x00007FF74D51AF86+4918] Microsoft::Applications::Events::EventProperty::~EventProperty [0x00007FF74D50F451+713073] BaseThreadInitThunk [0x00007FF87E917614+20] RtlUserThreadStart [0x00007FF87FD226B1+33] ``` ### Operating System Windows 10 Pro ### Selenium version Python 3.11.4 ### What are the browser(s) and version(s) where you see this issue? Microsoft Edge Version 116.0.1938.81 (Official build) (64-bit) ### What are the browser driver(s) and version(s) where you see this issue? msedgedriver Version: 116.0.1938.81 ### Are you using Selenium Grid? _No response_
1.0
[🐛 Bug]: Python selenium msedge does not work with headless mode disabled. Message: unknown error: msedge failed to start: crashed - ### What happened? I have a problem with a python selenium web scraper that uses microsoft edge as a webdriver. The msedgedriver works fine only when one of these is present: edge_options.add_argument("--user-data-dir=C:\\Users\\xxx\\Desktop\\Edge_user_data") edge_options.add_argument('--headless') In case i want to use both, the program crashes, i have tried: -giving msedge permissions -reseting msedge browser -duplicating the profile folder -many other solutions from stackoverflow of git, but nothing has worked ### How can we reproduce the issue? ```shell # Define Edge options edge_options = webdriver.EdgeOptions() edgepath = "C:\\Program Files (x86)\\Microsoft\\Edge\\Application\\msedge.exe" edge_options.setBinary = edgepath service = webdriver.EdgeService(service_args=['--log-level=SEVERE', '--disable-build-check']) # Using the user profile edge_options.add_argument("--user-data-dir=C:\\Users\\xxx\\Desktop\\Edge_user_data") # Options logger = logging.getLogger() logger.setLevel(logging.CRITICAL) #logging.basicConfig(level=logging.WARNING) edge_options.add_argument('--remote-debugging-port=0') edge_options.add_argument('--no-first-run') edge_options.add_argument('--no-default-browser-check') edge_options.add_argument('--headless') edge_options.add_argument('--log-level=3') edge_options.add_argument("--disable-logging") edge_options.add_argument('--start-maximized') edge_options.add_argument('--disable-infobars') edge_options.add_experimental_option('excludeSwitches', ['disable-popup-blocking']) driver = webdriver.Edge(options=edge_options, service=service) return driver ``` ### Relevant log output ```shell Traceback (most recent call last): File "C:\Users\xxx\Desktop\opus-test17.py", line 3414, in <module> slower_process(urls, result_file_path) File "C:\Users\xxx\Desktop\opus-test17.py", line 3338, in slower_process result = check_url(url) ^^^^^^^^^^^^^^ File "C:\Users\xxx\Desktop\opus-test17.py", line 42, in check_url result = 123_check(url, username) ^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Users\xxx\Desktop\opus-test17.py", line 534, in 123_check driver = initialize_driver5() ^^^^^^^^^^^^^^^^^^^^ File "C:\Users\xxx\Desktop\opus-test17.py", line 321, in initialize_driver5 driver = webdriver.Edge(options=edge_options, service=service) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Users\xxx\AppData\Local\Programs\Python\Python311\Lib\site-packages\selenium\webdriver\edge\webdriver.py", line 45, in __init__ super().__init__( File "C:\Users\xxx\AppData\Local\Programs\Python\Python311\Lib\site-packages\selenium\webdriver\chromium\webdriver.py", line 56, in __init__ super().__init__( File "C:\Users\xxx\AppData\Local\Programs\Python\Python311\Lib\site-packages\selenium\webdriver\remote\webdriver.py", line 206, in __init__ self.start_session(capabilities) File "C:\Users\xxx\AppData\Local\Programs\Python\Python311\Lib\site-packages\selenium\webdriver\remote\webdriver.py", line 290, in start_session response = self.execute(Command.NEW_SESSION, caps)["value"] ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Users\xxx\AppData\Local\Programs\Python\Python311\Lib\site-packages\selenium\webdriver\remote\webdriver.py", line 345, in execute self.error_handler.check_response(response) File "C:\Users\xxx\AppData\Local\Programs\Python\Python311\Lib\site-packages\selenium\webdriver\remote\errorhandler.py", line 229, in check_response raise exception_class(message, screen, stacktrace) selenium.common.exceptions.WebDriverException: Message: unknown error: msedge failed to start: crashed. (unknown error: DevToolsActivePort file doesn't exist) (The process started from msedge location C:\Program Files (x86)\Microsoft\Edge\Application\msedge.exe is no longer running, so msedgedriver is assuming that msedge has crashed.) Stacktrace: GetHandleVerifier [0x00007FF74D583DB2+61490] Microsoft::Applications::Events::EventProperty::~EventProperty [0x00007FF74D516002+740642] (No symbol) [0x00007FF74D2EB8AE] (No symbol) [0x00007FF74D31CBFE] (No symbol) [0x00007FF74D317F13] (No symbol) [0x00007FF74D35B7F8] (No symbol) [0x00007FF74D3536E3] (No symbol) [0x00007FF74D325EAA] (No symbol) [0x00007FF74D32518B] (No symbol) [0x00007FF74D326634] Microsoft::Applications::Events::ILogManager::DispatchEventBroadcast [0x00007FF74D748D69+1207369] (No symbol) [0x00007FF74D3A5304] Microsoft::Applications::Events::EventProperty::~EventProperty [0x00007FF74D4690F1+32273] Microsoft::Applications::Events::EventProperty::~EventProperty [0x00007FF74D4619E9+1801] Microsoft::Applications::Events::ILogManager::DispatchEventBroadcast [0x00007FF74D747944+1202212] Microsoft::Applications::Events::ILogConfiguration::operator* [0x00007FF74D51E998+19784] Microsoft::Applications::Events::ILogConfiguration::operator* [0x00007FF74D51AE54+4612] Microsoft::Applications::Events::ILogConfiguration::operator* [0x00007FF74D51AF86+4918] Microsoft::Applications::Events::EventProperty::~EventProperty [0x00007FF74D50F451+713073] BaseThreadInitThunk [0x00007FF87E917614+20] RtlUserThreadStart [0x00007FF87FD226B1+33] ``` ### Operating System Windows 10 Pro ### Selenium version Python 3.11.4 ### What are the browser(s) and version(s) where you see this issue? Microsoft Edge Version 116.0.1938.81 (Official build) (64-bit) ### What are the browser driver(s) and version(s) where you see this issue? msedgedriver Version: 116.0.1938.81 ### Are you using Selenium Grid? _No response_
non_test
python selenium msedge does not work with headless mode disabled message unknown error msedge failed to start crashed what happened i have a problem with a python selenium web scraper that uses microsoft edge as a webdriver the msedgedriver works fine only when one of these is present edge options add argument user data dir c users xxx desktop edge user data edge options add argument headless in case i want to use both the program crashes i have tried giving msedge permissions reseting msedge browser duplicating the profile folder many other solutions from stackoverflow of git but nothing has worked how can we reproduce the issue shell define edge options edge options webdriver edgeoptions edgepath c program files microsoft edge application msedge exe edge options setbinary edgepath service webdriver edgeservice service args using the user profile edge options add argument user data dir c users xxx desktop edge user data options logger logging getlogger logger setlevel logging critical logging basicconfig level logging warning edge options add argument remote debugging port edge options add argument no first run edge options add argument no default browser check edge options add argument headless edge options add argument log level edge options add argument disable logging edge options add argument start maximized edge options add argument disable infobars edge options add experimental option excludeswitches driver webdriver edge options edge options service service return driver relevant log output shell traceback most recent call last file c users xxx desktop opus py line in slower process urls result file path file c users xxx desktop opus py line in slower process result check url url file c users xxx desktop opus py line in check url result check url username file c users xxx desktop opus py line in check driver initialize file c users xxx desktop opus py line in initialize driver webdriver edge options edge options service service file c users xxx appdata local programs python lib site packages selenium webdriver edge webdriver py line in init super init file c users xxx appdata local programs python lib site packages selenium webdriver chromium webdriver py line in init super init file c users xxx appdata local programs python lib site packages selenium webdriver remote webdriver py line in init self start session capabilities file c users xxx appdata local programs python lib site packages selenium webdriver remote webdriver py line in start session response self execute command new session caps file c users xxx appdata local programs python lib site packages selenium webdriver remote webdriver py line in execute self error handler check response response file c users xxx appdata local programs python lib site packages selenium webdriver remote errorhandler py line in check response raise exception class message screen stacktrace selenium common exceptions webdriverexception message unknown error msedge failed to start crashed unknown error devtoolsactiveport file doesn t exist the process started from msedge location c program files microsoft edge application msedge exe is no longer running so msedgedriver is assuming that msedge has crashed stacktrace gethandleverifier microsoft applications events eventproperty eventproperty no symbol no symbol no symbol no symbol no symbol no symbol no symbol no symbol microsoft applications events ilogmanager dispatcheventbroadcast no symbol microsoft applications events eventproperty eventproperty microsoft applications events eventproperty eventproperty microsoft applications events ilogmanager dispatcheventbroadcast microsoft applications events ilogconfiguration operator microsoft applications events ilogconfiguration operator microsoft applications events ilogconfiguration operator microsoft applications events eventproperty eventproperty basethreadinitthunk rtluserthreadstart operating system windows pro selenium version python what are the browser s and version s where you see this issue microsoft edge version official build bit what are the browser driver s and version s where you see this issue msedgedriver version are you using selenium grid no response
0
339,555
30,455,656,220
IssuesEvent
2023-07-16 21:18:52
n4af/TR4W
https://api.github.com/repos/n4af/TR4W
closed
Add option to control offering TCP/IP socket for rig control for WSJT-X
enhancement WSJT-X In testing
Normally when using TR4W with WSJT-X, WSJT-X is set to use DXLab Commander as the radio. This in turn connects to a socket that TR4W offers to emulate the radio control protocol in Commander. IN certain cases such as the K4 in network mode, it is fine to let WSJT-X talk to the radio via TCP/IP such that the socket does not need to be offered. By not offering the socket, this allows the actual DXLab Commander to be running which is useful for those of us that use DXLabs. As this will not be the norm, create an option to a toggle offering this socket. The toggle should be true by default so we offer the socket when WSJT-X is active.
1.0
Add option to control offering TCP/IP socket for rig control for WSJT-X - Normally when using TR4W with WSJT-X, WSJT-X is set to use DXLab Commander as the radio. This in turn connects to a socket that TR4W offers to emulate the radio control protocol in Commander. IN certain cases such as the K4 in network mode, it is fine to let WSJT-X talk to the radio via TCP/IP such that the socket does not need to be offered. By not offering the socket, this allows the actual DXLab Commander to be running which is useful for those of us that use DXLabs. As this will not be the norm, create an option to a toggle offering this socket. The toggle should be true by default so we offer the socket when WSJT-X is active.
test
add option to control offering tcp ip socket for rig control for wsjt x normally when using with wsjt x wsjt x is set to use dxlab commander as the radio this in turn connects to a socket that offers to emulate the radio control protocol in commander in certain cases such as the in network mode it is fine to let wsjt x talk to the radio via tcp ip such that the socket does not need to be offered by not offering the socket this allows the actual dxlab commander to be running which is useful for those of us that use dxlabs as this will not be the norm create an option to a toggle offering this socket the toggle should be true by default so we offer the socket when wsjt x is active
1
114,648
11,852,930,206
IssuesEvent
2020-03-24 20:54:54
cyberark/conjur-oss-suite-release
https://api.github.com/repos/cyberark/conjur-oss-suite-release
opened
Changes should not be included for non-changed version between last release and current `suite.yml`
component/suite kind/documentation kind/enhancement team/community-and-integrations
Currently: - If our last-released version of repo `A` in `suite_<semver>.yml` is `x.y.z` - And the version of repo `A` in `suite.yml` is `x.y.z` We currently include all changes for repo `A` v`x.y.z` in the generated output. While we should include the version in the generated outputs as an artifacts, those changelog entries can be however ignored as nothing effectively changed. AC: - [ ] Unchanged repo versions between suite releases do not result in outputted changelog entries for that repo version.
1.0
Changes should not be included for non-changed version between last release and current `suite.yml` - Currently: - If our last-released version of repo `A` in `suite_<semver>.yml` is `x.y.z` - And the version of repo `A` in `suite.yml` is `x.y.z` We currently include all changes for repo `A` v`x.y.z` in the generated output. While we should include the version in the generated outputs as an artifacts, those changelog entries can be however ignored as nothing effectively changed. AC: - [ ] Unchanged repo versions between suite releases do not result in outputted changelog entries for that repo version.
non_test
changes should not be included for non changed version between last release and current suite yml currently if our last released version of repo a in suite yml is x y z and the version of repo a in suite yml is x y z we currently include all changes for repo a v x y z in the generated output while we should include the version in the generated outputs as an artifacts those changelog entries can be however ignored as nothing effectively changed ac unchanged repo versions between suite releases do not result in outputted changelog entries for that repo version
0
201,515
15,211,077,872
IssuesEvent
2021-02-17 08:33:23
apple/swift-nio
https://api.github.com/repos/apple/swift-nio
closed
Several SALChannelTest cases fail consistently on linux x86_64
flaky-test
### Expected behavior All tests to pass on Fedora Core 33, as they did on Ubuntu 20.04 over the last several months. ### Actual behavior Only the SALChannel tests fail, mostly with this error message `unexpected syscall whenReady(NIO.SelectorStrategy.now)`, which is some kind of race? I don't know if this is specific to Fedora, as I've since deleted my Ubuntu VPS. I see these errors occasionally on Android AArch64 too, which is why I then double-checked on linux, but they only randomly fail on Android when running the tests in `--parallel`, never serially. That flag makes no difference on Fedora, these tests always fail. ### Steps to reproduce ``` ../swift-5.3.3-RELEASE-centos8/usr/bin/swift test -j 2 --enable-test-discovery --filter SALC Test Suite 'Selected tests' started at 2021-02-13 06:59:37.867 Test Suite 'SALChannelTest' started at 2021-02-13 06:59:37.868 Test Case 'SALChannelTest.testBasicConnectWithClientBootstrap' started at 2021-02-13 06:59:37.868 /home/butta/swift-nio/Tests/NIOTests/SyscallAbstractionLayer.swift:506: error: SALChannelTest.testBasicConnectWithClientBootstrap : failed - unexpected syscall whenReady(NIO.SelectorStrategy.now) /home/butta/swift-nio/Tests/NIOTests/SALChannelTests.swift:241: error: SALChannelTest.testBasicConnectWithClientBootstrap : failed - couldn't make a channel Test Case 'SALChannelTest.testBasicConnectWithClientBootstrap' failed (10.003 seconds) Test Case 'SALChannelTest.testBasicConnectedChannel' started at 2021-02-13 06:59:47.871 /home/butta/swift-nio/Tests/NIOTests/SyscallAbstractionLayer.swift:506: error: SALChannelTest.testBasicConnectedChannel : failed - unexpected syscall whenReady(NIO.SelectorStrategy.now) <EXPR>:0: error: SALChannelTest.testBasicConnectedChannel : threw error "UnexpectedSyscall(syscall: NIOTests.UserToKernel.whenReady(NIO.SelectorStrategy.now))" Test Case 'SALChannelTest.testBasicConnectedChannel' failed (10.001 seconds) Test Case 'SALChannelTest.testBasicRead' started at 2021-02-13 06:59:57.872 /home/butta/swift-nio/Tests/NIOTests/SyscallAbstractionLayer.swift:506: error: SALChannelTest.testBasicRead : failed - unexpected syscall whenReady(NIO.SelectorStrategy.now) /home/butta/swift-nio/Tests/NIOTests/SALChannelTests.swift:216: error: SALChannelTest.testBasicRead : XCTAssertNoThrow failed: threw error "UnexpectedSyscall(syscall: NIOTests.UserToKernel.whenReady(NIO.SelectorStrategy.now))" - /home/butta/swift-nio/Tests/NIOTests/SALChannelTests.swift:219: error: SALChannelTest.testBasicRead : failed - couldn't construct channel Test Case 'SALChannelTest.testBasicRead' failed (10.001 seconds) Test Case 'SALChannelTest.testClientBootstrapBindIsDoneAfterSocketOptions' started at 2021-02-13 07:00:07.873 /home/butta/swift-nio/Tests/NIOTests/SyscallAbstractionLayer.swift:506: error: SALChannelTest.testClientBootstrapBindIsDoneAfterSocketOptions : failed - unexpected syscall whenReady(NIO.SelectorStrategy.now) /home/butta/swift-nio/Tests/NIOTests/SALChannelTests.swift:277: error: SALChannelTest.testClientBootstrapBindIsDoneAfterSocketOptions : failed - couldn't make a channel Test Case 'SALChannelTest.testClientBootstrapBindIsDoneAfterSocketOptions' failed (10.001 seconds) Test Case 'SALChannelTest.testWeSurviveIfIgnoringSIGPIPEFails' started at 2021-02-13 07:00:17.874 /home/butta/swift-nio/Tests/NIOTests/SyscallAbstractionLayer.swift:521: error: SALChannelTest.testWeSurviveIfIgnoringSIGPIPEFails : failed - unexpected syscall whenReady(NIO.SelectorStrategy.now) /home/butta/swift-nio/Tests/NIOTests/SALChannelTests.swift:187: error: SALChannelTest.testWeSurviveIfIgnoringSIGPIPEFails : XCTAssertEqual failed: ("Optional(22)") is not equal to ("nil") - Test Case 'SALChannelTest.testWeSurviveIfIgnoringSIGPIPEFails' failed (10.001 seconds) Test Case 'SALChannelTest.testWritesFromWritabilityNotificationsDoNotGetLostIfWePreviouslyWroteEverything' started at 2021-02-13 07:00:27.875 /home/butta/swift-nio/Tests/NIOTests/SyscallAbstractionLayer.swift:506: error: SALChannelTest.testWritesFromWritabilityNotificationsDoNotGetLostIfWePreviouslyWroteEverything : failed - unexpected syscall whenReady(NIO.SelectorStrategy.now) /home/butta/swift-nio/Tests/NIOTests/SALChannelTests.swift:117: error: SALChannelTest.testWritesFromWritabilityNotificationsDoNotGetLostIfWePreviouslyWroteEverything : XCTAssertNoThrow failed: threw error "UnexpectedSyscall(syscall: NIOTests.UserToKernel.whenReady(NIO.SelectorStrategy.now))" - /home/butta/swift-nio/Tests/NIOTests/SALChannelTests.swift:120: error: SALChannelTest.testWritesFromWritabilityNotificationsDoNotGetLostIfWePreviouslyWroteEverything : failed - couldn't construct channel Test Case 'SALChannelTest.testWritesFromWritabilityNotificationsDoNotGetLostIfWePreviouslyWroteEverything' failed (10.001 seconds) Test Suite 'SALChannelTest' failed at 2021-02-13 07:00:37.876 Executed 6 tests, with 14 failures (1 unexpected) in 60.008 (60.008) seconds Test Suite 'Selected tests' failed at 2021-02-13 07:00:37.876 Executed 6 tests, with 14 failures (1 unexpected) in 60.008 (60.008) seconds ``` ### SwiftNIO version/commit hash latest HEAD - ae3ea78 ### Swift & OS version ``` Swift version 5.3.3 (swift-5.3.3-RELEASE) Target: x86_64-unknown-linux-gnu ``` `uname -r` is `5.10.9-201.fc33.x86_64`.
1.0
Several SALChannelTest cases fail consistently on linux x86_64 - ### Expected behavior All tests to pass on Fedora Core 33, as they did on Ubuntu 20.04 over the last several months. ### Actual behavior Only the SALChannel tests fail, mostly with this error message `unexpected syscall whenReady(NIO.SelectorStrategy.now)`, which is some kind of race? I don't know if this is specific to Fedora, as I've since deleted my Ubuntu VPS. I see these errors occasionally on Android AArch64 too, which is why I then double-checked on linux, but they only randomly fail on Android when running the tests in `--parallel`, never serially. That flag makes no difference on Fedora, these tests always fail. ### Steps to reproduce ``` ../swift-5.3.3-RELEASE-centos8/usr/bin/swift test -j 2 --enable-test-discovery --filter SALC Test Suite 'Selected tests' started at 2021-02-13 06:59:37.867 Test Suite 'SALChannelTest' started at 2021-02-13 06:59:37.868 Test Case 'SALChannelTest.testBasicConnectWithClientBootstrap' started at 2021-02-13 06:59:37.868 /home/butta/swift-nio/Tests/NIOTests/SyscallAbstractionLayer.swift:506: error: SALChannelTest.testBasicConnectWithClientBootstrap : failed - unexpected syscall whenReady(NIO.SelectorStrategy.now) /home/butta/swift-nio/Tests/NIOTests/SALChannelTests.swift:241: error: SALChannelTest.testBasicConnectWithClientBootstrap : failed - couldn't make a channel Test Case 'SALChannelTest.testBasicConnectWithClientBootstrap' failed (10.003 seconds) Test Case 'SALChannelTest.testBasicConnectedChannel' started at 2021-02-13 06:59:47.871 /home/butta/swift-nio/Tests/NIOTests/SyscallAbstractionLayer.swift:506: error: SALChannelTest.testBasicConnectedChannel : failed - unexpected syscall whenReady(NIO.SelectorStrategy.now) <EXPR>:0: error: SALChannelTest.testBasicConnectedChannel : threw error "UnexpectedSyscall(syscall: NIOTests.UserToKernel.whenReady(NIO.SelectorStrategy.now))" Test Case 'SALChannelTest.testBasicConnectedChannel' failed (10.001 seconds) Test Case 'SALChannelTest.testBasicRead' started at 2021-02-13 06:59:57.872 /home/butta/swift-nio/Tests/NIOTests/SyscallAbstractionLayer.swift:506: error: SALChannelTest.testBasicRead : failed - unexpected syscall whenReady(NIO.SelectorStrategy.now) /home/butta/swift-nio/Tests/NIOTests/SALChannelTests.swift:216: error: SALChannelTest.testBasicRead : XCTAssertNoThrow failed: threw error "UnexpectedSyscall(syscall: NIOTests.UserToKernel.whenReady(NIO.SelectorStrategy.now))" - /home/butta/swift-nio/Tests/NIOTests/SALChannelTests.swift:219: error: SALChannelTest.testBasicRead : failed - couldn't construct channel Test Case 'SALChannelTest.testBasicRead' failed (10.001 seconds) Test Case 'SALChannelTest.testClientBootstrapBindIsDoneAfterSocketOptions' started at 2021-02-13 07:00:07.873 /home/butta/swift-nio/Tests/NIOTests/SyscallAbstractionLayer.swift:506: error: SALChannelTest.testClientBootstrapBindIsDoneAfterSocketOptions : failed - unexpected syscall whenReady(NIO.SelectorStrategy.now) /home/butta/swift-nio/Tests/NIOTests/SALChannelTests.swift:277: error: SALChannelTest.testClientBootstrapBindIsDoneAfterSocketOptions : failed - couldn't make a channel Test Case 'SALChannelTest.testClientBootstrapBindIsDoneAfterSocketOptions' failed (10.001 seconds) Test Case 'SALChannelTest.testWeSurviveIfIgnoringSIGPIPEFails' started at 2021-02-13 07:00:17.874 /home/butta/swift-nio/Tests/NIOTests/SyscallAbstractionLayer.swift:521: error: SALChannelTest.testWeSurviveIfIgnoringSIGPIPEFails : failed - unexpected syscall whenReady(NIO.SelectorStrategy.now) /home/butta/swift-nio/Tests/NIOTests/SALChannelTests.swift:187: error: SALChannelTest.testWeSurviveIfIgnoringSIGPIPEFails : XCTAssertEqual failed: ("Optional(22)") is not equal to ("nil") - Test Case 'SALChannelTest.testWeSurviveIfIgnoringSIGPIPEFails' failed (10.001 seconds) Test Case 'SALChannelTest.testWritesFromWritabilityNotificationsDoNotGetLostIfWePreviouslyWroteEverything' started at 2021-02-13 07:00:27.875 /home/butta/swift-nio/Tests/NIOTests/SyscallAbstractionLayer.swift:506: error: SALChannelTest.testWritesFromWritabilityNotificationsDoNotGetLostIfWePreviouslyWroteEverything : failed - unexpected syscall whenReady(NIO.SelectorStrategy.now) /home/butta/swift-nio/Tests/NIOTests/SALChannelTests.swift:117: error: SALChannelTest.testWritesFromWritabilityNotificationsDoNotGetLostIfWePreviouslyWroteEverything : XCTAssertNoThrow failed: threw error "UnexpectedSyscall(syscall: NIOTests.UserToKernel.whenReady(NIO.SelectorStrategy.now))" - /home/butta/swift-nio/Tests/NIOTests/SALChannelTests.swift:120: error: SALChannelTest.testWritesFromWritabilityNotificationsDoNotGetLostIfWePreviouslyWroteEverything : failed - couldn't construct channel Test Case 'SALChannelTest.testWritesFromWritabilityNotificationsDoNotGetLostIfWePreviouslyWroteEverything' failed (10.001 seconds) Test Suite 'SALChannelTest' failed at 2021-02-13 07:00:37.876 Executed 6 tests, with 14 failures (1 unexpected) in 60.008 (60.008) seconds Test Suite 'Selected tests' failed at 2021-02-13 07:00:37.876 Executed 6 tests, with 14 failures (1 unexpected) in 60.008 (60.008) seconds ``` ### SwiftNIO version/commit hash latest HEAD - ae3ea78 ### Swift & OS version ``` Swift version 5.3.3 (swift-5.3.3-RELEASE) Target: x86_64-unknown-linux-gnu ``` `uname -r` is `5.10.9-201.fc33.x86_64`.
test
several salchanneltest cases fail consistently on linux expected behavior all tests to pass on fedora core as they did on ubuntu over the last several months actual behavior only the salchannel tests fail mostly with this error message unexpected syscall whenready nio selectorstrategy now which is some kind of race i don t know if this is specific to fedora as i ve since deleted my ubuntu vps i see these errors occasionally on android too which is why i then double checked on linux but they only randomly fail on android when running the tests in parallel never serially that flag makes no difference on fedora these tests always fail steps to reproduce swift release usr bin swift test j enable test discovery filter salc test suite selected tests started at test suite salchanneltest started at test case salchanneltest testbasicconnectwithclientbootstrap started at home butta swift nio tests niotests syscallabstractionlayer swift error salchanneltest testbasicconnectwithclientbootstrap failed unexpected syscall whenready nio selectorstrategy now home butta swift nio tests niotests salchanneltests swift error salchanneltest testbasicconnectwithclientbootstrap failed couldn t make a channel test case salchanneltest testbasicconnectwithclientbootstrap failed seconds test case salchanneltest testbasicconnectedchannel started at home butta swift nio tests niotests syscallabstractionlayer swift error salchanneltest testbasicconnectedchannel failed unexpected syscall whenready nio selectorstrategy now error salchanneltest testbasicconnectedchannel threw error unexpectedsyscall syscall niotests usertokernel whenready nio selectorstrategy now test case salchanneltest testbasicconnectedchannel failed seconds test case salchanneltest testbasicread started at home butta swift nio tests niotests syscallabstractionlayer swift error salchanneltest testbasicread failed unexpected syscall whenready nio selectorstrategy now home butta swift nio tests niotests salchanneltests swift error salchanneltest testbasicread xctassertnothrow failed threw error unexpectedsyscall syscall niotests usertokernel whenready nio selectorstrategy now home butta swift nio tests niotests salchanneltests swift error salchanneltest testbasicread failed couldn t construct channel test case salchanneltest testbasicread failed seconds test case salchanneltest testclientbootstrapbindisdoneaftersocketoptions started at home butta swift nio tests niotests syscallabstractionlayer swift error salchanneltest testclientbootstrapbindisdoneaftersocketoptions failed unexpected syscall whenready nio selectorstrategy now home butta swift nio tests niotests salchanneltests swift error salchanneltest testclientbootstrapbindisdoneaftersocketoptions failed couldn t make a channel test case salchanneltest testclientbootstrapbindisdoneaftersocketoptions failed seconds test case salchanneltest testwesurviveifignoringsigpipefails started at home butta swift nio tests niotests syscallabstractionlayer swift error salchanneltest testwesurviveifignoringsigpipefails failed unexpected syscall whenready nio selectorstrategy now home butta swift nio tests niotests salchanneltests swift error salchanneltest testwesurviveifignoringsigpipefails xctassertequal failed optional is not equal to nil test case salchanneltest testwesurviveifignoringsigpipefails failed seconds test case salchanneltest testwritesfromwritabilitynotificationsdonotgetlostifwepreviouslywroteeverything started at home butta swift nio tests niotests syscallabstractionlayer swift error salchanneltest testwritesfromwritabilitynotificationsdonotgetlostifwepreviouslywroteeverything failed unexpected syscall whenready nio selectorstrategy now home butta swift nio tests niotests salchanneltests swift error salchanneltest testwritesfromwritabilitynotificationsdonotgetlostifwepreviouslywroteeverything xctassertnothrow failed threw error unexpectedsyscall syscall niotests usertokernel whenready nio selectorstrategy now home butta swift nio tests niotests salchanneltests swift error salchanneltest testwritesfromwritabilitynotificationsdonotgetlostifwepreviouslywroteeverything failed couldn t construct channel test case salchanneltest testwritesfromwritabilitynotificationsdonotgetlostifwepreviouslywroteeverything failed seconds test suite salchanneltest failed at executed tests with failures unexpected in seconds test suite selected tests failed at executed tests with failures unexpected in seconds swiftnio version commit hash latest head swift os version swift version swift release target unknown linux gnu uname r is
1
218,459
16,991,955,836
IssuesEvent
2021-06-30 21:57:24
microsoft/vscode
https://api.github.com/repos/microsoft/vscode
closed
Test CompletionItemLabel
testplan-item
Refs: https://github.com/microsoft/vscode/issues/39441 - [x] anyOS @mjbvz - [x] anyOS @meganrogge Complexity: 2 [Create Issue](https://github.com/microsoft/vscode/issues/new?body=Testing+%23127333%0A%0A) --- The `vscode.CompletionItem#label`-property can now, in addition to a string, be a `CompletionItemLabel`. This allows for more structured labels. Have an extension that implements a completion item provider, e.g start with https://github.com/microsoft/vscode-extension-samples/tree/main/completions-sample, and use the `CompletionItemLabel`. Test that (1) things work, (2) are documented properly, and (3) is good API
1.0
Test CompletionItemLabel - Refs: https://github.com/microsoft/vscode/issues/39441 - [x] anyOS @mjbvz - [x] anyOS @meganrogge Complexity: 2 [Create Issue](https://github.com/microsoft/vscode/issues/new?body=Testing+%23127333%0A%0A) --- The `vscode.CompletionItem#label`-property can now, in addition to a string, be a `CompletionItemLabel`. This allows for more structured labels. Have an extension that implements a completion item provider, e.g start with https://github.com/microsoft/vscode-extension-samples/tree/main/completions-sample, and use the `CompletionItemLabel`. Test that (1) things work, (2) are documented properly, and (3) is good API
test
test completionitemlabel refs anyos mjbvz anyos meganrogge complexity the vscode completionitem label property can now in addition to a string be a completionitemlabel this allows for more structured labels have an extension that implements a completion item provider e g start with and use the completionitemlabel test that things work are documented properly and is good api
1
348,719
10,451,875,137
IssuesEvent
2019-09-19 13:42:04
ntop/ntopng
https://api.github.com/repos/ntop/ntopng
closed
Invalid number formatting
in progress low-priority bug priority ticket
![Screenshot from 2019-09-19 10-06-08](https://user-images.githubusercontent.com/4493366/65225877-2c1d2c80-dac6-11e9-82e2-557e0bda5043.png) The number formatting (1233) is inconsistent across columns: it must be unified in all pages displayed.
2.0
Invalid number formatting - ![Screenshot from 2019-09-19 10-06-08](https://user-images.githubusercontent.com/4493366/65225877-2c1d2c80-dac6-11e9-82e2-557e0bda5043.png) The number formatting (1233) is inconsistent across columns: it must be unified in all pages displayed.
non_test
invalid number formatting the number formatting is inconsistent across columns it must be unified in all pages displayed
0
176,653
13,648,482,919
IssuesEvent
2020-09-26 09:27:40
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
sql/logictest: TestLogic/local-vec-auto/upsert/regression_35364 timed out
C-test-failure O-robot branch-release-20.2
[(sql/logictest).TestLogic/local-vec-auto/upsert/regression_35364 failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2324622&tab=buildLog) on [release-20.2@305ee38104e147c971fe21139082eb81614b4766](https://github.com/cockroachdb/cockroach/commits/305ee38104e147c971fe21139082eb81614b4766): ``` Slow failing tests: TestLogic/local-vec-auto/upsert/regression_35364 - 50.25s Slow passing tests: <none> ``` <details><summary>More</summary><p> ``` make stressrace TESTS=TestLogic/local-vec-auto/upsert/regression_35364 PKG=./pkg/sql/logictest TESTTIMEOUT=5m STRESSFLAGS='-timeout 5m' 2>&1 ``` [See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2ATestLogic%2Flocal-vec-auto%2Fupsert%2Fregression_35364.%2A&sort=title&restgroup=false&display=lastcommented+project) <sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
1.0
sql/logictest: TestLogic/local-vec-auto/upsert/regression_35364 timed out - [(sql/logictest).TestLogic/local-vec-auto/upsert/regression_35364 failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2324622&tab=buildLog) on [release-20.2@305ee38104e147c971fe21139082eb81614b4766](https://github.com/cockroachdb/cockroach/commits/305ee38104e147c971fe21139082eb81614b4766): ``` Slow failing tests: TestLogic/local-vec-auto/upsert/regression_35364 - 50.25s Slow passing tests: <none> ``` <details><summary>More</summary><p> ``` make stressrace TESTS=TestLogic/local-vec-auto/upsert/regression_35364 PKG=./pkg/sql/logictest TESTTIMEOUT=5m STRESSFLAGS='-timeout 5m' 2>&1 ``` [See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2ATestLogic%2Flocal-vec-auto%2Fupsert%2Fregression_35364.%2A&sort=title&restgroup=false&display=lastcommented+project) <sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
test
sql logictest testlogic local vec auto upsert regression timed out on slow failing tests testlogic local vec auto upsert regression slow passing tests more make stressrace tests testlogic local vec auto upsert regression pkg pkg sql logictest testtimeout stressflags timeout powered by
1
86,533
8,038,998,487
IssuesEvent
2018-07-30 16:58:29
zetkin/call.zetk.in
https://api.github.com/repos/zetkin/call.zetk.in
closed
Mobile: Trying to change address in target pane
user test
Several user looks for a way to report a change of address in the target pane (where read-only personal information is available). One user struggles to find the way to do it at all (the correct way is using "Organizer action needed" during the report). While looking for the feature, the user repeatedly returns to the target pane to look for it there. _This was observed during the July 2018 mobile user tests_
1.0
Mobile: Trying to change address in target pane - Several user looks for a way to report a change of address in the target pane (where read-only personal information is available). One user struggles to find the way to do it at all (the correct way is using "Organizer action needed" during the report). While looking for the feature, the user repeatedly returns to the target pane to look for it there. _This was observed during the July 2018 mobile user tests_
test
mobile trying to change address in target pane several user looks for a way to report a change of address in the target pane where read only personal information is available one user struggles to find the way to do it at all the correct way is using organizer action needed during the report while looking for the feature the user repeatedly returns to the target pane to look for it there this was observed during the july mobile user tests
1
46,466
9,948,707,898
IssuesEvent
2019-07-04 09:32:25
atomist/sdm-core
https://api.github.com/repos/atomist/sdm-core
closed
Code Inspection: Tslint on create-goals-on-sdm
bug code-inspection
### no-shadowed-variable - [`lib/machine/configure.ts:159`](https://github.com/atomist/sdm-core/blob/c0d7e3dc92163b0db693866983af7918d82090d1/lib/machine/configure.ts#L159): _(error)_ Shadowed name: 'goals' - [`lib/machine/configure.ts:161`](https://github.com/atomist/sdm-core/blob/c0d7e3dc92163b0db693866983af7918d82090d1/lib/machine/configure.ts#L161): _(error)_ Shadowed name: 'configurer' [atomist:code-inspection:create-goals-on-sdm=@atomist/atomist-sdm]
1.0
Code Inspection: Tslint on create-goals-on-sdm - ### no-shadowed-variable - [`lib/machine/configure.ts:159`](https://github.com/atomist/sdm-core/blob/c0d7e3dc92163b0db693866983af7918d82090d1/lib/machine/configure.ts#L159): _(error)_ Shadowed name: 'goals' - [`lib/machine/configure.ts:161`](https://github.com/atomist/sdm-core/blob/c0d7e3dc92163b0db693866983af7918d82090d1/lib/machine/configure.ts#L161): _(error)_ Shadowed name: 'configurer' [atomist:code-inspection:create-goals-on-sdm=@atomist/atomist-sdm]
non_test
code inspection tslint on create goals on sdm no shadowed variable error shadowed name goals error shadowed name configurer
0
137,590
30,719,098,673
IssuesEvent
2023-07-27 14:47:27
postmanlabs/postman-app-support
https://api.github.com/repos/postmanlabs/postman-app-support
closed
VSCode Extension no longer able to change view from top to bottom to side to side.
bug vscode
### Is there an existing issue for this? - [X] I have searched the tracker for existing similar issues and I know that duplicates will be closed ### Describe the Issue In the latest update (0.5.1) to the VS Code extension the button in the request or response panes that allowed toggling from a top to bottom view to a side to side view is no longer available. This is also called "Single Paned View" and "Two Paned View" ### Steps To Reproduce Update Postman extension in VS Code to 0.5.1 Send a request to a service. in request and response panes there is no longer an option to change panel layout. ### Screenshots or Videos _No response_ ### Operating System mac osX ### Postman Version 0.5.1 vs code postman extension ### Postman Platform Other ### User Account Type Signed In User ### Additional Context? _No response_
1.0
VSCode Extension no longer able to change view from top to bottom to side to side. - ### Is there an existing issue for this? - [X] I have searched the tracker for existing similar issues and I know that duplicates will be closed ### Describe the Issue In the latest update (0.5.1) to the VS Code extension the button in the request or response panes that allowed toggling from a top to bottom view to a side to side view is no longer available. This is also called "Single Paned View" and "Two Paned View" ### Steps To Reproduce Update Postman extension in VS Code to 0.5.1 Send a request to a service. in request and response panes there is no longer an option to change panel layout. ### Screenshots or Videos _No response_ ### Operating System mac osX ### Postman Version 0.5.1 vs code postman extension ### Postman Platform Other ### User Account Type Signed In User ### Additional Context? _No response_
non_test
vscode extension no longer able to change view from top to bottom to side to side is there an existing issue for this i have searched the tracker for existing similar issues and i know that duplicates will be closed describe the issue in the latest update to the vs code extension the button in the request or response panes that allowed toggling from a top to bottom view to a side to side view is no longer available this is also called single paned view and two paned view steps to reproduce update postman extension in vs code to send a request to a service in request and response panes there is no longer an option to change panel layout screenshots or videos no response operating system mac osx postman version vs code postman extension postman platform other user account type signed in user additional context no response
0
263,185
23,040,227,594
IssuesEvent
2022-07-23 03:11:53
FTBTeam/FTB-Modpack-Issues
https://api.github.com/repos/FTBTeam/FTB-Modpack-Issues
closed
[Bug]: Crafting Magebloom seed crashes the game
1: Mod Issue state: Testing required FTB One
### Modpack One ### Modpack version 1.3.0 ### Log Files [crash-2022-07-22_16.49.45-client.txt](https://github.com/FTBTeam/FTB-Modpack-Issues/files/9172173/crash-2022-07-22_16.49.45-client.txt) [crash-2022-07-22_16.59.07-client.txt](https://github.com/FTBTeam/FTB-Modpack-Issues/files/9172174/crash-2022-07-22_16.59.07-client.txt) ### Describe the bug Unsure what logs may be useful - attempting to make magebloom seed has resulted in successive crashes. Suspect that this might be Ars's enchanting apparatus, but not entirely sure. (Will need further validation). Also might be related to the placement of enchantment apparatus at same level as the pedestals. ### Steps to reproduce Place Arcane gems on pedestals, place wheat seed in Enchantment Apparatus. Sound plays, but game crashes to desktop. ### Expected behaviour Enchantment Apparatus should play short animation, then output correct recipie ### Screenshots _No response_ ### Additional information _No response_
1.0
[Bug]: Crafting Magebloom seed crashes the game - ### Modpack One ### Modpack version 1.3.0 ### Log Files [crash-2022-07-22_16.49.45-client.txt](https://github.com/FTBTeam/FTB-Modpack-Issues/files/9172173/crash-2022-07-22_16.49.45-client.txt) [crash-2022-07-22_16.59.07-client.txt](https://github.com/FTBTeam/FTB-Modpack-Issues/files/9172174/crash-2022-07-22_16.59.07-client.txt) ### Describe the bug Unsure what logs may be useful - attempting to make magebloom seed has resulted in successive crashes. Suspect that this might be Ars's enchanting apparatus, but not entirely sure. (Will need further validation). Also might be related to the placement of enchantment apparatus at same level as the pedestals. ### Steps to reproduce Place Arcane gems on pedestals, place wheat seed in Enchantment Apparatus. Sound plays, but game crashes to desktop. ### Expected behaviour Enchantment Apparatus should play short animation, then output correct recipie ### Screenshots _No response_ ### Additional information _No response_
test
crafting magebloom seed crashes the game modpack one modpack version log files describe the bug unsure what logs may be useful attempting to make magebloom seed has resulted in successive crashes suspect that this might be ars s enchanting apparatus but not entirely sure will need further validation also might be related to the placement of enchantment apparatus at same level as the pedestals steps to reproduce place arcane gems on pedestals place wheat seed in enchantment apparatus sound plays but game crashes to desktop expected behaviour enchantment apparatus should play short animation then output correct recipie screenshots no response additional information no response
1
47,607
5,904,962,397
IssuesEvent
2017-05-19 11:22:17
openSUSE/xmldiffng
https://api.github.com/repos/openSUSE/xmldiffng
opened
High-Level Overview of Improvements
doc enhancement help wanted testing
This is a high-level overview of what is needed: * [ ] Migrate to Python3, see #7 * [ ] Support lxml * [ ] Embrace TDD and pytest ## Codestyle * [ ] Follow [PEP257](https://www.python.org/dev/peps/pep-0257/) * [ ] Follow [PEP8](https://www.python.org/dev/peps/pep-0008/) * [ ] Follow Landscape, Scrutinizer, Codecov, or other code testing tools
1.0
High-Level Overview of Improvements - This is a high-level overview of what is needed: * [ ] Migrate to Python3, see #7 * [ ] Support lxml * [ ] Embrace TDD and pytest ## Codestyle * [ ] Follow [PEP257](https://www.python.org/dev/peps/pep-0257/) * [ ] Follow [PEP8](https://www.python.org/dev/peps/pep-0008/) * [ ] Follow Landscape, Scrutinizer, Codecov, or other code testing tools
test
high level overview of improvements this is a high level overview of what is needed migrate to see support lxml embrace tdd and pytest codestyle follow follow follow landscape scrutinizer codecov or other code testing tools
1
55,500
6,480,235,467
IssuesEvent
2017-08-18 12:56:08
openbmc/openbmc-test-automation
https://api.github.com/repos/openbmc/openbmc-test-automation
closed
Software version test case for BMC and PNOR
Test
With the new PNOR object in the software the existing test case needs to be little bit smart to differentiate between BMC and Host
1.0
Software version test case for BMC and PNOR - With the new PNOR object in the software the existing test case needs to be little bit smart to differentiate between BMC and Host
test
software version test case for bmc and pnor with the new pnor object in the software the existing test case needs to be little bit smart to differentiate between bmc and host
1
181,313
14,015,355,368
IssuesEvent
2020-10-29 13:13:37
numpy/numpy
https://api.github.com/repos/numpy/numpy
closed
TST: typing tests slow on shippable arm64 runs
05 - Testing component: numpy.typing static typing
Our arm64 runs are quite slow, so we add `--durations` to find the slowest tests. Here are the results from a [random run](https://app.shippable.com/github/numpy/numpy/dashboard) (choose a successful Job from "Latest Status" -> "build_ci" -> python runtests.py ... -> scroll to the end) . The typing tests are almost an order of magnitude slower than the rest, and take 201 of the 581 seconds total to run. On non-ARM test runs they do not show this slowdown. Any thoughts? ``` -- Docs: https://docs.pytest.org/en/latest/warnings.html - generated xml file: /root/src/github.com/numpy/numpy/shippable/testresults/tests.xml - ========================== slowest 10 test durations =========================== 122.43s call tests/test_typing.py::test_success[numerictypes.py] 88.92s call tests/test_typing.py::test_success[simple_py3.py] 14.69s call f2py/tests/test_compile_function.py::test_f2py_init_compile[extra_args0] 14.64s call f2py/tests/test_compile_function.py::test_f2py_init_compile[] 14.64s call f2py/tests/test_compile_function.py::test_f2py_init_compile[--noopt --debug] 14.19s call tests/test_typing.py::test_fail[fromnumeric.py] 11.50s call tests/test_typing.py::test_reveal[fromnumeric.py] 11.43s call tests/test_typing.py::test_success[fromnumeric.py] 7.93s setup core/tests/test_cython.py::test_get_datetime64_unit 7.93s setup f2py/tests/test_return_integer.py::TestF90ReturnInteger::test_all[t0] ```
1.0
TST: typing tests slow on shippable arm64 runs - Our arm64 runs are quite slow, so we add `--durations` to find the slowest tests. Here are the results from a [random run](https://app.shippable.com/github/numpy/numpy/dashboard) (choose a successful Job from "Latest Status" -> "build_ci" -> python runtests.py ... -> scroll to the end) . The typing tests are almost an order of magnitude slower than the rest, and take 201 of the 581 seconds total to run. On non-ARM test runs they do not show this slowdown. Any thoughts? ``` -- Docs: https://docs.pytest.org/en/latest/warnings.html - generated xml file: /root/src/github.com/numpy/numpy/shippable/testresults/tests.xml - ========================== slowest 10 test durations =========================== 122.43s call tests/test_typing.py::test_success[numerictypes.py] 88.92s call tests/test_typing.py::test_success[simple_py3.py] 14.69s call f2py/tests/test_compile_function.py::test_f2py_init_compile[extra_args0] 14.64s call f2py/tests/test_compile_function.py::test_f2py_init_compile[] 14.64s call f2py/tests/test_compile_function.py::test_f2py_init_compile[--noopt --debug] 14.19s call tests/test_typing.py::test_fail[fromnumeric.py] 11.50s call tests/test_typing.py::test_reveal[fromnumeric.py] 11.43s call tests/test_typing.py::test_success[fromnumeric.py] 7.93s setup core/tests/test_cython.py::test_get_datetime64_unit 7.93s setup f2py/tests/test_return_integer.py::TestF90ReturnInteger::test_all[t0] ```
test
tst typing tests slow on shippable runs our runs are quite slow so we add durations to find the slowest tests here are the results from a choose a successful job from latest status build ci python runtests py scroll to the end the typing tests are almost an order of magnitude slower than the rest and take of the seconds total to run on non arm test runs they do not show this slowdown any thoughts docs generated xml file root src github com numpy numpy shippable testresults tests xml slowest test durations call tests test typing py test success call tests test typing py test success call tests test compile function py test init compile call tests test compile function py test init compile call tests test compile function py test init compile call tests test typing py test fail call tests test typing py test reveal call tests test typing py test success setup core tests test cython py test get unit setup tests test return integer py test all
1
50,917
12,602,103,219
IssuesEvent
2020-06-11 11:05:04
spack/spack
https://api.github.com/repos/spack/spack
opened
Installation issue: mvapich2
build-error
<!-- Thanks for taking the time to report this build failure. To proceed with the report please: 1. Title the issue "Installation issue: <name-of-the-package>". 2. Provide the information required below. We encourage you to try, as much as possible, to reduce your problem to the minimal example that still reproduces the issue. That would help us a lot in fixing it quickly and effectively! --> I'm trying to build mvapich2 in CentOS 7 with gcc 7.3.1 and configured for slurm and nemesis. The build succeeds for mvapich2@2.3.3 but fails for mvapich2@2.3.4. ### Steps to reproduce the issue <!-- Fill in the exact spec you are trying to build and the relevant part of the error message --> ```console $ spack install mvapich2@2.3.4 process_managers=slurm fabrics=nemesis ... ==> 23457: Installing mvapich2 ==> Finding buildcaches in /home/vagrant/mirror/build_cache ==> Fetching file:///home/vagrant/mirror/build_cache/linux-centos7-sandybridge-gcc-8.3.1-mvapich2-2.3.4-a2ixxlfjcx62bu66ytl7corlsczgydxd.spec.yaml curl: (37) Couldn't open file /home/vagrant/mirror/build_cache/linux-centos7-sandybridge-gcc-8.3.1-mvapich2-2.3.4-a2ixxlfjcx62bu66ytl7corlsczgydxd.spec.yaml ==> Failed to fetch file from URL: file:///home/vagrant/mirror/build_cache/linux-centos7-sandybridge-gcc-8.3.1-mvapich2-2.3.4-a2ixxlfjcx62bu66ytl7corlsczgydxd.spec.yaml Curl failed with error 37 ==> Fetching from file:///home/vagrant/mirror/build_cache/linux-centos7-sandybridge-gcc-8.3.1-mvapich2-2.3.4-a2ixxlfjcx62bu66ytl7corlsczgydxd.spec.yaml failed. ==> Using cached archive: /home/vagrant/spack/var/spack/cache/_source-cache/archive/72/7226a45c7c98333c8e5d2888119cce186199b430c13b7b1dca1769909e68ea7a.tar.gz ==> Staging archive: /tmp/vagrant/spack-stage/spack-stage-mvapich2-2.3.4-a2ixxlfjcx62bu66ytl7corlsczgydxd/mvapich2-2.3.4.tar.gz ==> Created stage in /tmp/vagrant/spack-stage/spack-stage-mvapich2-2.3.4-a2ixxlfjcx62bu66ytl7corlsczgydxd ==> No patches needed for mvapich2 ==> 23457: mvapich2: Building mvapich2 [AutotoolsPackage] ==> 23457: mvapich2: Executing phase: 'autoreconf' ==> 23457: mvapich2: Executing phase: 'configure' ==> 23457: mvapich2: Executing phase: 'build' ==> Error: ProcessError: Command exited with status 2: 'make' '-j4' 1 error found in build log: 2897 mv -f src/mpi/coll/.deps/lib_libmpi_la-nbcutil.Tpo src/mpi/coll/.d eps/lib_libmpi_la-nbcutil.Plo 2898 /bin/sh ./libtool --tag=CC --mode=compile /home/vagrant/spack/l ib/spack/env/gcc/gcc -DHAVE_CONFIG_H -I. -I./src/include -I./src/ include -I./src/mpi/datatype -I./src/mpi/spawn -I./src/util/dbg -I src/util/logging/common -I./src/util/mem -I./src/util/thread -I./s rc/util/type -I./src/util/wrappers -I./src/util/wrappers -I./src/u til/assert -I./src/util/timers -I./src/util/pointer -I./src/bindin g/cxx -I/tmp/vagrant/spack-stage/spack-stage-mvapich2-2.3.4-a2ixxl fjcx62bu66ytl7corlsczgydxd/spack-src/src/binding/fortran/mpif_h -I src/binding/fortran/use_mpi -I./src/mpid/ch3/include -I./src/mpid /ch3/include -I/tmp/vagrant/spack-stage/spack-stage-mvapich2-2.3.4 -a2ixxlfjcx62bu66ytl7corlsczgydxd/spack-src/src/mpid/ch3/channels/ common/include -I./src/mpid/ch3/channels/nemesis/include -I./src/ mpid/ch3/channels/nemesis/include -I./src/mpid/ch3/channels/nemes is/utils/monitor -I./src/mpid/common/datatype -I./src /mpid/common/datatype -I./src/mpid/common/sched -I./src/mpid/com mon/thread -I./src/mpid/common/timers -I./src/pmi/simple -I/t mp/vagrant/spack-stage/spack-stage-mvapich2-2.3.4-a2ixxlfjcx62bu66 ytl7corlsczgydxd/spack-src/src/mpl/include -I/tmp/vagrant/spack-st age/spack-stage-mvapich2-2.3.4-a2ixxlfjcx62bu66ytl7corlsczgydxd/sp ack-src/src/mpl/include -I/tmp/vagrant/spack-stage/spack-stage-mva pich2-2.3.4-a2ixxlfjcx62bu66ytl7corlsczgydxd/spack-src/src/openpa/ src -I/tmp/vagrant/spack-stage/spack-stage-mvapich2-2.3.4-a2ixxlfj cx62bu66ytl7corlsczgydxd/spack-src/src/openpa/src -D_REENTRANT -I/ tmp/vagrant/spack-stage/spack-stage-mvapich2-2.3.4-a2ixxlfjcx62bu6 6ytl7corlsczgydxd/spack-src/src/mpi/romio/include -I/home/vagrant/ spack/opt/spack/linux-centos7-sandybridge/gcc-8.3.1/slurm-19-05-6- 1-qawb2uwcv3evmdj3gqmrybduqfze3tou/include -DNDEBUG -DNVALGRIND - O2 -MT src/mpi/errhan/lib_libmpi_la-errutil.lo -MD -MP -MF src/mpi /errhan/.deps/lib_libmpi_la-errutil.Tpo -c -o src/mpi/errhan/lib_l ibmpi_la-errutil.lo `test -f 'src/mpi/errhan/errutil.c' || echo '. /'`src/mpi/errhan/errutil.c 2899 libtool: compile: /home/vagrant/spack/lib/spack/env/gcc/gcc -DHAV E_CONFIG_H -I. -I./src/include -I./src/include -I./src/mpi/datatyp e -I./src/mpi/spawn -I./src/util/dbg -Isrc/util/logging/common -I. /src/util/mem -I./src/util/thread -I./src/util/type -I./src/util/w rappers -I./src/util/wrappers -I./src/util/assert -I./src/util/tim ers -I./src/util/pointer -I./src/binding/cxx -I/tmp/vagrant/spack- stage/spack-stage-mvapich2-2.3.4-a2ixxlfjcx62bu66ytl7corlsczgydxd/ spack-src/src/binding/fortran/mpif_h -Isrc/binding/fortran/use_mpi -I./src/mpid/ch3/include -I./src/mpid/ch3/include -I/tmp/vagrant/ spack-stage/spack-stage-mvapich2-2.3.4-a2ixxlfjcx62bu66ytl7corlscz gydxd/spack-src/src/mpid/ch3/channels/common/include -I./src/mpid/ ch3/channels/nemesis/include -I./src/mpid/ch3/channels/nemesis/inc lude -I./src/mpid/ch3/channels/nemesis/utils/monitor -I./src/mpid/ common/datatype -I./src/mpid/common/datatype -I./src/mpid/common/s ched -I./src/mpid/common/thread -I./src/mpid/common/timers -I./src /pmi/simple -I/tmp/vagrant/spack-stage/spack-stage-mvapich2-2.3.4- a2ixxlfjcx62bu66ytl7corlsczgydxd/spack-src/src/mpl/include -I/tmp/ vagrant/spack-stage/spack-stage-mvapich2-2.3.4-a2ixxlfjcx62bu66ytl 7corlsczgydxd/spack-src/src/mpl/include -I/tmp/vagrant/spack-stage /spack-stage-mvapich2-2.3.4-a2ixxlfjcx62bu66ytl7corlsczgydxd/spack -src/src/openpa/src -I/tmp/vagrant/spack-stage/spack-stage-mvapich 2-2.3.4-a2ixxlfjcx62bu66ytl7corlsczgydxd/spack-src/src/openpa/src -D_REENTRANT -I/tmp/vagrant/spack-stage/spack-stage-mvapich2-2.3.4 -a2ixxlfjcx62bu66ytl7corlsczgydxd/spack-src/src/mpi/romio/include -I/home/vagrant/spack/opt/spack/linux-centos7-sandybridge/gcc-8.3. 1/slurm-19-05-6-1-qawb2uwcv3evmdj3gqmrybduqfze3tou/include -DNDEBU G -DNVALGRIND -O2 -MT src/mpi/comm/lib_libmpi_la-commutil.lo -MD - MP -MF src/mpi/comm/.deps/lib_libmpi_la-commutil.Tpo -c src/mpi/co mm/commutil.c -fPIC -DPIC -o src/mpi/comm/.libs/lib_libmpi_la-com mutil.o 2900 libtool: compile: /home/vagrant/spack/lib/spack/env/gcc/gcc -DHAV E_CONFIG_H -I. -I./src/include -I./src/include -I./src/mpi/datatyp e -I./src/mpi/spawn -I./src/util/dbg -Isrc/util/logging/common -I. /src/util/mem -I./src/util/thread -I./src/util/type -I./src/util/w rappers -I./src/util/wrappers -I./src/util/assert -I./src/util/tim ers -I./src/util/pointer -I./src/binding/cxx -I/tmp/vagrant/spack- stage/spack-stage-mvapich2-2.3.4-a2ixxlfjcx62bu66ytl7corlsczgydxd/ spack-src/src/binding/fortran/mpif_h -Isrc/binding/fortran/use_mpi -I./src/mpid/ch3/include -I./src/mpid/ch3/include -I/tmp/vagrant/ spack-stage/spack-stage-mvapich2-2.3.4-a2ixxlfjcx62bu66ytl7corlscz gydxd/spack-src/src/mpid/ch3/channels/common/include -I./src/mpid/ ch3/channels/nemesis/include -I./src/mpid/ch3/channels/nemesis/inc lude -I./src/mpid/ch3/channels/nemesis/utils/monitor -I./src/mpid/ common/datatype -I./src/mpid/common/datatype -I./src/mpid/common/s ched -I./src/mpid/common/thread -I./src/mpid/common/timers -I./src /pmi/simple -I/tmp/vagrant/spack-stage/spack-stage-mvapich2-2.3.4- a2ixxlfjcx62bu66ytl7corlsczgydxd/spack-src/src/mpl/include -I/tmp/ vagrant/spack-stage/spack-stage-mvapich2-2.3.4-a2ixxlfjcx62bu66ytl 7corlsczgydxd/spack-src/src/mpl/include -I/tmp/vagrant/spack-stage /spack-stage-mvapich2-2.3.4-a2ixxlfjcx62bu66ytl7corlsczgydxd/spack -src/src/openpa/src -I/tmp/vagrant/spack-stage/spack-stage-mvapich 2-2.3.4-a2ixxlfjcx62bu66ytl7corlsczgydxd/spack-src/src/openpa/src -D_REENTRANT -I/tmp/vagrant/spack-stage/spack-stage-mvapich2-2.3.4 -a2ixxlfjcx62bu66ytl7corlsczgydxd/spack-src/src/mpi/romio/include -I/home/vagrant/spack/opt/spack/linux-centos7-sandybridge/gcc-8.3. 1/slurm-19-05-6-1-qawb2uwcv3evmdj3gqmrybduqfze3tou/include -DNDEBU G -DNVALGRIND -O2 -MT src/mpi/comm/lib_libmpi_la-contextid.lo -MD -MP -MF src/mpi/comm/.deps/lib_libmpi_la-contextid.Tpo -c src/mpi/ comm/contextid.c -fPIC -DPIC -o src/mpi/comm/.libs/lib_libmpi_la- contextid.o 2901 libtool: compile: /home/vagrant/spack/lib/spack/env/gcc/gcc -DHAV E_CONFIG_H -I. -I./src/include -I./src/include -I./src/mpi/datatyp e -I./src/mpi/spawn -I./src/util/dbg -Isrc/util/logging/common -I. /src/util/mem -I./src/util/thread -I./src/util/type -I./src/util/w rappers -I./src/util/wrappers -I./src/util/assert -I./src/util/tim ers -I./src/util/pointer -I./src/binding/cxx -I/tmp/vagrant/spack- stage/spack-stage-mvapich2-2.3.4-a2ixxlfjcx62bu66ytl7corlsczgydxd/ spack-src/src/binding/fortran/mpif_h -Isrc/binding/fortran/use_mpi -I./src/mpid/ch3/include -I./src/mpid/ch3/include -I/tmp/vagrant/ spack-stage/spack-stage-mvapich2-2.3.4-a2ixxlfjcx62bu66ytl7corlscz gydxd/spack-src/src/mpid/ch3/channels/common/include -I./src/mpid/ ch3/channels/nemesis/include -I./src/mpid/ch3/channels/nemesis/inc lude -I./src/mpid/ch3/channels/nemesis/utils/monitor -I./src/mpid/ common/datatype -I./src/mpid/common/datatype -I./src/mpid/common/s ched -I./src/mpid/common/thread -I./src/mpid/common/timers -I./src /pmi/simple -I/tmp/vagrant/spack-stage/spack-stage-mvapich2-2.3.4- a2ixxlfjcx62bu66ytl7corlsczgydxd/spack-src/src/mpl/include -I/tmp/ vagrant/spack-stage/spack-stage-mvapich2-2.3.4-a2ixxlfjcx62bu66ytl 7corlsczgydxd/spack-src/src/mpl/include -I/tmp/vagrant/spack-stage /spack-stage-mvapich2-2.3.4-a2ixxlfjcx62bu66ytl7corlsczgydxd/spack -src/src/openpa/src -I/tmp/vagrant/spack-stage/spack-stage-mvapich 2-2.3.4-a2ixxlfjcx62bu66ytl7corlsczgydxd/spack-src/src/openpa/src -D_REENTRANT -I/tmp/vagrant/spack-stage/spack-stage-mvapich2-2.3.4 -a2ixxlfjcx62bu66ytl7corlsczgydxd/spack-src/src/mpi/romio/include -I/home/vagrant/spack/opt/spack/linux-centos7-sandybridge/gcc-8.3. 1/slurm-19-05-6-1-qawb2uwcv3evmdj3gqmrybduqfze3tou/include -DNDEBU G -DNVALGRIND -O2 -MT src/mpi/datatype/lib_libmpi_la-typeutil.lo - MD -MP -MF src/mpi/datatype/.deps/lib_libmpi_la-typeutil.Tpo -c sr c/mpi/datatype/typeutil.c -fPIC -DPIC -o src/mpi/datatype/.libs/l ib_libmpi_la-typeutil.o 2902 libtool: compile: /home/vagrant/spack/lib/spack/env/gcc/gcc -DHAV E_CONFIG_H -I. -I./src/include -I./src/include -I./src/mpi/datatyp e -I./src/mpi/spawn -I./src/util/dbg -Isrc/util/logging/common -I. /src/util/mem -I./src/util/thread -I./src/util/type -I./src/util/w rappers -I./src/util/wrappers -I./src/util/assert -I./src/util/tim ers -I./src/util/pointer -I./src/binding/cxx -I/tmp/vagrant/spack- stage/spack-stage-mvapich2-2.3.4-a2ixxlfjcx62bu66ytl7corlsczgydxd/ spack-src/src/binding/fortran/mpif_h -Isrc/binding/fortran/use_mpi -I./src/mpid/ch3/include -I./src/mpid/ch3/include -I/tmp/vagrant/ spack-stage/spack-stage-mvapich2-2.3.4-a2ixxlfjcx62bu66ytl7corlscz gydxd/spack-src/src/mpid/ch3/channels/common/include -I./src/mpid/ ch3/channels/nemesis/include -I./src/mpid/ch3/channels/nemesis/inc lude -I./src/mpid/ch3/channels/nemesis/utils/monitor -I./src/mpid/ common/datatype -I./src/mpid/common/datatype -I./src/mpid/common/s ched -I./src/mpid/common/thread -I./src/mpid/common/timers -I./src /pmi/simple -I/tmp/vagrant/spack-stage/spack-stage-mvapich2-2.3.4- a2ixxlfjcx62bu66ytl7corlsczgydxd/spack-src/src/mpl/include -I/tmp/ vagrant/spack-stage/spack-stage-mvapich2-2.3.4-a2ixxlfjcx62bu66ytl 7corlsczgydxd/spack-src/src/mpl/include -I/tmp/vagrant/spack-stage /spack-stage-mvapich2-2.3.4-a2ixxlfjcx62bu66ytl7corlsczgydxd/spack -src/src/openpa/src -I/tmp/vagrant/spack-stage/spack-stage-mvapich 2-2.3.4-a2ixxlfjcx62bu66ytl7corlsczgydxd/spack-src/src/openpa/src -D_REENTRANT -I/tmp/vagrant/spack-stage/spack-stage-mvapich2-2.3.4 -a2ixxlfjcx62bu66ytl7corlsczgydxd/spack-src/src/mpi/romio/include -I/home/vagrant/spack/opt/spack/linux-centos7-sandybridge/gcc-8.3. 1/slurm-19-05-6-1-qawb2uwcv3evmdj3gqmrybduqfze3tou/include -DNDEBU G -DNVALGRIND -O2 -MT src/mpi/errhan/lib_libmpi_la-errutil.lo -MD -MP -MF src/mpi/errhan/.deps/lib_libmpi_la-errutil.Tpo -c src/mpi/ errhan/errutil.c -fPIC -DPIC -o src/mpi/errhan/.libs/lib_libmpi_l a-errutil.o >> 2903 src/mpi/comm/contextid.c:13:10: fatal error: helper_fns.h: No such file or directory 2904 #include "helper_fns.h" 2905 ^~~~~~~~~~~~~~ 2906 compilation terminated. ``` ### Information on your system <!-- Please include the output of `spack debug report` --> * **Spack:** 0.14.2-1390-19dba603a * **Python:** 2.7.5 * **Platform:** linux-centos7-sandybridge <!-- If you have any relevant configuration detail (custom `packages.yaml` or `modules.yaml`, etc.) you can add that here as well. --> ### Additional information <!-- Please upload the following files. They should be present in the stage directory of the failing build. Also upload any config.log or similar file if one exists. --> * [mvapich2-spack-build-env.txt](https://github.com/spack/spack/files/4764323/mvapich2-spack-build-env.txt) * [mvapich2-spack-build-out.txt](https://github.com/spack/spack/files/4764324/mvapich2-spack-build-out.txt) <!-- Some packages have maintainers who have volunteered to debug build failures. Run `spack maintainers <name-of-the-package>` and @mention them here if they exist. --> ### General information <!-- These boxes can be checked by replacing [ ] with [x] or by clicking them after submitting the issue. --> - [ ] I have run `spack debug report` and reported the version of Spack/Python/Platform - [ ] I have run `spack maintainers <name-of-the-package>` and @mentioned any maintainers - [ ] I have uploaded the build log and environment files - [ ] I have searched the issues of this repo and believe this is not a duplicate
1.0
Installation issue: mvapich2 - <!-- Thanks for taking the time to report this build failure. To proceed with the report please: 1. Title the issue "Installation issue: <name-of-the-package>". 2. Provide the information required below. We encourage you to try, as much as possible, to reduce your problem to the minimal example that still reproduces the issue. That would help us a lot in fixing it quickly and effectively! --> I'm trying to build mvapich2 in CentOS 7 with gcc 7.3.1 and configured for slurm and nemesis. The build succeeds for mvapich2@2.3.3 but fails for mvapich2@2.3.4. ### Steps to reproduce the issue <!-- Fill in the exact spec you are trying to build and the relevant part of the error message --> ```console $ spack install mvapich2@2.3.4 process_managers=slurm fabrics=nemesis ... ==> 23457: Installing mvapich2 ==> Finding buildcaches in /home/vagrant/mirror/build_cache ==> Fetching file:///home/vagrant/mirror/build_cache/linux-centos7-sandybridge-gcc-8.3.1-mvapich2-2.3.4-a2ixxlfjcx62bu66ytl7corlsczgydxd.spec.yaml curl: (37) Couldn't open file /home/vagrant/mirror/build_cache/linux-centos7-sandybridge-gcc-8.3.1-mvapich2-2.3.4-a2ixxlfjcx62bu66ytl7corlsczgydxd.spec.yaml ==> Failed to fetch file from URL: file:///home/vagrant/mirror/build_cache/linux-centos7-sandybridge-gcc-8.3.1-mvapich2-2.3.4-a2ixxlfjcx62bu66ytl7corlsczgydxd.spec.yaml Curl failed with error 37 ==> Fetching from file:///home/vagrant/mirror/build_cache/linux-centos7-sandybridge-gcc-8.3.1-mvapich2-2.3.4-a2ixxlfjcx62bu66ytl7corlsczgydxd.spec.yaml failed. ==> Using cached archive: /home/vagrant/spack/var/spack/cache/_source-cache/archive/72/7226a45c7c98333c8e5d2888119cce186199b430c13b7b1dca1769909e68ea7a.tar.gz ==> Staging archive: /tmp/vagrant/spack-stage/spack-stage-mvapich2-2.3.4-a2ixxlfjcx62bu66ytl7corlsczgydxd/mvapich2-2.3.4.tar.gz ==> Created stage in /tmp/vagrant/spack-stage/spack-stage-mvapich2-2.3.4-a2ixxlfjcx62bu66ytl7corlsczgydxd ==> No patches needed for mvapich2 ==> 23457: mvapich2: Building mvapich2 [AutotoolsPackage] ==> 23457: mvapich2: Executing phase: 'autoreconf' ==> 23457: mvapich2: Executing phase: 'configure' ==> 23457: mvapich2: Executing phase: 'build' ==> Error: ProcessError: Command exited with status 2: 'make' '-j4' 1 error found in build log: 2897 mv -f src/mpi/coll/.deps/lib_libmpi_la-nbcutil.Tpo src/mpi/coll/.d eps/lib_libmpi_la-nbcutil.Plo 2898 /bin/sh ./libtool --tag=CC --mode=compile /home/vagrant/spack/l ib/spack/env/gcc/gcc -DHAVE_CONFIG_H -I. -I./src/include -I./src/ include -I./src/mpi/datatype -I./src/mpi/spawn -I./src/util/dbg -I src/util/logging/common -I./src/util/mem -I./src/util/thread -I./s rc/util/type -I./src/util/wrappers -I./src/util/wrappers -I./src/u til/assert -I./src/util/timers -I./src/util/pointer -I./src/bindin g/cxx -I/tmp/vagrant/spack-stage/spack-stage-mvapich2-2.3.4-a2ixxl fjcx62bu66ytl7corlsczgydxd/spack-src/src/binding/fortran/mpif_h -I src/binding/fortran/use_mpi -I./src/mpid/ch3/include -I./src/mpid /ch3/include -I/tmp/vagrant/spack-stage/spack-stage-mvapich2-2.3.4 -a2ixxlfjcx62bu66ytl7corlsczgydxd/spack-src/src/mpid/ch3/channels/ common/include -I./src/mpid/ch3/channels/nemesis/include -I./src/ mpid/ch3/channels/nemesis/include -I./src/mpid/ch3/channels/nemes is/utils/monitor -I./src/mpid/common/datatype -I./src /mpid/common/datatype -I./src/mpid/common/sched -I./src/mpid/com mon/thread -I./src/mpid/common/timers -I./src/pmi/simple -I/t mp/vagrant/spack-stage/spack-stage-mvapich2-2.3.4-a2ixxlfjcx62bu66 ytl7corlsczgydxd/spack-src/src/mpl/include -I/tmp/vagrant/spack-st age/spack-stage-mvapich2-2.3.4-a2ixxlfjcx62bu66ytl7corlsczgydxd/sp ack-src/src/mpl/include -I/tmp/vagrant/spack-stage/spack-stage-mva pich2-2.3.4-a2ixxlfjcx62bu66ytl7corlsczgydxd/spack-src/src/openpa/ src -I/tmp/vagrant/spack-stage/spack-stage-mvapich2-2.3.4-a2ixxlfj cx62bu66ytl7corlsczgydxd/spack-src/src/openpa/src -D_REENTRANT -I/ tmp/vagrant/spack-stage/spack-stage-mvapich2-2.3.4-a2ixxlfjcx62bu6 6ytl7corlsczgydxd/spack-src/src/mpi/romio/include -I/home/vagrant/ spack/opt/spack/linux-centos7-sandybridge/gcc-8.3.1/slurm-19-05-6- 1-qawb2uwcv3evmdj3gqmrybduqfze3tou/include -DNDEBUG -DNVALGRIND - O2 -MT src/mpi/errhan/lib_libmpi_la-errutil.lo -MD -MP -MF src/mpi /errhan/.deps/lib_libmpi_la-errutil.Tpo -c -o src/mpi/errhan/lib_l ibmpi_la-errutil.lo `test -f 'src/mpi/errhan/errutil.c' || echo '. /'`src/mpi/errhan/errutil.c 2899 libtool: compile: /home/vagrant/spack/lib/spack/env/gcc/gcc -DHAV E_CONFIG_H -I. -I./src/include -I./src/include -I./src/mpi/datatyp e -I./src/mpi/spawn -I./src/util/dbg -Isrc/util/logging/common -I. /src/util/mem -I./src/util/thread -I./src/util/type -I./src/util/w rappers -I./src/util/wrappers -I./src/util/assert -I./src/util/tim ers -I./src/util/pointer -I./src/binding/cxx -I/tmp/vagrant/spack- stage/spack-stage-mvapich2-2.3.4-a2ixxlfjcx62bu66ytl7corlsczgydxd/ spack-src/src/binding/fortran/mpif_h -Isrc/binding/fortran/use_mpi -I./src/mpid/ch3/include -I./src/mpid/ch3/include -I/tmp/vagrant/ spack-stage/spack-stage-mvapich2-2.3.4-a2ixxlfjcx62bu66ytl7corlscz gydxd/spack-src/src/mpid/ch3/channels/common/include -I./src/mpid/ ch3/channels/nemesis/include -I./src/mpid/ch3/channels/nemesis/inc lude -I./src/mpid/ch3/channels/nemesis/utils/monitor -I./src/mpid/ common/datatype -I./src/mpid/common/datatype -I./src/mpid/common/s ched -I./src/mpid/common/thread -I./src/mpid/common/timers -I./src /pmi/simple -I/tmp/vagrant/spack-stage/spack-stage-mvapich2-2.3.4- a2ixxlfjcx62bu66ytl7corlsczgydxd/spack-src/src/mpl/include -I/tmp/ vagrant/spack-stage/spack-stage-mvapich2-2.3.4-a2ixxlfjcx62bu66ytl 7corlsczgydxd/spack-src/src/mpl/include -I/tmp/vagrant/spack-stage /spack-stage-mvapich2-2.3.4-a2ixxlfjcx62bu66ytl7corlsczgydxd/spack -src/src/openpa/src -I/tmp/vagrant/spack-stage/spack-stage-mvapich 2-2.3.4-a2ixxlfjcx62bu66ytl7corlsczgydxd/spack-src/src/openpa/src -D_REENTRANT -I/tmp/vagrant/spack-stage/spack-stage-mvapich2-2.3.4 -a2ixxlfjcx62bu66ytl7corlsczgydxd/spack-src/src/mpi/romio/include -I/home/vagrant/spack/opt/spack/linux-centos7-sandybridge/gcc-8.3. 1/slurm-19-05-6-1-qawb2uwcv3evmdj3gqmrybduqfze3tou/include -DNDEBU G -DNVALGRIND -O2 -MT src/mpi/comm/lib_libmpi_la-commutil.lo -MD - MP -MF src/mpi/comm/.deps/lib_libmpi_la-commutil.Tpo -c src/mpi/co mm/commutil.c -fPIC -DPIC -o src/mpi/comm/.libs/lib_libmpi_la-com mutil.o 2900 libtool: compile: /home/vagrant/spack/lib/spack/env/gcc/gcc -DHAV E_CONFIG_H -I. -I./src/include -I./src/include -I./src/mpi/datatyp e -I./src/mpi/spawn -I./src/util/dbg -Isrc/util/logging/common -I. /src/util/mem -I./src/util/thread -I./src/util/type -I./src/util/w rappers -I./src/util/wrappers -I./src/util/assert -I./src/util/tim ers -I./src/util/pointer -I./src/binding/cxx -I/tmp/vagrant/spack- stage/spack-stage-mvapich2-2.3.4-a2ixxlfjcx62bu66ytl7corlsczgydxd/ spack-src/src/binding/fortran/mpif_h -Isrc/binding/fortran/use_mpi -I./src/mpid/ch3/include -I./src/mpid/ch3/include -I/tmp/vagrant/ spack-stage/spack-stage-mvapich2-2.3.4-a2ixxlfjcx62bu66ytl7corlscz gydxd/spack-src/src/mpid/ch3/channels/common/include -I./src/mpid/ ch3/channels/nemesis/include -I./src/mpid/ch3/channels/nemesis/inc lude -I./src/mpid/ch3/channels/nemesis/utils/monitor -I./src/mpid/ common/datatype -I./src/mpid/common/datatype -I./src/mpid/common/s ched -I./src/mpid/common/thread -I./src/mpid/common/timers -I./src /pmi/simple -I/tmp/vagrant/spack-stage/spack-stage-mvapich2-2.3.4- a2ixxlfjcx62bu66ytl7corlsczgydxd/spack-src/src/mpl/include -I/tmp/ vagrant/spack-stage/spack-stage-mvapich2-2.3.4-a2ixxlfjcx62bu66ytl 7corlsczgydxd/spack-src/src/mpl/include -I/tmp/vagrant/spack-stage /spack-stage-mvapich2-2.3.4-a2ixxlfjcx62bu66ytl7corlsczgydxd/spack -src/src/openpa/src -I/tmp/vagrant/spack-stage/spack-stage-mvapich 2-2.3.4-a2ixxlfjcx62bu66ytl7corlsczgydxd/spack-src/src/openpa/src -D_REENTRANT -I/tmp/vagrant/spack-stage/spack-stage-mvapich2-2.3.4 -a2ixxlfjcx62bu66ytl7corlsczgydxd/spack-src/src/mpi/romio/include -I/home/vagrant/spack/opt/spack/linux-centos7-sandybridge/gcc-8.3. 1/slurm-19-05-6-1-qawb2uwcv3evmdj3gqmrybduqfze3tou/include -DNDEBU G -DNVALGRIND -O2 -MT src/mpi/comm/lib_libmpi_la-contextid.lo -MD -MP -MF src/mpi/comm/.deps/lib_libmpi_la-contextid.Tpo -c src/mpi/ comm/contextid.c -fPIC -DPIC -o src/mpi/comm/.libs/lib_libmpi_la- contextid.o 2901 libtool: compile: /home/vagrant/spack/lib/spack/env/gcc/gcc -DHAV E_CONFIG_H -I. -I./src/include -I./src/include -I./src/mpi/datatyp e -I./src/mpi/spawn -I./src/util/dbg -Isrc/util/logging/common -I. /src/util/mem -I./src/util/thread -I./src/util/type -I./src/util/w rappers -I./src/util/wrappers -I./src/util/assert -I./src/util/tim ers -I./src/util/pointer -I./src/binding/cxx -I/tmp/vagrant/spack- stage/spack-stage-mvapich2-2.3.4-a2ixxlfjcx62bu66ytl7corlsczgydxd/ spack-src/src/binding/fortran/mpif_h -Isrc/binding/fortran/use_mpi -I./src/mpid/ch3/include -I./src/mpid/ch3/include -I/tmp/vagrant/ spack-stage/spack-stage-mvapich2-2.3.4-a2ixxlfjcx62bu66ytl7corlscz gydxd/spack-src/src/mpid/ch3/channels/common/include -I./src/mpid/ ch3/channels/nemesis/include -I./src/mpid/ch3/channels/nemesis/inc lude -I./src/mpid/ch3/channels/nemesis/utils/monitor -I./src/mpid/ common/datatype -I./src/mpid/common/datatype -I./src/mpid/common/s ched -I./src/mpid/common/thread -I./src/mpid/common/timers -I./src /pmi/simple -I/tmp/vagrant/spack-stage/spack-stage-mvapich2-2.3.4- a2ixxlfjcx62bu66ytl7corlsczgydxd/spack-src/src/mpl/include -I/tmp/ vagrant/spack-stage/spack-stage-mvapich2-2.3.4-a2ixxlfjcx62bu66ytl 7corlsczgydxd/spack-src/src/mpl/include -I/tmp/vagrant/spack-stage /spack-stage-mvapich2-2.3.4-a2ixxlfjcx62bu66ytl7corlsczgydxd/spack -src/src/openpa/src -I/tmp/vagrant/spack-stage/spack-stage-mvapich 2-2.3.4-a2ixxlfjcx62bu66ytl7corlsczgydxd/spack-src/src/openpa/src -D_REENTRANT -I/tmp/vagrant/spack-stage/spack-stage-mvapich2-2.3.4 -a2ixxlfjcx62bu66ytl7corlsczgydxd/spack-src/src/mpi/romio/include -I/home/vagrant/spack/opt/spack/linux-centos7-sandybridge/gcc-8.3. 1/slurm-19-05-6-1-qawb2uwcv3evmdj3gqmrybduqfze3tou/include -DNDEBU G -DNVALGRIND -O2 -MT src/mpi/datatype/lib_libmpi_la-typeutil.lo - MD -MP -MF src/mpi/datatype/.deps/lib_libmpi_la-typeutil.Tpo -c sr c/mpi/datatype/typeutil.c -fPIC -DPIC -o src/mpi/datatype/.libs/l ib_libmpi_la-typeutil.o 2902 libtool: compile: /home/vagrant/spack/lib/spack/env/gcc/gcc -DHAV E_CONFIG_H -I. -I./src/include -I./src/include -I./src/mpi/datatyp e -I./src/mpi/spawn -I./src/util/dbg -Isrc/util/logging/common -I. /src/util/mem -I./src/util/thread -I./src/util/type -I./src/util/w rappers -I./src/util/wrappers -I./src/util/assert -I./src/util/tim ers -I./src/util/pointer -I./src/binding/cxx -I/tmp/vagrant/spack- stage/spack-stage-mvapich2-2.3.4-a2ixxlfjcx62bu66ytl7corlsczgydxd/ spack-src/src/binding/fortran/mpif_h -Isrc/binding/fortran/use_mpi -I./src/mpid/ch3/include -I./src/mpid/ch3/include -I/tmp/vagrant/ spack-stage/spack-stage-mvapich2-2.3.4-a2ixxlfjcx62bu66ytl7corlscz gydxd/spack-src/src/mpid/ch3/channels/common/include -I./src/mpid/ ch3/channels/nemesis/include -I./src/mpid/ch3/channels/nemesis/inc lude -I./src/mpid/ch3/channels/nemesis/utils/monitor -I./src/mpid/ common/datatype -I./src/mpid/common/datatype -I./src/mpid/common/s ched -I./src/mpid/common/thread -I./src/mpid/common/timers -I./src /pmi/simple -I/tmp/vagrant/spack-stage/spack-stage-mvapich2-2.3.4- a2ixxlfjcx62bu66ytl7corlsczgydxd/spack-src/src/mpl/include -I/tmp/ vagrant/spack-stage/spack-stage-mvapich2-2.3.4-a2ixxlfjcx62bu66ytl 7corlsczgydxd/spack-src/src/mpl/include -I/tmp/vagrant/spack-stage /spack-stage-mvapich2-2.3.4-a2ixxlfjcx62bu66ytl7corlsczgydxd/spack -src/src/openpa/src -I/tmp/vagrant/spack-stage/spack-stage-mvapich 2-2.3.4-a2ixxlfjcx62bu66ytl7corlsczgydxd/spack-src/src/openpa/src -D_REENTRANT -I/tmp/vagrant/spack-stage/spack-stage-mvapich2-2.3.4 -a2ixxlfjcx62bu66ytl7corlsczgydxd/spack-src/src/mpi/romio/include -I/home/vagrant/spack/opt/spack/linux-centos7-sandybridge/gcc-8.3. 1/slurm-19-05-6-1-qawb2uwcv3evmdj3gqmrybduqfze3tou/include -DNDEBU G -DNVALGRIND -O2 -MT src/mpi/errhan/lib_libmpi_la-errutil.lo -MD -MP -MF src/mpi/errhan/.deps/lib_libmpi_la-errutil.Tpo -c src/mpi/ errhan/errutil.c -fPIC -DPIC -o src/mpi/errhan/.libs/lib_libmpi_l a-errutil.o >> 2903 src/mpi/comm/contextid.c:13:10: fatal error: helper_fns.h: No such file or directory 2904 #include "helper_fns.h" 2905 ^~~~~~~~~~~~~~ 2906 compilation terminated. ``` ### Information on your system <!-- Please include the output of `spack debug report` --> * **Spack:** 0.14.2-1390-19dba603a * **Python:** 2.7.5 * **Platform:** linux-centos7-sandybridge <!-- If you have any relevant configuration detail (custom `packages.yaml` or `modules.yaml`, etc.) you can add that here as well. --> ### Additional information <!-- Please upload the following files. They should be present in the stage directory of the failing build. Also upload any config.log or similar file if one exists. --> * [mvapich2-spack-build-env.txt](https://github.com/spack/spack/files/4764323/mvapich2-spack-build-env.txt) * [mvapich2-spack-build-out.txt](https://github.com/spack/spack/files/4764324/mvapich2-spack-build-out.txt) <!-- Some packages have maintainers who have volunteered to debug build failures. Run `spack maintainers <name-of-the-package>` and @mention them here if they exist. --> ### General information <!-- These boxes can be checked by replacing [ ] with [x] or by clicking them after submitting the issue. --> - [ ] I have run `spack debug report` and reported the version of Spack/Python/Platform - [ ] I have run `spack maintainers <name-of-the-package>` and @mentioned any maintainers - [ ] I have uploaded the build log and environment files - [ ] I have searched the issues of this repo and believe this is not a duplicate
non_test
installation issue thanks for taking the time to report this build failure to proceed with the report please title the issue installation issue provide the information required below we encourage you to try as much as possible to reduce your problem to the minimal example that still reproduces the issue that would help us a lot in fixing it quickly and effectively i m trying to build in centos with gcc and configured for slurm and nemesis the build succeeds for but fails for steps to reproduce the issue console spack install process managers slurm fabrics nemesis installing finding buildcaches in home vagrant mirror build cache fetching file home vagrant mirror build cache linux sandybridge gcc spec yaml curl couldn t open file home vagrant mirror build cache linux sandybridge gcc spec yaml failed to fetch file from url file home vagrant mirror build cache linux sandybridge gcc spec yaml curl failed with error fetching from file home vagrant mirror build cache linux sandybridge gcc spec yaml failed using cached archive home vagrant spack var spack cache source cache archive tar gz staging archive tmp vagrant spack stage spack stage tar gz created stage in tmp vagrant spack stage spack stage no patches needed for building executing phase autoreconf executing phase configure executing phase build error processerror command exited with status make error found in build log mv f src mpi coll deps lib libmpi la nbcutil tpo src mpi coll d eps lib libmpi la nbcutil plo bin sh libtool tag cc mode compile home vagrant spack l ib spack env gcc gcc dhave config h i i src include i src include i src mpi datatype i src mpi spawn i src util dbg i src util logging common i src util mem i src util thread i s rc util type i src util wrappers i src util wrappers i src u til assert i src util timers i src util pointer i src bindin g cxx i tmp vagrant spack stage spack stage spack src src binding fortran mpif h i src binding fortran use mpi i src mpid include i src mpid include i tmp vagrant spack stage spack stage spack src src mpid channels common include i src mpid channels nemesis include i src mpid channels nemesis include i src mpid channels nemes is utils monitor i src mpid common datatype i src mpid common datatype i src mpid common sched i src mpid com mon thread i src mpid common timers i src pmi simple i t mp vagrant spack stage spack stage spack src src mpl include i tmp vagrant spack st age spack stage sp ack src src mpl include i tmp vagrant spack stage spack stage mva spack src src openpa src i tmp vagrant spack stage spack stage spack src src openpa src d reentrant i tmp vagrant spack stage spack stage spack src src mpi romio include i home vagrant spack opt spack linux sandybridge gcc slurm include dndebug dnvalgrind mt src mpi errhan lib libmpi la errutil lo md mp mf src mpi errhan deps lib libmpi la errutil tpo c o src mpi errhan lib l ibmpi la errutil lo test f src mpi errhan errutil c echo src mpi errhan errutil c libtool compile home vagrant spack lib spack env gcc gcc dhav e config h i i src include i src include i src mpi datatyp e i src mpi spawn i src util dbg isrc util logging common i src util mem i src util thread i src util type i src util w rappers i src util wrappers i src util assert i src util tim ers i src util pointer i src binding cxx i tmp vagrant spack stage spack stage spack src src binding fortran mpif h isrc binding fortran use mpi i src mpid include i src mpid include i tmp vagrant spack stage spack stage gydxd spack src src mpid channels common include i src mpid channels nemesis include i src mpid channels nemesis inc lude i src mpid channels nemesis utils monitor i src mpid common datatype i src mpid common datatype i src mpid common s ched i src mpid common thread i src mpid common timers i src pmi simple i tmp vagrant spack stage spack stage spack src src mpl include i tmp vagrant spack stage spack stage spack src src mpl include i tmp vagrant spack stage spack stage spack src src openpa src i tmp vagrant spack stage spack stage mvapich spack src src openpa src d reentrant i tmp vagrant spack stage spack stage spack src src mpi romio include i home vagrant spack opt spack linux sandybridge gcc slurm include dndebu g dnvalgrind mt src mpi comm lib libmpi la commutil lo md mp mf src mpi comm deps lib libmpi la commutil tpo c src mpi co mm commutil c fpic dpic o src mpi comm libs lib libmpi la com mutil o libtool compile home vagrant spack lib spack env gcc gcc dhav e config h i i src include i src include i src mpi datatyp e i src mpi spawn i src util dbg isrc util logging common i src util mem i src util thread i src util type i src util w rappers i src util wrappers i src util assert i src util tim ers i src util pointer i src binding cxx i tmp vagrant spack stage spack stage spack src src binding fortran mpif h isrc binding fortran use mpi i src mpid include i src mpid include i tmp vagrant spack stage spack stage gydxd spack src src mpid channels common include i src mpid channels nemesis include i src mpid channels nemesis inc lude i src mpid channels nemesis utils monitor i src mpid common datatype i src mpid common datatype i src mpid common s ched i src mpid common thread i src mpid common timers i src pmi simple i tmp vagrant spack stage spack stage spack src src mpl include i tmp vagrant spack stage spack stage spack src src mpl include i tmp vagrant spack stage spack stage spack src src openpa src i tmp vagrant spack stage spack stage mvapich spack src src openpa src d reentrant i tmp vagrant spack stage spack stage spack src src mpi romio include i home vagrant spack opt spack linux sandybridge gcc slurm include dndebu g dnvalgrind mt src mpi comm lib libmpi la contextid lo md mp mf src mpi comm deps lib libmpi la contextid tpo c src mpi comm contextid c fpic dpic o src mpi comm libs lib libmpi la contextid o libtool compile home vagrant spack lib spack env gcc gcc dhav e config h i i src include i src include i src mpi datatyp e i src mpi spawn i src util dbg isrc util logging common i src util mem i src util thread i src util type i src util w rappers i src util wrappers i src util assert i src util tim ers i src util pointer i src binding cxx i tmp vagrant spack stage spack stage spack src src binding fortran mpif h isrc binding fortran use mpi i src mpid include i src mpid include i tmp vagrant spack stage spack stage gydxd spack src src mpid channels common include i src mpid channels nemesis include i src mpid channels nemesis inc lude i src mpid channels nemesis utils monitor i src mpid common datatype i src mpid common datatype i src mpid common s ched i src mpid common thread i src mpid common timers i src pmi simple i tmp vagrant spack stage spack stage spack src src mpl include i tmp vagrant spack stage spack stage spack src src mpl include i tmp vagrant spack stage spack stage spack src src openpa src i tmp vagrant spack stage spack stage mvapich spack src src openpa src d reentrant i tmp vagrant spack stage spack stage spack src src mpi romio include i home vagrant spack opt spack linux sandybridge gcc slurm include dndebu g dnvalgrind mt src mpi datatype lib libmpi la typeutil lo md mp mf src mpi datatype deps lib libmpi la typeutil tpo c sr c mpi datatype typeutil c fpic dpic o src mpi datatype libs l ib libmpi la typeutil o libtool compile home vagrant spack lib spack env gcc gcc dhav e config h i i src include i src include i src mpi datatyp e i src mpi spawn i src util dbg isrc util logging common i src util mem i src util thread i src util type i src util w rappers i src util wrappers i src util assert i src util tim ers i src util pointer i src binding cxx i tmp vagrant spack stage spack stage spack src src binding fortran mpif h isrc binding fortran use mpi i src mpid include i src mpid include i tmp vagrant spack stage spack stage gydxd spack src src mpid channels common include i src mpid channels nemesis include i src mpid channels nemesis inc lude i src mpid channels nemesis utils monitor i src mpid common datatype i src mpid common datatype i src mpid common s ched i src mpid common thread i src mpid common timers i src pmi simple i tmp vagrant spack stage spack stage spack src src mpl include i tmp vagrant spack stage spack stage spack src src mpl include i tmp vagrant spack stage spack stage spack src src openpa src i tmp vagrant spack stage spack stage mvapich spack src src openpa src d reentrant i tmp vagrant spack stage spack stage spack src src mpi romio include i home vagrant spack opt spack linux sandybridge gcc slurm include dndebu g dnvalgrind mt src mpi errhan lib libmpi la errutil lo md mp mf src mpi errhan deps lib libmpi la errutil tpo c src mpi errhan errutil c fpic dpic o src mpi errhan libs lib libmpi l a errutil o src mpi comm contextid c fatal error helper fns h no such file or directory include helper fns h compilation terminated information on your system spack python platform linux sandybridge additional information and mention them here if they exist general information i have run spack debug report and reported the version of spack python platform i have run spack maintainers and mentioned any maintainers i have uploaded the build log and environment files i have searched the issues of this repo and believe this is not a duplicate
0
60,215
6,676,177,962
IssuesEvent
2017-10-05 03:33:34
statsmodels/statsmodels
https://api.github.com/repos/statsmodels/statsmodels
opened
SUMM/Topic Week: gofplots probability plots
comp-graphics type-bug type-test
several open issues and PRs it will be time consuming to work through all option combination I was looking briefly at #3547 and AFAICS we cannot turn off loc and scale fit I tried some `normal` examples and fit=False and fit=True plots look the same when I use the class, but maybe (?) not when using the qqplot helper function (which doesn't make sense base on the code) Thinking about fit cases again - no fit at all, requires fully specified distribution compare data with given distribution, no standardization, no estimation, not even loc and scale expected: this should show clear discrepancies if there is only a loc shift or loc-scale transformation e.g. straight line qqplot but not at 45 degree line - fit loc scale This is equivalent to full fit for loc-scale families, without shape parameters, e.g. normal. this seems to be the current default, or only solution if options are ignored. expected: normal data should always fit well. distributions with shape parameter should show discrepancy coming from the shape parameter, e.g. df in t-distribution - full fit only relevant in distribution with shape parameters. expectation: good fit if data is from the null distribution display axis, should be orthogonal to fit issue, but is not if we use standardized versions. ppplots are always in terms of probabilities on axis, not much choice qqplots y - axis - raw scale: don't transform data quantiles, available as prob_plots with raw scale y axis - sample standardized y-axis, this is the current `sample percentiles` (?), standardized by estimated loc and scale (to have this version might be a bug) - theoretical standardized sample quantiles: I think this is missing and should be the case when fit is False, i.e. standardize by given loc-scale (which might be estimated and then coincide with the previous case) `sample_quantiles` uses self.loc and self.scale, in contrast to sample_percentiles x - axis in qq plot is easier and should use the corresponding distribution under the null two possibilities - use loc-scale standardized distribution as xlabels - use raw distribution as xlabels Once the *content* of the plot is defined, then the axis can be transformed e.g. nonlinearly to match some probability paper which would often/mainly affect axis labeling. (Paul Hobson's newer work on axis scale in/for matplotlib) first task: create a pandas dataframe with the quantities for all options. then start to check plots for all cases and this needs unit tests (not smoke tests), otherwise I and the code will get confused again.
1.0
SUMM/Topic Week: gofplots probability plots - several open issues and PRs it will be time consuming to work through all option combination I was looking briefly at #3547 and AFAICS we cannot turn off loc and scale fit I tried some `normal` examples and fit=False and fit=True plots look the same when I use the class, but maybe (?) not when using the qqplot helper function (which doesn't make sense base on the code) Thinking about fit cases again - no fit at all, requires fully specified distribution compare data with given distribution, no standardization, no estimation, not even loc and scale expected: this should show clear discrepancies if there is only a loc shift or loc-scale transformation e.g. straight line qqplot but not at 45 degree line - fit loc scale This is equivalent to full fit for loc-scale families, without shape parameters, e.g. normal. this seems to be the current default, or only solution if options are ignored. expected: normal data should always fit well. distributions with shape parameter should show discrepancy coming from the shape parameter, e.g. df in t-distribution - full fit only relevant in distribution with shape parameters. expectation: good fit if data is from the null distribution display axis, should be orthogonal to fit issue, but is not if we use standardized versions. ppplots are always in terms of probabilities on axis, not much choice qqplots y - axis - raw scale: don't transform data quantiles, available as prob_plots with raw scale y axis - sample standardized y-axis, this is the current `sample percentiles` (?), standardized by estimated loc and scale (to have this version might be a bug) - theoretical standardized sample quantiles: I think this is missing and should be the case when fit is False, i.e. standardize by given loc-scale (which might be estimated and then coincide with the previous case) `sample_quantiles` uses self.loc and self.scale, in contrast to sample_percentiles x - axis in qq plot is easier and should use the corresponding distribution under the null two possibilities - use loc-scale standardized distribution as xlabels - use raw distribution as xlabels Once the *content* of the plot is defined, then the axis can be transformed e.g. nonlinearly to match some probability paper which would often/mainly affect axis labeling. (Paul Hobson's newer work on axis scale in/for matplotlib) first task: create a pandas dataframe with the quantities for all options. then start to check plots for all cases and this needs unit tests (not smoke tests), otherwise I and the code will get confused again.
test
summ topic week gofplots probability plots several open issues and prs it will be time consuming to work through all option combination i was looking briefly at and afaics we cannot turn off loc and scale fit i tried some normal examples and fit false and fit true plots look the same when i use the class but maybe not when using the qqplot helper function which doesn t make sense base on the code thinking about fit cases again no fit at all requires fully specified distribution compare data with given distribution no standardization no estimation not even loc and scale expected this should show clear discrepancies if there is only a loc shift or loc scale transformation e g straight line qqplot but not at degree line fit loc scale this is equivalent to full fit for loc scale families without shape parameters e g normal this seems to be the current default or only solution if options are ignored expected normal data should always fit well distributions with shape parameter should show discrepancy coming from the shape parameter e g df in t distribution full fit only relevant in distribution with shape parameters expectation good fit if data is from the null distribution display axis should be orthogonal to fit issue but is not if we use standardized versions ppplots are always in terms of probabilities on axis not much choice qqplots y axis raw scale don t transform data quantiles available as prob plots with raw scale y axis sample standardized y axis this is the current sample percentiles standardized by estimated loc and scale to have this version might be a bug theoretical standardized sample quantiles i think this is missing and should be the case when fit is false i e standardize by given loc scale which might be estimated and then coincide with the previous case sample quantiles uses self loc and self scale in contrast to sample percentiles x axis in qq plot is easier and should use the corresponding distribution under the null two possibilities use loc scale standardized distribution as xlabels use raw distribution as xlabels once the content of the plot is defined then the axis can be transformed e g nonlinearly to match some probability paper which would often mainly affect axis labeling paul hobson s newer work on axis scale in for matplotlib first task create a pandas dataframe with the quantities for all options then start to check plots for all cases and this needs unit tests not smoke tests otherwise i and the code will get confused again
1
732,903
25,279,614,060
IssuesEvent
2022-11-16 14:56:18
hack4bengal/hack4bengal.github.io
https://api.github.com/repos/hack4bengal/hack4bengal.github.io
opened
Remove background image
bug hacktoberfest 🛠 goal: fix 🟧 priority: high
### Description Make the know about us laptop image transparent ### Screenshots ![image](https://user-images.githubusercontent.com/72190187/202214591-1f43a6fe-2fda-4143-ab73-a76bb52892ad.png) ### Additional information _No response_
1.0
Remove background image - ### Description Make the know about us laptop image transparent ### Screenshots ![image](https://user-images.githubusercontent.com/72190187/202214591-1f43a6fe-2fda-4143-ab73-a76bb52892ad.png) ### Additional information _No response_
non_test
remove background image description make the know about us laptop image transparent screenshots additional information no response
0
66,316
20,150,191,582
IssuesEvent
2022-02-09 11:35:31
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
opened
The "Space home" and "Explore rooms" button do the same thing
T-Defect
### Steps to reproduce ![image](https://user-images.githubusercontent.com/57274151/153191556-1dc23c8a-6889-48ea-ad2f-3b8e64d1c691.png) ### Outcome #### What did you expect? Those two buttons should trigger two different actions #### What happened instead? Both `onClick={onExploreRoomsClick}` ### Operating system _No response_ ### Browser information _No response_ ### URL for webapp https://develop.element.io ### Application version Element version: 762fc53c6150-react-09b88d1fbe6d-js-41bf8c2d5f4a ### Homeserver _No response_ ### Will you send logs? No
1.0
The "Space home" and "Explore rooms" button do the same thing - ### Steps to reproduce ![image](https://user-images.githubusercontent.com/57274151/153191556-1dc23c8a-6889-48ea-ad2f-3b8e64d1c691.png) ### Outcome #### What did you expect? Those two buttons should trigger two different actions #### What happened instead? Both `onClick={onExploreRoomsClick}` ### Operating system _No response_ ### Browser information _No response_ ### URL for webapp https://develop.element.io ### Application version Element version: 762fc53c6150-react-09b88d1fbe6d-js-41bf8c2d5f4a ### Homeserver _No response_ ### Will you send logs? No
non_test
the space home and explore rooms button do the same thing steps to reproduce outcome what did you expect those two buttons should trigger two different actions what happened instead both onclick onexploreroomsclick operating system no response browser information no response url for webapp application version element version react js homeserver no response will you send logs no
0
2,861
28,356,661,233
IssuesEvent
2023-04-12 07:51:54
adoptium/infrastructure
https://api.github.com/repos/adoptium/infrastructure
closed
EPIC: Clean up Nagios so that we can use #infrastructure-bot
Epic reliability
Currently working on the following tasks related to this issue: Issue #2609 - PR #2638 Awaiting Review - Completed (29/09/2022 ) 1) Create ansible playbook to install Nagios Server Core on Ubuntu 22.04 via Ansible Issue [#2650](https://github.com/adoptium/infrastructure/issues/2650) - Completed ( 29/09/2022 ) 2) Tidy Up Existing Nagios Processeses, To Provide Base For Future Automation Issue [#2619](https://github.com/adoptium/infrastructure/issues/2619) - Completed ( 27/06/2022 ) 3) Upgrade O/S on Nagios Production Server To Ubuntu 22.04 Issue #2758 - Completed ( 29/09/2022 ) 4) Align Nagios & Ansible Inventory Files, Fix Hosts In Nagios But Not Ansible Issue #2759 - Completed ( 22/12/2022 ) 5) Align Nagios & Ansible Inventory Files, Fix Hosts In Ansible But Not Nagios PR https://github.com/adoptium/infrastructure/pull/2948 6) Automate Creation Of Nagios Configuration Elements From Ansible Inventory Issue [#2629](https://github.com/adoptium/infrastructure/issues/2629) - Draft PR Ready(31/01/2023) 6.1) Automate parsing of ansible inventory into Nagios inventory Issue [#2864](https://github.com/adoptium/infrastructure/issues/2864) 6.2) Automate parsing of ansible inventory to create hostgroups.cfg - Completed (10/01/2023) Issue [#2876](https://github.com/adoptium/infrastructure/issues/2876) 6.3) Automate parsing of ansible inventory to create servicegroups.cfg Issue [#2865](https://github.com/adoptium/infrastructure/issues/2865) - Completed (27/01/2023) 6.4) Automate population of machine configurations using service/types and j2 template mappings.. 6.4.1 ) Windows - Issue #2899 PR - https://github.com/adoptium/infrastructure/pull/2948 Issue [#2802](https://github.com/adoptium/infrastructure/issues/2802) - Completed (21/11/2022) 7.1) Investigate Automatic Adding Of Windows Hosts To Nagios & Document -- Need to create new template, windows version of remote tunnel etc. Issue [#2825](https://github.com/adoptium/infrastructure/issues/2825) - Completed ( 02/12/2022 ) 7.2) Update Ansible Windows Playbook To Install NSClient Required By Nagios PR: https://github.com/adoptium/infrastructure/pull/2827 Issue [#1718](https://github.com/adoptium/infrastructure/issues/1718) - Completed ( 02/12/2022 ) 7.3) Add Windows Machines To Nagios 8) Add additional monitoring for key infrastructure machines & dockerhosts ( these may need specialist individual setups and configurations creating to monitor system unique services etc ) Issue #2587 - Completed 16/12/2022 8.1) Configure dockerhosts to be monitored in Nagios Issue #2903 - Completed 02/2/2023 8.2) Configure dynamic docker-xxxx hosts to be monitored in Nagios 8.3) Expand Dockerhost Monitoring For Container/Workspace Usage - Requires New Nagios Plugin To Be Written Issue #2929 - Completed 01/2023 8.4) Expand Dockerhost Monitoring To Check For Non Running Containers - Requires New Nagios Plugin To Be Written 8.5) Configure Unique Systems List Of Unique Systems: **trss.adoptopenjdk.net - Completed** **nagios.adoptopenjdk.net - Completed** **ci.adoptopenjdk.net- Completed** 8.5.1) Configure TRSS - #2904 - Completed 02/12/2022 8.5.2) Configure Jenkins - #2905 - Completed 02/02/2023 PR https://github.com/adoptium/infrastructure/pull/2948 delivers these elements 9) Automate Installation & Configuration Of Nagios Nodes ( Requires 2 & 4 ) 9.2) Update server related global config files ( nagios.cfg ) 9.3) Create Commands.CFG file from source. ( commands.cfg ) 10) Additional Tasks Carried Out:- 10.1) Document process to update host group : https://github.com/adoptium/infrastructure/issues/2765 10.2) Document process to add additional host check : https://github.com/adoptium/infrastructure/issues/2767 11) Review and standardise host checks 11.1) Issue #2199 - Standardise & Rework The Timesync Role - Completed ( 02/2023 ) 11.2) Amend check_ntp_timesync script to work for both ntpd and chronyd services - Completed ( 02/2023 ) --- We have a slack channel for Nagios notifications (#infrastructure-bot) but it's a bit noisy, mostly due to some things like NTP not being in a suitable state. At the moment the channel is receiving hundreds of notifications a day which prevents it being useful. We need to clean up this output (fix the problems, or choose not to publish the notifications) so that we can take remedial action in a timely manner to reduce the impact to the build and test processes. This epic will have multiple individual issues underneath it to cover things that can be tackled individually. Follow on to #1229
True
EPIC: Clean up Nagios so that we can use #infrastructure-bot - Currently working on the following tasks related to this issue: Issue #2609 - PR #2638 Awaiting Review - Completed (29/09/2022 ) 1) Create ansible playbook to install Nagios Server Core on Ubuntu 22.04 via Ansible Issue [#2650](https://github.com/adoptium/infrastructure/issues/2650) - Completed ( 29/09/2022 ) 2) Tidy Up Existing Nagios Processeses, To Provide Base For Future Automation Issue [#2619](https://github.com/adoptium/infrastructure/issues/2619) - Completed ( 27/06/2022 ) 3) Upgrade O/S on Nagios Production Server To Ubuntu 22.04 Issue #2758 - Completed ( 29/09/2022 ) 4) Align Nagios & Ansible Inventory Files, Fix Hosts In Nagios But Not Ansible Issue #2759 - Completed ( 22/12/2022 ) 5) Align Nagios & Ansible Inventory Files, Fix Hosts In Ansible But Not Nagios PR https://github.com/adoptium/infrastructure/pull/2948 6) Automate Creation Of Nagios Configuration Elements From Ansible Inventory Issue [#2629](https://github.com/adoptium/infrastructure/issues/2629) - Draft PR Ready(31/01/2023) 6.1) Automate parsing of ansible inventory into Nagios inventory Issue [#2864](https://github.com/adoptium/infrastructure/issues/2864) 6.2) Automate parsing of ansible inventory to create hostgroups.cfg - Completed (10/01/2023) Issue [#2876](https://github.com/adoptium/infrastructure/issues/2876) 6.3) Automate parsing of ansible inventory to create servicegroups.cfg Issue [#2865](https://github.com/adoptium/infrastructure/issues/2865) - Completed (27/01/2023) 6.4) Automate population of machine configurations using service/types and j2 template mappings.. 6.4.1 ) Windows - Issue #2899 PR - https://github.com/adoptium/infrastructure/pull/2948 Issue [#2802](https://github.com/adoptium/infrastructure/issues/2802) - Completed (21/11/2022) 7.1) Investigate Automatic Adding Of Windows Hosts To Nagios & Document -- Need to create new template, windows version of remote tunnel etc. Issue [#2825](https://github.com/adoptium/infrastructure/issues/2825) - Completed ( 02/12/2022 ) 7.2) Update Ansible Windows Playbook To Install NSClient Required By Nagios PR: https://github.com/adoptium/infrastructure/pull/2827 Issue [#1718](https://github.com/adoptium/infrastructure/issues/1718) - Completed ( 02/12/2022 ) 7.3) Add Windows Machines To Nagios 8) Add additional monitoring for key infrastructure machines & dockerhosts ( these may need specialist individual setups and configurations creating to monitor system unique services etc ) Issue #2587 - Completed 16/12/2022 8.1) Configure dockerhosts to be monitored in Nagios Issue #2903 - Completed 02/2/2023 8.2) Configure dynamic docker-xxxx hosts to be monitored in Nagios 8.3) Expand Dockerhost Monitoring For Container/Workspace Usage - Requires New Nagios Plugin To Be Written Issue #2929 - Completed 01/2023 8.4) Expand Dockerhost Monitoring To Check For Non Running Containers - Requires New Nagios Plugin To Be Written 8.5) Configure Unique Systems List Of Unique Systems: **trss.adoptopenjdk.net - Completed** **nagios.adoptopenjdk.net - Completed** **ci.adoptopenjdk.net- Completed** 8.5.1) Configure TRSS - #2904 - Completed 02/12/2022 8.5.2) Configure Jenkins - #2905 - Completed 02/02/2023 PR https://github.com/adoptium/infrastructure/pull/2948 delivers these elements 9) Automate Installation & Configuration Of Nagios Nodes ( Requires 2 & 4 ) 9.2) Update server related global config files ( nagios.cfg ) 9.3) Create Commands.CFG file from source. ( commands.cfg ) 10) Additional Tasks Carried Out:- 10.1) Document process to update host group : https://github.com/adoptium/infrastructure/issues/2765 10.2) Document process to add additional host check : https://github.com/adoptium/infrastructure/issues/2767 11) Review and standardise host checks 11.1) Issue #2199 - Standardise & Rework The Timesync Role - Completed ( 02/2023 ) 11.2) Amend check_ntp_timesync script to work for both ntpd and chronyd services - Completed ( 02/2023 ) --- We have a slack channel for Nagios notifications (#infrastructure-bot) but it's a bit noisy, mostly due to some things like NTP not being in a suitable state. At the moment the channel is receiving hundreds of notifications a day which prevents it being useful. We need to clean up this output (fix the problems, or choose not to publish the notifications) so that we can take remedial action in a timely manner to reduce the impact to the build and test processes. This epic will have multiple individual issues underneath it to cover things that can be tackled individually. Follow on to #1229
non_test
epic clean up nagios so that we can use infrastructure bot currently working on the following tasks related to this issue issue pr awaiting review completed create ansible playbook to install nagios server core on ubuntu via ansible issue completed tidy up existing nagios processeses to provide base for future automation issue completed upgrade o s on nagios production server to ubuntu issue completed align nagios ansible inventory files fix hosts in nagios but not ansible issue completed align nagios ansible inventory files fix hosts in ansible but not nagios pr automate creation of nagios configuration elements from ansible inventory issue draft pr ready automate parsing of ansible inventory into nagios inventory issue automate parsing of ansible inventory to create hostgroups cfg completed issue automate parsing of ansible inventory to create servicegroups cfg issue completed automate population of machine configurations using service types and template mappings windows issue pr issue completed investigate automatic adding of windows hosts to nagios document need to create new template windows version of remote tunnel etc issue completed update ansible windows playbook to install nsclient required by nagios pr issue completed add windows machines to nagios add additional monitoring for key infrastructure machines dockerhosts these may need specialist individual setups and configurations creating to monitor system unique services etc issue completed configure dockerhosts to be monitored in nagios issue completed configure dynamic docker xxxx hosts to be monitored in nagios expand dockerhost monitoring for container workspace usage requires new nagios plugin to be written issue completed expand dockerhost monitoring to check for non running containers requires new nagios plugin to be written configure unique systems list of unique systems trss adoptopenjdk net completed nagios adoptopenjdk net completed ci adoptopenjdk net completed configure trss completed configure jenkins completed pr delivers these elements automate installation configuration of nagios nodes requires update server related global config files nagios cfg create commands cfg file from source commands cfg additional tasks carried out document process to update host group document process to add additional host check review and standardise host checks issue standardise rework the timesync role completed amend check ntp timesync script to work for both ntpd and chronyd services completed we have a slack channel for nagios notifications infrastructure bot but it s a bit noisy mostly due to some things like ntp not being in a suitable state at the moment the channel is receiving hundreds of notifications a day which prevents it being useful we need to clean up this output fix the problems or choose not to publish the notifications so that we can take remedial action in a timely manner to reduce the impact to the build and test processes this epic will have multiple individual issues underneath it to cover things that can be tackled individually follow on to
0
68,901
7,113,439,705
IssuesEvent
2018-01-17 20:28:42
LiskHQ/lisk
https://api.github.com/repos/LiskHQ/lisk
closed
Increase database logging information for test config
*easy database test
### Expected behaviour We have to update test config to log detailed information for database events. 1. query 2. connect 3. disconnect 4. error 5. tx 7. task ### Actual behavior Currently we only log 1 event 1. query This should enable us to have extensive logging on jenkins for debugging purpose.
1.0
Increase database logging information for test config - ### Expected behaviour We have to update test config to log detailed information for database events. 1. query 2. connect 3. disconnect 4. error 5. tx 7. task ### Actual behavior Currently we only log 1 event 1. query This should enable us to have extensive logging on jenkins for debugging purpose.
test
increase database logging information for test config expected behaviour we have to update test config to log detailed information for database events query connect disconnect error tx task actual behavior currently we only log event query this should enable us to have extensive logging on jenkins for debugging purpose
1
15,100
9,477,333,296
IssuesEvent
2019-04-19 18:14:59
aspnet/AspNetCore
https://api.github.com/repos/aspnet/AspNetCore
closed
AddJwtBearer not configured properly when using custom DI causing socket IO errors
area-security investigate
On request of @Tratcher copied from https://github.com/AzureAD/azure-activedirectory-identitymodel-extensions-for-dotnet/issues/1078 For .net core 2.1.x and 2.2.x When using the AddJwtBearer extension with custom DI such as Unity, the "options" lambda is called on every request instead of just startup.. According to @Tratcher this is not the expected behaviour. Eventually, in my web app, this causes two subsequent issues: - Socket exhaustion on HTTP Client when using the default settings for ConfigurationManager causing web site instability and performance issues - Continued invoking of the .well-known/openid-configuration endpoint incurring additional load on the host of that service _When using the inbuilt DI this problem does not occur._ The Unity team investigated this error and discovered a difference in behaviour between the resolution tree between Unity and the default DI - as identified https://github.com/AzureAD/azure-activedirectory-identitymodel-extensions-for-dotnet/issues/1078#issuecomment-457292191. They then ran the same test app with the latest compiled branch from https://github.com/aspnet/Extensions and found the error did not occur therefore they believe the issue is in the .net code itself and closed their issue. Unity team investigations are found at https://github.com/unitycontainer/microsoft-dependency-injection/issues/34 Therefore it appears that there is a difference in the way the AddJwtBearer logic handles dependencies in the 2.1 and 2.2 versions that is not in the 3.0 versions. ### To Reproduce A test app is available at https://github.com/grahambunce/unityinjectionjwtbearer configured to use Unity but it can be changed to use default DI via minor commenting out of one line. ### Expected behavior In the 2.1 (LTS) and 2.2 versions the AddJwtBearer lambda should only be called on startup instead of on each api request
True
AddJwtBearer not configured properly when using custom DI causing socket IO errors - On request of @Tratcher copied from https://github.com/AzureAD/azure-activedirectory-identitymodel-extensions-for-dotnet/issues/1078 For .net core 2.1.x and 2.2.x When using the AddJwtBearer extension with custom DI such as Unity, the "options" lambda is called on every request instead of just startup.. According to @Tratcher this is not the expected behaviour. Eventually, in my web app, this causes two subsequent issues: - Socket exhaustion on HTTP Client when using the default settings for ConfigurationManager causing web site instability and performance issues - Continued invoking of the .well-known/openid-configuration endpoint incurring additional load on the host of that service _When using the inbuilt DI this problem does not occur._ The Unity team investigated this error and discovered a difference in behaviour between the resolution tree between Unity and the default DI - as identified https://github.com/AzureAD/azure-activedirectory-identitymodel-extensions-for-dotnet/issues/1078#issuecomment-457292191. They then ran the same test app with the latest compiled branch from https://github.com/aspnet/Extensions and found the error did not occur therefore they believe the issue is in the .net code itself and closed their issue. Unity team investigations are found at https://github.com/unitycontainer/microsoft-dependency-injection/issues/34 Therefore it appears that there is a difference in the way the AddJwtBearer logic handles dependencies in the 2.1 and 2.2 versions that is not in the 3.0 versions. ### To Reproduce A test app is available at https://github.com/grahambunce/unityinjectionjwtbearer configured to use Unity but it can be changed to use default DI via minor commenting out of one line. ### Expected behavior In the 2.1 (LTS) and 2.2 versions the AddJwtBearer lambda should only be called on startup instead of on each api request
non_test
addjwtbearer not configured properly when using custom di causing socket io errors on request of tratcher copied from for net core x and x when using the addjwtbearer extension with custom di such as unity the options lambda is called on every request instead of just startup according to tratcher this is not the expected behaviour eventually in my web app this causes two subsequent issues socket exhaustion on http client when using the default settings for configurationmanager causing web site instability and performance issues continued invoking of the well known openid configuration endpoint incurring additional load on the host of that service when using the inbuilt di this problem does not occur the unity team investigated this error and discovered a difference in behaviour between the resolution tree between unity and the default di as identified they then ran the same test app with the latest compiled branch from and found the error did not occur therefore they believe the issue is in the net code itself and closed their issue unity team investigations are found at therefore it appears that there is a difference in the way the addjwtbearer logic handles dependencies in the and versions that is not in the versions to reproduce a test app is available at configured to use unity but it can be changed to use default di via minor commenting out of one line expected behavior in the lts and versions the addjwtbearer lambda should only be called on startup instead of on each api request
0
9,617
3,060,105,648
IssuesEvent
2015-08-14 18:43:16
ebassi/graphene
https://api.github.com/repos/ebassi/graphene
closed
make check fails
bug test suite
on Mac OS X Yosemite with graphene 1.2.4: ```` ERROR: quaternion - missing test plan ERROR: quaternion - exited with status 134 (terminated by signal 6?) ````
1.0
make check fails - on Mac OS X Yosemite with graphene 1.2.4: ```` ERROR: quaternion - missing test plan ERROR: quaternion - exited with status 134 (terminated by signal 6?) ````
test
make check fails on mac os x yosemite with graphene error quaternion missing test plan error quaternion exited with status terminated by signal
1
279,055
30,702,441,508
IssuesEvent
2023-07-27 01:30:29
artsking/packages_apps_settings_10.0.0_r33
https://api.github.com/repos/artsking/packages_apps_settings_10.0.0_r33
reopened
CVE-2022-20533 (Low) detected in Settingsandroid-10.0.0_r33
Mend: dependency security vulnerability
## CVE-2022-20533 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Settingsandroid-10.0.0_r33</b></p></summary> <p> <p>Library home page: <a href=https://android.googlesource.com/platform/packages/apps/Settings>https://android.googlesource.com/platform/packages/apps/Settings</a></p> <p>Found in HEAD commit: <a href="https://github.com/artsking/packages_apps_settings_10.0.0_r33/commit/081b5699d08adc751bd29d01eff86bb13c550019">081b5699d08adc751bd29d01eff86bb13c550019</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/src/com/android/settings/wifi/slice/WifiSlice.java</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> In getSlice of WifiSlice.java, there is a possible way to connect a new WiFi network from the guest mode due to a missing permission check. This could lead to local escalation of privilege with no additional execution privileges needed. User interaction is not needed for exploitation.Product: AndroidVersions: Android-13Android ID: A-232798363 <p>Publish Date: 2022-12-16 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-20533>CVE-2022-20533</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://android.googlesource.com/platform/packages/apps/Settings/+/5f421125abcdc78c73ef4af3da68ab623d2d95db">https://android.googlesource.com/platform/packages/apps/Settings/+/5f421125abcdc78c73ef4af3da68ab623d2d95db</a></p> <p>Release Date: 2022-12-16</p> <p>Fix Resolution: android-13.0.0_r16</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-20533 (Low) detected in Settingsandroid-10.0.0_r33 - ## CVE-2022-20533 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Settingsandroid-10.0.0_r33</b></p></summary> <p> <p>Library home page: <a href=https://android.googlesource.com/platform/packages/apps/Settings>https://android.googlesource.com/platform/packages/apps/Settings</a></p> <p>Found in HEAD commit: <a href="https://github.com/artsking/packages_apps_settings_10.0.0_r33/commit/081b5699d08adc751bd29d01eff86bb13c550019">081b5699d08adc751bd29d01eff86bb13c550019</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/src/com/android/settings/wifi/slice/WifiSlice.java</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> In getSlice of WifiSlice.java, there is a possible way to connect a new WiFi network from the guest mode due to a missing permission check. This could lead to local escalation of privilege with no additional execution privileges needed. User interaction is not needed for exploitation.Product: AndroidVersions: Android-13Android ID: A-232798363 <p>Publish Date: 2022-12-16 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-20533>CVE-2022-20533</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://android.googlesource.com/platform/packages/apps/Settings/+/5f421125abcdc78c73ef4af3da68ab623d2d95db">https://android.googlesource.com/platform/packages/apps/Settings/+/5f421125abcdc78c73ef4af3da68ab623d2d95db</a></p> <p>Release Date: 2022-12-16</p> <p>Fix Resolution: android-13.0.0_r16</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve low detected in settingsandroid cve low severity vulnerability vulnerable library settingsandroid library home page a href found in head commit a href found in base branch master vulnerable source files src com android settings wifi slice wifislice java vulnerability details in getslice of wifislice java there is a possible way to connect a new wifi network from the guest mode due to a missing permission check this could lead to local escalation of privilege with no additional execution privileges needed user interaction is not needed for exploitation product androidversions android id a publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution android step up your open source security game with mend
0
125,793
12,269,435,460
IssuesEvent
2020-05-07 14:04:27
clara-genomics/ClaraGenomicsAnalysis
https://api.github.com/repos/clara-genomics/ClaraGenomicsAnalysis
opened
[pycga] enable doc generation test for pyclaragenomics
documentation pyclaragenomics
Python doc generation only happens successfully when the bindings have been built. However the current sphinx command doesn't fail when bindings aren't built. This issue is to investigate the right way to build sphinx based documentation and enable it in the tests so doc generation is validated on every PR.
1.0
[pycga] enable doc generation test for pyclaragenomics - Python doc generation only happens successfully when the bindings have been built. However the current sphinx command doesn't fail when bindings aren't built. This issue is to investigate the right way to build sphinx based documentation and enable it in the tests so doc generation is validated on every PR.
non_test
enable doc generation test for pyclaragenomics python doc generation only happens successfully when the bindings have been built however the current sphinx command doesn t fail when bindings aren t built this issue is to investigate the right way to build sphinx based documentation and enable it in the tests so doc generation is validated on every pr
0
117,612
15,144,966,739
IssuesEvent
2021-02-11 02:55:48
microsoft/rushstack
https://api.github.com/repos/microsoft/rushstack
closed
[rush] make incremental build happen on file changes (watch mode)
enhancement needs design
I'd love a feature where you would type something like rush watch, and then it would invoke rush build whenever it sees that a file has been touched. This would be amazingly performant for a large repo. Behavior would be is somewhat similar to how nodemon works. Some suggestions by current lerna maintainer here https://github.com/lerna/lerna/issues/1722#issuecomment-431069978 The main motivation behind this is that we'd be able to skip having watch instances running for each package, and instead have one watch at the top. I think this would be a big step forward for scaled micro frontend monorepos. We have 30 packages now and the RAM usage of each is killing those poor developer laptops (for each package, npm is 20 mb, the process itself a minimum of 60mb). Since rush already features incremental builds, and building upstream and downstream, I think this would be the perfect thing to build onto what is in Rush. Full disclaimer, I'm not using rush yet but Lerna. This feature would tip the scale for me use rush, no doubt. Another minor thing, I'd love to be able to have this watcher actually start watch instances of webpack. I know this will block execution, so some sort of daemonization would be necessary, and likely not something that rush could offer without some complementary implementation in each micro app. I thought perhaps it would be neat if rush could help the daemonization process by either documenting it, or letting the micro apps do process.send() with some daemon activation message. This area isn't my strong suit but this is a need that we also have, and that I plan to document here later on when it's done in my project (currently using lerna but that doesn't matter in this context).
1.0
[rush] make incremental build happen on file changes (watch mode) - I'd love a feature where you would type something like rush watch, and then it would invoke rush build whenever it sees that a file has been touched. This would be amazingly performant for a large repo. Behavior would be is somewhat similar to how nodemon works. Some suggestions by current lerna maintainer here https://github.com/lerna/lerna/issues/1722#issuecomment-431069978 The main motivation behind this is that we'd be able to skip having watch instances running for each package, and instead have one watch at the top. I think this would be a big step forward for scaled micro frontend monorepos. We have 30 packages now and the RAM usage of each is killing those poor developer laptops (for each package, npm is 20 mb, the process itself a minimum of 60mb). Since rush already features incremental builds, and building upstream and downstream, I think this would be the perfect thing to build onto what is in Rush. Full disclaimer, I'm not using rush yet but Lerna. This feature would tip the scale for me use rush, no doubt. Another minor thing, I'd love to be able to have this watcher actually start watch instances of webpack. I know this will block execution, so some sort of daemonization would be necessary, and likely not something that rush could offer without some complementary implementation in each micro app. I thought perhaps it would be neat if rush could help the daemonization process by either documenting it, or letting the micro apps do process.send() with some daemon activation message. This area isn't my strong suit but this is a need that we also have, and that I plan to document here later on when it's done in my project (currently using lerna but that doesn't matter in this context).
non_test
make incremental build happen on file changes watch mode i d love a feature where you would type something like rush watch and then it would invoke rush build whenever it sees that a file has been touched this would be amazingly performant for a large repo behavior would be is somewhat similar to how nodemon works some suggestions by current lerna maintainer here the main motivation behind this is that we d be able to skip having watch instances running for each package and instead have one watch at the top i think this would be a big step forward for scaled micro frontend monorepos we have packages now and the ram usage of each is killing those poor developer laptops for each package npm is mb the process itself a minimum of since rush already features incremental builds and building upstream and downstream i think this would be the perfect thing to build onto what is in rush full disclaimer i m not using rush yet but lerna this feature would tip the scale for me use rush no doubt another minor thing i d love to be able to have this watcher actually start watch instances of webpack i know this will block execution so some sort of daemonization would be necessary and likely not something that rush could offer without some complementary implementation in each micro app i thought perhaps it would be neat if rush could help the daemonization process by either documenting it or letting the micro apps do process send with some daemon activation message this area isn t my strong suit but this is a need that we also have and that i plan to document here later on when it s done in my project currently using lerna but that doesn t matter in this context
0
220,914
7,372,090,382
IssuesEvent
2018-03-13 13:52:12
geosolutions-it/MapStore2-C040
https://api.github.com/repos/geosolutions-it/MapStore2-C040
closed
Check ZONE_STORICHE visibility
Internal Priority: Medium Project: C040 bug ready
### Description (reported) Il layer ZONE_STORICHE non compare disponibile nell’elenco del catalogo CSW sebbene sia presente invece nelle capabilities WMS. Questa problematica potrebbe coinvolgere anche altri layers. TODO: compile below: ### Please indicate if this issue is related to a bug or a new feature request - [x] Bug - [ ] New Feature ### In case of Bug (otherwise remove this paragraph) *Browser Affected* (use this site: https://www.whatsmybrowser.org/ for non expert users) - [ ] Internet Explorer - [ ] Chrome - [ ] Firefox - [ ] Safari *Browser Version Affected* - Indicate the browser version in which the issue has been found *Steps to reproduce* - A list of steps to reproduce the bug - ... *Expected Result* - Describe here the expected result *Current Result* - Describe here the current behavior ### In case of New Feature (otherwise remove this paragraph) *Acceptance Criteria - AC* - Describe the list of Acceptance Criteria - ... ### Other useful information (optional):
1.0
Check ZONE_STORICHE visibility - ### Description (reported) Il layer ZONE_STORICHE non compare disponibile nell’elenco del catalogo CSW sebbene sia presente invece nelle capabilities WMS. Questa problematica potrebbe coinvolgere anche altri layers. TODO: compile below: ### Please indicate if this issue is related to a bug or a new feature request - [x] Bug - [ ] New Feature ### In case of Bug (otherwise remove this paragraph) *Browser Affected* (use this site: https://www.whatsmybrowser.org/ for non expert users) - [ ] Internet Explorer - [ ] Chrome - [ ] Firefox - [ ] Safari *Browser Version Affected* - Indicate the browser version in which the issue has been found *Steps to reproduce* - A list of steps to reproduce the bug - ... *Expected Result* - Describe here the expected result *Current Result* - Describe here the current behavior ### In case of New Feature (otherwise remove this paragraph) *Acceptance Criteria - AC* - Describe the list of Acceptance Criteria - ... ### Other useful information (optional):
non_test
check zone storiche visibility description reported il layer zone storiche non compare disponibile nell’elenco del catalogo csw sebbene sia presente invece nelle capabilities wms questa problematica potrebbe coinvolgere anche altri layers todo compile below please indicate if this issue is related to a bug or a new feature request bug new feature in case of bug otherwise remove this paragraph browser affected use this site for non expert users internet explorer chrome firefox safari browser version affected indicate the browser version in which the issue has been found steps to reproduce a list of steps to reproduce the bug expected result describe here the expected result current result describe here the current behavior in case of new feature otherwise remove this paragraph acceptance criteria ac describe the list of acceptance criteria other useful information optional
0
782
3,226,356,945
IssuesEvent
2015-10-10 07:04:40
Yoast/wordpress-seo
https://api.github.com/repos/Yoast/wordpress-seo
closed
Domain Mapping Related Bug - Twitter/Opengraph image cards.
compatibility multisite wait for feedback
Environment: WP Multi-Site with the Domain Mapping plugin. The following meta commands are not processed correctly: meta property="og:image" content="https://subdomain.site1.com/wp-content/uploads/sites/4/2015/09/bridezilla.jpg" /> meta name="twitter:image" content="https://subdomain.site1.com/wp-content/uploads/sites/4/2015/09/bridezilla.jpg"/> The code in both class-opengraph.php and class-twitter.com call the raw image file and never processes the image name through a domain mapping routine.
True
Domain Mapping Related Bug - Twitter/Opengraph image cards. - Environment: WP Multi-Site with the Domain Mapping plugin. The following meta commands are not processed correctly: meta property="og:image" content="https://subdomain.site1.com/wp-content/uploads/sites/4/2015/09/bridezilla.jpg" /> meta name="twitter:image" content="https://subdomain.site1.com/wp-content/uploads/sites/4/2015/09/bridezilla.jpg"/> The code in both class-opengraph.php and class-twitter.com call the raw image file and never processes the image name through a domain mapping routine.
non_test
domain mapping related bug twitter opengraph image cards environment wp multi site with the domain mapping plugin the following meta commands are not processed correctly meta property og image content meta name twitter image content the code in both class opengraph php and class twitter com call the raw image file and never processes the image name through a domain mapping routine
0
54,446
7,888,237,290
IssuesEvent
2018-06-27 21:17:51
kubernetes-sigs/aws-alb-ingress-controller
https://api.github.com/repos/kubernetes-sigs/aws-alb-ingress-controller
closed
It seems the cluster it is running with Authorization enabled (like RBAC) and there is no permissions for the ingress controller. Please check the configuration
kind/documentation
I am using kubernetes 1.9 version getting below issue , the alb-ingress controller is going on crashloop It seems the cluster it is running with Authorization enabled (like RBAC) and there is no permissions for the ingress controller. Please check the configuration Can anyone let me how to resolve this issue ? Is there any permissions I need to set ?
1.0
It seems the cluster it is running with Authorization enabled (like RBAC) and there is no permissions for the ingress controller. Please check the configuration - I am using kubernetes 1.9 version getting below issue , the alb-ingress controller is going on crashloop It seems the cluster it is running with Authorization enabled (like RBAC) and there is no permissions for the ingress controller. Please check the configuration Can anyone let me how to resolve this issue ? Is there any permissions I need to set ?
non_test
it seems the cluster it is running with authorization enabled like rbac and there is no permissions for the ingress controller please check the configuration i am using kubernetes version getting below issue the alb ingress controller is going on crashloop it seems the cluster it is running with authorization enabled like rbac and there is no permissions for the ingress controller please check the configuration can anyone let me how to resolve this issue is there any permissions i need to set
0
294,273
22,144,357,167
IssuesEvent
2022-06-03 10:18:26
taurr/bevy-atlas-loader
https://api.github.com/repos/taurr/bevy-atlas-loader
closed
Enhance the documentation
documentation
* [ ] create link to the examples * [ ] touchup doc for all public structs * [ ] touchup dox for all public functions * [ ] include small examles? * [ ] add grafical links to crates.io info etc. (docs, GH, build status...)
1.0
Enhance the documentation - * [ ] create link to the examples * [ ] touchup doc for all public structs * [ ] touchup dox for all public functions * [ ] include small examles? * [ ] add grafical links to crates.io info etc. (docs, GH, build status...)
non_test
enhance the documentation create link to the examples touchup doc for all public structs touchup dox for all public functions include small examles add grafical links to crates io info etc docs gh build status
0
155,721
12,268,270,174
IssuesEvent
2020-05-07 12:14:11
inf112-v20/factoryRobots
https://api.github.com/repos/inf112-v20/factoryRobots
closed
Create more tests
help wanted test
Missing test classes are from what I can see: - [x] CardSlotTest - Can't make automated test because "Gdx test runner" cannot create a sprite batch. Ref. https://github.com/TomGrill/gdx-testing/issues/5 - [x] Rounds - Same with this class - [x] Sounds - [x] Robot - [x] Network tests if possible - Same with the class More classes are missing their respective test classes, but I haven't checked if we can create tests for them or if we should create manual tests. We should at least create manuals tests for the main menu
1.0
Create more tests - Missing test classes are from what I can see: - [x] CardSlotTest - Can't make automated test because "Gdx test runner" cannot create a sprite batch. Ref. https://github.com/TomGrill/gdx-testing/issues/5 - [x] Rounds - Same with this class - [x] Sounds - [x] Robot - [x] Network tests if possible - Same with the class More classes are missing their respective test classes, but I haven't checked if we can create tests for them or if we should create manual tests. We should at least create manuals tests for the main menu
test
create more tests missing test classes are from what i can see cardslottest can t make automated test because gdx test runner cannot create a sprite batch ref rounds same with this class sounds robot network tests if possible same with the class more classes are missing their respective test classes but i haven t checked if we can create tests for them or if we should create manual tests we should at least create manuals tests for the main menu
1
195,358
14,725,670,478
IssuesEvent
2021-01-06 05:17:29
pingcap/tidb
https://api.github.com/repos/pingcap/tidb
closed
TestLoadData fail in tidb_test
component/test severity/moderate status/can't-reproduce type/bug
## Bug Report Please answer these questions before submitting your issue. Thanks! 1. What did you do? If possible, provide a recipe for reproducing the error. ``` [2019-11-14T13:23:34.349Z] FAIL: tidb_test.go:115: TidbTestSuite.TestLoadData [2019-11-14T13:23:34.349Z] [2019-11-14T13:23:34.349Z] tidb_test.go:117: [2019-11-14T13:23:34.349Z] runTestLoadData(c, suite.server) [2019-11-14T13:23:34.349Z] server_test.go:734: [2019-11-14T13:23:34.349Z] dbt.Assert(ok, IsTrue) [2019-11-14T13:23:34.349Z] ... obtained bool = false [2019-11-14T13:23:34.349Z] ``` 2. What version of TiDB are you using (`tidb-server -V` or run `select tidb_version();` on TiDB)? https://internal.pingcap.net/idc-jenkins/blue/organizations/jenkins/tidb_ghpr_unit_test/detail/tidb_ghpr_unit_test/15676/pipeline/
1.0
TestLoadData fail in tidb_test - ## Bug Report Please answer these questions before submitting your issue. Thanks! 1. What did you do? If possible, provide a recipe for reproducing the error. ``` [2019-11-14T13:23:34.349Z] FAIL: tidb_test.go:115: TidbTestSuite.TestLoadData [2019-11-14T13:23:34.349Z] [2019-11-14T13:23:34.349Z] tidb_test.go:117: [2019-11-14T13:23:34.349Z] runTestLoadData(c, suite.server) [2019-11-14T13:23:34.349Z] server_test.go:734: [2019-11-14T13:23:34.349Z] dbt.Assert(ok, IsTrue) [2019-11-14T13:23:34.349Z] ... obtained bool = false [2019-11-14T13:23:34.349Z] ``` 2. What version of TiDB are you using (`tidb-server -V` or run `select tidb_version();` on TiDB)? https://internal.pingcap.net/idc-jenkins/blue/organizations/jenkins/tidb_ghpr_unit_test/detail/tidb_ghpr_unit_test/15676/pipeline/
test
testloaddata fail in tidb test bug report please answer these questions before submitting your issue thanks what did you do if possible provide a recipe for reproducing the error fail tidb test go tidbtestsuite testloaddata tidb test go runtestloaddata c suite server server test go dbt assert ok istrue obtained bool false what version of tidb are you using tidb server v or run select tidb version on tidb
1
41,077
5,331,904,366
IssuesEvent
2017-02-15 20:41:29
khartec/waltz
https://api.github.com/repos/khartec/waltz
closed
Surveys: template list page
enhancement fixed (test & close)
Grouped by target entity kind, links to show more details , show run configs and create new run configs.
1.0
Surveys: template list page - Grouped by target entity kind, links to show more details , show run configs and create new run configs.
test
surveys template list page grouped by target entity kind links to show more details show run configs and create new run configs
1
227,107
18,050,355,732
IssuesEvent
2021-09-19 16:33:38
ultimate-research/ssbh_lib
https://api.github.com/repos/ultimate-research/ssbh_lib
opened
improve error types
testing ssbh_data
Error types can be cleaned up with thiserror. Use the following code snippet instead of simply asserting for panics. ```rust assert!(matches!( result, Err(BoneTransformError::CycleDetected { index: 0 }) )); ```
1.0
improve error types - Error types can be cleaned up with thiserror. Use the following code snippet instead of simply asserting for panics. ```rust assert!(matches!( result, Err(BoneTransformError::CycleDetected { index: 0 }) )); ```
test
improve error types error types can be cleaned up with thiserror use the following code snippet instead of simply asserting for panics rust assert matches result err bonetransformerror cycledetected index
1
287,438
24,829,741,271
IssuesEvent
2022-10-26 01:33:12
LiamMahoney/ghrum
https://api.github.com/repos/LiamMahoney/ghrum
closed
Issue Unlabeled
stage: testing / review project: refactor 1
When an issue is unlabeled, the type of label needs to be determined so the proper action(s) can be taken. When a project label is removed from an issue, the project card in the project whose label was just removed should be deleted so that the project stays synced. --- **hook:** issue **action:** unlabeled
1.0
Issue Unlabeled - When an issue is unlabeled, the type of label needs to be determined so the proper action(s) can be taken. When a project label is removed from an issue, the project card in the project whose label was just removed should be deleted so that the project stays synced. --- **hook:** issue **action:** unlabeled
test
issue unlabeled when an issue is unlabeled the type of label needs to be determined so the proper action s can be taken when a project label is removed from an issue the project card in the project whose label was just removed should be deleted so that the project stays synced hook issue action unlabeled
1
155,132
13,612,158,042
IssuesEvent
2020-09-23 09:52:56
JaAlSaDev/Dream-Field
https://api.github.com/repos/JaAlSaDev/Dream-Field
closed
1- Title and Introduction
documentation
1. - [x] **Title**: Include the name of the project, and place the GA logo to left of the title. 1. - [x] **Introduction**: In this section, include the following: - [x] A brief definition/description of the game. - [x] Why you developed it. - [x] A link to a video showcasing the whole game. - [x] A GIF showcasing the battle screen. - [x] A link to the game.
1.0
1- Title and Introduction - 1. - [x] **Title**: Include the name of the project, and place the GA logo to left of the title. 1. - [x] **Introduction**: In this section, include the following: - [x] A brief definition/description of the game. - [x] Why you developed it. - [x] A link to a video showcasing the whole game. - [x] A GIF showcasing the battle screen. - [x] A link to the game.
non_test
title and introduction title include the name of the project and place the ga logo to left of the title introduction in this section include the following a brief definition description of the game why you developed it a link to a video showcasing the whole game a gif showcasing the battle screen a link to the game
0
62,753
12,238,965,059
IssuesEvent
2020-05-04 20:46:11
pnp/pnpjs
https://api.github.com/repos/pnp/pnpjs
closed
Copy bannerImageUrl from another file
area: code status: complete type: enhancement
### Category - [x] Enhancement - [ ] Bug - [ ] Question - [ ] Documentation gap/issue ### Expected / Desired Behavior / Question It would be nice to copy bannerImageUrl from another file (i.e. a template), currently when setting bannerImageUrl it uses getFileByServerRelativePath to get the image but it doesn't seems to work that well with _layouts/15/getpreview.ashx from the bannerImage url of the other file.
1.0
Copy bannerImageUrl from another file - ### Category - [x] Enhancement - [ ] Bug - [ ] Question - [ ] Documentation gap/issue ### Expected / Desired Behavior / Question It would be nice to copy bannerImageUrl from another file (i.e. a template), currently when setting bannerImageUrl it uses getFileByServerRelativePath to get the image but it doesn't seems to work that well with _layouts/15/getpreview.ashx from the bannerImage url of the other file.
non_test
copy bannerimageurl from another file category enhancement bug question documentation gap issue expected desired behavior question it would be nice to copy bannerimageurl from another file i e a template currently when setting bannerimageurl it uses getfilebyserverrelativepath to get the image but it doesn t seems to work that well with layouts getpreview ashx from the bannerimage url of the other file
0
5,576
5,795,297,738
IssuesEvent
2017-05-02 16:45:15
Lemoncode/LeanMood
https://api.github.com/repos/Lemoncode/LeanMood
closed
Migrate to awesome-typescript loader + babel
infrastructure
typescript-loader seems to be out of date, awesome-typescript-loader seems the way to go (also it has a good bab el integration).
1.0
Migrate to awesome-typescript loader + babel - typescript-loader seems to be out of date, awesome-typescript-loader seems the way to go (also it has a good bab el integration).
non_test
migrate to awesome typescript loader babel typescript loader seems to be out of date awesome typescript loader seems the way to go also it has a good bab el integration
0
99,474
8,700,983,814
IssuesEvent
2018-12-05 10:15:55
researchstudio-sat/webofneeds
https://api.github.com/repos/researchstudio-sat/webofneeds
closed
Implement Resend Verification Mail as State Reducer
GDPR prio: low refactoring opportunity suggestion testing
Currently we only call the resend verification mail ws-method from `login-form.js` and `slide-in.js` We might want to create a reducer that stores the initial invocation and the response of the ws-call in the state, right now we are not parsing the ws-response at all, we simply check if the button has been clicked and reset this after i think 2seconds again
1.0
Implement Resend Verification Mail as State Reducer - Currently we only call the resend verification mail ws-method from `login-form.js` and `slide-in.js` We might want to create a reducer that stores the initial invocation and the response of the ws-call in the state, right now we are not parsing the ws-response at all, we simply check if the button has been clicked and reset this after i think 2seconds again
test
implement resend verification mail as state reducer currently we only call the resend verification mail ws method from login form js and slide in js we might want to create a reducer that stores the initial invocation and the response of the ws call in the state right now we are not parsing the ws response at all we simply check if the button has been clicked and reset this after i think again
1
232,483
18,883,184,771
IssuesEvent
2021-11-15 02:43:41
milvus-io/milvus
https://api.github.com/repos/milvus-io/milvus
closed
[Bug]: [benchmark][cluster] search error: No Available Query node result
kind/bug priority/critical-urgent test/accuracy test/benchmark performance tuning
### Is there an existing issue for this? - [X] I have searched the existing issues ### Environment ```markdown - Milvus version: master-20211106-436d530 - Deployment mode(standalone or cluster): cluster - SDK version(e.g. pymilvus v2.0.0rc2): pymilvus-2.0.0rc8.dev5 - OS(Ubuntu or CentOS): - CPU/Memory: - GPU: - Others: ``` ### Current Behavior client log: ``` [2021-11-06 16:08:51,308] [ DEBUG] - Milvus create_index run in 109.44s (milvus_benchmark.client:52) [2021-11-06 16:08:51,312] [ INFO] - None (milvus_benchmark.client:290) [2021-11-06 16:08:51,312] [ INFO] - {'index_type': 'flat', 'metric_type': None, 'index_param': None} (milvus_benchmark.runners.accuracy:251) [2021-11-06 16:08:51,313] [ INFO] - Start load collection: glove_200_angular (milvus_benchmark.runners.accuracy:252) [2021-11-06 16:09:10,049] [ DEBUG] - Milvus load_collection run in 18.74s (milvus_benchmark.client:52) [2021-11-06 16:09:10,050] [ INFO] - End load collection: glove_200_angular (milvus_benchmark.runners.accuracy:255) [2021-11-06 16:09:10,288] [ INFO] - Start run case (milvus_benchmark.main:87) [2021-11-06 16:09:12,531] [ ERROR] - Error: <BaseException: (code=1, message=No Available Query node result, filter reason Search failed, C runtime error detected, error code = 1, error msg = Error in milvus::knowhere::ConfAd : id 428923895712975508)> (pymilvus.client.grpc_handler:59) [2021-11-06 16:09:12,532] [ ERROR] - Error: <BaseException: (code=1, message=No Available Query node result, filter reason Search failed, C runtime error detected, error code = 1, error msg = Error in milvus::knowhere::ConfAd : id 428923895712975508)> (pymilvus.client.grpc_handler:59) [2021-11-06 16:09:12,536] [ ERROR] - Traceback (most recent call last): File "main.py", line 94, in run_suite result = runner.run_case(case_metric, **case) File "/src/milvus_benchmark/runners/accuracy.py", line 265, in run_case self.milvus.query(case_param["vector_query"], filter_query=case_param["filter_query"]) File "/src/milvus_benchmark/client.py", line 48, in wrapper result = func(*args, **kwargs) File "/src/milvus_benchmark/client.py", line 317, in query result = self._milvus.search(tmp_collection_name, **params) File "/usr/local/lib/python3.6/site-packages/pymilvus/client/stub.py", line 61, in handler raise e File "/usr/local/lib/python3.6/site-packages/pymilvus/client/stub.py", line 45, in handler return func(self, *args, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/client/stub.py", line 1071, in search partition_names, output_fields, timeout, round_decimal, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 65, in handler raise e File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 57, in handler return func(self, *args, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 94, in handler return func(self, *args, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 620, in search return self._execute_search_requests(requests, timeout, **_kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 65, in handler raise e File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 57, in handler return func(self, *args, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 587, in _execute_search_requests raise pre_err File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 578, in _execute_search_requests raise BaseException(response.status.error_code, response.status.reason) pymilvus.client.exceptions.BaseException: <BaseException: (code=1, message=No Available Query node result, filter reason Search failed, C runtime error detected, error code = 1, error msg = Error in milvus::knowhere::ConfAd : id 428923895712975508)> (milvus_benchmark.main:97) [2021-11-06 16:09:12,537] [ INFO] - None (milvus_benchmark.main:98) [2021-11-06 16:09:12,537] [ DEBUG] - {'type': 'ann_accuracy', 'value': {}} (milvus_benchmark.main:106) [2021-11-06 16:09:12,538] [ DEBUG] - {'_version': '0.1', '_type': 'case', 'run_id': 1636214697, 'mode': 'local', 'server': <milvus_benchmark.metrics.models.server.Server object at 0x7f87a86115c0>, 'hardware': <milvus_benchmark.metrics.models.hardware.Hardware object at 0x7f876519cf98>, 'env': <milvus_benchmark.metrics.models.env.Env object at 0x7f876519ceb8>, 'status': 'RUN_FAILED', 'err_message': '<BaseException: (code=1, message=No Available Query node result, filter reason Search failed, C runtime error detected, error code = 1, error msg = Error in milvus::knowhere::ConfAd\n: id 428923895712975508)>\nTraceback (most recent call last):\n File "main.py", line 94, in run_suite\n result = runner.run_case(case_metric, **case)\n File "/src/milvus_benchmark/runners/accuracy.py", line 265, in run_case\n self.milvus.query(case_param["vector_query"], filter_query=case_param["filter_query"])\n File "/src/milvus_benchmark/client.py", line 48, in wrapper\n result = func(*args, **kwargs)\n File "/src/milvus_benchmark/client.py", line 317, in query\n result = self._milvus.search(tmp_collection_name, **params)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/stub.py", line 61, in handler\n raise e\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/stub.py", line 45, in handler\n return func(self, *args, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/stub.py", line 1071, in search\n partition_names, output_fields, timeout, round_decimal, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 65, in handler\n raise e\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 57, in handler\n return func(self, *args, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 94, in handler\n return func(self, *args, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 620, in search\n return self._execute_search_requests(requests, timeout, **_kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 65, in handler\n raise e\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 57, in handler\n return func(self, *args, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 587, in _execute_search_requests\n raise pre_err\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 578, in _execute_search_requests\n raise BaseException(response.status.error_code, response.status.reason)\npymilvus.client.exceptions.BaseException: <BaseException: (code=1, message=No Available Query node result, filter reason Search failed, C runtime error detected, error code = 1, error msg = Error in milvus::knowhere::ConfAd\n: id 428923895712975508)>\n', 'collection': {'dimension': 200, 'metric_type': 'ip', 'dataset_name': 'glove_200_angular', 'shards_num': None}, 'index': {'index_type': 'ivf_flat', 'index_param': {'nlist': 1024}}, 'search': {'nq': 10000, 'topk': 10, 'search_param': {'nprobe': 1}, 'filter': []}, 'run_params': None, 'metrics': {'type': 'ann_accuracy', 'value': {}}, 'datetime': '2021-11-06 16:04:57.516798', 'type': 'metric'} (milvus_benchmark.metric.api:29) ``` server: ``` NAME READY STATUS RESTARTS AGE IP NODE NOMINATED NODE READINESS GATES benchmark-cron-1636214400-5-etcd-0 1/1 Running 0 3m 10.97.17.154 qa-node014.zilliz.local <none> <none> benchmark-cron-1636214400-5-etcd-1 1/1 Running 0 3m 10.97.17.152 qa-node014.zilliz.local <none> <none> benchmark-cron-1636214400-5-etcd-2 1/1 Running 0 3m 10.97.17.157 qa-node014.zilliz.local <none> <none> benchmark-cron-1636214400-5-milvus-datacoord-5f69b9b594-5h6rs 1/1 Running 0 3m 10.97.8.4 qa-node006.zilliz.local <none> <none> benchmark-cron-1636214400-5-milvus-datanode-78b47db7bf-r49ml 1/1 Running 0 2m59s 10.97.11.11 qa-node009.zilliz.local <none> <none> benchmark-cron-1636214400-5-milvus-indexcoord-675795cf47-j7txs 1/1 Running 0 3m 10.97.10.181 qa-node008.zilliz.local <none> <none> benchmark-cron-1636214400-5-milvus-indexnode-68fd6b4b64-k4dws 1/1 Running 0 2m59s 10.97.17.139 qa-node014.zilliz.local <none> <none> benchmark-cron-1636214400-5-milvus-proxy-57876fc664-s94fg 1/1 Running 0 3m 10.97.10.176 qa-node008.zilliz.local <none> <none> benchmark-cron-1636214400-5-milvus-querycoord-8449d747b-b4jjk 1/1 Running 0 3m 10.97.10.175 qa-node008.zilliz.local <none> <none> benchmark-cron-1636214400-5-milvus-querynode-748c5946b-r9qk6 1/1 Running 0 3m 10.97.10.183 qa-node008.zilliz.local <none> <none> benchmark-cron-1636214400-5-milvus-rootcoord-78bfbd8f4b-dmh8v 1/1 Running 0 3m 10.97.8.2 qa-node006.zilliz.local <none> <none> benchmark-cron-1636214400-5-minio-0 1/1 Running 0 3m 10.97.10.174 qa-node008.zilliz.local <none> <none> benchmark-cron-1636214400-5-minio-1 1/1 Running 0 3m 10.97.10.177 qa-node008.zilliz.local <none> <none> benchmark-cron-1636214400-5-minio-2 1/1 Running 0 3m 10.97.10.178 qa-node008.zilliz.local <none> <none> benchmark-cron-1636214400-5-minio-3 1/1 Running 0 3m 10.97.10.179 qa-node008.zilliz.local <none> <none> benchmark-cron-1636214400-5-pulsar-autorecovery-7fbbd955c4vkc9z 1/1 Running 0 3m 10.97.8.5 qa-node006.zilliz.local <none> <none> benchmark-cron-1636214400-5-pulsar-bastion-5567946888-xr94b 1/1 Running 0 3m 10.97.8.3 qa-node006.zilliz.local <none> <none> benchmark-cron-1636214400-5-pulsar-bookkeeper-0 1/1 Running 0 3m 10.97.8.8 qa-node006.zilliz.local <none> <none> benchmark-cron-1636214400-5-pulsar-bookkeeper-1 1/1 Running 0 57s 10.97.6.179 qa-node004.zilliz.local <none> <none> benchmark-cron-1636214400-5-pulsar-broker-6d5ccd5df4-q5ch7 1/1 Running 0 2m59s 10.97.15.52 qa-node012.zilliz.local <none> <none> benchmark-cron-1636214400-5-pulsar-proxy-84dbfcc96-2n7w5 2/2 Running 0 2m59s 10.97.13.62 qa-node010.zilliz.local <none> <none> benchmark-cron-1636214400-5-pulsar-zookeeper-0 1/1 Running 0 3m 10.97.9.254 qa-node007.zilliz.local <none> <none> benchmark-cron-1636214400-5-pulsar-zookeeper-1 1/1 Running 0 2m42s 10.97.6.177 qa-node004.zilliz.local <none> <none> benchmark-cron-1636214400-5-pulsar-zookeeper-2 1/1 Running 0 2m28s 10.97.6.178 qa-node004.zilliz.local <none> <none> benchmark-cron-1636214400-5-pulsar-zookeeper-metadata-zzjf5 0/1 Completed 0 3m 10.97.10.180 qa-node008.zilliz.local <none> <none> ``` ### Expected Behavior _No response_ ### Steps To Reproduce ```markdown argo task:benchmark-cron-1636214400 test yaml: client-configmap:client-acc-glove-ivf-flat server-configmap:server-cluster-8c16m ``` ### Anything else? client-acc-glove-ivf-flat: ``` ann_accuracy: collections: - source_file: /test/milvus/ann_hdf5/glove-200-angular.hdf5 collection_name: glove_200_angular index_types: ['ivf_flat'] index_params: nlist: [1024] top_ks: [10] nqs: [10000] search_params: nprobe: [1, 2, 4, 8, 16, 32, 64, 128, 256, 512] ```
2.0
[Bug]: [benchmark][cluster] search error: No Available Query node result - ### Is there an existing issue for this? - [X] I have searched the existing issues ### Environment ```markdown - Milvus version: master-20211106-436d530 - Deployment mode(standalone or cluster): cluster - SDK version(e.g. pymilvus v2.0.0rc2): pymilvus-2.0.0rc8.dev5 - OS(Ubuntu or CentOS): - CPU/Memory: - GPU: - Others: ``` ### Current Behavior client log: ``` [2021-11-06 16:08:51,308] [ DEBUG] - Milvus create_index run in 109.44s (milvus_benchmark.client:52) [2021-11-06 16:08:51,312] [ INFO] - None (milvus_benchmark.client:290) [2021-11-06 16:08:51,312] [ INFO] - {'index_type': 'flat', 'metric_type': None, 'index_param': None} (milvus_benchmark.runners.accuracy:251) [2021-11-06 16:08:51,313] [ INFO] - Start load collection: glove_200_angular (milvus_benchmark.runners.accuracy:252) [2021-11-06 16:09:10,049] [ DEBUG] - Milvus load_collection run in 18.74s (milvus_benchmark.client:52) [2021-11-06 16:09:10,050] [ INFO] - End load collection: glove_200_angular (milvus_benchmark.runners.accuracy:255) [2021-11-06 16:09:10,288] [ INFO] - Start run case (milvus_benchmark.main:87) [2021-11-06 16:09:12,531] [ ERROR] - Error: <BaseException: (code=1, message=No Available Query node result, filter reason Search failed, C runtime error detected, error code = 1, error msg = Error in milvus::knowhere::ConfAd : id 428923895712975508)> (pymilvus.client.grpc_handler:59) [2021-11-06 16:09:12,532] [ ERROR] - Error: <BaseException: (code=1, message=No Available Query node result, filter reason Search failed, C runtime error detected, error code = 1, error msg = Error in milvus::knowhere::ConfAd : id 428923895712975508)> (pymilvus.client.grpc_handler:59) [2021-11-06 16:09:12,536] [ ERROR] - Traceback (most recent call last): File "main.py", line 94, in run_suite result = runner.run_case(case_metric, **case) File "/src/milvus_benchmark/runners/accuracy.py", line 265, in run_case self.milvus.query(case_param["vector_query"], filter_query=case_param["filter_query"]) File "/src/milvus_benchmark/client.py", line 48, in wrapper result = func(*args, **kwargs) File "/src/milvus_benchmark/client.py", line 317, in query result = self._milvus.search(tmp_collection_name, **params) File "/usr/local/lib/python3.6/site-packages/pymilvus/client/stub.py", line 61, in handler raise e File "/usr/local/lib/python3.6/site-packages/pymilvus/client/stub.py", line 45, in handler return func(self, *args, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/client/stub.py", line 1071, in search partition_names, output_fields, timeout, round_decimal, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 65, in handler raise e File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 57, in handler return func(self, *args, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 94, in handler return func(self, *args, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 620, in search return self._execute_search_requests(requests, timeout, **_kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 65, in handler raise e File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 57, in handler return func(self, *args, **kwargs) File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 587, in _execute_search_requests raise pre_err File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 578, in _execute_search_requests raise BaseException(response.status.error_code, response.status.reason) pymilvus.client.exceptions.BaseException: <BaseException: (code=1, message=No Available Query node result, filter reason Search failed, C runtime error detected, error code = 1, error msg = Error in milvus::knowhere::ConfAd : id 428923895712975508)> (milvus_benchmark.main:97) [2021-11-06 16:09:12,537] [ INFO] - None (milvus_benchmark.main:98) [2021-11-06 16:09:12,537] [ DEBUG] - {'type': 'ann_accuracy', 'value': {}} (milvus_benchmark.main:106) [2021-11-06 16:09:12,538] [ DEBUG] - {'_version': '0.1', '_type': 'case', 'run_id': 1636214697, 'mode': 'local', 'server': <milvus_benchmark.metrics.models.server.Server object at 0x7f87a86115c0>, 'hardware': <milvus_benchmark.metrics.models.hardware.Hardware object at 0x7f876519cf98>, 'env': <milvus_benchmark.metrics.models.env.Env object at 0x7f876519ceb8>, 'status': 'RUN_FAILED', 'err_message': '<BaseException: (code=1, message=No Available Query node result, filter reason Search failed, C runtime error detected, error code = 1, error msg = Error in milvus::knowhere::ConfAd\n: id 428923895712975508)>\nTraceback (most recent call last):\n File "main.py", line 94, in run_suite\n result = runner.run_case(case_metric, **case)\n File "/src/milvus_benchmark/runners/accuracy.py", line 265, in run_case\n self.milvus.query(case_param["vector_query"], filter_query=case_param["filter_query"])\n File "/src/milvus_benchmark/client.py", line 48, in wrapper\n result = func(*args, **kwargs)\n File "/src/milvus_benchmark/client.py", line 317, in query\n result = self._milvus.search(tmp_collection_name, **params)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/stub.py", line 61, in handler\n raise e\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/stub.py", line 45, in handler\n return func(self, *args, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/stub.py", line 1071, in search\n partition_names, output_fields, timeout, round_decimal, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 65, in handler\n raise e\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 57, in handler\n return func(self, *args, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 94, in handler\n return func(self, *args, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 620, in search\n return self._execute_search_requests(requests, timeout, **_kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 65, in handler\n raise e\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 57, in handler\n return func(self, *args, **kwargs)\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 587, in _execute_search_requests\n raise pre_err\n File "/usr/local/lib/python3.6/site-packages/pymilvus/client/grpc_handler.py", line 578, in _execute_search_requests\n raise BaseException(response.status.error_code, response.status.reason)\npymilvus.client.exceptions.BaseException: <BaseException: (code=1, message=No Available Query node result, filter reason Search failed, C runtime error detected, error code = 1, error msg = Error in milvus::knowhere::ConfAd\n: id 428923895712975508)>\n', 'collection': {'dimension': 200, 'metric_type': 'ip', 'dataset_name': 'glove_200_angular', 'shards_num': None}, 'index': {'index_type': 'ivf_flat', 'index_param': {'nlist': 1024}}, 'search': {'nq': 10000, 'topk': 10, 'search_param': {'nprobe': 1}, 'filter': []}, 'run_params': None, 'metrics': {'type': 'ann_accuracy', 'value': {}}, 'datetime': '2021-11-06 16:04:57.516798', 'type': 'metric'} (milvus_benchmark.metric.api:29) ``` server: ``` NAME READY STATUS RESTARTS AGE IP NODE NOMINATED NODE READINESS GATES benchmark-cron-1636214400-5-etcd-0 1/1 Running 0 3m 10.97.17.154 qa-node014.zilliz.local <none> <none> benchmark-cron-1636214400-5-etcd-1 1/1 Running 0 3m 10.97.17.152 qa-node014.zilliz.local <none> <none> benchmark-cron-1636214400-5-etcd-2 1/1 Running 0 3m 10.97.17.157 qa-node014.zilliz.local <none> <none> benchmark-cron-1636214400-5-milvus-datacoord-5f69b9b594-5h6rs 1/1 Running 0 3m 10.97.8.4 qa-node006.zilliz.local <none> <none> benchmark-cron-1636214400-5-milvus-datanode-78b47db7bf-r49ml 1/1 Running 0 2m59s 10.97.11.11 qa-node009.zilliz.local <none> <none> benchmark-cron-1636214400-5-milvus-indexcoord-675795cf47-j7txs 1/1 Running 0 3m 10.97.10.181 qa-node008.zilliz.local <none> <none> benchmark-cron-1636214400-5-milvus-indexnode-68fd6b4b64-k4dws 1/1 Running 0 2m59s 10.97.17.139 qa-node014.zilliz.local <none> <none> benchmark-cron-1636214400-5-milvus-proxy-57876fc664-s94fg 1/1 Running 0 3m 10.97.10.176 qa-node008.zilliz.local <none> <none> benchmark-cron-1636214400-5-milvus-querycoord-8449d747b-b4jjk 1/1 Running 0 3m 10.97.10.175 qa-node008.zilliz.local <none> <none> benchmark-cron-1636214400-5-milvus-querynode-748c5946b-r9qk6 1/1 Running 0 3m 10.97.10.183 qa-node008.zilliz.local <none> <none> benchmark-cron-1636214400-5-milvus-rootcoord-78bfbd8f4b-dmh8v 1/1 Running 0 3m 10.97.8.2 qa-node006.zilliz.local <none> <none> benchmark-cron-1636214400-5-minio-0 1/1 Running 0 3m 10.97.10.174 qa-node008.zilliz.local <none> <none> benchmark-cron-1636214400-5-minio-1 1/1 Running 0 3m 10.97.10.177 qa-node008.zilliz.local <none> <none> benchmark-cron-1636214400-5-minio-2 1/1 Running 0 3m 10.97.10.178 qa-node008.zilliz.local <none> <none> benchmark-cron-1636214400-5-minio-3 1/1 Running 0 3m 10.97.10.179 qa-node008.zilliz.local <none> <none> benchmark-cron-1636214400-5-pulsar-autorecovery-7fbbd955c4vkc9z 1/1 Running 0 3m 10.97.8.5 qa-node006.zilliz.local <none> <none> benchmark-cron-1636214400-5-pulsar-bastion-5567946888-xr94b 1/1 Running 0 3m 10.97.8.3 qa-node006.zilliz.local <none> <none> benchmark-cron-1636214400-5-pulsar-bookkeeper-0 1/1 Running 0 3m 10.97.8.8 qa-node006.zilliz.local <none> <none> benchmark-cron-1636214400-5-pulsar-bookkeeper-1 1/1 Running 0 57s 10.97.6.179 qa-node004.zilliz.local <none> <none> benchmark-cron-1636214400-5-pulsar-broker-6d5ccd5df4-q5ch7 1/1 Running 0 2m59s 10.97.15.52 qa-node012.zilliz.local <none> <none> benchmark-cron-1636214400-5-pulsar-proxy-84dbfcc96-2n7w5 2/2 Running 0 2m59s 10.97.13.62 qa-node010.zilliz.local <none> <none> benchmark-cron-1636214400-5-pulsar-zookeeper-0 1/1 Running 0 3m 10.97.9.254 qa-node007.zilliz.local <none> <none> benchmark-cron-1636214400-5-pulsar-zookeeper-1 1/1 Running 0 2m42s 10.97.6.177 qa-node004.zilliz.local <none> <none> benchmark-cron-1636214400-5-pulsar-zookeeper-2 1/1 Running 0 2m28s 10.97.6.178 qa-node004.zilliz.local <none> <none> benchmark-cron-1636214400-5-pulsar-zookeeper-metadata-zzjf5 0/1 Completed 0 3m 10.97.10.180 qa-node008.zilliz.local <none> <none> ``` ### Expected Behavior _No response_ ### Steps To Reproduce ```markdown argo task:benchmark-cron-1636214400 test yaml: client-configmap:client-acc-glove-ivf-flat server-configmap:server-cluster-8c16m ``` ### Anything else? client-acc-glove-ivf-flat: ``` ann_accuracy: collections: - source_file: /test/milvus/ann_hdf5/glove-200-angular.hdf5 collection_name: glove_200_angular index_types: ['ivf_flat'] index_params: nlist: [1024] top_ks: [10] nqs: [10000] search_params: nprobe: [1, 2, 4, 8, 16, 32, 64, 128, 256, 512] ```
test
search error no available query node result is there an existing issue for this i have searched the existing issues environment markdown milvus version master deployment mode standalone or cluster cluster sdk version e g pymilvus pymilvus os ubuntu or centos cpu memory gpu others current behavior client log milvus create index run in milvus benchmark client none milvus benchmark client index type flat metric type none index param none milvus benchmark runners accuracy start load collection glove angular milvus benchmark runners accuracy milvus load collection run in milvus benchmark client end load collection glove angular milvus benchmark runners accuracy start run case milvus benchmark main error baseexception code message no available query node result filter reason search failed c runtime error detected error code error msg error in milvus knowhere confad id pymilvus client grpc handler error baseexception code message no available query node result filter reason search failed c runtime error detected error code error msg error in milvus knowhere confad id pymilvus client grpc handler traceback most recent call last file main py line in run suite result runner run case case metric case file src milvus benchmark runners accuracy py line in run case self milvus query case param filter query case param file src milvus benchmark client py line in wrapper result func args kwargs file src milvus benchmark client py line in query result self milvus search tmp collection name params file usr local lib site packages pymilvus client stub py line in handler raise e file usr local lib site packages pymilvus client stub py line in handler return func self args kwargs file usr local lib site packages pymilvus client stub py line in search partition names output fields timeout round decimal kwargs file usr local lib site packages pymilvus client grpc handler py line in handler raise e file usr local lib site packages pymilvus client grpc handler py line in handler return func self args kwargs file usr local lib site packages pymilvus client grpc handler py line in handler return func self args kwargs file usr local lib site packages pymilvus client grpc handler py line in search return self execute search requests requests timeout kwargs file usr local lib site packages pymilvus client grpc handler py line in handler raise e file usr local lib site packages pymilvus client grpc handler py line in handler return func self args kwargs file usr local lib site packages pymilvus client grpc handler py line in execute search requests raise pre err file usr local lib site packages pymilvus client grpc handler py line in execute search requests raise baseexception response status error code response status reason pymilvus client exceptions baseexception baseexception code message no available query node result filter reason search failed c runtime error detected error code error msg error in milvus knowhere confad id milvus benchmark main none milvus benchmark main type ann accuracy value milvus benchmark main version type case run id mode local server hardware env status run failed err message ntraceback most recent call last n file main py line in run suite n result runner run case case metric case n file src milvus benchmark runners accuracy py line in run case n self milvus query case param filter query case param n file src milvus benchmark client py line in wrapper n result func args kwargs n file src milvus benchmark client py line in query n result self milvus search tmp collection name params n file usr local lib site packages pymilvus client stub py line in handler n raise e n file usr local lib site packages pymilvus client stub py line in handler n return func self args kwargs n file usr local lib site packages pymilvus client stub py line in search n partition names output fields timeout round decimal kwargs n file usr local lib site packages pymilvus client grpc handler py line in handler n raise e n file usr local lib site packages pymilvus client grpc handler py line in handler n return func self args kwargs n file usr local lib site packages pymilvus client grpc handler py line in handler n return func self args kwargs n file usr local lib site packages pymilvus client grpc handler py line in search n return self execute search requests requests timeout kwargs n file usr local lib site packages pymilvus client grpc handler py line in handler n raise e n file usr local lib site packages pymilvus client grpc handler py line in handler n return func self args kwargs n file usr local lib site packages pymilvus client grpc handler py line in execute search requests n raise pre err n file usr local lib site packages pymilvus client grpc handler py line in execute search requests n raise baseexception response status error code response status reason npymilvus client exceptions baseexception n collection dimension metric type ip dataset name glove angular shards num none index index type ivf flat index param nlist search nq topk search param nprobe filter run params none metrics type ann accuracy value datetime type metric milvus benchmark metric api server name ready status restarts age ip node nominated node readiness gates benchmark cron etcd running qa zilliz local benchmark cron etcd running qa zilliz local benchmark cron etcd running qa zilliz local benchmark cron milvus datacoord running qa zilliz local benchmark cron milvus datanode running qa zilliz local benchmark cron milvus indexcoord running qa zilliz local benchmark cron milvus indexnode running qa zilliz local benchmark cron milvus proxy running qa zilliz local benchmark cron milvus querycoord running qa zilliz local benchmark cron milvus querynode running qa zilliz local benchmark cron milvus rootcoord running qa zilliz local benchmark cron minio running qa zilliz local benchmark cron minio running qa zilliz local benchmark cron minio running qa zilliz local benchmark cron minio running qa zilliz local benchmark cron pulsar autorecovery running qa zilliz local benchmark cron pulsar bastion running qa zilliz local benchmark cron pulsar bookkeeper running qa zilliz local benchmark cron pulsar bookkeeper running qa zilliz local benchmark cron pulsar broker running qa zilliz local benchmark cron pulsar proxy running qa zilliz local benchmark cron pulsar zookeeper running qa zilliz local benchmark cron pulsar zookeeper running qa zilliz local benchmark cron pulsar zookeeper running qa zilliz local benchmark cron pulsar zookeeper metadata completed qa zilliz local expected behavior no response steps to reproduce markdown argo task benchmark cron test yaml client configmap client acc glove ivf flat server configmap server cluster anything else client acc glove ivf flat: ann accuracy collections source file test milvus ann glove angular collection name glove angular index types index params nlist top ks nqs search params nprobe
1
18,282
3,682,675,036
IssuesEvent
2016-02-24 10:43:04
IDgis/geoportaal-test
https://api.github.com/repos/IDgis/geoportaal-test
closed
Knoppen in huisstijl
gebruikerstest wens
Knoppen moeten conform huisstijl zijn, dus ook met icoontje voor de tekst. Is al wel in mockup nieuwe Geoportaal uitgewerkt. opmerking nr 12 n.a.v. testen 3-2-16 ![huisstijl_knoppen](https://cloud.githubusercontent.com/assets/9005444/12779885/f9fbc0e6-ca6a-11e5-9e1f-3f8bc419c761.jpg)
1.0
Knoppen in huisstijl - Knoppen moeten conform huisstijl zijn, dus ook met icoontje voor de tekst. Is al wel in mockup nieuwe Geoportaal uitgewerkt. opmerking nr 12 n.a.v. testen 3-2-16 ![huisstijl_knoppen](https://cloud.githubusercontent.com/assets/9005444/12779885/f9fbc0e6-ca6a-11e5-9e1f-3f8bc419c761.jpg)
test
knoppen in huisstijl knoppen moeten conform huisstijl zijn dus ook met icoontje voor de tekst is al wel in mockup nieuwe geoportaal uitgewerkt opmerking nr n a v testen
1
24,098
10,980,690,760
IssuesEvent
2019-11-30 16:14:10
openfoodfacts/openfoodfacts-server
https://api.github.com/repos/openfoodfacts/openfoodfacts-server
closed
Missing CORS headers on cgi/product_jqm2.pl endpoint
API bug security
As a public API the endpoints responsible for creation of different resources MUST return CORS headers - ref. [https://developer.mozilla.org/en-US/docs/Web/HTTP/CORS](https://developer.mozilla.org/en-US/docs/Web/HTTP/CORS) **Expected behavior** At least "Access-Control-Allow-Origin: *" value is expected as a header
True
Missing CORS headers on cgi/product_jqm2.pl endpoint - As a public API the endpoints responsible for creation of different resources MUST return CORS headers - ref. [https://developer.mozilla.org/en-US/docs/Web/HTTP/CORS](https://developer.mozilla.org/en-US/docs/Web/HTTP/CORS) **Expected behavior** At least "Access-Control-Allow-Origin: *" value is expected as a header
non_test
missing cors headers on cgi product pl endpoint as a public api the endpoints responsible for creation of different resources must return cors headers ref expected behavior at least access control allow origin value is expected as a header
0
461,408
13,229,811,384
IssuesEvent
2020-08-18 08:47:44
gambitph/Stackable
https://api.github.com/repos/gambitph/Stackable
closed
CSS code for Padding Top and Bottom didn't apply
bug high priority
CSS code for Margin Top and Bottom didn't apply, I think the PX is missing: ![chrome_2oJQN2t0tT](https://user-images.githubusercontent.com/51441886/89269117-0ceea580-d66c-11ea-881c-220f431fad03.png)
1.0
CSS code for Padding Top and Bottom didn't apply - CSS code for Margin Top and Bottom didn't apply, I think the PX is missing: ![chrome_2oJQN2t0tT](https://user-images.githubusercontent.com/51441886/89269117-0ceea580-d66c-11ea-881c-220f431fad03.png)
non_test
css code for padding top and bottom didn t apply css code for margin top and bottom didn t apply i think the px is missing
0
185,982
14,394,554,046
IssuesEvent
2020-12-03 01:34:01
github-vet/rangeclosure-findings
https://api.github.com/repos/github-vet/rangeclosure-findings
closed
moby/moby: daemon/logger/loggerutils/cache/log_cache_test.go; 11 LoC
fresh small test
Found a possible issue in [moby/moby](https://www.github.com/moby/moby) at [daemon/logger/loggerutils/cache/log_cache_test.go](https://github.com/moby/moby/blob/ca520bd1b807ace7c041bf416cb1bba385d968e5/daemon/logger/loggerutils/cache/log_cache_test.go#L59-L69) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call which takes a reference to msg at line 67 may start a goroutine [Click here to see the code in its original context.](https://github.com/moby/moby/blob/ca520bd1b807ace7c041bf416cb1bba385d968e5/daemon/logger/loggerutils/cache/log_cache_test.go#L59-L69) <details> <summary>Click here to show the 11 line(s) of Go which triggered the analyzer.</summary> ```go for _, msg := range messages { select { case <-ctx.Done(): return default: } m := logger.NewMessage() dumbCopyMessage(m, &msg) l.Log(m) } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: ca520bd1b807ace7c041bf416cb1bba385d968e5
1.0
moby/moby: daemon/logger/loggerutils/cache/log_cache_test.go; 11 LoC - Found a possible issue in [moby/moby](https://www.github.com/moby/moby) at [daemon/logger/loggerutils/cache/log_cache_test.go](https://github.com/moby/moby/blob/ca520bd1b807ace7c041bf416cb1bba385d968e5/daemon/logger/loggerutils/cache/log_cache_test.go#L59-L69) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call which takes a reference to msg at line 67 may start a goroutine [Click here to see the code in its original context.](https://github.com/moby/moby/blob/ca520bd1b807ace7c041bf416cb1bba385d968e5/daemon/logger/loggerutils/cache/log_cache_test.go#L59-L69) <details> <summary>Click here to show the 11 line(s) of Go which triggered the analyzer.</summary> ```go for _, msg := range messages { select { case <-ctx.Done(): return default: } m := logger.NewMessage() dumbCopyMessage(m, &msg) l.Log(m) } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: ca520bd1b807ace7c041bf416cb1bba385d968e5
test
moby moby daemon logger loggerutils cache log cache test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message function call which takes a reference to msg at line may start a goroutine click here to show the line s of go which triggered the analyzer go for msg range messages select case ctx done return default m logger newmessage dumbcopymessage m msg l log m leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
1