Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
4
112
repo_url
stringlengths
33
141
action
stringclasses
3 values
title
stringlengths
1
1.02k
labels
stringlengths
4
1.54k
body
stringlengths
1
262k
index
stringclasses
17 values
text_combine
stringlengths
95
262k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
166,695
26,396,225,092
IssuesEvent
2023-01-12 19:41:04
grafana/grafana
https://api.github.com/repos/grafana/grafana
opened
GLDS - TimeRangeInput breaks page formatting
area/ux timepicker design-system
**What happened**: When implementing a date selector component for the Faro plugin, the `TimeRangeInput` component breaks page formatting due to the date picker dropdown not using absolute positioning styling. **What you expected to happen**: The `TimeRangePicker` component uses absolute positioning and instead of breaking the page's formatting, the date picker sits nicely on top of the other elements on the page. **How to reproduce it (as minimally and precisely as possible)**: Simply use the `TimeRangeInput` component in any plugin (or observe its behavior in the [storybook](https://developers.grafana.com/ui/latest/index.html?path=/story/pickers-and-editors-timepickers-timerangeinput--basic)) and see that the component shifts around when you click to select the time range. **Anything else we need to know?**: **This very well could be expected behavior** but that does not seem likely. Additionally, this is likely related to [#60621](https://github.com/grafana/grafana/issues/60621) so this may already be on the radar. **Environment**: - Grafana version: `"grafana/ui": "^9.2.0"` - OS Grafana is installed on: macOS - User OS & Browser: Chrome, Firefox - Grafana plugins: Faro `TimeRangeInput` breaking formatting: https://user-images.githubusercontent.com/7052378/212161929-bc393589-e724-40ca-8ef8-da7f3f4e7a7c.mp4 `TimeRangePicker` in the same position working as expected: https://user-images.githubusercontent.com/7052378/212162144-c7538ddb-5327-482f-8f0b-d29725cc3a7a.mp4
1.0
GLDS - TimeRangeInput breaks page formatting - **What happened**: When implementing a date selector component for the Faro plugin, the `TimeRangeInput` component breaks page formatting due to the date picker dropdown not using absolute positioning styling. **What you expected to happen**: The `TimeRangePicker` component uses absolute positioning and instead of breaking the page's formatting, the date picker sits nicely on top of the other elements on the page. **How to reproduce it (as minimally and precisely as possible)**: Simply use the `TimeRangeInput` component in any plugin (or observe its behavior in the [storybook](https://developers.grafana.com/ui/latest/index.html?path=/story/pickers-and-editors-timepickers-timerangeinput--basic)) and see that the component shifts around when you click to select the time range. **Anything else we need to know?**: **This very well could be expected behavior** but that does not seem likely. Additionally, this is likely related to [#60621](https://github.com/grafana/grafana/issues/60621) so this may already be on the radar. **Environment**: - Grafana version: `"grafana/ui": "^9.2.0"` - OS Grafana is installed on: macOS - User OS & Browser: Chrome, Firefox - Grafana plugins: Faro `TimeRangeInput` breaking formatting: https://user-images.githubusercontent.com/7052378/212161929-bc393589-e724-40ca-8ef8-da7f3f4e7a7c.mp4 `TimeRangePicker` in the same position working as expected: https://user-images.githubusercontent.com/7052378/212162144-c7538ddb-5327-482f-8f0b-d29725cc3a7a.mp4
non_test
glds timerangeinput breaks page formatting what happened when implementing a date selector component for the faro plugin the timerangeinput component breaks page formatting due to the date picker dropdown not using absolute positioning styling what you expected to happen the timerangepicker component uses absolute positioning and instead of breaking the page s formatting the date picker sits nicely on top of the other elements on the page how to reproduce it as minimally and precisely as possible simply use the timerangeinput component in any plugin or observe its behavior in the and see that the component shifts around when you click to select the time range anything else we need to know this very well could be expected behavior but that does not seem likely additionally this is likely related to so this may already be on the radar environment grafana version grafana ui os grafana is installed on macos user os browser chrome firefox grafana plugins faro timerangeinput breaking formatting timerangepicker in the same position working as expected
0
292,487
25,217,585,736
IssuesEvent
2022-11-14 10:17:35
Test-Automation-Crash-Course-24-10-22/team_05
https://api.github.com/repos/Test-Automation-Crash-Course-24-10-22/team_05
opened
Rozetka [Positive] Сhecking the display of the "Warranty" page in the selected language
TestCase
**Descriptions:** Сhecking the display of the "Warranty" page in the selected language **Priority: Low** **Preconditions:** 1. Open https://rozetka.com.ua/ website and log in **Test steps** | Step | Test Data | Expected result | | ------------- | ------------- | ------------- | |1. Click on the "RU" button on a page header to switch the language to Russian | | | |2. Click on the "Warranty" button in the "Help" section | |The "Warranty" page is displayed in Russian |
1.0
Rozetka [Positive] Сhecking the display of the "Warranty" page in the selected language - **Descriptions:** Сhecking the display of the "Warranty" page in the selected language **Priority: Low** **Preconditions:** 1. Open https://rozetka.com.ua/ website and log in **Test steps** | Step | Test Data | Expected result | | ------------- | ------------- | ------------- | |1. Click on the "RU" button on a page header to switch the language to Russian | | | |2. Click on the "Warranty" button in the "Help" section | |The "Warranty" page is displayed in Russian |
test
rozetka сhecking the display of the warranty page in the selected language descriptions сhecking the display of the warranty page in the selected language priority low preconditions open website and log in test steps step test data expected result click on the ru button on a page header to switch the language to russian click on the warranty button in the help section the warranty page is displayed in russian
1
9,110
3,022,068,618
IssuesEvent
2015-07-31 18:10:04
couchbase/sync_gateway
https://api.github.com/repos/couchbase/sync_gateway
closed
Load balancer behavior results in high SG memory utilization
bug epic functional-test-needed P2: medium
Created to track [this issue raised by Brett Harrison on the mailing list](https://groups.google.com/d/msgid/mobile-couchbase/34cd2982-9d26-47c2-8ffd-85faa39acd7c%40googlegroups.com?utm_medium=email&utm_source=footer): > I am seeing a memory leaks in sync_gateway (1.0.4) > > I am not a Go programmer but I was hoping to gather some data that would be useful in finding these leaks. > I turned on profiling and took heap data snapshots about 12 hours apart. I hope I did this correctly. Is this data useful in finding the leaks? Is there anything else I could do to help?
1.0
Load balancer behavior results in high SG memory utilization - Created to track [this issue raised by Brett Harrison on the mailing list](https://groups.google.com/d/msgid/mobile-couchbase/34cd2982-9d26-47c2-8ffd-85faa39acd7c%40googlegroups.com?utm_medium=email&utm_source=footer): > I am seeing a memory leaks in sync_gateway (1.0.4) > > I am not a Go programmer but I was hoping to gather some data that would be useful in finding these leaks. > I turned on profiling and took heap data snapshots about 12 hours apart. I hope I did this correctly. Is this data useful in finding the leaks? Is there anything else I could do to help?
test
load balancer behavior results in high sg memory utilization created to track i am seeing a memory leaks in sync gateway i am not a go programmer but i was hoping to gather some data that would be useful in finding these leaks i turned on profiling and took heap data snapshots about hours apart i hope i did this correctly is this data useful in finding the leaks is there anything else i could do to help
1
168,173
14,141,150,756
IssuesEvent
2020-11-10 12:16:58
DADSILVER/Clipper-TwichtBot
https://api.github.com/repos/DADSILVER/Clipper-TwichtBot
closed
Escoger arquitectura del proyecto
documentation
Realizar un estudio de las distintas arquitecturas posible y escoger aquella que sea mas afín al proyecto.
1.0
Escoger arquitectura del proyecto - Realizar un estudio de las distintas arquitecturas posible y escoger aquella que sea mas afín al proyecto.
non_test
escoger arquitectura del proyecto realizar un estudio de las distintas arquitecturas posible y escoger aquella que sea mas afín al proyecto
0
76,050
7,517,130,032
IssuesEvent
2018-04-12 01:47:23
davidson16807/tectonics.js
https://api.github.com/repos/davidson16807/tectonics.js
closed
Log or console record of univariate stats
enhancement needs testing
I've been having trouble reconnecting the temperature model to altitude. I'm thinking this might be due to some problems with the values coming out of the altitude model itself. To test this, if it's not possible to export higher-order rasters(16-bit integer with at least min and max recorded, or optimally 32-bit floats…😌), it would be a real boon to have access to minimum, maximum, mean, median, possibly mode, and standard deviation to somewhere accessible(console output, a log record, an exported text file, or just as temporary text labels on the screen). This could perhaps be toggled at startup with a "univar" flag in the url… I'm thinking of this mainly for altitude, but I'm sure it would be useful for precipitation, temperature, biomass or any other quantitative values(again, toggled perhaps by "alt.univar," "prec.univar," etc. My preference would simply be an onscreen button menu to export a text file with the univariate stats for the selected parameter, but I'm flexible and overjoyed if you just decide to add this. Thank you, Colin
1.0
Log or console record of univariate stats - I've been having trouble reconnecting the temperature model to altitude. I'm thinking this might be due to some problems with the values coming out of the altitude model itself. To test this, if it's not possible to export higher-order rasters(16-bit integer with at least min and max recorded, or optimally 32-bit floats…😌), it would be a real boon to have access to minimum, maximum, mean, median, possibly mode, and standard deviation to somewhere accessible(console output, a log record, an exported text file, or just as temporary text labels on the screen). This could perhaps be toggled at startup with a "univar" flag in the url… I'm thinking of this mainly for altitude, but I'm sure it would be useful for precipitation, temperature, biomass or any other quantitative values(again, toggled perhaps by "alt.univar," "prec.univar," etc. My preference would simply be an onscreen button menu to export a text file with the univariate stats for the selected parameter, but I'm flexible and overjoyed if you just decide to add this. Thank you, Colin
test
log or console record of univariate stats i ve been having trouble reconnecting the temperature model to altitude i m thinking this might be due to some problems with the values coming out of the altitude model itself to test this if it s not possible to export higher order rasters bit integer with at least min and max recorded or optimally bit floats…😌 it would be a real boon to have access to minimum maximum mean median possibly mode and standard deviation to somewhere accessible console output a log record an exported text file or just as temporary text labels on the screen this could perhaps be toggled at startup with a univar flag in the url… i m thinking of this mainly for altitude but i m sure it would be useful for precipitation temperature biomass or any other quantitative values again toggled perhaps by alt univar prec univar etc my preference would simply be an onscreen button menu to export a text file with the univariate stats for the selected parameter but i m flexible and overjoyed if you just decide to add this thank you colin
1
190,649
14,566,762,115
IssuesEvent
2020-12-17 09:20:57
mozilla-mobile/fenix
https://api.github.com/repos/mozilla-mobile/fenix
opened
Intermittent UI test failure - BookmarksTest.editBookmarkTest
eng:intermittent-test eng:ui-test
### Firebase Test Run: https://console.firebase.google.com/u/0/project/moz-fenix/testlab/histories/bh.66b7091e15d53d45/matrices/9113695952861788340/executions/bs.a4c91089eeb2c61f/testcases/1/test-cases ### Stacktrace: `java.lang.NullPointerException: activityTestRule.activit…wById(R.id.bookmark_list) must not be null at org.mozilla.fenix.ui.BookmarksTest$editBookmarkTest$3.invoke(BookmarksTest.kt:162) at org.mozilla.fenix.ui.BookmarksTest$editBookmarkTest$3.invoke(BookmarksTest.kt:34) at org.mozilla.fenix.ui.robots.ThreeDotMenuMainRobot$Transition.openBookmarks(ThreeDotMenuMainRobot.kt:167) at org.mozilla.fenix.ui.BookmarksTest.editBookmarkTest(BookmarksTest.kt:160)` ### Build: Debug 12/16
2.0
Intermittent UI test failure - BookmarksTest.editBookmarkTest - ### Firebase Test Run: https://console.firebase.google.com/u/0/project/moz-fenix/testlab/histories/bh.66b7091e15d53d45/matrices/9113695952861788340/executions/bs.a4c91089eeb2c61f/testcases/1/test-cases ### Stacktrace: `java.lang.NullPointerException: activityTestRule.activit…wById(R.id.bookmark_list) must not be null at org.mozilla.fenix.ui.BookmarksTest$editBookmarkTest$3.invoke(BookmarksTest.kt:162) at org.mozilla.fenix.ui.BookmarksTest$editBookmarkTest$3.invoke(BookmarksTest.kt:34) at org.mozilla.fenix.ui.robots.ThreeDotMenuMainRobot$Transition.openBookmarks(ThreeDotMenuMainRobot.kt:167) at org.mozilla.fenix.ui.BookmarksTest.editBookmarkTest(BookmarksTest.kt:160)` ### Build: Debug 12/16
test
intermittent ui test failure bookmarkstest editbookmarktest firebase test run stacktrace java lang nullpointerexception activitytestrule activit…wbyid r id bookmark list must not be null at org mozilla fenix ui bookmarkstest editbookmarktest invoke bookmarkstest kt at org mozilla fenix ui bookmarkstest editbookmarktest invoke bookmarkstest kt at org mozilla fenix ui robots threedotmenumainrobot transition openbookmarks threedotmenumainrobot kt at org mozilla fenix ui bookmarkstest editbookmarktest bookmarkstest kt build debug
1
263,979
23,094,207,817
IssuesEvent
2022-07-26 17:51:39
microsoft/vscode
https://api.github.com/repos/microsoft/vscode
closed
Test: Sticky Scroll
testplan-item
Refs: https://github.com/microsoft/vscode/issues/26757 - [x] any OS @bpasero - [x] any OS @andreamah Complexity: 2 [Create Issue](https://github.com/microsoft/vscode/issues/new?body=Testing+%23156175%0A%0A&assignees=aiday-mar) --- It is possible to enable the sticky scroll and see in which class/method/constructor/interface/function you are in. You can also click on the sticky item to jump to the line. - Go to the settings and enable sticky scroll under `Editor > Sticky Scroll : Enable` - Open a file (so far I only tested on JavaScript files, so it would be a good idea to test other file types) - Scroll the file and check that the beginning lines corresponding to class/method/constructor/interface/function stick to the top of the editor correctly on the scrolling event, and appear and disappear smoothly. - Check the redirection on the onclick is correct. Any feedback is greatly appreciated. **Note** : Since the build, the code has been updated outside of the merge, in my personal repo, in an attempt to solve the following. - Incorrect indentation in the files. The indentation takes into account both tabs and spaces now. This is something I will continue working on this week. If interested, my updated personal repo is here : `https://github.com/aiday-mar/vscode/tree/aiday/semanticScroll` under the branch `aiday/semanticScroll`
1.0
Test: Sticky Scroll - Refs: https://github.com/microsoft/vscode/issues/26757 - [x] any OS @bpasero - [x] any OS @andreamah Complexity: 2 [Create Issue](https://github.com/microsoft/vscode/issues/new?body=Testing+%23156175%0A%0A&assignees=aiday-mar) --- It is possible to enable the sticky scroll and see in which class/method/constructor/interface/function you are in. You can also click on the sticky item to jump to the line. - Go to the settings and enable sticky scroll under `Editor > Sticky Scroll : Enable` - Open a file (so far I only tested on JavaScript files, so it would be a good idea to test other file types) - Scroll the file and check that the beginning lines corresponding to class/method/constructor/interface/function stick to the top of the editor correctly on the scrolling event, and appear and disappear smoothly. - Check the redirection on the onclick is correct. Any feedback is greatly appreciated. **Note** : Since the build, the code has been updated outside of the merge, in my personal repo, in an attempt to solve the following. - Incorrect indentation in the files. The indentation takes into account both tabs and spaces now. This is something I will continue working on this week. If interested, my updated personal repo is here : `https://github.com/aiday-mar/vscode/tree/aiday/semanticScroll` under the branch `aiday/semanticScroll`
test
test sticky scroll refs any os bpasero any os andreamah complexity it is possible to enable the sticky scroll and see in which class method constructor interface function you are in you can also click on the sticky item to jump to the line go to the settings and enable sticky scroll under editor sticky scroll enable open a file so far i only tested on javascript files so it would be a good idea to test other file types scroll the file and check that the beginning lines corresponding to class method constructor interface function stick to the top of the editor correctly on the scrolling event and appear and disappear smoothly check the redirection on the onclick is correct any feedback is greatly appreciated note since the build the code has been updated outside of the merge in my personal repo in an attempt to solve the following incorrect indentation in the files the indentation takes into account both tabs and spaces now this is something i will continue working on this week if interested my updated personal repo is here under the branch aiday semanticscroll
1
130,383
12,427,622,831
IssuesEvent
2020-05-25 03:09:04
OpenMined/GridNetwork
https://api.github.com/repos/OpenMined/GridNetwork
closed
Provide GridNetwork as a package
Type: Documentation :books: Type: Epic :call_me_hand: Type: New Feature :heavy_plus_sign:
**Is your feature request related to a problem? Please describe.** Right now GridNetwork can only be procured through Github releases and Docker images. It would be helpful to adapt this repository to make use of the power of the [Pypi Index].(https://pypi.org/). **Describe the solution you'd like** An intermediate change that keeps whatever deployment and application that uses GridNetwork working. A version of this package published to the [TestPyPi Index](https://test.pypi.org/). **Describe alternatives you've considered** A clean update to the repo, changing its structure to keep everything needed for the package in an inner directory [flask guidelines](https://flask.palletsprojects.com/en/1.1.x/patterns/packages/). Simply doing this might break whatever's currently hooked to this project. **Additional context** * https://flask.palletsprojects.com/en/1.1.x/tutorial/install/ * https://docs.python-guide.org/writing/structure/
1.0
Provide GridNetwork as a package - **Is your feature request related to a problem? Please describe.** Right now GridNetwork can only be procured through Github releases and Docker images. It would be helpful to adapt this repository to make use of the power of the [Pypi Index].(https://pypi.org/). **Describe the solution you'd like** An intermediate change that keeps whatever deployment and application that uses GridNetwork working. A version of this package published to the [TestPyPi Index](https://test.pypi.org/). **Describe alternatives you've considered** A clean update to the repo, changing its structure to keep everything needed for the package in an inner directory [flask guidelines](https://flask.palletsprojects.com/en/1.1.x/patterns/packages/). Simply doing this might break whatever's currently hooked to this project. **Additional context** * https://flask.palletsprojects.com/en/1.1.x/tutorial/install/ * https://docs.python-guide.org/writing/structure/
non_test
provide gridnetwork as a package is your feature request related to a problem please describe right now gridnetwork can only be procured through github releases and docker images it would be helpful to adapt this repository to make use of the power of the describe the solution you d like an intermediate change that keeps whatever deployment and application that uses gridnetwork working a version of this package published to the describe alternatives you ve considered a clean update to the repo changing its structure to keep everything needed for the package in an inner directory simply doing this might break whatever s currently hooked to this project additional context
0
487,849
14,060,491,213
IssuesEvent
2020-11-03 06:08:34
PyTorchLightning/pytorch-lightning
https://api.github.com/repos/PyTorchLightning/pytorch-lightning
closed
Issue with epoch count with repeated save/restore
Checkpoint Priority bug / fix help wanted
## 🐛 Bug I'm trying to save and restore the state of both a model and a pytorch-lightning trainer. I suspect the epoch count is wrong because I'm not able to save and restore several times with the same max_epoch count. Here's what I do: Step 1: run model for max_epochs = 1. Save checkpoint (gets saved as epoch=0.ckpt) Step 2: load previous checkpoint and rerun again with max_epochs = 1. No training is run (because 1 epoch was already run before). A checkpoint is saved again, however this is called epoch=1.ckpt. Step 3: load checkpoint from step 2 and rerun again with max_epochs = 1. Training fails because it believes step 2 was run for 2 epochs (and here max is 1) Output: ``` pytorch_lightning.utilities.exceptions.MisconfigurationException: you restored a checkpoint with current_epoch=2 but the Trainer(max_epochs=1) ``` Code below to reproduce. What am I doing wrong? this should be a possible scenario right? Thanks! ### To Reproduce Run code below 3 times from same location ``` import os import torch from torch import nn import torch.nn.functional as F from torchvision.datasets import MNIST from torch.utils.data import DataLoader, random_split from torchvision import transforms import pytorch_lightning as pl from pathlib import Path class LitAutoEncoder(pl.LightningModule): def __init__(self): super().__init__() self.encoder = nn.Sequential(nn.Linear(28 * 28, 128), nn.ReLU(), nn.Linear(128, 3)) self.decoder = nn.Sequential(nn.Linear(3, 128), nn.ReLU(), nn.Linear(128, 28 * 28)) def forward(self, x): # in lightning, forward defines the prediction/inference actions embedding = self.encoder(x) return embedding def training_step(self, batch, batch_idx): # training_step defined the train loop. It is independent of forward x, y = batch x = x.view(x.size(0), -1) z = self.encoder(x) x_hat = self.decoder(z) loss = F.mse_loss(x_hat, x) self.log('train_loss', loss) return loss def configure_optimizers(self): optimizer = torch.optim.Adam(self.parameters(), lr=1e-3) return optimizer dataset = MNIST(os.getcwd(), download=True, transform=transforms.ToTensor()) train, val, _ = random_split(dataset, [5500, 500, len(dataset) - 6000]) def get_last_checkpoint(checkpoint_folder): if os.path.exists(checkpoint_folder): past_experiments = sorted(Path(checkpoint_folder).iterdir(), key=os.path.getmtime) for experiment in past_experiments[::-1]: experiment_folder = os.path.join(experiment, "checkpoints") if os.path.exists(experiment_folder): checkpoints = os.listdir(experiment_folder) if len(checkpoints): checkpoints.sort() path = os.path.join(experiment_folder, checkpoints[-1]) return path return None chk = get_last_checkpoint('lightning_logs') if chk is not None: print("loading from ", chk) autoencoder = LitAutoEncoder.load_from_checkpoint(chk) else: autoencoder = LitAutoEncoder() trainer = pl.Trainer(max_epochs=1, resume_from_checkpoint=chk) trainer.fit(autoencoder, DataLoader(train), DataLoader(val)) ``` ### Expected behavior Should not increase epochs on second run. Should be able to load checkpoint and save unchanged (several times) ### Environment * CUDA: - GPU: - Quadro P2000 with Max-Q Design - available: True - version: 10.2 * Packages: - numpy: 1.19.2 - pyTorch_debug: True - pyTorch_version: 1.8.0.dev20201014 - pytorch-lightning: 1.0.2 - tqdm: 4.50.2 * System: - OS: Windows - architecture: - 64bit - WindowsPE - processor: Intel64 Family 6 Model 158 Stepping 10, GenuineIntel - python: 3.7.9 - version: 10.0.17763 ### Additional context <!-- Add any other context about the problem here. -->
1.0
Issue with epoch count with repeated save/restore - ## 🐛 Bug I'm trying to save and restore the state of both a model and a pytorch-lightning trainer. I suspect the epoch count is wrong because I'm not able to save and restore several times with the same max_epoch count. Here's what I do: Step 1: run model for max_epochs = 1. Save checkpoint (gets saved as epoch=0.ckpt) Step 2: load previous checkpoint and rerun again with max_epochs = 1. No training is run (because 1 epoch was already run before). A checkpoint is saved again, however this is called epoch=1.ckpt. Step 3: load checkpoint from step 2 and rerun again with max_epochs = 1. Training fails because it believes step 2 was run for 2 epochs (and here max is 1) Output: ``` pytorch_lightning.utilities.exceptions.MisconfigurationException: you restored a checkpoint with current_epoch=2 but the Trainer(max_epochs=1) ``` Code below to reproduce. What am I doing wrong? this should be a possible scenario right? Thanks! ### To Reproduce Run code below 3 times from same location ``` import os import torch from torch import nn import torch.nn.functional as F from torchvision.datasets import MNIST from torch.utils.data import DataLoader, random_split from torchvision import transforms import pytorch_lightning as pl from pathlib import Path class LitAutoEncoder(pl.LightningModule): def __init__(self): super().__init__() self.encoder = nn.Sequential(nn.Linear(28 * 28, 128), nn.ReLU(), nn.Linear(128, 3)) self.decoder = nn.Sequential(nn.Linear(3, 128), nn.ReLU(), nn.Linear(128, 28 * 28)) def forward(self, x): # in lightning, forward defines the prediction/inference actions embedding = self.encoder(x) return embedding def training_step(self, batch, batch_idx): # training_step defined the train loop. It is independent of forward x, y = batch x = x.view(x.size(0), -1) z = self.encoder(x) x_hat = self.decoder(z) loss = F.mse_loss(x_hat, x) self.log('train_loss', loss) return loss def configure_optimizers(self): optimizer = torch.optim.Adam(self.parameters(), lr=1e-3) return optimizer dataset = MNIST(os.getcwd(), download=True, transform=transforms.ToTensor()) train, val, _ = random_split(dataset, [5500, 500, len(dataset) - 6000]) def get_last_checkpoint(checkpoint_folder): if os.path.exists(checkpoint_folder): past_experiments = sorted(Path(checkpoint_folder).iterdir(), key=os.path.getmtime) for experiment in past_experiments[::-1]: experiment_folder = os.path.join(experiment, "checkpoints") if os.path.exists(experiment_folder): checkpoints = os.listdir(experiment_folder) if len(checkpoints): checkpoints.sort() path = os.path.join(experiment_folder, checkpoints[-1]) return path return None chk = get_last_checkpoint('lightning_logs') if chk is not None: print("loading from ", chk) autoencoder = LitAutoEncoder.load_from_checkpoint(chk) else: autoencoder = LitAutoEncoder() trainer = pl.Trainer(max_epochs=1, resume_from_checkpoint=chk) trainer.fit(autoencoder, DataLoader(train), DataLoader(val)) ``` ### Expected behavior Should not increase epochs on second run. Should be able to load checkpoint and save unchanged (several times) ### Environment * CUDA: - GPU: - Quadro P2000 with Max-Q Design - available: True - version: 10.2 * Packages: - numpy: 1.19.2 - pyTorch_debug: True - pyTorch_version: 1.8.0.dev20201014 - pytorch-lightning: 1.0.2 - tqdm: 4.50.2 * System: - OS: Windows - architecture: - 64bit - WindowsPE - processor: Intel64 Family 6 Model 158 Stepping 10, GenuineIntel - python: 3.7.9 - version: 10.0.17763 ### Additional context <!-- Add any other context about the problem here. -->
non_test
issue with epoch count with repeated save restore 🐛 bug i m trying to save and restore the state of both a model and a pytorch lightning trainer i suspect the epoch count is wrong because i m not able to save and restore several times with the same max epoch count here s what i do step run model for max epochs save checkpoint gets saved as epoch ckpt step load previous checkpoint and rerun again with max epochs no training is run because epoch was already run before a checkpoint is saved again however this is called epoch ckpt step load checkpoint from step and rerun again with max epochs training fails because it believes step was run for epochs and here max is output pytorch lightning utilities exceptions misconfigurationexception you restored a checkpoint with current epoch but the trainer max epochs code below to reproduce what am i doing wrong this should be a possible scenario right thanks to reproduce run code below times from same location import os import torch from torch import nn import torch nn functional as f from torchvision datasets import mnist from torch utils data import dataloader random split from torchvision import transforms import pytorch lightning as pl from pathlib import path class litautoencoder pl lightningmodule def init self super init self encoder nn sequential nn linear nn relu nn linear self decoder nn sequential nn linear nn relu nn linear def forward self x in lightning forward defines the prediction inference actions embedding self encoder x return embedding def training step self batch batch idx training step defined the train loop it is independent of forward x y batch x x view x size z self encoder x x hat self decoder z loss f mse loss x hat x self log train loss loss return loss def configure optimizers self optimizer torch optim adam self parameters lr return optimizer dataset mnist os getcwd download true transform transforms totensor train val random split dataset def get last checkpoint checkpoint folder if os path exists checkpoint folder past experiments sorted path checkpoint folder iterdir key os path getmtime for experiment in past experiments experiment folder os path join experiment checkpoints if os path exists experiment folder checkpoints os listdir experiment folder if len checkpoints checkpoints sort path os path join experiment folder checkpoints return path return none chk get last checkpoint lightning logs if chk is not none print loading from chk autoencoder litautoencoder load from checkpoint chk else autoencoder litautoencoder trainer pl trainer max epochs resume from checkpoint chk trainer fit autoencoder dataloader train dataloader val expected behavior should not increase epochs on second run should be able to load checkpoint and save unchanged several times environment cuda gpu quadro with max q design available true version packages numpy pytorch debug true pytorch version pytorch lightning tqdm system os windows architecture windowspe processor family model stepping genuineintel python version additional context
0
64,621
6,912,996,714
IssuesEvent
2017-11-28 13:59:35
pennsignals/mantis
https://api.github.com/repos/pennsignals/mantis
closed
Create BDD tests using Cucumber
testing
Use [Cucumber](https://cucumber.io/) and [Behave](http://pythonhosted.org/behave/) to write Behavior-driven development (BDD) tests.
1.0
Create BDD tests using Cucumber - Use [Cucumber](https://cucumber.io/) and [Behave](http://pythonhosted.org/behave/) to write Behavior-driven development (BDD) tests.
test
create bdd tests using cucumber use and to write behavior driven development bdd tests
1
210,346
23,752,998,234
IssuesEvent
2022-08-31 22:47:02
kube-tarian/sigrun
https://api.github.com/repos/kube-tarian/sigrun
closed
CVE-2021-41091 (Medium) detected in github.com/docker/docker-v20.10.7
security vulnerability no-issue-activity
## CVE-2021-41091 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>github.com/docker/docker-v20.10.7</b></p></summary> <p>Moby Project - a collaborative project for the container ecosystem to assemble container-based systems</p> <p> Dependency Hierarchy: - github.com/google/go-containerregistry-v0.6.0 (Root Library) - github.com/docker/cli-v20.10.7 - :x: **github.com/docker/docker-v20.10.7** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/kube-tarian/sigrun/commit/df1f7d3f67826e841793324e4796be4fbd91c00f">df1f7d3f67826e841793324e4796be4fbd91c00f</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Moby is an open-source project created by Docker to enable software containerization. A bug was found in Moby (Docker Engine) where the data directory (typically `/var/lib/docker`) contained subdirectories with insufficiently restricted permissions, allowing otherwise unprivileged Linux users to traverse directory contents and execute programs. When containers included executable programs with extended permission bits (such as `setuid`), unprivileged Linux users could discover and execute those programs. When the UID of an unprivileged Linux user on the host collided with the file owner or group inside a container, the unprivileged Linux user on the host could discover, read, and modify those files. This bug has been fixed in Moby (Docker Engine) 20.10.9. Users should update to this version as soon as possible. Running containers should be stopped and restarted for the permissions to be fixed. For users unable to upgrade limit access to the host to trusted users. Limit access to host volumes to trusted containers. <p>Publish Date: 2021-10-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-41091>CVE-2021-41091</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/moby/moby/security/advisories/GHSA-3fwx-pjgw-3558">https://github.com/moby/moby/security/advisories/GHSA-3fwx-pjgw-3558</a></p> <p>Release Date: 2021-10-04</p> <p>Fix Resolution: v20.10.9</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-41091 (Medium) detected in github.com/docker/docker-v20.10.7 - ## CVE-2021-41091 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>github.com/docker/docker-v20.10.7</b></p></summary> <p>Moby Project - a collaborative project for the container ecosystem to assemble container-based systems</p> <p> Dependency Hierarchy: - github.com/google/go-containerregistry-v0.6.0 (Root Library) - github.com/docker/cli-v20.10.7 - :x: **github.com/docker/docker-v20.10.7** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/kube-tarian/sigrun/commit/df1f7d3f67826e841793324e4796be4fbd91c00f">df1f7d3f67826e841793324e4796be4fbd91c00f</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Moby is an open-source project created by Docker to enable software containerization. A bug was found in Moby (Docker Engine) where the data directory (typically `/var/lib/docker`) contained subdirectories with insufficiently restricted permissions, allowing otherwise unprivileged Linux users to traverse directory contents and execute programs. When containers included executable programs with extended permission bits (such as `setuid`), unprivileged Linux users could discover and execute those programs. When the UID of an unprivileged Linux user on the host collided with the file owner or group inside a container, the unprivileged Linux user on the host could discover, read, and modify those files. This bug has been fixed in Moby (Docker Engine) 20.10.9. Users should update to this version as soon as possible. Running containers should be stopped and restarted for the permissions to be fixed. For users unable to upgrade limit access to the host to trusted users. Limit access to host volumes to trusted containers. <p>Publish Date: 2021-10-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-41091>CVE-2021-41091</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/moby/moby/security/advisories/GHSA-3fwx-pjgw-3558">https://github.com/moby/moby/security/advisories/GHSA-3fwx-pjgw-3558</a></p> <p>Release Date: 2021-10-04</p> <p>Fix Resolution: v20.10.9</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve medium detected in github com docker docker cve medium severity vulnerability vulnerable library github com docker docker moby project a collaborative project for the container ecosystem to assemble container based systems dependency hierarchy github com google go containerregistry root library github com docker cli x github com docker docker vulnerable library found in head commit a href found in base branch main vulnerability details moby is an open source project created by docker to enable software containerization a bug was found in moby docker engine where the data directory typically var lib docker contained subdirectories with insufficiently restricted permissions allowing otherwise unprivileged linux users to traverse directory contents and execute programs when containers included executable programs with extended permission bits such as setuid unprivileged linux users could discover and execute those programs when the uid of an unprivileged linux user on the host collided with the file owner or group inside a container the unprivileged linux user on the host could discover read and modify those files this bug has been fixed in moby docker engine users should update to this version as soon as possible running containers should be stopped and restarted for the permissions to be fixed for users unable to upgrade limit access to the host to trusted users limit access to host volumes to trusted containers publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope changed impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
23,134
7,290,600,363
IssuesEvent
2018-02-24 03:40:27
tensorflow/tensorflow
https://api.github.com/repos/tensorflow/tensorflow
closed
Error downloading nasm
type:build/install
As noted in issue #6950 nasm-2.12.02.tar.bz2 is currently unavailable. (www.nasm.us does not accept connections.) @trsaunders observed this for head of r0.12 and I can confirm this for v0.12.0. Workaround: Google the file to get it from another source and update URL in tensorflow/workspace.bzl (if you download a .gz file re-compress to bz2 and put it on a webserver) - sha256 fingerprinting should check that the file contents are correct; if you don't trust ./configure to check this, check the sha256sum yourself against the sha256 given in the .bzl file.
1.0
Error downloading nasm - As noted in issue #6950 nasm-2.12.02.tar.bz2 is currently unavailable. (www.nasm.us does not accept connections.) @trsaunders observed this for head of r0.12 and I can confirm this for v0.12.0. Workaround: Google the file to get it from another source and update URL in tensorflow/workspace.bzl (if you download a .gz file re-compress to bz2 and put it on a webserver) - sha256 fingerprinting should check that the file contents are correct; if you don't trust ./configure to check this, check the sha256sum yourself against the sha256 given in the .bzl file.
non_test
error downloading nasm as noted in issue nasm tar is currently unavailable does not accept connections trsaunders observed this for head of and i can confirm this for workaround google the file to get it from another source and update url in tensorflow workspace bzl if you download a gz file re compress to and put it on a webserver fingerprinting should check that the file contents are correct if you don t trust configure to check this check the yourself against the given in the bzl file
0
90,059
25,959,018,270
IssuesEvent
2022-12-18 16:33:00
CosmosOS/Cosmos
https://api.github.com/repos/CosmosOS/Cosmos
closed
System.IO.FileLoaderException: Could not load file or assembly Cosmos.Debug.Kernel
Pending User Response Area: Visual Studio Integration Area: Build
![image](https://user-images.githubusercontent.com/97028115/153651463-90002e1c-1ff7-456a-bec2-a2cb914dea59.png) Please help.
1.0
System.IO.FileLoaderException: Could not load file or assembly Cosmos.Debug.Kernel - ![image](https://user-images.githubusercontent.com/97028115/153651463-90002e1c-1ff7-456a-bec2-a2cb914dea59.png) Please help.
non_test
system io fileloaderexception could not load file or assembly cosmos debug kernel please help
0
273,489
23,758,731,895
IssuesEvent
2022-09-01 06:58:37
Kuadrant/testsuite
https://api.github.com/repos/Kuadrant/testsuite
closed
Add clusterwide tests
test-case Kuadrant
Authorino can be configured to reconcile resources across all namespaces
1.0
Add clusterwide tests - Authorino can be configured to reconcile resources across all namespaces
test
add clusterwide tests authorino can be configured to reconcile resources across all namespaces
1
260,916
8,216,862,466
IssuesEvent
2018-09-05 10:28:59
bio-tools/biotoolsRegistry
https://api.github.com/repos/bio-tools/biotoolsRegistry
closed
Suppress rendering of top-level EDAM annotations
bug content critical priority simple feature request
![capture](https://cloud.githubusercontent.com/assets/1506863/25635195/3100ca8a-2f75-11e7-9b56-131fb46f1501.PNG) this also goes for Topic of "Topic" ... both just make us look bad. While we await meaningful annotations, they need to be suppressed from the view.
1.0
Suppress rendering of top-level EDAM annotations - ![capture](https://cloud.githubusercontent.com/assets/1506863/25635195/3100ca8a-2f75-11e7-9b56-131fb46f1501.PNG) this also goes for Topic of "Topic" ... both just make us look bad. While we await meaningful annotations, they need to be suppressed from the view.
non_test
suppress rendering of top level edam annotations this also goes for topic of topic both just make us look bad while we await meaningful annotations they need to be suppressed from the view
0
110,364
9,449,320,086
IssuesEvent
2019-04-16 01:21:41
rancher/k3s
https://api.github.com/repos/rancher/k3s
closed
RPi broken on v0.4.0-rc1
status/blocker status/to-test
```Apr 11 23:41:18 k3s-base k3s[2084]: F0411 23:41:18.953434 2084 kubelet.go:1327] Failed to start ContainerManager failed to initialize top level QOS containers: failed to update top level Burstable QOS cgroup : failed to set supported cgroup subsystems for cgroup [kubepods burstable]: Failed to find subsystem mount for required subsystem: pids Apr 11 23:41:18 k3s-base k3s[2084]: goroutine 4913 [running]: Apr 11 23:41:18 k3s-base k3s[2084]: github.com/rancher/k3s/vendor/k8s.io/klog.stacks(0x4e82e00, 0x0, 0x138, 0x1d8) Apr 11 23:41:18 k3s-base k3s[2084]: /go/src/github.com/rancher/k3s/vendor/k8s.io/klog/klog.go:828 +0x94 Apr 11 23:41:18 k3s-base k3s[2084]: github.com/rancher/k3s/vendor/k8s.io/klog.(*loggingT).output(0x4e703c0, 0x3, 0x911a900, 0x4bcc54b, 0xa, 0x52f, 0x0) Apr 11 23:41:18 k3s-base k3s[2084]: /go/src/github.com/rancher/k3s/vendor/k8s.io/klog/klog.go:779 +0x2a0 Apr 11 23:41:18 k3s-base k3s[2084]: github.com/rancher/k3s/vendor/k8s.io/klog.(*loggingT).printf(0x4e703c0, 0x3, 0x2799ae0, 0x23, 0x85a9e7c, 0x1, 0x1) Apr 11 23:41:18 k3s-base k3s[2084]: /go/src/github.com/rancher/k3s/vendor/k8s.io/klog/klog.go:678 +0x110 Apr 11 23:41:18 k3s-base k3s[2084]: github.com/rancher/k3s/vendor/k8s.io/klog.Fatalf(...) Apr 11 23:41:18 k3s-base k3s[2084]: /go/src/github.com/rancher/k3s/vendor/k8s.io/klog/klog.go:1209 Apr 11 23:41:18 k3s-base k3s[2084]: github.com/rancher/k3s/vendor/k8s.io/kubernetes/pkg/kubelet.(*Kubelet).initializeRuntimeDependentModules(0x775d900) Apr 11 23:41:18 k3s-base k3s[2084]: /go/src/github.com/rancher/k3s/vendor/k8s.io/kubernetes/pkg/kubelet/kubelet.go:1327 +0x270 Apr 11 23:41:18 k3s-base k3s[2084]: sync.(*Once).Do(0x775dd20, 0x92c4f10) Apr 11 23:41:18 k3s-base k3s[2084]: /usr/local/go/src/sync/once.go:44 +0xb8 Apr 11 23:41:18 k3s-base k3s[2084]: github.com/rancher/k3s/vendor/k8s.io/kubernetes/pkg/kubelet.(*Kubelet).updateRuntimeUp(0x775d900) Apr 11 23:41:18 k3s-base k3s[2084]: /go/src/github.com/rancher/k3s/vendor/k8s.io/kubernetes/pkg/kubelet/kubelet.go:2153 +0x334 Apr 11 23:41:18 k3s-base k3s[2084]: github.com/rancher/k3s/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil.func1(0x92a2610) Apr 11 23:41:18 k3s-base k3s[2084]: /go/src/github.com/rancher/k3s/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:152 +0x4c Apr 11 23:41:18 k3s-base k3s[2084]: github.com/rancher/k3s/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil(0x92a2610, 0x2a05f200, 0x1, 0x0, 0x0, 0x1, 0x67158c0) Apr 11 23:41:18 k3s-base k3s[2084]: /go/src/github.com/rancher/k3s/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:153 +0xfc Apr 11 23:41:18 k3s-base k3s[2084]: github.com/rancher/k3s/vendor/k8s.io/apimachinery/pkg/util/wait.Until(0x92a2610, 0x2a05f200, 0x1, 0x67158c0) Apr 11 23:41:18 k3s-base k3s[2084]: /go/src/github.com/rancher/k3s/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:88 +0x48 Apr 11 23:41:18 k3s-base k3s[2084]: created by github.com/rancher/k3s/vendor/k8s.io/kubernetes/pkg/kubelet.(*Kubelet).Run Apr 11 23:41:18 k3s-base k3s[2084]: /go/src/github.com/rancher/k3s/vendor/k8s.io/kubernetes/pkg/kubelet/kubelet.go:1376 +0xfc Apr 11 23:41:21 k3s-base systemd[1]: k3s.service: Main process exited, code=exited, status=255/n/a Apr 11 23:41:21 k3s-base systemd[1]: k3s.service: Unit entered failed state. Apr 11 23:41:21 k3s-base systemd[1]: k3s.service: Failed with result 'exit-code'. ```
1.0
RPi broken on v0.4.0-rc1 - ```Apr 11 23:41:18 k3s-base k3s[2084]: F0411 23:41:18.953434 2084 kubelet.go:1327] Failed to start ContainerManager failed to initialize top level QOS containers: failed to update top level Burstable QOS cgroup : failed to set supported cgroup subsystems for cgroup [kubepods burstable]: Failed to find subsystem mount for required subsystem: pids Apr 11 23:41:18 k3s-base k3s[2084]: goroutine 4913 [running]: Apr 11 23:41:18 k3s-base k3s[2084]: github.com/rancher/k3s/vendor/k8s.io/klog.stacks(0x4e82e00, 0x0, 0x138, 0x1d8) Apr 11 23:41:18 k3s-base k3s[2084]: /go/src/github.com/rancher/k3s/vendor/k8s.io/klog/klog.go:828 +0x94 Apr 11 23:41:18 k3s-base k3s[2084]: github.com/rancher/k3s/vendor/k8s.io/klog.(*loggingT).output(0x4e703c0, 0x3, 0x911a900, 0x4bcc54b, 0xa, 0x52f, 0x0) Apr 11 23:41:18 k3s-base k3s[2084]: /go/src/github.com/rancher/k3s/vendor/k8s.io/klog/klog.go:779 +0x2a0 Apr 11 23:41:18 k3s-base k3s[2084]: github.com/rancher/k3s/vendor/k8s.io/klog.(*loggingT).printf(0x4e703c0, 0x3, 0x2799ae0, 0x23, 0x85a9e7c, 0x1, 0x1) Apr 11 23:41:18 k3s-base k3s[2084]: /go/src/github.com/rancher/k3s/vendor/k8s.io/klog/klog.go:678 +0x110 Apr 11 23:41:18 k3s-base k3s[2084]: github.com/rancher/k3s/vendor/k8s.io/klog.Fatalf(...) Apr 11 23:41:18 k3s-base k3s[2084]: /go/src/github.com/rancher/k3s/vendor/k8s.io/klog/klog.go:1209 Apr 11 23:41:18 k3s-base k3s[2084]: github.com/rancher/k3s/vendor/k8s.io/kubernetes/pkg/kubelet.(*Kubelet).initializeRuntimeDependentModules(0x775d900) Apr 11 23:41:18 k3s-base k3s[2084]: /go/src/github.com/rancher/k3s/vendor/k8s.io/kubernetes/pkg/kubelet/kubelet.go:1327 +0x270 Apr 11 23:41:18 k3s-base k3s[2084]: sync.(*Once).Do(0x775dd20, 0x92c4f10) Apr 11 23:41:18 k3s-base k3s[2084]: /usr/local/go/src/sync/once.go:44 +0xb8 Apr 11 23:41:18 k3s-base k3s[2084]: github.com/rancher/k3s/vendor/k8s.io/kubernetes/pkg/kubelet.(*Kubelet).updateRuntimeUp(0x775d900) Apr 11 23:41:18 k3s-base k3s[2084]: /go/src/github.com/rancher/k3s/vendor/k8s.io/kubernetes/pkg/kubelet/kubelet.go:2153 +0x334 Apr 11 23:41:18 k3s-base k3s[2084]: github.com/rancher/k3s/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil.func1(0x92a2610) Apr 11 23:41:18 k3s-base k3s[2084]: /go/src/github.com/rancher/k3s/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:152 +0x4c Apr 11 23:41:18 k3s-base k3s[2084]: github.com/rancher/k3s/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil(0x92a2610, 0x2a05f200, 0x1, 0x0, 0x0, 0x1, 0x67158c0) Apr 11 23:41:18 k3s-base k3s[2084]: /go/src/github.com/rancher/k3s/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:153 +0xfc Apr 11 23:41:18 k3s-base k3s[2084]: github.com/rancher/k3s/vendor/k8s.io/apimachinery/pkg/util/wait.Until(0x92a2610, 0x2a05f200, 0x1, 0x67158c0) Apr 11 23:41:18 k3s-base k3s[2084]: /go/src/github.com/rancher/k3s/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:88 +0x48 Apr 11 23:41:18 k3s-base k3s[2084]: created by github.com/rancher/k3s/vendor/k8s.io/kubernetes/pkg/kubelet.(*Kubelet).Run Apr 11 23:41:18 k3s-base k3s[2084]: /go/src/github.com/rancher/k3s/vendor/k8s.io/kubernetes/pkg/kubelet/kubelet.go:1376 +0xfc Apr 11 23:41:21 k3s-base systemd[1]: k3s.service: Main process exited, code=exited, status=255/n/a Apr 11 23:41:21 k3s-base systemd[1]: k3s.service: Unit entered failed state. Apr 11 23:41:21 k3s-base systemd[1]: k3s.service: Failed with result 'exit-code'. ```
test
rpi broken on apr base kubelet go failed to start containermanager failed to initialize top level qos containers failed to update top level burstable qos cgroup failed to set supported cgroup subsystems for cgroup failed to find subsystem mount for required subsystem pids apr base goroutine apr base github com rancher vendor io klog stacks apr base go src github com rancher vendor io klog klog go apr base github com rancher vendor io klog loggingt output apr base go src github com rancher vendor io klog klog go apr base github com rancher vendor io klog loggingt printf apr base go src github com rancher vendor io klog klog go apr base github com rancher vendor io klog fatalf apr base go src github com rancher vendor io klog klog go apr base github com rancher vendor io kubernetes pkg kubelet kubelet initializeruntimedependentmodules apr base go src github com rancher vendor io kubernetes pkg kubelet kubelet go apr base sync once do apr base usr local go src sync once go apr base github com rancher vendor io kubernetes pkg kubelet kubelet updateruntimeup apr base go src github com rancher vendor io kubernetes pkg kubelet kubelet go apr base github com rancher vendor io apimachinery pkg util wait jitteruntil apr base go src github com rancher vendor io apimachinery pkg util wait wait go apr base github com rancher vendor io apimachinery pkg util wait jitteruntil apr base go src github com rancher vendor io apimachinery pkg util wait wait go apr base github com rancher vendor io apimachinery pkg util wait until apr base go src github com rancher vendor io apimachinery pkg util wait wait go apr base created by github com rancher vendor io kubernetes pkg kubelet kubelet run apr base go src github com rancher vendor io kubernetes pkg kubelet kubelet go apr base systemd service main process exited code exited status n a apr base systemd service unit entered failed state apr base systemd service failed with result exit code
1
99,978
12,495,179,845
IssuesEvent
2020-06-01 12:43:28
NYCPlanning/labs-applicant-portal
https://api.github.com/repos/NYCPlanning/labs-applicant-portal
opened
RWCDS Submission confirmation modal
Design User: Applicant
<!-- Don't forget to add labels to this issue! --> [Azure Requirement](#) <!-- For Feature Requests/Enhancements --> ## User Story / User Journey ### User Story As an applicant, I want to be asked if I'm sure I want to "Submit" so that I can be sure I have included everything that's needed and didn't make a mistake. ## Front End ### [Wireframes](#) Provide link to Figma Paste in screenshots
1.0
RWCDS Submission confirmation modal - <!-- Don't forget to add labels to this issue! --> [Azure Requirement](#) <!-- For Feature Requests/Enhancements --> ## User Story / User Journey ### User Story As an applicant, I want to be asked if I'm sure I want to "Submit" so that I can be sure I have included everything that's needed and didn't make a mistake. ## Front End ### [Wireframes](#) Provide link to Figma Paste in screenshots
non_test
rwcds submission confirmation modal user story user journey user story as an applicant i want to be asked if i m sure i want to submit so that i can be sure i have included everything that s needed and didn t make a mistake front end provide link to figma paste in screenshots
0
56,048
23,687,535,592
IssuesEvent
2022-08-29 07:50:36
Azure/azure-cli
https://api.github.com/repos/Azure/azure-cli
closed
ERROR: Scope must be a valid resource id
bug Service Attention Monitor customer-reported needs-author-feedback Auto-Assign Azure CLI Team
> ### `az feedback` auto-generates most of the information requested below, as of CLI version 2.0.62 **Related command** az monitor metrics alert create **Describe the bug** When trying to create alert for Logic App failure, I am getting ERROR no matter what is actually provided as scopes **To Reproduce** use following SHELL script: ` ALERT_NAME="Scheduling` $(LOGIC_APP_NAME) failed" ALERT_DESCRYPTION="$ALERT_NAME at least 3 times in last 15 minutes" RESOURCE_ID=`az resource show -n '$(LOGIC_APP_NAME)' --resource-group $(RESOURCE_GROUP) --resource-type Microsoft.Logic/workflows --query id` az monitor metrics alert create \ --name "$ALERT_NAME" \ --description "$ALERT_DESCRYPTION" \ --resource-group $(RESOURCE_GROUP) \ --scopes $RESOURCE_ID \ --condition "total Runs Failed > 2" \ --action $(ACTION) \ --severity 4 ` **Expected behavior** <!--- A clear and concise description of what you expected to happen. ---> **Additional context** I am using Azure Resource Manager connection
1.0
ERROR: Scope must be a valid resource id - > ### `az feedback` auto-generates most of the information requested below, as of CLI version 2.0.62 **Related command** az monitor metrics alert create **Describe the bug** When trying to create alert for Logic App failure, I am getting ERROR no matter what is actually provided as scopes **To Reproduce** use following SHELL script: ` ALERT_NAME="Scheduling` $(LOGIC_APP_NAME) failed" ALERT_DESCRYPTION="$ALERT_NAME at least 3 times in last 15 minutes" RESOURCE_ID=`az resource show -n '$(LOGIC_APP_NAME)' --resource-group $(RESOURCE_GROUP) --resource-type Microsoft.Logic/workflows --query id` az monitor metrics alert create \ --name "$ALERT_NAME" \ --description "$ALERT_DESCRYPTION" \ --resource-group $(RESOURCE_GROUP) \ --scopes $RESOURCE_ID \ --condition "total Runs Failed > 2" \ --action $(ACTION) \ --severity 4 ` **Expected behavior** <!--- A clear and concise description of what you expected to happen. ---> **Additional context** I am using Azure Resource Manager connection
non_test
error scope must be a valid resource id az feedback auto generates most of the information requested below as of cli version related command az monitor metrics alert create describe the bug when trying to create alert for logic app failure i am getting error no matter what is actually provided as scopes to reproduce use following shell script alert name scheduling logic app name failed alert descryption alert name at least times in last minutes resource id az resource show n logic app name resource group resource group resource type microsoft logic workflows query id az monitor metrics alert create name alert name description alert descryption resource group resource group scopes resource id condition total runs failed action action severity expected behavior additional context i am using azure resource manager connection
0
99,119
20,886,159,610
IssuesEvent
2022-03-23 05:33:08
appsmithorg/appsmith
https://api.github.com/repos/appsmithorg/appsmith
closed
[Bug]: Revamp delete confirmation for all destructive actions
Bug Actions Pod High Needs Triaging BE Coders Pod
### Is there an existing issue for this? - [X] I have searched the existing issues ### Description For all deletion operations, we make use of clicking the delete button twice instead of a modal ### Steps To Reproduce try to delete a datasource, see the modal pop up ### Public Sample App _No response_ ### Version 1.66
1.0
[Bug]: Revamp delete confirmation for all destructive actions - ### Is there an existing issue for this? - [X] I have searched the existing issues ### Description For all deletion operations, we make use of clicking the delete button twice instead of a modal ### Steps To Reproduce try to delete a datasource, see the modal pop up ### Public Sample App _No response_ ### Version 1.66
non_test
revamp delete confirmation for all destructive actions is there an existing issue for this i have searched the existing issues description for all deletion operations we make use of clicking the delete button twice instead of a modal steps to reproduce try to delete a datasource see the modal pop up public sample app no response version
0
266,560
8,371,443,681
IssuesEvent
2018-10-05 00:41:35
containous/traefik
https://api.github.com/repos/containous/traefik
closed
Traefik not responsive after number of requests
kind/bug/possible platform/windows priority/P3
<!-- DO NOT FILE ISSUES FOR GENERAL SUPPORT QUESTIONS. The issue tracker is for reporting bugs and feature requests only. For end-user related support questions, please refer to one of the following: - Stack Overflow (using the "traefik" tag): https://stackoverflow.com/questions/tagged/traefik - the Traefik community Slack channel: https://slack.traefik.io --> ### Do you want to request a *feature* or report a *bug*? Bug ### What did you do? We're running Traefik 1.6.4 on Windows Server Core 1709 (Also tested on 1803) in Docker Swarm Mode. After a certain number of requests from a single machine, Traefik will no longer accept traffic from this machine. Other machines are able to connect and will be routed to the right container, but the original machine will not until Traefik is restarted. We have reproduced this on several machines, including on-premises, and cloud IaaS machines. ### What did you expect to see? Traefik should keep routing requests ### What did you see instead? After a certain amount of connections from a single IP, Traefik will no longer proxy connections from that IP. ### Output of `traefik version`: (_What version of Traefik are you using?_) ``` Version: v1.6.4 Codename: tetedemoine Go version: go1.10.3 Built: 2018-06-15_03:12:50PM OS/Arch: windows/amd64 ``` ### What is your environment & configuration (arguments, toml, provider, platform, ...)? Windows Version ``` Microsoft Windows [Version 10.0.16299.492] Windows Server Core 1709 ``` docker version ``` Version: 17.06.2-ee-13 API version: 1.30 Go version: go1.8.7 Git commit: ac44d73 Built: Mon Jun 4 16:46:59 2018 OS/Arch: windows/amd64 Server: Engine: Version: 17.06.2-ee-13 API version: 1.30 (minimum version 1.24) Go version: go1.8.7 Git commit: ac44d73 Built: Mon Jun 4 16:58:47 2018 OS/Arch: windows/amd64 Experimental: false ``` Docker daemon.json ```json { "fixed-cidr": "192.168.1.254/24", "hosts": [ "tcp://0.0.0.0:2375", "npipe://" ] } ``` Traefik Dockerfile, based on https://github.com/StefanScherer/dockerfiles-windows/blob/master/traefik/Dockerfile.1709 ```dockerfile FROM microsoft/windowsservercore:1709 as core FROM microsoft/nanoserver:1709 COPY --from=core /windows/system32/netapi32.dll /windows/system32/netapi32.dll ENV TRAEFIK_VERSION 1.6.4 ADD https://github.com/containous/traefik/releases/download/v${TRAEFIK_VERSION}/traefik_windows-amd64.exe /traefik.exe VOLUME C:/etc/traefik VOLUME C:/etc/ssl EXPOSE 80 ENTRYPOINT ["/traefik", "--configfile=C:/etc/traefik/traefik.toml"] # Metadata LABEL org.label-schema.vendor="Containous" \ org.label-schema.url="https://traefik.io" \ org.label-schema.name="Traefik" \ org.label-schema.description="A modern reverse-proxy" \ org.label-schema.version="v${TRAEFIK_VERSION}" \ org.label-schema.docker.schema-version="1.0" ``` docker-compose.yml ```yml version: '3.3' services: traefik: image: ivanjosipovic/public:traefik-v1.6.4-1709 command: - "--logLevel=DEBUG" - "--docker" - "--docker.domain=docker.local" - "--docker.endpoint=tcp://192.168.1.1:2375" - "--docker.swarmMode" - "--docker.watch" - "--docker.exposedbydefault=false" - "--entrypoints=Name:http Address::80" - "--defaultentrypoints=http" ports: - target: 80 published: 80 mode: host webapp: image: microsoft/dotnet-samples:aspnetapp-nanoserver-1709 depends_on: - traefik deploy: endpoint_mode: dnsrr labels: - "traefik.enable=true" - "traefik.port=80" - "traefik.frontend.rule=Host:tst-app.com" - "traefik.backend.loadbalancer.method=drr" - "traefik.docker.network: traefik" ports: - target: 80 published: 8080 mode: host networks: default: external: name: traefik ``` ### Repro Steps #### On Docker Host: Install Windows Server Core 1709 ``` Install-Module DockerMsftProvider -Force Install-Package Docker -ProviderName DockerMsftProvider -Force ``` Copy `daemon.json` above to `C:\ProgramData\docker\config` ``` start-service docker docker swarm init docker network create --driver=overlay traefik docker stack deploy --compose-file docker-compose.yml --prune Base ```` #### On a remote machine: Add a entry to the hosts file for, {docker host ip } tst-app.com Download bombardier, https://github.com/codesenberg/bombardier/releases Run: ``` .\bombardier-windows-amd64.exe -c 100 -n 5000 http://tst-app.com ``` Once this is complete, try loading the site through a browser, it will not work. However, if you try http://tst-app.com:8080 this will work as it bypasses Traefik ### If applicable, please paste the log output in DEBUG level (`--logLevel=DEBUG` switch) Log is too large to paste here, a zip can be grabbed here, https://1drv.ms/u/s!ArrmDf0NziYphOZhvuh8p6Xk9eflqw
1.0
Traefik not responsive after number of requests - <!-- DO NOT FILE ISSUES FOR GENERAL SUPPORT QUESTIONS. The issue tracker is for reporting bugs and feature requests only. For end-user related support questions, please refer to one of the following: - Stack Overflow (using the "traefik" tag): https://stackoverflow.com/questions/tagged/traefik - the Traefik community Slack channel: https://slack.traefik.io --> ### Do you want to request a *feature* or report a *bug*? Bug ### What did you do? We're running Traefik 1.6.4 on Windows Server Core 1709 (Also tested on 1803) in Docker Swarm Mode. After a certain number of requests from a single machine, Traefik will no longer accept traffic from this machine. Other machines are able to connect and will be routed to the right container, but the original machine will not until Traefik is restarted. We have reproduced this on several machines, including on-premises, and cloud IaaS machines. ### What did you expect to see? Traefik should keep routing requests ### What did you see instead? After a certain amount of connections from a single IP, Traefik will no longer proxy connections from that IP. ### Output of `traefik version`: (_What version of Traefik are you using?_) ``` Version: v1.6.4 Codename: tetedemoine Go version: go1.10.3 Built: 2018-06-15_03:12:50PM OS/Arch: windows/amd64 ``` ### What is your environment & configuration (arguments, toml, provider, platform, ...)? Windows Version ``` Microsoft Windows [Version 10.0.16299.492] Windows Server Core 1709 ``` docker version ``` Version: 17.06.2-ee-13 API version: 1.30 Go version: go1.8.7 Git commit: ac44d73 Built: Mon Jun 4 16:46:59 2018 OS/Arch: windows/amd64 Server: Engine: Version: 17.06.2-ee-13 API version: 1.30 (minimum version 1.24) Go version: go1.8.7 Git commit: ac44d73 Built: Mon Jun 4 16:58:47 2018 OS/Arch: windows/amd64 Experimental: false ``` Docker daemon.json ```json { "fixed-cidr": "192.168.1.254/24", "hosts": [ "tcp://0.0.0.0:2375", "npipe://" ] } ``` Traefik Dockerfile, based on https://github.com/StefanScherer/dockerfiles-windows/blob/master/traefik/Dockerfile.1709 ```dockerfile FROM microsoft/windowsservercore:1709 as core FROM microsoft/nanoserver:1709 COPY --from=core /windows/system32/netapi32.dll /windows/system32/netapi32.dll ENV TRAEFIK_VERSION 1.6.4 ADD https://github.com/containous/traefik/releases/download/v${TRAEFIK_VERSION}/traefik_windows-amd64.exe /traefik.exe VOLUME C:/etc/traefik VOLUME C:/etc/ssl EXPOSE 80 ENTRYPOINT ["/traefik", "--configfile=C:/etc/traefik/traefik.toml"] # Metadata LABEL org.label-schema.vendor="Containous" \ org.label-schema.url="https://traefik.io" \ org.label-schema.name="Traefik" \ org.label-schema.description="A modern reverse-proxy" \ org.label-schema.version="v${TRAEFIK_VERSION}" \ org.label-schema.docker.schema-version="1.0" ``` docker-compose.yml ```yml version: '3.3' services: traefik: image: ivanjosipovic/public:traefik-v1.6.4-1709 command: - "--logLevel=DEBUG" - "--docker" - "--docker.domain=docker.local" - "--docker.endpoint=tcp://192.168.1.1:2375" - "--docker.swarmMode" - "--docker.watch" - "--docker.exposedbydefault=false" - "--entrypoints=Name:http Address::80" - "--defaultentrypoints=http" ports: - target: 80 published: 80 mode: host webapp: image: microsoft/dotnet-samples:aspnetapp-nanoserver-1709 depends_on: - traefik deploy: endpoint_mode: dnsrr labels: - "traefik.enable=true" - "traefik.port=80" - "traefik.frontend.rule=Host:tst-app.com" - "traefik.backend.loadbalancer.method=drr" - "traefik.docker.network: traefik" ports: - target: 80 published: 8080 mode: host networks: default: external: name: traefik ``` ### Repro Steps #### On Docker Host: Install Windows Server Core 1709 ``` Install-Module DockerMsftProvider -Force Install-Package Docker -ProviderName DockerMsftProvider -Force ``` Copy `daemon.json` above to `C:\ProgramData\docker\config` ``` start-service docker docker swarm init docker network create --driver=overlay traefik docker stack deploy --compose-file docker-compose.yml --prune Base ```` #### On a remote machine: Add a entry to the hosts file for, {docker host ip } tst-app.com Download bombardier, https://github.com/codesenberg/bombardier/releases Run: ``` .\bombardier-windows-amd64.exe -c 100 -n 5000 http://tst-app.com ``` Once this is complete, try loading the site through a browser, it will not work. However, if you try http://tst-app.com:8080 this will work as it bypasses Traefik ### If applicable, please paste the log output in DEBUG level (`--logLevel=DEBUG` switch) Log is too large to paste here, a zip can be grabbed here, https://1drv.ms/u/s!ArrmDf0NziYphOZhvuh8p6Xk9eflqw
non_test
traefik not responsive after number of requests do not file issues for general support questions the issue tracker is for reporting bugs and feature requests only for end user related support questions please refer to one of the following stack overflow using the traefik tag the traefik community slack channel do you want to request a feature or report a bug bug what did you do we re running traefik on windows server core also tested on in docker swarm mode after a certain number of requests from a single machine traefik will no longer accept traffic from this machine other machines are able to connect and will be routed to the right container but the original machine will not until traefik is restarted we have reproduced this on several machines including on premises and cloud iaas machines what did you expect to see traefik should keep routing requests what did you see instead after a certain amount of connections from a single ip traefik will no longer proxy connections from that ip output of traefik version what version of traefik are you using version codename tetedemoine go version built os arch windows what is your environment configuration arguments toml provider platform windows version microsoft windows windows server core docker version version ee api version go version git commit built mon jun os arch windows server engine version ee api version minimum version go version git commit built mon jun os arch windows experimental false docker daemon json json fixed cidr hosts tcp npipe traefik dockerfile based on dockerfile from microsoft windowsservercore as core from microsoft nanoserver copy from core windows dll windows dll env traefik version add traefik exe volume c etc traefik volume c etc ssl expose entrypoint metadata label org label schema vendor containous org label schema url org label schema name traefik org label schema description a modern reverse proxy org label schema version v traefik version org label schema docker schema version docker compose yml yml version services traefik image ivanjosipovic public traefik command loglevel debug docker docker domain docker local docker endpoint tcp docker swarmmode docker watch docker exposedbydefault false entrypoints name http address defaultentrypoints http ports target published mode host webapp image microsoft dotnet samples aspnetapp nanoserver depends on traefik deploy endpoint mode dnsrr labels traefik enable true traefik port traefik frontend rule host tst app com traefik backend loadbalancer method drr traefik docker network traefik ports target published mode host networks default external name traefik repro steps on docker host install windows server core install module dockermsftprovider force install package docker providername dockermsftprovider force copy daemon json above to c programdata docker config start service docker docker swarm init docker network create driver overlay traefik docker stack deploy compose file docker compose yml prune base on a remote machine add a entry to the hosts file for docker host ip tst app com download bombardier run bombardier windows exe c n once this is complete try loading the site through a browser it will not work however if you try this will work as it bypasses traefik if applicable please paste the log output in debug level loglevel debug switch log is too large to paste here a zip can be grabbed here
0
121,036
10,147,905,413
IssuesEvent
2019-08-05 11:51:26
bitcoin/bitcoin
https://api.github.com/repos/bitcoin/bitcoin
closed
test: feature_pruning fails with AssertionError (intermittent)
Tests
Current master. Excerpt: ``` node0 2019-08-02T18:17:40.159357Z [http] Received a POST request for / from 127.0.0.1:41390 node0 2019-08-02T18:17:40.206391Z [httpworker.3] ThreadRPCServer method=submitblock user=__cookie__ node0 2019-08-02T18:17:40.365855Z [httpworker.3] PeerLogicValidation::NewPoWValidBlock sending header-and-ids 6db3371c4f0ad084a24c512bb7a3c0de35280a5feb1037b6ba4ceae5bf579114 to peer=27 node0 2019-08-02T18:17:40.367280Z [httpworker.3] sending cmpctblock (950159 bytes) peer=27 node0 2019-08-02T18:17:40.381896Z [httpworker.3] PeerLogicValidation::NewPoWValidBlock sending header-and-ids 6db3371c4f0ad084a24c512bb7a3c0de35280a5feb1037b6ba4ceae5bf579114 to peer=28 node0 2019-08-02T18:17:40.383189Z [httpworker.3] sending cmpctblock (950159 bytes) peer=28 node0 2019-08-02T18:17:40.406853Z [httpworker.3] - Load block from disk: 0.00ms [0.05s] node0 2019-08-02T18:17:40.407052Z [httpworker.3] - Sanity checks: 0.01ms [0.03s (0.01ms/blk)] node0 2019-08-02T18:17:40.407149Z [httpworker.3] - Fork checks: 0.11ms [0.19s (0.09ms/blk)] node1 2019-08-02T18:17:40.417969Z [msghand] received: cmpctblock (950159 bytes) peer=14 node1 2019-08-02T18:17:40.437712Z [msghand] Initialized PartiallyDownloadedBlock for block 6db3371c4f0ad084a24c512bb7a3c0de35280a5feb1037b6ba4ceae5bf579114 using a cmpctblock of size 950159 node1 2019-08-02T18:17:40.437830Z [msghand] received: blocktxn (33 bytes) peer=14 node0 2019-08-02T18:17:40.461279Z [httpworker.3] - Connect 1 transactions: 54.05ms (54.045ms/tx, 0.000ms/txin) [71.60s (35.98ms/blk)] node0 2019-08-02T18:17:40.461408Z [httpworker.3] - Verify 0 txins: 54.25ms (0.000ms/txin) [71.94s (36.15ms/blk)] node0 2019-08-02T18:17:40.461652Z [httpworker.3] - Index writing: 0.25ms [0.50s (0.25ms/blk)] node0 2019-08-02T18:17:40.461701Z [httpworker.3] - Callbacks: 0.05ms [0.10s (0.05ms/blk)] node0 2019-08-02T18:17:40.461781Z [httpworker.3] - Connect total: 54.98ms [73.14s (36.75ms/blk)] node0 2019-08-02T18:17:40.461831Z [httpworker.3] - Flush: 0.04ms [0.11s (0.06ms/blk)] node0 2019-08-02T18:17:40.461903Z [httpworker.3] - Writing chainstate: 0.08ms [0.16s (0.08ms/blk)] node0 2019-08-02T18:17:40.463359Z [httpworker.3] Blockpolicy estimates updated by 0 of 0 block txs, since last block 0 of 0 tracked, mempool map size 0, max target 0 from current node0 2019-08-02T18:17:40.463950Z [httpworker.3] UpdateTip: new best=6db3371c4f0ad084a24c512bb7a3c0de35280a5feb1037b6ba4ceae5bf579114 height=1552 version=0x20000000 log2_work=11.600842 tx=1553 date='2019-08-02T18:19:52Z' progress=1.000000 cache=0.1MiB(650txo) node0 2019-08-02T18:17:40.464006Z [httpworker.3] - Connect postprocess: 2.10ms [3.91s (1.96ms/blk)] node0 2019-08-02T18:17:40.464046Z [httpworker.3] - Connect block: 57.22ms [77.36s (38.88ms/blk)] node0 2019-08-02T18:17:40.464129Z [httpworker.3] Checking mempool with 0 transactions and 0 inputs test 2019-08-02T18:17:40.472000Z TestFramework (INFO): Usage should be below target: 560 node1 2019-08-02T18:17:40.486452Z [msghand] Successfully reconstructed block 6db3371c4f0ad084a24c512bb7a3c0de35280a5feb1037b6ba4ceae5bf579114 with 1 txn prefilled, 0 txn from mempool (incl at least 0 from extra pool) and 0 txn requested node1 2019-08-02T18:17:40.492254Z [msghand] PeerLogicValidation::NewPoWValidBlock sending header-and-ids 6db3371c4f0ad084a24c512bb7a3c0de35280a5feb1037b6ba4ceae5bf579114 to peer=15 node1 2019-08-02T18:17:40.493497Z [msghand] sending cmpctblock (950159 bytes) peer=15 node1 2019-08-02T18:17:40.516284Z [msghand] - Load block from disk: 0.01ms [0.01s] node1 2019-08-02T18:17:40.516417Z [msghand] - Sanity checks: 0.01ms [0.02s (0.01ms/blk)] node1 2019-08-02T18:17:40.516502Z [msghand] - Fork checks: 0.09ms [0.25s (0.11ms/blk)] node1 2019-08-02T18:17:40.568948Z [msghand] - Connect 1 transactions: 52.38ms (52.379ms/tx, 0.000ms/txin) [69.38s (29.65ms/blk)] node1 2019-08-02T18:17:40.569060Z [msghand] - Verify 0 txins: 52.55ms (0.000ms/txin) [69.83s (29.84ms/blk)] node1 2019-08-02T18:17:40.569290Z [msghand] - Index writing: 0.23ms [0.44s (0.19ms/blk)] node1 2019-08-02T18:17:40.569332Z [msghand] - Callbacks: 0.05ms [0.12s (0.05ms/blk)] node1 2019-08-02T18:17:40.569409Z [msghand] - Connect total: 53.17ms [70.93s (30.31ms/blk)] node1 2019-08-02T18:17:40.569454Z [msghand] - Flush: 0.04ms [0.10s (0.04ms/blk)] node1 2019-08-02T18:17:40.569527Z [msghand] - Writing chainstate: 0.07ms [0.14s (0.06ms/blk)] node1 2019-08-02T18:17:40.570938Z [msghand] Blockpolicy estimates updated by 0 of 0 block txs, since last block 0 of 0 tracked, mempool map size 0, max target 0 from current node1 2019-08-02T18:17:40.571506Z [msghand] UpdateTip: new best=6db3371c4f0ad084a24c512bb7a3c0de35280a5feb1037b6ba4ceae5bf579114 height=1552 version=0x20000000 log2_work=11.600842 tx=1553 date='2019-08-02T18:19:52Z' progress=1.000000 cache=0.1MiB(650txo) node1 2019-08-02T18:17:40.571560Z [msghand] - Connect postprocess: 2.03ms [3.66s (1.57ms/blk)] node1 2019-08-02T18:17:40.571599Z [msghand] - Connect block: 55.33ms [74.84s (31.98ms/blk)] node1 2019-08-02T18:17:40.571675Z [msghand] Checking mempool with 0 transactions and 0 inputs test 2019-08-02T18:17:43.227000Z TestFramework (ERROR): Assertion failed Traceback (most recent call last): File "/root/workspace/bitcoin-core/test/functional/test_framework/test_framework.py", line 193, in main self.run_test() File "/root/workspace/bitcoin-core/test/functional/feature_pruning.py", line 396, in run_test self.reorg_test() # (1033, ) File "/root/workspace/bitcoin-core/test/functional/feature_pruning.py", line 210, in reorg_test assert_greater_than(550, usage) File "/root/workspace/bitcoin-core/test/functional/test_framework/util.py", line 44, in assert_greater_than raise AssertionError("%s <= %s" % (str(thing1), str(thing2))) AssertionError: 550 <= 560.1718864440918 node2 2019-08-02T18:17:44.476610Z [msghand] Prune: UnlinkPrunedFiles deleted blk/rev (00004) node2 2019-08-02T18:17:44.476787Z [msghand] Writing final batch of 0.00 MiB node2 2019-08-02T18:17:44.477697Z [msghand] Committed 0 changed transaction outputs (out of 0) to coin database... node2 2019-08-02T18:17:44.484856Z [msghand] - Load block from disk: 0.01ms [0.04s] node2 2019-08-02T18:17:44.485366Z [msghand] - Sanity checks: 0.11ms [0.02s (0.01ms/blk)] node2 2019-08-02T18:17:44.485479Z [msghand] - Fork checks: 0.14ms [0.18s (0.10ms/blk)] test 2019-08-02T18:17:44.522000Z TestFramework (DEBUG): Closing down network thread node2 2019-08-02T18:17:44.544340Z [msghand] - Connect 1 transactions: 58.78ms (58.779ms/tx, 0.000ms/txin) [70.18s (38.69ms/blk)] node2 2019-08-02T18:17:44.544504Z [msghand] - Verify 0 txins: 59.02ms (0.000ms/txin) [70.48s (38.86ms/blk)] node2 2019-08-02T18:17:44.544743Z [msghand] - Index writing: 0.24ms [0.45s (0.25ms/blk)] node2 2019-08-02T18:17:44.544786Z [msghand] - Callbacks: 0.05ms [0.09s (0.05ms/blk)] node2 2019-08-02T18:17:44.544871Z [msghand] - Connect total: 60.10ms [71.70s (39.52ms/blk)] node2 2019-08-02T18:17:44.544915Z [msghand] - Flush: 0.05ms [0.09s (0.05ms/blk)] node2 2019-08-02T18:17:44.544985Z [msghand] - Writing chainstate: 0.07ms [0.14s (0.08ms/blk)] node2 2019-08-02T18:17:44.546402Z [msghand] Blockpolicy estimates updated by 0 of 0 block txs, since last block 0 of 0 tracked, mempool map size 0, max target 0 from current node2 2019-08-02T18:17:44.546973Z [msghand] UpdateTip: new best=3fa2e1b6704dc4c0ea7238313abe4242915ce18351455222a8510ad0fbefd19f height=1526 version=0x20000000 log2_work=11.576484 tx=1527 date='2019-08-02T18:19:26Z' progress=1.000000 cache=0.0MiB(0txo) node2 2019-08-02T18:17:44.547032Z [msghand] - Connect postprocess: 2.04ms [3.64s (2.01ms/blk)] node2 2019-08-02T18:17:44.547072Z [msghand] - Connect block: 62.27ms [75.61s (41.68ms/blk)] node2 2019-08-02T18:17:44.547154Z [msghand] Checking mempool with 0 transactions and 0 inputs node2 2019-08-02T18:17:44.554409Z [msghand] received: cmpctblock (950159 bytes) peer=14 node2 2019-08-02T18:17:44.575317Z [msghand] sending ping (8 bytes) peer=14 node1 2019-08-02T18:17:44.575900Z [msghand] received: ping (8 bytes) peer=15 node2 2019-08-02T18:17:44.576024Z [msghand] received: cmpctblock (950159 bytes) peer=13 node1 2019-08-02T18:17:44.576117Z [msghand] sending pong (8 bytes) peer=15 test 2019-08-02T18:17:44.587000Z TestFramework (INFO): Stopping nodes
1.0
test: feature_pruning fails with AssertionError (intermittent) - Current master. Excerpt: ``` node0 2019-08-02T18:17:40.159357Z [http] Received a POST request for / from 127.0.0.1:41390 node0 2019-08-02T18:17:40.206391Z [httpworker.3] ThreadRPCServer method=submitblock user=__cookie__ node0 2019-08-02T18:17:40.365855Z [httpworker.3] PeerLogicValidation::NewPoWValidBlock sending header-and-ids 6db3371c4f0ad084a24c512bb7a3c0de35280a5feb1037b6ba4ceae5bf579114 to peer=27 node0 2019-08-02T18:17:40.367280Z [httpworker.3] sending cmpctblock (950159 bytes) peer=27 node0 2019-08-02T18:17:40.381896Z [httpworker.3] PeerLogicValidation::NewPoWValidBlock sending header-and-ids 6db3371c4f0ad084a24c512bb7a3c0de35280a5feb1037b6ba4ceae5bf579114 to peer=28 node0 2019-08-02T18:17:40.383189Z [httpworker.3] sending cmpctblock (950159 bytes) peer=28 node0 2019-08-02T18:17:40.406853Z [httpworker.3] - Load block from disk: 0.00ms [0.05s] node0 2019-08-02T18:17:40.407052Z [httpworker.3] - Sanity checks: 0.01ms [0.03s (0.01ms/blk)] node0 2019-08-02T18:17:40.407149Z [httpworker.3] - Fork checks: 0.11ms [0.19s (0.09ms/blk)] node1 2019-08-02T18:17:40.417969Z [msghand] received: cmpctblock (950159 bytes) peer=14 node1 2019-08-02T18:17:40.437712Z [msghand] Initialized PartiallyDownloadedBlock for block 6db3371c4f0ad084a24c512bb7a3c0de35280a5feb1037b6ba4ceae5bf579114 using a cmpctblock of size 950159 node1 2019-08-02T18:17:40.437830Z [msghand] received: blocktxn (33 bytes) peer=14 node0 2019-08-02T18:17:40.461279Z [httpworker.3] - Connect 1 transactions: 54.05ms (54.045ms/tx, 0.000ms/txin) [71.60s (35.98ms/blk)] node0 2019-08-02T18:17:40.461408Z [httpworker.3] - Verify 0 txins: 54.25ms (0.000ms/txin) [71.94s (36.15ms/blk)] node0 2019-08-02T18:17:40.461652Z [httpworker.3] - Index writing: 0.25ms [0.50s (0.25ms/blk)] node0 2019-08-02T18:17:40.461701Z [httpworker.3] - Callbacks: 0.05ms [0.10s (0.05ms/blk)] node0 2019-08-02T18:17:40.461781Z [httpworker.3] - Connect total: 54.98ms [73.14s (36.75ms/blk)] node0 2019-08-02T18:17:40.461831Z [httpworker.3] - Flush: 0.04ms [0.11s (0.06ms/blk)] node0 2019-08-02T18:17:40.461903Z [httpworker.3] - Writing chainstate: 0.08ms [0.16s (0.08ms/blk)] node0 2019-08-02T18:17:40.463359Z [httpworker.3] Blockpolicy estimates updated by 0 of 0 block txs, since last block 0 of 0 tracked, mempool map size 0, max target 0 from current node0 2019-08-02T18:17:40.463950Z [httpworker.3] UpdateTip: new best=6db3371c4f0ad084a24c512bb7a3c0de35280a5feb1037b6ba4ceae5bf579114 height=1552 version=0x20000000 log2_work=11.600842 tx=1553 date='2019-08-02T18:19:52Z' progress=1.000000 cache=0.1MiB(650txo) node0 2019-08-02T18:17:40.464006Z [httpworker.3] - Connect postprocess: 2.10ms [3.91s (1.96ms/blk)] node0 2019-08-02T18:17:40.464046Z [httpworker.3] - Connect block: 57.22ms [77.36s (38.88ms/blk)] node0 2019-08-02T18:17:40.464129Z [httpworker.3] Checking mempool with 0 transactions and 0 inputs test 2019-08-02T18:17:40.472000Z TestFramework (INFO): Usage should be below target: 560 node1 2019-08-02T18:17:40.486452Z [msghand] Successfully reconstructed block 6db3371c4f0ad084a24c512bb7a3c0de35280a5feb1037b6ba4ceae5bf579114 with 1 txn prefilled, 0 txn from mempool (incl at least 0 from extra pool) and 0 txn requested node1 2019-08-02T18:17:40.492254Z [msghand] PeerLogicValidation::NewPoWValidBlock sending header-and-ids 6db3371c4f0ad084a24c512bb7a3c0de35280a5feb1037b6ba4ceae5bf579114 to peer=15 node1 2019-08-02T18:17:40.493497Z [msghand] sending cmpctblock (950159 bytes) peer=15 node1 2019-08-02T18:17:40.516284Z [msghand] - Load block from disk: 0.01ms [0.01s] node1 2019-08-02T18:17:40.516417Z [msghand] - Sanity checks: 0.01ms [0.02s (0.01ms/blk)] node1 2019-08-02T18:17:40.516502Z [msghand] - Fork checks: 0.09ms [0.25s (0.11ms/blk)] node1 2019-08-02T18:17:40.568948Z [msghand] - Connect 1 transactions: 52.38ms (52.379ms/tx, 0.000ms/txin) [69.38s (29.65ms/blk)] node1 2019-08-02T18:17:40.569060Z [msghand] - Verify 0 txins: 52.55ms (0.000ms/txin) [69.83s (29.84ms/blk)] node1 2019-08-02T18:17:40.569290Z [msghand] - Index writing: 0.23ms [0.44s (0.19ms/blk)] node1 2019-08-02T18:17:40.569332Z [msghand] - Callbacks: 0.05ms [0.12s (0.05ms/blk)] node1 2019-08-02T18:17:40.569409Z [msghand] - Connect total: 53.17ms [70.93s (30.31ms/blk)] node1 2019-08-02T18:17:40.569454Z [msghand] - Flush: 0.04ms [0.10s (0.04ms/blk)] node1 2019-08-02T18:17:40.569527Z [msghand] - Writing chainstate: 0.07ms [0.14s (0.06ms/blk)] node1 2019-08-02T18:17:40.570938Z [msghand] Blockpolicy estimates updated by 0 of 0 block txs, since last block 0 of 0 tracked, mempool map size 0, max target 0 from current node1 2019-08-02T18:17:40.571506Z [msghand] UpdateTip: new best=6db3371c4f0ad084a24c512bb7a3c0de35280a5feb1037b6ba4ceae5bf579114 height=1552 version=0x20000000 log2_work=11.600842 tx=1553 date='2019-08-02T18:19:52Z' progress=1.000000 cache=0.1MiB(650txo) node1 2019-08-02T18:17:40.571560Z [msghand] - Connect postprocess: 2.03ms [3.66s (1.57ms/blk)] node1 2019-08-02T18:17:40.571599Z [msghand] - Connect block: 55.33ms [74.84s (31.98ms/blk)] node1 2019-08-02T18:17:40.571675Z [msghand] Checking mempool with 0 transactions and 0 inputs test 2019-08-02T18:17:43.227000Z TestFramework (ERROR): Assertion failed Traceback (most recent call last): File "/root/workspace/bitcoin-core/test/functional/test_framework/test_framework.py", line 193, in main self.run_test() File "/root/workspace/bitcoin-core/test/functional/feature_pruning.py", line 396, in run_test self.reorg_test() # (1033, ) File "/root/workspace/bitcoin-core/test/functional/feature_pruning.py", line 210, in reorg_test assert_greater_than(550, usage) File "/root/workspace/bitcoin-core/test/functional/test_framework/util.py", line 44, in assert_greater_than raise AssertionError("%s <= %s" % (str(thing1), str(thing2))) AssertionError: 550 <= 560.1718864440918 node2 2019-08-02T18:17:44.476610Z [msghand] Prune: UnlinkPrunedFiles deleted blk/rev (00004) node2 2019-08-02T18:17:44.476787Z [msghand] Writing final batch of 0.00 MiB node2 2019-08-02T18:17:44.477697Z [msghand] Committed 0 changed transaction outputs (out of 0) to coin database... node2 2019-08-02T18:17:44.484856Z [msghand] - Load block from disk: 0.01ms [0.04s] node2 2019-08-02T18:17:44.485366Z [msghand] - Sanity checks: 0.11ms [0.02s (0.01ms/blk)] node2 2019-08-02T18:17:44.485479Z [msghand] - Fork checks: 0.14ms [0.18s (0.10ms/blk)] test 2019-08-02T18:17:44.522000Z TestFramework (DEBUG): Closing down network thread node2 2019-08-02T18:17:44.544340Z [msghand] - Connect 1 transactions: 58.78ms (58.779ms/tx, 0.000ms/txin) [70.18s (38.69ms/blk)] node2 2019-08-02T18:17:44.544504Z [msghand] - Verify 0 txins: 59.02ms (0.000ms/txin) [70.48s (38.86ms/blk)] node2 2019-08-02T18:17:44.544743Z [msghand] - Index writing: 0.24ms [0.45s (0.25ms/blk)] node2 2019-08-02T18:17:44.544786Z [msghand] - Callbacks: 0.05ms [0.09s (0.05ms/blk)] node2 2019-08-02T18:17:44.544871Z [msghand] - Connect total: 60.10ms [71.70s (39.52ms/blk)] node2 2019-08-02T18:17:44.544915Z [msghand] - Flush: 0.05ms [0.09s (0.05ms/blk)] node2 2019-08-02T18:17:44.544985Z [msghand] - Writing chainstate: 0.07ms [0.14s (0.08ms/blk)] node2 2019-08-02T18:17:44.546402Z [msghand] Blockpolicy estimates updated by 0 of 0 block txs, since last block 0 of 0 tracked, mempool map size 0, max target 0 from current node2 2019-08-02T18:17:44.546973Z [msghand] UpdateTip: new best=3fa2e1b6704dc4c0ea7238313abe4242915ce18351455222a8510ad0fbefd19f height=1526 version=0x20000000 log2_work=11.576484 tx=1527 date='2019-08-02T18:19:26Z' progress=1.000000 cache=0.0MiB(0txo) node2 2019-08-02T18:17:44.547032Z [msghand] - Connect postprocess: 2.04ms [3.64s (2.01ms/blk)] node2 2019-08-02T18:17:44.547072Z [msghand] - Connect block: 62.27ms [75.61s (41.68ms/blk)] node2 2019-08-02T18:17:44.547154Z [msghand] Checking mempool with 0 transactions and 0 inputs node2 2019-08-02T18:17:44.554409Z [msghand] received: cmpctblock (950159 bytes) peer=14 node2 2019-08-02T18:17:44.575317Z [msghand] sending ping (8 bytes) peer=14 node1 2019-08-02T18:17:44.575900Z [msghand] received: ping (8 bytes) peer=15 node2 2019-08-02T18:17:44.576024Z [msghand] received: cmpctblock (950159 bytes) peer=13 node1 2019-08-02T18:17:44.576117Z [msghand] sending pong (8 bytes) peer=15 test 2019-08-02T18:17:44.587000Z TestFramework (INFO): Stopping nodes
test
test feature pruning fails with assertionerror intermittent current master excerpt received a post request for from threadrpcserver method submitblock user cookie peerlogicvalidation newpowvalidblock sending header and ids to peer sending cmpctblock bytes peer peerlogicvalidation newpowvalidblock sending header and ids to peer sending cmpctblock bytes peer load block from disk sanity checks fork checks received cmpctblock bytes peer initialized partiallydownloadedblock for block using a cmpctblock of size received blocktxn bytes peer connect transactions tx txin verify txins txin index writing callbacks connect total flush writing chainstate blockpolicy estimates updated by of block txs since last block of tracked mempool map size max target from current updatetip new best height version work tx date progress cache connect postprocess connect block checking mempool with transactions and inputs test testframework info usage should be below target successfully reconstructed block with txn prefilled txn from mempool incl at least from extra pool and txn requested peerlogicvalidation newpowvalidblock sending header and ids to peer sending cmpctblock bytes peer load block from disk sanity checks fork checks connect transactions tx txin verify txins txin index writing callbacks connect total flush writing chainstate blockpolicy estimates updated by of block txs since last block of tracked mempool map size max target from current updatetip new best height version work tx date progress cache connect postprocess connect block checking mempool with transactions and inputs test testframework error assertion failed traceback most recent call last file root workspace bitcoin core test functional test framework test framework py line in main self run test file root workspace bitcoin core test functional feature pruning py line in run test self reorg test file root workspace bitcoin core test functional feature pruning py line in reorg test assert greater than usage file root workspace bitcoin core test functional test framework util py line in assert greater than raise assertionerror s s str str assertionerror prune unlinkprunedfiles deleted blk rev writing final batch of mib committed changed transaction outputs out of to coin database load block from disk sanity checks fork checks test testframework debug closing down network thread connect transactions tx txin verify txins txin index writing callbacks connect total flush writing chainstate blockpolicy estimates updated by of block txs since last block of tracked mempool map size max target from current updatetip new best height version work tx date progress cache connect postprocess connect block checking mempool with transactions and inputs received cmpctblock bytes peer sending ping bytes peer received ping bytes peer received cmpctblock bytes peer sending pong bytes peer test testframework info stopping nodes
1
32,533
8,876,043,845
IssuesEvent
2019-01-12 10:59:04
App-vNext/Simmy
https://api.github.com/repos/App-vNext/Simmy
closed
Add cake build and nuget packaging
CI/build help wanted
#### Add cake build and nuget packaging Add cake build and packaging. After #2 . Can just copy the cake build from https://github.com/App-vNext/Polly/pull/504 and rename everything Polly to Simmy.
1.0
Add cake build and nuget packaging - #### Add cake build and nuget packaging Add cake build and packaging. After #2 . Can just copy the cake build from https://github.com/App-vNext/Polly/pull/504 and rename everything Polly to Simmy.
non_test
add cake build and nuget packaging add cake build and nuget packaging add cake build and packaging after can just copy the cake build from and rename everything polly to simmy
0
163,452
12,730,832,030
IssuesEvent
2020-06-25 08:06:36
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachtest: django failed
C-test-failure O-roachtest O-robot branch-provisional_202006230817_v20.1.3 release-blocker
[(roachtest).django failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2032995&tab=buildLog) on [provisional_202006230817_v20.1.3@7fd454f880f386cdd0eda6b21b12f6532c14f0db](https://github.com/cockroachdb/cockroach/commits/7fd454f880f386cdd0eda6b21b12f6532c14f0db): ``` The test failed on branch=provisional_202006230817_v20.1.3, cloud=gce: test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/django/run_1 test_runner.go:800: test timed out (10h0m0s) django.go:192,django.go:212,test_runner.go:753: cluster.RunL: context canceled (1) attached stack trace | main.(*cluster).RunL | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2121 | main.registerDjango.func1 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/django.go:189 | main.registerDjango.func2 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/django.go:212 | main.(*testRunner).runTest.func2 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/test_runner.go:753 | runtime.goexit | /usr/local/go/src/runtime/asm_amd64.s:1357 Wraps: (2) cluster.RunL Wraps: (3) context canceled Error types: (1) *withstack.withStack (2) *errutil.withMessage (3) *errors.errorString cluster.go:1512,context.go:135,cluster.go:1501,test_runner.go:825: dead node detection: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod monitor teamcity-2032995-1592933600-35-n1cpu16 --oneshot --ignore-empty-nodes: exit status 1 1: dead Error: UNCLASSIFIED_PROBLEM: 1: dead (1) UNCLASSIFIED_PROBLEM Wraps: (2) attached stack trace | main.glob..func13 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1115 | main.wrap.func1 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:266 | github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra.(*Command).execute | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:766 | github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra.(*Command).ExecuteC | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:852 | github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra.(*Command).Execute | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:800 | main.main | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1789 | runtime.main | /usr/local/go/src/runtime/proc.go:203 | runtime.goexit | /usr/local/go/src/runtime/asm_amd64.s:1357 Wraps: (3) 3 safe details enclosed Wraps: (4) 1: dead Error types: (1) errors.Unclassified (2) *withstack.withStack (3) *safedetails.withSafeDetails (4) *errors.errorString ``` <details><summary>More</summary><p> Artifacts: [/django](https://teamcity.cockroachdb.com/viewLog.html?buildId=2032995&tab=artifacts#/django) Related: - #50548 roachtest: django failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202006230817_v19.2.8](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202006230817_v19.2.8) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) - #50474 roachtest: django failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202006220937_v19.2.8](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202006220937_v19.2.8) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) - #50040 roachtest: django failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202006091546_v20.1.2](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202006091546_v20.1.2) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) - #46800 roachtest: django failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-20.1](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-20.1) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) - #44477 roachtest: django failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-19.2](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-19.2) [See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Adjango.%2A&sort=title&restgroup=false&display=lastcommented+project) <sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
2.0
roachtest: django failed - [(roachtest).django failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2032995&tab=buildLog) on [provisional_202006230817_v20.1.3@7fd454f880f386cdd0eda6b21b12f6532c14f0db](https://github.com/cockroachdb/cockroach/commits/7fd454f880f386cdd0eda6b21b12f6532c14f0db): ``` The test failed on branch=provisional_202006230817_v20.1.3, cloud=gce: test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/django/run_1 test_runner.go:800: test timed out (10h0m0s) django.go:192,django.go:212,test_runner.go:753: cluster.RunL: context canceled (1) attached stack trace | main.(*cluster).RunL | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2121 | main.registerDjango.func1 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/django.go:189 | main.registerDjango.func2 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/django.go:212 | main.(*testRunner).runTest.func2 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/test_runner.go:753 | runtime.goexit | /usr/local/go/src/runtime/asm_amd64.s:1357 Wraps: (2) cluster.RunL Wraps: (3) context canceled Error types: (1) *withstack.withStack (2) *errutil.withMessage (3) *errors.errorString cluster.go:1512,context.go:135,cluster.go:1501,test_runner.go:825: dead node detection: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod monitor teamcity-2032995-1592933600-35-n1cpu16 --oneshot --ignore-empty-nodes: exit status 1 1: dead Error: UNCLASSIFIED_PROBLEM: 1: dead (1) UNCLASSIFIED_PROBLEM Wraps: (2) attached stack trace | main.glob..func13 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1115 | main.wrap.func1 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:266 | github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra.(*Command).execute | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:766 | github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra.(*Command).ExecuteC | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:852 | github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra.(*Command).Execute | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:800 | main.main | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1789 | runtime.main | /usr/local/go/src/runtime/proc.go:203 | runtime.goexit | /usr/local/go/src/runtime/asm_amd64.s:1357 Wraps: (3) 3 safe details enclosed Wraps: (4) 1: dead Error types: (1) errors.Unclassified (2) *withstack.withStack (3) *safedetails.withSafeDetails (4) *errors.errorString ``` <details><summary>More</summary><p> Artifacts: [/django](https://teamcity.cockroachdb.com/viewLog.html?buildId=2032995&tab=artifacts#/django) Related: - #50548 roachtest: django failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202006230817_v19.2.8](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202006230817_v19.2.8) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) - #50474 roachtest: django failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202006220937_v19.2.8](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202006220937_v19.2.8) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) - #50040 roachtest: django failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202006091546_v20.1.2](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202006091546_v20.1.2) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) - #46800 roachtest: django failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-20.1](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-20.1) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) - #44477 roachtest: django failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-19.2](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-19.2) [See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Adjango.%2A&sort=title&restgroup=false&display=lastcommented+project) <sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
test
roachtest django failed on the test failed on branch provisional cloud gce test artifacts and logs in home agent work go src github com cockroachdb cockroach artifacts django run test runner go test timed out django go django go test runner go cluster runl context canceled attached stack trace main cluster runl home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go main registerdjango home agent work go src github com cockroachdb cockroach pkg cmd roachtest django go main registerdjango home agent work go src github com cockroachdb cockroach pkg cmd roachtest django go main testrunner runtest home agent work go src github com cockroachdb cockroach pkg cmd roachtest test runner go runtime goexit usr local go src runtime asm s wraps cluster runl wraps context canceled error types withstack withstack errutil withmessage errors errorstring cluster go context go cluster go test runner go dead node detection home agent work go src github com cockroachdb cockroach bin roachprod monitor teamcity oneshot ignore empty nodes exit status dead error unclassified problem dead unclassified problem wraps attached stack trace main glob home agent work go src github com cockroachdb cockroach pkg cmd roachprod main go main wrap home agent work go src github com cockroachdb cockroach pkg cmd roachprod main go github com cockroachdb cockroach vendor github com cobra command execute home agent work go src github com cockroachdb cockroach vendor github com cobra command go github com cockroachdb cockroach vendor github com cobra command executec home agent work go src github com cockroachdb cockroach vendor github com cobra command go github com cockroachdb cockroach vendor github com cobra command execute home agent work go src github com cockroachdb cockroach vendor github com cobra command go main main home agent work go src github com cockroachdb cockroach pkg cmd roachprod main go runtime main usr local go src runtime proc go runtime goexit usr local go src runtime asm s wraps safe details enclosed wraps dead error types errors unclassified withstack withstack safedetails withsafedetails errors errorstring more artifacts related roachtest django failed roachtest django failed roachtest django failed roachtest django failed roachtest django failed powered by
1
275,585
23,923,591,827
IssuesEvent
2022-09-09 19:35:59
ValveSoftware/steam-for-linux
https://api.github.com/repos/ValveSoftware/steam-for-linux
closed
Microstutters every second
Need Retest
#### Your system information using nvidia-440.59 * Steam client version (build number or date): feb 11 2020 * Distribution (e.g. Ubuntu): void * Opted into Steam client beta?: [Yes/No] no * Have you checked for system updates?: [Yes/No] yes #### Please describe your issue in as much detail as possible: My steam client and some games hang every second. I've had this issue for ~10 days. The hang only lasts for a little while (tens / hundred ms), but it makes games unplayable. I have encountered this in CS;GO, and 7Days to die, while Warframe runs perfectly fine. I have tried different settings in the nvidia control panel, and switched my cpu frequency governor to 'performance', but nothing has fixed this so far.
1.0
Microstutters every second - #### Your system information using nvidia-440.59 * Steam client version (build number or date): feb 11 2020 * Distribution (e.g. Ubuntu): void * Opted into Steam client beta?: [Yes/No] no * Have you checked for system updates?: [Yes/No] yes #### Please describe your issue in as much detail as possible: My steam client and some games hang every second. I've had this issue for ~10 days. The hang only lasts for a little while (tens / hundred ms), but it makes games unplayable. I have encountered this in CS;GO, and 7Days to die, while Warframe runs perfectly fine. I have tried different settings in the nvidia control panel, and switched my cpu frequency governor to 'performance', but nothing has fixed this so far.
test
microstutters every second your system information using nvidia steam client version build number or date feb distribution e g ubuntu void opted into steam client beta no have you checked for system updates yes please describe your issue in as much detail as possible my steam client and some games hang every second i ve had this issue for days the hang only lasts for a little while tens hundred ms but it makes games unplayable i have encountered this in cs go and to die while warframe runs perfectly fine i have tried different settings in the nvidia control panel and switched my cpu frequency governor to performance but nothing has fixed this so far
1
176,458
13,642,831,326
IssuesEvent
2020-09-25 16:06:53
wazuh/wazuh-qa
https://api.github.com/repos/wazuh/wazuh-qa
closed
IT Wazuh-logtest - Invalid syntax tests
core/logtest
Hello team! To test Wazuh-logtest can treat invalid syntax of rules and decoders, we implement a test that loads invalid rules and decoders in the logtest (writing them in `local_rules.xml` and `local_decoders.xml` and creating logtest client session). Then, check error or warning message is expected. Best regards, Eva
1.0
IT Wazuh-logtest - Invalid syntax tests - Hello team! To test Wazuh-logtest can treat invalid syntax of rules and decoders, we implement a test that loads invalid rules and decoders in the logtest (writing them in `local_rules.xml` and `local_decoders.xml` and creating logtest client session). Then, check error or warning message is expected. Best regards, Eva
test
it wazuh logtest invalid syntax tests hello team to test wazuh logtest can treat invalid syntax of rules and decoders we implement a test that loads invalid rules and decoders in the logtest writing them in local rules xml and local decoders xml and creating logtest client session then check error or warning message is expected best regards eva
1
57,937
8,214,584,059
IssuesEvent
2018-09-05 00:12:43
EBWiki/EBWiki
https://api.github.com/repos/EBWiki/EBWiki
closed
Add section on agencies to Guide to Contributing
documentation enhancement
Note: This is issue 3 of 3, and the requested changes must be merged before you can move to the next one in the project. In the Guide to Contributing to EBWiki, add a section titled "Agencies". This section should contain two subsections: one for adding an agency and one for updating an agency. Have a heading for the section, and collapse the subsections by default - selecting the corresponding heading should expand and collapse the subsection.
1.0
Add section on agencies to Guide to Contributing - Note: This is issue 3 of 3, and the requested changes must be merged before you can move to the next one in the project. In the Guide to Contributing to EBWiki, add a section titled "Agencies". This section should contain two subsections: one for adding an agency and one for updating an agency. Have a heading for the section, and collapse the subsections by default - selecting the corresponding heading should expand and collapse the subsection.
non_test
add section on agencies to guide to contributing note this is issue of and the requested changes must be merged before you can move to the next one in the project in the guide to contributing to ebwiki add a section titled agencies this section should contain two subsections one for adding an agency and one for updating an agency have a heading for the section and collapse the subsections by default selecting the corresponding heading should expand and collapse the subsection
0
241,330
20,116,944,474
IssuesEvent
2022-02-07 20:36:19
cagov/design-system
https://api.github.com/repos/cagov/design-system
closed
Add sitewide tests to design system site
Testing P1 Accessibility
- Should load a url representing each template on the site and run axe accessibility checker on them
1.0
Add sitewide tests to design system site - - Should load a url representing each template on the site and run axe accessibility checker on them
test
add sitewide tests to design system site should load a url representing each template on the site and run axe accessibility checker on them
1
8,955
8,482,618,496
IssuesEvent
2018-10-25 19:03:28
terraform-providers/terraform-provider-aws
https://api.github.com/repos/terraform-providers/terraform-provider-aws
closed
aws_flow_log - add support for sending to S3
enhancement service/ec2
<!--- Please keep this note for the community ---> ### Community Note * Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Description Support S3 as a delivery destination for VPC Flow Logs. As of 2018-08-08, AWS announced support for [sending vpc flow logs to s3](https://aws.amazon.com/about-aws/whats-new/2018/08/amazon-vpc-flow-logs-can-now-be-delivered-to-s3/) (as oppose to the currently supported CloudWatch Logs group). The [API docs](https://docs.aws.amazon.com/AWSEC2/latest/APIReference/API_CreateFlowLogs.html) are updated to include this support, and the [`aws-sdk-go`](https://github.com/aws/aws-sdk-go/pull/2091/) has been updated to support this. The interface has changed a bit for this API, and different parameters are now used. ### New or Affected Resource(s) * aws_flow_log ### Potential Terraform Configuration <!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code ---> ```hcl resource "aws_flow_log" "test_flow_log" { log_destination_type = "s3" log_destination = "${aws_s3_bucket.test_bucket.arn}" iam_role_arn = "${aws_iam_role.test_role.arn}" resource_type = "VPC" resource_ids = ["${aws_vpc.default.id}"] traffic_type = "ALL" } ``` ### References * N/A
1.0
aws_flow_log - add support for sending to S3 - <!--- Please keep this note for the community ---> ### Community Note * Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Description Support S3 as a delivery destination for VPC Flow Logs. As of 2018-08-08, AWS announced support for [sending vpc flow logs to s3](https://aws.amazon.com/about-aws/whats-new/2018/08/amazon-vpc-flow-logs-can-now-be-delivered-to-s3/) (as oppose to the currently supported CloudWatch Logs group). The [API docs](https://docs.aws.amazon.com/AWSEC2/latest/APIReference/API_CreateFlowLogs.html) are updated to include this support, and the [`aws-sdk-go`](https://github.com/aws/aws-sdk-go/pull/2091/) has been updated to support this. The interface has changed a bit for this API, and different parameters are now used. ### New or Affected Resource(s) * aws_flow_log ### Potential Terraform Configuration <!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code ---> ```hcl resource "aws_flow_log" "test_flow_log" { log_destination_type = "s3" log_destination = "${aws_s3_bucket.test_bucket.arn}" iam_role_arn = "${aws_iam_role.test_role.arn}" resource_type = "VPC" resource_ids = ["${aws_vpc.default.id}"] traffic_type = "ALL" } ``` ### References * N/A
non_test
aws flow log add support for sending to community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment description support as a delivery destination for vpc flow logs as of aws announced support for as oppose to the currently supported cloudwatch logs group the are updated to include this support and the has been updated to support this the interface has changed a bit for this api and different parameters are now used new or affected resource s aws flow log potential terraform configuration hcl resource aws flow log test flow log log destination type log destination aws bucket test bucket arn iam role arn aws iam role test role arn resource type vpc resource ids traffic type all references n a
0
208,194
16,106,035,723
IssuesEvent
2021-04-27 15:02:00
empress-php/empress
https://api.github.com/repos/empress-php/empress
opened
Documentation
documentation enhancement help wanted
I need to start creating documentation for the project. Either in-repo wiki or GH pages seem like a good fit.
1.0
Documentation - I need to start creating documentation for the project. Either in-repo wiki or GH pages seem like a good fit.
non_test
documentation i need to start creating documentation for the project either in repo wiki or gh pages seem like a good fit
0
452,354
32,057,743,971
IssuesEvent
2023-09-24 09:39:54
privacy-scaling-explorations/bandada
https://api.github.com/repos/privacy-scaling-explorations/bandada
closed
Add the `x-api-key` parameter as required in the API Docs.
documentation :book: refactoring :recycle:
Add the `x-api-key` parameter as required in `addMembers` in the API Docs.
1.0
Add the `x-api-key` parameter as required in the API Docs. - Add the `x-api-key` parameter as required in `addMembers` in the API Docs.
non_test
add the x api key parameter as required in the api docs add the x api key parameter as required in addmembers in the api docs
0
338,314
30,291,975,333
IssuesEvent
2023-07-09 11:53:48
Realm667/WolfenDoom
https://api.github.com/repos/Realm667/WolfenDoom
closed
invisible wall
playtesting mapping
In c3m6_b, when you cross this line (just in front of the gun), there's a big slowdown as if you bump into an invicible wall. This is mostly the first time when go through this passage, but can happen when you walk here later on. ![Screenshot_Doom_20230519_162730](https://github.com/Realm667/WolfenDoom/assets/77861029/6000f6a0-0c1b-41a4-a85e-af385e404843)
1.0
invisible wall - In c3m6_b, when you cross this line (just in front of the gun), there's a big slowdown as if you bump into an invicible wall. This is mostly the first time when go through this passage, but can happen when you walk here later on. ![Screenshot_Doom_20230519_162730](https://github.com/Realm667/WolfenDoom/assets/77861029/6000f6a0-0c1b-41a4-a85e-af385e404843)
test
invisible wall in b when you cross this line just in front of the gun there s a big slowdown as if you bump into an invicible wall this is mostly the first time when go through this passage but can happen when you walk here later on
1
190,723
14,570,416,074
IssuesEvent
2020-12-17 14:21:56
zeebe-io/zeebe
https://api.github.com/repos/zeebe-io/zeebe
closed
RaftFailOverTest.shouldNotJoinAfterDataLoss flaky
Priority: Mid Status: Planned Type: Unstable Test
**Summary** - How often does the test fail? at least once so far - Does it block your work? no - Do we suspect that it is a real failure? unknwon **Failures** <details><summary>Example assertion failure</summary> <pre> java.util.concurrent.TimeoutException at java.base/java.util.concurrent.CompletableFuture.timedGet(CompletableFuture.java:1886) at java.base/java.util.concurrent.CompletableFuture.get(CompletableFuture.java:2021) at io.atomix.raft.RaftRule$TestAppendListener.awaitCommit(RaftRule.java:643) at io.atomix.raft.RaftRule.appendEntry(RaftRule.java:542) at io.atomix.raft.RaftRule.appendEntry(RaftRule.java:535) at io.atomix.raft.RaftRule.appendEntries(RaftRule.java:526) at io.atomix.raft.RaftFailOverTest.shouldReplicateEntriesAfterSnapshotAfterDataLoss(RaftFailOverTest.java:267) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54) at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.apache.maven.surefire.junitcore.JUnitCore.run(JUnitCore.java:55) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.createRequestAndRun(JUnitCoreWrapper.java:137) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.executeLazy(JUnitCoreWrapper.java:119) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:87) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:75) at org.apache.maven.surefire.junitcore.JUnitCoreProvider.invoke(JUnitCoreProvider.java:157) at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:428) at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162) at org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:562) at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:548) </pre> </details> **Hypotheses** **Logs** [io.atomix.raft.RaftFailOverTest-java-output.zip](https://github.com/zeebe-io/zeebe/files/5015403/io.atomix.raft.RaftFailOverTest-java-output.zip)
1.0
RaftFailOverTest.shouldNotJoinAfterDataLoss flaky - **Summary** - How often does the test fail? at least once so far - Does it block your work? no - Do we suspect that it is a real failure? unknwon **Failures** <details><summary>Example assertion failure</summary> <pre> java.util.concurrent.TimeoutException at java.base/java.util.concurrent.CompletableFuture.timedGet(CompletableFuture.java:1886) at java.base/java.util.concurrent.CompletableFuture.get(CompletableFuture.java:2021) at io.atomix.raft.RaftRule$TestAppendListener.awaitCommit(RaftRule.java:643) at io.atomix.raft.RaftRule.appendEntry(RaftRule.java:542) at io.atomix.raft.RaftRule.appendEntry(RaftRule.java:535) at io.atomix.raft.RaftRule.appendEntries(RaftRule.java:526) at io.atomix.raft.RaftFailOverTest.shouldReplicateEntriesAfterSnapshotAfterDataLoss(RaftFailOverTest.java:267) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54) at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.apache.maven.surefire.junitcore.JUnitCore.run(JUnitCore.java:55) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.createRequestAndRun(JUnitCoreWrapper.java:137) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.executeLazy(JUnitCoreWrapper.java:119) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:87) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:75) at org.apache.maven.surefire.junitcore.JUnitCoreProvider.invoke(JUnitCoreProvider.java:157) at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:428) at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162) at org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:562) at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:548) </pre> </details> **Hypotheses** **Logs** [io.atomix.raft.RaftFailOverTest-java-output.zip](https://github.com/zeebe-io/zeebe/files/5015403/io.atomix.raft.RaftFailOverTest-java-output.zip)
test
raftfailovertest shouldnotjoinafterdataloss flaky summary how often does the test fail at least once so far does it block your work no do we suspect that it is a real failure unknwon failures example assertion failure java util concurrent timeoutexception at java base java util concurrent completablefuture timedget completablefuture java at java base java util concurrent completablefuture get completablefuture java at io atomix raft raftrule testappendlistener awaitcommit raftrule java at io atomix raft raftrule appendentry raftrule java at io atomix raft raftrule appendentry raftrule java at io atomix raft raftrule appendentries raftrule java at io atomix raft raftfailovertest shouldreplicateentriesaftersnapshotafterdataloss raftfailovertest java at java base jdk internal reflect nativemethodaccessorimpl native method at java base jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at java base jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java base java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at org junit rules externalresource evaluate externalresource java at org junit rules externalresource evaluate externalresource java at org junit runners parentrunner evaluate parentrunner java at org junit runners evaluate java at org junit runners parentrunner runleaf parentrunner java at org junit runners runchild java at org junit runners runchild java at org junit runners parentrunner run parentrunner java at org junit runners parentrunner schedule parentrunner java at org junit runners parentrunner runchildren parentrunner java at org junit runners parentrunner access parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner run parentrunner java at org junit runners suite runchild suite java at org junit runners suite runchild suite java at org junit runners parentrunner run parentrunner java at org junit runners parentrunner schedule parentrunner java at org junit runners parentrunner runchildren parentrunner java at org junit runners parentrunner access parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner run parentrunner java at org junit runners suite runchild suite java at org junit runners suite runchild suite java at org junit runners parentrunner run parentrunner java at org junit runners parentrunner schedule parentrunner java at org junit runners parentrunner runchildren parentrunner java at org junit runners parentrunner access parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner run parentrunner java at org apache maven surefire junitcore junitcore run junitcore java at org apache maven surefire junitcore junitcorewrapper createrequestandrun junitcorewrapper java at org apache maven surefire junitcore junitcorewrapper executelazy junitcorewrapper java at org apache maven surefire junitcore junitcorewrapper execute junitcorewrapper java at org apache maven surefire junitcore junitcorewrapper execute junitcorewrapper java at org apache maven surefire junitcore junitcoreprovider invoke junitcoreprovider java at org apache maven surefire booter forkedbooter runsuitesinprocess forkedbooter java at org apache maven surefire booter forkedbooter execute forkedbooter java at org apache maven surefire booter forkedbooter run forkedbooter java at org apache maven surefire booter forkedbooter main forkedbooter java hypotheses logs
1
26,007
4,188,399,996
IssuesEvent
2016-06-23 20:37:32
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
closed
System.Net.Http.Functional.Tests.PostScenarioTest.PostUsingUsingConflictingSemantics_UsesChunkedSemantics failed in CI
System.Net test bug
http://dotnet-ci.cloudapp.net/job/dotnet_corefx/job/master/job/windows_nt_release_prtest/1179/testReport/junit/System.Net.Http.Functional.Tests/PostScenarioTest/PostUsingUsingConflictingSemantics_UsesChunkedSemantics_serverUri__https___corefx_net_cloudapp_net_Echo_ashx_/ >MESSAGE: System.Net.Http.HttpRequestException : An error occurred while sending the request.\r\n---- System.Net.Http.WinHttpException : The operation has been canceled +++++++++++++++++++ STACK TRACE: at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task) at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task) at System.Net.Http.HttpClient.<FinishSendAsync>d__58.MoveNext() in D:\j\workspace\windows_nt_re---37265eab\src\System.Net.Http\src\System\Net\Http\HttpClient.cs:line 392 --- End of stack trace from previous location where exception was thrown --- at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task) at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task) at System.Net.Http.Functional.Tests.PostScenarioTest.<PostHelper>d__22.MoveNext() in D:\j\workspace\windows_nt_re---37265eab\src\System.Net.Http\tests\FunctionalTests\PostScenarioTest.cs:line 171 --- End of stack trace from previous location where exception was thrown --- at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task) at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task) at System.Net.Http.Functional.Tests.PostScenarioTest.<PostUsingUsingConflictingSemantics_UsesChunkedSemantics>d__16.MoveNext() in D:\j\workspace\windows_nt_re---37265eab\src\System.Net.Http\tests\FunctionalTests\PostScenarioTest.cs:line 99 --- End of stack trace from previous location where exception was thrown --- at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task) at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task) --- End of stack trace from previous location where exception was thrown --- at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task) at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task) --- End of stack trace from previous location where exception was thrown --- at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task) at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task) ----- Inner Stack Trace ----- at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task) at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task) at System.Net.Http.WinHttpHandler.<StartRequest>d__102.MoveNext() in D:\j\workspace\windows_nt_re---37265eab\src\System.Net.Http.WinHttpHandler\src\System\Net\Http\WinHttpHandler.cs:line 843
1.0
System.Net.Http.Functional.Tests.PostScenarioTest.PostUsingUsingConflictingSemantics_UsesChunkedSemantics failed in CI - http://dotnet-ci.cloudapp.net/job/dotnet_corefx/job/master/job/windows_nt_release_prtest/1179/testReport/junit/System.Net.Http.Functional.Tests/PostScenarioTest/PostUsingUsingConflictingSemantics_UsesChunkedSemantics_serverUri__https___corefx_net_cloudapp_net_Echo_ashx_/ >MESSAGE: System.Net.Http.HttpRequestException : An error occurred while sending the request.\r\n---- System.Net.Http.WinHttpException : The operation has been canceled +++++++++++++++++++ STACK TRACE: at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task) at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task) at System.Net.Http.HttpClient.<FinishSendAsync>d__58.MoveNext() in D:\j\workspace\windows_nt_re---37265eab\src\System.Net.Http\src\System\Net\Http\HttpClient.cs:line 392 --- End of stack trace from previous location where exception was thrown --- at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task) at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task) at System.Net.Http.Functional.Tests.PostScenarioTest.<PostHelper>d__22.MoveNext() in D:\j\workspace\windows_nt_re---37265eab\src\System.Net.Http\tests\FunctionalTests\PostScenarioTest.cs:line 171 --- End of stack trace from previous location where exception was thrown --- at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task) at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task) at System.Net.Http.Functional.Tests.PostScenarioTest.<PostUsingUsingConflictingSemantics_UsesChunkedSemantics>d__16.MoveNext() in D:\j\workspace\windows_nt_re---37265eab\src\System.Net.Http\tests\FunctionalTests\PostScenarioTest.cs:line 99 --- End of stack trace from previous location where exception was thrown --- at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task) at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task) --- End of stack trace from previous location where exception was thrown --- at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task) at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task) --- End of stack trace from previous location where exception was thrown --- at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task) at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task) ----- Inner Stack Trace ----- at System.Runtime.CompilerServices.TaskAwaiter.ThrowForNonSuccess(Task task) at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task) at System.Net.Http.WinHttpHandler.<StartRequest>d__102.MoveNext() in D:\j\workspace\windows_nt_re---37265eab\src\System.Net.Http.WinHttpHandler\src\System\Net\Http\WinHttpHandler.cs:line 843
test
system net http functional tests postscenariotest postusingusingconflictingsemantics useschunkedsemantics failed in ci message system net http httprequestexception an error occurred while sending the request r n system net http winhttpexception the operation has been canceled stack trace at system runtime compilerservices taskawaiter throwfornonsuccess task task at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task task at system net http httpclient d movenext in d j workspace windows nt re src system net http src system net http httpclient cs line end of stack trace from previous location where exception was thrown at system runtime compilerservices taskawaiter throwfornonsuccess task task at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task task at system net http functional tests postscenariotest d movenext in d j workspace windows nt re src system net http tests functionaltests postscenariotest cs line end of stack trace from previous location where exception was thrown at system runtime compilerservices taskawaiter throwfornonsuccess task task at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task task at system net http functional tests postscenariotest d movenext in d j workspace windows nt re src system net http tests functionaltests postscenariotest cs line end of stack trace from previous location where exception was thrown at system runtime compilerservices taskawaiter throwfornonsuccess task task at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task task end of stack trace from previous location where exception was thrown at system runtime compilerservices taskawaiter throwfornonsuccess task task at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task task end of stack trace from previous location where exception was thrown at system runtime compilerservices taskawaiter throwfornonsuccess task task at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task task inner stack trace at system runtime compilerservices taskawaiter throwfornonsuccess task task at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task task at system net http winhttphandler d movenext in d j workspace windows nt re src system net http winhttphandler src system net http winhttphandler cs line
1
72,894
7,314,640,885
IssuesEvent
2018-03-01 08:11:03
researchstudio-sat/webofneeds
https://api.github.com/repos/researchstudio-sat/webofneeds
closed
Matching is too eager
bug testing
When no location is given, finds almost anything (on integration-test), also very lax even with specified location
1.0
Matching is too eager - When no location is given, finds almost anything (on integration-test), also very lax even with specified location
test
matching is too eager when no location is given finds almost anything on integration test also very lax even with specified location
1
345,194
30,792,641,391
IssuesEvent
2023-07-31 17:19:30
MD-Anderson-Bioinformatics/NG-CHM
https://api.github.com/repos/MD-Anderson-Bioinformatics/NG-CHM
closed
Undefined plugin error regularly generated with particular sequence of multiple saves of heatmap
bug passed retest show-stopper 2.23.1
If a map is created with a panels for 2D and 3D heatmaps has all panels but those closed is saved then is reloaded, it will load as expected with no errors on the screen or console. If you then open panels beneath those panels and display detail and summary maps below them, they will display correctly. If you save that map (4 panels) and try to reload it, the two scatter plot panes will show up as undefined. There will be no errors in the console, but both of the scatterplot panels will display as blank with the following text: "This restored panel requires the "undefined" plugin, which is not available. Adding the plugin will initialize the panel."
1.0
Undefined plugin error regularly generated with particular sequence of multiple saves of heatmap - If a map is created with a panels for 2D and 3D heatmaps has all panels but those closed is saved then is reloaded, it will load as expected with no errors on the screen or console. If you then open panels beneath those panels and display detail and summary maps below them, they will display correctly. If you save that map (4 panels) and try to reload it, the two scatter plot panes will show up as undefined. There will be no errors in the console, but both of the scatterplot panels will display as blank with the following text: "This restored panel requires the "undefined" plugin, which is not available. Adding the plugin will initialize the panel."
test
undefined plugin error regularly generated with particular sequence of multiple saves of heatmap if a map is created with a panels for and heatmaps has all panels but those closed is saved then is reloaded it will load as expected with no errors on the screen or console if you then open panels beneath those panels and display detail and summary maps below them they will display correctly if you save that map panels and try to reload it the two scatter plot panes will show up as undefined there will be no errors in the console but both of the scatterplot panels will display as blank with the following text this restored panel requires the undefined plugin which is not available adding the plugin will initialize the panel
1
205,880
15,696,867,054
IssuesEvent
2021-03-26 03:05:50
microsoft/AzureStorageExplorer
https://api.github.com/repos/microsoft/AzureStorageExplorer
closed
Miss the legal hold properties and policy properties for one new created blob container
:gear: blobs :heavy_check_mark: merged 🧪 testing
**Storage Explorer Version:** 1.19.0-dev **Build Number**: 20210323.4 **Branch**: main **Platform/OS**: Windows 10/ Linux Ubuntu 18.04/ MacOS Catalina **Architecture**: ia32/x64 **Regression From**: Not a regression ## Steps to Reproduce ## 1. Expand one storage account -> Blob Containers. 2. Create a new blob container -> Observe its properties in the 'Properties' panel. 3. Check the legal hold properties and policy properties show well. ## Expected Experience ## The legal hold properties and policy properties show well. ## Actual Experience ## Miss the legal hold properties and policy properties. ## Additional Context ## 1. This issue also reproduces for one cloned blob container. 2. This issue doesn't reproduce after manually refreshing.
1.0
Miss the legal hold properties and policy properties for one new created blob container - **Storage Explorer Version:** 1.19.0-dev **Build Number**: 20210323.4 **Branch**: main **Platform/OS**: Windows 10/ Linux Ubuntu 18.04/ MacOS Catalina **Architecture**: ia32/x64 **Regression From**: Not a regression ## Steps to Reproduce ## 1. Expand one storage account -> Blob Containers. 2. Create a new blob container -> Observe its properties in the 'Properties' panel. 3. Check the legal hold properties and policy properties show well. ## Expected Experience ## The legal hold properties and policy properties show well. ## Actual Experience ## Miss the legal hold properties and policy properties. ## Additional Context ## 1. This issue also reproduces for one cloned blob container. 2. This issue doesn't reproduce after manually refreshing.
test
miss the legal hold properties and policy properties for one new created blob container storage explorer version dev build number branch main platform os windows linux ubuntu macos catalina architecture regression from not a regression steps to reproduce expand one storage account blob containers create a new blob container observe its properties in the properties panel check the legal hold properties and policy properties show well expected experience the legal hold properties and policy properties show well actual experience miss the legal hold properties and policy properties additional context this issue also reproduces for one cloned blob container this issue doesn t reproduce after manually refreshing
1
39,064
12,627,480,482
IssuesEvent
2020-06-14 21:43:41
Vachok/ftpplus
https://api.github.com/repos/Vachok/ftpplus
closed
CVE-2019-14892 (High) detected in jackson-databind-2.9.6.jar
security vulnerability
## CVE-2019-14892 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.6.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /tmp/ws-scm/ftpplus/modules/networker/build.gradle</p> <p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.6/cfa4f316351a91bfd95cb0644c6a2c95f52db1fc/jackson-databind-2.9.6.jar,/root/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.6/cfa4f316351a91bfd95cb0644c6a2c95f52db1fc/jackson-databind-2.9.6.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.0.4.RELEASE.jar (Root Library) - spring-boot-starter-json-2.0.4.RELEASE.jar - :x: **jackson-databind-2.9.6.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Vachok/ftpplus/commit/e405ceee35d80b48d7dc32b60490d0289b30fe77">e405ceee35d80b48d7dc32b60490d0289b30fe77</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A flaw was discovered in jackson-databind in versions before 2.9.10, 2.8.11.5 and 2.6.7.3, where it would permit polymorphic deserialization of a malicious object using commons-configuration 1 and 2 JNDI classes. An attacker could use this flaw to execute arbitrary code. <p>Publish Date: 2020-03-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14892>CVE-2019-14892</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-14892">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-14892</a></p> <p>Release Date: 2020-03-02</p> <p>Fix Resolution: com.fasterxml.jackson:jackson-base:jackson-databind-2.8.11.5,com.fasterxml.jackson:jackson-base:jackson-databind-2.9.10,com.fasterxml.jackson:jackson-base:jackson-databind-2.10.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-14892 (High) detected in jackson-databind-2.9.6.jar - ## CVE-2019-14892 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.6.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /tmp/ws-scm/ftpplus/modules/networker/build.gradle</p> <p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.6/cfa4f316351a91bfd95cb0644c6a2c95f52db1fc/jackson-databind-2.9.6.jar,/root/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.6/cfa4f316351a91bfd95cb0644c6a2c95f52db1fc/jackson-databind-2.9.6.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.0.4.RELEASE.jar (Root Library) - spring-boot-starter-json-2.0.4.RELEASE.jar - :x: **jackson-databind-2.9.6.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Vachok/ftpplus/commit/e405ceee35d80b48d7dc32b60490d0289b30fe77">e405ceee35d80b48d7dc32b60490d0289b30fe77</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A flaw was discovered in jackson-databind in versions before 2.9.10, 2.8.11.5 and 2.6.7.3, where it would permit polymorphic deserialization of a malicious object using commons-configuration 1 and 2 JNDI classes. An attacker could use this flaw to execute arbitrary code. <p>Publish Date: 2020-03-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14892>CVE-2019-14892</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-14892">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-14892</a></p> <p>Release Date: 2020-03-02</p> <p>Fix Resolution: com.fasterxml.jackson:jackson-base:jackson-databind-2.8.11.5,com.fasterxml.jackson:jackson-base:jackson-databind-2.9.10,com.fasterxml.jackson:jackson-base:jackson-databind-2.10.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file tmp ws scm ftpplus modules networker build gradle path to vulnerable library root gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar root gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring boot starter web release jar root library spring boot starter json release jar x jackson databind jar vulnerable library found in head commit a href vulnerability details a flaw was discovered in jackson databind in versions before and where it would permit polymorphic deserialization of a malicious object using commons configuration and jndi classes an attacker could use this flaw to execute arbitrary code publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson jackson base jackson databind com fasterxml jackson jackson base jackson databind com fasterxml jackson jackson base jackson databind step up your open source security game with whitesource
0
54,921
6,417,620,098
IssuesEvent
2017-08-08 17:09:24
vmware/vic
https://api.github.com/repos/vmware/vic
closed
Volume Test Case for Inspect-Restart VCH-Inspect
component/test priority/high
As an engineer, I want to make sure that re-starting a VCH will produce the same inspect information on a volume (and NFS volume as well). **Acceptance Criteria:** Add test case for standard and NFS volumes where we compare the output from a volume using inspect, restart the VCH, and run inspect again. Info should be the same.
1.0
Volume Test Case for Inspect-Restart VCH-Inspect - As an engineer, I want to make sure that re-starting a VCH will produce the same inspect information on a volume (and NFS volume as well). **Acceptance Criteria:** Add test case for standard and NFS volumes where we compare the output from a volume using inspect, restart the VCH, and run inspect again. Info should be the same.
test
volume test case for inspect restart vch inspect as an engineer i want to make sure that re starting a vch will produce the same inspect information on a volume and nfs volume as well acceptance criteria add test case for standard and nfs volumes where we compare the output from a volume using inspect restart the vch and run inspect again info should be the same
1
327,448
28,064,284,503
IssuesEvent
2023-03-29 14:28:05
zephyrproject-rtos/test_results
https://api.github.com/repos/zephyrproject-rtos/test_results
closed
IPv4 Fragment Fragment: 512 MTU. Order: 1 2 3 ... N. Overlap: By 16 bytes. error
area: Tests
**Describe the bug** Fragment: 512 MTU. Order: 1 2 3 ... N. Overlap: By 16 bytes. test is Fail on Zephyr3.0.0 on qemu_x86 **References** RFC 791: section 2.3 RFC 791: section 3.2 RFC 1122: section 3.3.2 {Able to reassemble incoming datagrams} RFC 1122: section 3.2.1.4 {Support reassembly} RFC 1122: section 3.3.2 {At least 576 byte datagrams} **Results** FAIL: icmp.v4 got no echo response, which is not an expected result.; SKIPPED: udp.v4: Packets exchanged with the DUT did not match those needed to either perform or grade the test. (Traffic source reports: echo response.); SKIPPED: tcp.v4: Packets exchanged with the DUT did not match those needed to either perform or grade the test. (Traffic source reports: echo response.) **Environment (please complete the following information):** - OS: (e.g. Linux ) - Toolchain (e.g Zephyr SDK) - Commit SHA or Version used: Zephyr3.0.0
1.0
IPv4 Fragment Fragment: 512 MTU. Order: 1 2 3 ... N. Overlap: By 16 bytes. error - **Describe the bug** Fragment: 512 MTU. Order: 1 2 3 ... N. Overlap: By 16 bytes. test is Fail on Zephyr3.0.0 on qemu_x86 **References** RFC 791: section 2.3 RFC 791: section 3.2 RFC 1122: section 3.3.2 {Able to reassemble incoming datagrams} RFC 1122: section 3.2.1.4 {Support reassembly} RFC 1122: section 3.3.2 {At least 576 byte datagrams} **Results** FAIL: icmp.v4 got no echo response, which is not an expected result.; SKIPPED: udp.v4: Packets exchanged with the DUT did not match those needed to either perform or grade the test. (Traffic source reports: echo response.); SKIPPED: tcp.v4: Packets exchanged with the DUT did not match those needed to either perform or grade the test. (Traffic source reports: echo response.) **Environment (please complete the following information):** - OS: (e.g. Linux ) - Toolchain (e.g Zephyr SDK) - Commit SHA or Version used: Zephyr3.0.0
test
fragment fragment mtu order n overlap by bytes error describe the bug fragment mtu order n overlap by bytes test is fail on on qemu references rfc section rfc section rfc section able to reassemble incoming datagrams rfc section support reassembly rfc section at least byte datagrams results fail icmp got no echo response which is not an expected result skipped udp packets exchanged with the dut did not match those needed to either perform or grade the test traffic source reports echo response skipped tcp packets exchanged with the dut did not match those needed to either perform or grade the test traffic source reports echo response environment please complete the following information os e g linux toolchain e g zephyr sdk commit sha or version used
1
171,921
13,253,170,643
IssuesEvent
2020-08-20 07:05:08
MachoThemes/strong-testimonials
https://api.github.com/repos/MachoThemes/strong-testimonials
closed
Setting tab for forms
tested
**Is your feature request related to a problem? Please describe.** No **Describe the solution you'd like** Move form&email tab from general settings to Forms.
1.0
Setting tab for forms - **Is your feature request related to a problem? Please describe.** No **Describe the solution you'd like** Move form&email tab from general settings to Forms.
test
setting tab for forms is your feature request related to a problem please describe no describe the solution you d like move form email tab from general settings to forms
1
525,700
15,263,260,369
IssuesEvent
2021-02-22 02:13:11
swlegion/tts
https://api.github.com/repos/swlegion/tts
opened
Simplify miniInfo.*
priority 2: soon type: maintenance
Specifically: 1. We should be able to infer `collider` from the `unitInfo.baseSize`, right? 2. We should be able to infer / merge `name` with the upgrade card or unit card. 3. It appears in almost every or every case `red` and `blue` are the same. 4. We should just have `scale` _always_ be `{1, 1, 1}`.
1.0
Simplify miniInfo.* - Specifically: 1. We should be able to infer `collider` from the `unitInfo.baseSize`, right? 2. We should be able to infer / merge `name` with the upgrade card or unit card. 3. It appears in almost every or every case `red` and `blue` are the same. 4. We should just have `scale` _always_ be `{1, 1, 1}`.
non_test
simplify miniinfo specifically we should be able to infer collider from the unitinfo basesize right we should be able to infer merge name with the upgrade card or unit card it appears in almost every or every case red and blue are the same we should just have scale always be
0
49,332
12,323,402,229
IssuesEvent
2020-05-13 12:07:02
ballerina-platform/ballerina-lang
https://api.github.com/repos/ballerina-platform/ballerina-lang
closed
Build/run prints an additional error "error: compilation contains errors"
Area/BuildTools Priority/High Type/Task
**Description:** $title. We've avoided printing this error by special-casing `BLangCompilerException`. See https://github.com/ballerina-platform/ballerina-lang/blob/master/cli/ballerina-tool/src/main/java/org/ballerinalang/tool/Main.java#L62 But the `CompileTask` throws a `BLauncherException` with this error, and thus it is printed. https://github.com/ballerina-platform/ballerina-lang/blob/master/cli/ballerina-packerina/src/main/java/org/ballerinalang/packerina/task/CompileTask.java#L79 The exception thrown here hasn't changed recently though. Seems like some other change has caused this previously un-executed code to execute now. If that is intended, we should be able to fix this by simply changing the exception thrown.
1.0
Build/run prints an additional error "error: compilation contains errors" - **Description:** $title. We've avoided printing this error by special-casing `BLangCompilerException`. See https://github.com/ballerina-platform/ballerina-lang/blob/master/cli/ballerina-tool/src/main/java/org/ballerinalang/tool/Main.java#L62 But the `CompileTask` throws a `BLauncherException` with this error, and thus it is printed. https://github.com/ballerina-platform/ballerina-lang/blob/master/cli/ballerina-packerina/src/main/java/org/ballerinalang/packerina/task/CompileTask.java#L79 The exception thrown here hasn't changed recently though. Seems like some other change has caused this previously un-executed code to execute now. If that is intended, we should be able to fix this by simply changing the exception thrown.
non_test
build run prints an additional error error compilation contains errors description title we ve avoided printing this error by special casing blangcompilerexception see but the compiletask throws a blauncherexception with this error and thus it is printed the exception thrown here hasn t changed recently though seems like some other change has caused this previously un executed code to execute now if that is intended we should be able to fix this by simply changing the exception thrown
0
165,384
12,839,683,640
IssuesEvent
2020-07-07 19:43:23
astropy/astropy
https://api.github.com/repos/astropy/astropy
opened
TST: Test against system libraries (cfitsio, wcslib)
Enhancement io.fits testing wcs
To avoid a repeat of issues like #10534 , where things are broken when package is built against system libraries over the bundled ones. cc @astrofrog @saimn @nden
1.0
TST: Test against system libraries (cfitsio, wcslib) - To avoid a repeat of issues like #10534 , where things are broken when package is built against system libraries over the bundled ones. cc @astrofrog @saimn @nden
test
tst test against system libraries cfitsio wcslib to avoid a repeat of issues like where things are broken when package is built against system libraries over the bundled ones cc astrofrog saimn nden
1
145,656
11,701,393,993
IssuesEvent
2020-03-06 19:35:16
pierre-ernst/snyk-github-issue-creator
https://api.github.com/repos/pierre-ernst/snyk-github-issue-creator
closed
goof - Regular Expression Denial of Service (ReDoS) in moment 2.15.1
snyk test
This issue has been created automatically by a source code scanner SNYKUID:npm:moment:20161019 ## Third party component with known security vulnerabilities Introduced to goof through: * goof > moment@2.15.1 ## Overview [`moment`](https://www.npmjs.com/package/moment) is a lightweight JavaScript date library for parsing, validating, manipulating, and formatting dates. Affected versions of the package are vulnerable to Regular Expression Denial of Service (ReDoS) attacks for any locale that has separate format and standalone options and `format` input can be controlled by the user. An attacker can provide a specially crafted input to the `format` function, which nearly matches the pattern being matched. This will cause the regular expression matching to take a long time, all the while occupying the event loop and preventing it from processing other requests and making the server unavailable (a Denial of Service attack). ## Disclosure Timeline - October 19th, 2016 - Reported the issue to package owner. - October 19th, 2016 - Issue acknowledged by package owner. - October 24th, 2016 - Issue fixed and version `2.15.2` released. ## Details Denial of Service (DoS) describes a family of attacks, all aimed at making a system inaccessible to its original and legitimate users. There are many types of DoS attacks, ranging from trying to clog the network pipes to the system by generating a large volume of traffic from many machines (a Distributed Denial of Service - DDoS - attack) to sending crafted requests that cause a system to crash or take a disproportional amount of time to process. The Regular expression Denial of Service (ReDoS) is a type of Denial of Service attack. Regular expressions are incredibly powerful, but they aren't very intuitive and can ultimately end up making it easy for attackers to take your site down. Let’s take the following regular expression as an example: ```js regex = /A(B|C+)+D/ ``` This regular expression accomplishes the following: - `A` The string must start with the letter 'A' - `(B|C+)+` The string must then follow the letter A with either the letter 'B' or some number of occurrences of the letter 'C' (the `+` matches one or more times). The `+` at the end of this section states that we can look for one or more matches of this section. - `D` Finally, we ensure this section of the string ends with a 'D' The expression would match inputs such as `ABBD`, `ABCCCCD`, `ABCBCCCD` and `ACCCCCD` It most cases, it doesn't take very long for a regex engine to find a match: ```bash $ time node -e '/A(B|C+)+D/.test("ACCCCCCCCCCCCCCCCCCCCCCCCCCCCD")' 0.04s user 0.01s system 95% cpu 0.052 total $ time node -e '/A(B|C+)+D/.test("ACCCCCCCCCCCCCCCCCCCCCCCCCCCCX")' 1.79s user 0.02s system 99% cpu 1.812 total ``` The entire process of testing it against a 30 characters long string takes around ~52ms. But when given an invalid string, it takes nearly two seconds to complete the test, over ten times as long as it took to test a valid string. The dramatic difference is due to the way regular expressions get evaluated. Most Regex engines will work very similarly (with minor differences). The engine will match the first possible way to accept the current character and proceed to the next one. If it then fails to match the next one, it will backtrack and see if there was another way to digest the previous character. If it goes too far down the rabbit hole only to find out the string doesn’t match in the end, and if many characters have multiple valid regex paths, the number of backtracking steps can become very large, resulting in what is known as _catastrophic backtracking_. Let's look at how our expression runs into this problem, using a shorter string: "ACCCX". While it seems fairly straightforward, there are still four different ways that the engine could match those three C's: 1. CCC 2. CC+C 3. C+CC 4. C+C+C. The engine has to try each of those combinations to see if any of them potentially match against the expression. When you combine that with the other steps the engine must take, we can use [RegEx 101 debugger](https://regex101.com/debugger) to see the engine has to take a total of 38 steps before it can determine the string doesn't match. From there, the number of steps the engine must use to validate a string just continues to grow. | String | Number of C's | Number of steps | | -------|-------------:| -----:| | ACCCX | 3 | 38 | ACCCCX | 4 | 71 | ACCCCCX | 5 | 136 | ACCCCCCCCCCCCCCX | 14 | 65,553 By the time the string includes 14 C's, the engine has to take over 65,000 steps just to see if the string is valid. These extreme situations can cause them to work very slowly (exponentially related to input size, as shown above), allowing an attacker to exploit this and can cause the service to excessively consume CPU, resulting in a Denial of Service. ## References - [Proof of concept](https://gist.github.com/grnd/50192ce22681848a7de812d95241b7fc) - [Fix commit](https://github.com/moment/moment/commit/663f33e333212b3800b63592cd8e237ac8fabdb9) - [npm:moment:20161019](https://snyk.io/vuln/npm:moment:20161019)
1.0
goof - Regular Expression Denial of Service (ReDoS) in moment 2.15.1 - This issue has been created automatically by a source code scanner SNYKUID:npm:moment:20161019 ## Third party component with known security vulnerabilities Introduced to goof through: * goof > moment@2.15.1 ## Overview [`moment`](https://www.npmjs.com/package/moment) is a lightweight JavaScript date library for parsing, validating, manipulating, and formatting dates. Affected versions of the package are vulnerable to Regular Expression Denial of Service (ReDoS) attacks for any locale that has separate format and standalone options and `format` input can be controlled by the user. An attacker can provide a specially crafted input to the `format` function, which nearly matches the pattern being matched. This will cause the regular expression matching to take a long time, all the while occupying the event loop and preventing it from processing other requests and making the server unavailable (a Denial of Service attack). ## Disclosure Timeline - October 19th, 2016 - Reported the issue to package owner. - October 19th, 2016 - Issue acknowledged by package owner. - October 24th, 2016 - Issue fixed and version `2.15.2` released. ## Details Denial of Service (DoS) describes a family of attacks, all aimed at making a system inaccessible to its original and legitimate users. There are many types of DoS attacks, ranging from trying to clog the network pipes to the system by generating a large volume of traffic from many machines (a Distributed Denial of Service - DDoS - attack) to sending crafted requests that cause a system to crash or take a disproportional amount of time to process. The Regular expression Denial of Service (ReDoS) is a type of Denial of Service attack. Regular expressions are incredibly powerful, but they aren't very intuitive and can ultimately end up making it easy for attackers to take your site down. Let’s take the following regular expression as an example: ```js regex = /A(B|C+)+D/ ``` This regular expression accomplishes the following: - `A` The string must start with the letter 'A' - `(B|C+)+` The string must then follow the letter A with either the letter 'B' or some number of occurrences of the letter 'C' (the `+` matches one or more times). The `+` at the end of this section states that we can look for one or more matches of this section. - `D` Finally, we ensure this section of the string ends with a 'D' The expression would match inputs such as `ABBD`, `ABCCCCD`, `ABCBCCCD` and `ACCCCCD` It most cases, it doesn't take very long for a regex engine to find a match: ```bash $ time node -e '/A(B|C+)+D/.test("ACCCCCCCCCCCCCCCCCCCCCCCCCCCCD")' 0.04s user 0.01s system 95% cpu 0.052 total $ time node -e '/A(B|C+)+D/.test("ACCCCCCCCCCCCCCCCCCCCCCCCCCCCX")' 1.79s user 0.02s system 99% cpu 1.812 total ``` The entire process of testing it against a 30 characters long string takes around ~52ms. But when given an invalid string, it takes nearly two seconds to complete the test, over ten times as long as it took to test a valid string. The dramatic difference is due to the way regular expressions get evaluated. Most Regex engines will work very similarly (with minor differences). The engine will match the first possible way to accept the current character and proceed to the next one. If it then fails to match the next one, it will backtrack and see if there was another way to digest the previous character. If it goes too far down the rabbit hole only to find out the string doesn’t match in the end, and if many characters have multiple valid regex paths, the number of backtracking steps can become very large, resulting in what is known as _catastrophic backtracking_. Let's look at how our expression runs into this problem, using a shorter string: "ACCCX". While it seems fairly straightforward, there are still four different ways that the engine could match those three C's: 1. CCC 2. CC+C 3. C+CC 4. C+C+C. The engine has to try each of those combinations to see if any of them potentially match against the expression. When you combine that with the other steps the engine must take, we can use [RegEx 101 debugger](https://regex101.com/debugger) to see the engine has to take a total of 38 steps before it can determine the string doesn't match. From there, the number of steps the engine must use to validate a string just continues to grow. | String | Number of C's | Number of steps | | -------|-------------:| -----:| | ACCCX | 3 | 38 | ACCCCX | 4 | 71 | ACCCCCX | 5 | 136 | ACCCCCCCCCCCCCCX | 14 | 65,553 By the time the string includes 14 C's, the engine has to take over 65,000 steps just to see if the string is valid. These extreme situations can cause them to work very slowly (exponentially related to input size, as shown above), allowing an attacker to exploit this and can cause the service to excessively consume CPU, resulting in a Denial of Service. ## References - [Proof of concept](https://gist.github.com/grnd/50192ce22681848a7de812d95241b7fc) - [Fix commit](https://github.com/moment/moment/commit/663f33e333212b3800b63592cd8e237ac8fabdb9) - [npm:moment:20161019](https://snyk.io/vuln/npm:moment:20161019)
test
goof regular expression denial of service redos in moment this issue has been created automatically by a source code scanner snykuid npm moment third party component with known security vulnerabilities introduced to goof through goof moment overview is a lightweight javascript date library for parsing validating manipulating and formatting dates affected versions of the package are vulnerable to regular expression denial of service redos attacks for any locale that has separate format and standalone options and format input can be controlled by the user an attacker can provide a specially crafted input to the format function which nearly matches the pattern being matched this will cause the regular expression matching to take a long time all the while occupying the event loop and preventing it from processing other requests and making the server unavailable a denial of service attack disclosure timeline october reported the issue to package owner october issue acknowledged by package owner october issue fixed and version released details denial of service dos describes a family of attacks all aimed at making a system inaccessible to its original and legitimate users there are many types of dos attacks ranging from trying to clog the network pipes to the system by generating a large volume of traffic from many machines a distributed denial of service ddos attack to sending crafted requests that cause a system to crash or take a disproportional amount of time to process the regular expression denial of service redos is a type of denial of service attack regular expressions are incredibly powerful but they aren t very intuitive and can ultimately end up making it easy for attackers to take your site down let’s take the following regular expression as an example js regex a b c d this regular expression accomplishes the following a the string must start with the letter a b c the string must then follow the letter a with either the letter b or some number of occurrences of the letter c the matches one or more times the at the end of this section states that we can look for one or more matches of this section d finally we ensure this section of the string ends with a d the expression would match inputs such as abbd abccccd abcbcccd and acccccd it most cases it doesn t take very long for a regex engine to find a match bash time node e a b c d test accccccccccccccccccccccccccccd user system cpu total time node e a b c d test accccccccccccccccccccccccccccx user system cpu total the entire process of testing it against a characters long string takes around but when given an invalid string it takes nearly two seconds to complete the test over ten times as long as it took to test a valid string the dramatic difference is due to the way regular expressions get evaluated most regex engines will work very similarly with minor differences the engine will match the first possible way to accept the current character and proceed to the next one if it then fails to match the next one it will backtrack and see if there was another way to digest the previous character if it goes too far down the rabbit hole only to find out the string doesn’t match in the end and if many characters have multiple valid regex paths the number of backtracking steps can become very large resulting in what is known as catastrophic backtracking let s look at how our expression runs into this problem using a shorter string acccx while it seems fairly straightforward there are still four different ways that the engine could match those three c s ccc cc c c cc c c c the engine has to try each of those combinations to see if any of them potentially match against the expression when you combine that with the other steps the engine must take we can use to see the engine has to take a total of steps before it can determine the string doesn t match from there the number of steps the engine must use to validate a string just continues to grow string number of c s number of steps acccx accccx acccccx accccccccccccccx by the time the string includes c s the engine has to take over steps just to see if the string is valid these extreme situations can cause them to work very slowly exponentially related to input size as shown above allowing an attacker to exploit this and can cause the service to excessively consume cpu resulting in a denial of service references
1
14,008
16,814,616,075
IssuesEvent
2021-06-17 05:24:05
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
GitLab source control integration?
Pri2 automation/svc awaiting-product-team-response cxp process-automation/subsvc product-question triaged
[ Hi, Is there a way to use GitLab for SCI with Azure Automation? ] --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 83c90e64-b615-711f-a53d-fc76606e2ecd * Version Independent ID: 2d164036-6886-4440-50f7-369f99f41cea * Content: [Use source control integration in Azure Automation](https://docs.microsoft.com/en-us/azure/automation/source-control-integration) * Content Source: [articles/automation/source-control-integration.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/source-control-integration.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @MGoedtel * Microsoft Alias: **magoedte**
1.0
GitLab source control integration? - [ Hi, Is there a way to use GitLab for SCI with Azure Automation? ] --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 83c90e64-b615-711f-a53d-fc76606e2ecd * Version Independent ID: 2d164036-6886-4440-50f7-369f99f41cea * Content: [Use source control integration in Azure Automation](https://docs.microsoft.com/en-us/azure/automation/source-control-integration) * Content Source: [articles/automation/source-control-integration.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/source-control-integration.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @MGoedtel * Microsoft Alias: **magoedte**
non_test
gitlab source control integration hi is there a way to use gitlab for sci with azure automation document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login mgoedtel microsoft alias magoedte
0
231,244
7,625,220,304
IssuesEvent
2018-05-03 20:40:33
OregonDigital/ControlledVocabularyManager
https://api.github.com/repos/OregonDigital/ControlledVocabularyManager
opened
Button for admins to clear cached HTML for Term, Vocabulary or Predicate
Priority: Low enhancement
Occasionally certain changes happen and the cached HTML gets stuck, requiring logging in to the server and deleting the HTML file, forcing a new version to be created. Add a button on a Term/Predicate/Vocabulary show page to force removing the HTML and reload, titled "Clear Cached HTML"
1.0
Button for admins to clear cached HTML for Term, Vocabulary or Predicate - Occasionally certain changes happen and the cached HTML gets stuck, requiring logging in to the server and deleting the HTML file, forcing a new version to be created. Add a button on a Term/Predicate/Vocabulary show page to force removing the HTML and reload, titled "Clear Cached HTML"
non_test
button for admins to clear cached html for term vocabulary or predicate occasionally certain changes happen and the cached html gets stuck requiring logging in to the server and deleting the html file forcing a new version to be created add a button on a term predicate vocabulary show page to force removing the html and reload titled clear cached html
0
12,227
3,258,394,252
IssuesEvent
2015-10-20 22:07:13
docker/docker
https://api.github.com/repos/docker/docker
opened
Flaky test: DockerSuite.TestExecInteractiveStdinClose
kind/flakytest
Found in: #17202 ``` ---------------------------------------------------------------------- FAIL: docker_cli_exec_unix_test.go:18: DockerSuite.TestExecInteractiveStdinClose docker_cli_exec_unix_test.go:37: c.Assert(strings.TrimSpace(output), checker.Equals, "hello") ... obtained string = "" ... expected string = "hello" ---------------------------------------------------------------------- ```
1.0
Flaky test: DockerSuite.TestExecInteractiveStdinClose - Found in: #17202 ``` ---------------------------------------------------------------------- FAIL: docker_cli_exec_unix_test.go:18: DockerSuite.TestExecInteractiveStdinClose docker_cli_exec_unix_test.go:37: c.Assert(strings.TrimSpace(output), checker.Equals, "hello") ... obtained string = "" ... expected string = "hello" ---------------------------------------------------------------------- ```
test
flaky test dockersuite testexecinteractivestdinclose found in fail docker cli exec unix test go dockersuite testexecinteractivestdinclose docker cli exec unix test go c assert strings trimspace output checker equals hello obtained string expected string hello
1
33,458
27,456,431,895
IssuesEvent
2023-03-02 21:47:52
cds-snc/simplify-privacy-statements-V2
https://api.github.com/repos/cds-snc/simplify-privacy-statements-V2
closed
Terraform: create `ecr` module
infrastructure
## Description Create a `/terragrunt/aws/ecr` Terraform module and the `/terragrunt/env/production/ecr` config that creates the Elastic Container Registry (ECR) that will be used by the Lambda function. ## Todo - [ ] [ECR example](https://github.com/cds-snc/scan-files/blob/ece4d21257be07ced3cb84e3d13c5ff28e675e49/terragrunt/aws/api/ecr.tf) - [ ] Create an output with the ECR's `repository_url` attribute so we can use it when creating the [Lambda function](https://github.com/cds-snc/scan-files/blob/ece4d21257be07ced3cb84e3d13c5ff28e675e49/terragrunt/aws/api/lambda.tf#L7). ## Related - #207
1.0
Terraform: create `ecr` module - ## Description Create a `/terragrunt/aws/ecr` Terraform module and the `/terragrunt/env/production/ecr` config that creates the Elastic Container Registry (ECR) that will be used by the Lambda function. ## Todo - [ ] [ECR example](https://github.com/cds-snc/scan-files/blob/ece4d21257be07ced3cb84e3d13c5ff28e675e49/terragrunt/aws/api/ecr.tf) - [ ] Create an output with the ECR's `repository_url` attribute so we can use it when creating the [Lambda function](https://github.com/cds-snc/scan-files/blob/ece4d21257be07ced3cb84e3d13c5ff28e675e49/terragrunt/aws/api/lambda.tf#L7). ## Related - #207
non_test
terraform create ecr module description create a terragrunt aws ecr terraform module and the terragrunt env production ecr config that creates the elastic container registry ecr that will be used by the lambda function todo create an output with the ecr s repository url attribute so we can use it when creating the related
0
5,213
2,765,509,281
IssuesEvent
2015-04-29 20:57:32
fossology/fossology
https://api.github.com/repos/fossology/fossology
closed
License not found
Category: Nomos Component: Rank Component: Tester Priority: High Status: New Tracker: Bug
--- Author Name: **Mary Laser** Original Redmine Issue: 7701, http://www.fossology.org/issues/7701 Original Date: 2014/09/09 Original Assignee: larry shi --- None
1.0
License not found - --- Author Name: **Mary Laser** Original Redmine Issue: 7701, http://www.fossology.org/issues/7701 Original Date: 2014/09/09 Original Assignee: larry shi --- None
test
license not found author name mary laser original redmine issue original date original assignee larry shi none
1
229,856
18,445,387,639
IssuesEvent
2021-10-15 00:45:30
backend-br/vagas
https://api.github.com/repos/backend-br/vagas
closed
[Remoto] Back-end Developer @ Z1
PJ Pleno Remoto MongoDB TypeScript NoSQL Testes Unitários Stale
<!-- ================================================== Caso a vaga for remoto durante a pandemia informar no texto "Remoto durante o covid" ================================================== --> <!-- ================================================== POR FAVOR, SÓ POSTE SE A VAGA FOR PARA BACK-END! Não faça distinção de gênero no título da vaga. Use: "Back-End Developer" ao invés de "Desenvolvedor Back-End" \o/ Exemplo: `[São Paulo] Back-End Developer @ NOME DA EMPRESA` ================================================== --> <!-- ================================================== Caso a vaga for remoto durante a pandemia deixar a linha abaixo ================================================== --> > ## Nossa empresa A Z1 é uma conta digital para adolescentes. Nosso propósito é dar autonomia financeira para adolescentes que não podem ter conta no banco nem cartão. A Z1 atua através da educação financeira, ou seja, criamos funcionalidades para adolescentes aprenderem a ter responsabilidade financeira e acesso à economia digital com independência. Começamos a nossa história em 2019 e lançamos o nosso app em janeiro de 2021. Desde então, estamos crescendo a uma taxa de 30% por semana e 200% por mês, e já recebemos mais de R$18 milhões em aportes de investidores de primeira linha, como a aceleradora Y Combinator e os fundos de venture capital MAYA Capital, Homebrew, Mantis e Clocktower, entre outros. Saiba mais em http://z1.app/ e nessas matérias no Brazil Journal, Exame e Estadão. Falando sobre diversidade, essa é uma das principais batalhas da Z1. A diversidade e a luta pela equidade é um pilar central da Z1. Queremos ser a empresa mais diversa do Brasil. Usamos a linguagem neutra e respeitamos os pronomes escolhidos por cada pessoa. A linguagem neutra é utilizada para evitar generalizações no masculino e incluir quem não se sente confortável em se associar a limitação masculina e feminina da língua portuguesa tradicional. ## Descrição da vaga - Sugerir, implementar e seguir processos para garantir a qualidade e evolução da nossa área de backend; - Escrever códigos Typescript limpos, eficientes e testáveis que são fáceis de manter e depurar; - Escrever testes garantindo cobertura do código e também realizar testes manuais para garantir a qualidade da entrega; - Revisar o código das pull requests do time, debater melhorias, e prover feedbacks; - Compartilhar conhecimentos e ajudar a levantar a barra técnica do time; - Participar ativamente da construção de nossa cultura. ## Local Durante a pandemia estamos trabalhando de forma 100% remota. Quando a pandemia terminar, teremos um escritório em São Paulo/SP, mas a presença será sempre opcional. Você pode se candidatar de qualquer lugar do Brasil. :) ## Requisitos **Obrigatórios:** - Vaga exclusiva para pessoas pretas, trans e/ou mulheres. - 2+ anos de experiência; - Conhecimento de Typescript com Node.js; - Conhecimento de Clean Code: aplicar técnicas simples que visam facilitar a escrita e leitura do código; - Conhecimento de bancos de dados NoSQL, dynamo, mongodb, ou outros; - Experiência na escrita de testes unitários; - Conhecimento prático de metodologias ágeis. ## Benefícios - Plano de saúde ## Contratação PJ a combinar ## Como se candidatar Você pode se candidatar [clicando aqui](https://jobs.lever.co/z1.app/124fb73a-93aa-42f2-8b24-1c89ca7c65fc) ## Tempo médio de feedbacks Costumamos enviar feedbacks em até 15 dias após cada processo. ## Labels <!-- retire os labels que não fazem sentido à vaga --> #### Alocação - Remoto #### Regime - PJ #### Nível - Pleno
1.0
[Remoto] Back-end Developer @ Z1 - <!-- ================================================== Caso a vaga for remoto durante a pandemia informar no texto "Remoto durante o covid" ================================================== --> <!-- ================================================== POR FAVOR, SÓ POSTE SE A VAGA FOR PARA BACK-END! Não faça distinção de gênero no título da vaga. Use: "Back-End Developer" ao invés de "Desenvolvedor Back-End" \o/ Exemplo: `[São Paulo] Back-End Developer @ NOME DA EMPRESA` ================================================== --> <!-- ================================================== Caso a vaga for remoto durante a pandemia deixar a linha abaixo ================================================== --> > ## Nossa empresa A Z1 é uma conta digital para adolescentes. Nosso propósito é dar autonomia financeira para adolescentes que não podem ter conta no banco nem cartão. A Z1 atua através da educação financeira, ou seja, criamos funcionalidades para adolescentes aprenderem a ter responsabilidade financeira e acesso à economia digital com independência. Começamos a nossa história em 2019 e lançamos o nosso app em janeiro de 2021. Desde então, estamos crescendo a uma taxa de 30% por semana e 200% por mês, e já recebemos mais de R$18 milhões em aportes de investidores de primeira linha, como a aceleradora Y Combinator e os fundos de venture capital MAYA Capital, Homebrew, Mantis e Clocktower, entre outros. Saiba mais em http://z1.app/ e nessas matérias no Brazil Journal, Exame e Estadão. Falando sobre diversidade, essa é uma das principais batalhas da Z1. A diversidade e a luta pela equidade é um pilar central da Z1. Queremos ser a empresa mais diversa do Brasil. Usamos a linguagem neutra e respeitamos os pronomes escolhidos por cada pessoa. A linguagem neutra é utilizada para evitar generalizações no masculino e incluir quem não se sente confortável em se associar a limitação masculina e feminina da língua portuguesa tradicional. ## Descrição da vaga - Sugerir, implementar e seguir processos para garantir a qualidade e evolução da nossa área de backend; - Escrever códigos Typescript limpos, eficientes e testáveis que são fáceis de manter e depurar; - Escrever testes garantindo cobertura do código e também realizar testes manuais para garantir a qualidade da entrega; - Revisar o código das pull requests do time, debater melhorias, e prover feedbacks; - Compartilhar conhecimentos e ajudar a levantar a barra técnica do time; - Participar ativamente da construção de nossa cultura. ## Local Durante a pandemia estamos trabalhando de forma 100% remota. Quando a pandemia terminar, teremos um escritório em São Paulo/SP, mas a presença será sempre opcional. Você pode se candidatar de qualquer lugar do Brasil. :) ## Requisitos **Obrigatórios:** - Vaga exclusiva para pessoas pretas, trans e/ou mulheres. - 2+ anos de experiência; - Conhecimento de Typescript com Node.js; - Conhecimento de Clean Code: aplicar técnicas simples que visam facilitar a escrita e leitura do código; - Conhecimento de bancos de dados NoSQL, dynamo, mongodb, ou outros; - Experiência na escrita de testes unitários; - Conhecimento prático de metodologias ágeis. ## Benefícios - Plano de saúde ## Contratação PJ a combinar ## Como se candidatar Você pode se candidatar [clicando aqui](https://jobs.lever.co/z1.app/124fb73a-93aa-42f2-8b24-1c89ca7c65fc) ## Tempo médio de feedbacks Costumamos enviar feedbacks em até 15 dias após cada processo. ## Labels <!-- retire os labels que não fazem sentido à vaga --> #### Alocação - Remoto #### Regime - PJ #### Nível - Pleno
test
back end developer caso a vaga for remoto durante a pandemia informar no texto remoto durante o covid por favor só poste se a vaga for para back end não faça distinção de gênero no título da vaga use back end developer ao invés de desenvolvedor back end o exemplo back end developer nome da empresa caso a vaga for remoto durante a pandemia deixar a linha abaixo nossa empresa a é uma conta digital para adolescentes nosso propósito é dar autonomia financeira para adolescentes que não podem ter conta no banco nem cartão a atua através da educação financeira ou seja criamos funcionalidades para adolescentes aprenderem a ter responsabilidade financeira e acesso à economia digital com independência começamos a nossa história em e lançamos o nosso app em janeiro de desde então estamos crescendo a uma taxa de por semana e por mês e já recebemos mais de r milhões em aportes de investidores de primeira linha como a aceleradora y combinator e os fundos de venture capital maya capital homebrew mantis e clocktower entre outros saiba mais em e nessas matérias no brazil journal exame e estadão falando sobre diversidade essa é uma das principais batalhas da a diversidade e a luta pela equidade é um pilar central da queremos ser a empresa mais diversa do brasil usamos a linguagem neutra e respeitamos os pronomes escolhidos por cada pessoa a linguagem neutra é utilizada para evitar generalizações no masculino e incluir quem não se sente confortável em se associar a limitação masculina e feminina da língua portuguesa tradicional descrição da vaga sugerir implementar e seguir processos para garantir a qualidade e evolução da nossa área de backend escrever códigos typescript limpos eficientes e testáveis que são fáceis de manter e depurar escrever testes garantindo cobertura do código e também realizar testes manuais para garantir a qualidade da entrega revisar o código das pull requests do time debater melhorias e prover feedbacks compartilhar conhecimentos e ajudar a levantar a barra técnica do time participar ativamente da construção de nossa cultura local durante a pandemia estamos trabalhando de forma remota quando a pandemia terminar teremos um escritório em são paulo sp mas a presença será sempre opcional você pode se candidatar de qualquer lugar do brasil requisitos obrigatórios vaga exclusiva para pessoas pretas trans e ou mulheres anos de experiência conhecimento de typescript com node js conhecimento de clean code aplicar técnicas simples que visam facilitar a escrita e leitura do código conhecimento de bancos de dados nosql dynamo mongodb ou outros experiência na escrita de testes unitários conhecimento prático de metodologias ágeis benefícios plano de saúde contratação pj a combinar como se candidatar você pode se candidatar tempo médio de feedbacks costumamos enviar feedbacks em até dias após cada processo labels alocação remoto regime pj nível pleno
1
26,007
19,561,193,531
IssuesEvent
2022-01-03 16:25:33
platten/enarx-test
https://api.github.com/repos/platten/enarx-test
opened
Automatically apply labels to PRs based on files modified
good first issue infrastructure
<a href="https://github.com/npmccallum"><img src="https://avatars.githubusercontent.com/u/288304?v=4" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [npmccallum](https://github.com/npmccallum)** _Sat Dec 21 15:55:22 2019_ ---- Get a list of the files modified. Do a regular expression search for SGX. If found, label it with `technology/sgx`. Or likewise for other projects.
1.0
Automatically apply labels to PRs based on files modified - <a href="https://github.com/npmccallum"><img src="https://avatars.githubusercontent.com/u/288304?v=4" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [npmccallum](https://github.com/npmccallum)** _Sat Dec 21 15:55:22 2019_ ---- Get a list of the files modified. Do a regular expression search for SGX. If found, label it with `technology/sgx`. Or likewise for other projects.
non_test
automatically apply labels to prs based on files modified issue by sat dec get a list of the files modified do a regular expression search for sgx if found label it with technology sgx or likewise for other projects
0
187,245
14,427,276,259
IssuesEvent
2020-12-06 03:00:40
kalexmills/github-vet-tests-dec2020
https://api.github.com/repos/kalexmills/github-vet-tests-dec2020
closed
mongodb/ftdc: vendor/go.mongodb.org/mongo-driver/mongo/crud_spec_test.go; 56 LoC
fresh medium test vendored
Found a possible issue in [mongodb/ftdc](https://www.github.com/mongodb/ftdc) at [vendor/go.mongodb.org/mongo-driver/mongo/crud_spec_test.go](https://github.com/mongodb/ftdc/blob/d5dbbae38209b053448f889ce16f57f1bee3d296/vendor/go.mongodb.org/mongo-driver/mongo/crud_spec_test.go#L132-L187) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call which takes a reference to test at line 157 may start a goroutine [Click here to see the code in its original context.](https://github.com/mongodb/ftdc/blob/d5dbbae38209b053448f889ce16f57f1bee3d296/vendor/go.mongodb.org/mongo-driver/mongo/crud_spec_test.go#L132-L187) <details> <summary>Click here to show the 56 line(s) of Go which triggered the analyzer.</summary> ```go for _, test := range testfile.Tests { collName := sanitizeCollectionName("crud-spec-tests", test.Description) _ = db.RunCommand( context.Background(), bsonx.Doc{{"drop", bsonx.String(collName)}}, ) if test.Outcome.Collection != nil && len(test.Outcome.Collection.Name) > 0 { _ = db.RunCommand( context.Background(), bsonx.Doc{{"drop", bsonx.String(test.Outcome.Collection.Name)}}, ) } coll := db.Collection(collName) docsToInsert := docSliceToInterfaceSlice(docSliceFromRaw(t, testfile.Data)) wcColl, err := coll.Clone(options.Collection().SetWriteConcern(writeconcern.New(writeconcern.WMajority()))) require.NoError(t, err) _, err = wcColl.InsertMany(context.Background(), docsToInsert) require.NoError(t, err) switch test.Operation.Name { case "aggregate": aggregateTest(t, db, coll, &test) case "bulkWrite": bulkWriteTest(t, wcColl, &test) case "count": countTest(t, coll, &test) case "distinct": distinctTest(t, coll, &test) case "find": findTest(t, coll, &test) case "deleteMany": deleteManyTest(t, coll, &test) case "deleteOne": deleteOneTest(t, coll, &test) case "findOneAndDelete": findOneAndDeleteTest(t, coll, &test) case "findOneAndReplace": findOneAndReplaceTest(t, coll, &test) case "findOneAndUpdate": findOneAndUpdateTest(t, coll, &test) case "insertMany": insertManyTest(t, coll, &test) case "insertOne": insertOneTest(t, coll, &test) case "replaceOne": replaceOneTest(t, coll, &test) case "updateMany": updateManyTest(t, coll, &test) case "updateOne": updateOneTest(t, coll, &test) } } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: d5dbbae38209b053448f889ce16f57f1bee3d296
1.0
mongodb/ftdc: vendor/go.mongodb.org/mongo-driver/mongo/crud_spec_test.go; 56 LoC - Found a possible issue in [mongodb/ftdc](https://www.github.com/mongodb/ftdc) at [vendor/go.mongodb.org/mongo-driver/mongo/crud_spec_test.go](https://github.com/mongodb/ftdc/blob/d5dbbae38209b053448f889ce16f57f1bee3d296/vendor/go.mongodb.org/mongo-driver/mongo/crud_spec_test.go#L132-L187) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call which takes a reference to test at line 157 may start a goroutine [Click here to see the code in its original context.](https://github.com/mongodb/ftdc/blob/d5dbbae38209b053448f889ce16f57f1bee3d296/vendor/go.mongodb.org/mongo-driver/mongo/crud_spec_test.go#L132-L187) <details> <summary>Click here to show the 56 line(s) of Go which triggered the analyzer.</summary> ```go for _, test := range testfile.Tests { collName := sanitizeCollectionName("crud-spec-tests", test.Description) _ = db.RunCommand( context.Background(), bsonx.Doc{{"drop", bsonx.String(collName)}}, ) if test.Outcome.Collection != nil && len(test.Outcome.Collection.Name) > 0 { _ = db.RunCommand( context.Background(), bsonx.Doc{{"drop", bsonx.String(test.Outcome.Collection.Name)}}, ) } coll := db.Collection(collName) docsToInsert := docSliceToInterfaceSlice(docSliceFromRaw(t, testfile.Data)) wcColl, err := coll.Clone(options.Collection().SetWriteConcern(writeconcern.New(writeconcern.WMajority()))) require.NoError(t, err) _, err = wcColl.InsertMany(context.Background(), docsToInsert) require.NoError(t, err) switch test.Operation.Name { case "aggregate": aggregateTest(t, db, coll, &test) case "bulkWrite": bulkWriteTest(t, wcColl, &test) case "count": countTest(t, coll, &test) case "distinct": distinctTest(t, coll, &test) case "find": findTest(t, coll, &test) case "deleteMany": deleteManyTest(t, coll, &test) case "deleteOne": deleteOneTest(t, coll, &test) case "findOneAndDelete": findOneAndDeleteTest(t, coll, &test) case "findOneAndReplace": findOneAndReplaceTest(t, coll, &test) case "findOneAndUpdate": findOneAndUpdateTest(t, coll, &test) case "insertMany": insertManyTest(t, coll, &test) case "insertOne": insertOneTest(t, coll, &test) case "replaceOne": replaceOneTest(t, coll, &test) case "updateMany": updateManyTest(t, coll, &test) case "updateOne": updateOneTest(t, coll, &test) } } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: d5dbbae38209b053448f889ce16f57f1bee3d296
test
mongodb ftdc vendor go mongodb org mongo driver mongo crud spec test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message function call which takes a reference to test at line may start a goroutine click here to show the line s of go which triggered the analyzer go for test range testfile tests collname sanitizecollectionname crud spec tests test description db runcommand context background bsonx doc drop bsonx string collname if test outcome collection nil len test outcome collection name db runcommand context background bsonx doc drop bsonx string test outcome collection name coll db collection collname docstoinsert docslicetointerfaceslice docslicefromraw t testfile data wccoll err coll clone options collection setwriteconcern writeconcern new writeconcern wmajority require noerror t err err wccoll insertmany context background docstoinsert require noerror t err switch test operation name case aggregate aggregatetest t db coll test case bulkwrite bulkwritetest t wccoll test case count counttest t coll test case distinct distincttest t coll test case find findtest t coll test case deletemany deletemanytest t coll test case deleteone deleteonetest t coll test case findoneanddelete findoneanddeletetest t coll test case findoneandreplace findoneandreplacetest t coll test case findoneandupdate findoneandupdatetest t coll test case insertmany insertmanytest t coll test case insertone insertonetest t coll test case replaceone replaceonetest t coll test case updatemany updatemanytest t coll test case updateone updateonetest t coll test leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
1
27,765
5,096,985,224
IssuesEvent
2017-01-03 19:58:46
cakephp/cakephp
https://api.github.com/repos/cakephp/cakephp
closed
A Lot of preg_match() warning No ending matching delimiter and No ending delimiter when baking new plugin
Defect
This is a (multiple allowed): * [x] bug * [ ] enhancement * [ ] feature-discussion (RFC) * CakePHP Version: 3.3.11 * Platform and Target: Apache 2, MySQL, PHP 5.5.9, GNU/Linux Mint Rebecca. ### What you did I bake plugin ### What happened ```text Warning Error: preg_match(): No ending delimiter '?' found in [/var/www/html/global-arch/vendor/cakephp/bake/src/View/BakeView.php, line 248] 2017-01-01 14:11:51 Warning: Warning (2): preg_match(): No ending delimiter '?' found in [/var/www/html/global-arch/vendor/cakephp/bake/src/View/BakeView.php, line 248] Trace: Cake\Error\BaseErrorHandler::handleError() - CORE/src/Error/BaseErrorHandler.php, line 146 preg_match - [internal], line ?? Bake\View\BakeView::_isRegex() - ROOT/vendor/cakephp/bake/src/View/BakeView.php, line 248 Bake\View\BakeView::_evaluate() - ROOT/vendor/cakephp/bake/src/View/BakeView.php, line 186 Cake\View\View::_render() - CORE/src/View/View.php, line 973 Bake\View\BakeView::render() - ROOT/vendor/cakephp/bake/src/View/BakeView.php, line 136 Bake\Shell\Task\BakeTemplateTask::generate() - ROOT/vendor/cakephp/bake/src/Shell/Task/BakeTemplateTask.php, line 89 Bake\Shell\Task\PluginTask::_generateFile() - ROOT/vendor/cakephp/bake/src/Shell/Task/PluginTask.php, line 212 Bake\Shell\Task\PluginTask::_generateFiles() - ROOT/vendor/cakephp/bake/src/Shell/Task/PluginTask.php, line 198 Bake\Shell\Task\PluginTask::bake() - ROOT/vendor/cakephp/bake/src/Shell/Task/PluginTask.php, line 109 Bake\Shell\Task\PluginTask::main() - ROOT/vendor/cakephp/bake/src/Shell/Task/PluginTask.php, line 80 Cake\Console\Shell::runCommand() - CORE/src/Console/Shell.php, line 473 Cake\Console\Shell::runCommand() - CORE/src/Console/Shell.php, line 466 Cake\Console\ShellDispatcher::_dispatch() - CORE/src/Console/ShellDispatcher.php, line 227 Cake\Console\ShellDispatcher::dispatch() - CORE/src/Console/ShellDispatcher.php, line 182 Cake\Console\ShellDispatcher::run() - CORE/src/Console/ShellDispatcher.php, line 128 [main] - ROOT/bin/cake.php, line 34 ``` ```text Warning Error: preg_match(): No ending matching delimiter '>' found in [/var/www/html/global-arch/vendor/cakephp/bake/src/View/BakeView.php, line 248] 2017-01-01 14:11:51 Warning: Warning (2): preg_match(): No ending matching delimiter '>' found in [/var/www/html/global-arch/vendor/cakephp/bake/src/View/BakeView.php, line 248] Trace: Cake\Error\BaseErrorHandler::handleError() - CORE/src/Error/BaseErrorHandler.php, line 146 preg_match - [internal], line ?? Bake\View\BakeView::_isRegex() - ROOT/vendor/cakephp/bake/src/View/BakeView.php, line 248 Bake\View\BakeView::_evaluate() - ROOT/vendor/cakephp/bake/src/View/BakeView.php, line 186 Cake\View\View::_render() - CORE/src/View/View.php, line 973 Bake\View\BakeView::render() - ROOT/vendor/cakephp/bake/src/View/BakeView.php, line 136 Bake\Shell\Task\BakeTemplateTask::generate() - ROOT/vendor/cakephp/bake/src/Shell/Task/BakeTemplateTask.php, line 89 Bake\Shell\Task\PluginTask::_generateFile() - ROOT/vendor/cakephp/bake/src/Shell/Task/PluginTask.php, line 212 Bake\Shell\Task\PluginTask::_generateFiles() - ROOT/vendor/cakephp/bake/src/Shell/Task/PluginTask.php, line 198 Bake\Shell\Task\PluginTask::bake() - ROOT/vendor/cakephp/bake/src/Shell/Task/PluginTask.php, line 109 Bake\Shell\Task\PluginTask::main() - ROOT/vendor/cakephp/bake/src/Shell/Task/PluginTask.php, line 80 Cake\Console\Shell::runCommand() - CORE/src/Console/Shell.php, line 473 Cake\Console\Shell::runCommand() - CORE/src/Console/Shell.php, line 466 Cake\Console\ShellDispatcher::_dispatch() - CORE/src/Console/ShellDispatcher.php, line 227 Cake\Console\ShellDispatcher::dispatch() - CORE/src/Console/ShellDispatcher.php, line 182 Cake\Console\ShellDispatcher::run() - CORE/src/Console/ShellDispatcher.php, line 128 [main] - ROOT/bin/cake.php, line 34 ```
1.0
A Lot of preg_match() warning No ending matching delimiter and No ending delimiter when baking new plugin - This is a (multiple allowed): * [x] bug * [ ] enhancement * [ ] feature-discussion (RFC) * CakePHP Version: 3.3.11 * Platform and Target: Apache 2, MySQL, PHP 5.5.9, GNU/Linux Mint Rebecca. ### What you did I bake plugin ### What happened ```text Warning Error: preg_match(): No ending delimiter '?' found in [/var/www/html/global-arch/vendor/cakephp/bake/src/View/BakeView.php, line 248] 2017-01-01 14:11:51 Warning: Warning (2): preg_match(): No ending delimiter '?' found in [/var/www/html/global-arch/vendor/cakephp/bake/src/View/BakeView.php, line 248] Trace: Cake\Error\BaseErrorHandler::handleError() - CORE/src/Error/BaseErrorHandler.php, line 146 preg_match - [internal], line ?? Bake\View\BakeView::_isRegex() - ROOT/vendor/cakephp/bake/src/View/BakeView.php, line 248 Bake\View\BakeView::_evaluate() - ROOT/vendor/cakephp/bake/src/View/BakeView.php, line 186 Cake\View\View::_render() - CORE/src/View/View.php, line 973 Bake\View\BakeView::render() - ROOT/vendor/cakephp/bake/src/View/BakeView.php, line 136 Bake\Shell\Task\BakeTemplateTask::generate() - ROOT/vendor/cakephp/bake/src/Shell/Task/BakeTemplateTask.php, line 89 Bake\Shell\Task\PluginTask::_generateFile() - ROOT/vendor/cakephp/bake/src/Shell/Task/PluginTask.php, line 212 Bake\Shell\Task\PluginTask::_generateFiles() - ROOT/vendor/cakephp/bake/src/Shell/Task/PluginTask.php, line 198 Bake\Shell\Task\PluginTask::bake() - ROOT/vendor/cakephp/bake/src/Shell/Task/PluginTask.php, line 109 Bake\Shell\Task\PluginTask::main() - ROOT/vendor/cakephp/bake/src/Shell/Task/PluginTask.php, line 80 Cake\Console\Shell::runCommand() - CORE/src/Console/Shell.php, line 473 Cake\Console\Shell::runCommand() - CORE/src/Console/Shell.php, line 466 Cake\Console\ShellDispatcher::_dispatch() - CORE/src/Console/ShellDispatcher.php, line 227 Cake\Console\ShellDispatcher::dispatch() - CORE/src/Console/ShellDispatcher.php, line 182 Cake\Console\ShellDispatcher::run() - CORE/src/Console/ShellDispatcher.php, line 128 [main] - ROOT/bin/cake.php, line 34 ``` ```text Warning Error: preg_match(): No ending matching delimiter '>' found in [/var/www/html/global-arch/vendor/cakephp/bake/src/View/BakeView.php, line 248] 2017-01-01 14:11:51 Warning: Warning (2): preg_match(): No ending matching delimiter '>' found in [/var/www/html/global-arch/vendor/cakephp/bake/src/View/BakeView.php, line 248] Trace: Cake\Error\BaseErrorHandler::handleError() - CORE/src/Error/BaseErrorHandler.php, line 146 preg_match - [internal], line ?? Bake\View\BakeView::_isRegex() - ROOT/vendor/cakephp/bake/src/View/BakeView.php, line 248 Bake\View\BakeView::_evaluate() - ROOT/vendor/cakephp/bake/src/View/BakeView.php, line 186 Cake\View\View::_render() - CORE/src/View/View.php, line 973 Bake\View\BakeView::render() - ROOT/vendor/cakephp/bake/src/View/BakeView.php, line 136 Bake\Shell\Task\BakeTemplateTask::generate() - ROOT/vendor/cakephp/bake/src/Shell/Task/BakeTemplateTask.php, line 89 Bake\Shell\Task\PluginTask::_generateFile() - ROOT/vendor/cakephp/bake/src/Shell/Task/PluginTask.php, line 212 Bake\Shell\Task\PluginTask::_generateFiles() - ROOT/vendor/cakephp/bake/src/Shell/Task/PluginTask.php, line 198 Bake\Shell\Task\PluginTask::bake() - ROOT/vendor/cakephp/bake/src/Shell/Task/PluginTask.php, line 109 Bake\Shell\Task\PluginTask::main() - ROOT/vendor/cakephp/bake/src/Shell/Task/PluginTask.php, line 80 Cake\Console\Shell::runCommand() - CORE/src/Console/Shell.php, line 473 Cake\Console\Shell::runCommand() - CORE/src/Console/Shell.php, line 466 Cake\Console\ShellDispatcher::_dispatch() - CORE/src/Console/ShellDispatcher.php, line 227 Cake\Console\ShellDispatcher::dispatch() - CORE/src/Console/ShellDispatcher.php, line 182 Cake\Console\ShellDispatcher::run() - CORE/src/Console/ShellDispatcher.php, line 128 [main] - ROOT/bin/cake.php, line 34 ```
non_test
a lot of preg match warning no ending matching delimiter and no ending delimiter when baking new plugin this is a multiple allowed bug enhancement feature discussion rfc cakephp version platform and target apache mysql php gnu linux mint rebecca what you did i bake plugin what happened text warning error preg match no ending delimiter found in warning warning preg match no ending delimiter found in trace cake error baseerrorhandler handleerror core src error baseerrorhandler php line preg match line bake view bakeview isregex root vendor cakephp bake src view bakeview php line bake view bakeview evaluate root vendor cakephp bake src view bakeview php line cake view view render core src view view php line bake view bakeview render root vendor cakephp bake src view bakeview php line bake shell task baketemplatetask generate root vendor cakephp bake src shell task baketemplatetask php line bake shell task plugintask generatefile root vendor cakephp bake src shell task plugintask php line bake shell task plugintask generatefiles root vendor cakephp bake src shell task plugintask php line bake shell task plugintask bake root vendor cakephp bake src shell task plugintask php line bake shell task plugintask main root vendor cakephp bake src shell task plugintask php line cake console shell runcommand core src console shell php line cake console shell runcommand core src console shell php line cake console shelldispatcher dispatch core src console shelldispatcher php line cake console shelldispatcher dispatch core src console shelldispatcher php line cake console shelldispatcher run core src console shelldispatcher php line root bin cake php line text warning error preg match no ending matching delimiter found in warning warning preg match no ending matching delimiter found in trace cake error baseerrorhandler handleerror core src error baseerrorhandler php line preg match line bake view bakeview isregex root vendor cakephp bake src view bakeview php line bake view bakeview evaluate root vendor cakephp bake src view bakeview php line cake view view render core src view view php line bake view bakeview render root vendor cakephp bake src view bakeview php line bake shell task baketemplatetask generate root vendor cakephp bake src shell task baketemplatetask php line bake shell task plugintask generatefile root vendor cakephp bake src shell task plugintask php line bake shell task plugintask generatefiles root vendor cakephp bake src shell task plugintask php line bake shell task plugintask bake root vendor cakephp bake src shell task plugintask php line bake shell task plugintask main root vendor cakephp bake src shell task plugintask php line cake console shell runcommand core src console shell php line cake console shell runcommand core src console shell php line cake console shelldispatcher dispatch core src console shelldispatcher php line cake console shelldispatcher dispatch core src console shelldispatcher php line cake console shelldispatcher run core src console shelldispatcher php line root bin cake php line
0
116,395
14,952,026,722
IssuesEvent
2021-01-26 15:04:05
owncloud/ios-app
https://api.github.com/repos/owncloud/ios-app
closed
[BUG] Menu enhancements
Design UX bug enhancement
### Expected behavior The menu should have the same spacing everywhere (1). The folders are not completely hidden by the menu, the view that results is confusing (2). ### Actual behavior The spaces of the first and last cell look truncated (1). The folders are still visible behind the menu which gives the impression that they are part of the entries (2). ![image](https://user-images.githubusercontent.com/33026403/104772621-fe0c3300-5773-11eb-9104-635f6d0654a6.png) ### Client **iOS version:** 14.2 **ownCloud app version:** 11.4.5 **Device model:** iPhone 11
1.0
[BUG] Menu enhancements - ### Expected behavior The menu should have the same spacing everywhere (1). The folders are not completely hidden by the menu, the view that results is confusing (2). ### Actual behavior The spaces of the first and last cell look truncated (1). The folders are still visible behind the menu which gives the impression that they are part of the entries (2). ![image](https://user-images.githubusercontent.com/33026403/104772621-fe0c3300-5773-11eb-9104-635f6d0654a6.png) ### Client **iOS version:** 14.2 **ownCloud app version:** 11.4.5 **Device model:** iPhone 11
non_test
menu enhancements expected behavior the menu should have the same spacing everywhere the folders are not completely hidden by the menu the view that results is confusing actual behavior the spaces of the first and last cell look truncated the folders are still visible behind the menu which gives the impression that they are part of the entries client ios version owncloud app version device model iphone
0
98,441
8,677,263,497
IssuesEvent
2018-11-30 16:18:33
mozilla-services/syncstorage-rs
https://api.github.com/repos/mozilla-services/syncstorage-rs
closed
Improper response status codes
e2e tests in progress
Several requests return a 400 instead of 415, or a 400 instead of 404, or a 200 instead of a 400.
1.0
Improper response status codes - Several requests return a 400 instead of 415, or a 400 instead of 404, or a 200 instead of a 400.
test
improper response status codes several requests return a instead of or a instead of or a instead of a
1
44,041
5,582,062,379
IssuesEvent
2017-03-28 20:22:06
palantir/tslint
https://api.github.com/repos/palantir/tslint
closed
Broken test: Exit with code ??? if --project is not passed
Domain: Unit Testing P1 Status: Accepting PRs Type: Bug
The skipped test "exits with code 1 if --project is not passed" has asserts that the error code should be 2.
1.0
Broken test: Exit with code ??? if --project is not passed - The skipped test "exits with code 1 if --project is not passed" has asserts that the error code should be 2.
test
broken test exit with code if project is not passed the skipped test exits with code if project is not passed has asserts that the error code should be
1
45,033
18,346,897,141
IssuesEvent
2021-10-08 07:40:36
hashicorp/terraform-provider-azurerm
https://api.github.com/repos/hashicorp/terraform-provider-azurerm
closed
Improve validation for azurerm_firewall_network_rule_collection protocols argument
enhancement good first issue service/firewall
<!--- Please keep this note for the community ---> ### Community Note * Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Description Currently the azurerm provider does not appear to perform a full validation of values for the protocols argument in the azurerm_firewall_network_rule_collection resource, so if an invalid value is provided (such as "*") terraform plan and terraform validate will appear correct however the apply stage will fail. Any chance we can get improved validation for this? Currently using azurerm version 2.63 but looking through the changelogs there should be no changes relevant to this in future patches <!--- Please leave a helpful description of the feature request here. ---> ### New or Affected Resource(s) <!--- Please list the new or affected resources and data sources. ---> * azurerm_firewall_network_rule_collection ### Potential Terraform Configuration <!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code ---> ```hcl # Copy-paste your Terraform configurations here - for large Terraform configs, # please use a service like Dropbox and share a link to the ZIP file. For # security, you can also encrypt the files using our GPG public key. ``` ### References <!--- Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation? For example: * https://azure.microsoft.com/en-us/roadmap/virtual-network-service-endpoint-for-azure-cosmos-db/ ---> * #0000
1.0
Improve validation for azurerm_firewall_network_rule_collection protocols argument - <!--- Please keep this note for the community ---> ### Community Note * Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Description Currently the azurerm provider does not appear to perform a full validation of values for the protocols argument in the azurerm_firewall_network_rule_collection resource, so if an invalid value is provided (such as "*") terraform plan and terraform validate will appear correct however the apply stage will fail. Any chance we can get improved validation for this? Currently using azurerm version 2.63 but looking through the changelogs there should be no changes relevant to this in future patches <!--- Please leave a helpful description of the feature request here. ---> ### New or Affected Resource(s) <!--- Please list the new or affected resources and data sources. ---> * azurerm_firewall_network_rule_collection ### Potential Terraform Configuration <!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code ---> ```hcl # Copy-paste your Terraform configurations here - for large Terraform configs, # please use a service like Dropbox and share a link to the ZIP file. For # security, you can also encrypt the files using our GPG public key. ``` ### References <!--- Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation? For example: * https://azure.microsoft.com/en-us/roadmap/virtual-network-service-endpoint-for-azure-cosmos-db/ ---> * #0000
non_test
improve validation for azurerm firewall network rule collection protocols argument community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment description currently the azurerm provider does not appear to perform a full validation of values for the protocols argument in the azurerm firewall network rule collection resource so if an invalid value is provided such as terraform plan and terraform validate will appear correct however the apply stage will fail any chance we can get improved validation for this currently using azurerm version but looking through the changelogs there should be no changes relevant to this in future patches new or affected resource s azurerm firewall network rule collection potential terraform configuration hcl copy paste your terraform configurations here for large terraform configs please use a service like dropbox and share a link to the zip file for security you can also encrypt the files using our gpg public key references information about referencing github issues are there any other github issues open or closed or pull requests that should be linked here vendor blog posts or documentation for example
0
433,150
12,502,726,236
IssuesEvent
2020-06-02 05:28:38
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
m.facebook.com - see bug description
browser-fenix engine-gecko ml-needsdiagnosis-false ml-probability-high priority-critical
<!-- @browser: Firefox Mobile 78.0 --> <!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:78.0) Gecko/78.0 Firefox/78.0 --> <!-- @reported_with: --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/53608 --> <!-- @extra_labels: browser-fenix --> **URL**: https://m.facebook.com/androidauthority **Browser / Version**: Firefox Mobile 78.0 **Operating System**: Android **Tested Another Browser**: No **Problem type**: Something else **Description**: unable to open Facebook app after navigating to Facebook website. **Steps to Reproduce**: unable to open Facebook app after navigating to Facebook website. <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
m.facebook.com - see bug description - <!-- @browser: Firefox Mobile 78.0 --> <!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:78.0) Gecko/78.0 Firefox/78.0 --> <!-- @reported_with: --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/53608 --> <!-- @extra_labels: browser-fenix --> **URL**: https://m.facebook.com/androidauthority **Browser / Version**: Firefox Mobile 78.0 **Operating System**: Android **Tested Another Browser**: No **Problem type**: Something else **Description**: unable to open Facebook app after navigating to Facebook website. **Steps to Reproduce**: unable to open Facebook app after navigating to Facebook website. <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_test
m facebook com see bug description url browser version firefox mobile operating system android tested another browser no problem type something else description unable to open facebook app after navigating to facebook website steps to reproduce unable to open facebook app after navigating to facebook website browser configuration none from with ❤️
0
150,584
11,967,526,624
IssuesEvent
2020-04-06 06:53:32
ubtue/DatenProbleme
https://api.github.com/repos/ubtue/DatenProbleme
closed
ISSN 1872-0226 Contemporary Islam: Dynamics of Muslim Life Mix aus Latest Artikel und Standardartikel
Zotero_AUTO_RSS blocked ready for testing
Daten enthalten Titel, die keinem Heft zugeordnet sind, als auch solche, die dies sind
1.0
ISSN 1872-0226 Contemporary Islam: Dynamics of Muslim Life Mix aus Latest Artikel und Standardartikel - Daten enthalten Titel, die keinem Heft zugeordnet sind, als auch solche, die dies sind
test
issn contemporary islam dynamics of muslim life mix aus latest artikel und standardartikel daten enthalten titel die keinem heft zugeordnet sind als auch solche die dies sind
1
283,821
30,913,553,370
IssuesEvent
2023-08-05 02:13:22
panasalap/linux-4.19.72_test1
https://api.github.com/repos/panasalap/linux-4.19.72_test1
reopened
CVE-2022-3526 (High) detected in linux-yoctov5.4.51
Mend: dependency security vulnerability
## CVE-2022-3526 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yoctov5.4.51</b></p></summary> <p> <p>Yocto Linux Embedded kernel</p> <p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto>https://git.yoctoproject.org/git/linux-yocto</a></p> <p>Found in HEAD commit: <a href="https://github.com/panasalap/linux-4.19.72/commit/f1b7c617b9b8f4135ab2f75a0c407cc44d43683f">f1b7c617b9b8f4135ab2f75a0c407cc44d43683f</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/macvlan.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/macvlan.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> A vulnerability classified as problematic was found in Linux Kernel. This vulnerability affects the function macvlan_handle_frame of the file drivers/net/macvlan.c of the component skb. The manipulation leads to memory leak. The attack can be initiated remotely. It is recommended to apply a patch to fix this issue. The identifier of this vulnerability is VDB-211024. <p>Publish Date: 2022-10-16 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-3526>CVE-2022-3526</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-3526">https://www.linuxkernelcves.com/cves/CVE-2022-3526</a></p> <p>Release Date: 2022-10-16</p> <p>Fix Resolution: v5.15.35,v5.17.4,v5.18-rc3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-3526 (High) detected in linux-yoctov5.4.51 - ## CVE-2022-3526 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yoctov5.4.51</b></p></summary> <p> <p>Yocto Linux Embedded kernel</p> <p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto>https://git.yoctoproject.org/git/linux-yocto</a></p> <p>Found in HEAD commit: <a href="https://github.com/panasalap/linux-4.19.72/commit/f1b7c617b9b8f4135ab2f75a0c407cc44d43683f">f1b7c617b9b8f4135ab2f75a0c407cc44d43683f</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/macvlan.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/macvlan.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> A vulnerability classified as problematic was found in Linux Kernel. This vulnerability affects the function macvlan_handle_frame of the file drivers/net/macvlan.c of the component skb. The manipulation leads to memory leak. The attack can be initiated remotely. It is recommended to apply a patch to fix this issue. The identifier of this vulnerability is VDB-211024. <p>Publish Date: 2022-10-16 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-3526>CVE-2022-3526</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-3526">https://www.linuxkernelcves.com/cves/CVE-2022-3526</a></p> <p>Release Date: 2022-10-16</p> <p>Fix Resolution: v5.15.35,v5.17.4,v5.18-rc3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in linux cve high severity vulnerability vulnerable library linux yocto linux embedded kernel library home page a href found in head commit a href found in base branch master vulnerable source files drivers net macvlan c drivers net macvlan c vulnerability details a vulnerability classified as problematic was found in linux kernel this vulnerability affects the function macvlan handle frame of the file drivers net macvlan c of the component skb the manipulation leads to memory leak the attack can be initiated remotely it is recommended to apply a patch to fix this issue the identifier of this vulnerability is vdb publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
521,785
15,115,730,515
IssuesEvent
2021-02-09 05:13:25
brave/qa-resources
https://api.github.com/repos/brave/qa-resources
closed
1.19.x Release #5
ARM OS/Android priority/P1 x86
### <img src="https://www.rebron.org/wordpress/wp-content/uploads/2019/06/Android.png">`1.19.x Release #5` #### Release Date/Target: * Release Date: **`February 5, 2021`** #### Summary: Includes the following features/fixes: * **`88.0.4324.150`** - https://chromereleases.googleblog.com/2021/02/stable-channel-update-for-desktop_4.html **Milestone:** * https://github.com/brave/brave-browser/milestone/170 **Current Progress:** **`QA Complete & Signed off`** * ARM --> **`Completed`** * Tablet --> **`Completed`** * x86 --> **`Completed`** **Manual Passes:** * ARM --> ~~https://github.com/brave/brave-browser/issues/13976~~ **`Completed`** * Tablet --> ~~https://github.com/brave/brave-browser/issues/13977~~ **`Completed`** * x86 --> ~~https://github.com/brave/brave-browser/issues/13978~~ **`Completed`**
1.0
1.19.x Release #5 - ### <img src="https://www.rebron.org/wordpress/wp-content/uploads/2019/06/Android.png">`1.19.x Release #5` #### Release Date/Target: * Release Date: **`February 5, 2021`** #### Summary: Includes the following features/fixes: * **`88.0.4324.150`** - https://chromereleases.googleblog.com/2021/02/stable-channel-update-for-desktop_4.html **Milestone:** * https://github.com/brave/brave-browser/milestone/170 **Current Progress:** **`QA Complete & Signed off`** * ARM --> **`Completed`** * Tablet --> **`Completed`** * x86 --> **`Completed`** **Manual Passes:** * ARM --> ~~https://github.com/brave/brave-browser/issues/13976~~ **`Completed`** * Tablet --> ~~https://github.com/brave/brave-browser/issues/13977~~ **`Completed`** * x86 --> ~~https://github.com/brave/brave-browser/issues/13978~~ **`Completed`**
non_test
x release img src release release date target release date february summary includes the following features fixes milestone current progress qa complete signed off arm completed tablet completed completed manual passes arm completed tablet completed completed
0
324,107
27,778,291,926
IssuesEvent
2023-03-16 18:53:43
dart-lang/co19
https://api.github.com/repos/dart-lang/co19
closed
Missing line in a patterns test
bad-test
In the test . . . the last `for` statement doesn't follow the same principle as all of the previous `for` statements above it: setting the variable `visited` to `true` at the end of the loop body. It causes the test to fail at runtime. https://github.com/dart-lang/co19/blob/f6f980bd5e2c26085b0394c66fb3665a91802783/LanguageFeatures/Patterns/variable_declaration_A04_t03.dart#L79-L85
1.0
Missing line in a patterns test - In the test . . . the last `for` statement doesn't follow the same principle as all of the previous `for` statements above it: setting the variable `visited` to `true` at the end of the loop body. It causes the test to fail at runtime. https://github.com/dart-lang/co19/blob/f6f980bd5e2c26085b0394c66fb3665a91802783/LanguageFeatures/Patterns/variable_declaration_A04_t03.dart#L79-L85
test
missing line in a patterns test in the test the last for statement doesn t follow the same principle as all of the previous for statements above it setting the variable visited to true at the end of the loop body it causes the test to fail at runtime
1
247,650
20,987,411,468
IssuesEvent
2022-03-29 05:41:09
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
roachtest: import/nodeShutdown/worker failed
C-test-failure O-robot O-roachtest branch-master release-blocker
roachtest.import/nodeShutdown/worker [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=4713654&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=4713654&tab=artifacts#/import/nodeShutdown/worker) on master @ [29716850b181718594663889ddb5f479fef7a305](https://github.com/cockroachdb/cockroach/commits/29716850b181718594663889ddb5f479fef7a305): ``` The test failed on branch=master, cloud=gce: test artifacts and logs in: /artifacts/import/nodeShutdown/worker/run_1 cluster.go:1868,import.go:90,test_runner.go:875: one or more parallel execution failure (1) attached stack trace -- stack trace: | github.com/cockroachdb/cockroach/pkg/roachprod/install.(*SyncedCluster).ParallelE | github.com/cockroachdb/cockroach/pkg/roachprod/install/cluster_synced.go:2042 | github.com/cockroachdb/cockroach/pkg/roachprod/install.(*SyncedCluster).Parallel | github.com/cockroachdb/cockroach/pkg/roachprod/install/cluster_synced.go:1923 | github.com/cockroachdb/cockroach/pkg/roachprod/install.(*SyncedCluster).Start | github.com/cockroachdb/cockroach/pkg/roachprod/install/cockroach.go:167 | github.com/cockroachdb/cockroach/pkg/roachprod.Start | github.com/cockroachdb/cockroach/pkg/roachprod/roachprod.go:660 | main.(*clusterImpl).StartE | main/pkg/cmd/roachtest/cluster.go:1826 | main.(*clusterImpl).Start | main/pkg/cmd/roachtest/cluster.go:1867 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.registerImportNodeShutdown.func2 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/import.go:90 | main.(*testRunner).runTest.func2 | main/pkg/cmd/roachtest/test_runner.go:875 | runtime.goexit | GOROOT/src/runtime/asm_amd64.s:1581 Wraps: (2) one or more parallel execution failure Error types: (1) *withstack.withStack (2) *errutil.leafError ``` <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/bulk-io <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*import/nodeShutdown/worker.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
2.0
roachtest: import/nodeShutdown/worker failed - roachtest.import/nodeShutdown/worker [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=4713654&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=4713654&tab=artifacts#/import/nodeShutdown/worker) on master @ [29716850b181718594663889ddb5f479fef7a305](https://github.com/cockroachdb/cockroach/commits/29716850b181718594663889ddb5f479fef7a305): ``` The test failed on branch=master, cloud=gce: test artifacts and logs in: /artifacts/import/nodeShutdown/worker/run_1 cluster.go:1868,import.go:90,test_runner.go:875: one or more parallel execution failure (1) attached stack trace -- stack trace: | github.com/cockroachdb/cockroach/pkg/roachprod/install.(*SyncedCluster).ParallelE | github.com/cockroachdb/cockroach/pkg/roachprod/install/cluster_synced.go:2042 | github.com/cockroachdb/cockroach/pkg/roachprod/install.(*SyncedCluster).Parallel | github.com/cockroachdb/cockroach/pkg/roachprod/install/cluster_synced.go:1923 | github.com/cockroachdb/cockroach/pkg/roachprod/install.(*SyncedCluster).Start | github.com/cockroachdb/cockroach/pkg/roachprod/install/cockroach.go:167 | github.com/cockroachdb/cockroach/pkg/roachprod.Start | github.com/cockroachdb/cockroach/pkg/roachprod/roachprod.go:660 | main.(*clusterImpl).StartE | main/pkg/cmd/roachtest/cluster.go:1826 | main.(*clusterImpl).Start | main/pkg/cmd/roachtest/cluster.go:1867 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.registerImportNodeShutdown.func2 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/import.go:90 | main.(*testRunner).runTest.func2 | main/pkg/cmd/roachtest/test_runner.go:875 | runtime.goexit | GOROOT/src/runtime/asm_amd64.s:1581 Wraps: (2) one or more parallel execution failure Error types: (1) *withstack.withStack (2) *errutil.leafError ``` <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/bulk-io <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*import/nodeShutdown/worker.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
test
roachtest import nodeshutdown worker failed roachtest import nodeshutdown worker with on master the test failed on branch master cloud gce test artifacts and logs in artifacts import nodeshutdown worker run cluster go import go test runner go one or more parallel execution failure attached stack trace stack trace github com cockroachdb cockroach pkg roachprod install syncedcluster parallele github com cockroachdb cockroach pkg roachprod install cluster synced go github com cockroachdb cockroach pkg roachprod install syncedcluster parallel github com cockroachdb cockroach pkg roachprod install cluster synced go github com cockroachdb cockroach pkg roachprod install syncedcluster start github com cockroachdb cockroach pkg roachprod install cockroach go github com cockroachdb cockroach pkg roachprod start github com cockroachdb cockroach pkg roachprod roachprod go main clusterimpl starte main pkg cmd roachtest cluster go main clusterimpl start main pkg cmd roachtest cluster go github com cockroachdb cockroach pkg cmd roachtest tests registerimportnodeshutdown github com cockroachdb cockroach pkg cmd roachtest tests import go main testrunner runtest main pkg cmd roachtest test runner go runtime goexit goroot src runtime asm s wraps one or more parallel execution failure error types withstack withstack errutil leaferror help see see cc cockroachdb bulk io
1
195,467
14,730,316,335
IssuesEvent
2021-01-06 13:00:47
elastic/elasticsearch
https://api.github.com/repos/elastic/elasticsearch
opened
[CI] AsyncSearchActionIT.testUpdateStoreKeepAlive fails
:Search/Search >test-failure v7.12.0
The test `AsyncSearchActionIT.testUpdateStoreKeepAlive` failed today on 7.x. **Build scan**: https://gradle-enterprise.elastic.co/s/ee56ukpaummx4 **Repro line**: ``` ./gradlew ':x-pack:plugin:async-search:internalClusterTest' --tests "org.elasticsearch.xpack.search.AsyncSearchActionIT.testUpdateStoreKeepAlive" -Dtests.seed=44FC88CE59837C9C -Dtests.security.manager=true -Dtests.locale=is -Dtests.timezone=Asia/Tehran -Druntime.java=8 ``` **Reproduces locally?**: No **Applicable branches**: 7.x **Failure history**: Failed once today. **Failure excerpt**: ``` java.util.concurrent.ExecutionException: RemoteTransportException[[node_s0][127.0.0.1:38963][indices:data/read/async_search/get]]; nested: RemoteTransportException[[node_s1][127.0.0.1:43749][indices:data/read/async_search/get]]; nested: ResourceNotFoundException[FmlmQVI5RUlmVFZlSEhqR2tRbFpyaGcaMTNUM21sY2lTeEtYZGNDVzZUN1IxUToyMjI=]; at __randomizedtesting.SeedInfo.seed([44FC88CE59837C9C:7D414CB2B4D0C3F7]:0) at org.elasticsearch.common.util.concurrent.BaseFuture$Sync.getValue(BaseFuture.java:273) at org.elasticsearch.common.util.concurrent.BaseFuture$Sync.get(BaseFuture.java:260) at org.elasticsearch.common.util.concurrent.BaseFuture.get(BaseFuture.java:87) at org.elasticsearch.xpack.search.AsyncSearchIntegTestCase.getAsyncSearch(AsyncSearchIntegTestCase.java:170) at org.elasticsearch.xpack.search.AsyncSearchActionIT.testUpdateStoreKeepAlive(AsyncSearchActionIT.java:400) Caused by: RemoteTransportException[[node_s0][127.0.0.1:38963][indices:data/read/async_search/get]]; nested: RemoteTransportException[[node_s1][127.0.0.1:43749][indices:data/read/async_search/get]]; nested: ResourceNotFoundException[FmlmQVI5RUlmVFZlSEhqR2tRbFpyaGcaMTNUM21sY2lTeEtYZGNDVzZUN1IxUToyMjI=]; Caused by: RemoteTransportException[[node_s1][127.0.0.1:43749][indices:data/read/async_search/get]]; nested: ResourceNotFoundException[FmlmQVI5RUlmVFZlSEhqR2tRbFpyaGcaMTNUM21sY2lTeEtYZGNDVzZUN1IxUToyMjI=]; Caused by: ResourceNotFoundException[FmlmQVI5RUlmVFZlSEhqR2tRbFpyaGcaMTNUM21sY2lTeEtYZGNDVzZUN1IxUToyMjI=] at org.elasticsearch.xpack.core.async.AsyncTaskIndexService.lambda$getEncodedResponse$6(AsyncTaskIndexService.java:277) at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:128) at org.elasticsearch.action.support.ContextPreservingActionListener.onResponse(ContextPreservingActionListener.java:43) at org.elasticsearch.action.support.TransportAction$1.onResponse(TransportAction.java:94) at org.elasticsearch.action.support.TransportAction$1.onResponse(TransportAction.java:88) at org.elasticsearch.action.support.single.shard.TransportSingleShardAction$AsyncSingleAction$2.handleResponse(TransportSingleShardAction.java:251) at org.elasticsearch.action.support.single.shard.TransportSingleShardAction$AsyncSingleAction$2.handleResponse(TransportSingleShardAction.java:242) at org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.handleResponse(TransportService.java:1289) at org.elasticsearch.transport.InboundHandler.doHandleResponse(InboundHandler.java:253) at org.elasticsearch.transport.InboundHandler.handleResponse(InboundHandler.java:245) at org.elasticsearch.transport.InboundHandler.messageReceived(InboundHandler.java:133) at org.elasticsearch.transport.InboundHandler.inboundMessage(InboundHandler.java:89) at org.elasticsearch.transport.TcpTransport.inboundMessage(TcpTransport.java:700) at org.elasticsearch.transport.InboundPipeline.forwardFragments(InboundPipeline.java:142) at org.elasticsearch.transport.InboundPipeline.doHandleBytes(InboundPipeline.java:117) at org.elasticsearch.transport.InboundPipeline.handleBytes(InboundPipeline.java:82) at org.elasticsearch.transport.nio.MockNioTransport$MockTcpReadWriteHandler.consumeReads(MockNioTransport.java:296) at org.elasticsearch.nio.SocketChannelContext.handleReadBytes(SocketChannelContext.java:228) at org.elasticsearch.nio.BytesChannelContext.read(BytesChannelContext.java:40) at org.elasticsearch.nio.EventHandler.handleRead(EventHandler.java:139) at org.elasticsearch.transport.nio.TestEventHandler.handleRead(TestEventHandler.java:151) at org.elasticsearch.nio.NioSelector.handleRead(NioSelector.java:420) at org.elasticsearch.nio.NioSelector.processKey(NioSelector.java:246) at org.elasticsearch.nio.NioSelector.singleLoop(NioSelector.java:174) at org.elasticsearch.nio.NioSelector.runLoop(NioSelector.java:131) at java.lang.Thread.run(Thread.java:748) ```
1.0
[CI] AsyncSearchActionIT.testUpdateStoreKeepAlive fails - The test `AsyncSearchActionIT.testUpdateStoreKeepAlive` failed today on 7.x. **Build scan**: https://gradle-enterprise.elastic.co/s/ee56ukpaummx4 **Repro line**: ``` ./gradlew ':x-pack:plugin:async-search:internalClusterTest' --tests "org.elasticsearch.xpack.search.AsyncSearchActionIT.testUpdateStoreKeepAlive" -Dtests.seed=44FC88CE59837C9C -Dtests.security.manager=true -Dtests.locale=is -Dtests.timezone=Asia/Tehran -Druntime.java=8 ``` **Reproduces locally?**: No **Applicable branches**: 7.x **Failure history**: Failed once today. **Failure excerpt**: ``` java.util.concurrent.ExecutionException: RemoteTransportException[[node_s0][127.0.0.1:38963][indices:data/read/async_search/get]]; nested: RemoteTransportException[[node_s1][127.0.0.1:43749][indices:data/read/async_search/get]]; nested: ResourceNotFoundException[FmlmQVI5RUlmVFZlSEhqR2tRbFpyaGcaMTNUM21sY2lTeEtYZGNDVzZUN1IxUToyMjI=]; at __randomizedtesting.SeedInfo.seed([44FC88CE59837C9C:7D414CB2B4D0C3F7]:0) at org.elasticsearch.common.util.concurrent.BaseFuture$Sync.getValue(BaseFuture.java:273) at org.elasticsearch.common.util.concurrent.BaseFuture$Sync.get(BaseFuture.java:260) at org.elasticsearch.common.util.concurrent.BaseFuture.get(BaseFuture.java:87) at org.elasticsearch.xpack.search.AsyncSearchIntegTestCase.getAsyncSearch(AsyncSearchIntegTestCase.java:170) at org.elasticsearch.xpack.search.AsyncSearchActionIT.testUpdateStoreKeepAlive(AsyncSearchActionIT.java:400) Caused by: RemoteTransportException[[node_s0][127.0.0.1:38963][indices:data/read/async_search/get]]; nested: RemoteTransportException[[node_s1][127.0.0.1:43749][indices:data/read/async_search/get]]; nested: ResourceNotFoundException[FmlmQVI5RUlmVFZlSEhqR2tRbFpyaGcaMTNUM21sY2lTeEtYZGNDVzZUN1IxUToyMjI=]; Caused by: RemoteTransportException[[node_s1][127.0.0.1:43749][indices:data/read/async_search/get]]; nested: ResourceNotFoundException[FmlmQVI5RUlmVFZlSEhqR2tRbFpyaGcaMTNUM21sY2lTeEtYZGNDVzZUN1IxUToyMjI=]; Caused by: ResourceNotFoundException[FmlmQVI5RUlmVFZlSEhqR2tRbFpyaGcaMTNUM21sY2lTeEtYZGNDVzZUN1IxUToyMjI=] at org.elasticsearch.xpack.core.async.AsyncTaskIndexService.lambda$getEncodedResponse$6(AsyncTaskIndexService.java:277) at org.elasticsearch.action.ActionListener$1.onResponse(ActionListener.java:128) at org.elasticsearch.action.support.ContextPreservingActionListener.onResponse(ContextPreservingActionListener.java:43) at org.elasticsearch.action.support.TransportAction$1.onResponse(TransportAction.java:94) at org.elasticsearch.action.support.TransportAction$1.onResponse(TransportAction.java:88) at org.elasticsearch.action.support.single.shard.TransportSingleShardAction$AsyncSingleAction$2.handleResponse(TransportSingleShardAction.java:251) at org.elasticsearch.action.support.single.shard.TransportSingleShardAction$AsyncSingleAction$2.handleResponse(TransportSingleShardAction.java:242) at org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.handleResponse(TransportService.java:1289) at org.elasticsearch.transport.InboundHandler.doHandleResponse(InboundHandler.java:253) at org.elasticsearch.transport.InboundHandler.handleResponse(InboundHandler.java:245) at org.elasticsearch.transport.InboundHandler.messageReceived(InboundHandler.java:133) at org.elasticsearch.transport.InboundHandler.inboundMessage(InboundHandler.java:89) at org.elasticsearch.transport.TcpTransport.inboundMessage(TcpTransport.java:700) at org.elasticsearch.transport.InboundPipeline.forwardFragments(InboundPipeline.java:142) at org.elasticsearch.transport.InboundPipeline.doHandleBytes(InboundPipeline.java:117) at org.elasticsearch.transport.InboundPipeline.handleBytes(InboundPipeline.java:82) at org.elasticsearch.transport.nio.MockNioTransport$MockTcpReadWriteHandler.consumeReads(MockNioTransport.java:296) at org.elasticsearch.nio.SocketChannelContext.handleReadBytes(SocketChannelContext.java:228) at org.elasticsearch.nio.BytesChannelContext.read(BytesChannelContext.java:40) at org.elasticsearch.nio.EventHandler.handleRead(EventHandler.java:139) at org.elasticsearch.transport.nio.TestEventHandler.handleRead(TestEventHandler.java:151) at org.elasticsearch.nio.NioSelector.handleRead(NioSelector.java:420) at org.elasticsearch.nio.NioSelector.processKey(NioSelector.java:246) at org.elasticsearch.nio.NioSelector.singleLoop(NioSelector.java:174) at org.elasticsearch.nio.NioSelector.runLoop(NioSelector.java:131) at java.lang.Thread.run(Thread.java:748) ```
test
asyncsearchactionit testupdatestorekeepalive fails the test asyncsearchactionit testupdatestorekeepalive failed today on x build scan repro line gradlew x pack plugin async search internalclustertest tests org elasticsearch xpack search asyncsearchactionit testupdatestorekeepalive dtests seed dtests security manager true dtests locale is dtests timezone asia tehran druntime java reproduces locally no applicable branches x failure history failed once today failure excerpt java util concurrent executionexception remotetransportexception nested remotetransportexception nested resourcenotfoundexception at randomizedtesting seedinfo seed at org elasticsearch common util concurrent basefuture sync getvalue basefuture java at org elasticsearch common util concurrent basefuture sync get basefuture java at org elasticsearch common util concurrent basefuture get basefuture java at org elasticsearch xpack search asyncsearchintegtestcase getasyncsearch asyncsearchintegtestcase java at org elasticsearch xpack search asyncsearchactionit testupdatestorekeepalive asyncsearchactionit java caused by remotetransportexception nested remotetransportexception nested resourcenotfoundexception caused by remotetransportexception nested resourcenotfoundexception caused by resourcenotfoundexception at org elasticsearch xpack core async asynctaskindexservice lambda getencodedresponse asynctaskindexservice java at org elasticsearch action actionlistener onresponse actionlistener java at org elasticsearch action support contextpreservingactionlistener onresponse contextpreservingactionlistener java at org elasticsearch action support transportaction onresponse transportaction java at org elasticsearch action support transportaction onresponse transportaction java at org elasticsearch action support single shard transportsingleshardaction asyncsingleaction handleresponse transportsingleshardaction java at org elasticsearch action support single shard transportsingleshardaction asyncsingleaction handleresponse transportsingleshardaction java at org elasticsearch transport transportservice contextrestoreresponsehandler handleresponse transportservice java at org elasticsearch transport inboundhandler dohandleresponse inboundhandler java at org elasticsearch transport inboundhandler handleresponse inboundhandler java at org elasticsearch transport inboundhandler messagereceived inboundhandler java at org elasticsearch transport inboundhandler inboundmessage inboundhandler java at org elasticsearch transport tcptransport inboundmessage tcptransport java at org elasticsearch transport inboundpipeline forwardfragments inboundpipeline java at org elasticsearch transport inboundpipeline dohandlebytes inboundpipeline java at org elasticsearch transport inboundpipeline handlebytes inboundpipeline java at org elasticsearch transport nio mockniotransport mocktcpreadwritehandler consumereads mockniotransport java at org elasticsearch nio socketchannelcontext handlereadbytes socketchannelcontext java at org elasticsearch nio byteschannelcontext read byteschannelcontext java at org elasticsearch nio eventhandler handleread eventhandler java at org elasticsearch transport nio testeventhandler handleread testeventhandler java at org elasticsearch nio nioselector handleread nioselector java at org elasticsearch nio nioselector processkey nioselector java at org elasticsearch nio nioselector singleloop nioselector java at org elasticsearch nio nioselector runloop nioselector java at java lang thread run thread java
1
172,407
6,505,182,178
IssuesEvent
2017-08-24 01:04:44
openshift/openshift-ansible
https://api.github.com/repos/openshift/openshift-ansible
closed
improper dnsmasq config - pods don't have dns to outside world
affects_3.6 kind/bug priority/P2
When using `ansible-2.3.1.0-1.fc25.noarch` to run openshift-ansible @ a095a16 (with `openshift_release=v1.5.1`) against a set of Fedora 26 Atomic Host's in DigitalOcean I don't get a proper dns set up (running pods within openshift can't get dns). DigitalOcean uses a static networking configuration (not from DHCP) and the network-scripts files that get populated don't have dns information in them, but the `/etc/resolv.conf` on the host does have the proper dns config in it. ``` [root@origin-node-1 ~]# cat /etc/resolv.conf ; Created by cloud-init on instance boot automatically, do not edit. ; nameserver 8.8.8.8 nameserver 8.8.4.4 [root@origin-node-1 ~]# ls /etc/dnsmasq.d/ node-dnsmasq.conf origin-dns.conf ``` I've been told I should have another file in that directory. To work around this I removed the `no-resolv` line from `/etc/dnsmasq.d/origin-dns.conf`. A possible solution is to try to gather dns information from network files and if it doesn't exist in there then take it from resolv.conf before editing resolv.conf to point back to dnsmasq (will this persist across reboots, though, it depends on if cloud-init populates it every time or not). More info is below: <details> network-scripts files ``` [root@origin-node-1 ~]# cat /etc/sysconfig/network NETWORKING=yes NOZEROCONF=yes [root@origin-node-1 ~]# cat /etc/sysconfig/network-scripts/ifcfg-eth0 # Created by cloud-init on instance boot automatically, do not edit. # BOOTPROTO=none DEVICE=eth0 HWADDR=f2:91:9f:e3:4a:41 MTU=1500 ONBOOT=yes TYPE=Ethernet USERCTL=no [root@origin-node-1 ~]# cat /etc/sysconfig/network-scripts/ifcfg-eth0\: ifcfg-eth0:0 ifcfg-eth0:1 [root@origin-node-1 ~]# cat /etc/sysconfig/network-scripts/ifcfg-eth0\:0 # Created by cloud-init on instance boot automatically, do not edit. # BOOTPROTO=static DEFROUTE=yes DEVICE=eth0:0 GATEWAY=174.138.48.1 HWADDR=f2:91:9f:e3:4a:41 IPADDR=174.138.56.164 MTU=1500 NETMASK=255.255.240.0 ONBOOT=yes TYPE=Ethernet USERCTL=no [root@origin-node-1 ~]# cat /etc/sysconfig/network-scripts/ifcfg-eth0\:1 # Created by cloud-init on instance boot automatically, do not edit. # BOOTPROTO=static DEVICE=eth0:1 HWADDR=f2:91:9f:e3:4a:41 IPADDR=10.17.0.6 MTU=1500 NETMASK=255.255.0.0 ONBOOT=yes TYPE=Ethernet USERCTL=no ```
1.0
improper dnsmasq config - pods don't have dns to outside world - When using `ansible-2.3.1.0-1.fc25.noarch` to run openshift-ansible @ a095a16 (with `openshift_release=v1.5.1`) against a set of Fedora 26 Atomic Host's in DigitalOcean I don't get a proper dns set up (running pods within openshift can't get dns). DigitalOcean uses a static networking configuration (not from DHCP) and the network-scripts files that get populated don't have dns information in them, but the `/etc/resolv.conf` on the host does have the proper dns config in it. ``` [root@origin-node-1 ~]# cat /etc/resolv.conf ; Created by cloud-init on instance boot automatically, do not edit. ; nameserver 8.8.8.8 nameserver 8.8.4.4 [root@origin-node-1 ~]# ls /etc/dnsmasq.d/ node-dnsmasq.conf origin-dns.conf ``` I've been told I should have another file in that directory. To work around this I removed the `no-resolv` line from `/etc/dnsmasq.d/origin-dns.conf`. A possible solution is to try to gather dns information from network files and if it doesn't exist in there then take it from resolv.conf before editing resolv.conf to point back to dnsmasq (will this persist across reboots, though, it depends on if cloud-init populates it every time or not). More info is below: <details> network-scripts files ``` [root@origin-node-1 ~]# cat /etc/sysconfig/network NETWORKING=yes NOZEROCONF=yes [root@origin-node-1 ~]# cat /etc/sysconfig/network-scripts/ifcfg-eth0 # Created by cloud-init on instance boot automatically, do not edit. # BOOTPROTO=none DEVICE=eth0 HWADDR=f2:91:9f:e3:4a:41 MTU=1500 ONBOOT=yes TYPE=Ethernet USERCTL=no [root@origin-node-1 ~]# cat /etc/sysconfig/network-scripts/ifcfg-eth0\: ifcfg-eth0:0 ifcfg-eth0:1 [root@origin-node-1 ~]# cat /etc/sysconfig/network-scripts/ifcfg-eth0\:0 # Created by cloud-init on instance boot automatically, do not edit. # BOOTPROTO=static DEFROUTE=yes DEVICE=eth0:0 GATEWAY=174.138.48.1 HWADDR=f2:91:9f:e3:4a:41 IPADDR=174.138.56.164 MTU=1500 NETMASK=255.255.240.0 ONBOOT=yes TYPE=Ethernet USERCTL=no [root@origin-node-1 ~]# cat /etc/sysconfig/network-scripts/ifcfg-eth0\:1 # Created by cloud-init on instance boot automatically, do not edit. # BOOTPROTO=static DEVICE=eth0:1 HWADDR=f2:91:9f:e3:4a:41 IPADDR=10.17.0.6 MTU=1500 NETMASK=255.255.0.0 ONBOOT=yes TYPE=Ethernet USERCTL=no ```
non_test
improper dnsmasq config pods don t have dns to outside world when using ansible noarch to run openshift ansible with openshift release against a set of fedora atomic host s in digitalocean i don t get a proper dns set up running pods within openshift can t get dns digitalocean uses a static networking configuration not from dhcp and the network scripts files that get populated don t have dns information in them but the etc resolv conf on the host does have the proper dns config in it cat etc resolv conf created by cloud init on instance boot automatically do not edit nameserver nameserver ls etc dnsmasq d node dnsmasq conf origin dns conf i ve been told i should have another file in that directory to work around this i removed the no resolv line from etc dnsmasq d origin dns conf a possible solution is to try to gather dns information from network files and if it doesn t exist in there then take it from resolv conf before editing resolv conf to point back to dnsmasq will this persist across reboots though it depends on if cloud init populates it every time or not more info is below network scripts files cat etc sysconfig network networking yes nozeroconf yes cat etc sysconfig network scripts ifcfg created by cloud init on instance boot automatically do not edit bootproto none device hwaddr mtu onboot yes type ethernet userctl no cat etc sysconfig network scripts ifcfg ifcfg ifcfg cat etc sysconfig network scripts ifcfg created by cloud init on instance boot automatically do not edit bootproto static defroute yes device gateway hwaddr ipaddr mtu netmask onboot yes type ethernet userctl no cat etc sysconfig network scripts ifcfg created by cloud init on instance boot automatically do not edit bootproto static device hwaddr ipaddr mtu netmask onboot yes type ethernet userctl no
0
558,594
16,538,258,917
IssuesEvent
2021-05-27 14:04:44
fossasia/open-event-frontend
https://api.github.com/repos/fossasia/open-event-frontend
reopened
Edit Video Page: When there is only one recording, recording is not shown.
Priority: High bug next step
**Describe the bug** <!-- A clear and concise description of what the bug is. --> When there is only one BBB recording, then it is not shown in table and there is an error in console. **To Reproduce** Steps to reproduce the behaviour: 1. Go to 'video edit page of any BBB video' 2. See error **Expected behaviour** <!-- A clear and concise description of what you expected to happen. --> it should show recording. **Screenshots** <!-- If applicable, add screenshots to help explain your problem. --> ![1](https://user-images.githubusercontent.com/72552281/115452042-8ef31e80-a23b-11eb-9251-366e1b924d98.png)
1.0
Edit Video Page: When there is only one recording, recording is not shown. - **Describe the bug** <!-- A clear and concise description of what the bug is. --> When there is only one BBB recording, then it is not shown in table and there is an error in console. **To Reproduce** Steps to reproduce the behaviour: 1. Go to 'video edit page of any BBB video' 2. See error **Expected behaviour** <!-- A clear and concise description of what you expected to happen. --> it should show recording. **Screenshots** <!-- If applicable, add screenshots to help explain your problem. --> ![1](https://user-images.githubusercontent.com/72552281/115452042-8ef31e80-a23b-11eb-9251-366e1b924d98.png)
non_test
edit video page when there is only one recording recording is not shown describe the bug when there is only one bbb recording then it is not shown in table and there is an error in console to reproduce steps to reproduce the behaviour go to video edit page of any bbb video see error expected behaviour it should show recording screenshots
0
479,350
13,795,100,765
IssuesEvent
2020-10-09 17:25:06
ngageoint/hootenanny
https://api.github.com/repos/ngageoint/hootenanny
closed
North Vegas C&R bad way snaps across multiple task grid cell replacements - 9/2/20
Category: Algorithms Priority: High Type: Bug
* ~~`road_marking=solid_stop_line` snapping to roads~~ * ~~`road_marking=solid_stop_line` snapping to each other~~ * ~~`highway=footway` between West Flamingo Road and West Nevso Drive~~ * ~~`barrier=wall` north of Jockey Club~~ * ~~`highway=footway` North of Bally's~~ * ~~service roads by tunnel near Olive and Fremont~~
1.0
North Vegas C&R bad way snaps across multiple task grid cell replacements - 9/2/20 - * ~~`road_marking=solid_stop_line` snapping to roads~~ * ~~`road_marking=solid_stop_line` snapping to each other~~ * ~~`highway=footway` between West Flamingo Road and West Nevso Drive~~ * ~~`barrier=wall` north of Jockey Club~~ * ~~`highway=footway` North of Bally's~~ * ~~service roads by tunnel near Olive and Fremont~~
non_test
north vegas c r bad way snaps across multiple task grid cell replacements road marking solid stop line snapping to roads road marking solid stop line snapping to each other highway footway between west flamingo road and west nevso drive barrier wall north of jockey club highway footway north of bally s service roads by tunnel near olive and fremont
0
35,580
14,749,871,703
IssuesEvent
2021-01-08 00:34:09
Azure/azure-rest-api-specs
https://api.github.com/repos/Azure/azure-rest-api-specs
closed
webapp:Setting date in App Service app setting converts date to different format
App Services Service Attention customer-reported
When using the `az webapp config appsettings set` command to set an App Service app setting to an ISO-8601 date string (YYYY-MM-DDTHH:mm:ssZ), the date string gets converted into a date that loses time zone information and introduces uncertainty into day and month ordering. ## To reproduce: 1) Create App Service. 2) Run the following command: `az webapp config appsettings set --name <app-service-name> --resource-group <resource-group> --settings=BUILD_DATE=2019-12-09T19:42:25Z` ## Observed behaviour: Output from above command: ``` [ { "name": "BUILD_DATE", "slotSetting": false, "value": "12/09/2019 19:42:25" } ] ``` When observing the app setting via the portal, the date is also in the incorrect format. ## Expected behaviour: ``` [ { "name": "BUILD_DATE", "slotSetting": false, "value": "2019-12-09T19:42:25Z" } ] ``` ## Environment Summary ``` Darwin-18.7.0-x86_64-i386-64bit Python 3.7.5 Shell: bash azure-cli 2.0.76 * Extensions: front-door 1.0.2 storage-preview 0.2.9 ```
2.0
webapp:Setting date in App Service app setting converts date to different format - When using the `az webapp config appsettings set` command to set an App Service app setting to an ISO-8601 date string (YYYY-MM-DDTHH:mm:ssZ), the date string gets converted into a date that loses time zone information and introduces uncertainty into day and month ordering. ## To reproduce: 1) Create App Service. 2) Run the following command: `az webapp config appsettings set --name <app-service-name> --resource-group <resource-group> --settings=BUILD_DATE=2019-12-09T19:42:25Z` ## Observed behaviour: Output from above command: ``` [ { "name": "BUILD_DATE", "slotSetting": false, "value": "12/09/2019 19:42:25" } ] ``` When observing the app setting via the portal, the date is also in the incorrect format. ## Expected behaviour: ``` [ { "name": "BUILD_DATE", "slotSetting": false, "value": "2019-12-09T19:42:25Z" } ] ``` ## Environment Summary ``` Darwin-18.7.0-x86_64-i386-64bit Python 3.7.5 Shell: bash azure-cli 2.0.76 * Extensions: front-door 1.0.2 storage-preview 0.2.9 ```
non_test
webapp setting date in app service app setting converts date to different format when using the az webapp config appsettings set command to set an app service app setting to an iso date string yyyy mm ddthh mm ssz the date string gets converted into a date that loses time zone information and introduces uncertainty into day and month ordering to reproduce create app service run the following command az webapp config appsettings set name resource group settings build date observed behaviour output from above command name build date slotsetting false value when observing the app setting via the portal the date is also in the incorrect format expected behaviour name build date slotsetting false value environment summary darwin python shell bash azure cli extensions front door storage preview
0
139,478
11,271,095,592
IssuesEvent
2020-01-14 12:19:36
mozilla/iris_firefox
https://api.github.com/repos/mozilla/iris_firefox
closed
Test case: Bug 1403951 - [Win10] Browser window doesn't display when Dark/Light theme is enabled on Windows 10.
Test: new
https://testrail.stage.mozaws.net/index.php?/cases/view/118731 note: test case feature ID is 1998 and this test belongs in iris/tests/toolbars_window_controls
1.0
Test case: Bug 1403951 - [Win10] Browser window doesn't display when Dark/Light theme is enabled on Windows 10. - https://testrail.stage.mozaws.net/index.php?/cases/view/118731 note: test case feature ID is 1998 and this test belongs in iris/tests/toolbars_window_controls
test
test case bug browser window doesn t display when dark light theme is enabled on windows note test case feature id is and this test belongs in iris tests toolbars window controls
1
309,950
9,482,064,997
IssuesEvent
2019-04-21 11:46:39
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
closed
[7.7.0] Stairs can’t be placed on a wall in certain direction
Medium Priority
Version: v0.7.7.0-beta-staging-13e8d143 Steps to reproduce: Try to build stairs on a wall, because a fourth row can not be reached in height. Expected behavior: The stairs should place as direction is selected. This could be done in former versions. Actual behavior: The stairs flip direction in the moment of placing and in addition it is depending on point of compass. before placing ![stairs4](https://user-images.githubusercontent.com/38510829/44401715-d0dddb00-a54f-11e8-83a8-d2ca50378e00.jpg) after placing ![stairs3](https://user-images.githubusercontent.com/38510829/44401740-e94df580-a54f-11e8-9ab7-b501197d8121.jpg)
1.0
[7.7.0] Stairs can’t be placed on a wall in certain direction - Version: v0.7.7.0-beta-staging-13e8d143 Steps to reproduce: Try to build stairs on a wall, because a fourth row can not be reached in height. Expected behavior: The stairs should place as direction is selected. This could be done in former versions. Actual behavior: The stairs flip direction in the moment of placing and in addition it is depending on point of compass. before placing ![stairs4](https://user-images.githubusercontent.com/38510829/44401715-d0dddb00-a54f-11e8-83a8-d2ca50378e00.jpg) after placing ![stairs3](https://user-images.githubusercontent.com/38510829/44401740-e94df580-a54f-11e8-9ab7-b501197d8121.jpg)
non_test
stairs can’t be placed on a wall in certain direction version beta staging steps to reproduce try to build stairs on a wall because a fourth row can not be reached in height expected behavior the stairs should place as direction is selected this could be done in former versions actual behavior the stairs flip direction in the moment of placing and in addition it is depending on point of compass before placing after placing
0
1,960
2,580,172,546
IssuesEvent
2015-02-13 16:02:11
girldevelopit/gdi-new-site
https://api.github.com/repos/girldevelopit/gdi-new-site
closed
Fix margins on the "our story" section of /about
UX/Design Needed
**Description** The "our story" section of the about page currently runs into the text surrounding it. **Ask** Add margins to the "our story" section. **Additional Info** ![Fix it :)](https://cloud.githubusercontent.com/assets/876210/6181848/44582672-b2ff-11e4-8c74-16f8e1779f08.png)
1.0
Fix margins on the "our story" section of /about - **Description** The "our story" section of the about page currently runs into the text surrounding it. **Ask** Add margins to the "our story" section. **Additional Info** ![Fix it :)](https://cloud.githubusercontent.com/assets/876210/6181848/44582672-b2ff-11e4-8c74-16f8e1779f08.png)
non_test
fix margins on the our story section of about description the our story section of the about page currently runs into the text surrounding it ask add margins to the our story section additional info
0
260,263
22,605,651,424
IssuesEvent
2022-06-29 13:03:53
scylladb/scylla
https://api.github.com/repos/scylladb/scylla
closed
test/cql-pytest/test_tools.py some tests use the wrong schema file
test
While debugging issue #10924 I noticed an even bigger bug, which will be this issue. Both bugs will probably be fixed together. The bug is that the `scylla_sstable` fixture in `test/cql-pytest/test_tools.py` is called multiple times for different parameterization, but each time, it writes to the **same** schema file, and returns it. As far as I understand, the first test that uses this fixture is `test_scylla_sstable_dump` and it causes all instances of `scylla_sstable` to be run once. The next test that uses this fixture is `test_scylla_sstable_dump_merge`, and that uses the already existing module-scoped fixtures. So all of them use the same schema file - the last one written by the last `scylla_sstable` fixture instantiated! Unless I'm missing something, I'm worried that this bug didn't cause this test to fail - it means that the test is not actually sensitive to what the schema file contains. So why do we need this file at all? CC @denesb
1.0
test/cql-pytest/test_tools.py some tests use the wrong schema file - While debugging issue #10924 I noticed an even bigger bug, which will be this issue. Both bugs will probably be fixed together. The bug is that the `scylla_sstable` fixture in `test/cql-pytest/test_tools.py` is called multiple times for different parameterization, but each time, it writes to the **same** schema file, and returns it. As far as I understand, the first test that uses this fixture is `test_scylla_sstable_dump` and it causes all instances of `scylla_sstable` to be run once. The next test that uses this fixture is `test_scylla_sstable_dump_merge`, and that uses the already existing module-scoped fixtures. So all of them use the same schema file - the last one written by the last `scylla_sstable` fixture instantiated! Unless I'm missing something, I'm worried that this bug didn't cause this test to fail - it means that the test is not actually sensitive to what the schema file contains. So why do we need this file at all? CC @denesb
test
test cql pytest test tools py some tests use the wrong schema file while debugging issue i noticed an even bigger bug which will be this issue both bugs will probably be fixed together the bug is that the scylla sstable fixture in test cql pytest test tools py is called multiple times for different parameterization but each time it writes to the same schema file and returns it as far as i understand the first test that uses this fixture is test scylla sstable dump and it causes all instances of scylla sstable to be run once the next test that uses this fixture is test scylla sstable dump merge and that uses the already existing module scoped fixtures so all of them use the same schema file the last one written by the last scylla sstable fixture instantiated unless i m missing something i m worried that this bug didn t cause this test to fail it means that the test is not actually sensitive to what the schema file contains so why do we need this file at all cc denesb
1
118,323
9,982,645,472
IssuesEvent
2019-07-10 10:20:36
microsoft/azure-tools-for-java
https://api.github.com/repos/microsoft/azure-tools-for-java
opened
[IntelliJ]Can not save login status.
HDInsight IntelliJ Internal Test
Build: dev EAP 1172 Repro Steps: 1. Start intelliJ. 2. Azure sign in. 3. Close intelliJ. 4. Start intelliJ. Result: It change to sign out state.
1.0
[IntelliJ]Can not save login status. - Build: dev EAP 1172 Repro Steps: 1. Start intelliJ. 2. Azure sign in. 3. Close intelliJ. 4. Start intelliJ. Result: It change to sign out state.
test
can not save login status build dev eap repro steps start intellij azure sign in close intellij start intellij result it change to sign out state
1
825,472
31,391,730,738
IssuesEvent
2023-08-26 12:29:45
cesium/atomic
https://api.github.com/repos/cesium/atomic
closed
Review when to show "You must log in to access this page."
bug backend priority:medium
Basically, when these two situations happen, we don't want to show the user the mentioned notification: - ~~It's the first time the user opens the application and it's not logged in~~ - The user just logged out
1.0
Review when to show "You must log in to access this page." - Basically, when these two situations happen, we don't want to show the user the mentioned notification: - ~~It's the first time the user opens the application and it's not logged in~~ - The user just logged out
non_test
review when to show you must log in to access this page basically when these two situations happen we don t want to show the user the mentioned notification it s the first time the user opens the application and it s not logged in the user just logged out
0
17,042
3,591,078,818
IssuesEvent
2016-02-01 10:01:58
xyzdevuy/S9-Beta-Issue-Tracker
https://api.github.com/repos/xyzdevuy/S9-Beta-Issue-Tracker
closed
Existen items con sockets repetidas, resetearselas
Migración Requiere más testing y feedback Urgente
Resetear las sockets de los items que tienen repetidas, al parecer se migran tan cual y eso no está permitido, rompe el pvp. ![](http://i.imgur.com/udN4HaT.png)
1.0
Existen items con sockets repetidas, resetearselas - Resetear las sockets de los items que tienen repetidas, al parecer se migran tan cual y eso no está permitido, rompe el pvp. ![](http://i.imgur.com/udN4HaT.png)
test
existen items con sockets repetidas resetearselas resetear las sockets de los items que tienen repetidas al parecer se migran tan cual y eso no está permitido rompe el pvp
1
380,933
11,271,490,416
IssuesEvent
2020-01-14 13:10:14
googleapis/elixir-google-api
https://api.github.com/repos/googleapis/elixir-google-api
opened
Synthesis failed for Fitness
autosynth failure priority: p1 type: bug
Hello! Autosynth couldn't regenerate Fitness. :broken_heart: Here's the output from running `synth.py`: ``` Cloning into 'working_repo'... Switched to branch 'autosynth-fitness' Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 256, in <module> main() File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 196, in main last_synth_commit_hash = get_last_metadata_commit(args.metadata_path) File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 149, in get_last_metadata_commit text=True, File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/subprocess.py", line 403, in run with Popen(*popenargs, **kwargs) as process: TypeError: __init__() got an unexpected keyword argument 'text' ``` Google internal developers can see the full log [here](https://sponge/164839b8-e54e-409d-bb48-012d4d868e86).
1.0
Synthesis failed for Fitness - Hello! Autosynth couldn't regenerate Fitness. :broken_heart: Here's the output from running `synth.py`: ``` Cloning into 'working_repo'... Switched to branch 'autosynth-fitness' Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 256, in <module> main() File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 196, in main last_synth_commit_hash = get_last_metadata_commit(args.metadata_path) File "/tmpfs/src/git/autosynth/autosynth/synth.py", line 149, in get_last_metadata_commit text=True, File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/subprocess.py", line 403, in run with Popen(*popenargs, **kwargs) as process: TypeError: __init__() got an unexpected keyword argument 'text' ``` Google internal developers can see the full log [here](https://sponge/164839b8-e54e-409d-bb48-012d4d868e86).
non_test
synthesis failed for fitness hello autosynth couldn t regenerate fitness broken heart here s the output from running synth py cloning into working repo switched to branch autosynth fitness traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src git autosynth autosynth synth py line in main file tmpfs src git autosynth autosynth synth py line in main last synth commit hash get last metadata commit args metadata path file tmpfs src git autosynth autosynth synth py line in get last metadata commit text true file home kbuilder pyenv versions lib subprocess py line in run with popen popenargs kwargs as process typeerror init got an unexpected keyword argument text google internal developers can see the full log
0
245,705
20,793,948,227
IssuesEvent
2022-03-17 07:07:04
streamnative/pulsar
https://api.github.com/repos/streamnative/pulsar
closed
ISSUE-14465: Flaky-test: ResendRequestTest. testExclusiveCumulativeAckedNormalTopic
component/test flaky-tests release/blocker scope/test
Original Issue: apache/pulsar#14465 --- <!--- Instructions for reporting a flaky test using this issue template: 1. Replace test_class in title and body with the short test class name WITHOUT the package name. 2. Replace test_method in title and body with the test method that failed. Multiple methods are flaky, remove the content that refers to the test method. 3. Replace "url here" with a url to an example failure. In the Github Actions workflow run logs, you can right click on the line number to copy a link to the line. Example of such url is https://github.com/apache/pulsar/pull/8892/checks?check_run_id=1531075794#step:9:377 . The logs are available for a limited amount of time (usually for a few weeks). 4. Replace "relevant parts of the exception stacktrace here" with the a few lines of the stack trace that shows at least the exception message and the line of test code where the stacktrace occurred. 5. Replace "full exception stacktrace here" with the full exception stacktrace from logs. This section will be hidden by default. 6. Remove all unused fields / content to unclutter the reported issue. Remove this comment too. --> ResendRequestTest.testExclusiveCumulativeAckedNormalTopic is flaky. It fails sporadically. [example failure](https://github.com/apache/pulsar/runs/5327552441?check_suite_focus=true#step:9:2632) ``` Error: testExclusiveCumulativeAckedNormalTopic(org.apache.pulsar.broker.service.ResendRequestTest) Time elapsed: 60.166 s <<< FAILURE! org.testng.internal.thread.ThreadTimeoutException: Method org.apache.pulsar.broker.service.ResendRequestTest.testExclusiveCumulativeAckedNormalTopic() didn't finish within the time-out 60000 at org.testng.internal.MethodInvocationHelper.invokeWithTimeoutWithNewExecutor(MethodInvocationHelper.java:371) at org.testng.internal.MethodInvocationHelper.invokeWithTimeout(MethodInvocationHelper.java:282) at org.testng.internal.TestInvoker.invokeMethod(TestInvoker.java:605) at org.testng.internal.TestInvoker.retryFailed(TestInvoker.java:214) at org.testng.internal.MethodRunner.runInSequence(MethodRunner.java:58) at org.testng.internal.TestInvoker$MethodInvocationAgent.invoke(TestInvoker.java:822) at org.testng.internal.TestInvoker.invokeTestMethods(TestInvoker.java:147) at org.testng.internal.TestMethodWorker.invokeTestMethods(TestMethodWorker.java:146) at org.testng.internal.TestMethodWorker.run(TestMethodWorker.java:128) at java.base/java.util.ArrayList.forEach(ArrayList.java:1541) at org.testng.TestRunner.privateRun(TestRunner.java:764) at org.testng.TestRunner.run(TestRunner.java:585) at org.testng.SuiteRunner.runTest(SuiteRunner.java:384) at org.testng.SuiteRunner.runSequentially(SuiteRunner.java:378) at org.testng.SuiteRunner.privateRun(SuiteRunner.java:337) at org.testng.SuiteRunner.run(SuiteRunner.java:286) at org.testng.SuiteRunnerWorker.runSuite(SuiteRunnerWorker.java:53) at org.testng.SuiteRunnerWorker.run(SuiteRunnerWorker.java:96) at org.testng.TestNG.runSuitesSequentially(TestNG.java:1218) at org.testng.TestNG.runSuitesLocally(TestNG.java:1140) at org.testng.TestNG.runSuites(TestNG.java:1069) at org.testng.TestNG.run(TestNG.java:1037) at org.apache.maven.surefire.testng.TestNGExecutor.run(TestNGExecutor.java:135) at org.apache.maven.surefire.testng.TestNGDirectoryTestSuite.executeSingleClass(TestNGDirectoryTestSuite.java:112) at org.apache.maven.surefire.testng.TestNGDirectoryTestSuite.executeLazy(TestNGDirectoryTestSuite.java:123) at org.apache.maven.surefire.testng.TestNGDirectoryTestSuite.execute(TestNGDirectoryTestSuite.java:90) at org.apache.maven.surefire.testng.TestNGProvider.invoke(TestNGProvider.java:146) at org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:384) at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:345) at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:126) at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:418) ```
3.0
ISSUE-14465: Flaky-test: ResendRequestTest. testExclusiveCumulativeAckedNormalTopic - Original Issue: apache/pulsar#14465 --- <!--- Instructions for reporting a flaky test using this issue template: 1. Replace test_class in title and body with the short test class name WITHOUT the package name. 2. Replace test_method in title and body with the test method that failed. Multiple methods are flaky, remove the content that refers to the test method. 3. Replace "url here" with a url to an example failure. In the Github Actions workflow run logs, you can right click on the line number to copy a link to the line. Example of such url is https://github.com/apache/pulsar/pull/8892/checks?check_run_id=1531075794#step:9:377 . The logs are available for a limited amount of time (usually for a few weeks). 4. Replace "relevant parts of the exception stacktrace here" with the a few lines of the stack trace that shows at least the exception message and the line of test code where the stacktrace occurred. 5. Replace "full exception stacktrace here" with the full exception stacktrace from logs. This section will be hidden by default. 6. Remove all unused fields / content to unclutter the reported issue. Remove this comment too. --> ResendRequestTest.testExclusiveCumulativeAckedNormalTopic is flaky. It fails sporadically. [example failure](https://github.com/apache/pulsar/runs/5327552441?check_suite_focus=true#step:9:2632) ``` Error: testExclusiveCumulativeAckedNormalTopic(org.apache.pulsar.broker.service.ResendRequestTest) Time elapsed: 60.166 s <<< FAILURE! org.testng.internal.thread.ThreadTimeoutException: Method org.apache.pulsar.broker.service.ResendRequestTest.testExclusiveCumulativeAckedNormalTopic() didn't finish within the time-out 60000 at org.testng.internal.MethodInvocationHelper.invokeWithTimeoutWithNewExecutor(MethodInvocationHelper.java:371) at org.testng.internal.MethodInvocationHelper.invokeWithTimeout(MethodInvocationHelper.java:282) at org.testng.internal.TestInvoker.invokeMethod(TestInvoker.java:605) at org.testng.internal.TestInvoker.retryFailed(TestInvoker.java:214) at org.testng.internal.MethodRunner.runInSequence(MethodRunner.java:58) at org.testng.internal.TestInvoker$MethodInvocationAgent.invoke(TestInvoker.java:822) at org.testng.internal.TestInvoker.invokeTestMethods(TestInvoker.java:147) at org.testng.internal.TestMethodWorker.invokeTestMethods(TestMethodWorker.java:146) at org.testng.internal.TestMethodWorker.run(TestMethodWorker.java:128) at java.base/java.util.ArrayList.forEach(ArrayList.java:1541) at org.testng.TestRunner.privateRun(TestRunner.java:764) at org.testng.TestRunner.run(TestRunner.java:585) at org.testng.SuiteRunner.runTest(SuiteRunner.java:384) at org.testng.SuiteRunner.runSequentially(SuiteRunner.java:378) at org.testng.SuiteRunner.privateRun(SuiteRunner.java:337) at org.testng.SuiteRunner.run(SuiteRunner.java:286) at org.testng.SuiteRunnerWorker.runSuite(SuiteRunnerWorker.java:53) at org.testng.SuiteRunnerWorker.run(SuiteRunnerWorker.java:96) at org.testng.TestNG.runSuitesSequentially(TestNG.java:1218) at org.testng.TestNG.runSuitesLocally(TestNG.java:1140) at org.testng.TestNG.runSuites(TestNG.java:1069) at org.testng.TestNG.run(TestNG.java:1037) at org.apache.maven.surefire.testng.TestNGExecutor.run(TestNGExecutor.java:135) at org.apache.maven.surefire.testng.TestNGDirectoryTestSuite.executeSingleClass(TestNGDirectoryTestSuite.java:112) at org.apache.maven.surefire.testng.TestNGDirectoryTestSuite.executeLazy(TestNGDirectoryTestSuite.java:123) at org.apache.maven.surefire.testng.TestNGDirectoryTestSuite.execute(TestNGDirectoryTestSuite.java:90) at org.apache.maven.surefire.testng.TestNGProvider.invoke(TestNGProvider.java:146) at org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:384) at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:345) at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:126) at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:418) ```
test
issue flaky test resendrequesttest testexclusivecumulativeackednormaltopic original issue apache pulsar instructions for reporting a flaky test using this issue template replace test class in title and body with the short test class name without the package name replace test method in title and body with the test method that failed multiple methods are flaky remove the content that refers to the test method replace url here with a url to an example failure in the github actions workflow run logs you can right click on the line number to copy a link to the line example of such url is the logs are available for a limited amount of time usually for a few weeks replace relevant parts of the exception stacktrace here with the a few lines of the stack trace that shows at least the exception message and the line of test code where the stacktrace occurred replace full exception stacktrace here with the full exception stacktrace from logs this section will be hidden by default remove all unused fields content to unclutter the reported issue remove this comment too resendrequesttest testexclusivecumulativeackednormaltopic is flaky it fails sporadically error testexclusivecumulativeackednormaltopic org apache pulsar broker service resendrequesttest time elapsed s failure org testng internal thread threadtimeoutexception method org apache pulsar broker service resendrequesttest testexclusivecumulativeackednormaltopic didn t finish within the time out at org testng internal methodinvocationhelper invokewithtimeoutwithnewexecutor methodinvocationhelper java at org testng internal methodinvocationhelper invokewithtimeout methodinvocationhelper java at org testng internal testinvoker invokemethod testinvoker java at org testng internal testinvoker retryfailed testinvoker java at org testng internal methodrunner runinsequence methodrunner java at org testng internal testinvoker methodinvocationagent invoke testinvoker java at org testng internal testinvoker invoketestmethods testinvoker java at org testng internal testmethodworker invoketestmethods testmethodworker java at org testng internal testmethodworker run testmethodworker java at java base java util arraylist foreach arraylist java at org testng testrunner privaterun testrunner java at org testng testrunner run testrunner java at org testng suiterunner runtest suiterunner java at org testng suiterunner runsequentially suiterunner java at org testng suiterunner privaterun suiterunner java at org testng suiterunner run suiterunner java at org testng suiterunnerworker runsuite suiterunnerworker java at org testng suiterunnerworker run suiterunnerworker java at org testng testng runsuitessequentially testng java at org testng testng runsuiteslocally testng java at org testng testng runsuites testng java at org testng testng run testng java at org apache maven surefire testng testngexecutor run testngexecutor java at org apache maven surefire testng testngdirectorytestsuite executesingleclass testngdirectorytestsuite java at org apache maven surefire testng testngdirectorytestsuite executelazy testngdirectorytestsuite java at org apache maven surefire testng testngdirectorytestsuite execute testngdirectorytestsuite java at org apache maven surefire testng testngprovider invoke testngprovider java at org apache maven surefire booter forkedbooter invokeproviderinsameclassloader forkedbooter java at org apache maven surefire booter forkedbooter runsuitesinprocess forkedbooter java at org apache maven surefire booter forkedbooter execute forkedbooter java at org apache maven surefire booter forkedbooter main forkedbooter java
1
151,887
5,829,394,478
IssuesEvent
2017-05-08 14:29:31
dotkom/onlineweb4
https://api.github.com/repos/dotkom/onlineweb4
closed
Not possible to edit event in Django administration
Location: Admin Package: Event Priority: High Status: Available Type: Bug
## What kind of an issue is this? - [x] Bug report ## What is the expected behaviour? It should be possible to save an event on the event administration page, having filled out all the required fields, and make it not produce a validation error. ## What is the current behaviour? When saving an event in the administration page it does not validate properly, so saving fails. ## How do you reproduce this problem? Try to create an event on the administration page, or edit an existing event. ## Other information Missing all relevant fields for an event, but all the inline fields are included. ![screenshot 2017-03-07 14 43 29](https://cloud.githubusercontent.com/assets/5422571/23658908/8c8c2d7a-0344-11e7-9fe8-06afdc047ceb.png)
1.0
Not possible to edit event in Django administration - ## What kind of an issue is this? - [x] Bug report ## What is the expected behaviour? It should be possible to save an event on the event administration page, having filled out all the required fields, and make it not produce a validation error. ## What is the current behaviour? When saving an event in the administration page it does not validate properly, so saving fails. ## How do you reproduce this problem? Try to create an event on the administration page, or edit an existing event. ## Other information Missing all relevant fields for an event, but all the inline fields are included. ![screenshot 2017-03-07 14 43 29](https://cloud.githubusercontent.com/assets/5422571/23658908/8c8c2d7a-0344-11e7-9fe8-06afdc047ceb.png)
non_test
not possible to edit event in django administration what kind of an issue is this bug report what is the expected behaviour it should be possible to save an event on the event administration page having filled out all the required fields and make it not produce a validation error what is the current behaviour when saving an event in the administration page it does not validate properly so saving fails how do you reproduce this problem try to create an event on the administration page or edit an existing event other information missing all relevant fields for an event but all the inline fields are included
0
126,979
18,010,210,582
IssuesEvent
2021-09-16 07:42:53
maddyCode23/linux-4.1.15
https://api.github.com/repos/maddyCode23/linux-4.1.15
opened
CVE-2018-14611 (Medium) detected in linux-stable-rtv4.1.33
security vulnerability
## CVE-2018-14611 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p> </p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/btrfs/volumes.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/btrfs/volumes.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in the Linux kernel through 4.17.10. There is a use-after-free in try_merge_free_space() when mounting a crafted btrfs image, because of a lack of chunk type flag checks in btrfs_check_chunk_valid in fs/btrfs/volumes.c. <p>Publish Date: 2018-07-27 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14611>CVE-2018-14611</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2018-14611">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2018-14611</a></p> <p>Release Date: 2018-07-27</p> <p>Fix Resolution: v4.19-rc1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-14611 (Medium) detected in linux-stable-rtv4.1.33 - ## CVE-2018-14611 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p> </p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/btrfs/volumes.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/btrfs/volumes.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in the Linux kernel through 4.17.10. There is a use-after-free in try_merge_free_space() when mounting a crafted btrfs image, because of a lack of chunk type flag checks in btrfs_check_chunk_valid in fs/btrfs/volumes.c. <p>Publish Date: 2018-07-27 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14611>CVE-2018-14611</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2018-14611">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2018-14611</a></p> <p>Release Date: 2018-07-27</p> <p>Fix Resolution: v4.19-rc1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve medium detected in linux stable cve medium severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href vulnerable source files fs btrfs volumes c fs btrfs volumes c vulnerability details an issue was discovered in the linux kernel through there is a use after free in try merge free space when mounting a crafted btrfs image because of a lack of chunk type flag checks in btrfs check chunk valid in fs btrfs volumes c publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
29,677
8,390,165,306
IssuesEvent
2018-10-09 11:49:58
carla-simulator/carla
https://api.github.com/repos/carla-simulator/carla
closed
Segmentation Fault while executing CarlaUE4.sh
build system support
Hi, While executing ./CarlaUE4.sh in Xubuntu to run Carla the terminal prints me the following error: _Increasing per-process limit of core file size to infinity. Signal 11 caught. Malloc Size=131076 LargeMemoryPoolOffset=131092 CommonLinuxCrashHandler: Signal=11 Malloc Size=65535 LargeMemoryPoolOffset=196655 Malloc Size=78304 LargeMemoryPoolOffset=274976 Segmentation fault (core dumped)_ Do you know what's happening and why do I get this error?
1.0
Segmentation Fault while executing CarlaUE4.sh - Hi, While executing ./CarlaUE4.sh in Xubuntu to run Carla the terminal prints me the following error: _Increasing per-process limit of core file size to infinity. Signal 11 caught. Malloc Size=131076 LargeMemoryPoolOffset=131092 CommonLinuxCrashHandler: Signal=11 Malloc Size=65535 LargeMemoryPoolOffset=196655 Malloc Size=78304 LargeMemoryPoolOffset=274976 Segmentation fault (core dumped)_ Do you know what's happening and why do I get this error?
non_test
segmentation fault while executing sh hi while executing sh in xubuntu to run carla the terminal prints me the following error increasing per process limit of core file size to infinity signal caught malloc size largememorypooloffset commonlinuxcrashhandler signal malloc size largememorypooloffset malloc size largememorypooloffset segmentation fault core dumped do you know what s happening and why do i get this error
0
246,431
20,864,564,510
IssuesEvent
2022-03-22 05:01:42
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachtest: clock/jump/large_backward_enabled failed [Span use after finish]
C-test-failure O-robot O-roachtest branch-master release-blocker T-kv
roachtest.clock/jump/large_backward_enabled [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=4389357&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=4389357&tab=artifacts#/clock/jump/large_backward_enabled) on master @ [6d9ceec64fa120113c484d17e3f61dfd6737b221](https://github.com/cockroachdb/cockroach/commits/6d9ceec64fa120113c484d17e3f61dfd6737b221): ``` | runtime.gopanic | GOROOT/src/runtime/panic.go:1038 | main.(*testImpl).Fatalf | main/pkg/cmd/roachtest/test_impl.go:254 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.runClockJump | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/clock_jump_crash.go:86 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.registerClockJumpTests.func1 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/clock_jump_crash.go:143 | main.(*testRunner).runTest.func2 | main/pkg/cmd/roachtest/test_runner.go:779 | runtime.goexit | GOROOT/src/runtime/asm_amd64.s:1581 Wraps: (2) cluster.RunE Wraps: (3) context canceled Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *errors.errorString cluster.go:1981,clock_util.go:112,clock_jump_crash.go:67,panic.go:1038,test_impl.go:248,cluster.go:1981,clock_util.go:112,panic.go:1038,test_impl.go:254,clock_jump_crash.go:86,clock_jump_crash.go:143,test_runner.go:779: cluster.RunE: context canceled (1) attached stack trace -- stack trace: | main.(*clusterImpl).RunE | main/pkg/cmd/roachtest/cluster.go:1998 | main.(*clusterImpl).Run | main/pkg/cmd/roachtest/cluster.go:1979 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.(*offsetInjector).recover | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/clock_util.go:112 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.runClockJump.func1 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/clock_jump_crash.go:67 | runtime.gopanic | GOROOT/src/runtime/panic.go:1038 | main.(*testImpl).Fatal | main/pkg/cmd/roachtest/test_impl.go:248 | main.(*clusterImpl).Run | main/pkg/cmd/roachtest/cluster.go:1981 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.(*offsetInjector).recover | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/clock_util.go:112 | runtime.gopanic | GOROOT/src/runtime/panic.go:1038 | main.(*testImpl).Fatalf | main/pkg/cmd/roachtest/test_impl.go:254 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.runClockJump | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/clock_jump_crash.go:86 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.registerClockJumpTests.func1 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/clock_jump_crash.go:143 | main.(*testRunner).runTest.func2 | main/pkg/cmd/roachtest/test_runner.go:779 | runtime.goexit | GOROOT/src/runtime/asm_amd64.s:1581 Wraps: (2) cluster.RunE Wraps: (3) context canceled Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *errors.errorString ``` <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/kv-triage <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*clock/jump/large_backward_enabled.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-13204
2.0
roachtest: clock/jump/large_backward_enabled failed [Span use after finish] - roachtest.clock/jump/large_backward_enabled [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=4389357&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=4389357&tab=artifacts#/clock/jump/large_backward_enabled) on master @ [6d9ceec64fa120113c484d17e3f61dfd6737b221](https://github.com/cockroachdb/cockroach/commits/6d9ceec64fa120113c484d17e3f61dfd6737b221): ``` | runtime.gopanic | GOROOT/src/runtime/panic.go:1038 | main.(*testImpl).Fatalf | main/pkg/cmd/roachtest/test_impl.go:254 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.runClockJump | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/clock_jump_crash.go:86 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.registerClockJumpTests.func1 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/clock_jump_crash.go:143 | main.(*testRunner).runTest.func2 | main/pkg/cmd/roachtest/test_runner.go:779 | runtime.goexit | GOROOT/src/runtime/asm_amd64.s:1581 Wraps: (2) cluster.RunE Wraps: (3) context canceled Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *errors.errorString cluster.go:1981,clock_util.go:112,clock_jump_crash.go:67,panic.go:1038,test_impl.go:248,cluster.go:1981,clock_util.go:112,panic.go:1038,test_impl.go:254,clock_jump_crash.go:86,clock_jump_crash.go:143,test_runner.go:779: cluster.RunE: context canceled (1) attached stack trace -- stack trace: | main.(*clusterImpl).RunE | main/pkg/cmd/roachtest/cluster.go:1998 | main.(*clusterImpl).Run | main/pkg/cmd/roachtest/cluster.go:1979 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.(*offsetInjector).recover | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/clock_util.go:112 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.runClockJump.func1 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/clock_jump_crash.go:67 | runtime.gopanic | GOROOT/src/runtime/panic.go:1038 | main.(*testImpl).Fatal | main/pkg/cmd/roachtest/test_impl.go:248 | main.(*clusterImpl).Run | main/pkg/cmd/roachtest/cluster.go:1981 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.(*offsetInjector).recover | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/clock_util.go:112 | runtime.gopanic | GOROOT/src/runtime/panic.go:1038 | main.(*testImpl).Fatalf | main/pkg/cmd/roachtest/test_impl.go:254 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.runClockJump | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/clock_jump_crash.go:86 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.registerClockJumpTests.func1 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/clock_jump_crash.go:143 | main.(*testRunner).runTest.func2 | main/pkg/cmd/roachtest/test_runner.go:779 | runtime.goexit | GOROOT/src/runtime/asm_amd64.s:1581 Wraps: (2) cluster.RunE Wraps: (3) context canceled Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *errors.errorString ``` <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/kv-triage <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*clock/jump/large_backward_enabled.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-13204
test
roachtest clock jump large backward enabled failed roachtest clock jump large backward enabled with on master runtime gopanic goroot src runtime panic go main testimpl fatalf main pkg cmd roachtest test impl go github com cockroachdb cockroach pkg cmd roachtest tests runclockjump github com cockroachdb cockroach pkg cmd roachtest tests clock jump crash go github com cockroachdb cockroach pkg cmd roachtest tests registerclockjumptests github com cockroachdb cockroach pkg cmd roachtest tests clock jump crash go main testrunner runtest main pkg cmd roachtest test runner go runtime goexit goroot src runtime asm s wraps cluster rune wraps context canceled error types withstack withstack errutil withprefix errors errorstring cluster go clock util go clock jump crash go panic go test impl go cluster go clock util go panic go test impl go clock jump crash go clock jump crash go test runner go cluster rune context canceled attached stack trace stack trace main clusterimpl rune main pkg cmd roachtest cluster go main clusterimpl run main pkg cmd roachtest cluster go github com cockroachdb cockroach pkg cmd roachtest tests offsetinjector recover github com cockroachdb cockroach pkg cmd roachtest tests clock util go github com cockroachdb cockroach pkg cmd roachtest tests runclockjump github com cockroachdb cockroach pkg cmd roachtest tests clock jump crash go runtime gopanic goroot src runtime panic go main testimpl fatal main pkg cmd roachtest test impl go main clusterimpl run main pkg cmd roachtest cluster go github com cockroachdb cockroach pkg cmd roachtest tests offsetinjector recover github com cockroachdb cockroach pkg cmd roachtest tests clock util go runtime gopanic goroot src runtime panic go main testimpl fatalf main pkg cmd roachtest test impl go github com cockroachdb cockroach pkg cmd roachtest tests runclockjump github com cockroachdb cockroach pkg cmd roachtest tests clock jump crash go github com cockroachdb cockroach pkg cmd roachtest tests registerclockjumptests github com cockroachdb cockroach pkg cmd roachtest tests clock jump crash go main testrunner runtest main pkg cmd roachtest test runner go runtime goexit goroot src runtime asm s wraps cluster rune wraps context canceled error types withstack withstack errutil withprefix errors errorstring help see see cc cockroachdb kv triage jira issue crdb
1
95,794
8,577,751,302
IssuesEvent
2018-11-13 01:30:58
ODM2/ODM2ControlledVocabularies
https://api.github.com/repos/ODM2/ODM2ControlledVocabularies
closed
Need to modify term view page to be a little more consistent
ready for testing
On the view for an an individual term (e.g., http://vocabulary.odm2.org/actiontype/cruise/), the text that lists the term and it's definition at the top is small. I would prefer that these just be included in the table with the other attributes as the first two rows. This should be implemented across the views for all of the terms.
1.0
Need to modify term view page to be a little more consistent - On the view for an an individual term (e.g., http://vocabulary.odm2.org/actiontype/cruise/), the text that lists the term and it's definition at the top is small. I would prefer that these just be included in the table with the other attributes as the first two rows. This should be implemented across the views for all of the terms.
test
need to modify term view page to be a little more consistent on the view for an an individual term e g the text that lists the term and it s definition at the top is small i would prefer that these just be included in the table with the other attributes as the first two rows this should be implemented across the views for all of the terms
1
22,637
2,649,607,801
IssuesEvent
2015-03-15 03:00:12
garysoed/protoboard
https://api.github.com/repos/garysoed/protoboard
closed
Reword rect grid ... again
feature priority-high
Should support scrolling and infinite dimension. To support games like Among the Stars
1.0
Reword rect grid ... again - Should support scrolling and infinite dimension. To support games like Among the Stars
non_test
reword rect grid again should support scrolling and infinite dimension to support games like among the stars
0
344,480
30,748,341,423
IssuesEvent
2023-07-28 16:51:30
elastic/kibana
https://api.github.com/repos/elastic/kibana
opened
Failing ES Promotion: FTR Configs #28 / Agents fleet_uploads should get agent uploads
bug blocker skipped-test failed-es-promotion Team:Fleet v8.10.0
**X-Pack Fleet Agent API Integration Tests x-pack/test/fleet_api_integration/apis/agents/uploads.ts** **Agents fleet_uploads should get agent uploads** This failure is preventing the promotion of the current Elasticsearch snapshot. For more information on the Elasticsearch snapshot process including how to reproduce using the unverified ES build please read the [failed promotion annotation](https://buildkite.com/elastic/kibana-elasticsearch-snapshot-verify/builds/2891#annotation-es-snapshot-manifest). Other important information can be found at: * [Failed promotion job](https://buildkite.com/elastic/kibana-elasticsearch-snapshot-verify/builds/2891) * [Test Failure](https://buildkite.com/elastic/kibana-elasticsearch-snapshot-verify/builds/2891#01899ce1-58bd-44ef-92d4-456150cdc507) ``` Error: expected 200 "OK", got 500 "Internal Server Error" at Test._assertStatus (node_modules/supertest/lib/test.js:268:12) at Test._assertFunction (node_modules/supertest/lib/test.js:283:11) at Test.assert (node_modules/supertest/lib/test.js:173:18) at localAssert (node_modules/supertest/lib/test.js:131:12) at /var/lib/buildkite-agent/builds/kb-n2-4-spot-c783f087a87149c4/elastic/kibana-elasticsearch-snapshot-verify/kibana/node_modules/supertest/lib/test.js:128:5 at Test.Request.callback (node_modules/superagent/lib/node/index.js:728:3) at /var/lib/buildkite-agent/builds/kb-n2-4-spot-c783f087a87149c4/elastic/kibana-elasticsearch-snapshot-verify/kibana/node_modules/superagent/lib/node/index.js:916:18 at IncomingMessage.<anonymous> (node_modules/superagent/lib/node/parsers/json.js:19:7) at IncomingMessage.emit (node:events:526:35) at endReadableNT (node:internal/streams/readable:1359:12) at processTicksAndRejections (node:internal/process/task_queues:82:21) ```
1.0
Failing ES Promotion: FTR Configs #28 / Agents fleet_uploads should get agent uploads - **X-Pack Fleet Agent API Integration Tests x-pack/test/fleet_api_integration/apis/agents/uploads.ts** **Agents fleet_uploads should get agent uploads** This failure is preventing the promotion of the current Elasticsearch snapshot. For more information on the Elasticsearch snapshot process including how to reproduce using the unverified ES build please read the [failed promotion annotation](https://buildkite.com/elastic/kibana-elasticsearch-snapshot-verify/builds/2891#annotation-es-snapshot-manifest). Other important information can be found at: * [Failed promotion job](https://buildkite.com/elastic/kibana-elasticsearch-snapshot-verify/builds/2891) * [Test Failure](https://buildkite.com/elastic/kibana-elasticsearch-snapshot-verify/builds/2891#01899ce1-58bd-44ef-92d4-456150cdc507) ``` Error: expected 200 "OK", got 500 "Internal Server Error" at Test._assertStatus (node_modules/supertest/lib/test.js:268:12) at Test._assertFunction (node_modules/supertest/lib/test.js:283:11) at Test.assert (node_modules/supertest/lib/test.js:173:18) at localAssert (node_modules/supertest/lib/test.js:131:12) at /var/lib/buildkite-agent/builds/kb-n2-4-spot-c783f087a87149c4/elastic/kibana-elasticsearch-snapshot-verify/kibana/node_modules/supertest/lib/test.js:128:5 at Test.Request.callback (node_modules/superagent/lib/node/index.js:728:3) at /var/lib/buildkite-agent/builds/kb-n2-4-spot-c783f087a87149c4/elastic/kibana-elasticsearch-snapshot-verify/kibana/node_modules/superagent/lib/node/index.js:916:18 at IncomingMessage.<anonymous> (node_modules/superagent/lib/node/parsers/json.js:19:7) at IncomingMessage.emit (node:events:526:35) at endReadableNT (node:internal/streams/readable:1359:12) at processTicksAndRejections (node:internal/process/task_queues:82:21) ```
test
failing es promotion ftr configs agents fleet uploads should get agent uploads x pack fleet agent api integration tests x pack test fleet api integration apis agents uploads ts agents fleet uploads should get agent uploads this failure is preventing the promotion of the current elasticsearch snapshot for more information on the elasticsearch snapshot process including how to reproduce using the unverified es build please read the other important information can be found at error expected ok got internal server error at test assertstatus node modules supertest lib test js at test assertfunction node modules supertest lib test js at test assert node modules supertest lib test js at localassert node modules supertest lib test js at var lib buildkite agent builds kb spot elastic kibana elasticsearch snapshot verify kibana node modules supertest lib test js at test request callback node modules superagent lib node index js at var lib buildkite agent builds kb spot elastic kibana elasticsearch snapshot verify kibana node modules superagent lib node index js at incomingmessage node modules superagent lib node parsers json js at incomingmessage emit node events at endreadablent node internal streams readable at processticksandrejections node internal process task queues
1
708,305
24,337,298,176
IssuesEvent
2022-10-01 08:23:46
EndurionNetwork/Endurion
https://api.github.com/repos/EndurionNetwork/Endurion
closed
[Bug] Message when performing a moderating action is not shown correctly in the Discord under #automod-log
bug: open priority: low
![grafik](https://user-images.githubusercontent.com/36358673/188270891-98a6d0b4-85af-419a-9f33-30f665150358.png) Triton does not seem to work, in addition the chat formatting does not work.
1.0
[Bug] Message when performing a moderating action is not shown correctly in the Discord under #automod-log - ![grafik](https://user-images.githubusercontent.com/36358673/188270891-98a6d0b4-85af-419a-9f33-30f665150358.png) Triton does not seem to work, in addition the chat formatting does not work.
non_test
message when performing a moderating action is not shown correctly in the discord under automod log triton does not seem to work in addition the chat formatting does not work
0
253,710
27,300,805,754
IssuesEvent
2023-02-24 01:39:35
panasalap/linux-4.19.72_1
https://api.github.com/repos/panasalap/linux-4.19.72_1
closed
CVE-2022-4379 (High) detected in linux-yoctov5.4.51 - autoclosed
security vulnerability
## CVE-2022-4379 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yoctov5.4.51</b></p></summary> <p> <p>Yocto Linux Embedded kernel</p> <p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto>https://git.yoctoproject.org/git/linux-yocto</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/nfsd/nfs4proc.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/nfsd/nfs4proc.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A use-after-free vulnerability was found in __nfs42_ssc_open() in fs/nfs/nfs4file.c in the Linux kernel. This flaw allows an attacker to conduct a remote denial <p>Publish Date: 2023-01-10 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-4379>CVE-2022-4379</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-4379">https://www.linuxkernelcves.com/cves/CVE-2022-4379</a></p> <p>Release Date: 2023-01-10</p> <p>Fix Resolution: v6.1.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-4379 (High) detected in linux-yoctov5.4.51 - autoclosed - ## CVE-2022-4379 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yoctov5.4.51</b></p></summary> <p> <p>Yocto Linux Embedded kernel</p> <p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto>https://git.yoctoproject.org/git/linux-yocto</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/nfsd/nfs4proc.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/nfsd/nfs4proc.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A use-after-free vulnerability was found in __nfs42_ssc_open() in fs/nfs/nfs4file.c in the Linux kernel. This flaw allows an attacker to conduct a remote denial <p>Publish Date: 2023-01-10 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-4379>CVE-2022-4379</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-4379">https://www.linuxkernelcves.com/cves/CVE-2022-4379</a></p> <p>Release Date: 2023-01-10</p> <p>Fix Resolution: v6.1.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in linux autoclosed cve high severity vulnerability vulnerable library linux yocto linux embedded kernel library home page a href found in base branch master vulnerable source files fs nfsd c fs nfsd c vulnerability details a use after free vulnerability was found in ssc open in fs nfs c in the linux kernel this flaw allows an attacker to conduct a remote denial publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
6,609
2,853,756,517
IssuesEvent
2015-06-01 20:14:52
kdahlquist/GRNmap
https://api.github.com/repos/kdahlquist/GRNmap
reopened
Try all 16 possible combinations of input sheets
priority 0 testing
Combined with the 3 test cases from @bengfitzpatrick, make dummy cases of excel sheets with different combinations (e.g fix b, est p, graph, or fix b, fix p, no graph, etc.)
1.0
Try all 16 possible combinations of input sheets - Combined with the 3 test cases from @bengfitzpatrick, make dummy cases of excel sheets with different combinations (e.g fix b, est p, graph, or fix b, fix p, no graph, etc.)
test
try all possible combinations of input sheets combined with the test cases from bengfitzpatrick make dummy cases of excel sheets with different combinations e g fix b est p graph or fix b fix p no graph etc
1
84,698
10,554,549,085
IssuesEvent
2019-10-03 19:43:38
telus/tds-community
https://api.github.com/repos/telus/tds-community
opened
Input enhancement: CanadaPost AutoComplete dropdown
priority: medium status: DPA design review needed type: feature
<!-- ### IMPORTANT SECURITY NOTE ### When opening issues, be sure NOT to include any private or personal information such as secrets, passwords, or any source code that involves data retrieval. Also, do not include links to sites on staging. --> ## Problem statement As a customer, I would like to enter a Canadian address so that I can complete an e-commerce flow (e.g. checkout) process, enter an address to access products, and/or complete other activities on T.com that require address completion. In the case where an address is not found on CanadaPost's AutoComplete, I would like an option to manually enter the address as well. ## Recommendation The UPC team is building out and would like to contribute a reusable component using the CanadaPost AutoComplete dropdown logic, which is attached to a TDS-core input component. ## Design intent <!-- Note: designs must be approved by the Digital Platform Ambassadors Before any code or pull requests can be submitted As a new customer, I want an Address AutoComplete so that I can enter it in my checkout flow as a billing, shipping, and/or service address. As a business customer, I want an Address AutoComplete so that I can enter it as a business/service address. As a new customer, I want an Address AutoComplete so that I can view Home Service services applicable to my location. --> A customer should be allowed to type in and select their address simply from a list instead of inputting street/unit #, street name, postal code etc. manually. If the address is not found by CanadaPost, there should also be a method for them to input the necessary fields to complete an address. ## Designs - https://telus.invisionapp.com/share/FCU82AE68XT - https://telusdigital.atlassian.net/browse/TELCOM-1370 ## Acceptance criteria <!-- TELUS guideline on writing acceptance criteria for development and QA Can be found on the Reference Architecture: https://github.com/telus/reference-architecture/blob/master/process/user-stories.md#how DO NOT REMOVE THIS SECTION Things to consider when writing acceptance criteria (ACs): Things to think about: @yinbai -> What needs to be done to make this accessible? A: The input should already have a11y methods built in. The dropdown will be treated differently with a spec provided for tab order & script. -> Visual properties A: Only outstanding question would be hover colour for dropdown list -> Interactions A: N/A -> Documentation A: WIP -> Considerations if changes need to be mirrored to tds-core and tds-community A: IDK -> Are there any responsive behaviours or visual differences? A: It is responsive; the dropdown list text stacks -> French considerations, do symbols or text need to be rendered in a certain order or manner? A: IDK -> Text considerations, do we need SVGs, symbols, superscript text, or other special formatting? A: Powered By CanadaPost should be included in the dropdown -> Make sure the ACs do not overly specify names of props. Try to stay generic and descriptive. Describe the interactions and behaviours of your component. Given... When... (And...) Then... (And...) Example: Given I am displaying the community-testimonial component When I supply an image And I supply testimonial text And I supply info Then the testimonial is displayed within a card with testimonial text And the testimonial image is displayed to the left of its info When I do not supply an image And I supply testimonial text And I supply info Then the testimonial is displayed within a card with testimonial text --> This section is intentionally left blank. See [story writing process](https://github.com/telus/tds-community/blob/master/guide/Community_Backlog.md#5-story-writing). ### User stories (This section is intentionally left blank. User stories and acceptance criteria go here.) **Example:** Scenario: Displaying the component Given text is provided Then the text is displayed as a child And the text size matches core-text `medium` ### Out of scope N/A ## Meta - Include component name and version: vX.Y.Z - Willing to design solution: **Yes**/No - Willing to develop solution: **Yes**/No - Has workaround: Yes/No - Do any similar components already exist in TDS (core or community): **Yes**/No (https://tds.telus.com/components/index.html#input) - Which teams/applications do you think would use this improvement? **all** - High impact: Yes/No
1.0
Input enhancement: CanadaPost AutoComplete dropdown - <!-- ### IMPORTANT SECURITY NOTE ### When opening issues, be sure NOT to include any private or personal information such as secrets, passwords, or any source code that involves data retrieval. Also, do not include links to sites on staging. --> ## Problem statement As a customer, I would like to enter a Canadian address so that I can complete an e-commerce flow (e.g. checkout) process, enter an address to access products, and/or complete other activities on T.com that require address completion. In the case where an address is not found on CanadaPost's AutoComplete, I would like an option to manually enter the address as well. ## Recommendation The UPC team is building out and would like to contribute a reusable component using the CanadaPost AutoComplete dropdown logic, which is attached to a TDS-core input component. ## Design intent <!-- Note: designs must be approved by the Digital Platform Ambassadors Before any code or pull requests can be submitted As a new customer, I want an Address AutoComplete so that I can enter it in my checkout flow as a billing, shipping, and/or service address. As a business customer, I want an Address AutoComplete so that I can enter it as a business/service address. As a new customer, I want an Address AutoComplete so that I can view Home Service services applicable to my location. --> A customer should be allowed to type in and select their address simply from a list instead of inputting street/unit #, street name, postal code etc. manually. If the address is not found by CanadaPost, there should also be a method for them to input the necessary fields to complete an address. ## Designs - https://telus.invisionapp.com/share/FCU82AE68XT - https://telusdigital.atlassian.net/browse/TELCOM-1370 ## Acceptance criteria <!-- TELUS guideline on writing acceptance criteria for development and QA Can be found on the Reference Architecture: https://github.com/telus/reference-architecture/blob/master/process/user-stories.md#how DO NOT REMOVE THIS SECTION Things to consider when writing acceptance criteria (ACs): Things to think about: @yinbai -> What needs to be done to make this accessible? A: The input should already have a11y methods built in. The dropdown will be treated differently with a spec provided for tab order & script. -> Visual properties A: Only outstanding question would be hover colour for dropdown list -> Interactions A: N/A -> Documentation A: WIP -> Considerations if changes need to be mirrored to tds-core and tds-community A: IDK -> Are there any responsive behaviours or visual differences? A: It is responsive; the dropdown list text stacks -> French considerations, do symbols or text need to be rendered in a certain order or manner? A: IDK -> Text considerations, do we need SVGs, symbols, superscript text, or other special formatting? A: Powered By CanadaPost should be included in the dropdown -> Make sure the ACs do not overly specify names of props. Try to stay generic and descriptive. Describe the interactions and behaviours of your component. Given... When... (And...) Then... (And...) Example: Given I am displaying the community-testimonial component When I supply an image And I supply testimonial text And I supply info Then the testimonial is displayed within a card with testimonial text And the testimonial image is displayed to the left of its info When I do not supply an image And I supply testimonial text And I supply info Then the testimonial is displayed within a card with testimonial text --> This section is intentionally left blank. See [story writing process](https://github.com/telus/tds-community/blob/master/guide/Community_Backlog.md#5-story-writing). ### User stories (This section is intentionally left blank. User stories and acceptance criteria go here.) **Example:** Scenario: Displaying the component Given text is provided Then the text is displayed as a child And the text size matches core-text `medium` ### Out of scope N/A ## Meta - Include component name and version: vX.Y.Z - Willing to design solution: **Yes**/No - Willing to develop solution: **Yes**/No - Has workaround: Yes/No - Do any similar components already exist in TDS (core or community): **Yes**/No (https://tds.telus.com/components/index.html#input) - Which teams/applications do you think would use this improvement? **all** - High impact: Yes/No
non_test
input enhancement canadapost autocomplete dropdown important security note when opening issues be sure not to include any private or personal information such as secrets passwords or any source code that involves data retrieval also do not include links to sites on staging problem statement as a customer i would like to enter a canadian address so that i can complete an e commerce flow e g checkout process enter an address to access products and or complete other activities on t com that require address completion in the case where an address is not found on canadapost s autocomplete i would like an option to manually enter the address as well recommendation the upc team is building out and would like to contribute a reusable component using the canadapost autocomplete dropdown logic which is attached to a tds core input component design intent note designs must be approved by the digital platform ambassadors before any code or pull requests can be submitted as a new customer i want an address autocomplete so that i can enter it in my checkout flow as a billing shipping and or service address as a business customer i want an address autocomplete so that i can enter it as a business service address as a new customer i want an address autocomplete so that i can view home service services applicable to my location a customer should be allowed to type in and select their address simply from a list instead of inputting street unit street name postal code etc manually if the address is not found by canadapost there should also be a method for them to input the necessary fields to complete an address designs acceptance criteria telus guideline on writing acceptance criteria for development and qa can be found on the reference architecture do not remove this section things to consider when writing acceptance criteria acs things to think about yinbai what needs to be done to make this accessible a the input should already have methods built in the dropdown will be treated differently with a spec provided for tab order script visual properties a only outstanding question would be hover colour for dropdown list interactions a n a documentation a wip considerations if changes need to be mirrored to tds core and tds community a idk are there any responsive behaviours or visual differences a it is responsive the dropdown list text stacks french considerations do symbols or text need to be rendered in a certain order or manner a idk text considerations do we need svgs symbols superscript text or other special formatting a powered by canadapost should be included in the dropdown make sure the acs do not overly specify names of props try to stay generic and descriptive describe the interactions and behaviours of your component given when and then and example given i am displaying the community testimonial component when i supply an image and i supply testimonial text and i supply info then the testimonial is displayed within a card with testimonial text and the testimonial image is displayed to the left of its info when i do not supply an image and i supply testimonial text and i supply info then the testimonial is displayed within a card with testimonial text this section is intentionally left blank see user stories this section is intentionally left blank user stories and acceptance criteria go here example scenario displaying the component given text is provided then the text is displayed as a child and the text size matches core text medium out of scope n a meta include component name and version vx y z willing to design solution yes no willing to develop solution yes no has workaround yes no do any similar components already exist in tds core or community yes no which teams applications do you think would use this improvement all high impact yes no
0
688,474
23,583,871,208
IssuesEvent
2022-08-23 09:53:49
naev/naev
https://api.github.com/repos/naev/naev
closed
Point of Interest Expansion
Priority-Low
This is mainly an issue for tracking the upcoming 0.10.0 release. With points of interest being implemented and very modular in nature, it's time to flesh it a bit more out for the release. In particular the following things would be nice to have: 1. More events covering different aspects of the Naev Lore. In particular, things like history of different factions and things about cultures could be interesting to have. 2. Repeatable events that give some sort of currency to allow the player to get special outfits and/or ships. Maybe having Misi accept the tokens in exchange for things could be useful. Non-repeatable events could also give tokens instead of outfits directly so that the rewards don't get diluted as many POI get added.
1.0
Point of Interest Expansion - This is mainly an issue for tracking the upcoming 0.10.0 release. With points of interest being implemented and very modular in nature, it's time to flesh it a bit more out for the release. In particular the following things would be nice to have: 1. More events covering different aspects of the Naev Lore. In particular, things like history of different factions and things about cultures could be interesting to have. 2. Repeatable events that give some sort of currency to allow the player to get special outfits and/or ships. Maybe having Misi accept the tokens in exchange for things could be useful. Non-repeatable events could also give tokens instead of outfits directly so that the rewards don't get diluted as many POI get added.
non_test
point of interest expansion this is mainly an issue for tracking the upcoming release with points of interest being implemented and very modular in nature it s time to flesh it a bit more out for the release in particular the following things would be nice to have more events covering different aspects of the naev lore in particular things like history of different factions and things about cultures could be interesting to have repeatable events that give some sort of currency to allow the player to get special outfits and or ships maybe having misi accept the tokens in exchange for things could be useful non repeatable events could also give tokens instead of outfits directly so that the rewards don t get diluted as many poi get added
0
204,316
23,239,476,890
IssuesEvent
2022-08-03 14:28:22
turkdevops/angular
https://api.github.com/repos/turkdevops/angular
closed
CVE-2022-21680 (High) detected in marked-0.7.0.tgz - autoclosed
security vulnerability
## CVE-2022-21680 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>marked-0.7.0.tgz</b></p></summary> <p>A markdown parser built for speed</p> <p>Library home page: <a href="https://registry.npmjs.org/marked/-/marked-0.7.0.tgz">https://registry.npmjs.org/marked/-/marked-0.7.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/marked/package.json,/aio/node_modules/marked/package.json</p> <p> Dependency Hierarchy: - firebase-tools-7.16.2.tgz (Root Library) - :x: **marked-0.7.0.tgz** (Vulnerable Library) <p>Found in base branch: <b>labs/router</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Marked is a markdown parser and compiler. Prior to version 4.0.10, the regular expression `block.def` may cause catastrophic backtracking against some strings and lead to a regular expression denial of service (ReDoS). Anyone who runs untrusted markdown through a vulnerable version of marked and does not use a worker with a time limit may be affected. This issue is patched in version 4.0.10. As a workaround, avoid running untrusted markdown through marked or run marked on a worker thread and set a reasonable time limit to prevent draining resources. <p>Publish Date: 2022-01-14 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-21680>CVE-2022-21680</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-rrrm-qjm4-v8hf">https://github.com/advisories/GHSA-rrrm-qjm4-v8hf</a></p> <p>Release Date: 2022-01-14</p> <p>Fix Resolution (marked): 4.0.10</p> <p>Direct dependency fix Resolution (firebase-tools): 10.1.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-21680 (High) detected in marked-0.7.0.tgz - autoclosed - ## CVE-2022-21680 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>marked-0.7.0.tgz</b></p></summary> <p>A markdown parser built for speed</p> <p>Library home page: <a href="https://registry.npmjs.org/marked/-/marked-0.7.0.tgz">https://registry.npmjs.org/marked/-/marked-0.7.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/marked/package.json,/aio/node_modules/marked/package.json</p> <p> Dependency Hierarchy: - firebase-tools-7.16.2.tgz (Root Library) - :x: **marked-0.7.0.tgz** (Vulnerable Library) <p>Found in base branch: <b>labs/router</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Marked is a markdown parser and compiler. Prior to version 4.0.10, the regular expression `block.def` may cause catastrophic backtracking against some strings and lead to a regular expression denial of service (ReDoS). Anyone who runs untrusted markdown through a vulnerable version of marked and does not use a worker with a time limit may be affected. This issue is patched in version 4.0.10. As a workaround, avoid running untrusted markdown through marked or run marked on a worker thread and set a reasonable time limit to prevent draining resources. <p>Publish Date: 2022-01-14 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-21680>CVE-2022-21680</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-rrrm-qjm4-v8hf">https://github.com/advisories/GHSA-rrrm-qjm4-v8hf</a></p> <p>Release Date: 2022-01-14</p> <p>Fix Resolution (marked): 4.0.10</p> <p>Direct dependency fix Resolution (firebase-tools): 10.1.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in marked tgz autoclosed cve high severity vulnerability vulnerable library marked tgz a markdown parser built for speed library home page a href path to dependency file package json path to vulnerable library node modules marked package json aio node modules marked package json dependency hierarchy firebase tools tgz root library x marked tgz vulnerable library found in base branch labs router vulnerability details marked is a markdown parser and compiler prior to version the regular expression block def may cause catastrophic backtracking against some strings and lead to a regular expression denial of service redos anyone who runs untrusted markdown through a vulnerable version of marked and does not use a worker with a time limit may be affected this issue is patched in version as a workaround avoid running untrusted markdown through marked or run marked on a worker thread and set a reasonable time limit to prevent draining resources publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution marked direct dependency fix resolution firebase tools step up your open source security game with mend
0
194,321
14,675,668,293
IssuesEvent
2020-12-30 18:10:24
github-vet/rangeloop-pointer-findings
https://api.github.com/repos/github-vet/rangeloop-pointer-findings
closed
remind101/pkg: httpx/middleware/r101newrelic/tracer_test.go; 3 LoC
fresh test tiny
Found a possible issue in [remind101/pkg](https://www.github.com/remind101/pkg) at [httpx/middleware/r101newrelic/tracer_test.go](https://github.com/remind101/pkg/blob/fc2554d37998d796b28af24e5f4dc769fa691b82/httpx/middleware/r101newrelic/tracer_test.go#L82-L84) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call which takes a reference to tt at line 83 may start a goroutine [Click here to see the code in its original context.](https://github.com/remind101/pkg/blob/fc2554d37998d796b28af24e5f4dc769fa691b82/httpx/middleware/r101newrelic/tracer_test.go#L82-L84) <details> <summary>Click here to show the 3 line(s) of Go which triggered the analyzer.</summary> ```go for _, tt := range tracerTests { traceTest(t, &tt) } ``` </details> <details> <summary>Click here to show extra information the analyzer produced.</summary> ``` The following graphviz dot graph describes paths through the callgraph that could lead to a function calling a goroutine: digraph G { "(ServeHTTPContext, 3)" -> {} "(traceTest, 2)" -> {"(ServeHTTPContext, 3)";} } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: fc2554d37998d796b28af24e5f4dc769fa691b82
1.0
remind101/pkg: httpx/middleware/r101newrelic/tracer_test.go; 3 LoC - Found a possible issue in [remind101/pkg](https://www.github.com/remind101/pkg) at [httpx/middleware/r101newrelic/tracer_test.go](https://github.com/remind101/pkg/blob/fc2554d37998d796b28af24e5f4dc769fa691b82/httpx/middleware/r101newrelic/tracer_test.go#L82-L84) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call which takes a reference to tt at line 83 may start a goroutine [Click here to see the code in its original context.](https://github.com/remind101/pkg/blob/fc2554d37998d796b28af24e5f4dc769fa691b82/httpx/middleware/r101newrelic/tracer_test.go#L82-L84) <details> <summary>Click here to show the 3 line(s) of Go which triggered the analyzer.</summary> ```go for _, tt := range tracerTests { traceTest(t, &tt) } ``` </details> <details> <summary>Click here to show extra information the analyzer produced.</summary> ``` The following graphviz dot graph describes paths through the callgraph that could lead to a function calling a goroutine: digraph G { "(ServeHTTPContext, 3)" -> {} "(traceTest, 2)" -> {"(ServeHTTPContext, 3)";} } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: fc2554d37998d796b28af24e5f4dc769fa691b82
test
pkg httpx middleware tracer test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message function call which takes a reference to tt at line may start a goroutine click here to show the line s of go which triggered the analyzer go for tt range tracertests tracetest t tt click here to show extra information the analyzer produced the following graphviz dot graph describes paths through the callgraph that could lead to a function calling a goroutine digraph g servehttpcontext tracetest servehttpcontext leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
1
20,189
5,997,349,629
IssuesEvent
2017-06-03 23:08:50
open-austin/budgetparty
https://api.github.com/repos/open-austin/budgetparty
closed
Double values in css
code
I am not sure if this is an issue. The problem with hover not working was the 2 background-color values in the stylesheet. I was wondering if you want me to go through and remove all the double values. Example http://imgur.com/a/MoHwr Wasn't looking at sass so I might be completely off.
1.0
Double values in css - I am not sure if this is an issue. The problem with hover not working was the 2 background-color values in the stylesheet. I was wondering if you want me to go through and remove all the double values. Example http://imgur.com/a/MoHwr Wasn't looking at sass so I might be completely off.
non_test
double values in css i am not sure if this is an issue the problem with hover not working was the background color values in the stylesheet i was wondering if you want me to go through and remove all the double values example wasn t looking at sass so i might be completely off
0
67,446
7,048,203,641
IssuesEvent
2018-01-02 16:42:43
MajkiIT/polish-ads-filter
https://api.github.com/repos/MajkiIT/polish-ads-filter
closed
sokolka.tv
dodać reguły gotowe/testowanie reklama
![2018-01-02 10_34_16-wydarzenia](https://user-images.githubusercontent.com/5884000/34479690-88b7eda0-efa8-11e7-9936-d3d253369b52.png) moje filtry easylist + polskie filtry Nano Adblocker 1.0.0.23 Nano Defender 13.16 Chrome 63.0.3239.108
1.0
sokolka.tv - ![2018-01-02 10_34_16-wydarzenia](https://user-images.githubusercontent.com/5884000/34479690-88b7eda0-efa8-11e7-9936-d3d253369b52.png) moje filtry easylist + polskie filtry Nano Adblocker 1.0.0.23 Nano Defender 13.16 Chrome 63.0.3239.108
test
sokolka tv moje filtry easylist polskie filtry nano adblocker nano defender chrome
1
95,671
8,570,332,537
IssuesEvent
2018-11-11 19:13:32
vladbbb/mathApp
https://api.github.com/repos/vladbbb/mathApp
closed
Initial usability tests
medium priority test
By the end of 3rd iteration, its necessary to see how user friendly the on-implementation system is and take notes what must be fixed in 4th iteration. [Check NF requirements](https://github.com/vladbbb/mathApp/wiki/Non-functional-requirements) Estimated time: 4h
1.0
Initial usability tests - By the end of 3rd iteration, its necessary to see how user friendly the on-implementation system is and take notes what must be fixed in 4th iteration. [Check NF requirements](https://github.com/vladbbb/mathApp/wiki/Non-functional-requirements) Estimated time: 4h
test
initial usability tests by the end of iteration its necessary to see how user friendly the on implementation system is and take notes what must be fixed in iteration estimated time
1
157,365
19,957,111,645
IssuesEvent
2022-01-28 01:25:47
panasalap/linux-4.1.15
https://api.github.com/repos/panasalap/linux-4.1.15
opened
CVE-2017-14489 (Medium) detected in linuxlinux-4.1.17
security vulnerability
## CVE-2017-14489 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.1.17</b></p></summary> <p> <p>Apache Software Foundation (ASF)</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/scsi/scsi_transport_iscsi.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The iscsi_if_rx function in drivers/scsi/scsi_transport_iscsi.c in the Linux kernel through 4.13.2 allows local users to cause a denial of service (panic) by leveraging incorrect length validation. <p>Publish Date: 2017-09-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-14489>CVE-2017-14489</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-14489">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-14489</a></p> <p>Release Date: 2017-09-15</p> <p>Fix Resolution: v4.14-rc3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2017-14489 (Medium) detected in linuxlinux-4.1.17 - ## CVE-2017-14489 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.1.17</b></p></summary> <p> <p>Apache Software Foundation (ASF)</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/scsi/scsi_transport_iscsi.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The iscsi_if_rx function in drivers/scsi/scsi_transport_iscsi.c in the Linux kernel through 4.13.2 allows local users to cause a denial of service (panic) by leveraging incorrect length validation. <p>Publish Date: 2017-09-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-14489>CVE-2017-14489</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-14489">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-14489</a></p> <p>Release Date: 2017-09-15</p> <p>Fix Resolution: v4.14-rc3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve medium detected in linuxlinux cve medium severity vulnerability vulnerable library linuxlinux apache software foundation asf library home page a href found in base branch master vulnerable source files drivers scsi scsi transport iscsi c vulnerability details the iscsi if rx function in drivers scsi scsi transport iscsi c in the linux kernel through allows local users to cause a denial of service panic by leveraging incorrect length validation publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
58,033
8,224,053,580
IssuesEvent
2018-09-06 12:39:58
TimCastelijns/Room15Bot
https://api.github.com/repos/TimCastelijns/Room15Bot
closed
Provide some info about available commands and usage
documentation
Probably a help command, or notes in the readme
1.0
Provide some info about available commands and usage - Probably a help command, or notes in the readme
non_test
provide some info about available commands and usage probably a help command or notes in the readme
0
53,607
13,184,616,051
IssuesEvent
2020-08-12 19:45:58
curl/curl
https://api.github.com/repos/curl/curl
closed
Windows 10 with MinGW: Protocol "https" not supported or disabled in libcurl
Windows build not-a-bug
<!-- Only file bugs here! Ask questions on the mailing lists https://curl.haxx.se/mail/ SECURITY RELATED? Post it here: https://hackerone.com/curl There are collections of known issues to be aware of: https://curl.haxx.se/docs/knownbugs.html https://curl.haxx.se/docs/todo.html --> ### I did this I installed libcurl with the following commands: ```bash $ git clone https://github.com/curl/curl $ cd curl $ ./buildconf.bat $ mingw32-make mingw32 $ cp lib/libcurl.a /c/MinGW/lib $ cp -r include/curl/ /c/MinGW/include ``` I wrote this program to test it out: ```cpp // main.cpp #include <iostream> #include <curl/curl.h> int main(void) { CURL *curl; curl_global_init(CURL_GLOBAL_ALL); curl = curl_easy_init(); curl_easy_setopt(curl, CURLOPT_VERBOSE, 1L); curl_easy_setopt(curl, CURLOPT_URL, "https://api.ecs.rocks/v0/util/removeMessageHeader"); curl_easy_perform(curl); curl_free(curl); return 0; } ``` I built the C++ program with the following command: ```bash $ g++ main.cpp -o curltest -m32 -DCURL_STATICLIB -lcurl -lws2_32 -lwldap32 -lwsock32 -lwinmm ``` And I ran it: ```bash $ ./curltest.exe ``` ### I expected the following A response of some kind from https://api.ecs.rocks/v0/util/removeMessageHeader ### I got the following ``` * Protocol "https" not supported or disabled in libcurl * Closing connection -1 ``` ### curl/libcurl version ``` curl 7.71.1 (x86_64-w64-mingw32) libcurl/7.71.1 OpenSSL/1.1.1g (Schannel) zlib/1.2.11 brotli/1.0.7 libidn2/2.3.0 libssh2/1.9.0 nghttp2/1.41.0 Release-Date: 2020-07-01 Protocols: dict file ftp ftps gopher http https imap imaps ldap ldaps pop3 pop3s rtsp scp sftp smtp smtps telnet tftp Features: AsynchDNS brotli HTTP2 HTTPS-proxy IDN IPv6 Kerberos Largefile libz Metalink MultiSSL NTLM SPNEGO SSL SSPI TLS-SRP ``` That's the version of `curl` I have, but my version of `libcurl` should be the most recent version because I built from source. ### operating system MINGW64_NT-10.0-18363 ECSS-L01 3.1.6-340.x86_64 2020-07-09 14:33 UTC x86_64 Msys
1.0
Windows 10 with MinGW: Protocol "https" not supported or disabled in libcurl - <!-- Only file bugs here! Ask questions on the mailing lists https://curl.haxx.se/mail/ SECURITY RELATED? Post it here: https://hackerone.com/curl There are collections of known issues to be aware of: https://curl.haxx.se/docs/knownbugs.html https://curl.haxx.se/docs/todo.html --> ### I did this I installed libcurl with the following commands: ```bash $ git clone https://github.com/curl/curl $ cd curl $ ./buildconf.bat $ mingw32-make mingw32 $ cp lib/libcurl.a /c/MinGW/lib $ cp -r include/curl/ /c/MinGW/include ``` I wrote this program to test it out: ```cpp // main.cpp #include <iostream> #include <curl/curl.h> int main(void) { CURL *curl; curl_global_init(CURL_GLOBAL_ALL); curl = curl_easy_init(); curl_easy_setopt(curl, CURLOPT_VERBOSE, 1L); curl_easy_setopt(curl, CURLOPT_URL, "https://api.ecs.rocks/v0/util/removeMessageHeader"); curl_easy_perform(curl); curl_free(curl); return 0; } ``` I built the C++ program with the following command: ```bash $ g++ main.cpp -o curltest -m32 -DCURL_STATICLIB -lcurl -lws2_32 -lwldap32 -lwsock32 -lwinmm ``` And I ran it: ```bash $ ./curltest.exe ``` ### I expected the following A response of some kind from https://api.ecs.rocks/v0/util/removeMessageHeader ### I got the following ``` * Protocol "https" not supported or disabled in libcurl * Closing connection -1 ``` ### curl/libcurl version ``` curl 7.71.1 (x86_64-w64-mingw32) libcurl/7.71.1 OpenSSL/1.1.1g (Schannel) zlib/1.2.11 brotli/1.0.7 libidn2/2.3.0 libssh2/1.9.0 nghttp2/1.41.0 Release-Date: 2020-07-01 Protocols: dict file ftp ftps gopher http https imap imaps ldap ldaps pop3 pop3s rtsp scp sftp smtp smtps telnet tftp Features: AsynchDNS brotli HTTP2 HTTPS-proxy IDN IPv6 Kerberos Largefile libz Metalink MultiSSL NTLM SPNEGO SSL SSPI TLS-SRP ``` That's the version of `curl` I have, but my version of `libcurl` should be the most recent version because I built from source. ### operating system MINGW64_NT-10.0-18363 ECSS-L01 3.1.6-340.x86_64 2020-07-09 14:33 UTC x86_64 Msys
non_test
windows with mingw protocol https not supported or disabled in libcurl only file bugs here ask questions on the mailing lists security related post it here there are collections of known issues to be aware of i did this i installed libcurl with the following commands bash git clone cd curl buildconf bat make cp lib libcurl a c mingw lib cp r include curl c mingw include i wrote this program to test it out cpp main cpp include include int main void curl curl curl global init curl global all curl curl easy init curl easy setopt curl curlopt verbose curl easy setopt curl curlopt url curl easy perform curl curl free curl return i built the c program with the following command bash g main cpp o curltest dcurl staticlib lcurl lwinmm and i ran it bash curltest exe i expected the following a response of some kind from i got the following protocol https not supported or disabled in libcurl closing connection curl libcurl version curl libcurl openssl schannel zlib brotli release date protocols dict file ftp ftps gopher http https imap imaps ldap ldaps rtsp scp sftp smtp smtps telnet tftp features asynchdns brotli https proxy idn kerberos largefile libz metalink multissl ntlm spnego ssl sspi tls srp that s the version of curl i have but my version of libcurl should be the most recent version because i built from source operating system nt ecss utc msys
0
815,456
30,555,538,388
IssuesEvent
2023-07-20 11:26:20
informalsystems/quint
https://api.github.com/repos/informalsystems/quint
closed
Error reporter fails during test
bug W1 product-priority-audits
When running `quint test` on [this spec](https://github.com/informalsystems/partnership-heliax/tree/manuel/quint-report-issue/2023/Q2/artifacts/PoS-quint), a test fails and the error reporter crashes. Quint version: 0.12.0 The quint test command: `quint test --main=namada --max-samples=10 --match=slashValidatorTest --verbosity=3 namada-redelegation.qnt` The error: ``` RangeError: Invalid count value at String.repeat (<anonymous>) at formatLine (/opt/homebrew/lib/node_modules/@informalsystems/quint/dist/src/errorReporter.js:64:19) at /opt/homebrew/lib/node_modules/@informalsystems/quint/dist/src/errorReporter.js:50:23 at Array.reduce (<anonymous>) at formatError (/opt/homebrew/lib/node_modules/@informalsystems/quint/dist/src/errorReporter.js:34:25) at /opt/homebrew/lib/node_modules/@informalsystems/quint/dist/src/cliCommands.js:272:65 at Array.forEach (<anonymous>) at runTests (/opt/homebrew/lib/node_modules/@informalsystems/quint/dist/src/cliCommands.js:271:25) at EitherConstructor.asyncChain (/opt/homebrew/lib/node_modules/@informalsystems/quint/node_modules/@sweet-monads/either/cjs/index.js:125:16) at /opt/homebrew/lib/node_modules/@informalsystems/quint/dist/src/cli.js:33:39 ``` Line 64 of errorReporter.js: `output += '^'.repeat(1 + endCol - startCol);`
1.0
Error reporter fails during test - When running `quint test` on [this spec](https://github.com/informalsystems/partnership-heliax/tree/manuel/quint-report-issue/2023/Q2/artifacts/PoS-quint), a test fails and the error reporter crashes. Quint version: 0.12.0 The quint test command: `quint test --main=namada --max-samples=10 --match=slashValidatorTest --verbosity=3 namada-redelegation.qnt` The error: ``` RangeError: Invalid count value at String.repeat (<anonymous>) at formatLine (/opt/homebrew/lib/node_modules/@informalsystems/quint/dist/src/errorReporter.js:64:19) at /opt/homebrew/lib/node_modules/@informalsystems/quint/dist/src/errorReporter.js:50:23 at Array.reduce (<anonymous>) at formatError (/opt/homebrew/lib/node_modules/@informalsystems/quint/dist/src/errorReporter.js:34:25) at /opt/homebrew/lib/node_modules/@informalsystems/quint/dist/src/cliCommands.js:272:65 at Array.forEach (<anonymous>) at runTests (/opt/homebrew/lib/node_modules/@informalsystems/quint/dist/src/cliCommands.js:271:25) at EitherConstructor.asyncChain (/opt/homebrew/lib/node_modules/@informalsystems/quint/node_modules/@sweet-monads/either/cjs/index.js:125:16) at /opt/homebrew/lib/node_modules/@informalsystems/quint/dist/src/cli.js:33:39 ``` Line 64 of errorReporter.js: `output += '^'.repeat(1 + endCol - startCol);`
non_test
error reporter fails during test when running quint test on a test fails and the error reporter crashes quint version the quint test command quint test main namada max samples match slashvalidatortest verbosity namada redelegation qnt the error rangeerror invalid count value at string repeat at formatline opt homebrew lib node modules informalsystems quint dist src errorreporter js at opt homebrew lib node modules informalsystems quint dist src errorreporter js at array reduce at formaterror opt homebrew lib node modules informalsystems quint dist src errorreporter js at opt homebrew lib node modules informalsystems quint dist src clicommands js at array foreach at runtests opt homebrew lib node modules informalsystems quint dist src clicommands js at eitherconstructor asyncchain opt homebrew lib node modules informalsystems quint node modules sweet monads either cjs index js at opt homebrew lib node modules informalsystems quint dist src cli js line of errorreporter js output repeat endcol startcol
0
735,340
25,390,281,525
IssuesEvent
2022-11-22 03:00:19
spidernet-io/spiderpool
https://api.github.com/repos/spidernet-io/spiderpool
closed
improve statefulset ipam cmdDel
priority/important-soon kind/bug
**Describe the bug** In the dual stack situation, once the IPv4 IP address was allocated but failed to allocate IPv6 address and the rollback function failed. In the next IPAM cmdAdd, the statefulset will retrieve from SpiderEndpoint but just get one IPv4 address. And we can't process it next.
1.0
improve statefulset ipam cmdDel - **Describe the bug** In the dual stack situation, once the IPv4 IP address was allocated but failed to allocate IPv6 address and the rollback function failed. In the next IPAM cmdAdd, the statefulset will retrieve from SpiderEndpoint but just get one IPv4 address. And we can't process it next.
non_test
improve statefulset ipam cmddel describe the bug in the dual stack situation once the ip address was allocated but failed to allocate address and the rollback function failed in the next ipam cmdadd the statefulset will retrieve from spiderendpoint but just get one address and we can t process it next
0
69,219
14,978,974,132
IssuesEvent
2021-01-28 11:34:52
AlexRogalskiy/java4you
https://api.github.com/repos/AlexRogalskiy/java4you
opened
CVE-2020-11996 (High) detected in tomcat-embed-core-9.0.17.jar
security vulnerability
## CVE-2020-11996 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-core-9.0.17.jar</b></p></summary> <p>Core Tomcat implementation</p> <p>Path to dependency file: java4you/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/tomcat/embed/tomcat-embed-core/9.0.17/tomcat-embed-core-9.0.17.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.1.4.RELEASE.jar (Root Library) - spring-boot-starter-tomcat-2.1.4.RELEASE.jar - :x: **tomcat-embed-core-9.0.17.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/java4you/commit/5ccabe003dfd14ccbb13bba969fdcdd318f2be81">5ccabe003dfd14ccbb13bba969fdcdd318f2be81</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A specially crafted sequence of HTTP/2 requests sent to Apache Tomcat 10.0.0-M1 to 10.0.0-M5, 9.0.0.M1 to 9.0.35 and 8.5.0 to 8.5.55 could trigger high CPU usage for several seconds. If a sufficient number of such requests were made on concurrent HTTP/2 connections, the server could become unresponsive. <p>Publish Date: 2020-06-26 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11996>CVE-2020-11996</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://lists.apache.org/thread.html/r5541ef6b6b68b49f76fc4c45695940116da2bcbe0312ef204a00a2e0%40%3Cannounce.tomcat.apache.org%3E,http://tomcat.apache.org/security-10.html">https://lists.apache.org/thread.html/r5541ef6b6b68b49f76fc4c45695940116da2bcbe0312ef204a00a2e0%40%3Cannounce.tomcat.apache.org%3E,http://tomcat.apache.org/security-10.html</a></p> <p>Release Date: 2020-06-26</p> <p>Fix Resolution: org.apache.tomcat:tomcat-coyote:10.0.0-M6,9.0.36,8.5.56,org.apache.tomcat.embed:org.apache.tomcat.embed:10.0.0-M6,9.0.36,8.5.56</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-11996 (High) detected in tomcat-embed-core-9.0.17.jar - ## CVE-2020-11996 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-core-9.0.17.jar</b></p></summary> <p>Core Tomcat implementation</p> <p>Path to dependency file: java4you/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/tomcat/embed/tomcat-embed-core/9.0.17/tomcat-embed-core-9.0.17.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.1.4.RELEASE.jar (Root Library) - spring-boot-starter-tomcat-2.1.4.RELEASE.jar - :x: **tomcat-embed-core-9.0.17.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/java4you/commit/5ccabe003dfd14ccbb13bba969fdcdd318f2be81">5ccabe003dfd14ccbb13bba969fdcdd318f2be81</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A specially crafted sequence of HTTP/2 requests sent to Apache Tomcat 10.0.0-M1 to 10.0.0-M5, 9.0.0.M1 to 9.0.35 and 8.5.0 to 8.5.55 could trigger high CPU usage for several seconds. If a sufficient number of such requests were made on concurrent HTTP/2 connections, the server could become unresponsive. <p>Publish Date: 2020-06-26 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11996>CVE-2020-11996</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://lists.apache.org/thread.html/r5541ef6b6b68b49f76fc4c45695940116da2bcbe0312ef204a00a2e0%40%3Cannounce.tomcat.apache.org%3E,http://tomcat.apache.org/security-10.html">https://lists.apache.org/thread.html/r5541ef6b6b68b49f76fc4c45695940116da2bcbe0312ef204a00a2e0%40%3Cannounce.tomcat.apache.org%3E,http://tomcat.apache.org/security-10.html</a></p> <p>Release Date: 2020-06-26</p> <p>Fix Resolution: org.apache.tomcat:tomcat-coyote:10.0.0-M6,9.0.36,8.5.56,org.apache.tomcat.embed:org.apache.tomcat.embed:10.0.0-M6,9.0.36,8.5.56</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in tomcat embed core jar cve high severity vulnerability vulnerable library tomcat embed core jar core tomcat implementation path to dependency file pom xml path to vulnerable library home wss scanner repository org apache tomcat embed tomcat embed core tomcat embed core jar dependency hierarchy spring boot starter web release jar root library spring boot starter tomcat release jar x tomcat embed core jar vulnerable library found in head commit a href found in base branch master vulnerability details a specially crafted sequence of http requests sent to apache tomcat to to and to could trigger high cpu usage for several seconds if a sufficient number of such requests were made on concurrent http connections the server could become unresponsive publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache tomcat tomcat coyote org apache tomcat embed org apache tomcat embed step up your open source security game with whitesource
0