Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
449,014
31,822,420,068
IssuesEvent
2023-09-14 04:13:06
CABLE-LSM/benchcab
https://api.github.com/repos/CABLE-LSM/benchcab
closed
Documentation errors
documentation
Some documentation errors that need to be fixed: - [ ] In user_guide.md - 'Directory structure and files', `tasks` should be `runs/fluxsite/tasks`. - [ ] In config_options.md - "NRI Land testing" should be "benchcab-evaluation" - [ ] In internal.py - "NRI Land testing" should be "benchcab-evaluation"
1.0
Documentation errors - Some documentation errors that need to be fixed: - [ ] In user_guide.md - 'Directory structure and files', `tasks` should be `runs/fluxsite/tasks`. - [ ] In config_options.md - "NRI Land testing" should be "benchcab-evaluation" - [ ] In internal.py - "NRI Land testing" should be "benchcab-evaluation"
non_process
documentation errors some documentation errors that need to be fixed in user guide md directory structure and files tasks should be runs fluxsite tasks in config options md nri land testing should be benchcab evaluation in internal py nri land testing should be benchcab evaluation
0
21,566
29,923,515,295
IssuesEvent
2023-06-22 02:00:08
lizhihao6/get-daily-arxiv-noti
https://api.github.com/repos/lizhihao6/get-daily-arxiv-noti
opened
New submissions for Thu, 22 Jun 23
event camera white balance isp compression image signal processing image signal process raw raw image events camera color contrast events AWB
## Keyword: events ### Self-Distilled Masked Auto-Encoders are Efficient Video Anomaly Detectors - **Authors:** Nicolae-Catalin Ristea, Florinel-Alin Croitoru, Radu Tudor Ionescu, Marius Popescu, Fahad Shahbaz Khan, Mubarak Shah - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG) - **Arxiv link:** https://arxiv.org/abs/2306.12041 - **Pdf link:** https://arxiv.org/pdf/2306.12041 - **Abstract** We propose an efficient abnormal event detection model based on a lightweight masked auto-encoder (AE) applied at the video frame level. The novelty of the proposed model is threefold. First, we introduce an approach to weight tokens based on motion gradients, thus avoiding learning to reconstruct the static background scene. Second, we integrate a teacher decoder and a student decoder into our architecture, leveraging the discrepancy between the outputs given by the two decoders to improve anomaly detection. Third, we generate synthetic abnormal events to augment the training videos, and task the masked AE model to jointly reconstruct the original frames (without anomalies) and the corresponding pixel-level anomaly maps. Our design leads to an efficient and effective model, as demonstrated by the extensive experiments carried out on three benchmarks: Avenue, ShanghaiTech and UCSD Ped2. The empirical results show that our model achieves an excellent trade-off between speed and accuracy, obtaining competitive AUC scores, while processing 1670 FPS. Hence, our model is between 8 and 70 times faster than competing methods. We also conduct an ablation study to justify our design. ### Wildfire Detection Via Transfer Learning: A Survey - **Authors:** Ziliang Hong, Emadeldeen Hamdan, Yifei Zhao, Tianxiao Ye, Hongyi Pan, A. Enis Cetin - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2306.12276 - **Pdf link:** https://arxiv.org/pdf/2306.12276 - **Abstract** This paper surveys different publicly available neural network models used for detecting wildfires using regular visible-range cameras which are placed on hilltops or forest lookout towers. The neural network models are pre-trained on ImageNet-1K and fine-tuned on a custom wildfire dataset. The performance of these models is evaluated on a diverse set of wildfire images, and the survey provides useful information for those interested in using transfer learning for wildfire detection. Swin Transformer-tiny has the highest AUC value but ConvNext-tiny detects all the wildfire events and has the lowest false alarm rate in our dataset. ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB There is no result ## Keyword: ISP ### NILUT: Conditional Neural Implicit 3D Lookup Tables for Image Enhancement - **Authors:** Marcos V. Conde, Javier Vazquez-Corral, Michael S. Brown, Radu Timofte - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2306.11920 - **Pdf link:** https://arxiv.org/pdf/2306.11920 - **Abstract** 3D lookup tables (3D LUTs) are a key component for image enhancement. Modern image signal processors (ISPs) have dedicated support for these as part of the camera rendering pipeline. Cameras typically provide multiple options for picture styles, where each style is usually obtained by applying a unique handcrafted 3D LUT. Current approaches for learning and applying 3D LUTs are notably fast, yet not so memory-efficient, as storing multiple 3D LUTs is required. For this reason and other implementation limitations, their use on mobile devices is less popular. In this work, we propose a Neural Implicit LUT (NILUT), an implicitly defined continuous 3D color transformation parameterized by a neural network. We show that NILUTs are capable of accurately emulating real 3D LUTs. Moreover, a NILUT can be extended to incorporate multiple styles into a single network with the ability to blend styles implicitly. Our novel approach is memory-efficient, controllable and can complement previous methods, including learned ISPs. Code, models and dataset available at: https://github.com/mv-lab/nilut ### HSR-Diff:Hyperspectral Image Super-Resolution via Conditional Diffusion Models - **Authors:** Chanyue Wu, Dong Wang, Hanyu Mao, Ying Li - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2306.12085 - **Pdf link:** https://arxiv.org/pdf/2306.12085 - **Abstract** Despite the proven significance of hyperspectral images (HSIs) in performing various computer vision tasks, its potential is adversely affected by the low-resolution (LR) property in the spatial domain, resulting from multiple physical factors. Inspired by recent advancements in deep generative models, we propose an HSI Super-resolution (SR) approach with Conditional Diffusion Models (HSR-Diff) that merges a high-resolution (HR) multispectral image (MSI) with the corresponding LR-HSI. HSR-Diff generates an HR-HSI via repeated refinement, in which the HR-HSI is initialized with pure Gaussian noise and iteratively refined. At each iteration, the noise is removed with a Conditional Denoising Transformer (CDF ormer) that is trained on denoising at different noise levels, conditioned on the hierarchical feature maps of HR-MSI and LR-HSI. In addition, a progressive learning strategy is employed to exploit the global information of full-resolution images. Systematic experiments have been conducted on four public datasets, demonstrating that HSR-Diff outperforms state-of-the-art methods. ## Keyword: image signal processing There is no result ## Keyword: image signal process ### NILUT: Conditional Neural Implicit 3D Lookup Tables for Image Enhancement - **Authors:** Marcos V. Conde, Javier Vazquez-Corral, Michael S. Brown, Radu Timofte - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2306.11920 - **Pdf link:** https://arxiv.org/pdf/2306.11920 - **Abstract** 3D lookup tables (3D LUTs) are a key component for image enhancement. Modern image signal processors (ISPs) have dedicated support for these as part of the camera rendering pipeline. Cameras typically provide multiple options for picture styles, where each style is usually obtained by applying a unique handcrafted 3D LUT. Current approaches for learning and applying 3D LUTs are notably fast, yet not so memory-efficient, as storing multiple 3D LUTs is required. For this reason and other implementation limitations, their use on mobile devices is less popular. In this work, we propose a Neural Implicit LUT (NILUT), an implicitly defined continuous 3D color transformation parameterized by a neural network. We show that NILUTs are capable of accurately emulating real 3D LUTs. Moreover, a NILUT can be extended to incorporate multiple styles into a single network with the ability to blend styles implicitly. Our novel approach is memory-efficient, controllable and can complement previous methods, including learned ISPs. Code, models and dataset available at: https://github.com/mv-lab/nilut ## Keyword: compression ### Spiking Neural Network for Ultra-low-latency and High-accurate Object Detection - **Authors:** Jinye Qu, Zeyu Gao, Tielin Zhang, Yanfeng Lu, Huajin Tang, Hong Qiao - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Neural and Evolutionary Computing (cs.NE) - **Arxiv link:** https://arxiv.org/abs/2306.12010 - **Pdf link:** https://arxiv.org/pdf/2306.12010 - **Abstract** Spiking Neural Networks (SNNs) have garnered widespread interest for their energy efficiency and brain-inspired event-driven properties. While recent methods like Spiking-YOLO have expanded the SNNs to more challenging object detection tasks, they often suffer from high latency and low detection accuracy, making them difficult to deploy on latency sensitive mobile platforms. Furthermore, the conversion method from Artificial Neural Networks (ANNs) to SNNs is hard to maintain the complete structure of the ANNs, resulting in poor feature representation and high conversion errors. To address these challenges, we propose two methods: timesteps compression and spike-time-dependent integrated (STDI) coding. The former reduces the timesteps required in ANN-SNN conversion by compressing information, while the latter sets a time-varying threshold to expand the information holding capacity. We also present a SNN-based ultra-low latency and high accurate object detection model (SUHD) that achieves state-of-the-art performance on nontrivial datasets like PASCAL VOC and MS COCO, with about remarkable 750x fewer timesteps and 30% mean average precision (mAP) improvement, compared to the Spiking-YOLO on MS COCO datasets. To the best of our knowledge, SUHD is the deepest spike-based object detection model to date that achieves ultra low timesteps to complete the lossless conversion. ### Beyond Learned Metadata-based Raw Image Reconstruction - **Authors:** Yufei Wang, Yi Yu, Wenhan Yang, Lanqing Guo, Lap-Pui Chau, Alex C. Kot, Bihan Wen - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2306.12058 - **Pdf link:** https://arxiv.org/pdf/2306.12058 - **Abstract** While raw images have distinct advantages over sRGB images, e.g., linearity and fine-grained quantization levels, they are not widely adopted by general users due to their substantial storage requirements. Very recent studies propose to compress raw images by designing sampling masks within the pixel space of the raw image. However, these approaches often leave space for pursuing more effective image representations and compact metadata. In this work, we propose a novel framework that learns a compact representation in the latent space, serving as metadata, in an end-to-end manner. Compared with lossy image compression, we analyze the intrinsic difference of the raw image reconstruction task caused by rich information from the sRGB image. Based on the analysis, a novel backbone design with asymmetric and hybrid spatial feature resolutions is proposed, which significantly improves the rate-distortion performance. Besides, we propose a novel design of the context model, which can better predict the order masks of encoding/decoding based on both the sRGB image and the masks of already processed features. Benefited from the better modeling of the correlation between order masks, the already processed information can be better utilized. Moreover, a novel sRGB-guided adaptive quantization precision strategy, which dynamically assigns varying levels of quantization precision to different regions, further enhances the representation ability of the model. Finally, based on the iterative properties of the proposed context model, we propose a novel strategy to achieve variable bit rates using a single model. This strategy allows for the continuous convergence of a wide range of bit rates. Extensive experimental results demonstrate that the proposed method can achieve better reconstruction quality with a smaller metadata size. ## Keyword: RAW ### Self-supervised Multi-task Learning Framework for Safety and Health-Oriented Connected Driving Environment Perception using Onboard Camera - **Authors:** Shaocheng Jia, Wei Yao - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2306.11822 - **Pdf link:** https://arxiv.org/pdf/2306.11822 - **Abstract** Cutting-edge connected vehicle (CV) technologies have drawn much attention in recent years. The real-time traffic data captured by a CV can be shared with other CVs and data centers so as to open new possibilities for solving diverse transportation problems. However, imagery captured by onboard cameras in a connected environment, are not sufficiently investigated, especially for safety and health-oriented visual perception. In this paper, a bidirectional process of image synthesis and decomposition (BPISD) approach is proposed, and thus a novel self-supervised multi-task learning framework, to simultaneously estimate depth map, atmospheric visibility, airlight, and PM2.5 mass concentration, in which depth map and visibility are considered highly associated with traffic safety, while airlight and PM2.5 mass concentration are directly correlated with human health. Both the training and testing phases of the proposed system solely require a single image as input. Due to the innovative training pipeline, the depth estimation network can manage various levels of visibility conditions and overcome inherent problems in current image-synthesis-based depth estimation, thereby generating high-quality depth maps even in low-visibility situations and further benefiting accurate estimations of visibility, airlight, and PM2.5 mass concentration. Extensive experiments on the synthesized data from the KITTI and real-world data collected in Beijing demonstrate that the proposed method can (1) achieve performance competitive in depth estimation as compared with state-of-the-art methods when taking clear images as input; (2) predict vivid depth map for images contaminated by various levels of haze; and (3) accurately estimate visibility, airlight, and PM2.5 mass concentrations. Beneficial applications can be developed based on the presented work to improve traffic safety, air quality, and public health. ### Beyond Learned Metadata-based Raw Image Reconstruction - **Authors:** Yufei Wang, Yi Yu, Wenhan Yang, Lanqing Guo, Lap-Pui Chau, Alex C. Kot, Bihan Wen - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2306.12058 - **Pdf link:** https://arxiv.org/pdf/2306.12058 - **Abstract** While raw images have distinct advantages over sRGB images, e.g., linearity and fine-grained quantization levels, they are not widely adopted by general users due to their substantial storage requirements. Very recent studies propose to compress raw images by designing sampling masks within the pixel space of the raw image. However, these approaches often leave space for pursuing more effective image representations and compact metadata. In this work, we propose a novel framework that learns a compact representation in the latent space, serving as metadata, in an end-to-end manner. Compared with lossy image compression, we analyze the intrinsic difference of the raw image reconstruction task caused by rich information from the sRGB image. Based on the analysis, a novel backbone design with asymmetric and hybrid spatial feature resolutions is proposed, which significantly improves the rate-distortion performance. Besides, we propose a novel design of the context model, which can better predict the order masks of encoding/decoding based on both the sRGB image and the masks of already processed features. Benefited from the better modeling of the correlation between order masks, the already processed information can be better utilized. Moreover, a novel sRGB-guided adaptive quantization precision strategy, which dynamically assigns varying levels of quantization precision to different regions, further enhances the representation ability of the model. Finally, based on the iterative properties of the proposed context model, we propose a novel strategy to achieve variable bit rates using a single model. This strategy allows for the continuous convergence of a wide range of bit rates. Extensive experimental results demonstrate that the proposed method can achieve better reconstruction quality with a smaller metadata size. ## Keyword: raw image ### Beyond Learned Metadata-based Raw Image Reconstruction - **Authors:** Yufei Wang, Yi Yu, Wenhan Yang, Lanqing Guo, Lap-Pui Chau, Alex C. Kot, Bihan Wen - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2306.12058 - **Pdf link:** https://arxiv.org/pdf/2306.12058 - **Abstract** While raw images have distinct advantages over sRGB images, e.g., linearity and fine-grained quantization levels, they are not widely adopted by general users due to their substantial storage requirements. Very recent studies propose to compress raw images by designing sampling masks within the pixel space of the raw image. However, these approaches often leave space for pursuing more effective image representations and compact metadata. In this work, we propose a novel framework that learns a compact representation in the latent space, serving as metadata, in an end-to-end manner. Compared with lossy image compression, we analyze the intrinsic difference of the raw image reconstruction task caused by rich information from the sRGB image. Based on the analysis, a novel backbone design with asymmetric and hybrid spatial feature resolutions is proposed, which significantly improves the rate-distortion performance. Besides, we propose a novel design of the context model, which can better predict the order masks of encoding/decoding based on both the sRGB image and the masks of already processed features. Benefited from the better modeling of the correlation between order masks, the already processed information can be better utilized. Moreover, a novel sRGB-guided adaptive quantization precision strategy, which dynamically assigns varying levels of quantization precision to different regions, further enhances the representation ability of the model. Finally, based on the iterative properties of the proposed context model, we propose a novel strategy to achieve variable bit rates using a single model. This strategy allows for the continuous convergence of a wide range of bit rates. Extensive experimental results demonstrate that the proposed method can achieve better reconstruction quality with a smaller metadata size.
2.0
New submissions for Thu, 22 Jun 23 - ## Keyword: events ### Self-Distilled Masked Auto-Encoders are Efficient Video Anomaly Detectors - **Authors:** Nicolae-Catalin Ristea, Florinel-Alin Croitoru, Radu Tudor Ionescu, Marius Popescu, Fahad Shahbaz Khan, Mubarak Shah - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG) - **Arxiv link:** https://arxiv.org/abs/2306.12041 - **Pdf link:** https://arxiv.org/pdf/2306.12041 - **Abstract** We propose an efficient abnormal event detection model based on a lightweight masked auto-encoder (AE) applied at the video frame level. The novelty of the proposed model is threefold. First, we introduce an approach to weight tokens based on motion gradients, thus avoiding learning to reconstruct the static background scene. Second, we integrate a teacher decoder and a student decoder into our architecture, leveraging the discrepancy between the outputs given by the two decoders to improve anomaly detection. Third, we generate synthetic abnormal events to augment the training videos, and task the masked AE model to jointly reconstruct the original frames (without anomalies) and the corresponding pixel-level anomaly maps. Our design leads to an efficient and effective model, as demonstrated by the extensive experiments carried out on three benchmarks: Avenue, ShanghaiTech and UCSD Ped2. The empirical results show that our model achieves an excellent trade-off between speed and accuracy, obtaining competitive AUC scores, while processing 1670 FPS. Hence, our model is between 8 and 70 times faster than competing methods. We also conduct an ablation study to justify our design. ### Wildfire Detection Via Transfer Learning: A Survey - **Authors:** Ziliang Hong, Emadeldeen Hamdan, Yifei Zhao, Tianxiao Ye, Hongyi Pan, A. Enis Cetin - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2306.12276 - **Pdf link:** https://arxiv.org/pdf/2306.12276 - **Abstract** This paper surveys different publicly available neural network models used for detecting wildfires using regular visible-range cameras which are placed on hilltops or forest lookout towers. The neural network models are pre-trained on ImageNet-1K and fine-tuned on a custom wildfire dataset. The performance of these models is evaluated on a diverse set of wildfire images, and the survey provides useful information for those interested in using transfer learning for wildfire detection. Swin Transformer-tiny has the highest AUC value but ConvNext-tiny detects all the wildfire events and has the lowest false alarm rate in our dataset. ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB There is no result ## Keyword: ISP ### NILUT: Conditional Neural Implicit 3D Lookup Tables for Image Enhancement - **Authors:** Marcos V. Conde, Javier Vazquez-Corral, Michael S. Brown, Radu Timofte - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2306.11920 - **Pdf link:** https://arxiv.org/pdf/2306.11920 - **Abstract** 3D lookup tables (3D LUTs) are a key component for image enhancement. Modern image signal processors (ISPs) have dedicated support for these as part of the camera rendering pipeline. Cameras typically provide multiple options for picture styles, where each style is usually obtained by applying a unique handcrafted 3D LUT. Current approaches for learning and applying 3D LUTs are notably fast, yet not so memory-efficient, as storing multiple 3D LUTs is required. For this reason and other implementation limitations, their use on mobile devices is less popular. In this work, we propose a Neural Implicit LUT (NILUT), an implicitly defined continuous 3D color transformation parameterized by a neural network. We show that NILUTs are capable of accurately emulating real 3D LUTs. Moreover, a NILUT can be extended to incorporate multiple styles into a single network with the ability to blend styles implicitly. Our novel approach is memory-efficient, controllable and can complement previous methods, including learned ISPs. Code, models and dataset available at: https://github.com/mv-lab/nilut ### HSR-Diff:Hyperspectral Image Super-Resolution via Conditional Diffusion Models - **Authors:** Chanyue Wu, Dong Wang, Hanyu Mao, Ying Li - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2306.12085 - **Pdf link:** https://arxiv.org/pdf/2306.12085 - **Abstract** Despite the proven significance of hyperspectral images (HSIs) in performing various computer vision tasks, its potential is adversely affected by the low-resolution (LR) property in the spatial domain, resulting from multiple physical factors. Inspired by recent advancements in deep generative models, we propose an HSI Super-resolution (SR) approach with Conditional Diffusion Models (HSR-Diff) that merges a high-resolution (HR) multispectral image (MSI) with the corresponding LR-HSI. HSR-Diff generates an HR-HSI via repeated refinement, in which the HR-HSI is initialized with pure Gaussian noise and iteratively refined. At each iteration, the noise is removed with a Conditional Denoising Transformer (CDF ormer) that is trained on denoising at different noise levels, conditioned on the hierarchical feature maps of HR-MSI and LR-HSI. In addition, a progressive learning strategy is employed to exploit the global information of full-resolution images. Systematic experiments have been conducted on four public datasets, demonstrating that HSR-Diff outperforms state-of-the-art methods. ## Keyword: image signal processing There is no result ## Keyword: image signal process ### NILUT: Conditional Neural Implicit 3D Lookup Tables for Image Enhancement - **Authors:** Marcos V. Conde, Javier Vazquez-Corral, Michael S. Brown, Radu Timofte - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2306.11920 - **Pdf link:** https://arxiv.org/pdf/2306.11920 - **Abstract** 3D lookup tables (3D LUTs) are a key component for image enhancement. Modern image signal processors (ISPs) have dedicated support for these as part of the camera rendering pipeline. Cameras typically provide multiple options for picture styles, where each style is usually obtained by applying a unique handcrafted 3D LUT. Current approaches for learning and applying 3D LUTs are notably fast, yet not so memory-efficient, as storing multiple 3D LUTs is required. For this reason and other implementation limitations, their use on mobile devices is less popular. In this work, we propose a Neural Implicit LUT (NILUT), an implicitly defined continuous 3D color transformation parameterized by a neural network. We show that NILUTs are capable of accurately emulating real 3D LUTs. Moreover, a NILUT can be extended to incorporate multiple styles into a single network with the ability to blend styles implicitly. Our novel approach is memory-efficient, controllable and can complement previous methods, including learned ISPs. Code, models and dataset available at: https://github.com/mv-lab/nilut ## Keyword: compression ### Spiking Neural Network for Ultra-low-latency and High-accurate Object Detection - **Authors:** Jinye Qu, Zeyu Gao, Tielin Zhang, Yanfeng Lu, Huajin Tang, Hong Qiao - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Neural and Evolutionary Computing (cs.NE) - **Arxiv link:** https://arxiv.org/abs/2306.12010 - **Pdf link:** https://arxiv.org/pdf/2306.12010 - **Abstract** Spiking Neural Networks (SNNs) have garnered widespread interest for their energy efficiency and brain-inspired event-driven properties. While recent methods like Spiking-YOLO have expanded the SNNs to more challenging object detection tasks, they often suffer from high latency and low detection accuracy, making them difficult to deploy on latency sensitive mobile platforms. Furthermore, the conversion method from Artificial Neural Networks (ANNs) to SNNs is hard to maintain the complete structure of the ANNs, resulting in poor feature representation and high conversion errors. To address these challenges, we propose two methods: timesteps compression and spike-time-dependent integrated (STDI) coding. The former reduces the timesteps required in ANN-SNN conversion by compressing information, while the latter sets a time-varying threshold to expand the information holding capacity. We also present a SNN-based ultra-low latency and high accurate object detection model (SUHD) that achieves state-of-the-art performance on nontrivial datasets like PASCAL VOC and MS COCO, with about remarkable 750x fewer timesteps and 30% mean average precision (mAP) improvement, compared to the Spiking-YOLO on MS COCO datasets. To the best of our knowledge, SUHD is the deepest spike-based object detection model to date that achieves ultra low timesteps to complete the lossless conversion. ### Beyond Learned Metadata-based Raw Image Reconstruction - **Authors:** Yufei Wang, Yi Yu, Wenhan Yang, Lanqing Guo, Lap-Pui Chau, Alex C. Kot, Bihan Wen - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2306.12058 - **Pdf link:** https://arxiv.org/pdf/2306.12058 - **Abstract** While raw images have distinct advantages over sRGB images, e.g., linearity and fine-grained quantization levels, they are not widely adopted by general users due to their substantial storage requirements. Very recent studies propose to compress raw images by designing sampling masks within the pixel space of the raw image. However, these approaches often leave space for pursuing more effective image representations and compact metadata. In this work, we propose a novel framework that learns a compact representation in the latent space, serving as metadata, in an end-to-end manner. Compared with lossy image compression, we analyze the intrinsic difference of the raw image reconstruction task caused by rich information from the sRGB image. Based on the analysis, a novel backbone design with asymmetric and hybrid spatial feature resolutions is proposed, which significantly improves the rate-distortion performance. Besides, we propose a novel design of the context model, which can better predict the order masks of encoding/decoding based on both the sRGB image and the masks of already processed features. Benefited from the better modeling of the correlation between order masks, the already processed information can be better utilized. Moreover, a novel sRGB-guided adaptive quantization precision strategy, which dynamically assigns varying levels of quantization precision to different regions, further enhances the representation ability of the model. Finally, based on the iterative properties of the proposed context model, we propose a novel strategy to achieve variable bit rates using a single model. This strategy allows for the continuous convergence of a wide range of bit rates. Extensive experimental results demonstrate that the proposed method can achieve better reconstruction quality with a smaller metadata size. ## Keyword: RAW ### Self-supervised Multi-task Learning Framework for Safety and Health-Oriented Connected Driving Environment Perception using Onboard Camera - **Authors:** Shaocheng Jia, Wei Yao - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2306.11822 - **Pdf link:** https://arxiv.org/pdf/2306.11822 - **Abstract** Cutting-edge connected vehicle (CV) technologies have drawn much attention in recent years. The real-time traffic data captured by a CV can be shared with other CVs and data centers so as to open new possibilities for solving diverse transportation problems. However, imagery captured by onboard cameras in a connected environment, are not sufficiently investigated, especially for safety and health-oriented visual perception. In this paper, a bidirectional process of image synthesis and decomposition (BPISD) approach is proposed, and thus a novel self-supervised multi-task learning framework, to simultaneously estimate depth map, atmospheric visibility, airlight, and PM2.5 mass concentration, in which depth map and visibility are considered highly associated with traffic safety, while airlight and PM2.5 mass concentration are directly correlated with human health. Both the training and testing phases of the proposed system solely require a single image as input. Due to the innovative training pipeline, the depth estimation network can manage various levels of visibility conditions and overcome inherent problems in current image-synthesis-based depth estimation, thereby generating high-quality depth maps even in low-visibility situations and further benefiting accurate estimations of visibility, airlight, and PM2.5 mass concentration. Extensive experiments on the synthesized data from the KITTI and real-world data collected in Beijing demonstrate that the proposed method can (1) achieve performance competitive in depth estimation as compared with state-of-the-art methods when taking clear images as input; (2) predict vivid depth map for images contaminated by various levels of haze; and (3) accurately estimate visibility, airlight, and PM2.5 mass concentrations. Beneficial applications can be developed based on the presented work to improve traffic safety, air quality, and public health. ### Beyond Learned Metadata-based Raw Image Reconstruction - **Authors:** Yufei Wang, Yi Yu, Wenhan Yang, Lanqing Guo, Lap-Pui Chau, Alex C. Kot, Bihan Wen - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2306.12058 - **Pdf link:** https://arxiv.org/pdf/2306.12058 - **Abstract** While raw images have distinct advantages over sRGB images, e.g., linearity and fine-grained quantization levels, they are not widely adopted by general users due to their substantial storage requirements. Very recent studies propose to compress raw images by designing sampling masks within the pixel space of the raw image. However, these approaches often leave space for pursuing more effective image representations and compact metadata. In this work, we propose a novel framework that learns a compact representation in the latent space, serving as metadata, in an end-to-end manner. Compared with lossy image compression, we analyze the intrinsic difference of the raw image reconstruction task caused by rich information from the sRGB image. Based on the analysis, a novel backbone design with asymmetric and hybrid spatial feature resolutions is proposed, which significantly improves the rate-distortion performance. Besides, we propose a novel design of the context model, which can better predict the order masks of encoding/decoding based on both the sRGB image and the masks of already processed features. Benefited from the better modeling of the correlation between order masks, the already processed information can be better utilized. Moreover, a novel sRGB-guided adaptive quantization precision strategy, which dynamically assigns varying levels of quantization precision to different regions, further enhances the representation ability of the model. Finally, based on the iterative properties of the proposed context model, we propose a novel strategy to achieve variable bit rates using a single model. This strategy allows for the continuous convergence of a wide range of bit rates. Extensive experimental results demonstrate that the proposed method can achieve better reconstruction quality with a smaller metadata size. ## Keyword: raw image ### Beyond Learned Metadata-based Raw Image Reconstruction - **Authors:** Yufei Wang, Yi Yu, Wenhan Yang, Lanqing Guo, Lap-Pui Chau, Alex C. Kot, Bihan Wen - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2306.12058 - **Pdf link:** https://arxiv.org/pdf/2306.12058 - **Abstract** While raw images have distinct advantages over sRGB images, e.g., linearity and fine-grained quantization levels, they are not widely adopted by general users due to their substantial storage requirements. Very recent studies propose to compress raw images by designing sampling masks within the pixel space of the raw image. However, these approaches often leave space for pursuing more effective image representations and compact metadata. In this work, we propose a novel framework that learns a compact representation in the latent space, serving as metadata, in an end-to-end manner. Compared with lossy image compression, we analyze the intrinsic difference of the raw image reconstruction task caused by rich information from the sRGB image. Based on the analysis, a novel backbone design with asymmetric and hybrid spatial feature resolutions is proposed, which significantly improves the rate-distortion performance. Besides, we propose a novel design of the context model, which can better predict the order masks of encoding/decoding based on both the sRGB image and the masks of already processed features. Benefited from the better modeling of the correlation between order masks, the already processed information can be better utilized. Moreover, a novel sRGB-guided adaptive quantization precision strategy, which dynamically assigns varying levels of quantization precision to different regions, further enhances the representation ability of the model. Finally, based on the iterative properties of the proposed context model, we propose a novel strategy to achieve variable bit rates using a single model. This strategy allows for the continuous convergence of a wide range of bit rates. Extensive experimental results demonstrate that the proposed method can achieve better reconstruction quality with a smaller metadata size.
process
new submissions for thu jun keyword events self distilled masked auto encoders are efficient video anomaly detectors authors nicolae catalin ristea florinel alin croitoru radu tudor ionescu marius popescu fahad shahbaz khan mubarak shah subjects computer vision and pattern recognition cs cv machine learning cs lg arxiv link pdf link abstract we propose an efficient abnormal event detection model based on a lightweight masked auto encoder ae applied at the video frame level the novelty of the proposed model is threefold first we introduce an approach to weight tokens based on motion gradients thus avoiding learning to reconstruct the static background scene second we integrate a teacher decoder and a student decoder into our architecture leveraging the discrepancy between the outputs given by the two decoders to improve anomaly detection third we generate synthetic abnormal events to augment the training videos and task the masked ae model to jointly reconstruct the original frames without anomalies and the corresponding pixel level anomaly maps our design leads to an efficient and effective model as demonstrated by the extensive experiments carried out on three benchmarks avenue shanghaitech and ucsd the empirical results show that our model achieves an excellent trade off between speed and accuracy obtaining competitive auc scores while processing fps hence our model is between and times faster than competing methods we also conduct an ablation study to justify our design wildfire detection via transfer learning a survey authors ziliang hong emadeldeen hamdan yifei zhao tianxiao ye hongyi pan a enis cetin subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract this paper surveys different publicly available neural network models used for detecting wildfires using regular visible range cameras which are placed on hilltops or forest lookout towers the neural network models are pre trained on imagenet and fine tuned on a custom wildfire dataset the performance of these models is evaluated on a diverse set of wildfire images and the survey provides useful information for those interested in using transfer learning for wildfire detection swin transformer tiny has the highest auc value but convnext tiny detects all the wildfire events and has the lowest false alarm rate in our dataset keyword event camera there is no result keyword events camera there is no result keyword white balance there is no result keyword color contrast there is no result keyword awb there is no result keyword isp nilut conditional neural implicit lookup tables for image enhancement authors marcos v conde javier vazquez corral michael s brown radu timofte subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract lookup tables luts are a key component for image enhancement modern image signal processors isps have dedicated support for these as part of the camera rendering pipeline cameras typically provide multiple options for picture styles where each style is usually obtained by applying a unique handcrafted lut current approaches for learning and applying luts are notably fast yet not so memory efficient as storing multiple luts is required for this reason and other implementation limitations their use on mobile devices is less popular in this work we propose a neural implicit lut nilut an implicitly defined continuous color transformation parameterized by a neural network we show that niluts are capable of accurately emulating real luts moreover a nilut can be extended to incorporate multiple styles into a single network with the ability to blend styles implicitly our novel approach is memory efficient controllable and can complement previous methods including learned isps code models and dataset available at hsr diff hyperspectral image super resolution via conditional diffusion models authors chanyue wu dong wang hanyu mao ying li subjects computer vision and pattern recognition cs cv image and video processing eess iv arxiv link pdf link abstract despite the proven significance of hyperspectral images hsis in performing various computer vision tasks its potential is adversely affected by the low resolution lr property in the spatial domain resulting from multiple physical factors inspired by recent advancements in deep generative models we propose an hsi super resolution sr approach with conditional diffusion models hsr diff that merges a high resolution hr multispectral image msi with the corresponding lr hsi hsr diff generates an hr hsi via repeated refinement in which the hr hsi is initialized with pure gaussian noise and iteratively refined at each iteration the noise is removed with a conditional denoising transformer cdf ormer that is trained on denoising at different noise levels conditioned on the hierarchical feature maps of hr msi and lr hsi in addition a progressive learning strategy is employed to exploit the global information of full resolution images systematic experiments have been conducted on four public datasets demonstrating that hsr diff outperforms state of the art methods keyword image signal processing there is no result keyword image signal process nilut conditional neural implicit lookup tables for image enhancement authors marcos v conde javier vazquez corral michael s brown radu timofte subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract lookup tables luts are a key component for image enhancement modern image signal processors isps have dedicated support for these as part of the camera rendering pipeline cameras typically provide multiple options for picture styles where each style is usually obtained by applying a unique handcrafted lut current approaches for learning and applying luts are notably fast yet not so memory efficient as storing multiple luts is required for this reason and other implementation limitations their use on mobile devices is less popular in this work we propose a neural implicit lut nilut an implicitly defined continuous color transformation parameterized by a neural network we show that niluts are capable of accurately emulating real luts moreover a nilut can be extended to incorporate multiple styles into a single network with the ability to blend styles implicitly our novel approach is memory efficient controllable and can complement previous methods including learned isps code models and dataset available at keyword compression spiking neural network for ultra low latency and high accurate object detection authors jinye qu zeyu gao tielin zhang yanfeng lu huajin tang hong qiao subjects computer vision and pattern recognition cs cv neural and evolutionary computing cs ne arxiv link pdf link abstract spiking neural networks snns have garnered widespread interest for their energy efficiency and brain inspired event driven properties while recent methods like spiking yolo have expanded the snns to more challenging object detection tasks they often suffer from high latency and low detection accuracy making them difficult to deploy on latency sensitive mobile platforms furthermore the conversion method from artificial neural networks anns to snns is hard to maintain the complete structure of the anns resulting in poor feature representation and high conversion errors to address these challenges we propose two methods timesteps compression and spike time dependent integrated stdi coding the former reduces the timesteps required in ann snn conversion by compressing information while the latter sets a time varying threshold to expand the information holding capacity we also present a snn based ultra low latency and high accurate object detection model suhd that achieves state of the art performance on nontrivial datasets like pascal voc and ms coco with about remarkable fewer timesteps and mean average precision map improvement compared to the spiking yolo on ms coco datasets to the best of our knowledge suhd is the deepest spike based object detection model to date that achieves ultra low timesteps to complete the lossless conversion beyond learned metadata based raw image reconstruction authors yufei wang yi yu wenhan yang lanqing guo lap pui chau alex c kot bihan wen subjects computer vision and pattern recognition cs cv image and video processing eess iv arxiv link pdf link abstract while raw images have distinct advantages over srgb images e g linearity and fine grained quantization levels they are not widely adopted by general users due to their substantial storage requirements very recent studies propose to compress raw images by designing sampling masks within the pixel space of the raw image however these approaches often leave space for pursuing more effective image representations and compact metadata in this work we propose a novel framework that learns a compact representation in the latent space serving as metadata in an end to end manner compared with lossy image compression we analyze the intrinsic difference of the raw image reconstruction task caused by rich information from the srgb image based on the analysis a novel backbone design with asymmetric and hybrid spatial feature resolutions is proposed which significantly improves the rate distortion performance besides we propose a novel design of the context model which can better predict the order masks of encoding decoding based on both the srgb image and the masks of already processed features benefited from the better modeling of the correlation between order masks the already processed information can be better utilized moreover a novel srgb guided adaptive quantization precision strategy which dynamically assigns varying levels of quantization precision to different regions further enhances the representation ability of the model finally based on the iterative properties of the proposed context model we propose a novel strategy to achieve variable bit rates using a single model this strategy allows for the continuous convergence of a wide range of bit rates extensive experimental results demonstrate that the proposed method can achieve better reconstruction quality with a smaller metadata size keyword raw self supervised multi task learning framework for safety and health oriented connected driving environment perception using onboard camera authors shaocheng jia wei yao subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract cutting edge connected vehicle cv technologies have drawn much attention in recent years the real time traffic data captured by a cv can be shared with other cvs and data centers so as to open new possibilities for solving diverse transportation problems however imagery captured by onboard cameras in a connected environment are not sufficiently investigated especially for safety and health oriented visual perception in this paper a bidirectional process of image synthesis and decomposition bpisd approach is proposed and thus a novel self supervised multi task learning framework to simultaneously estimate depth map atmospheric visibility airlight and mass concentration in which depth map and visibility are considered highly associated with traffic safety while airlight and mass concentration are directly correlated with human health both the training and testing phases of the proposed system solely require a single image as input due to the innovative training pipeline the depth estimation network can manage various levels of visibility conditions and overcome inherent problems in current image synthesis based depth estimation thereby generating high quality depth maps even in low visibility situations and further benefiting accurate estimations of visibility airlight and mass concentration extensive experiments on the synthesized data from the kitti and real world data collected in beijing demonstrate that the proposed method can achieve performance competitive in depth estimation as compared with state of the art methods when taking clear images as input predict vivid depth map for images contaminated by various levels of haze and accurately estimate visibility airlight and mass concentrations beneficial applications can be developed based on the presented work to improve traffic safety air quality and public health beyond learned metadata based raw image reconstruction authors yufei wang yi yu wenhan yang lanqing guo lap pui chau alex c kot bihan wen subjects computer vision and pattern recognition cs cv image and video processing eess iv arxiv link pdf link abstract while raw images have distinct advantages over srgb images e g linearity and fine grained quantization levels they are not widely adopted by general users due to their substantial storage requirements very recent studies propose to compress raw images by designing sampling masks within the pixel space of the raw image however these approaches often leave space for pursuing more effective image representations and compact metadata in this work we propose a novel framework that learns a compact representation in the latent space serving as metadata in an end to end manner compared with lossy image compression we analyze the intrinsic difference of the raw image reconstruction task caused by rich information from the srgb image based on the analysis a novel backbone design with asymmetric and hybrid spatial feature resolutions is proposed which significantly improves the rate distortion performance besides we propose a novel design of the context model which can better predict the order masks of encoding decoding based on both the srgb image and the masks of already processed features benefited from the better modeling of the correlation between order masks the already processed information can be better utilized moreover a novel srgb guided adaptive quantization precision strategy which dynamically assigns varying levels of quantization precision to different regions further enhances the representation ability of the model finally based on the iterative properties of the proposed context model we propose a novel strategy to achieve variable bit rates using a single model this strategy allows for the continuous convergence of a wide range of bit rates extensive experimental results demonstrate that the proposed method can achieve better reconstruction quality with a smaller metadata size keyword raw image beyond learned metadata based raw image reconstruction authors yufei wang yi yu wenhan yang lanqing guo lap pui chau alex c kot bihan wen subjects computer vision and pattern recognition cs cv image and video processing eess iv arxiv link pdf link abstract while raw images have distinct advantages over srgb images e g linearity and fine grained quantization levels they are not widely adopted by general users due to their substantial storage requirements very recent studies propose to compress raw images by designing sampling masks within the pixel space of the raw image however these approaches often leave space for pursuing more effective image representations and compact metadata in this work we propose a novel framework that learns a compact representation in the latent space serving as metadata in an end to end manner compared with lossy image compression we analyze the intrinsic difference of the raw image reconstruction task caused by rich information from the srgb image based on the analysis a novel backbone design with asymmetric and hybrid spatial feature resolutions is proposed which significantly improves the rate distortion performance besides we propose a novel design of the context model which can better predict the order masks of encoding decoding based on both the srgb image and the masks of already processed features benefited from the better modeling of the correlation between order masks the already processed information can be better utilized moreover a novel srgb guided adaptive quantization precision strategy which dynamically assigns varying levels of quantization precision to different regions further enhances the representation ability of the model finally based on the iterative properties of the proposed context model we propose a novel strategy to achieve variable bit rates using a single model this strategy allows for the continuous convergence of a wide range of bit rates extensive experimental results demonstrate that the proposed method can achieve better reconstruction quality with a smaller metadata size
1
20,866
27,645,774,705
IssuesEvent
2023-03-10 22:50:24
cse442-at-ub/project_s23-iweatherify
https://api.github.com/repos/cse442-at-ub/project_s23-iweatherify
closed
Create tables for database
Processing Task Sprint 2
**Task Tests** *Test 1* 1) Go to https://www-student.cse.buffalo.edu/tools/db/phpmyadmin 2) For the username, type in vwong27 3) For the password, type in 50342607 3) Ensure that the server name host is oceanus 4) Click "Go" 5) On the left hand panel click on "cse442_2023_spring_team_a_db" 6) Verify that there exists at least one table in the database
1.0
Create tables for database - **Task Tests** *Test 1* 1) Go to https://www-student.cse.buffalo.edu/tools/db/phpmyadmin 2) For the username, type in vwong27 3) For the password, type in 50342607 3) Ensure that the server name host is oceanus 4) Click "Go" 5) On the left hand panel click on "cse442_2023_spring_team_a_db" 6) Verify that there exists at least one table in the database
process
create tables for database task tests test go to for the username type in for the password type in ensure that the server name host is oceanus click go on the left hand panel click on spring team a db verify that there exists at least one table in the database
1
34,564
30,180,770,963
IssuesEvent
2023-07-04 08:44:46
Chaste/Chaste
https://api.github.com/repos/Chaste/Chaste
closed
#3053 - Improve performance of Chaste webserver
component: infrastructure priority: high
This issue continues the discussion for legacy trac ticket 3053: --- [jmpf](https://github.com/orgs/Chaste/people/jmpf) created the following ticket on 2020-10-22 at 11:40:57, it is owned by [jmpf](https://github.com/orgs/Chaste/people/jmpf) Improve performance of Chaste web-server. Most nights it's running out of memory and swapping. This is disruptive to people working in other time zones. --- Comment by [jmpf](https://github.com/orgs/Chaste/people/jmpf) on 2020-10-22 at 11:48:32 Since Feb 2019 I have had the machine record its load-average and swap usage in half-hourly intervals. (See crontab for chaste@chaste.) I have also had the machine attempt to recover swap space at 8am every morning. If the nightly backup is still going then this may fail. (See crontab for root@chaste.) I have also produced email alerts for lack of webserver on an hourly basis. This is done using wget and looking for "Internal Server Error". (See crontab for jmpf@his-work-desktop.) --- Comment by [jmpf](https://github.com/orgs/Chaste/people/jmpf) on 2020-10-22 at 11:54:37 Record of the biggest files being backed up: ``` 405,957 21-10-2020 00:29:16 Normal File--> 405,957 /var/www/scratch/buildbot/master/public_html/Nightly Google Profile/profile1127/Test2DMeshBasedCryptRepresentativeSimulation.prof [Sent] 409,535 21-10-2020 00:29:12 Normal File--> 409,535 /var/www/scratch/buildbot/master/public_html/Nightly Coverage/coverage1174/mesh/src/vertex/MutableVertexMesh.cpp.gcov.html [Sent] 479,846 21-10-2020 00:29:17 Normal File--> 479,846 /var/www/scratch/buildbot/master/public_html/Nightly Google Profile/profile1127/Test3dBidomainProblemForEfficiencyWithFasterOdes.prof [Sent] 511,487 21-10-2020 00:29:17 Normal File--> 511,487 /var/www/scratch/buildbot/master/public_html/Nightly Google Profile/profile1127/Test3dBidomainProblemWithMetisForEfficiency.prof [Sent] 525,068 21-10-2020 00:29:17 Normal File--> 525,068 /var/www/scratch/buildbot/master/public_html/Nightly Google Profile/profile1127/Test3dBidomainProblemWithPermForEfficiency.prof [Sent] 577,478 21-10-2020 00:33:27 Normal File--> 577,478 /var/log/apache2/error.log [Sent] 964,604 21-10-2020 00:29:17 Normal File--> 964,604 /var/www/scratch/buildbot/master/public_html/Nightly Google Profile/profile1127/Test3dBidomainProblemForEfficiency.prof [Sent] 990,654 21-10-2020 00:29:18 Normal File--> 990,654 /var/www/scratch/buildbot/master/public_html/Nightly Google Profile/profile1127/Test3dOffLatticeRepresentativeSimulation.prof [Sent] 1,951,255 21-10-2020 00:29:17 Normal File--> 1,951,255 /var/www/scratch/buildbot/master/public_html/Nightly Google Profile/profile1127/Test2dVertexBasedSimulationWithFreeBoundary.prof [Sent] 2,292,032 21-10-2020 00:29:16 Normal File--> 2,292,032 /var/www/scratch/buildbot/master/public_html/Nightly Google Profile/profile1127/Test2DVertexBasedCryptRepresentativeSimulation.prof [Sent] 2,811,807 21-10-2020 00:29:18 Normal File--> 2,811,807 /var/www/scratch/buildbot/master/public_html/Nightly Google Profile/profile1127/TestLongPostprocessing.prof [Sent] 3,860,718 21-10-2020 00:32:27 Normal File--> 3,860,718 /var/lib/apt-xapian-index/cataloged_times.p [Sent] 4,730,249 21-10-2020 00:32:26 Normal File--> 4,730,249 /home/chaste/forensic [Sent] 5,242,880 21-10-2020 00:33:16 Normal File--> 5,242,880 /var/lib/mysql/ib_logfile0 [Sent] 5,242,880 21-10-2020 00:33:17 Normal File--> 5,242,880 /var/lib/mysql/ib_logfile1 [Sent] 6,141,311 21-10-2020 00:33:21 Normal File--> 6,141,311 /var/log/dsminstr.log [Sent] 14,514,672 21-10-2020 00:33:23 Normal File--> 14,514,672 /var/log/dsmsched.log Changed 20,806,656 21-10-2020 00:31:09 Normal File--> 20,806,656 /var/www/scratch/chaste.bak/db/rep-cache.db [Sent] 25,640,960 21-10-2020 00:14:39 Normal File--> 25,640,960 /var/www/scratch/buildbot/master/state.sqlite [Sent] 26,422,578 21-10-2020 00:33:27 Normal File--> 26,422,578 /var/log/apache2/access.log Changed 27,262,976 21-10-2020 00:33:21 Normal File--> 27,262,976 /var/lib/mysql/ibdata1 [Sent] 46,948,528 21-10-2020 00:37:09 Normal File--> 46,948,528 /var/mail/svn [Sent] 699,631,490 21-10-2020 00:33:15 Normal File--> 699,631,490 /var/lib/mlocate/mlocate.db [Sent] 1,625,510,849 21-10-2020 00:37:01 Normal File--> 1,625,510,849 /var/mail/root [Sent] 2,570,941,138 21-10-2020 00:25:11 Normal File--> 2,570,941,138 /var/www/scratch/buildbot/master/twistd.log Changed 10,642,110,311 21-10-2020 00:14:33 Normal File--> 10,642,110,311 /var/www/scratch/buildbot/master/http.log [Sent] ``` Looks like backups weren't properly configured. A particular problem is * /var/log/dsmsched.log (which records all transactions) is backed up * backups launched via crontab send a record of all transactions to root@chaste --- Comment by [jmpf](https://github.com/orgs/Chaste/people/jmpf) on 2020-10-22 at 11:57:59 ``` sudo rm /var/mail/root ``` --- Comment by [jmpf](https://github.com/orgs/Chaste/people/jmpf) on 2020-10-22 at 12:12:30 Reconfigure backup ``` $ diff -u /opt/tivoli/tsm/client/ba/bin/dsm.sys.0 /opt/tivoli/tsm/client/ba/bin/dsm.sys --- /opt/tivoli/tsm/client/ba/bin/dsm.sys.0 2020-10-22 12:57:13.272333478 +0100 +++ /opt/tivoli/tsm/client/ba/bin/dsm.sys 2020-10-22 13:06:41.739554442 +0100 @@ -132,3 +132,9 @@ *Exclude.fs /tmp *Exclude.fs /var/tmp +* Don't backup log files (largest is currently ~40Mb) +Exclude.dir /var/log + +* Don't backup buildbot's logs (10Gb and 2Gb and stored on spinning rust drive). +Exclude /var/www/scratch/buildbot/master/http.log +Exclude /var/www/scratch/buildbot/master/twistd.log ``` ``` $ sudo /opt/tivoli/tsm/client/ba/bin/HFSscheduler stop $ sudo /opt/tivoli/tsm/client/ba/bin/HFSscheduler start ``` --- Comment by [jmpf](https://github.com/orgs/Chaste/people/jmpf) on 2020-10-22 at 12:13:22 Stop stdout of the cron launched backup going to email (it's logged already). ``` diff -u crontab.old crontab.new --- crontab.old 2020-10-22 13:11:07.433060187 +0100 +++ crontab.new 2020-10-22 13:11:51.645310653 +0100 @@ -20,5 +20,5 @@ # For more information see the manual pages of crontab(5) and cron(8) # # m h dom mon dow command -15 0 * * 1,2,4,5,6 dsmc incremental # Back up everything after midnight -* 8 * * * /sbin/swapoff -a && /sbin/swapon -a # Recover any swap at 8am +15 0 * * 1,2,4,5,6 dsmc incremental > /dev/null # Back up everything after midnight +* 8 * * * /sbin/swapoff -a && /sbin/swapon -a # Recover any swap at 8am ``` --- Comment by [jmpf](https://github.com/orgs/Chaste/people/jmpf) on 2020-11-05 at 16:20:47 This morning Thu 5 November the Chaste Trac was unavailable in the late morning. Machine was still completing the Wednesday dsm backup at lunchtime and swap space had not been recovered. Trac killed. Apache restarted. Backup left to complete. Machine settled down. --- Comment by [jmpf](https://github.com/orgs/Chaste/people/jmpf) on 2020-11-11 at 17:50:20 Last night (Tues 10 Nov) the automatic scheduler failed to trigger and I had to restart the HFSscheduler service. --- Comment by [jmpf](https://github.com/orgs/Chaste/people/jmpf) on 2020-11-25 at 10:40:01 Last night (Tues 25 Nov) at 22:10 the website was unavailable. Scheduled backup happened normally but the machine was still swapping this morning. I'm suspicious of ``` 24-11-2020 22:38:47 Normal File--> 2,493,455,360 /var/www/chaste_test_data/.stats-nightly.db [Sent] ``` which may have filled the swap sometime after 10pm. Excluding: ``` chaste@chaste:~$ diff -u /opt/tivoli/tsm/client/ba/bin/dsm.sys.1 /opt/tivoli/tsm/client/ba/bin/dsm.sys --- /opt/tivoli/tsm/client/ba/bin/dsm.sys.1 2020-11-25 10:36:25.410851734 +0000 +++ /opt/tivoli/tsm/client/ba/bin/dsm.sys 2020-11-25 10:38:57.867730857 +0000 @@ -138,3 +138,5 @@ * Don't backup buildbot's logs (10Gb and 2Gb and stored on spinning rust drive). Exclude /var/www/scratch/buildbot/master/http.log Exclude /var/www/scratch/buildbot/master/twistd.log +* Don't backup nightly test database (2.5Gb stored on spinning rust drive). +Exclude /var/www/chaste_test_data/.stats-nightly.db ```
1.0
#3053 - Improve performance of Chaste webserver - This issue continues the discussion for legacy trac ticket 3053: --- [jmpf](https://github.com/orgs/Chaste/people/jmpf) created the following ticket on 2020-10-22 at 11:40:57, it is owned by [jmpf](https://github.com/orgs/Chaste/people/jmpf) Improve performance of Chaste web-server. Most nights it's running out of memory and swapping. This is disruptive to people working in other time zones. --- Comment by [jmpf](https://github.com/orgs/Chaste/people/jmpf) on 2020-10-22 at 11:48:32 Since Feb 2019 I have had the machine record its load-average and swap usage in half-hourly intervals. (See crontab for chaste@chaste.) I have also had the machine attempt to recover swap space at 8am every morning. If the nightly backup is still going then this may fail. (See crontab for root@chaste.) I have also produced email alerts for lack of webserver on an hourly basis. This is done using wget and looking for "Internal Server Error". (See crontab for jmpf@his-work-desktop.) --- Comment by [jmpf](https://github.com/orgs/Chaste/people/jmpf) on 2020-10-22 at 11:54:37 Record of the biggest files being backed up: ``` 405,957 21-10-2020 00:29:16 Normal File--> 405,957 /var/www/scratch/buildbot/master/public_html/Nightly Google Profile/profile1127/Test2DMeshBasedCryptRepresentativeSimulation.prof [Sent] 409,535 21-10-2020 00:29:12 Normal File--> 409,535 /var/www/scratch/buildbot/master/public_html/Nightly Coverage/coverage1174/mesh/src/vertex/MutableVertexMesh.cpp.gcov.html [Sent] 479,846 21-10-2020 00:29:17 Normal File--> 479,846 /var/www/scratch/buildbot/master/public_html/Nightly Google Profile/profile1127/Test3dBidomainProblemForEfficiencyWithFasterOdes.prof [Sent] 511,487 21-10-2020 00:29:17 Normal File--> 511,487 /var/www/scratch/buildbot/master/public_html/Nightly Google Profile/profile1127/Test3dBidomainProblemWithMetisForEfficiency.prof [Sent] 525,068 21-10-2020 00:29:17 Normal File--> 525,068 /var/www/scratch/buildbot/master/public_html/Nightly Google Profile/profile1127/Test3dBidomainProblemWithPermForEfficiency.prof [Sent] 577,478 21-10-2020 00:33:27 Normal File--> 577,478 /var/log/apache2/error.log [Sent] 964,604 21-10-2020 00:29:17 Normal File--> 964,604 /var/www/scratch/buildbot/master/public_html/Nightly Google Profile/profile1127/Test3dBidomainProblemForEfficiency.prof [Sent] 990,654 21-10-2020 00:29:18 Normal File--> 990,654 /var/www/scratch/buildbot/master/public_html/Nightly Google Profile/profile1127/Test3dOffLatticeRepresentativeSimulation.prof [Sent] 1,951,255 21-10-2020 00:29:17 Normal File--> 1,951,255 /var/www/scratch/buildbot/master/public_html/Nightly Google Profile/profile1127/Test2dVertexBasedSimulationWithFreeBoundary.prof [Sent] 2,292,032 21-10-2020 00:29:16 Normal File--> 2,292,032 /var/www/scratch/buildbot/master/public_html/Nightly Google Profile/profile1127/Test2DVertexBasedCryptRepresentativeSimulation.prof [Sent] 2,811,807 21-10-2020 00:29:18 Normal File--> 2,811,807 /var/www/scratch/buildbot/master/public_html/Nightly Google Profile/profile1127/TestLongPostprocessing.prof [Sent] 3,860,718 21-10-2020 00:32:27 Normal File--> 3,860,718 /var/lib/apt-xapian-index/cataloged_times.p [Sent] 4,730,249 21-10-2020 00:32:26 Normal File--> 4,730,249 /home/chaste/forensic [Sent] 5,242,880 21-10-2020 00:33:16 Normal File--> 5,242,880 /var/lib/mysql/ib_logfile0 [Sent] 5,242,880 21-10-2020 00:33:17 Normal File--> 5,242,880 /var/lib/mysql/ib_logfile1 [Sent] 6,141,311 21-10-2020 00:33:21 Normal File--> 6,141,311 /var/log/dsminstr.log [Sent] 14,514,672 21-10-2020 00:33:23 Normal File--> 14,514,672 /var/log/dsmsched.log Changed 20,806,656 21-10-2020 00:31:09 Normal File--> 20,806,656 /var/www/scratch/chaste.bak/db/rep-cache.db [Sent] 25,640,960 21-10-2020 00:14:39 Normal File--> 25,640,960 /var/www/scratch/buildbot/master/state.sqlite [Sent] 26,422,578 21-10-2020 00:33:27 Normal File--> 26,422,578 /var/log/apache2/access.log Changed 27,262,976 21-10-2020 00:33:21 Normal File--> 27,262,976 /var/lib/mysql/ibdata1 [Sent] 46,948,528 21-10-2020 00:37:09 Normal File--> 46,948,528 /var/mail/svn [Sent] 699,631,490 21-10-2020 00:33:15 Normal File--> 699,631,490 /var/lib/mlocate/mlocate.db [Sent] 1,625,510,849 21-10-2020 00:37:01 Normal File--> 1,625,510,849 /var/mail/root [Sent] 2,570,941,138 21-10-2020 00:25:11 Normal File--> 2,570,941,138 /var/www/scratch/buildbot/master/twistd.log Changed 10,642,110,311 21-10-2020 00:14:33 Normal File--> 10,642,110,311 /var/www/scratch/buildbot/master/http.log [Sent] ``` Looks like backups weren't properly configured. A particular problem is * /var/log/dsmsched.log (which records all transactions) is backed up * backups launched via crontab send a record of all transactions to root@chaste --- Comment by [jmpf](https://github.com/orgs/Chaste/people/jmpf) on 2020-10-22 at 11:57:59 ``` sudo rm /var/mail/root ``` --- Comment by [jmpf](https://github.com/orgs/Chaste/people/jmpf) on 2020-10-22 at 12:12:30 Reconfigure backup ``` $ diff -u /opt/tivoli/tsm/client/ba/bin/dsm.sys.0 /opt/tivoli/tsm/client/ba/bin/dsm.sys --- /opt/tivoli/tsm/client/ba/bin/dsm.sys.0 2020-10-22 12:57:13.272333478 +0100 +++ /opt/tivoli/tsm/client/ba/bin/dsm.sys 2020-10-22 13:06:41.739554442 +0100 @@ -132,3 +132,9 @@ *Exclude.fs /tmp *Exclude.fs /var/tmp +* Don't backup log files (largest is currently ~40Mb) +Exclude.dir /var/log + +* Don't backup buildbot's logs (10Gb and 2Gb and stored on spinning rust drive). +Exclude /var/www/scratch/buildbot/master/http.log +Exclude /var/www/scratch/buildbot/master/twistd.log ``` ``` $ sudo /opt/tivoli/tsm/client/ba/bin/HFSscheduler stop $ sudo /opt/tivoli/tsm/client/ba/bin/HFSscheduler start ``` --- Comment by [jmpf](https://github.com/orgs/Chaste/people/jmpf) on 2020-10-22 at 12:13:22 Stop stdout of the cron launched backup going to email (it's logged already). ``` diff -u crontab.old crontab.new --- crontab.old 2020-10-22 13:11:07.433060187 +0100 +++ crontab.new 2020-10-22 13:11:51.645310653 +0100 @@ -20,5 +20,5 @@ # For more information see the manual pages of crontab(5) and cron(8) # # m h dom mon dow command -15 0 * * 1,2,4,5,6 dsmc incremental # Back up everything after midnight -* 8 * * * /sbin/swapoff -a && /sbin/swapon -a # Recover any swap at 8am +15 0 * * 1,2,4,5,6 dsmc incremental > /dev/null # Back up everything after midnight +* 8 * * * /sbin/swapoff -a && /sbin/swapon -a # Recover any swap at 8am ``` --- Comment by [jmpf](https://github.com/orgs/Chaste/people/jmpf) on 2020-11-05 at 16:20:47 This morning Thu 5 November the Chaste Trac was unavailable in the late morning. Machine was still completing the Wednesday dsm backup at lunchtime and swap space had not been recovered. Trac killed. Apache restarted. Backup left to complete. Machine settled down. --- Comment by [jmpf](https://github.com/orgs/Chaste/people/jmpf) on 2020-11-11 at 17:50:20 Last night (Tues 10 Nov) the automatic scheduler failed to trigger and I had to restart the HFSscheduler service. --- Comment by [jmpf](https://github.com/orgs/Chaste/people/jmpf) on 2020-11-25 at 10:40:01 Last night (Tues 25 Nov) at 22:10 the website was unavailable. Scheduled backup happened normally but the machine was still swapping this morning. I'm suspicious of ``` 24-11-2020 22:38:47 Normal File--> 2,493,455,360 /var/www/chaste_test_data/.stats-nightly.db [Sent] ``` which may have filled the swap sometime after 10pm. Excluding: ``` chaste@chaste:~$ diff -u /opt/tivoli/tsm/client/ba/bin/dsm.sys.1 /opt/tivoli/tsm/client/ba/bin/dsm.sys --- /opt/tivoli/tsm/client/ba/bin/dsm.sys.1 2020-11-25 10:36:25.410851734 +0000 +++ /opt/tivoli/tsm/client/ba/bin/dsm.sys 2020-11-25 10:38:57.867730857 +0000 @@ -138,3 +138,5 @@ * Don't backup buildbot's logs (10Gb and 2Gb and stored on spinning rust drive). Exclude /var/www/scratch/buildbot/master/http.log Exclude /var/www/scratch/buildbot/master/twistd.log +* Don't backup nightly test database (2.5Gb stored on spinning rust drive). +Exclude /var/www/chaste_test_data/.stats-nightly.db ```
non_process
improve performance of chaste webserver this issue continues the discussion for legacy trac ticket created the following ticket on at it is owned by improve performance of chaste web server most nights it s running out of memory and swapping this is disruptive to people working in other time zones comment by on at since feb i have had the machine record its load average and swap usage in half hourly intervals see crontab for chaste chaste i have also had the machine attempt to recover swap space at every morning if the nightly backup is still going then this may fail see crontab for root chaste i have also produced email alerts for lack of webserver on an hourly basis this is done using wget and looking for internal server error see crontab for jmpf his work desktop comment by on at record of the biggest files being backed up normal file var www scratch buildbot master public html nightly google profile prof normal file var www scratch buildbot master public html nightly coverage mesh src vertex mutablevertexmesh cpp gcov html normal file var www scratch buildbot master public html nightly google profile prof normal file var www scratch buildbot master public html nightly google profile prof normal file var www scratch buildbot master public html nightly google profile prof normal file var log error log normal file var www scratch buildbot master public html nightly google profile prof normal file var www scratch buildbot master public html nightly google profile prof normal file var www scratch buildbot master public html nightly google profile prof normal file var www scratch buildbot master public html nightly google profile prof normal file var www scratch buildbot master public html nightly google profile testlongpostprocessing prof normal file var lib apt xapian index cataloged times p normal file home chaste forensic normal file var lib mysql ib normal file var lib mysql ib normal file var log dsminstr log normal file var log dsmsched log changed normal file var www scratch chaste bak db rep cache db normal file var www scratch buildbot master state sqlite normal file var log access log changed normal file var lib mysql normal file var mail svn normal file var lib mlocate mlocate db normal file var mail root normal file var www scratch buildbot master twistd log changed normal file var www scratch buildbot master http log looks like backups weren t properly configured a particular problem is var log dsmsched log which records all transactions is backed up backups launched via crontab send a record of all transactions to root chaste comment by on at sudo rm var mail root comment by on at reconfigure backup diff u opt tivoli tsm client ba bin dsm sys opt tivoli tsm client ba bin dsm sys opt tivoli tsm client ba bin dsm sys opt tivoli tsm client ba bin dsm sys exclude fs tmp exclude fs var tmp don t backup log files largest is currently exclude dir var log don t backup buildbot s logs and and stored on spinning rust drive exclude var www scratch buildbot master http log exclude var www scratch buildbot master twistd log sudo opt tivoli tsm client ba bin hfsscheduler stop sudo opt tivoli tsm client ba bin hfsscheduler start comment by on at stop stdout of the cron launched backup going to email it s logged already diff u crontab old crontab new crontab old crontab new for more information see the manual pages of crontab and cron m h dom mon dow command dsmc incremental back up everything after midnight sbin swapoff a sbin swapon a recover any swap at dsmc incremental dev null back up everything after midnight sbin swapoff a sbin swapon a recover any swap at comment by on at this morning thu november the chaste trac was unavailable in the late morning machine was still completing the wednesday dsm backup at lunchtime and swap space had not been recovered trac killed apache restarted backup left to complete machine settled down comment by on at last night tues nov the automatic scheduler failed to trigger and i had to restart the hfsscheduler service comment by on at last night tues nov at the website was unavailable scheduled backup happened normally but the machine was still swapping this morning i m suspicious of normal file var www chaste test data stats nightly db which may have filled the swap sometime after excluding chaste chaste diff u opt tivoli tsm client ba bin dsm sys opt tivoli tsm client ba bin dsm sys opt tivoli tsm client ba bin dsm sys opt tivoli tsm client ba bin dsm sys don t backup buildbot s logs and and stored on spinning rust drive exclude var www scratch buildbot master http log exclude var www scratch buildbot master twistd log don t backup nightly test database stored on spinning rust drive exclude var www chaste test data stats nightly db
0
21,364
29,194,080,011
IssuesEvent
2023-05-20 00:31:44
devssa/onde-codar-em-salvador
https://api.github.com/repos/devssa/onde-codar-em-salvador
closed
[Hibrido / Caieiras, São Paulo, Brazil] Front-end Developer (Júnior) (Híbrido) na Coodesh
SALVADOR FRONT-END PJ BANCO DE DADOS JAVASCRIPT FULL-STACK HTML SQL ANGULAR REACT VUE REQUISITOS PROCESSOS GITHUB UMA AUTOMAÇÃO DE PROCESSOS HIBRIDO ALOCADO Stale
## Descrição da vaga: Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios. Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/vagas/fullstack-developer-netangular-hibrido-142048389?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋 <p>A <strong>TRIFORSYS</strong> está em busca de <strong><ins>Front-end Developer (Júnior)</ins></strong> para integrar seu time!</p> <p>Fábrica Softwares e soluções em TI de forma personalizada, atendendo a necessidade do nosso cliente com foco na automação de processos.</p> <p><strong>Sobre a empresa</strong></p> <p>Somos sua melhor experiência com softwares inteligentes e descomplicados que resolvem sua necessidade, apostamos na forma simples e prática de falar e tratar dos teus processos, considerando e respeitando as particularidades e características que fazem a diferença no seu negócio.</p> <p>Cuidamos de seus objetivos e metas para melhores soluções de seu negócio. Construindo desde um software completo até ferramentas acessórias totalmente integradas ao seu sistema atual, tudo isso para que você tenha controle total sobre a gestão de sua empresa e de sua operação. Tratamos cada cliente como único e especial, considerando o tamanho da empresa, ramo de atividade no mercado, produto, serviço e principalmente sua história, que faz sua empresa ser o que realmente é.</p> <p></p> <p>Venha com a gente fazer seu negócio!</p> ## Triforsys: <p>Somos sua melhor experiencia com softwares inteligentes e descomplicados que resolvem sua necessidade, apostamos na forma simples e prática de falar e tratar dos teus processos, considerando e respeitando as particularidades e características que fazem a diferença no seu negócio.</p> <p>Cuidamos de seus objetivos e metas para melhores soluções de seu negócio. Construindo desde um software completo até ferramentas acessórias totalmente integradas ao seu sistema atual, tudo isso para que você tenha controle total sobre a gestão de sua empresa e de sua operação. Tratamos cada cliente como único e especial, considerando o tamanho da empresa, ramo de atividade no mercado, produto, serviço e principalmente sua história, que faz sua empresa ser o que realmente é.</p><a href='https://coodesh.com/empresas/triforsys-desenvolvimento-de-softwares-e-aplicativos'>Veja mais no site</a> ## Habilidades: - Angular - HTML - CSS - React.js - Banco de dados relacionais (SQL) ## Local: Caieiras, São Paulo, Brazil ## Requisitos: - Residir em Caieiras ou proximidades (Formato híbrido); - Conhecimentos em Front-end (HTML, CSS, Javascript). ## Diferenciais: - Conhecimentos em Banco de dados(SqlServer); - Conhecimentos em React/Angular/Vue. ## Como se candidatar: Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Front-end Developer (Júnior) (Híbrido) na Triforsys](https://coodesh.com/vagas/fullstack-developer-netangular-hibrido-142048389?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação. ## Labels #### Alocação Alocado #### Regime PJ #### Categoria Full-Stack
2.0
[Hibrido / Caieiras, São Paulo, Brazil] Front-end Developer (Júnior) (Híbrido) na Coodesh - ## Descrição da vaga: Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios. Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/vagas/fullstack-developer-netangular-hibrido-142048389?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋 <p>A <strong>TRIFORSYS</strong> está em busca de <strong><ins>Front-end Developer (Júnior)</ins></strong> para integrar seu time!</p> <p>Fábrica Softwares e soluções em TI de forma personalizada, atendendo a necessidade do nosso cliente com foco na automação de processos.</p> <p><strong>Sobre a empresa</strong></p> <p>Somos sua melhor experiência com softwares inteligentes e descomplicados que resolvem sua necessidade, apostamos na forma simples e prática de falar e tratar dos teus processos, considerando e respeitando as particularidades e características que fazem a diferença no seu negócio.</p> <p>Cuidamos de seus objetivos e metas para melhores soluções de seu negócio. Construindo desde um software completo até ferramentas acessórias totalmente integradas ao seu sistema atual, tudo isso para que você tenha controle total sobre a gestão de sua empresa e de sua operação. Tratamos cada cliente como único e especial, considerando o tamanho da empresa, ramo de atividade no mercado, produto, serviço e principalmente sua história, que faz sua empresa ser o que realmente é.</p> <p></p> <p>Venha com a gente fazer seu negócio!</p> ## Triforsys: <p>Somos sua melhor experiencia com softwares inteligentes e descomplicados que resolvem sua necessidade, apostamos na forma simples e prática de falar e tratar dos teus processos, considerando e respeitando as particularidades e características que fazem a diferença no seu negócio.</p> <p>Cuidamos de seus objetivos e metas para melhores soluções de seu negócio. Construindo desde um software completo até ferramentas acessórias totalmente integradas ao seu sistema atual, tudo isso para que você tenha controle total sobre a gestão de sua empresa e de sua operação. Tratamos cada cliente como único e especial, considerando o tamanho da empresa, ramo de atividade no mercado, produto, serviço e principalmente sua história, que faz sua empresa ser o que realmente é.</p><a href='https://coodesh.com/empresas/triforsys-desenvolvimento-de-softwares-e-aplicativos'>Veja mais no site</a> ## Habilidades: - Angular - HTML - CSS - React.js - Banco de dados relacionais (SQL) ## Local: Caieiras, São Paulo, Brazil ## Requisitos: - Residir em Caieiras ou proximidades (Formato híbrido); - Conhecimentos em Front-end (HTML, CSS, Javascript). ## Diferenciais: - Conhecimentos em Banco de dados(SqlServer); - Conhecimentos em React/Angular/Vue. ## Como se candidatar: Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Front-end Developer (Júnior) (Híbrido) na Triforsys](https://coodesh.com/vagas/fullstack-developer-netangular-hibrido-142048389?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação. ## Labels #### Alocação Alocado #### Regime PJ #### Categoria Full-Stack
process
front end developer júnior híbrido na coodesh descrição da vaga esta é uma vaga de um parceiro da plataforma coodesh ao candidatar se você terá acesso as informações completas sobre a empresa e benefícios fique atento ao redirecionamento que vai te levar para uma url com o pop up personalizado de candidatura 👋 a triforsys está em busca de front end developer júnior para integrar seu time fábrica softwares e soluções em ti de forma personalizada atendendo a necessidade do nosso cliente com foco na automação de processos sobre a empresa somos sua melhor experiência com softwares inteligentes e descomplicados que resolvem sua necessidade apostamos na forma simples e prática de falar e tratar dos teus processos considerando e respeitando as particularidades e características que fazem a diferença no seu negócio cuidamos de seus objetivos e metas para melhores soluções de seu negócio construindo desde um software completo até ferramentas acessórias totalmente integradas ao seu sistema atual tudo isso para que você tenha controle total sobre a gestão de sua empresa e de sua operação tratamos cada cliente como único e especial considerando o tamanho da empresa ramo de atividade no mercado produto serviço e principalmente sua história que faz sua empresa ser o que realmente é venha com a gente fazer seu negócio triforsys somos sua melhor experiencia com softwares inteligentes e descomplicados que resolvem sua necessidade apostamos na forma simples e prática de falar e tratar dos teus processos considerando e respeitando as particularidades e características que fazem a diferença no seu negócio cuidamos de seus objetivos e metas para melhores soluções de seu negócio construindo desde um software completo até ferramentas acessórias totalmente integradas ao seu sistema atual tudo isso para que você tenha controle total sobre a gestão de sua empresa e de sua operação tratamos cada cliente como único e especial considerando o tamanho da empresa ramo de atividade no mercado produto serviço e principalmente sua história que faz sua empresa ser o que realmente é habilidades angular html css react js banco de dados relacionais sql local caieiras são paulo brazil requisitos residir em caieiras ou proximidades formato híbrido conhecimentos em front end html css javascript diferenciais conhecimentos em banco de dados sqlserver conhecimentos em react angular vue como se candidatar candidatar se exclusivamente através da plataforma coodesh no link a seguir após candidatar se via plataforma coodesh e validar o seu login você poderá acompanhar e receber todas as interações do processo por lá utilize a opção pedir feedback entre uma etapa e outra na vaga que se candidatou isso fará com que a pessoa recruiter responsável pelo processo na empresa receba a notificação labels alocação alocado regime pj categoria full stack
1
2,906
2,533,941,786
IssuesEvent
2015-01-24 11:52:31
zendframework/modules.zendframework.com
https://api.github.com/repos/zendframework/modules.zendframework.com
closed
Performance Improvement
high priority
Currently parts of the page are facing huge loading times due to several repeated requests to the github api. Locate performance bottlenecks and implement caching or over improvements.
1.0
Performance Improvement - Currently parts of the page are facing huge loading times due to several repeated requests to the github api. Locate performance bottlenecks and implement caching or over improvements.
non_process
performance improvement currently parts of the page are facing huge loading times due to several repeated requests to the github api locate performance bottlenecks and implement caching or over improvements
0
17,074
22,575,021,268
IssuesEvent
2022-06-28 06:25:03
weiquany/KTVAnywhere
https://api.github.com/repos/weiquany/KTVAnywhere
closed
Lyrics retrieval
feature: song preprocessing
Automatically retrieve lyrics file for a song online if possible. ## User story As a user, * I want to have lyrics for KTV, and I want them to be provided for me. ### Acceptance criteria The application should: - [x] Be able to retrieve lyrics if they can be found - [x] Allow users to choose to let the application retrieve lyrics file ## Complexities * Finding the appropriate lyrics file given the song name and artist
1.0
Lyrics retrieval - Automatically retrieve lyrics file for a song online if possible. ## User story As a user, * I want to have lyrics for KTV, and I want them to be provided for me. ### Acceptance criteria The application should: - [x] Be able to retrieve lyrics if they can be found - [x] Allow users to choose to let the application retrieve lyrics file ## Complexities * Finding the appropriate lyrics file given the song name and artist
process
lyrics retrieval automatically retrieve lyrics file for a song online if possible user story as a user i want to have lyrics for ktv and i want them to be provided for me acceptance criteria the application should be able to retrieve lyrics if they can be found allow users to choose to let the application retrieve lyrics file complexities finding the appropriate lyrics file given the song name and artist
1
21,037
27,979,263,476
IssuesEvent
2023-03-26 00:20:10
darktable-org/darktable
https://api.github.com/repos/darktable-org/darktable
closed
Enhanced dehazing algorithm (with new code)
feature: enhancement wip scope: image processing no-issue-activity
First I'll pre-apologize for not being GIT trained enough to do this myself. I have a little write up of an enhanced version of the haze removal tool (attached). I simply overwrote the previous src/iop/hazeremoval.c with my changes, which is attached. But I think you'd want a whole new module name if you think it's worthy and I have no idea how to do it. I don't have opencl versions of the algorithm, and am probably not the person to attempt it. I've been using this over a year now and am pretty happy with it, but it is in need of some help from someone who knows darktable coding protocols well (which is not me). I'll be happy to help in any way my skill set will allow :-D Cheers, Jeff Welty [An updated algorithm for haze removal.pdf](https://github.com/darktable-org/darktable/files/7261734/An.updated.algorithm.for.haze.removal.pdf) [hazeremoval.c.txt](https://github.com/darktable-org/darktable/files/7312329/hazeremoval.c.txt) [MoreDehazeExamples.pdf](https://github.com/darktable-org/darktable/files/7312332/MoreDehazeExamples.pdf) [Screenshots_ehanced_module.pdf](https://github.com/darktable-org/darktable/files/7312333/Screenshots_ehanced_module.pdf)
1.0
Enhanced dehazing algorithm (with new code) - First I'll pre-apologize for not being GIT trained enough to do this myself. I have a little write up of an enhanced version of the haze removal tool (attached). I simply overwrote the previous src/iop/hazeremoval.c with my changes, which is attached. But I think you'd want a whole new module name if you think it's worthy and I have no idea how to do it. I don't have opencl versions of the algorithm, and am probably not the person to attempt it. I've been using this over a year now and am pretty happy with it, but it is in need of some help from someone who knows darktable coding protocols well (which is not me). I'll be happy to help in any way my skill set will allow :-D Cheers, Jeff Welty [An updated algorithm for haze removal.pdf](https://github.com/darktable-org/darktable/files/7261734/An.updated.algorithm.for.haze.removal.pdf) [hazeremoval.c.txt](https://github.com/darktable-org/darktable/files/7312329/hazeremoval.c.txt) [MoreDehazeExamples.pdf](https://github.com/darktable-org/darktable/files/7312332/MoreDehazeExamples.pdf) [Screenshots_ehanced_module.pdf](https://github.com/darktable-org/darktable/files/7312333/Screenshots_ehanced_module.pdf)
process
enhanced dehazing algorithm with new code first i ll pre apologize for not being git trained enough to do this myself i have a little write up of an enhanced version of the haze removal tool attached i simply overwrote the previous src iop hazeremoval c with my changes which is attached but i think you d want a whole new module name if you think it s worthy and i have no idea how to do it i don t have opencl versions of the algorithm and am probably not the person to attempt it i ve been using this over a year now and am pretty happy with it but it is in need of some help from someone who knows darktable coding protocols well which is not me i ll be happy to help in any way my skill set will allow d cheers jeff welty
1
201,092
7,022,393,617
IssuesEvent
2017-12-22 10:21:47
Mandiklopper/People-Connect
https://api.github.com/repos/Mandiklopper/People-Connect
closed
Leave Balance Request rejection
High Priority
Days "Taken" should be reversed /returned back to the staff and the leave balance records should be updated once request is rejected
1.0
Leave Balance Request rejection - Days "Taken" should be reversed /returned back to the staff and the leave balance records should be updated once request is rejected
non_process
leave balance request rejection days taken should be reversed returned back to the staff and the leave balance records should be updated once request is rejected
0
95,528
27,533,620,434
IssuesEvent
2023-03-07 00:46:37
CasparCG/server
https://api.github.com/repos/CasparCG/server
closed
Make build-scripts checkout themselves when a build is triggered
type/enhancement build
### Expected Behaviour When a build is triggered automatically, the build-server should checkout the build-scripts itself from a repository. By going this approach, it would be enough with a pull request against the build-scripts repository to update the scripts when needed. ### Current Behaviour When changes are made to CasparCG/server that affect files, folders, and versions of media-scanner that needs to be packaged to the final build (for example), it requires changes to the build-scripts located on the build-server. Those changes are dependent ob those who have access to the actual building server, which is bad.
1.0
Make build-scripts checkout themselves when a build is triggered - ### Expected Behaviour When a build is triggered automatically, the build-server should checkout the build-scripts itself from a repository. By going this approach, it would be enough with a pull request against the build-scripts repository to update the scripts when needed. ### Current Behaviour When changes are made to CasparCG/server that affect files, folders, and versions of media-scanner that needs to be packaged to the final build (for example), it requires changes to the build-scripts located on the build-server. Those changes are dependent ob those who have access to the actual building server, which is bad.
non_process
make build scripts checkout themselves when a build is triggered expected behaviour when a build is triggered automatically the build server should checkout the build scripts itself from a repository by going this approach it would be enough with a pull request against the build scripts repository to update the scripts when needed current behaviour when changes are made to casparcg server that affect files folders and versions of media scanner that needs to be packaged to the final build for example it requires changes to the build scripts located on the build server those changes are dependent ob those who have access to the actual building server which is bad
0
17,707
23,596,256,774
IssuesEvent
2022-08-23 19:33:01
open-telemetry/opentelemetry-collector-contrib
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
closed
[processor/transform] Add ability to concat strings
good first issue priority:p2 processor/transform
**Is your feature request related to a problem? Please describe.** There are situations where a user might need to combine existing fields to be used as an argument in an Invocation. [Here is an example.](https://cloud-native.slack.com/archives/C01N6P7KR6W/p1657799193277499). At the moment the TQL has no build-in capability to do this and the transform processor has not function that can do it. **Describe the solution you'd like** I think we have 2 options. 1. **Add a "Concat" factory to the transform processor that takes 2 string arguments and returns a concatenation of the 2 string.** This is a solution that should be implemented now. 2. ~~Add something to the TQL so it knows how to do this natively.~~ We should do this, but in the future. For now, to enhance the transform processor, we'll use a Factory. I'm leaning towards option 1, as it is simpler to implement and factories are already a defined pattern.
1.0
[processor/transform] Add ability to concat strings - **Is your feature request related to a problem? Please describe.** There are situations where a user might need to combine existing fields to be used as an argument in an Invocation. [Here is an example.](https://cloud-native.slack.com/archives/C01N6P7KR6W/p1657799193277499). At the moment the TQL has no build-in capability to do this and the transform processor has not function that can do it. **Describe the solution you'd like** I think we have 2 options. 1. **Add a "Concat" factory to the transform processor that takes 2 string arguments and returns a concatenation of the 2 string.** This is a solution that should be implemented now. 2. ~~Add something to the TQL so it knows how to do this natively.~~ We should do this, but in the future. For now, to enhance the transform processor, we'll use a Factory. I'm leaning towards option 1, as it is simpler to implement and factories are already a defined pattern.
process
add ability to concat strings is your feature request related to a problem please describe there are situations where a user might need to combine existing fields to be used as an argument in an invocation at the moment the tql has no build in capability to do this and the transform processor has not function that can do it describe the solution you d like i think we have options add a concat factory to the transform processor that takes string arguments and returns a concatenation of the string this is a solution that should be implemented now add something to the tql so it knows how to do this natively we should do this but in the future for now to enhance the transform processor we ll use a factory i m leaning towards option as it is simpler to implement and factories are already a defined pattern
1
511,830
14,882,712,438
IssuesEvent
2021-01-20 12:18:18
Hypothesize/standard.js
https://api.github.com/repos/Hypothesize/standard.js
closed
Reimplement the outlier filter
high priority
Currently the data table filter function doesn't filter values when the operator is `outlier`. It should be reimplemented.
1.0
Reimplement the outlier filter - Currently the data table filter function doesn't filter values when the operator is `outlier`. It should be reimplemented.
non_process
reimplement the outlier filter currently the data table filter function doesn t filter values when the operator is outlier it should be reimplemented
0
211,409
23,818,045,127
IssuesEvent
2022-09-05 08:41:24
sast-automation-dev/BenchmarkJava-333
https://api.github.com/repos/sast-automation-dev/BenchmarkJava-333
opened
bootstrap-3.3.4.js: 6 vulnerabilities (highest severity is: 6.1)
security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.4.js</b></p></summary> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.4/js/bootstrap.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.4/js/bootstrap.js</a></p> <p>Path to vulnerable library: /scorecard/content/js/bootstrap.js</p> <p> <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/BenchmarkJava-333/commit/3c6dcf8b9e2f8272e0ae17dea63a48ce3ac0245b">3c6dcf8b9e2f8272e0ae17dea63a48ce3ac0245b</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2019-8331](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-8331) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-3.3.4.js | Direct | bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1 | &#10060; | | [CVE-2018-14040](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14040) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-3.3.4.js | Direct | org.webjars.npm:bootstrap:4.1.2,org.webjars:bootstrap:3.4.0 | &#10060; | | [CVE-2018-20677](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20677) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-3.3.4.js | Direct | Bootstrap - v3.4.0;NorDroN.AngularTemplate - 0.1.6;Dynamic.NET.Express.ProjectTemplates - 0.8.0;dotnetng.template - 1.0.0.4;ZNxtApp.Core.Module.Theme - 1.0.9-Beta;JMeter - 5.0.0 | &#10060; | | [CVE-2018-20676](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20676) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-3.3.4.js | Direct | bootstrap - 3.4.0 | &#10060; | | [CVE-2018-14042](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14042) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-3.3.4.js | Direct | org.webjars.npm:bootstrap:4.1.2.org.webjars:bootstrap:3.4.0 | &#10060; | | [CVE-2016-10735](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10735) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-3.3.4.js | Direct | 3.4.0 | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2019-8331</summary> ### Vulnerable Library - <b>bootstrap-3.3.4.js</b></p> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.4/js/bootstrap.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.4/js/bootstrap.js</a></p> <p>Path to vulnerable library: /scorecard/content/js/bootstrap.js</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.4.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/BenchmarkJava-333/commit/3c6dcf8b9e2f8272e0ae17dea63a48ce3ac0245b">3c6dcf8b9e2f8272e0ae17dea63a48ce3ac0245b</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> In Bootstrap before 3.4.1 and 4.3.x before 4.3.1, XSS is possible in the tooltip or popover data-template attribute. <p>Publish Date: 2019-02-20 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-8331>CVE-2019-8331</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2019-02-20</p> <p>Fix Resolution: bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2018-14040</summary> ### Vulnerable Library - <b>bootstrap-3.3.4.js</b></p> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.4/js/bootstrap.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.4/js/bootstrap.js</a></p> <p>Path to vulnerable library: /scorecard/content/js/bootstrap.js</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.4.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/BenchmarkJava-333/commit/3c6dcf8b9e2f8272e0ae17dea63a48ce3ac0245b">3c6dcf8b9e2f8272e0ae17dea63a48ce3ac0245b</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> In Bootstrap before 4.1.2, XSS is possible in the collapse data-parent attribute. <p>Publish Date: 2018-07-13 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14040>CVE-2018-14040</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2018-07-13</p> <p>Fix Resolution: org.webjars.npm:bootstrap:4.1.2,org.webjars:bootstrap:3.4.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2018-20677</summary> ### Vulnerable Library - <b>bootstrap-3.3.4.js</b></p> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.4/js/bootstrap.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.4/js/bootstrap.js</a></p> <p>Path to vulnerable library: /scorecard/content/js/bootstrap.js</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.4.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/BenchmarkJava-333/commit/3c6dcf8b9e2f8272e0ae17dea63a48ce3ac0245b">3c6dcf8b9e2f8272e0ae17dea63a48ce3ac0245b</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> In Bootstrap before 3.4.0, XSS is possible in the affix configuration target property. <p>Publish Date: 2019-01-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20677>CVE-2018-20677</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20677">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20677</a></p> <p>Release Date: 2019-01-09</p> <p>Fix Resolution: Bootstrap - v3.4.0;NorDroN.AngularTemplate - 0.1.6;Dynamic.NET.Express.ProjectTemplates - 0.8.0;dotnetng.template - 1.0.0.4;ZNxtApp.Core.Module.Theme - 1.0.9-Beta;JMeter - 5.0.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2018-20676</summary> ### Vulnerable Library - <b>bootstrap-3.3.4.js</b></p> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.4/js/bootstrap.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.4/js/bootstrap.js</a></p> <p>Path to vulnerable library: /scorecard/content/js/bootstrap.js</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.4.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/BenchmarkJava-333/commit/3c6dcf8b9e2f8272e0ae17dea63a48ce3ac0245b">3c6dcf8b9e2f8272e0ae17dea63a48ce3ac0245b</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> In Bootstrap before 3.4.0, XSS is possible in the tooltip data-viewport attribute. <p>Publish Date: 2019-01-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20676>CVE-2018-20676</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20676">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20676</a></p> <p>Release Date: 2019-01-09</p> <p>Fix Resolution: bootstrap - 3.4.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2018-14042</summary> ### Vulnerable Library - <b>bootstrap-3.3.4.js</b></p> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.4/js/bootstrap.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.4/js/bootstrap.js</a></p> <p>Path to vulnerable library: /scorecard/content/js/bootstrap.js</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.4.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/BenchmarkJava-333/commit/3c6dcf8b9e2f8272e0ae17dea63a48ce3ac0245b">3c6dcf8b9e2f8272e0ae17dea63a48ce3ac0245b</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> In Bootstrap before 4.1.2, XSS is possible in the data-container property of tooltip. <p>Publish Date: 2018-07-13 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14042>CVE-2018-14042</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2018-07-13</p> <p>Fix Resolution: org.webjars.npm:bootstrap:4.1.2.org.webjars:bootstrap:3.4.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2016-10735</summary> ### Vulnerable Library - <b>bootstrap-3.3.4.js</b></p> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.4/js/bootstrap.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.4/js/bootstrap.js</a></p> <p>Path to vulnerable library: /scorecard/content/js/bootstrap.js</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.4.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/BenchmarkJava-333/commit/3c6dcf8b9e2f8272e0ae17dea63a48ce3ac0245b">3c6dcf8b9e2f8272e0ae17dea63a48ce3ac0245b</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> In Bootstrap 3.x before 3.4.0 and 4.x-beta before 4.0.0-beta.2, XSS is possible in the data-target attribute, a different vulnerability than CVE-2018-14041. <p>Publish Date: 2019-01-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10735>CVE-2016-10735</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2019-01-09</p> <p>Fix Resolution: 3.4.0</p> </p> <p></p> </details>
True
bootstrap-3.3.4.js: 6 vulnerabilities (highest severity is: 6.1) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.4.js</b></p></summary> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.4/js/bootstrap.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.4/js/bootstrap.js</a></p> <p>Path to vulnerable library: /scorecard/content/js/bootstrap.js</p> <p> <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/BenchmarkJava-333/commit/3c6dcf8b9e2f8272e0ae17dea63a48ce3ac0245b">3c6dcf8b9e2f8272e0ae17dea63a48ce3ac0245b</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2019-8331](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-8331) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-3.3.4.js | Direct | bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1 | &#10060; | | [CVE-2018-14040](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14040) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-3.3.4.js | Direct | org.webjars.npm:bootstrap:4.1.2,org.webjars:bootstrap:3.4.0 | &#10060; | | [CVE-2018-20677](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20677) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-3.3.4.js | Direct | Bootstrap - v3.4.0;NorDroN.AngularTemplate - 0.1.6;Dynamic.NET.Express.ProjectTemplates - 0.8.0;dotnetng.template - 1.0.0.4;ZNxtApp.Core.Module.Theme - 1.0.9-Beta;JMeter - 5.0.0 | &#10060; | | [CVE-2018-20676](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20676) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-3.3.4.js | Direct | bootstrap - 3.4.0 | &#10060; | | [CVE-2018-14042](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14042) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-3.3.4.js | Direct | org.webjars.npm:bootstrap:4.1.2.org.webjars:bootstrap:3.4.0 | &#10060; | | [CVE-2016-10735](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10735) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | bootstrap-3.3.4.js | Direct | 3.4.0 | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2019-8331</summary> ### Vulnerable Library - <b>bootstrap-3.3.4.js</b></p> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.4/js/bootstrap.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.4/js/bootstrap.js</a></p> <p>Path to vulnerable library: /scorecard/content/js/bootstrap.js</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.4.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/BenchmarkJava-333/commit/3c6dcf8b9e2f8272e0ae17dea63a48ce3ac0245b">3c6dcf8b9e2f8272e0ae17dea63a48ce3ac0245b</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> In Bootstrap before 3.4.1 and 4.3.x before 4.3.1, XSS is possible in the tooltip or popover data-template attribute. <p>Publish Date: 2019-02-20 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-8331>CVE-2019-8331</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2019-02-20</p> <p>Fix Resolution: bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2018-14040</summary> ### Vulnerable Library - <b>bootstrap-3.3.4.js</b></p> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.4/js/bootstrap.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.4/js/bootstrap.js</a></p> <p>Path to vulnerable library: /scorecard/content/js/bootstrap.js</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.4.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/BenchmarkJava-333/commit/3c6dcf8b9e2f8272e0ae17dea63a48ce3ac0245b">3c6dcf8b9e2f8272e0ae17dea63a48ce3ac0245b</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> In Bootstrap before 4.1.2, XSS is possible in the collapse data-parent attribute. <p>Publish Date: 2018-07-13 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14040>CVE-2018-14040</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2018-07-13</p> <p>Fix Resolution: org.webjars.npm:bootstrap:4.1.2,org.webjars:bootstrap:3.4.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2018-20677</summary> ### Vulnerable Library - <b>bootstrap-3.3.4.js</b></p> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.4/js/bootstrap.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.4/js/bootstrap.js</a></p> <p>Path to vulnerable library: /scorecard/content/js/bootstrap.js</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.4.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/BenchmarkJava-333/commit/3c6dcf8b9e2f8272e0ae17dea63a48ce3ac0245b">3c6dcf8b9e2f8272e0ae17dea63a48ce3ac0245b</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> In Bootstrap before 3.4.0, XSS is possible in the affix configuration target property. <p>Publish Date: 2019-01-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20677>CVE-2018-20677</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20677">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20677</a></p> <p>Release Date: 2019-01-09</p> <p>Fix Resolution: Bootstrap - v3.4.0;NorDroN.AngularTemplate - 0.1.6;Dynamic.NET.Express.ProjectTemplates - 0.8.0;dotnetng.template - 1.0.0.4;ZNxtApp.Core.Module.Theme - 1.0.9-Beta;JMeter - 5.0.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2018-20676</summary> ### Vulnerable Library - <b>bootstrap-3.3.4.js</b></p> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.4/js/bootstrap.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.4/js/bootstrap.js</a></p> <p>Path to vulnerable library: /scorecard/content/js/bootstrap.js</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.4.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/BenchmarkJava-333/commit/3c6dcf8b9e2f8272e0ae17dea63a48ce3ac0245b">3c6dcf8b9e2f8272e0ae17dea63a48ce3ac0245b</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> In Bootstrap before 3.4.0, XSS is possible in the tooltip data-viewport attribute. <p>Publish Date: 2019-01-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20676>CVE-2018-20676</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20676">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20676</a></p> <p>Release Date: 2019-01-09</p> <p>Fix Resolution: bootstrap - 3.4.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2018-14042</summary> ### Vulnerable Library - <b>bootstrap-3.3.4.js</b></p> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.4/js/bootstrap.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.4/js/bootstrap.js</a></p> <p>Path to vulnerable library: /scorecard/content/js/bootstrap.js</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.4.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/BenchmarkJava-333/commit/3c6dcf8b9e2f8272e0ae17dea63a48ce3ac0245b">3c6dcf8b9e2f8272e0ae17dea63a48ce3ac0245b</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> In Bootstrap before 4.1.2, XSS is possible in the data-container property of tooltip. <p>Publish Date: 2018-07-13 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14042>CVE-2018-14042</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2018-07-13</p> <p>Fix Resolution: org.webjars.npm:bootstrap:4.1.2.org.webjars:bootstrap:3.4.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2016-10735</summary> ### Vulnerable Library - <b>bootstrap-3.3.4.js</b></p> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.4/js/bootstrap.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.4/js/bootstrap.js</a></p> <p>Path to vulnerable library: /scorecard/content/js/bootstrap.js</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.4.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/BenchmarkJava-333/commit/3c6dcf8b9e2f8272e0ae17dea63a48ce3ac0245b">3c6dcf8b9e2f8272e0ae17dea63a48ce3ac0245b</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> In Bootstrap 3.x before 3.4.0 and 4.x-beta before 4.0.0-beta.2, XSS is possible in the data-target attribute, a different vulnerability than CVE-2018-14041. <p>Publish Date: 2019-01-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10735>CVE-2016-10735</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2019-01-09</p> <p>Fix Resolution: 3.4.0</p> </p> <p></p> </details>
non_process
bootstrap js vulnerabilities highest severity is vulnerable library bootstrap js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to vulnerable library scorecard content js bootstrap js found in head commit a href vulnerabilities cve severity cvss dependency type fixed in remediation available medium bootstrap js direct bootstrap bootstrap sass medium bootstrap js direct org webjars npm bootstrap org webjars bootstrap medium bootstrap js direct bootstrap nordron angulartemplate dynamic net express projecttemplates dotnetng template znxtapp core module theme beta jmeter medium bootstrap js direct bootstrap medium bootstrap js direct org webjars npm bootstrap org webjars bootstrap medium bootstrap js direct details cve vulnerable library bootstrap js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to vulnerable library scorecard content js bootstrap js dependency hierarchy x bootstrap js vulnerable library found in head commit a href found in base branch master vulnerability details in bootstrap before and x before xss is possible in the tooltip or popover data template attribute publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version release date fix resolution bootstrap bootstrap sass cve vulnerable library bootstrap js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to vulnerable library scorecard content js bootstrap js dependency hierarchy x bootstrap js vulnerable library found in head commit a href found in base branch master vulnerability details in bootstrap before xss is possible in the collapse data parent attribute publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version release date fix resolution org webjars npm bootstrap org webjars bootstrap cve vulnerable library bootstrap js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to vulnerable library scorecard content js bootstrap js dependency hierarchy x bootstrap js vulnerable library found in head commit a href found in base branch master vulnerability details in bootstrap before xss is possible in the affix configuration target property publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution bootstrap nordron angulartemplate dynamic net express projecttemplates dotnetng template znxtapp core module theme beta jmeter cve vulnerable library bootstrap js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to vulnerable library scorecard content js bootstrap js dependency hierarchy x bootstrap js vulnerable library found in head commit a href found in base branch master vulnerability details in bootstrap before xss is possible in the tooltip data viewport attribute publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution bootstrap cve vulnerable library bootstrap js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to vulnerable library scorecard content js bootstrap js dependency hierarchy x bootstrap js vulnerable library found in head commit a href found in base branch master vulnerability details in bootstrap before xss is possible in the data container property of tooltip publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version release date fix resolution org webjars npm bootstrap org webjars bootstrap cve vulnerable library bootstrap js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to vulnerable library scorecard content js bootstrap js dependency hierarchy x bootstrap js vulnerable library found in head commit a href found in base branch master vulnerability details in bootstrap x before and x beta before beta xss is possible in the data target attribute a different vulnerability than cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version release date fix resolution
0
7,769
10,889,738,104
IssuesEvent
2019-11-18 18:54:07
ORNL-AMO/AMO-Tools-Desktop
https://api.github.com/repos/ORNL-AMO/AMO-Tools-Desktop
closed
O2 enrichment metric bug
Calculator Process Heating bug
Using the example: Imperial - ![image.png](https://images.zenhubusercontent.com/5cd48a2af8cffa5a19122d27/05a6e968-c653-42de-b7b1-bf09b6362de6) Metric - ![image.png](https://images.zenhubusercontent.com/5cd48a2af8cffa5a19122d27/9f6669ce-655d-4abe-a7f6-61a5e41a93f8) Not sure what the screwy unit conversion is but something is causing fuel consumption to be way to high yet available heat and efficiency are okay
1.0
O2 enrichment metric bug - Using the example: Imperial - ![image.png](https://images.zenhubusercontent.com/5cd48a2af8cffa5a19122d27/05a6e968-c653-42de-b7b1-bf09b6362de6) Metric - ![image.png](https://images.zenhubusercontent.com/5cd48a2af8cffa5a19122d27/9f6669ce-655d-4abe-a7f6-61a5e41a93f8) Not sure what the screwy unit conversion is but something is causing fuel consumption to be way to high yet available heat and efficiency are okay
process
enrichment metric bug using the example imperial metric not sure what the screwy unit conversion is but something is causing fuel consumption to be way to high yet available heat and efficiency are okay
1
18,481
24,550,742,761
IssuesEvent
2022-10-12 12:25:29
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[iOS] Unable to enroll to the study
Bug Blocker P0 iOS Android Process: Fixed Process: Tested dev
Participants are not able to enroll to the studies > Getting continuous loading after the consent flow Note: Issue observed in the latest build 3.0(49) App ID: GCPMOB001
2.0
[iOS] Unable to enroll to the study - Participants are not able to enroll to the studies > Getting continuous loading after the consent flow Note: Issue observed in the latest build 3.0(49) App ID: GCPMOB001
process
unable to enroll to the study participants are not able to enroll to the studies getting continuous loading after the consent flow note issue observed in the latest build app id
1
614,846
19,190,796,829
IssuesEvent
2021-12-05 23:35:50
ondryaso/kachna-online
https://api.github.com/repos/ondryaso/kachna-online
opened
Accept externally provided KIS refresh tokens to log in
enhancement frontend low-priority
### Proposal Add a frontend route that accepts a KIS refresh token and uses it to log in. Use this _frontend endpoint_ to enable users to open the app directly from KIS Administration or Operator without having to log in again.
1.0
Accept externally provided KIS refresh tokens to log in - ### Proposal Add a frontend route that accepts a KIS refresh token and uses it to log in. Use this _frontend endpoint_ to enable users to open the app directly from KIS Administration or Operator without having to log in again.
non_process
accept externally provided kis refresh tokens to log in proposal add a frontend route that accepts a kis refresh token and uses it to log in use this frontend endpoint to enable users to open the app directly from kis administration or operator without having to log in again
0
47,469
7,328,358,002
IssuesEvent
2018-03-04 19:58:15
liballeg/allegro5
https://api.github.com/repos/liballeg/allegro5
opened
Display options inconsistently set on platforms
Core Library Documentation
This affects pretty much every platform, and unfortunately likely needs to be fixed in the platform specific code. The idea is that you should get meaningful values for `al_get_display_option`, even if that option does nothing for that platform. See https://github.com/liballeg/allegro5/issues/887 for an example issue. Related this is also documenting what all those display options default to.
1.0
Display options inconsistently set on platforms - This affects pretty much every platform, and unfortunately likely needs to be fixed in the platform specific code. The idea is that you should get meaningful values for `al_get_display_option`, even if that option does nothing for that platform. See https://github.com/liballeg/allegro5/issues/887 for an example issue. Related this is also documenting what all those display options default to.
non_process
display options inconsistently set on platforms this affects pretty much every platform and unfortunately likely needs to be fixed in the platform specific code the idea is that you should get meaningful values for al get display option even if that option does nothing for that platform see for an example issue related this is also documenting what all those display options default to
0
288,052
31,856,946,666
IssuesEvent
2023-09-15 08:10:30
nidhi7598/linux-4.19.72_CVE-2022-3564
https://api.github.com/repos/nidhi7598/linux-4.19.72_CVE-2022-3564
closed
CVE-2020-29370 (High) detected in linuxlinux-4.19.294 - autoclosed
Mend: dependency security vulnerability
## CVE-2020-29370 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.294</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-4.19.72_CVE-2022-3564/commit/454c7dacf6fa9a6de86d4067f5a08f25cffa519b">454c7dacf6fa9a6de86d4067f5a08f25cffa519b</a></p> <p>Found in base branch: <b>main</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/mm/slub.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/mm/slub.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in kmem_cache_alloc_bulk in mm/slub.c in the Linux kernel before 5.5.11. The slowpath lacks the required TID increment, aka CID-fd4d9c7d0c71. <p>Publish Date: 2020-11-28 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-29370>CVE-2020-29370</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.0</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-29370">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-29370</a></p> <p>Release Date: 2020-11-28</p> <p>Fix Resolution: v5.6-rc7,v5.5.11,v5.4.28</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-29370 (High) detected in linuxlinux-4.19.294 - autoclosed - ## CVE-2020-29370 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.294</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-4.19.72_CVE-2022-3564/commit/454c7dacf6fa9a6de86d4067f5a08f25cffa519b">454c7dacf6fa9a6de86d4067f5a08f25cffa519b</a></p> <p>Found in base branch: <b>main</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/mm/slub.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/mm/slub.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in kmem_cache_alloc_bulk in mm/slub.c in the Linux kernel before 5.5.11. The slowpath lacks the required TID increment, aka CID-fd4d9c7d0c71. <p>Publish Date: 2020-11-28 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-29370>CVE-2020-29370</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.0</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-29370">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-29370</a></p> <p>Release Date: 2020-11-28</p> <p>Fix Resolution: v5.6-rc7,v5.5.11,v5.4.28</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in linuxlinux autoclosed cve high severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch main vulnerable source files mm slub c mm slub c vulnerability details an issue was discovered in kmem cache alloc bulk in mm slub c in the linux kernel before the slowpath lacks the required tid increment aka cid publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
21,411
29,351,206,386
IssuesEvent
2023-05-27 00:34:48
devssa/onde-codar-em-salvador
https://api.github.com/repos/devssa/onde-codar-em-salvador
closed
[Remoto] C#/Xamarin Developer (Sênior) na Coodesh
SALVADOR BACK-END PJ XAMARIN SCRUM SENIOR AGILE STARTUP MOBILE REQUISITOS IOS REMOTO ANDROID PROCESSOS GITHUB INGLÊS UMA C Stale
## Descrição da vaga: Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios. Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/vagas/desenvolvedor-cxamarin-senior-195750387?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋 <p>A <strong>Jornada Quest</strong> está em busca de <strong><ins>C#/Xamarin Developer (Sênior)</ins></strong> para compor seu time!</p> <p>Empresa de outsourcing que atua com projetos internacionais está em busca de um membro sênior para o seu time. Você integrará um squad responsável pelo desenvolvimento de tecnologia de empresa canadense. Se você é conectado com a linguagem C# e deseja focar num projeto único e com potencial de conhecer novas culturas e oportunidades então essa vaga é sua.&nbsp;</p> <p><strong>Seus desafios e contribuições:</strong></p> <ul> <li>Se você deseja aprender novas tecnologias e desenvolver novos módulos e componentes, com potencial de crescimento coletivo, então vem compor essa squad. Com direção horizontal, você terá acesso a colegas, líder técnico e contato direto com CTO e CEO das empresas para criação de tecnologia opensource.&nbsp;</li> </ul> <p><strong>Conquiste seu lugar com a gente!</strong></p> <p>Aqui acreditamos na igualdade de oportunidades. Independentemente de gênero, idade, raça, cor, credo, orientação sexual e deficiência, todos os candidatos qualificados serão considerados em nossos processos.</p> <p><span style="font-size: 16px;"><strong>Para cruzar a sua jornada com a nossa, você precisa seguir essa trilha:</strong></span></p> <p><span style="font-size: 16px;">Etapa Gamificada, sim nós temos uma sessão de jogo; [35 min] - em português;</span></p> <p><span style="font-size: 16px;">Entrevista com gestor da área [40 min] - em português;</span></p> <p><span style="font-size: 16px;">Entrevista com o Cliente (canadense) - [25 min] - em inglês;</span></p> <p><span style="font-size: 16px;">Entrevista com CEO/Diretoria [40 minutos] - em inglês</span></p> <p><span style="font-size: 16px;">Entrevista com o Cliente (canadense) - [25 min] - em inglês.</span></p> ## Jornada Quest: <p>Somos a Jornada Quest, uma startup com recurso próprio que acredita que os melhores não são recrutáveis, são atraídos!</p> <p>Nossos esforços estão em alinhar o ‘timing’ do seu projeto profissional com o da empresa que contrata, para juntos garantir uma trilha de desenvolvimento e crescimento coletivos.</p><a href='https://coodesh.com/empresas/jornada-quest'>Veja mais no site</a> ## Habilidades: - C# - Inglês - Xamarin - Agile ## Local: 100% Remoto ## Requisitos: - Experiência com metodologia ágil (scrum); - Comunicação e objetividade na exposição de ideias; - Experiência sólida com C#; - Experiência em XAML; - Inglês fluente; - Conhecimento de arquitetura e habilidades de desenvolvimento móvel; - Comprometimento com entregas e trabalhar colaborativamente; - Ser aberto a feedback. ## Diferenciais: - Experiência com desenvolvimento mobile (nativo ou multiplataforma) iOS e/ou Android; - Graduação. ## Benefícios: - 15 dias férias remuneradas. ## Como se candidatar: Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [C#/Xamarin Developer (Sênior) na Jornada Quest](https://coodesh.com/vagas/desenvolvedor-cxamarin-senior-195750387?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação. ## Labels #### Alocação Remoto #### Regime PJ #### Categoria Back-End
1.0
[Remoto] C#/Xamarin Developer (Sênior) na Coodesh - ## Descrição da vaga: Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios. Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/vagas/desenvolvedor-cxamarin-senior-195750387?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋 <p>A <strong>Jornada Quest</strong> está em busca de <strong><ins>C#/Xamarin Developer (Sênior)</ins></strong> para compor seu time!</p> <p>Empresa de outsourcing que atua com projetos internacionais está em busca de um membro sênior para o seu time. Você integrará um squad responsável pelo desenvolvimento de tecnologia de empresa canadense. Se você é conectado com a linguagem C# e deseja focar num projeto único e com potencial de conhecer novas culturas e oportunidades então essa vaga é sua.&nbsp;</p> <p><strong>Seus desafios e contribuições:</strong></p> <ul> <li>Se você deseja aprender novas tecnologias e desenvolver novos módulos e componentes, com potencial de crescimento coletivo, então vem compor essa squad. Com direção horizontal, você terá acesso a colegas, líder técnico e contato direto com CTO e CEO das empresas para criação de tecnologia opensource.&nbsp;</li> </ul> <p><strong>Conquiste seu lugar com a gente!</strong></p> <p>Aqui acreditamos na igualdade de oportunidades. Independentemente de gênero, idade, raça, cor, credo, orientação sexual e deficiência, todos os candidatos qualificados serão considerados em nossos processos.</p> <p><span style="font-size: 16px;"><strong>Para cruzar a sua jornada com a nossa, você precisa seguir essa trilha:</strong></span></p> <p><span style="font-size: 16px;">Etapa Gamificada, sim nós temos uma sessão de jogo; [35 min] - em português;</span></p> <p><span style="font-size: 16px;">Entrevista com gestor da área [40 min] - em português;</span></p> <p><span style="font-size: 16px;">Entrevista com o Cliente (canadense) - [25 min] - em inglês;</span></p> <p><span style="font-size: 16px;">Entrevista com CEO/Diretoria [40 minutos] - em inglês</span></p> <p><span style="font-size: 16px;">Entrevista com o Cliente (canadense) - [25 min] - em inglês.</span></p> ## Jornada Quest: <p>Somos a Jornada Quest, uma startup com recurso próprio que acredita que os melhores não são recrutáveis, são atraídos!</p> <p>Nossos esforços estão em alinhar o ‘timing’ do seu projeto profissional com o da empresa que contrata, para juntos garantir uma trilha de desenvolvimento e crescimento coletivos.</p><a href='https://coodesh.com/empresas/jornada-quest'>Veja mais no site</a> ## Habilidades: - C# - Inglês - Xamarin - Agile ## Local: 100% Remoto ## Requisitos: - Experiência com metodologia ágil (scrum); - Comunicação e objetividade na exposição de ideias; - Experiência sólida com C#; - Experiência em XAML; - Inglês fluente; - Conhecimento de arquitetura e habilidades de desenvolvimento móvel; - Comprometimento com entregas e trabalhar colaborativamente; - Ser aberto a feedback. ## Diferenciais: - Experiência com desenvolvimento mobile (nativo ou multiplataforma) iOS e/ou Android; - Graduação. ## Benefícios: - 15 dias férias remuneradas. ## Como se candidatar: Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [C#/Xamarin Developer (Sênior) na Jornada Quest](https://coodesh.com/vagas/desenvolvedor-cxamarin-senior-195750387?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação. ## Labels #### Alocação Remoto #### Regime PJ #### Categoria Back-End
process
c xamarin developer sênior na coodesh descrição da vaga esta é uma vaga de um parceiro da plataforma coodesh ao candidatar se você terá acesso as informações completas sobre a empresa e benefícios fique atento ao redirecionamento que vai te levar para uma url com o pop up personalizado de candidatura 👋 a jornada quest está em busca de c xamarin developer sênior para compor seu time empresa de outsourcing que atua com projetos internacionais está em busca de um membro sênior para o seu time você integrará um squad responsável pelo desenvolvimento de tecnologia de empresa canadense se você é conectado com a linguagem c e deseja focar num projeto único e com potencial de conhecer novas culturas e oportunidades então essa vaga é sua nbsp seus desafios e contribuições se você deseja aprender novas tecnologias e desenvolver novos módulos e componentes com potencial de crescimento coletivo então vem compor essa squad com direção horizontal você terá acesso a colegas líder técnico e contato direto com cto e ceo das empresas para criação de tecnologia opensource nbsp conquiste seu lugar com a gente aqui acreditamos na igualdade de oportunidades independentemente de gênero idade raça cor credo orientação sexual e deficiência todos os candidatos qualificados serão considerados em nossos processos para cruzar a sua jornada com a nossa você precisa seguir essa trilha etapa gamificada sim nós temos uma sessão de jogo em português entrevista com gestor da área em português entrevista com o cliente canadense em inglês entrevista com ceo diretoria em inglês entrevista com o cliente canadense em inglês jornada quest somos a jornada quest uma startup com recurso próprio que acredita que os melhores não são recrutáveis são atraídos nossos esforços estão em alinhar o ‘timing’ do seu projeto profissional com o da empresa que contrata para juntos garantir uma trilha de desenvolvimento e crescimento coletivos habilidades c inglês xamarin agile local remoto requisitos experiência com metodologia ágil scrum comunicação e objetividade na exposição de ideias experiência sólida com c experiência em xaml inglês fluente conhecimento de arquitetura e habilidades de desenvolvimento móvel comprometimento com entregas e trabalhar colaborativamente ser aberto a feedback diferenciais experiência com desenvolvimento mobile nativo ou multiplataforma ios e ou android graduação benefícios dias férias remuneradas como se candidatar candidatar se exclusivamente através da plataforma coodesh no link a seguir após candidatar se via plataforma coodesh e validar o seu login você poderá acompanhar e receber todas as interações do processo por lá utilize a opção pedir feedback entre uma etapa e outra na vaga que se candidatou isso fará com que a pessoa recruiter responsável pelo processo na empresa receba a notificação labels alocação remoto regime pj categoria back end
1
24,710
5,098,043,512
IssuesEvent
2017-01-03 23:41:54
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
closed
hack/verify-gendocs.sh is slow
kind/documentation priority/backlog team/ux (deprecated - do not use)
The slowest part of every commit is waiting for hooks. I added some printouts to find the culprit - see $SUBJECT. Can we not run this test on every commit and rebase? If we switched to using pre-commit instead of prepare-commit-msg we could force-override it (--no-verify) if needed. @lavalamp
1.0
hack/verify-gendocs.sh is slow - The slowest part of every commit is waiting for hooks. I added some printouts to find the culprit - see $SUBJECT. Can we not run this test on every commit and rebase? If we switched to using pre-commit instead of prepare-commit-msg we could force-override it (--no-verify) if needed. @lavalamp
non_process
hack verify gendocs sh is slow the slowest part of every commit is waiting for hooks i added some printouts to find the culprit see subject can we not run this test on every commit and rebase if we switched to using pre commit instead of prepare commit msg we could force override it no verify if needed lavalamp
0
407,702
11,936,030,788
IssuesEvent
2020-04-02 09:36:42
project-koku/koku
https://api.github.com/repos/project-koku/koku
closed
Sources: editing Azure billing_source can result in failure
bug priority - medium
**Describe the bug** Updating a single billing-source field (resource-group or storage-account) results in validation failure. The error message says that the field which was not updated is missing. **To Reproduce** Steps to reproduce the behavior: 1. Create Azure source 2. Edit the source. Change resource-group. 3. Click Save. 4. See attached screenshot error **Expected behavior** Editing a single billing-source field is successful. **Screenshots** <img width="695" alt="Screen Shot 2020-03-18 at 4 22 49 PM" src="https://user-images.githubusercontent.com/19751919/77004352-534d9d00-6935-11ea-8d9f-4948ade277e9.png">
1.0
Sources: editing Azure billing_source can result in failure - **Describe the bug** Updating a single billing-source field (resource-group or storage-account) results in validation failure. The error message says that the field which was not updated is missing. **To Reproduce** Steps to reproduce the behavior: 1. Create Azure source 2. Edit the source. Change resource-group. 3. Click Save. 4. See attached screenshot error **Expected behavior** Editing a single billing-source field is successful. **Screenshots** <img width="695" alt="Screen Shot 2020-03-18 at 4 22 49 PM" src="https://user-images.githubusercontent.com/19751919/77004352-534d9d00-6935-11ea-8d9f-4948ade277e9.png">
non_process
sources editing azure billing source can result in failure describe the bug updating a single billing source field resource group or storage account results in validation failure the error message says that the field which was not updated is missing to reproduce steps to reproduce the behavior create azure source edit the source change resource group click save see attached screenshot error expected behavior editing a single billing source field is successful screenshots img width alt screen shot at pm src
0
7,010
10,151,609,760
IssuesEvent
2019-08-05 20:47:50
jtablesaw/tablesaw
https://api.github.com/repos/jtablesaw/tablesaw
closed
Code format standards
process
Having thought more on the google format option, I would propose this: - We switch over to using the google format tool, including going back to 2 spaces. Staying with 4 spaces would mean people having to configure the tool in various ways depending on whether or not they're using a plugin. - We publish it as the standard way to format code - We don't force occasional submitters to use it. We can run on the source as a whole for an occasional cleanup. It's fine to recommend it, though. - We discontinue comments on formatting in code reviews, unless it's something that's outside of the google format yet is still important (but I can't imagine what that might be). thoughts?
1.0
Code format standards - Having thought more on the google format option, I would propose this: - We switch over to using the google format tool, including going back to 2 spaces. Staying with 4 spaces would mean people having to configure the tool in various ways depending on whether or not they're using a plugin. - We publish it as the standard way to format code - We don't force occasional submitters to use it. We can run on the source as a whole for an occasional cleanup. It's fine to recommend it, though. - We discontinue comments on formatting in code reviews, unless it's something that's outside of the google format yet is still important (but I can't imagine what that might be). thoughts?
process
code format standards having thought more on the google format option i would propose this we switch over to using the google format tool including going back to spaces staying with spaces would mean people having to configure the tool in various ways depending on whether or not they re using a plugin we publish it as the standard way to format code we don t force occasional submitters to use it we can run on the source as a whole for an occasional cleanup it s fine to recommend it though we discontinue comments on formatting in code reviews unless it s something that s outside of the google format yet is still important but i can t imagine what that might be thoughts
1
11,736
13,808,060,476
IssuesEvent
2020-10-12 01:03:32
storybookjs/storybook
https://api.github.com/repos/storybookjs/storybook
closed
"babel-plugin-react-docgen" breaks Storybook v6 builds when using the Flow spread operator
bug compatibility with other tools has workaround inactive
**Observed behavior** Storybook does not build successfully when it parses components using the Flow spread operator. **Expected behavior** I expect Storybook to build successfully. **Details** After upgrading to Storybook v6 and installing the Create React App preset, Storybook fails to build due to an error in `babel-loader`. Since there were no problems in previous Storybook releases, I think it is safe to assume this is due to new additions in Storybook v6, specifically the addition of `react-docgen`. The specifics of the error message are detailed further below, but here is a snapshot of the component and LOC that is raising the issue. Note this component makes heavy use of [Flow](https://flow.org/) features, in particular the spread operator. **Example Component** ```js type Props = {| ...ComponentProps, ...ExtractReturn<typeof mapStateToProps>, ...typeof mapDispatchToProps, |} class IntroOfferMapObject extends React.PureComponent<Props> { ``` **Resulting Error Trace** ``` yarn run storybook Module build failed (from ./node_modules/babel-loader/lib/index.js): TypeError: /MyUser/MyProject/src/components/Map/IntroOfferMapObject.js: Property elements[1] of ArrayExpression expected node to be of a type ["null","Expression","SpreadElement"] but instead got undefined at validate (/MyUser/MyProject/node_modules/@babel/types/lib/definitions/utils.js:148:11) at validator (/MyUser/MyProject/node_modules/@babel/types/lib/definitions/utils.js:103:7) at Object.validate (/MyUser/MyProject/node_modules/@babel/types/lib/definitions/utils.js:229:7) at validateField (/MyUser/MyProject/node_modules/@babel/types/lib/validators/validate.js:24:9) at validate (/MyUser/MyProject/node_modules/@babel/types/lib/validators/validate.js:17:3) at builder (/MyUser/MyProject/node_modules/@babel/types/lib/builders/builder.js:38:27) at Object.arrayExpression (/MyUser/MyProject/node_modules/@babel/types/lib/builders/generated/index.js:254:31) at buildObjectExpression (/MyUser/MyProject/node_modules/babel-plugin-react-docgen/lib/index.js:146:14) at buildObjectExpression (/MyUser/MyProject/node_modules/babel-plugin-react-docgen/lib/index.js:131:60) at /MyUser/MyProject/node_modules/babel-plugin-react-docgen/lib/index.js:95:19 at Array.forEach (<anonymous>) at injectReactDocgenInfo (/MyUser/MyProject/node_modules/babel-plugin-react-docgen/lib/index.js:87:17) at PluginPass.exit (/MyUser/MyProject/node_modules/babel-plugin-react-docgen/lib/index.js:44:11) at newFn (/MyUser/MyProject/node_modules/@babel/traverse/lib/visitors.js:175:21) at NodePath._call (/MyUser/MyProject/node_modules/@babel/traverse/lib/path/context.js:55:20) at NodePath.call (/MyUser/MyProject/node_modules/@babel/traverse/lib/path/context.js:42:17) ``` **System:** ``` Environment Info: System: OS: macOS Mojave 10.14.6 CPU: (12) x64 Intel(R) Core(TM) i7-8850H CPU @ 2.60GHz Binaries: Node: 12.9.0 - /usr/local/bin/node Yarn: 1.21.1 - ~/.yarn/bin/yarn npm: 6.10.2 - /usr/local/bin/npm Browsers: Chrome: 84.0.4147.125 Safari: 13.1.2 npmPackages: @storybook/addon-actions: ^6.0.5 => 6.0.5 @storybook/addon-knobs: ^6.0.5 => 6.0.5 @storybook/addon-links: ^6.0.5 => 6.0.5 @storybook/addon-viewport: ^6.0.5 => 6.0.5 @storybook/addons: ^6.0.5 => 6.0.5 @storybook/preset-create-react-app: ^3.1.4 => 3.1.4 @storybook/react: ^6.0.5 => 6.0.5 ``` **Additional context** This build issue can be resolved by removing the `react-docgen` plugin manually from the webpack config. Observe the following `.storybook/main.js` file, which removes the `react-docgen` override from the Storybook webpack configuration: ```js module.exports = { webpackFinal: config => { // Manually remove the react-docgen plugin that breaks w/ the // Flow spread operator. config.module.rules[3].oneOf[1].options.overrides.pop() return config }, addons: ['@storybook/preset-create-react-app'], } ``` By removing `config.module.rules[3].oneOf[1].options.overrides` (the `react-docgen` override), this issue is resolved. I got this workaround from the following issue: https://github.com/storybookjs/storybook/issues/4873#issuecomment-458497220. There are a number of other threads that report this same issue, but they have all been closed with no proper resolution.
True
"babel-plugin-react-docgen" breaks Storybook v6 builds when using the Flow spread operator - **Observed behavior** Storybook does not build successfully when it parses components using the Flow spread operator. **Expected behavior** I expect Storybook to build successfully. **Details** After upgrading to Storybook v6 and installing the Create React App preset, Storybook fails to build due to an error in `babel-loader`. Since there were no problems in previous Storybook releases, I think it is safe to assume this is due to new additions in Storybook v6, specifically the addition of `react-docgen`. The specifics of the error message are detailed further below, but here is a snapshot of the component and LOC that is raising the issue. Note this component makes heavy use of [Flow](https://flow.org/) features, in particular the spread operator. **Example Component** ```js type Props = {| ...ComponentProps, ...ExtractReturn<typeof mapStateToProps>, ...typeof mapDispatchToProps, |} class IntroOfferMapObject extends React.PureComponent<Props> { ``` **Resulting Error Trace** ``` yarn run storybook Module build failed (from ./node_modules/babel-loader/lib/index.js): TypeError: /MyUser/MyProject/src/components/Map/IntroOfferMapObject.js: Property elements[1] of ArrayExpression expected node to be of a type ["null","Expression","SpreadElement"] but instead got undefined at validate (/MyUser/MyProject/node_modules/@babel/types/lib/definitions/utils.js:148:11) at validator (/MyUser/MyProject/node_modules/@babel/types/lib/definitions/utils.js:103:7) at Object.validate (/MyUser/MyProject/node_modules/@babel/types/lib/definitions/utils.js:229:7) at validateField (/MyUser/MyProject/node_modules/@babel/types/lib/validators/validate.js:24:9) at validate (/MyUser/MyProject/node_modules/@babel/types/lib/validators/validate.js:17:3) at builder (/MyUser/MyProject/node_modules/@babel/types/lib/builders/builder.js:38:27) at Object.arrayExpression (/MyUser/MyProject/node_modules/@babel/types/lib/builders/generated/index.js:254:31) at buildObjectExpression (/MyUser/MyProject/node_modules/babel-plugin-react-docgen/lib/index.js:146:14) at buildObjectExpression (/MyUser/MyProject/node_modules/babel-plugin-react-docgen/lib/index.js:131:60) at /MyUser/MyProject/node_modules/babel-plugin-react-docgen/lib/index.js:95:19 at Array.forEach (<anonymous>) at injectReactDocgenInfo (/MyUser/MyProject/node_modules/babel-plugin-react-docgen/lib/index.js:87:17) at PluginPass.exit (/MyUser/MyProject/node_modules/babel-plugin-react-docgen/lib/index.js:44:11) at newFn (/MyUser/MyProject/node_modules/@babel/traverse/lib/visitors.js:175:21) at NodePath._call (/MyUser/MyProject/node_modules/@babel/traverse/lib/path/context.js:55:20) at NodePath.call (/MyUser/MyProject/node_modules/@babel/traverse/lib/path/context.js:42:17) ``` **System:** ``` Environment Info: System: OS: macOS Mojave 10.14.6 CPU: (12) x64 Intel(R) Core(TM) i7-8850H CPU @ 2.60GHz Binaries: Node: 12.9.0 - /usr/local/bin/node Yarn: 1.21.1 - ~/.yarn/bin/yarn npm: 6.10.2 - /usr/local/bin/npm Browsers: Chrome: 84.0.4147.125 Safari: 13.1.2 npmPackages: @storybook/addon-actions: ^6.0.5 => 6.0.5 @storybook/addon-knobs: ^6.0.5 => 6.0.5 @storybook/addon-links: ^6.0.5 => 6.0.5 @storybook/addon-viewport: ^6.0.5 => 6.0.5 @storybook/addons: ^6.0.5 => 6.0.5 @storybook/preset-create-react-app: ^3.1.4 => 3.1.4 @storybook/react: ^6.0.5 => 6.0.5 ``` **Additional context** This build issue can be resolved by removing the `react-docgen` plugin manually from the webpack config. Observe the following `.storybook/main.js` file, which removes the `react-docgen` override from the Storybook webpack configuration: ```js module.exports = { webpackFinal: config => { // Manually remove the react-docgen plugin that breaks w/ the // Flow spread operator. config.module.rules[3].oneOf[1].options.overrides.pop() return config }, addons: ['@storybook/preset-create-react-app'], } ``` By removing `config.module.rules[3].oneOf[1].options.overrides` (the `react-docgen` override), this issue is resolved. I got this workaround from the following issue: https://github.com/storybookjs/storybook/issues/4873#issuecomment-458497220. There are a number of other threads that report this same issue, but they have all been closed with no proper resolution.
non_process
babel plugin react docgen breaks storybook builds when using the flow spread operator observed behavior storybook does not build successfully when it parses components using the flow spread operator expected behavior i expect storybook to build successfully details after upgrading to storybook and installing the create react app preset storybook fails to build due to an error in babel loader since there were no problems in previous storybook releases i think it is safe to assume this is due to new additions in storybook specifically the addition of react docgen the specifics of the error message are detailed further below but here is a snapshot of the component and loc that is raising the issue note this component makes heavy use of features in particular the spread operator example component js type props componentprops extractreturn typeof mapdispatchtoprops class introoffermapobject extends react purecomponent resulting error trace yarn run storybook module build failed from node modules babel loader lib index js typeerror myuser myproject src components map introoffermapobject js property elements of arrayexpression expected node to be of a type but instead got undefined at validate myuser myproject node modules babel types lib definitions utils js at validator myuser myproject node modules babel types lib definitions utils js at object validate myuser myproject node modules babel types lib definitions utils js at validatefield myuser myproject node modules babel types lib validators validate js at validate myuser myproject node modules babel types lib validators validate js at builder myuser myproject node modules babel types lib builders builder js at object arrayexpression myuser myproject node modules babel types lib builders generated index js at buildobjectexpression myuser myproject node modules babel plugin react docgen lib index js at buildobjectexpression myuser myproject node modules babel plugin react docgen lib index js at myuser myproject node modules babel plugin react docgen lib index js at array foreach at injectreactdocgeninfo myuser myproject node modules babel plugin react docgen lib index js at pluginpass exit myuser myproject node modules babel plugin react docgen lib index js at newfn myuser myproject node modules babel traverse lib visitors js at nodepath call myuser myproject node modules babel traverse lib path context js at nodepath call myuser myproject node modules babel traverse lib path context js system environment info system os macos mojave cpu intel r core tm cpu binaries node usr local bin node yarn yarn bin yarn npm usr local bin npm browsers chrome safari npmpackages storybook addon actions storybook addon knobs storybook addon links storybook addon viewport storybook addons storybook preset create react app storybook react additional context this build issue can be resolved by removing the react docgen plugin manually from the webpack config observe the following storybook main js file which removes the react docgen override from the storybook webpack configuration js module exports webpackfinal config manually remove the react docgen plugin that breaks w the flow spread operator config module rules oneof options overrides pop return config addons by removing config module rules oneof options overrides the react docgen override this issue is resolved i got this workaround from the following issue there are a number of other threads that report this same issue but they have all been closed with no proper resolution
0
155,604
5,957,542,699
IssuesEvent
2017-05-29 02:53:48
wordpress-mobile/AztecEditor-Android
https://api.github.com/repos/wordpress-mobile/AztecEditor-Android
closed
Image-loading scrolls the editor to the cursor
bug high priority
When an image is being uploaded and a user scrolls away from the cursor, the editor automatically scrolls back to the cursor on every image-loading progress bar update.
1.0
Image-loading scrolls the editor to the cursor - When an image is being uploaded and a user scrolls away from the cursor, the editor automatically scrolls back to the cursor on every image-loading progress bar update.
non_process
image loading scrolls the editor to the cursor when an image is being uploaded and a user scrolls away from the cursor the editor automatically scrolls back to the cursor on every image loading progress bar update
0
211,675
23,835,724,781
IssuesEvent
2022-09-06 05:34:46
ioana-nicolae/keycloak
https://api.github.com/repos/ioana-nicolae/keycloak
opened
CVE-2022-38749 (Medium) detected in multiple libraries
security vulnerability
## CVE-2022-38749 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>snakeyaml-1.27.jar</b>, <b>snakeyaml-1.19.jar</b>, <b>snakeyaml-1.17.jar</b>, <b>snakeyaml-1.14.jar</b></p></summary> <p> <details><summary><b>snakeyaml-1.27.jar</b></p></summary> <p>YAML 1.1 parser and emitter for Java</p> <p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p> <p>Path to dependency file: /quarkus/runtime/pom.xml</p> <p>Path to vulnerable library: /2/repository/org/yaml/snakeyaml/1.27/snakeyaml-1.27.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.27/snakeyaml-1.27.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.27/snakeyaml-1.27.jar</p> <p> Dependency Hierarchy: - :x: **snakeyaml-1.27.jar** (Vulnerable Library) </details> <details><summary><b>snakeyaml-1.19.jar</b></p></summary> <p>YAML 1.1 parser and emitter for Java</p> <p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p> <p>Path to dependency file: /misc/spring-boot-starter/keycloak-spring-boot-starter/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.19/snakeyaml-1.19.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.19/snakeyaml-1.19.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.0.5.RELEASE.jar (Root Library) - spring-boot-starter-2.0.5.RELEASE.jar - :x: **snakeyaml-1.19.jar** (Vulnerable Library) </details> <details><summary><b>snakeyaml-1.17.jar</b></p></summary> <p>YAML 1.1 parser and emitter for Java</p> <p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p> <p>Path to dependency file: /adapters/oidc/spring-boot-adapter-core/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.17/snakeyaml-1.17.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.17/snakeyaml-1.17.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.17/snakeyaml-1.17.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.17/snakeyaml-1.17.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-1.5.16.RELEASE.jar (Root Library) - spring-boot-starter-1.5.16.RELEASE.jar - :x: **snakeyaml-1.17.jar** (Vulnerable Library) </details> <details><summary><b>snakeyaml-1.14.jar</b></p></summary> <p>YAML 1.1 parser and emitter for Java</p> <p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p> <p>Path to dependency file: /testsuite/integration-arquillian/servers/auth-server/undertow/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.14/snakeyaml-1.14.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.14/snakeyaml-1.14.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.14/snakeyaml-1.14.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.14/snakeyaml-1.14.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.14/snakeyaml-1.14.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.14/snakeyaml-1.14.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.14/snakeyaml-1.14.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.14/snakeyaml-1.14.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.14/snakeyaml-1.14.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.14/snakeyaml-1.14.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.14/snakeyaml-1.14.jar</p> <p> Dependency Hierarchy: - integration-arquillian-testsuite-providers-13.0.0-SNAPSHOT.jar (Root Library) - keycloak-dependencies-server-all-13.0.0-SNAPSHOT.pom - openshift-restclient-java-8.0.0.Final.jar - :x: **snakeyaml-1.14.jar** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/ioana-nicolae/keycloak/commit/083dc472750fe560cb65e68c8ff4c9e6d9246bb3">083dc472750fe560cb65e68c8ff4c9e6d9246bb3</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Using snakeYAML to parse untrusted YAML files may be vulnerable to Denial of Service attacks (DOS). If the parser is running on user supplied input, an attacker may supply content that causes the parser to crash by stackoverflow. <p>Publish Date: 2022-09-05 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-38749>CVE-2022-38749</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bitbucket.org/snakeyaml/snakeyaml/issues/525/got-stackoverflowerror-for-many-open">https://bitbucket.org/snakeyaml/snakeyaml/issues/525/got-stackoverflowerror-for-many-open</a></p> <p>Release Date: 2022-09-05</p> <p>Fix Resolution: org.yaml:snakeyaml:1.31</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue
True
CVE-2022-38749 (Medium) detected in multiple libraries - ## CVE-2022-38749 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>snakeyaml-1.27.jar</b>, <b>snakeyaml-1.19.jar</b>, <b>snakeyaml-1.17.jar</b>, <b>snakeyaml-1.14.jar</b></p></summary> <p> <details><summary><b>snakeyaml-1.27.jar</b></p></summary> <p>YAML 1.1 parser and emitter for Java</p> <p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p> <p>Path to dependency file: /quarkus/runtime/pom.xml</p> <p>Path to vulnerable library: /2/repository/org/yaml/snakeyaml/1.27/snakeyaml-1.27.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.27/snakeyaml-1.27.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.27/snakeyaml-1.27.jar</p> <p> Dependency Hierarchy: - :x: **snakeyaml-1.27.jar** (Vulnerable Library) </details> <details><summary><b>snakeyaml-1.19.jar</b></p></summary> <p>YAML 1.1 parser and emitter for Java</p> <p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p> <p>Path to dependency file: /misc/spring-boot-starter/keycloak-spring-boot-starter/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.19/snakeyaml-1.19.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.19/snakeyaml-1.19.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.0.5.RELEASE.jar (Root Library) - spring-boot-starter-2.0.5.RELEASE.jar - :x: **snakeyaml-1.19.jar** (Vulnerable Library) </details> <details><summary><b>snakeyaml-1.17.jar</b></p></summary> <p>YAML 1.1 parser and emitter for Java</p> <p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p> <p>Path to dependency file: /adapters/oidc/spring-boot-adapter-core/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.17/snakeyaml-1.17.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.17/snakeyaml-1.17.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.17/snakeyaml-1.17.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.17/snakeyaml-1.17.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-1.5.16.RELEASE.jar (Root Library) - spring-boot-starter-1.5.16.RELEASE.jar - :x: **snakeyaml-1.17.jar** (Vulnerable Library) </details> <details><summary><b>snakeyaml-1.14.jar</b></p></summary> <p>YAML 1.1 parser and emitter for Java</p> <p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p> <p>Path to dependency file: /testsuite/integration-arquillian/servers/auth-server/undertow/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.14/snakeyaml-1.14.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.14/snakeyaml-1.14.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.14/snakeyaml-1.14.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.14/snakeyaml-1.14.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.14/snakeyaml-1.14.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.14/snakeyaml-1.14.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.14/snakeyaml-1.14.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.14/snakeyaml-1.14.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.14/snakeyaml-1.14.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.14/snakeyaml-1.14.jar,/home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.14/snakeyaml-1.14.jar</p> <p> Dependency Hierarchy: - integration-arquillian-testsuite-providers-13.0.0-SNAPSHOT.jar (Root Library) - keycloak-dependencies-server-all-13.0.0-SNAPSHOT.pom - openshift-restclient-java-8.0.0.Final.jar - :x: **snakeyaml-1.14.jar** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/ioana-nicolae/keycloak/commit/083dc472750fe560cb65e68c8ff4c9e6d9246bb3">083dc472750fe560cb65e68c8ff4c9e6d9246bb3</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Using snakeYAML to parse untrusted YAML files may be vulnerable to Denial of Service attacks (DOS). If the parser is running on user supplied input, an attacker may supply content that causes the parser to crash by stackoverflow. <p>Publish Date: 2022-09-05 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-38749>CVE-2022-38749</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bitbucket.org/snakeyaml/snakeyaml/issues/525/got-stackoverflowerror-for-many-open">https://bitbucket.org/snakeyaml/snakeyaml/issues/525/got-stackoverflowerror-for-many-open</a></p> <p>Release Date: 2022-09-05</p> <p>Fix Resolution: org.yaml:snakeyaml:1.31</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue
non_process
cve medium detected in multiple libraries cve medium severity vulnerability vulnerable libraries snakeyaml jar snakeyaml jar snakeyaml jar snakeyaml jar snakeyaml jar yaml parser and emitter for java library home page a href path to dependency file quarkus runtime pom xml path to vulnerable library repository org yaml snakeyaml snakeyaml jar home wss scanner repository org yaml snakeyaml snakeyaml jar home wss scanner repository org yaml snakeyaml snakeyaml jar dependency hierarchy x snakeyaml jar vulnerable library snakeyaml jar yaml parser and emitter for java library home page a href path to dependency file misc spring boot starter keycloak spring boot starter pom xml path to vulnerable library home wss scanner repository org yaml snakeyaml snakeyaml jar home wss scanner repository org yaml snakeyaml snakeyaml jar dependency hierarchy spring boot starter web release jar root library spring boot starter release jar x snakeyaml jar vulnerable library snakeyaml jar yaml parser and emitter for java library home page a href path to dependency file adapters oidc spring boot adapter core pom xml path to vulnerable library home wss scanner repository org yaml snakeyaml snakeyaml jar home wss scanner repository org yaml snakeyaml snakeyaml jar home wss scanner repository org yaml snakeyaml snakeyaml jar home wss scanner repository org yaml snakeyaml snakeyaml jar dependency hierarchy spring boot starter web release jar root library spring boot starter release jar x snakeyaml jar vulnerable library snakeyaml jar yaml parser and emitter for java library home page a href path to dependency file testsuite integration arquillian servers auth server undertow pom xml path to vulnerable library home wss scanner repository org yaml snakeyaml snakeyaml jar home wss scanner repository org yaml snakeyaml snakeyaml jar home wss scanner repository org yaml snakeyaml snakeyaml jar home wss scanner repository org yaml snakeyaml snakeyaml jar home wss scanner repository org yaml snakeyaml snakeyaml jar home wss scanner repository org yaml snakeyaml snakeyaml jar home wss scanner repository org yaml snakeyaml snakeyaml jar home wss scanner repository org yaml snakeyaml snakeyaml jar home wss scanner repository org yaml snakeyaml snakeyaml jar home wss scanner repository org yaml snakeyaml snakeyaml jar home wss scanner repository org yaml snakeyaml snakeyaml jar dependency hierarchy integration arquillian testsuite providers snapshot jar root library keycloak dependencies server all snapshot pom openshift restclient java final jar x snakeyaml jar vulnerable library found in head commit a href found in base branch master vulnerability details using snakeyaml to parse untrusted yaml files may be vulnerable to denial of service attacks dos if the parser is running on user supplied input an attacker may supply content that causes the parser to crash by stackoverflow publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org yaml snakeyaml rescue worker helmet automatic remediation is available for this issue
0
42,878
11,349,483,947
IssuesEvent
2020-01-24 05:09:06
idaholab/raven
https://api.github.com/repos/idaholab/raven
closed
If illegal XML named variables are present, the code crashes
defect devel priority_critical
-------- Issue Description -------- ##### What did you expect to see happen? Ability to load CSV files with invalid XML variables (e.g. containing "/", etc) ##### What did you see instead? If a variable has an invalid XML character (e.g. "/", "$", etc.), the XML utilis methods (used to write out the metadata) is going to replace it with a "." when constructing the XML metadata (node <dim>). Since the new variables do not match any of the output variables, those variables are not written in the auxiliary metadata XML file. This causes the code to crash when trying to reload the CSVs. When the variables are read back, the code does not find the Index of the "illegal" variables (not written in the XML) defined in the <Output> block. For example, When we dump the following variables we get this warning: ``` XML UTILS: Replacing illegal tag characters in "u-234_atoms/barn-cm": u-234_atoms.barn-cm XML UTILS: Replacing illegal tag characters in "ag-107_atoms/barn-cm": ag-107_atoms.barn-cm XML UTILS: Replacing illegal tag characters in "u-235_atoms/barn-cm": u-235_atoms.barn-cm ``` When they are read back, the illegal named variables (e.g. ag-107_atoms/barn-cm) are not found in the XML and consequentially are not stored in the dict ```self._pivotParams```. This leaves variables in the output space without any index linked to them and consequentially, when the method ```asDataset``` gets called, an unexpected crash happens: ``` ( 43.79 sec) HistorySet : DEBUG -> Reading data from "/Users/alfoa/projects/raven_github/raven/dh_chenj/samples/originalModelCSV/history_time_evolution_2017.csv.csv" ( 43.81 sec) HistorySet : DEBUG -> Reading data from "/Users/alfoa/projects/raven_github/raven/dh_chenj/samples/originalModelCSV/history_time_evolution_2018.csv.csv" ( 43.83 sec) HistorySet : DEBUG -> Reading data from "/Users/alfoa/projects/raven_github/raven/dh_chenj/samples/originalModelCSV/history_time_evolution_2019.csv.csv" Traceback (most recent call last): File "/Users/alfoa/projects/raven_github/raven/framework/Driver.py", line 281, in <module> raven() File "/Users/alfoa/projects/raven_github/raven/framework/Driver.py", line 234, in raven simulation.run() File "/Users/alfoa/projects/raven_github/raven/framework/Simulation.py", line 798, in run stepInstance.takeAstep(stepInputDict) File "/Users/alfoa/projects/raven_github/raven/framework/Steps.py", line 312, in takeAstep self._localTakeAstepRun(inDictionary) File "/Users/alfoa/projects/raven_github/raven/framework/Steps.py", line 947, in _localTakeAstepRun outputs[i].load(inDictionary['Input'][i].getPath(),'csv',**options) File "/Users/alfoa/projects/raven_github/raven/framework/DataObjects/DataSet.py", line 433, in load self._fromCSV(fileName,**kwargs) File "/Users/alfoa/projects/raven_github/raven/framework/DataObjects/HistorySet.py", line 154, in _fromCSV self.load(data,style='dict',dims=self.getDimensions()) File "/Users/alfoa/projects/raven_github/raven/framework/DataObjects/DataSet.py", line 435, in load self._fromDict(fileName,**kwargs) File "/Users/alfoa/projects/raven_github/raven/framework/DataObjects/DataSet.py", line 1257, in _fromDict self.asDataset() File "/Users/alfoa/projects/raven_github/raven/framework/DataObjects/DataSet.py", line 260, in asDataset return self._convertToXrDataset() File "/Users/alfoa/projects/raven_github/raven/framework/DataObjects/DataSet.py", line 1079, in _convertToXrDataset arrays[var][self.sampleTag] += firstSample KeyError: u'u-234_atoms/barn-cm' ``` ##### Do you have a suggested fix for the development team? Either the illegal XML character should be replaced with a "token" recognizable in the loading stage => we will able to identify it and connect them back with the original variables or we move from : ``` <dims> <totals_watts>time</totals_watts> </dims> ``` to ``` <dims> <var name="totals_watts">time</var> <var name="ag-107_atoms/barn-cm">time</var> </dims> ``` or change the DataObject.py method ```setPivotParameter``` to: ``` def setPivotParams(self,params): """ Sets the pivot parameters for variables. @ In, params, dict, var:[params] as str:list(str) @ Out, None """ # TODO typechecking, assertions coords = set().union(*params.values()) for coord in coords: if coord not in self._pivotParams: self._pivotParams[coord] = list(var for var in params.keys() if coord in params[var]) else: self._pivotParams[coord] = list(set(list(var for var in params.keys() if coord in params[var]) + self._pivotParams[coord])) ``` ##### Please attach the input file(s) that generate this error. The simpler the input, the faster we can find the issue. ---------------- For Change Control Board: Issue Review ---------------- This review should occur before any development is performed as a response to this issue. - [x] 1. Is it tagged with a type: defect or improvement? - [x] 2. Is it tagged with a priority: critical, normal or minor? - [x] 3. If it will impact requirements or requirements tests, is it tagged with requirements? - [x] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users. - [x] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.) ------- For Change Control Board: Issue Closure ------- This review should occur when the issue is imminently going to be closed. - [ ] 1. If the issue is a defect, is the defect fixed? - [ ] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.) - [ ] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)? - [ ] 4. If the issue is a defect, does it impact the latest stable branch? If yes, is there any issue tagged with stable (create if needed)? - [x] 5. If the issue is being closed without a merge request, has an explanation of why it is being closed been provided?
1.0
If illegal XML named variables are present, the code crashes - -------- Issue Description -------- ##### What did you expect to see happen? Ability to load CSV files with invalid XML variables (e.g. containing "/", etc) ##### What did you see instead? If a variable has an invalid XML character (e.g. "/", "$", etc.), the XML utilis methods (used to write out the metadata) is going to replace it with a "." when constructing the XML metadata (node <dim>). Since the new variables do not match any of the output variables, those variables are not written in the auxiliary metadata XML file. This causes the code to crash when trying to reload the CSVs. When the variables are read back, the code does not find the Index of the "illegal" variables (not written in the XML) defined in the <Output> block. For example, When we dump the following variables we get this warning: ``` XML UTILS: Replacing illegal tag characters in "u-234_atoms/barn-cm": u-234_atoms.barn-cm XML UTILS: Replacing illegal tag characters in "ag-107_atoms/barn-cm": ag-107_atoms.barn-cm XML UTILS: Replacing illegal tag characters in "u-235_atoms/barn-cm": u-235_atoms.barn-cm ``` When they are read back, the illegal named variables (e.g. ag-107_atoms/barn-cm) are not found in the XML and consequentially are not stored in the dict ```self._pivotParams```. This leaves variables in the output space without any index linked to them and consequentially, when the method ```asDataset``` gets called, an unexpected crash happens: ``` ( 43.79 sec) HistorySet : DEBUG -> Reading data from "/Users/alfoa/projects/raven_github/raven/dh_chenj/samples/originalModelCSV/history_time_evolution_2017.csv.csv" ( 43.81 sec) HistorySet : DEBUG -> Reading data from "/Users/alfoa/projects/raven_github/raven/dh_chenj/samples/originalModelCSV/history_time_evolution_2018.csv.csv" ( 43.83 sec) HistorySet : DEBUG -> Reading data from "/Users/alfoa/projects/raven_github/raven/dh_chenj/samples/originalModelCSV/history_time_evolution_2019.csv.csv" Traceback (most recent call last): File "/Users/alfoa/projects/raven_github/raven/framework/Driver.py", line 281, in <module> raven() File "/Users/alfoa/projects/raven_github/raven/framework/Driver.py", line 234, in raven simulation.run() File "/Users/alfoa/projects/raven_github/raven/framework/Simulation.py", line 798, in run stepInstance.takeAstep(stepInputDict) File "/Users/alfoa/projects/raven_github/raven/framework/Steps.py", line 312, in takeAstep self._localTakeAstepRun(inDictionary) File "/Users/alfoa/projects/raven_github/raven/framework/Steps.py", line 947, in _localTakeAstepRun outputs[i].load(inDictionary['Input'][i].getPath(),'csv',**options) File "/Users/alfoa/projects/raven_github/raven/framework/DataObjects/DataSet.py", line 433, in load self._fromCSV(fileName,**kwargs) File "/Users/alfoa/projects/raven_github/raven/framework/DataObjects/HistorySet.py", line 154, in _fromCSV self.load(data,style='dict',dims=self.getDimensions()) File "/Users/alfoa/projects/raven_github/raven/framework/DataObjects/DataSet.py", line 435, in load self._fromDict(fileName,**kwargs) File "/Users/alfoa/projects/raven_github/raven/framework/DataObjects/DataSet.py", line 1257, in _fromDict self.asDataset() File "/Users/alfoa/projects/raven_github/raven/framework/DataObjects/DataSet.py", line 260, in asDataset return self._convertToXrDataset() File "/Users/alfoa/projects/raven_github/raven/framework/DataObjects/DataSet.py", line 1079, in _convertToXrDataset arrays[var][self.sampleTag] += firstSample KeyError: u'u-234_atoms/barn-cm' ``` ##### Do you have a suggested fix for the development team? Either the illegal XML character should be replaced with a "token" recognizable in the loading stage => we will able to identify it and connect them back with the original variables or we move from : ``` <dims> <totals_watts>time</totals_watts> </dims> ``` to ``` <dims> <var name="totals_watts">time</var> <var name="ag-107_atoms/barn-cm">time</var> </dims> ``` or change the DataObject.py method ```setPivotParameter``` to: ``` def setPivotParams(self,params): """ Sets the pivot parameters for variables. @ In, params, dict, var:[params] as str:list(str) @ Out, None """ # TODO typechecking, assertions coords = set().union(*params.values()) for coord in coords: if coord not in self._pivotParams: self._pivotParams[coord] = list(var for var in params.keys() if coord in params[var]) else: self._pivotParams[coord] = list(set(list(var for var in params.keys() if coord in params[var]) + self._pivotParams[coord])) ``` ##### Please attach the input file(s) that generate this error. The simpler the input, the faster we can find the issue. ---------------- For Change Control Board: Issue Review ---------------- This review should occur before any development is performed as a response to this issue. - [x] 1. Is it tagged with a type: defect or improvement? - [x] 2. Is it tagged with a priority: critical, normal or minor? - [x] 3. If it will impact requirements or requirements tests, is it tagged with requirements? - [x] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users. - [x] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.) ------- For Change Control Board: Issue Closure ------- This review should occur when the issue is imminently going to be closed. - [ ] 1. If the issue is a defect, is the defect fixed? - [ ] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.) - [ ] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)? - [ ] 4. If the issue is a defect, does it impact the latest stable branch? If yes, is there any issue tagged with stable (create if needed)? - [x] 5. If the issue is being closed without a merge request, has an explanation of why it is being closed been provided?
non_process
if illegal xml named variables are present the code crashes issue description what did you expect to see happen ability to load csv files with invalid xml variables e g containing etc what did you see instead if a variable has an invalid xml character e g etc the xml utilis methods used to write out the metadata is going to replace it with a when constructing the xml metadata node since the new variables do not match any of the output variables those variables are not written in the auxiliary metadata xml file this causes the code to crash when trying to reload the csvs when the variables are read back the code does not find the index of the illegal variables not written in the xml defined in the block for example when we dump the following variables we get this warning xml utils replacing illegal tag characters in u atoms barn cm u atoms barn cm xml utils replacing illegal tag characters in ag atoms barn cm ag atoms barn cm xml utils replacing illegal tag characters in u atoms barn cm u atoms barn cm when they are read back the illegal named variables e g ag atoms barn cm are not found in the xml and consequentially are not stored in the dict self pivotparams this leaves variables in the output space without any index linked to them and consequentially when the method asdataset gets called an unexpected crash happens sec historyset debug reading data from users alfoa projects raven github raven dh chenj samples originalmodelcsv history time evolution csv csv sec historyset debug reading data from users alfoa projects raven github raven dh chenj samples originalmodelcsv history time evolution csv csv sec historyset debug reading data from users alfoa projects raven github raven dh chenj samples originalmodelcsv history time evolution csv csv traceback most recent call last file users alfoa projects raven github raven framework driver py line in raven file users alfoa projects raven github raven framework driver py line in raven simulation run file users alfoa projects raven github raven framework simulation py line in run stepinstance takeastep stepinputdict file users alfoa projects raven github raven framework steps py line in takeastep self localtakeasteprun indictionary file users alfoa projects raven github raven framework steps py line in localtakeasteprun outputs load indictionary getpath csv options file users alfoa projects raven github raven framework dataobjects dataset py line in load self fromcsv filename kwargs file users alfoa projects raven github raven framework dataobjects historyset py line in fromcsv self load data style dict dims self getdimensions file users alfoa projects raven github raven framework dataobjects dataset py line in load self fromdict filename kwargs file users alfoa projects raven github raven framework dataobjects dataset py line in fromdict self asdataset file users alfoa projects raven github raven framework dataobjects dataset py line in asdataset return self converttoxrdataset file users alfoa projects raven github raven framework dataobjects dataset py line in converttoxrdataset arrays firstsample keyerror u u atoms barn cm do you have a suggested fix for the development team either the illegal xml character should be replaced with a token recognizable in the loading stage we will able to identify it and connect them back with the original variables or we move from time to time time or change the dataobject py method setpivotparameter to def setpivotparams self params sets the pivot parameters for variables in params dict var as str list str out none todo typechecking assertions coords set union params values for coord in coords if coord not in self pivotparams self pivotparams list var for var in params keys if coord in params else self pivotparams list set list var for var in params keys if coord in params self pivotparams please attach the input file s that generate this error the simpler the input the faster we can find the issue for change control board issue review this review should occur before any development is performed as a response to this issue is it tagged with a type defect or improvement is it tagged with a priority critical normal or minor if it will impact requirements or requirements tests is it tagged with requirements if it is a defect can it cause wrong results for users if so an email needs to be sent to the users is a rationale provided such as explaining why the improvement is needed or why current code is wrong for change control board issue closure this review should occur when the issue is imminently going to be closed if the issue is a defect is the defect fixed if the issue is a defect is the defect tested for in the regression test system if not explain why not if the issue can impact users has an email to the users group been written the email should specify if the defect impacts stable or master if the issue is a defect does it impact the latest stable branch if yes is there any issue tagged with stable create if needed if the issue is being closed without a merge request has an explanation of why it is being closed been provided
0
349,744
10,472,662,125
IssuesEvent
2019-09-23 10:44:25
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
opened
[master-preview] Hammer filled type glitch
Low Priority QA
Steps: 1) Take material 2) Take iron hammer 3) Take point fill and some form 4) Build 5) Press ctrl 6) Build It should still build with point, but after Ctrl was pressed fill type became rectangle without any changes in the UI (Ctrl is for Crouch in the Controls). https://drive.google.com/file/d/1HFtc-phduveU39zmcDi4UjG-ChFjsJ7j/view?usp=sharing
1.0
[master-preview] Hammer filled type glitch - Steps: 1) Take material 2) Take iron hammer 3) Take point fill and some form 4) Build 5) Press ctrl 6) Build It should still build with point, but after Ctrl was pressed fill type became rectangle without any changes in the UI (Ctrl is for Crouch in the Controls). https://drive.google.com/file/d/1HFtc-phduveU39zmcDi4UjG-ChFjsJ7j/view?usp=sharing
non_process
hammer filled type glitch steps take material take iron hammer take point fill and some form build press ctrl build it should still build with point but after ctrl was pressed fill type became rectangle without any changes in the ui ctrl is for crouch in the controls
0
326,180
9,948,440,899
IssuesEvent
2019-07-04 08:56:18
kubermatic/machine-controller
https://api.github.com/repos/kubermatic/machine-controller
closed
Add support for OpenStack TENANT_ID
priority/important-soon team/lifecycle
In some environments the tenant name is not unique and will lead to errors. To avoid this, the tenant ID must be used. **Acceptance criteria**: - [ ] The machine-controller can create instances when using the tenant id instead of the name
1.0
Add support for OpenStack TENANT_ID - In some environments the tenant name is not unique and will lead to errors. To avoid this, the tenant ID must be used. **Acceptance criteria**: - [ ] The machine-controller can create instances when using the tenant id instead of the name
non_process
add support for openstack tenant id in some environments the tenant name is not unique and will lead to errors to avoid this the tenant id must be used acceptance criteria the machine controller can create instances when using the tenant id instead of the name
0
12,611
15,013,584,895
IssuesEvent
2021-02-01 04:49:29
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
Multiple record for same study in studyState endpoint
Bug P1 Process: Fixed
The studyState needs to send only the latest state of the study in the studySate endpoint. The front end application has no way to determine which among those should be used if multiple records are coming ![unnamed](https://user-images.githubusercontent.com/59356289/106239173-2857f900-6228-11eb-8e69-35420decfbce.png)
1.0
Multiple record for same study in studyState endpoint - The studyState needs to send only the latest state of the study in the studySate endpoint. The front end application has no way to determine which among those should be used if multiple records are coming ![unnamed](https://user-images.githubusercontent.com/59356289/106239173-2857f900-6228-11eb-8e69-35420decfbce.png)
process
multiple record for same study in studystate endpoint the studystate needs to send only the latest state of the study in the studysate endpoint the front end application has no way to determine which among those should be used if multiple records are coming
1
125,946
10,371,138,181
IssuesEvent
2019-09-08 18:03:49
Vachok/ftpplus
https://api.github.com/repos/Vachok/ftpplus
closed
testDelTemp - Fail [D106]
Medium TestQuality bug resolution_Cant Reproduce
Execute FileSystemWorkerTest::testDelTemp**testDelTemp** *FileSystemWorkerTest* *G:\My\_Proj\FtpClientPlus\modules\networker\DeleterTemp.txt did not expect to find [true] but found [false]* *java.lang.AssertionError*
1.0
testDelTemp - Fail [D106] - Execute FileSystemWorkerTest::testDelTemp**testDelTemp** *FileSystemWorkerTest* *G:\My\_Proj\FtpClientPlus\modules\networker\DeleterTemp.txt did not expect to find [true] but found [false]* *java.lang.AssertionError*
non_process
testdeltemp fail execute filesystemworkertest testdeltemp testdeltemp filesystemworkertest g my proj ftpclientplus modules networker deletertemp txt did not expect to find but found java lang assertionerror
0
701,713
24,105,308,171
IssuesEvent
2022-09-20 06:54:34
kubesphere/console
https://api.github.com/repos/kubesphere/console
closed
There is a delete button for project management permissions to view the details of grayscale publishing tasks
kind/bug kind/need-to-verify priority/low
**Describe the bug** 1、 User cc has platform-regular, enterprise space viewer role and project viewer role 2、Use the user to log in to the platform, enter the project details, click the grayscale release - "grayscale release task, click the existing task to view the information, and there is a delete button. ![image](https://user-images.githubusercontent.com/88183150/185064815-1927102f-200c-4a5a-90ac-ffaee56f2c5c.png) **Versions used(KubeSphere/Kubernetes)** KubeSphere: v3.3.1-rc.0 **Expected behavior** Delete button is unavailable or does not exist /kind bug /priority low
1.0
There is a delete button for project management permissions to view the details of grayscale publishing tasks - **Describe the bug** 1、 User cc has platform-regular, enterprise space viewer role and project viewer role 2、Use the user to log in to the platform, enter the project details, click the grayscale release - "grayscale release task, click the existing task to view the information, and there is a delete button. ![image](https://user-images.githubusercontent.com/88183150/185064815-1927102f-200c-4a5a-90ac-ffaee56f2c5c.png) **Versions used(KubeSphere/Kubernetes)** KubeSphere: v3.3.1-rc.0 **Expected behavior** Delete button is unavailable or does not exist /kind bug /priority low
non_process
there is a delete button for project management permissions to view the details of grayscale publishing tasks describe the bug 、 user cc has platform regular enterprise space viewer role and project viewer role 、use the user to log in to the platform enter the project details click the grayscale release grayscale release task click the existing task to view the information and there is a delete button versions used kubesphere kubernetes kubesphere rc expected behavior delete button is unavailable or does not exist kind bug priority low
0
79,834
15,286,988,473
IssuesEvent
2021-02-23 15:17:34
yakra/DataProcessing
https://api.github.com/repos/yakra/DataProcessing
opened
contains template
C++ code organization efficiency
`contains` could possibly return an iterator to the found object or to haystack::end. Would this slow down any of the instances where it's used? Where might this help?
1.0
contains template - `contains` could possibly return an iterator to the found object or to haystack::end. Would this slow down any of the instances where it's used? Where might this help?
non_process
contains template contains could possibly return an iterator to the found object or to haystack end would this slow down any of the instances where it s used where might this help
0
11,908
14,699,378,345
IssuesEvent
2021-01-04 08:27:14
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
App participant registry > Number of count is reducing in Studies column
Bug P1 Participant manager Process: Fixed Process: Tested dev
Steps 1. Join a open study 2. Withdrawn from the study ( Observe number of studies count in App participant registry) 3. For te same study fails the eligibility test 4. Click on Apps tab >Navigate to App participant registry and observe the number of count AR : Number of Studies count is reducing ER : Count should not be reduced
2.0
App participant registry > Number of count is reducing in Studies column - Steps 1. Join a open study 2. Withdrawn from the study ( Observe number of studies count in App participant registry) 3. For te same study fails the eligibility test 4. Click on Apps tab >Navigate to App participant registry and observe the number of count AR : Number of Studies count is reducing ER : Count should not be reduced
process
app participant registry number of count is reducing in studies column steps join a open study withdrawn from the study observe number of studies count in app participant registry for te same study fails the eligibility test click on apps tab navigate to app participant registry and observe the number of count ar number of studies count is reducing er count should not be reduced
1
11,539
14,427,290,493
IssuesEvent
2020-12-06 03:05:40
cggos/cggos.github.io
https://api.github.com/repos/cggos/cggos.github.io
opened
图像分析之高斯滤波 - Gavin Gao's Blog
Gitalk image-process-gauss-filter
https://cggos.github.io/image-process-gauss-filter.html [TOC]高斯函数一维高斯函数\[f(x) = A \cdot e^{-\frac{(x-\mu)^2}{2{\sigma}^2}}\]多维高斯函数\[f_{X}\left(x_{1}, x_{2}, \cdots, x_{k}\right) =A \cdot \exp \left(-\frac{1}{2}(X-...
1.0
图像分析之高斯滤波 - Gavin Gao's Blog - https://cggos.github.io/image-process-gauss-filter.html [TOC]高斯函数一维高斯函数\[f(x) = A \cdot e^{-\frac{(x-\mu)^2}{2{\sigma}^2}}\]多维高斯函数\[f_{X}\left(x_{1}, x_{2}, \cdots, x_{k}\right) =A \cdot \exp \left(-\frac{1}{2}(X-...
process
图像分析之高斯滤波 gavin gao s blog 高斯函数一维高斯函数 多维高斯函数 f x left x x cdots x k right a cdot exp left frac x
1
63,217
8,665,603,204
IssuesEvent
2018-11-29 00:03:51
fga-eps-mds/2018.2-Roles
https://api.github.com/repos/fga-eps-mds/2018.2-Roles
opened
Atualizar documentação de metodologia
Documentation Eps Mds
**Descrição** Deverá ser atualizado o documento de plano metodológico. ***Requisitos:*** Deverão ser revisados os erros ortográficos.
1.0
Atualizar documentação de metodologia - **Descrição** Deverá ser atualizado o documento de plano metodológico. ***Requisitos:*** Deverão ser revisados os erros ortográficos.
non_process
atualizar documentação de metodologia descrição deverá ser atualizado o documento de plano metodológico requisitos deverão ser revisados os erros ortográficos
0
20,306
26,947,013,727
IssuesEvent
2023-02-08 08:55:44
prisma/prisma-engines
https://api.github.com/repos/prisma/prisma-engines
opened
Fix Postgres introspection of partition tables false positives on inherited tables
process/candidate kind/improvement tech/engines/introspection engine team/schema
I kept the PG10+ SQL as if we were also using it for CockroachDb, which no longer is the case. I can now verify if the table is a partition table correctly, rather than rely on whether it uses inheritance. See https://github.com/prisma/prisma-engines/blob/main/libs/sql-schema-describer/src/postgres/tables_query.sql ; specifically I think we can simplify it to ```sql SELECT tbl.relname AS table_name, namespace.nspname as namespace, tbl.relispartition FROM pg_class AS tbl INNER JOIN pg_namespace AS namespace ON namespace.oid = tbl.relnamespace WHERE ( -- grab tables when -- it's an oRdinary table ('r') and is not a partition; (tbl.relkind = 'r' AND tbl.relispartition = 'f') OR -- when it's a partition tbl.relkind = 'p' AND tbl.relispartition = 't' ) AND namespace.nspname = ANY ( $1 ) ORDER BY namespace, table_name; ```
1.0
Fix Postgres introspection of partition tables false positives on inherited tables - I kept the PG10+ SQL as if we were also using it for CockroachDb, which no longer is the case. I can now verify if the table is a partition table correctly, rather than rely on whether it uses inheritance. See https://github.com/prisma/prisma-engines/blob/main/libs/sql-schema-describer/src/postgres/tables_query.sql ; specifically I think we can simplify it to ```sql SELECT tbl.relname AS table_name, namespace.nspname as namespace, tbl.relispartition FROM pg_class AS tbl INNER JOIN pg_namespace AS namespace ON namespace.oid = tbl.relnamespace WHERE ( -- grab tables when -- it's an oRdinary table ('r') and is not a partition; (tbl.relkind = 'r' AND tbl.relispartition = 'f') OR -- when it's a partition tbl.relkind = 'p' AND tbl.relispartition = 't' ) AND namespace.nspname = ANY ( $1 ) ORDER BY namespace, table_name; ```
process
fix postgres introspection of partition tables false positives on inherited tables i kept the sql as if we were also using it for cockroachdb which no longer is the case i can now verify if the table is a partition table correctly rather than rely on whether it uses inheritance see specifically i think we can simplify it to sql select tbl relname as table name namespace nspname as namespace tbl relispartition from pg class as tbl inner join pg namespace as namespace on namespace oid tbl relnamespace where grab tables when it s an ordinary table r and is not a partition tbl relkind r and tbl relispartition f or when it s a partition tbl relkind p and tbl relispartition t and namespace nspname any order by namespace table name
1
9,674
6,412,310,255
IssuesEvent
2017-08-08 02:38:40
FReBOmusic/FReBO
https://api.github.com/repos/FReBOmusic/FReBO
opened
Borrowing Listitem
Usability
In the event that the user navigates to the Borrowing Listview, on the Inventory Screen. **Expected Response**: The Borrowing Listview should be populated with records of the active borrowing transactions made by the user.
True
Borrowing Listitem - In the event that the user navigates to the Borrowing Listview, on the Inventory Screen. **Expected Response**: The Borrowing Listview should be populated with records of the active borrowing transactions made by the user.
non_process
borrowing listitem in the event that the user navigates to the borrowing listview on the inventory screen expected response the borrowing listview should be populated with records of the active borrowing transactions made by the user
0
176,009
14,547,987,763
IssuesEvent
2020-12-16 00:10:23
pi-top/pi-top-Python-SDK
https://api.github.com/repos/pi-top/pi-top-Python-SDK
closed
Reorganize example sections
documentation good first issue
- [x] Rename sections: use recipes instead of examples - [x] Move component basic examples to the respective class documentation - [ ] Add low-hanging fruit examples where they are obviously not there
1.0
Reorganize example sections - - [x] Rename sections: use recipes instead of examples - [x] Move component basic examples to the respective class documentation - [ ] Add low-hanging fruit examples where they are obviously not there
non_process
reorganize example sections rename sections use recipes instead of examples move component basic examples to the respective class documentation add low hanging fruit examples where they are obviously not there
0
69,259
17,613,102,378
IssuesEvent
2021-08-18 06:00:38
ballerina-platform/ballerina-release
https://api.github.com/repos/ballerina-platform/ballerina-release
opened
Create a GitHub pages website to show dependencies graph
Type/Task Team/BuildPipeline
**Description:** Currently dependency graph is represented in a dot file. We need to create a GitHub pages website to show dependencies graph
1.0
Create a GitHub pages website to show dependencies graph - **Description:** Currently dependency graph is represented in a dot file. We need to create a GitHub pages website to show dependencies graph
non_process
create a github pages website to show dependencies graph description currently dependency graph is represented in a dot file we need to create a github pages website to show dependencies graph
0
19,772
26,146,482,188
IssuesEvent
2022-12-30 06:08:58
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
[Mirror] jdk 19 archives
P2 type: process team-OSS mirror request
### Please list the URLs of the archives you'd like to mirror: https://cdn.azul.com/zulu/bin/zulu19.30.11-ca-jdk19.0.1-linux_aarch64.tar.gz https://cdn.azul.com/zulu/bin/zulu19.30.11-ca-jdk19.0.1-linux_x64.tar.gz https://cdn.azul.com/zulu/bin/zulu19.30.11-ca-jdk19.0.1-win_x64.zip https://cdn.azul.com/zulu/bin/zulu19.30.11-ca-jdk19.0.1-macosx_x64.tar.gz https://cdn.azul.com/zulu/bin/zulu19.30.11-ca-jdk19.0.1-macosx_aarch64.tar.gz
1.0
[Mirror] jdk 19 archives - ### Please list the URLs of the archives you'd like to mirror: https://cdn.azul.com/zulu/bin/zulu19.30.11-ca-jdk19.0.1-linux_aarch64.tar.gz https://cdn.azul.com/zulu/bin/zulu19.30.11-ca-jdk19.0.1-linux_x64.tar.gz https://cdn.azul.com/zulu/bin/zulu19.30.11-ca-jdk19.0.1-win_x64.zip https://cdn.azul.com/zulu/bin/zulu19.30.11-ca-jdk19.0.1-macosx_x64.tar.gz https://cdn.azul.com/zulu/bin/zulu19.30.11-ca-jdk19.0.1-macosx_aarch64.tar.gz
process
jdk archives please list the urls of the archives you d like to mirror
1
234,351
25,842,054,964
IssuesEvent
2022-12-13 01:39:36
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
System.Security.Cryptography.Xml 7.0.0 package doesn't work on .NET Framework
area-System.Security untriaged
### Description When compiling against System.Security.Cryptography.Xml 7.0.0 for netstandard2.0 and running on .NET Framework, you get a TypeLoadException. This is because the net462 version of the assembly in the package is missing the type fowarding attributes to the .NET Framework assembly System.Sercurity.dll. It doesn't even reference System.Sercurity.dll. ### Reproduction Steps Create a netstandard2,0 library that references the 7.0 version of the nuget package System.Security.Cryptography.Xml. Create a dummy class that references SignedXml in a static method (e.g. instantiates it). From a .NET Framework application, reference this library and call the static method. Compile and run. ### Expected behavior Able to reference and use SignedXml in a .NET Standard library referenced from a .NET Framework application. ### Actual behavior Throws the following exception: ``` System.TypeLoadException: Could not load type 'System.Security.Cryptography.Xml.SignedXml' from assembly 'System.Security.Cryptography.Xml, Version=7.0.0.0, Culture=neutral, PublicKeyToken=cc7b13ffcd2ddd51'. ``` ### Regression? Yes, this works with the latest 6.0.x version (currently 6.0.1). In that version, the .NET Framework lib is for net461 (7.0.0 is for net462), it references System.Security.dll and includes all the necessary type forwarding for the various types. ### Known Workarounds Reference the 6.0.1 version of the package. ### Configuration _No response_ ### Other information _No response_
True
System.Security.Cryptography.Xml 7.0.0 package doesn't work on .NET Framework - ### Description When compiling against System.Security.Cryptography.Xml 7.0.0 for netstandard2.0 and running on .NET Framework, you get a TypeLoadException. This is because the net462 version of the assembly in the package is missing the type fowarding attributes to the .NET Framework assembly System.Sercurity.dll. It doesn't even reference System.Sercurity.dll. ### Reproduction Steps Create a netstandard2,0 library that references the 7.0 version of the nuget package System.Security.Cryptography.Xml. Create a dummy class that references SignedXml in a static method (e.g. instantiates it). From a .NET Framework application, reference this library and call the static method. Compile and run. ### Expected behavior Able to reference and use SignedXml in a .NET Standard library referenced from a .NET Framework application. ### Actual behavior Throws the following exception: ``` System.TypeLoadException: Could not load type 'System.Security.Cryptography.Xml.SignedXml' from assembly 'System.Security.Cryptography.Xml, Version=7.0.0.0, Culture=neutral, PublicKeyToken=cc7b13ffcd2ddd51'. ``` ### Regression? Yes, this works with the latest 6.0.x version (currently 6.0.1). In that version, the .NET Framework lib is for net461 (7.0.0 is for net462), it references System.Security.dll and includes all the necessary type forwarding for the various types. ### Known Workarounds Reference the 6.0.1 version of the package. ### Configuration _No response_ ### Other information _No response_
non_process
system security cryptography xml package doesn t work on net framework description when compiling against system security cryptography xml for and running on net framework you get a typeloadexception this is because the version of the assembly in the package is missing the type fowarding attributes to the net framework assembly system sercurity dll it doesn t even reference system sercurity dll reproduction steps create a library that references the version of the nuget package system security cryptography xml create a dummy class that references signedxml in a static method e g instantiates it from a net framework application reference this library and call the static method compile and run expected behavior able to reference and use signedxml in a net standard library referenced from a net framework application actual behavior throws the following exception system typeloadexception could not load type system security cryptography xml signedxml from assembly system security cryptography xml version culture neutral publickeytoken regression yes this works with the latest x version currently in that version the net framework lib is for is for it references system security dll and includes all the necessary type forwarding for the various types known workarounds reference the version of the package configuration no response other information no response
0
41,423
5,356,124,311
IssuesEvent
2017-02-20 14:53:35
mozilla/fxa-content-server
https://api.github.com/repos/mozilla/fxa-content-server
closed
test - sign in with incorrect email case before normalization fix, on second attempt canonical form is used
tests waffle:review
The following test is unstable on CI: ``` × firefox on any platform - sign_in cached - sign in with incorrect email case before normalization fix, on second attempt canonical form is used (3.373s) ``` https://circleci.com/gh/mozilla/fxa-content-server/6113
1.0
test - sign in with incorrect email case before normalization fix, on second attempt canonical form is used - The following test is unstable on CI: ``` × firefox on any platform - sign_in cached - sign in with incorrect email case before normalization fix, on second attempt canonical form is used (3.373s) ``` https://circleci.com/gh/mozilla/fxa-content-server/6113
non_process
test sign in with incorrect email case before normalization fix on second attempt canonical form is used the following test is unstable on ci × firefox on any platform sign in cached sign in with incorrect email case before normalization fix on second attempt canonical form is used
0
13,287
15,764,977,166
IssuesEvent
2021-03-31 13:44:21
dhh1128/ctwg
https://api.github.com/repos/dhh1128/ctwg
opened
[PROCESS] Hyperlink transformation
process
## Need The current [hyperlinks-document](https://github.com/dhh1128/ctwg/blob/master/docs/hyperlinks.md) not only tries to inventory the various kinds of hyperlinks that need to be considered, but also how they change in the various transformations that we already have defined, and of which I expect others will follow. ## Proposed Solution Define the set of hyperlinks as they are used within the corpus, and as they may be used to refer into the corpus i.e.: - Local links - Fully qualified links - Cross-scope links - Links from issues - Transverse links - perhaps also, but I'm not sure of that: Transformer links Ensure that in the documentation of any [transformation tool](https://github.com/dhh1128/ctwg/issues/4) it is documented how these links are being modified, which is of course specific for the kind of tool and the purposes it serves.
1.0
[PROCESS] Hyperlink transformation - ## Need The current [hyperlinks-document](https://github.com/dhh1128/ctwg/blob/master/docs/hyperlinks.md) not only tries to inventory the various kinds of hyperlinks that need to be considered, but also how they change in the various transformations that we already have defined, and of which I expect others will follow. ## Proposed Solution Define the set of hyperlinks as they are used within the corpus, and as they may be used to refer into the corpus i.e.: - Local links - Fully qualified links - Cross-scope links - Links from issues - Transverse links - perhaps also, but I'm not sure of that: Transformer links Ensure that in the documentation of any [transformation tool](https://github.com/dhh1128/ctwg/issues/4) it is documented how these links are being modified, which is of course specific for the kind of tool and the purposes it serves.
process
hyperlink transformation need the current not only tries to inventory the various kinds of hyperlinks that need to be considered but also how they change in the various transformations that we already have defined and of which i expect others will follow proposed solution define the set of hyperlinks as they are used within the corpus and as they may be used to refer into the corpus i e local links fully qualified links cross scope links links from issues transverse links perhaps also but i m not sure of that transformer links ensure that in the documentation of any it is documented how these links are being modified which is of course specific for the kind of tool and the purposes it serves
1
689
3,175,950,934
IssuesEvent
2015-09-24 04:54:30
dita-ot/dita-ot
https://api.github.com/repos/dita-ot/dita-ot
closed
DITA-OT 2.0 - Build Error (Windows) - Illegal character - keyref target
bug P2 platform-dependent preprocess/keyref
Attempting to produce output using DITA-OT 2.0 on Windows results in build failure with the following message: > C:\ditaot\plugins\org.dita.base\build_preprocess.xml:285: java.lang.IllegalArgumentException: Illegal character in path at index 3: bcc\concepts\notification_events.dita The map file is the following: ```xml <!-- configuration.ditamap --> <topicref navtitle = "Notification Events" href = "../concepts/notification_events.dita" type = "concept"> <topicmeta><navtitle>Notification Events</navtitle></topicmeta></topicref> <keydef href = "images/notification_events.png" keys = "notification_events_image" format = "png"><topicmeta> <navtitle>NO TITLE</navtitle></topicmeta></keydef> ``` The topic in question: ```xml <!-- notification_events.dita --> <fig frame = "all"><image href = "../dhcp/images/notification_events.png" height="289.440pt" width="281.280pt" placement="break" keyref = "notification_events_image" /></fig> ```
1.0
DITA-OT 2.0 - Build Error (Windows) - Illegal character - keyref target - Attempting to produce output using DITA-OT 2.0 on Windows results in build failure with the following message: > C:\ditaot\plugins\org.dita.base\build_preprocess.xml:285: java.lang.IllegalArgumentException: Illegal character in path at index 3: bcc\concepts\notification_events.dita The map file is the following: ```xml <!-- configuration.ditamap --> <topicref navtitle = "Notification Events" href = "../concepts/notification_events.dita" type = "concept"> <topicmeta><navtitle>Notification Events</navtitle></topicmeta></topicref> <keydef href = "images/notification_events.png" keys = "notification_events_image" format = "png"><topicmeta> <navtitle>NO TITLE</navtitle></topicmeta></keydef> ``` The topic in question: ```xml <!-- notification_events.dita --> <fig frame = "all"><image href = "../dhcp/images/notification_events.png" height="289.440pt" width="281.280pt" placement="break" keyref = "notification_events_image" /></fig> ```
process
dita ot build error windows illegal character keyref target attempting to produce output using dita ot on windows results in build failure with the following message c ditaot plugins org dita base build preprocess xml java lang illegalargumentexception illegal character in path at index bcc concepts notification events dita the map file is the following xml topicref navtitle notification events href concepts notification events dita type concept notification events keydef href images notification events png keys notification events image format png no title the topic in question xml image href dhcp images notification events png height width placement break keyref notification events image
1
12,653
15,024,756,308
IssuesEvent
2021-02-01 20:05:42
elastic/beats
https://api.github.com/repos/elastic/beats
opened
Improve handling of missing field in dissect processor
:Processors Functionbeat
7.9.3 ``` - dissect: tokenizer: "Init Duration: %{init_duration} ms" field: "lambda.extra" target_prefix: "lambda" ignore_failure: true ``` When there is a dissect processor referencing a field but the event does not contain the field, we can only tell that something is wrong if we increase logging to the DEBUG level: ``` 2021-01-26T10:59:05.167-07:00 2021-01-26T17:59:05.149Z DEBUG [processors] processing/processors.go:112 Fail to apply processor global{drop_fields={"Fields":["agent","event_source_arn","event_name","event_id","event_version","kinesis_partition_key","kinesis_schema_version","kinesis_sequence_number","kinesis_encryption_type"],"IgnoreMissing":true}, dissect=%{type} RequestId: %{request_id} 2021/01/26/[$LATEST]<uuid> field=message,target_prefix=lambda, condition=regexp: map[], dissect=%{type} RequestId: %{request_id} Version: %{version} 2021/01/26/[$LATEST]<uuid> field=message,target_prefix=lambda, condition=regexp: map[], dissect=%{type} RequestId: %{request_id} Duration: %{duration_ms} ms Billed Duration: %{billed_duration_ms} ms Memory Size: %{memory_size_mb} MB Max Memory Used: %{max_memory_used_mb} MB %{extra},field=message,target_prefix=lambda, condition=regexp: map[], dissect=Init Duration: %{init_duration} ms,field=lambda.extra,target_prefix=lambda, dissect=%{timestamp} %{request_id} %{log_level} %{message} 2021/01/26/[$LATEST]<uuid> field=message,target_prefix=nodejs, condition=regexp: map[], decode_json_fields=message, nodejs.message}: key not found ``` I will be helpful to call out explicitly the field that is missing in the DEBUG log entry. Also, it will be helpful if we can add an `ignore_missing` option to the `dissect` processor similar to what we have for drop_fields processor so that users will not have to write their own check. ``` - dissect: when: has_fields: ["lambda.extra"] tokenizer: "Init Duration: %{init_duration} ms" field: "lambda.extra" target_prefix: "lambda" ```
1.0
Improve handling of missing field in dissect processor - 7.9.3 ``` - dissect: tokenizer: "Init Duration: %{init_duration} ms" field: "lambda.extra" target_prefix: "lambda" ignore_failure: true ``` When there is a dissect processor referencing a field but the event does not contain the field, we can only tell that something is wrong if we increase logging to the DEBUG level: ``` 2021-01-26T10:59:05.167-07:00 2021-01-26T17:59:05.149Z DEBUG [processors] processing/processors.go:112 Fail to apply processor global{drop_fields={"Fields":["agent","event_source_arn","event_name","event_id","event_version","kinesis_partition_key","kinesis_schema_version","kinesis_sequence_number","kinesis_encryption_type"],"IgnoreMissing":true}, dissect=%{type} RequestId: %{request_id} 2021/01/26/[$LATEST]<uuid> field=message,target_prefix=lambda, condition=regexp: map[], dissect=%{type} RequestId: %{request_id} Version: %{version} 2021/01/26/[$LATEST]<uuid> field=message,target_prefix=lambda, condition=regexp: map[], dissect=%{type} RequestId: %{request_id} Duration: %{duration_ms} ms Billed Duration: %{billed_duration_ms} ms Memory Size: %{memory_size_mb} MB Max Memory Used: %{max_memory_used_mb} MB %{extra},field=message,target_prefix=lambda, condition=regexp: map[], dissect=Init Duration: %{init_duration} ms,field=lambda.extra,target_prefix=lambda, dissect=%{timestamp} %{request_id} %{log_level} %{message} 2021/01/26/[$LATEST]<uuid> field=message,target_prefix=nodejs, condition=regexp: map[], decode_json_fields=message, nodejs.message}: key not found ``` I will be helpful to call out explicitly the field that is missing in the DEBUG log entry. Also, it will be helpful if we can add an `ignore_missing` option to the `dissect` processor similar to what we have for drop_fields processor so that users will not have to write their own check. ``` - dissect: when: has_fields: ["lambda.extra"] tokenizer: "Init Duration: %{init_duration} ms" field: "lambda.extra" target_prefix: "lambda" ```
process
improve handling of missing field in dissect processor dissect tokenizer init duration init duration ms field lambda extra target prefix lambda ignore failure true when there is a dissect processor referencing a field but the event does not contain the field we can only tell that something is wrong if we increase logging to the debug level debug processing processors go fail to apply processor global drop fields fields ignoremissing true dissect type requestid request id field message target prefix lambda condition regexp map dissect type requestid request id version version field message target prefix lambda condition regexp map dissect type requestid request id duration duration ms ms billed duration billed duration ms ms memory size memory size mb mb max memory used max memory used mb mb extra field message target prefix lambda condition regexp map dissect init duration init duration ms field lambda extra target prefix lambda dissect timestamp request id log level message field message target prefix nodejs condition regexp map decode json fields message nodejs message key not found i will be helpful to call out explicitly the field that is missing in the debug log entry also it will be helpful if we can add an ignore missing option to the dissect processor similar to what we have for drop fields processor so that users will not have to write their own check dissect when has fields tokenizer init duration init duration ms field lambda extra target prefix lambda
1
1,362
3,921,626,255
IssuesEvent
2016-04-22 00:14:54
18F/FEC
https://api.github.com/repos/18F/FEC
closed
Fonts are varying size on "Tax ID and bank account" page
bug processed
## General feedback? https://beta.fec.gov/registration-and-reporting/get-tax-id-and-bank-account/&#34;,&#34;https://beta.fec.gov/registration-and-reporting/get-tax-id-and-bank-account/ &#34;You&#39;ll list the name and address of the bank...&#34; is a smaller font than the rest of the page. Verified in Chrome and IE ## Details * URL: https://beta.fec.gov/ * User Agent: Mozilla/5.0 (Windows NT 6.1; rv:43.0) Gecko/20100101 Firefox/43.0
1.0
Fonts are varying size on "Tax ID and bank account" page - ## General feedback? https://beta.fec.gov/registration-and-reporting/get-tax-id-and-bank-account/&#34;,&#34;https://beta.fec.gov/registration-and-reporting/get-tax-id-and-bank-account/ &#34;You&#39;ll list the name and address of the bank...&#34; is a smaller font than the rest of the page. Verified in Chrome and IE ## Details * URL: https://beta.fec.gov/ * User Agent: Mozilla/5.0 (Windows NT 6.1; rv:43.0) Gecko/20100101 Firefox/43.0
process
fonts are varying size on tax id and bank account page general feedback you ll list the name and address of the bank is a smaller font than the rest of the page verified in chrome and ie details url user agent mozilla windows nt rv gecko firefox
1
283,980
30,913,577,277
IssuesEvent
2023-08-05 02:17:36
Satheesh575555/linux-4.1.15_CVE-2022-45934
https://api.github.com/repos/Satheesh575555/linux-4.1.15_CVE-2022-45934
reopened
CVE-2022-1943 (High) detected in linuxlinux-4.6
Mend: dependency security vulnerability
## CVE-2022-1943 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/Satheesh575555/linux-4.1.15_CVE-2022-45934/commit/7c0b143b43394df131d83e9aecb3c5518edc127a">7c0b143b43394df131d83e9aecb3c5518edc127a</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> A flaw out of bounds memory write in the Linux kernel UDF file system functionality was found in the way user triggers some file operation which triggers udf_write_fi(). A local user could use this flaw to crash the system or potentially <p>Publish Date: 2022-06-02 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-1943>CVE-2022-1943</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-1943">https://www.linuxkernelcves.com/cves/CVE-2022-1943</a></p> <p>Release Date: 2022-06-02</p> <p>Fix Resolution: v5.15.40,v5.17.8,v5.18-rc7</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-1943 (High) detected in linuxlinux-4.6 - ## CVE-2022-1943 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/Satheesh575555/linux-4.1.15_CVE-2022-45934/commit/7c0b143b43394df131d83e9aecb3c5518edc127a">7c0b143b43394df131d83e9aecb3c5518edc127a</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> A flaw out of bounds memory write in the Linux kernel UDF file system functionality was found in the way user triggers some file operation which triggers udf_write_fi(). A local user could use this flaw to crash the system or potentially <p>Publish Date: 2022-06-02 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-1943>CVE-2022-1943</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-1943">https://www.linuxkernelcves.com/cves/CVE-2022-1943</a></p> <p>Release Date: 2022-06-02</p> <p>Fix Resolution: v5.15.40,v5.17.8,v5.18-rc7</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in linuxlinux cve high severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch master vulnerable source files vulnerability details a flaw out of bounds memory write in the linux kernel udf file system functionality was found in the way user triggers some file operation which triggers udf write fi a local user could use this flaw to crash the system or potentially publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
578,713
17,150,517,826
IssuesEvent
2021-07-13 19:55:37
cagov/cannabis.ca.gov
https://api.github.com/repos/cagov/cannabis.ca.gov
opened
Design: Announcements page - Show only top 5 and paginate the rest + styling considerations
1 pt Low Priority - Non-critical
Feedback on styling: Maybe color code differences between Announcements, Press Releases, etc.
1.0
Design: Announcements page - Show only top 5 and paginate the rest + styling considerations - Feedback on styling: Maybe color code differences between Announcements, Press Releases, etc.
non_process
design announcements page show only top and paginate the rest styling considerations feedback on styling maybe color code differences between announcements press releases etc
0
587,842
17,632,673,994
IssuesEvent
2021-08-19 09:56:02
massenergize/api
https://api.github.com/repos/massenergize/api
closed
Enable Teams to span multiple communities
db model change priority 2
This was part of API #145 but not done at that time. An example is a religious congregation team with members from neighboring towns. Easiest part of this is to change the single ForeignKey field "Community" to a ManyToMany field "Communities" and updating the data so the list of Communities includes the original Community. **Question: should we maintain the Community field to denote the Primary Community?** Would allow the admin from the primary community to add or remove other communities (with proper notification) Otherwise, super admin required to add a second community to a team, on request from a community admin from the new community. Change to Model definition, related API changes (for team modification and download, backfill teams
1.0
Enable Teams to span multiple communities - This was part of API #145 but not done at that time. An example is a religious congregation team with members from neighboring towns. Easiest part of this is to change the single ForeignKey field "Community" to a ManyToMany field "Communities" and updating the data so the list of Communities includes the original Community. **Question: should we maintain the Community field to denote the Primary Community?** Would allow the admin from the primary community to add or remove other communities (with proper notification) Otherwise, super admin required to add a second community to a team, on request from a community admin from the new community. Change to Model definition, related API changes (for team modification and download, backfill teams
non_process
enable teams to span multiple communities this was part of api but not done at that time an example is a religious congregation team with members from neighboring towns easiest part of this is to change the single foreignkey field community to a manytomany field communities and updating the data so the list of communities includes the original community question should we maintain the community field to denote the primary community would allow the admin from the primary community to add or remove other communities with proper notification otherwise super admin required to add a second community to a team on request from a community admin from the new community change to model definition related api changes for team modification and download backfill teams
0
14,755
18,025,388,642
IssuesEvent
2021-09-17 03:19:33
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Build Virtual Raster tool creates an additional layer when selecting and reordering layers
Processing Bug
The Build Virtual Raster can create one additional, duplicate layer when reordering input files, leading to unexpected behaviour. It is almost certainly a bug in the Input Layers selection window itself. This seems to only happen when selecting (any/all) layers **first** and then reordering with drag-and-drop. Re-ordering first, or deselecting and reselecting will not cause this problem. It will only create 1 additional layer, no matter the number of reordering steps taken. I have found the resulting band order to be somewhat inconsistent, or at least is affected by the reordering steps performed. **How to Reproduce** 1. Add multiple raster layers to the project 2. Open "Build Virtual Raster" tool 3. Select the ellipsis (...) to modify the input layers 4. Select some/all layers 5. Drag at least one layer to reorder the list 6. Press ok 7. The number of selected layers listed in the "Input layers" box should be 1 more than previously selected (optional) 8. Check "Place each input file into a separate band 9. Press "Run" 10. Observe that the log notes a duplicate layer in an unexpected order and the output file contains an additional band **QGIS and OS versions** QGIS version 3.16.5-Hannover QGIS code revision 58ba7c1ed6 Compiled against Qt 5.11.2 Running against Qt 5.11.2 Compiled against GDAL/OGR 3.1.4 Running against GDAL/OGR 3.1.4 Compiled against GEOS 3.8.1-CAPI-1.13.3 Running against GEOS 3.8.1-CAPI-1.13.3 Compiled against SQLite 3.29.0 Running against SQLite 3.29.0 PostgreSQL Client Version 11.5 SpatiaLite Version 4.3.0 QWT Version 6.1.3 QScintilla2 Version 2.10.8 Compiled against PROJ 6.3.2 Running against PROJ Rel. 6.3.2, May 1st, 2020 OS Version Windows 10 (10.0) Active python plugins pointsamplingtool; QGIS3-getWKT; QuickWKT; quick_map_services; TerrainShading; db_manager; MetaSearch; processing **Additionally tested on :** QGIS version 3.20.0-Odense QGIS code branch Release 3.20 Qt version 5.15.2 Python version 3.9.5 GDAL/OGR version 3.3.0 PROJ version 8.0.1 EPSG Registry database version v10.018 (2021-04-02) GEOS version 3.9.1-CAPI-1.14.2 SQLite version 3.36.0 PDAL version 2.2.0 PostgreSQL client version 13.3 SpatiaLite version 5.0.1 QWT version 6.1.6 QScintilla2 version 2.12.1 OS version Arch Linux Active Python plugins QuickWKT QGIS3-getWKT processing MetaSearch db_manager
1.0
Build Virtual Raster tool creates an additional layer when selecting and reordering layers - The Build Virtual Raster can create one additional, duplicate layer when reordering input files, leading to unexpected behaviour. It is almost certainly a bug in the Input Layers selection window itself. This seems to only happen when selecting (any/all) layers **first** and then reordering with drag-and-drop. Re-ordering first, or deselecting and reselecting will not cause this problem. It will only create 1 additional layer, no matter the number of reordering steps taken. I have found the resulting band order to be somewhat inconsistent, or at least is affected by the reordering steps performed. **How to Reproduce** 1. Add multiple raster layers to the project 2. Open "Build Virtual Raster" tool 3. Select the ellipsis (...) to modify the input layers 4. Select some/all layers 5. Drag at least one layer to reorder the list 6. Press ok 7. The number of selected layers listed in the "Input layers" box should be 1 more than previously selected (optional) 8. Check "Place each input file into a separate band 9. Press "Run" 10. Observe that the log notes a duplicate layer in an unexpected order and the output file contains an additional band **QGIS and OS versions** QGIS version 3.16.5-Hannover QGIS code revision 58ba7c1ed6 Compiled against Qt 5.11.2 Running against Qt 5.11.2 Compiled against GDAL/OGR 3.1.4 Running against GDAL/OGR 3.1.4 Compiled against GEOS 3.8.1-CAPI-1.13.3 Running against GEOS 3.8.1-CAPI-1.13.3 Compiled against SQLite 3.29.0 Running against SQLite 3.29.0 PostgreSQL Client Version 11.5 SpatiaLite Version 4.3.0 QWT Version 6.1.3 QScintilla2 Version 2.10.8 Compiled against PROJ 6.3.2 Running against PROJ Rel. 6.3.2, May 1st, 2020 OS Version Windows 10 (10.0) Active python plugins pointsamplingtool; QGIS3-getWKT; QuickWKT; quick_map_services; TerrainShading; db_manager; MetaSearch; processing **Additionally tested on :** QGIS version 3.20.0-Odense QGIS code branch Release 3.20 Qt version 5.15.2 Python version 3.9.5 GDAL/OGR version 3.3.0 PROJ version 8.0.1 EPSG Registry database version v10.018 (2021-04-02) GEOS version 3.9.1-CAPI-1.14.2 SQLite version 3.36.0 PDAL version 2.2.0 PostgreSQL client version 13.3 SpatiaLite version 5.0.1 QWT version 6.1.6 QScintilla2 version 2.12.1 OS version Arch Linux Active Python plugins QuickWKT QGIS3-getWKT processing MetaSearch db_manager
process
build virtual raster tool creates an additional layer when selecting and reordering layers the build virtual raster can create one additional duplicate layer when reordering input files leading to unexpected behaviour it is almost certainly a bug in the input layers selection window itself this seems to only happen when selecting any all layers first and then reordering with drag and drop re ordering first or deselecting and reselecting will not cause this problem it will only create additional layer no matter the number of reordering steps taken i have found the resulting band order to be somewhat inconsistent or at least is affected by the reordering steps performed how to reproduce add multiple raster layers to the project open build virtual raster tool select the ellipsis to modify the input layers select some all layers drag at least one layer to reorder the list press ok the number of selected layers listed in the input layers box should be more than previously selected optional check place each input file into a separate band press run observe that the log notes a duplicate layer in an unexpected order and the output file contains an additional band qgis and os versions qgis version hannover qgis code revision compiled against qt running against qt compiled against gdal ogr running against gdal ogr compiled against geos capi running against geos capi compiled against sqlite running against sqlite postgresql client version spatialite version qwt version version compiled against proj running against proj rel may os version windows active python plugins pointsamplingtool getwkt quickwkt quick map services terrainshading db manager metasearch processing additionally tested on qgis version odense qgis code branch release qt version python version gdal ogr version proj version epsg registry database version geos version capi sqlite version pdal version postgresql client version spatialite version qwt version version os version arch linux active python plugins quickwkt getwkt processing metasearch db manager
1
20,959
27,817,509,513
IssuesEvent
2023-03-18 21:19:03
cse442-at-ub/project_s23-iweatherify
https://api.github.com/repos/cse442-at-ub/project_s23-iweatherify
closed
Load unit and temperature setting from the database to the respective page when visited
Processing Task Sprint 2
**Task Tests** *Test 1* 1. Go to the following URL: https://github.com/cse442-at-ub/project_s23-iweatherify/tree/dev 2. Click on the green `<> Code` button and download the ZIP file. ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/75c741f7-5ca3-4f64-879f-df960ad51a8b) 3. Unzip the downloaded file to a folder on your computer. 4. Open a terminal and navigate to the git repository folder using the `cd` command. 5. Run the `npm install` command in the terminal to install the necessary dependencies. 6. Run the `npm start` command in the terminal to start the application. 7. Check the output from the npm start command for the URL to access the application. The URL should be a localhost address (e.g., http://localhost:8080). 8. Navigate to http://localhost:8080/#/login 9. Ensure you have logged in to our app to see the page use UserID: `UB442` and Password:`Myub442@!` to login 10. Go to URL: http://localhost:8080/#/unitsSettings 11. Verify that the units page is displayed ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/3f9cbb63-e582-43f0-91bf-16970daf57eb) 12. Change the temperature unit to Celsius (°C) 13. Change the wind unit to mph 14. Change the pressure unit to mm 15. Change the distance unit to mi 16. Click the save button 17. Open a different tab and go to: https://www-student.cse.buffalo.edu/tools/db/phpmyadmin/index.php 18. Input username: `jpan26` and password: `50314999` 19. Make sure the server choice is `oceanus.cse.buffalo.edu:3306` 20. Click go and you should see this page ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/443bf59a-f5fd-4dbe-88a8-645493eaa713) 21. Click `cse442_2023_spring_team_a_db` first and then `saved_units` on the left side of the page ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/9cb87c0a-e93f-4327-9640-8f6c0478c3a2) 22. Verify you see a row with the exact same information shown by the picture ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/1f326a8d-6adc-4863-ba2f-2dac935fe23b) 23. Go back to the units page you can press 'F5' to refresh the page and you should see the same units you saved previously on the the page 24. You can log out and log back in and go to units page 25. You should still see the saved same unit settings on the page *Test 2* 1. Repeat steps 1 to 9 from `Test 1` 2. Go to URL: http://localhost:8080/#/tempSettings 3. Verify that the temperature setting page is displayed ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/1feae072-4846-46ae-8639-9958248e9158) 4. Change the hot temperature to 100, you can either use the slider or input box and click save 5. Change the warm temperature to 75, you can either use the slider or input box and click save 6. Change the ideal temperature to 25, you can either use the slider or input box and click save 7. Change the chilly temperature to 10, you can either use the slider or input box and click save 8. Change the cold temperature to -25, you can either use the slider or input box and click save 9. Change the freezing temperature to -75, you can either use the slider or input box and click save 10. Repeat steps 17 to 20 from `Test 1` 11. Click `cse442_2023_spring_team_a_db` first and then `saved_temperatures` on the left side of the page ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/ae9e6ed8-1291-4690-89ab-ce8cb4223843) 12. Verify you see a row with the exact same information as shown by the picture ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/16b1d57f-5c20-4bb0-b4fa-82833da6ca95) 23. Go back to the temperature setting page you can press 'F5' to refresh the page and you should see the same temperatures you saved previously on the the page 24. You can log out and log back in and go to temperature settings page 25. You should still see the saved same temperatures on the page
1.0
Load unit and temperature setting from the database to the respective page when visited - **Task Tests** *Test 1* 1. Go to the following URL: https://github.com/cse442-at-ub/project_s23-iweatherify/tree/dev 2. Click on the green `<> Code` button and download the ZIP file. ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/75c741f7-5ca3-4f64-879f-df960ad51a8b) 3. Unzip the downloaded file to a folder on your computer. 4. Open a terminal and navigate to the git repository folder using the `cd` command. 5. Run the `npm install` command in the terminal to install the necessary dependencies. 6. Run the `npm start` command in the terminal to start the application. 7. Check the output from the npm start command for the URL to access the application. The URL should be a localhost address (e.g., http://localhost:8080). 8. Navigate to http://localhost:8080/#/login 9. Ensure you have logged in to our app to see the page use UserID: `UB442` and Password:`Myub442@!` to login 10. Go to URL: http://localhost:8080/#/unitsSettings 11. Verify that the units page is displayed ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/3f9cbb63-e582-43f0-91bf-16970daf57eb) 12. Change the temperature unit to Celsius (°C) 13. Change the wind unit to mph 14. Change the pressure unit to mm 15. Change the distance unit to mi 16. Click the save button 17. Open a different tab and go to: https://www-student.cse.buffalo.edu/tools/db/phpmyadmin/index.php 18. Input username: `jpan26` and password: `50314999` 19. Make sure the server choice is `oceanus.cse.buffalo.edu:3306` 20. Click go and you should see this page ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/443bf59a-f5fd-4dbe-88a8-645493eaa713) 21. Click `cse442_2023_spring_team_a_db` first and then `saved_units` on the left side of the page ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/9cb87c0a-e93f-4327-9640-8f6c0478c3a2) 22. Verify you see a row with the exact same information shown by the picture ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/1f326a8d-6adc-4863-ba2f-2dac935fe23b) 23. Go back to the units page you can press 'F5' to refresh the page and you should see the same units you saved previously on the the page 24. You can log out and log back in and go to units page 25. You should still see the saved same unit settings on the page *Test 2* 1. Repeat steps 1 to 9 from `Test 1` 2. Go to URL: http://localhost:8080/#/tempSettings 3. Verify that the temperature setting page is displayed ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/1feae072-4846-46ae-8639-9958248e9158) 4. Change the hot temperature to 100, you can either use the slider or input box and click save 5. Change the warm temperature to 75, you can either use the slider or input box and click save 6. Change the ideal temperature to 25, you can either use the slider or input box and click save 7. Change the chilly temperature to 10, you can either use the slider or input box and click save 8. Change the cold temperature to -25, you can either use the slider or input box and click save 9. Change the freezing temperature to -75, you can either use the slider or input box and click save 10. Repeat steps 17 to 20 from `Test 1` 11. Click `cse442_2023_spring_team_a_db` first and then `saved_temperatures` on the left side of the page ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/ae9e6ed8-1291-4690-89ab-ce8cb4223843) 12. Verify you see a row with the exact same information as shown by the picture ![image.png](https://images.zenhubusercontent.com/63e1796387907702186b8c6a/16b1d57f-5c20-4bb0-b4fa-82833da6ca95) 23. Go back to the temperature setting page you can press 'F5' to refresh the page and you should see the same temperatures you saved previously on the the page 24. You can log out and log back in and go to temperature settings page 25. You should still see the saved same temperatures on the page
process
load unit and temperature setting from the database to the respective page when visited task tests test go to the following url click on the green code button and download the zip file unzip the downloaded file to a folder on your computer open a terminal and navigate to the git repository folder using the cd command run the npm install command in the terminal to install the necessary dependencies run the npm start command in the terminal to start the application check the output from the npm start command for the url to access the application the url should be a localhost address e g navigate to ensure you have logged in to our app to see the page use userid and password to login go to url verify that the units page is displayed change the temperature unit to celsius °c change the wind unit to mph change the pressure unit to mm change the distance unit to mi click the save button open a different tab and go to input username and password make sure the server choice is oceanus cse buffalo edu click go and you should see this page click spring team a db first and then saved units on the left side of the page verify you see a row with the exact same information shown by the picture go back to the units page you can press to refresh the page and you should see the same units you saved previously on the the page you can log out and log back in and go to units page you should still see the saved same unit settings on the page test repeat steps to from test go to url verify that the temperature setting page is displayed change the hot temperature to you can either use the slider or input box and click save change the warm temperature to you can either use the slider or input box and click save change the ideal temperature to you can either use the slider or input box and click save change the chilly temperature to you can either use the slider or input box and click save change the cold temperature to you can either use the slider or input box and click save change the freezing temperature to you can either use the slider or input box and click save repeat steps to from test click spring team a db first and then saved temperatures on the left side of the page verify you see a row with the exact same information as shown by the picture go back to the temperature setting page you can press to refresh the page and you should see the same temperatures you saved previously on the the page you can log out and log back in and go to temperature settings page you should still see the saved same temperatures on the page
1
8,980
12,099,692,119
IssuesEvent
2020-04-20 12:38:02
googleapis/google-cloud-dotnet
https://api.github.com/repos/googleapis/google-cloud-dotnet
closed
We may be able to stop skipping UploadObject_BucketHasDefaultKmsKey_UploadWithCsek
type: process
(Will investigate when I have time.)
1.0
We may be able to stop skipping UploadObject_BucketHasDefaultKmsKey_UploadWithCsek - (Will investigate when I have time.)
process
we may be able to stop skipping uploadobject buckethasdefaultkmskey uploadwithcsek will investigate when i have time
1
13,140
15,558,728,279
IssuesEvent
2021-03-16 10:38:39
prisma/prisma
https://api.github.com/repos/prisma/prisma
closed
`prisma migrate deploy` sometimes errors with `Database error: Error querying the database: db error: ERROR: prepared statement "s0" does not exist` with pgbouncer
bug/0-needs-info kind/bug process/candidate team/migrations topic: pgbouncer
<!-- Thanks for helping us improve Prisma! 🙏 Please follow the sections in the template and provide as much information as possible about your problem, e.g. by setting the `DEBUG="*"` environment variable and enabling additional logging output in Prisma Client. Learn more about writing proper bug reports here: https://pris.ly/d/bug-reports --> ## Bug description <!-- A clear and concise description of what the bug is. --> https://prisma.slack.com/archives/CA491RJH0/p1615370492400800?thread_ts=1614161039.079900&cid=CA491RJH0 User says: Funny things is that when i redeploy once or twice it works :man-shrugging: always running the same `NODE_ENV=production blitz prisma migrate deploy --preview-feature` - prisma & prisma-client are at 2.18.0 I just tested prisma for a few weeks so i only tested 2.17 - 2.18 Db is hosted on digital-ocean(london 1) with pgBouncer enabled. Actually it seems to be two errors - in most cases it’s just ``` 21:35:05.746 Error: P1002 21:35:05.746 The database server at `xx.xxx.xx.xxx`:`25061` was reached but timed out. ``` But at least once it also was: ``` Response: Database error: Error querying the database: db error: ERROR: prepared statement "s0" does not exist ``` ## How to reproduce <!-- Steps to reproduce the behavior: 1. Go to '...' 2. Change '....' 3. Run '....' 4. See error --> ## Expected behavior <!-- A clear and concise description of what you expected to happen. --> ## Prisma information <!-- Your Prisma schema, Prisma Client queries, ... Do not include your database credentials when sharing your Prisma schema! --> ## Environment & setup <!-- In which environment does the problem occur --> - OS: <!--[e.g. Mac OS, Windows, Debian, CentOS, ...]--> - Database: <!--[PostgreSQL, MySQL, MariaDB or SQLite]--> - Node.js version: <!--[Run `node -v` to see your Node.js version]--> - Prisma version: <!--[Run `prisma -v` to see your Prisma version and paste it between the ´´´]--> ``` ```
1.0
`prisma migrate deploy` sometimes errors with `Database error: Error querying the database: db error: ERROR: prepared statement "s0" does not exist` with pgbouncer - <!-- Thanks for helping us improve Prisma! 🙏 Please follow the sections in the template and provide as much information as possible about your problem, e.g. by setting the `DEBUG="*"` environment variable and enabling additional logging output in Prisma Client. Learn more about writing proper bug reports here: https://pris.ly/d/bug-reports --> ## Bug description <!-- A clear and concise description of what the bug is. --> https://prisma.slack.com/archives/CA491RJH0/p1615370492400800?thread_ts=1614161039.079900&cid=CA491RJH0 User says: Funny things is that when i redeploy once or twice it works :man-shrugging: always running the same `NODE_ENV=production blitz prisma migrate deploy --preview-feature` - prisma & prisma-client are at 2.18.0 I just tested prisma for a few weeks so i only tested 2.17 - 2.18 Db is hosted on digital-ocean(london 1) with pgBouncer enabled. Actually it seems to be two errors - in most cases it’s just ``` 21:35:05.746 Error: P1002 21:35:05.746 The database server at `xx.xxx.xx.xxx`:`25061` was reached but timed out. ``` But at least once it also was: ``` Response: Database error: Error querying the database: db error: ERROR: prepared statement "s0" does not exist ``` ## How to reproduce <!-- Steps to reproduce the behavior: 1. Go to '...' 2. Change '....' 3. Run '....' 4. See error --> ## Expected behavior <!-- A clear and concise description of what you expected to happen. --> ## Prisma information <!-- Your Prisma schema, Prisma Client queries, ... Do not include your database credentials when sharing your Prisma schema! --> ## Environment & setup <!-- In which environment does the problem occur --> - OS: <!--[e.g. Mac OS, Windows, Debian, CentOS, ...]--> - Database: <!--[PostgreSQL, MySQL, MariaDB or SQLite]--> - Node.js version: <!--[Run `node -v` to see your Node.js version]--> - Prisma version: <!--[Run `prisma -v` to see your Prisma version and paste it between the ´´´]--> ``` ```
process
prisma migrate deploy sometimes errors with database error error querying the database db error error prepared statement does not exist with pgbouncer thanks for helping us improve prisma 🙏 please follow the sections in the template and provide as much information as possible about your problem e g by setting the debug environment variable and enabling additional logging output in prisma client learn more about writing proper bug reports here bug description user says funny things is that when i redeploy once or twice it works man shrugging always running the same node env production blitz prisma migrate deploy preview feature prisma prisma client are at i just tested prisma for a few weeks so i only tested db is hosted on digital ocean london with pgbouncer enabled actually it seems to be two errors in most cases it’s just error the database server at xx xxx xx xxx was reached but timed out but at least once it also was response database error error querying the database db error error prepared statement does not exist how to reproduce steps to reproduce the behavior go to change run see error expected behavior prisma information your prisma schema prisma client queries do not include your database credentials when sharing your prisma schema environment setup os database node js version prisma version
1
5,740
8,580,836,332
IssuesEvent
2018-11-13 13:10:50
easy-software-ufal/annotations_repos
https://api.github.com/repos/easy-software-ufal/annotations_repos
opened
NakedObjectsGroup/NakedObjectsFramework Table view not rendering properly if field in a column is not visible
C# RMA test wrong processing
Issue: `https://github.com/NakedObjectsGroup/NakedObjectsFramework/issues/41` PR: `https://github.com/NakedObjectsGroup/NakedObjectsFramework/commit/5630bcfca67216b4871ecf53b0a984057af6b077` Multiple pull requests.
1.0
NakedObjectsGroup/NakedObjectsFramework Table view not rendering properly if field in a column is not visible - Issue: `https://github.com/NakedObjectsGroup/NakedObjectsFramework/issues/41` PR: `https://github.com/NakedObjectsGroup/NakedObjectsFramework/commit/5630bcfca67216b4871ecf53b0a984057af6b077` Multiple pull requests.
process
nakedobjectsgroup nakedobjectsframework table view not rendering properly if field in a column is not visible issue pr multiple pull requests
1
769,899
27,021,609,035
IssuesEvent
2023-02-11 03:54:57
okTurtles/group-income
https://api.github.com/repos/okTurtles/group-income
closed
Archive payments in indexed storage using pagination
Kind:Enhancement App:Frontend Priority:High
### Problem We have `gi.contracts/group/archivePayments` method in group contract which archives historical payments in Indexed Storage of the browser. What this function does is to clone the state of historical `payments` and `paymentsByPeriod` to the indexed storage. We need the payments in pagination, so we should find another way to save historical `payments` and `paymentsByPeriod` so that it could be easy to get them for a specific page and also in types(`Received` and `Sent`) easily. Discussed here: https://github.com/okTurtles/group-income/pull/1434#discussion_r1008971602 ### Solution Need to divide `payments` and `paymentsByPeriod` into several parts using different keys so that it's easy to pick payments easily if necessary. Should TAKE CARE of the payments types of `Received` and `Sent`. I think it's good to make keys for each of `Received` and `Sent`.
1.0
Archive payments in indexed storage using pagination - ### Problem We have `gi.contracts/group/archivePayments` method in group contract which archives historical payments in Indexed Storage of the browser. What this function does is to clone the state of historical `payments` and `paymentsByPeriod` to the indexed storage. We need the payments in pagination, so we should find another way to save historical `payments` and `paymentsByPeriod` so that it could be easy to get them for a specific page and also in types(`Received` and `Sent`) easily. Discussed here: https://github.com/okTurtles/group-income/pull/1434#discussion_r1008971602 ### Solution Need to divide `payments` and `paymentsByPeriod` into several parts using different keys so that it's easy to pick payments easily if necessary. Should TAKE CARE of the payments types of `Received` and `Sent`. I think it's good to make keys for each of `Received` and `Sent`.
non_process
archive payments in indexed storage using pagination problem we have gi contracts group archivepayments method in group contract which archives historical payments in indexed storage of the browser what this function does is to clone the state of historical payments and paymentsbyperiod to the indexed storage we need the payments in pagination so we should find another way to save historical payments and paymentsbyperiod so that it could be easy to get them for a specific page and also in types received and sent easily discussed here solution need to divide payments and paymentsbyperiod into several parts using different keys so that it s easy to pick payments easily if necessary should take care of the payments types of received and sent i think it s good to make keys for each of received and sent
0
14,647
17,774,412,730
IssuesEvent
2021-08-30 17:17:07
julioPlaceres/A_Helping_Hand
https://api.github.com/repos/julioPlaceres/A_Helping_Hand
opened
Home Page
in process
Create Home Page (Initial page that will be seeing, with options to log in and sign up)
1.0
Home Page - Create Home Page (Initial page that will be seeing, with options to log in and sign up)
process
home page create home page initial page that will be seeing with options to log in and sign up
1
560,024
16,583,248,263
IssuesEvent
2021-05-31 14:39:20
asteca/ASteCA
https://api.github.com/repos/asteca/ASteCA
closed
Explore chaospy as a possible addition to the best fit process
priority:low type:best-fit type:packages
* Article: [Chaospy: An open source tool for designing methods of uncertainty quantification](https://www.sciencedirect.com/science/article/pii/S1877750315300119?via%3Dihub) * Repo: https://github.com/jonathf/chaospy * Docs: https://chaospy.readthedocs.io/en/master/index.html > Chaospy is a numerical tool for performing uncertainty quantification using polynomial chaos expansions and Monte Carlo methods.
1.0
Explore chaospy as a possible addition to the best fit process - * Article: [Chaospy: An open source tool for designing methods of uncertainty quantification](https://www.sciencedirect.com/science/article/pii/S1877750315300119?via%3Dihub) * Repo: https://github.com/jonathf/chaospy * Docs: https://chaospy.readthedocs.io/en/master/index.html > Chaospy is a numerical tool for performing uncertainty quantification using polynomial chaos expansions and Monte Carlo methods.
non_process
explore chaospy as a possible addition to the best fit process article repo docs chaospy is a numerical tool for performing uncertainty quantification using polynomial chaos expansions and monte carlo methods
0
8,886
11,984,094,180
IssuesEvent
2020-04-07 15:21:52
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
closed
NTR: effector-mediated suppression of effector-triggered immunity
multi-species process
A process mediated by a molecule secreted by a symbiont that results in the suppression of host effector-triggered immune response. The host is defined as the larger of the organisms involved in a symbiotic interaction. descendant of GO:0140403 effector-mediated suppression of host innate immune response by symbiont PMID:31804478 @CuzickA this sounds a bit weird. I guess it is correct as it's the third part of the zig-zag-zig model (I.e and effector mediating a response to an effector-mediated response)
1.0
NTR: effector-mediated suppression of effector-triggered immunity - A process mediated by a molecule secreted by a symbiont that results in the suppression of host effector-triggered immune response. The host is defined as the larger of the organisms involved in a symbiotic interaction. descendant of GO:0140403 effector-mediated suppression of host innate immune response by symbiont PMID:31804478 @CuzickA this sounds a bit weird. I guess it is correct as it's the third part of the zig-zag-zig model (I.e and effector mediating a response to an effector-mediated response)
process
ntr effector mediated suppression of effector triggered immunity a process mediated by a molecule secreted by a symbiont that results in the suppression of host effector triggered immune response the host is defined as the larger of the organisms involved in a symbiotic interaction descendant of go effector mediated suppression of host innate immune response by symbiont pmid cuzicka this sounds a bit weird i guess it is correct as it s the third part of the zig zag zig model i e and effector mediating a response to an effector mediated response
1
546
3,005,947,173
IssuesEvent
2015-07-27 06:41:56
HotCore-Studio/hot-project
https://api.github.com/repos/HotCore-Studio/hot-project
closed
основной 3D рендерер
In process
Инициализация, работа, рендеринг, установка вьюпортов, свап чеинов, контекста и тд. Выполняет [ToxikCoder](https://github.com/ToxikCoder)
1.0
основной 3D рендерер - Инициализация, работа, рендеринг, установка вьюпортов, свап чеинов, контекста и тд. Выполняет [ToxikCoder](https://github.com/ToxikCoder)
process
основной рендерер инициализация работа рендеринг установка вьюпортов свап чеинов контекста и тд выполняет
1
469,553
13,520,692,757
IssuesEvent
2020-09-15 05:29:53
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.facebook.com - site is not usable
browser-firefox engine-gecko ml-needsdiagnosis-false ml-probability-high priority-critical
<!-- @browser: Firefox 81.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 6.1; Win64; x64; rv:81.0) Gecko/20100101 Firefox/81.0 --> <!-- @reported_with: desktop-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/58248 --> **URL**: https://www.facebook.com/rsrc.php/v3/yx/r/pyNVUg5EM0j.png **Browser / Version**: Firefox 81.0 **Operating System**: Windows 7 **Tested Another Browser**: Yes Internet Explorer **Problem type**: Site is not usable **Description**: Buttons or links not working **Steps to Reproduce**: <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2020/9/f2bda1c7-95d0-4218-ada0-b52b08dc3314.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200910180444</li><li>channel: beta</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2020/9/8693db54-5c66-46d5-8552-98e00145f6af) _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.facebook.com - site is not usable - <!-- @browser: Firefox 81.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 6.1; Win64; x64; rv:81.0) Gecko/20100101 Firefox/81.0 --> <!-- @reported_with: desktop-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/58248 --> **URL**: https://www.facebook.com/rsrc.php/v3/yx/r/pyNVUg5EM0j.png **Browser / Version**: Firefox 81.0 **Operating System**: Windows 7 **Tested Another Browser**: Yes Internet Explorer **Problem type**: Site is not usable **Description**: Buttons or links not working **Steps to Reproduce**: <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2020/9/f2bda1c7-95d0-4218-ada0-b52b08dc3314.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200910180444</li><li>channel: beta</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2020/9/8693db54-5c66-46d5-8552-98e00145f6af) _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
site is not usable url browser version firefox operating system windows tested another browser yes internet explorer problem type site is not usable description buttons or links not working steps to reproduce view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel beta hastouchscreen false mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
0
465,018
13,349,998,046
IssuesEvent
2020-08-30 05:04:37
internetarchive/openlibrary
https://api.github.com/repos/internetarchive/openlibrary
closed
There is a mismatch of availability of a book in a carousel to its true status
Good First Issue Lead: @mekarpeles Priority: 3 Type: Bug
### Example: https://openlibrary.org/books/OL26571490M/Flush https://openlibrary.org/books/OL26571491M/A_haunted_house_and_other_short_stories ![image](https://user-images.githubusercontent.com/55302685/77956345-92111a80-7286-11ea-8a9b-e17431ff2ccd.png) ### Problem: Works in carousel say 'read', but when clicked on says 'no ebook available'. The expected result should be a read button there. ### Solution: Change the carousel to match the availability of the edition (as what shows on the editions page is the true status) ### Stakeholders Users - this issue creates confusion and frustration for someone hoping to see a book only to find out it's not available.
1.0
There is a mismatch of availability of a book in a carousel to its true status - ### Example: https://openlibrary.org/books/OL26571490M/Flush https://openlibrary.org/books/OL26571491M/A_haunted_house_and_other_short_stories ![image](https://user-images.githubusercontent.com/55302685/77956345-92111a80-7286-11ea-8a9b-e17431ff2ccd.png) ### Problem: Works in carousel say 'read', but when clicked on says 'no ebook available'. The expected result should be a read button there. ### Solution: Change the carousel to match the availability of the edition (as what shows on the editions page is the true status) ### Stakeholders Users - this issue creates confusion and frustration for someone hoping to see a book only to find out it's not available.
non_process
there is a mismatch of availability of a book in a carousel to its true status example problem works in carousel say read but when clicked on says no ebook available the expected result should be a read button there solution change the carousel to match the availability of the edition as what shows on the editions page is the true status stakeholders users this issue creates confusion and frustration for someone hoping to see a book only to find out it s not available
0
13,774
16,531,090,175
IssuesEvent
2021-05-27 06:07:55
aodn/imos-toolbox
https://api.github.com/repos/aodn/imos-toolbox
opened
WorkhorseParser - wrong assignment of velocity components
Type:Reprocessing Type:bug Unit:Instrument Reader Unit:TimeSeries
This bug affects v2.6.11 & v2.6.12 When reading ENU datasets with the newly refactored workhorse Parser, the variable mappings are reversed and a wrong assignment is being done. The current bug lies in assigning `velocity1->VCUR` and `velocity2->UCUR`, while the correct is the reverse. This only occurs for ENU datasets. The problem is located in the import_mappings on the recently refactored Workhorse parser (v2.6.11+). It went undetected because even the tests got the typo (see `+Workhorse/import_mappings.m`). We are also missing a content regression test against ENU files, which would have picked the problem. The origin is likely related to a wrong copy/paste/edit since the original workhorseParser firstly defined a VCUR variable, then a UCUR variable, but with the correct assignments.
1.0
WorkhorseParser - wrong assignment of velocity components - This bug affects v2.6.11 & v2.6.12 When reading ENU datasets with the newly refactored workhorse Parser, the variable mappings are reversed and a wrong assignment is being done. The current bug lies in assigning `velocity1->VCUR` and `velocity2->UCUR`, while the correct is the reverse. This only occurs for ENU datasets. The problem is located in the import_mappings on the recently refactored Workhorse parser (v2.6.11+). It went undetected because even the tests got the typo (see `+Workhorse/import_mappings.m`). We are also missing a content regression test against ENU files, which would have picked the problem. The origin is likely related to a wrong copy/paste/edit since the original workhorseParser firstly defined a VCUR variable, then a UCUR variable, but with the correct assignments.
process
workhorseparser wrong assignment of velocity components this bug affects when reading enu datasets with the newly refactored workhorse parser the variable mappings are reversed and a wrong assignment is being done the current bug lies in assigning vcur and ucur while the correct is the reverse this only occurs for enu datasets the problem is located in the import mappings on the recently refactored workhorse parser it went undetected because even the tests got the typo see workhorse import mappings m we are also missing a content regression test against enu files which would have picked the problem the origin is likely related to a wrong copy paste edit since the original workhorseparser firstly defined a vcur variable then a ucur variable but with the correct assignments
1
2,429
4,807,573,149
IssuesEvent
2016-11-02 21:52:45
bfpepper/little-shop-1608
https://api.github.com/repos/bfpepper/little-shop-1608
closed
Retired Items
Base requirements
As a user if I visit an item page and that item has been retired Then I should still be able to access the item page And I should not be able to add the item to their cart And I should see in place of the "Add to Cart" button or link - "Item Retired"
1.0
Retired Items - As a user if I visit an item page and that item has been retired Then I should still be able to access the item page And I should not be able to add the item to their cart And I should see in place of the "Add to Cart" button or link - "Item Retired"
non_process
retired items as a user if i visit an item page and that item has been retired then i should still be able to access the item page and i should not be able to add the item to their cart and i should see in place of the add to cart button or link item retired
0
21,425
29,359,593,475
IssuesEvent
2023-05-28 00:37:11
devssa/onde-codar-em-salvador
https://api.github.com/repos/devssa/onde-codar-em-salvador
closed
[Hibrido / São Paulo, São Paulo, Brazil] Fullstack Developer (Java) (Híbrido - São Paulo) na Coodesh
SALVADOR BACK-END FRONT-END JAVA FULL-STACK ANGULAR REACT REQUISITOS PROCESSOS BACKEND GITHUB UMA QUALIDADE R NEGÓCIOS ALOCADO Stale
## Descrição da vaga: Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios. Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/jobs/consultant-java-developer-hibrida-sao-paulo-172418087?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋 <p>A <strong>KLB Group</strong> busca <strong><ins>Fullstack Developer (Java)</ins></strong> para compor seu time!</p> <p></p> <p>A KLB Group é especializada na implementação de projetos em empresas públicas e privadas. Seja em projetos de desenvolvimento, produção ou transformação, a KLB Group assegura a implementação eficaz, mobilizando rapidamente uma equipe de especialistas de várias funções (compras, cadeia de suprimentos, qualidade, engenharia, TI, finanças, etc.), com uma combinação única de experiência em design, implementação e operação. A KLB Group tem mais de 500 funcionários na Europa, América e Ásia.</p> <p></p> <p><strong>Atividades e Atribuições: </strong></p> <ul> <li>&nbsp;Realizar atendimentos das demandas dos Portais de Sinistros;</li> <li>&nbsp;Analisar as demandas visando atender às necessidades dos clientes;</li> <li>&nbsp;Levantamento de requisitos;</li> <li>Prever impactos das alteraçõs, garantindo a integridade do sistema.</li> </ul> ## KLB GROUP BRASIL : <p>A KLB Group é especializada na implementação de projetos em empresas públicas e privadas. Seja em projetos de desenvolvimento, produção ou transformação, a KLB Group assegura a implementação eficaz, mobilizando rapidamente uma equipe de especialistas de várias funções (compras, cadeia de suprimentos, qualidade, engenharia, TI, finanças, etc.), com uma combinação única de experiência em design, implementação e operação. A KLB Group tem mais de 500 funcionários na Europa, América e Ásia.</p> </p> ## Habilidades: - Java - Front-end System Design - Back-end System Design ## Local: São Paulo, São Paulo, Brazil ## Requisitos: - Ensino Superior Completo; - Conhecimento de tecnologias para desenvolvimento Java no Backend; - Estudo da Arquitetura das aplicações para propor soluções; - Entendimento das soluções sistêmicas visando a solução do problemas ou a melhoria dos processos de negócios; - Conhecimento de tecnologias para desenvolvimento com Angular, React, entre outras no Frontend. ## Diferenciais: - Desejável conhecimento de Seguros. ## Benefícios: - Vale-refeição R$35,43 dia/ média R$779,70; - Vale-alimentação R$ 602,40; - Vale-transporte (6% de desconto); - Assistência médica/medicina em grupo; - Assistência odontológica R$10,87; - Auxílio Babá - R$ 855,16 até 6 meses; - Auxílio Creche - R$ 460,85 a partir de 7 meses até 5 anos e 11 meses; - Auxílio Babá - R$ 460,85 a partir de 7 meses até 5 anos e 11 meses; - PLR; - Seguro de vida; - Empréstimo; - Plano de previdência privada; - Programa de bolsa de estudo; - Total pass. ## Como se candidatar: Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Fullstack Developer (Java) (Híbrido - São Paulo) na KLB GROUP BRASIL ](https://coodesh.com/jobs/consultant-java-developer-hibrida-sao-paulo-172418087?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação. ## Labels #### Alocação Alocado #### Regime CLT #### Categoria Full-Stack
1.0
[Hibrido / São Paulo, São Paulo, Brazil] Fullstack Developer (Java) (Híbrido - São Paulo) na Coodesh - ## Descrição da vaga: Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios. Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/jobs/consultant-java-developer-hibrida-sao-paulo-172418087?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋 <p>A <strong>KLB Group</strong> busca <strong><ins>Fullstack Developer (Java)</ins></strong> para compor seu time!</p> <p></p> <p>A KLB Group é especializada na implementação de projetos em empresas públicas e privadas. Seja em projetos de desenvolvimento, produção ou transformação, a KLB Group assegura a implementação eficaz, mobilizando rapidamente uma equipe de especialistas de várias funções (compras, cadeia de suprimentos, qualidade, engenharia, TI, finanças, etc.), com uma combinação única de experiência em design, implementação e operação. A KLB Group tem mais de 500 funcionários na Europa, América e Ásia.</p> <p></p> <p><strong>Atividades e Atribuições: </strong></p> <ul> <li>&nbsp;Realizar atendimentos das demandas dos Portais de Sinistros;</li> <li>&nbsp;Analisar as demandas visando atender às necessidades dos clientes;</li> <li>&nbsp;Levantamento de requisitos;</li> <li>Prever impactos das alteraçõs, garantindo a integridade do sistema.</li> </ul> ## KLB GROUP BRASIL : <p>A KLB Group é especializada na implementação de projetos em empresas públicas e privadas. Seja em projetos de desenvolvimento, produção ou transformação, a KLB Group assegura a implementação eficaz, mobilizando rapidamente uma equipe de especialistas de várias funções (compras, cadeia de suprimentos, qualidade, engenharia, TI, finanças, etc.), com uma combinação única de experiência em design, implementação e operação. A KLB Group tem mais de 500 funcionários na Europa, América e Ásia.</p> </p> ## Habilidades: - Java - Front-end System Design - Back-end System Design ## Local: São Paulo, São Paulo, Brazil ## Requisitos: - Ensino Superior Completo; - Conhecimento de tecnologias para desenvolvimento Java no Backend; - Estudo da Arquitetura das aplicações para propor soluções; - Entendimento das soluções sistêmicas visando a solução do problemas ou a melhoria dos processos de negócios; - Conhecimento de tecnologias para desenvolvimento com Angular, React, entre outras no Frontend. ## Diferenciais: - Desejável conhecimento de Seguros. ## Benefícios: - Vale-refeição R$35,43 dia/ média R$779,70; - Vale-alimentação R$ 602,40; - Vale-transporte (6% de desconto); - Assistência médica/medicina em grupo; - Assistência odontológica R$10,87; - Auxílio Babá - R$ 855,16 até 6 meses; - Auxílio Creche - R$ 460,85 a partir de 7 meses até 5 anos e 11 meses; - Auxílio Babá - R$ 460,85 a partir de 7 meses até 5 anos e 11 meses; - PLR; - Seguro de vida; - Empréstimo; - Plano de previdência privada; - Programa de bolsa de estudo; - Total pass. ## Como se candidatar: Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Fullstack Developer (Java) (Híbrido - São Paulo) na KLB GROUP BRASIL ](https://coodesh.com/jobs/consultant-java-developer-hibrida-sao-paulo-172418087?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação. ## Labels #### Alocação Alocado #### Regime CLT #### Categoria Full-Stack
process
fullstack developer java híbrido são paulo na coodesh descrição da vaga esta é uma vaga de um parceiro da plataforma coodesh ao candidatar se você terá acesso as informações completas sobre a empresa e benefícios fique atento ao redirecionamento que vai te levar para uma url com o pop up personalizado de candidatura 👋 a klb group busca fullstack developer java para compor seu time a klb group é especializada na implementação de projetos em empresas públicas e privadas seja em projetos de desenvolvimento produção ou transformação a klb group assegura a implementação eficaz mobilizando rapidamente uma equipe de especialistas de várias funções compras cadeia de suprimentos qualidade engenharia ti finanças etc com uma combinação única de experiência em design implementação e operação a klb group tem mais de funcionários na europa américa e ásia atividades e atribuições nbsp realizar atendimentos das demandas dos portais de sinistros nbsp analisar as demandas visando atender às necessidades dos clientes nbsp levantamento de requisitos prever impactos das alteraçõs garantindo a integridade do sistema klb group brasil a klb group é especializada na implementação de projetos em empresas públicas e privadas seja em projetos de desenvolvimento produção ou transformação a klb group assegura a implementação eficaz mobilizando rapidamente uma equipe de especialistas de várias funções compras cadeia de suprimentos qualidade engenharia ti finanças etc com uma combinação única de experiência em design implementação e operação a klb group tem mais de funcionários na europa américa e ásia habilidades java front end system design back end system design local são paulo são paulo brazil requisitos ensino superior completo conhecimento de tecnologias para desenvolvimento java no backend estudo da arquitetura das aplicações para propor soluções entendimento das soluções sistêmicas visando a solução do problemas ou a melhoria dos processos de negócios conhecimento de tecnologias para desenvolvimento com angular react entre outras no frontend diferenciais desejável conhecimento de seguros benefícios vale refeição r dia média r vale alimentação r vale transporte de desconto assistência médica medicina em grupo assistência odontológica r auxílio babá r até meses auxílio creche r a partir de meses até anos e meses auxílio babá r a partir de meses até anos e meses plr seguro de vida empréstimo plano de previdência privada programa de bolsa de estudo total pass como se candidatar candidatar se exclusivamente através da plataforma coodesh no link a seguir após candidatar se via plataforma coodesh e validar o seu login você poderá acompanhar e receber todas as interações do processo por lá utilize a opção pedir feedback entre uma etapa e outra na vaga que se candidatou isso fará com que a pessoa recruiter responsável pelo processo na empresa receba a notificação labels alocação alocado regime clt categoria full stack
1
467,184
13,442,977,733
IssuesEvent
2020-09-08 07:39:16
teamforus/general
https://api.github.com/repos/teamforus/general
closed
Implement feedback Noordoostpolder webshop
Approval: Granted Priority: Must have Scope: Small Status: In progress Type: Change request project-89
Learn more about change requests here: https://bit.ly/39CWeEE ### Requested by: - ### Change description https://www.figma.com/file/pUWeZwelwsJK8W58ZGip9r/Webshop-Noordoostpolder?node-id=658%3A88
1.0
Implement feedback Noordoostpolder webshop - Learn more about change requests here: https://bit.ly/39CWeEE ### Requested by: - ### Change description https://www.figma.com/file/pUWeZwelwsJK8W58ZGip9r/Webshop-Noordoostpolder?node-id=658%3A88
non_process
implement feedback noordoostpolder webshop learn more about change requests here requested by change description
0
405,912
11,884,275,403
IssuesEvent
2020-03-27 17:21:24
RobotLocomotion/drake
https://api.github.com/repos/RobotLocomotion/drake
closed
Some energy and power related calculation functions are missing in MBP
priority: low team: dynamics type: bug
Functions that are missing in MBP include 1. CalcKineticEnergy() 2. CalcNonConservativePower()
1.0
Some energy and power related calculation functions are missing in MBP - Functions that are missing in MBP include 1. CalcKineticEnergy() 2. CalcNonConservativePower()
non_process
some energy and power related calculation functions are missing in mbp functions that are missing in mbp include calckineticenergy calcnonconservativepower
0
1,639
4,259,886,123
IssuesEvent
2016-07-11 12:44:09
nodejs/node
https://api.github.com/repos/nodejs/node
closed
child_process.spawn execution is significantly slower than running command directly from shell
child_process performance unconfirmed
This issue came up in Glavin001/atom-beautify#893. I have made a minimal test case for the issue: [phpcbf_spawn_test.zip](https://github.com/nodejs/node/files/356751/phpcbf_spawn_test.zip), but it requires you to have `phpcbf` from [PHP-Code-Sniffer](https://github.com/squizlabs/PHP_CodeSniffer) installed on your system. The problem is, I want to execute `phpcbf` command with some parameters. When doing it directly from shell it runs in under **1 second**: ``` > phpcbf --no-patch --standard=PSR2 test.php Changing into directory /home/arman/phpcbf_spawn_test Processing test.php [PHP => 18 tokens in 6 lines]... DONE in 3ms (0 fixable violations) Fixed 0 files Time: 49ms; Memory: 4Mb ``` When doing it in node via `child_process.spawn` it takes about **1 minute**: ``` > node beautify.js stdout: Changing into directory /home/arman/phpcbf_spawn_test Processing test.php [PHP => 18 tokens in 6 lines]... DONE in 2ms (0 fixable violations) Fixed 0 files Time: 1 mins, 0.13 secs; Memory: 4Mb child process exited with code 1 ``` Note that the file `test.php` is minimal, it shouldn't take much time to process it. * **Version**: 6.2.1 on Linux & 6.3.0 on Mac * **Platform**: Linux 3.10.0-123.8.1.el7.x86_64 & Darwin 15.5.0 (OS X 10.11.5) * If it might be useful my php versions are: 7.0.7 on Linux & 7.0.8 on Mac
1.0
child_process.spawn execution is significantly slower than running command directly from shell - This issue came up in Glavin001/atom-beautify#893. I have made a minimal test case for the issue: [phpcbf_spawn_test.zip](https://github.com/nodejs/node/files/356751/phpcbf_spawn_test.zip), but it requires you to have `phpcbf` from [PHP-Code-Sniffer](https://github.com/squizlabs/PHP_CodeSniffer) installed on your system. The problem is, I want to execute `phpcbf` command with some parameters. When doing it directly from shell it runs in under **1 second**: ``` > phpcbf --no-patch --standard=PSR2 test.php Changing into directory /home/arman/phpcbf_spawn_test Processing test.php [PHP => 18 tokens in 6 lines]... DONE in 3ms (0 fixable violations) Fixed 0 files Time: 49ms; Memory: 4Mb ``` When doing it in node via `child_process.spawn` it takes about **1 minute**: ``` > node beautify.js stdout: Changing into directory /home/arman/phpcbf_spawn_test Processing test.php [PHP => 18 tokens in 6 lines]... DONE in 2ms (0 fixable violations) Fixed 0 files Time: 1 mins, 0.13 secs; Memory: 4Mb child process exited with code 1 ``` Note that the file `test.php` is minimal, it shouldn't take much time to process it. * **Version**: 6.2.1 on Linux & 6.3.0 on Mac * **Platform**: Linux 3.10.0-123.8.1.el7.x86_64 & Darwin 15.5.0 (OS X 10.11.5) * If it might be useful my php versions are: 7.0.7 on Linux & 7.0.8 on Mac
process
child process spawn execution is significantly slower than running command directly from shell this issue came up in atom beautify i have made a minimal test case for the issue but it requires you to have phpcbf from installed on your system the problem is i want to execute phpcbf command with some parameters when doing it directly from shell it runs in under second phpcbf no patch standard test php changing into directory home arman phpcbf spawn test processing test php done in fixable violations fixed files time memory when doing it in node via child process spawn it takes about minute node beautify js stdout changing into directory home arman phpcbf spawn test processing test php done in fixable violations fixed files time mins secs memory child process exited with code note that the file test php is minimal it shouldn t take much time to process it version on linux on mac platform linux darwin os x if it might be useful my php versions are on linux on mac
1
47,412
12,031,500,958
IssuesEvent
2020-04-13 09:49:27
tensorflow/tensorflow
https://api.github.com/repos/tensorflow/tensorflow
opened
ERROR: Config value download_clang is not defined in any .rc file
type:build/install
tensorflow:r2.2 bazel :2.0.0 ERROR: Config value download_clang is not defined in any .rc file print info: `bazel build -c opt //tensorflow/contrib/android:libtensorflow_inference.so --crosstool_top=//external:android/crosstool --host_crosstool_top=@bazel_tools//tools/cpp:toolchain --cpu=armeabi-v7a INFO: Options provided by the client: Inherited 'common' options: --isatty=1 --terminal_columns=80 INFO: Reading rc options for 'build' from /home/wushengqi/ai/tensorflow/.bazelrc: Inherited 'common' options: --experimental_repo_remote_exec INFO: Reading rc options for 'build' from /home/wushengqi/ai/tensorflow/.bazelrc: 'build' options: --apple_platform_type=macos --define framework_shared_object=true --define open_source_build=true --java_toolchain=//third_party/toolchains/java:tf_java_toolchain --host_java_toolchain=//third_party/toolchains/java:tf_java_toolchain --define=use_fast_cpp_protos=true --define=allow_oversize_protos=true --spawn_strategy=standalone -c opt --announce_rc --define=grpc_no_ares=true --noincompatible_remove_legacy_whole_archive --noincompatible_prohibit_aapt1 --enable_platform_specific_config --config=v2 INFO: Reading rc options for 'build' from /home/wushengqi/ai/tensorflow/.tf_configure.bazelrc: 'build' options: --action_env PYTHON_BIN_PATH=/home/wushengqi/anaconda3/bin/python --action_env PYTHON_LIB_PATH=/home/wushengqi/anaconda3/lib/python3.7/site-packages --python_path=/home/wushengqi/anaconda3/bin/python --config=xla --config=rocm --config=download_clang --action_env ANDROID_NDK_HOME=/home/wushengqi/android/android-ndk-r14b --action_env ANDROID_NDK_API_LEVEL=19 --action_env ANDROID_BUILD_TOOLS_VERSION=29.0.2 --action_env ANDROID_SDK_API_LEVEL=29 --action_env ANDROID_SDK_HOME=/home/wushengqi/android/Sdk --action_env TF_CONFIGURE_IOS=0 INFO: Found applicable config definition build:v2 in file /home/wushengqi/ai/tensorflow/.bazelrc: --define=tf_api_version=2 --action_env=TF2_BEHAVIOR=1 INFO: Found applicable config definition build:xla in file /home/wushengqi/ai/tensorflow/.bazelrc: --action_env=TF_ENABLE_XLA=1 --define=with_xla_support=true INFO: Found applicable config definition build:rocm in file /home/wushengqi/ai/tensorflow/.bazelrc: --crosstool_top=@local_config_rocm//crosstool:toolchain --define=using_rocm=true --define=using_rocm_hipcc=true --action_env TF_NEED_ROCM=1 ERROR: Config value download_clang is not defined in any .rc file ` If I'm using a lower version of bazel(1.2.1): `bazel build -c opt //tensorflow/contrib/android:libtensorflow_inference.so --crosstool_top=//external:android/crosstool --host_crosstool_top=@bazel_tools//tools/cpp:toolchain --cpu=armeabi-v7a INFO: Options provided by the client: Inherited 'common' options: --isatty=1 --terminal_columns=80 INFO: Reading rc options for 'build' from /home/wushengqi/ai/tensorflow/.bazelrc: Inherited 'common' options: --experimental_repo_remote_exec ERROR: Unrecognized option: --experimental_repo_remote_exec `
1.0
ERROR: Config value download_clang is not defined in any .rc file - tensorflow:r2.2 bazel :2.0.0 ERROR: Config value download_clang is not defined in any .rc file print info: `bazel build -c opt //tensorflow/contrib/android:libtensorflow_inference.so --crosstool_top=//external:android/crosstool --host_crosstool_top=@bazel_tools//tools/cpp:toolchain --cpu=armeabi-v7a INFO: Options provided by the client: Inherited 'common' options: --isatty=1 --terminal_columns=80 INFO: Reading rc options for 'build' from /home/wushengqi/ai/tensorflow/.bazelrc: Inherited 'common' options: --experimental_repo_remote_exec INFO: Reading rc options for 'build' from /home/wushengqi/ai/tensorflow/.bazelrc: 'build' options: --apple_platform_type=macos --define framework_shared_object=true --define open_source_build=true --java_toolchain=//third_party/toolchains/java:tf_java_toolchain --host_java_toolchain=//third_party/toolchains/java:tf_java_toolchain --define=use_fast_cpp_protos=true --define=allow_oversize_protos=true --spawn_strategy=standalone -c opt --announce_rc --define=grpc_no_ares=true --noincompatible_remove_legacy_whole_archive --noincompatible_prohibit_aapt1 --enable_platform_specific_config --config=v2 INFO: Reading rc options for 'build' from /home/wushengqi/ai/tensorflow/.tf_configure.bazelrc: 'build' options: --action_env PYTHON_BIN_PATH=/home/wushengqi/anaconda3/bin/python --action_env PYTHON_LIB_PATH=/home/wushengqi/anaconda3/lib/python3.7/site-packages --python_path=/home/wushengqi/anaconda3/bin/python --config=xla --config=rocm --config=download_clang --action_env ANDROID_NDK_HOME=/home/wushengqi/android/android-ndk-r14b --action_env ANDROID_NDK_API_LEVEL=19 --action_env ANDROID_BUILD_TOOLS_VERSION=29.0.2 --action_env ANDROID_SDK_API_LEVEL=29 --action_env ANDROID_SDK_HOME=/home/wushengqi/android/Sdk --action_env TF_CONFIGURE_IOS=0 INFO: Found applicable config definition build:v2 in file /home/wushengqi/ai/tensorflow/.bazelrc: --define=tf_api_version=2 --action_env=TF2_BEHAVIOR=1 INFO: Found applicable config definition build:xla in file /home/wushengqi/ai/tensorflow/.bazelrc: --action_env=TF_ENABLE_XLA=1 --define=with_xla_support=true INFO: Found applicable config definition build:rocm in file /home/wushengqi/ai/tensorflow/.bazelrc: --crosstool_top=@local_config_rocm//crosstool:toolchain --define=using_rocm=true --define=using_rocm_hipcc=true --action_env TF_NEED_ROCM=1 ERROR: Config value download_clang is not defined in any .rc file ` If I'm using a lower version of bazel(1.2.1): `bazel build -c opt //tensorflow/contrib/android:libtensorflow_inference.so --crosstool_top=//external:android/crosstool --host_crosstool_top=@bazel_tools//tools/cpp:toolchain --cpu=armeabi-v7a INFO: Options provided by the client: Inherited 'common' options: --isatty=1 --terminal_columns=80 INFO: Reading rc options for 'build' from /home/wushengqi/ai/tensorflow/.bazelrc: Inherited 'common' options: --experimental_repo_remote_exec ERROR: Unrecognized option: --experimental_repo_remote_exec `
non_process
error config value download clang is not defined in any rc file tensorflow: bazel : error config value download clang is not defined in any rc file print info: bazel build c opt tensorflow contrib android libtensorflow inference so crosstool top external android crosstool host crosstool top bazel tools tools cpp toolchain cpu armeabi info options provided by the client inherited common options isatty terminal columns info reading rc options for build from home wushengqi ai tensorflow bazelrc inherited common options experimental repo remote exec info reading rc options for build from home wushengqi ai tensorflow bazelrc build options apple platform type macos define framework shared object true define open source build true java toolchain third party toolchains java tf java toolchain host java toolchain third party toolchains java tf java toolchain define use fast cpp protos true define allow oversize protos true spawn strategy standalone c opt announce rc define grpc no ares true noincompatible remove legacy whole archive noincompatible prohibit enable platform specific config config info reading rc options for build from home wushengqi ai tensorflow tf configure bazelrc build options action env python bin path home wushengqi bin python action env python lib path home wushengqi lib site packages python path home wushengqi bin python config xla config rocm config download clang action env android ndk home home wushengqi android android ndk action env android ndk api level action env android build tools version action env android sdk api level action env android sdk home home wushengqi android sdk action env tf configure ios info found applicable config definition build in file home wushengqi ai tensorflow bazelrc define tf api version action env behavior info found applicable config definition build xla in file home wushengqi ai tensorflow bazelrc action env tf enable xla define with xla support true info found applicable config definition build rocm in file home wushengqi ai tensorflow bazelrc crosstool top local config rocm crosstool toolchain define using rocm true define using rocm hipcc true action env tf need rocm error config value download clang is not defined in any rc file if i m using a lower version of bazel( ): bazel build c opt tensorflow contrib android libtensorflow inference so crosstool top external android crosstool host crosstool top bazel tools tools cpp toolchain cpu armeabi info options provided by the client inherited common options isatty terminal columns info reading rc options for build from home wushengqi ai tensorflow bazelrc inherited common options experimental repo remote exec error unrecognized option experimental repo remote exec
0
5,596
8,453,014,344
IssuesEvent
2018-10-20 11:13:40
SerialLain3170/GAN-papers
https://api.github.com/repos/SerialLain3170/GAN-papers
opened
Parallel-Data-Free Voice Conversion Using Cycle-Consistent Adversarial Networks
Speech Processing
# Paper [Parallel-Data-Free Voice Conversion Using Cycle-Consistent Adversarial Networks](https://arxiv.org/pdf/1711.11293.pdf) # Summary - Voice ConversionにCycleGANを適用、Adversarial lossとCycle consistency lossに加え、identity mapping lossも考慮している。 - メルケプストラム24次元を変換、基本周波数は線形変換。 - Network ArchitectureにはGated CNNを適用 ![screenshot from 2018-10-20 20-09-27](https://user-images.githubusercontent.com/32360147/47254993-8a035780-d4a4-11e8-8a8b-0f0fd18e9157.png) # Date 2017/11/30
1.0
Parallel-Data-Free Voice Conversion Using Cycle-Consistent Adversarial Networks - # Paper [Parallel-Data-Free Voice Conversion Using Cycle-Consistent Adversarial Networks](https://arxiv.org/pdf/1711.11293.pdf) # Summary - Voice ConversionにCycleGANを適用、Adversarial lossとCycle consistency lossに加え、identity mapping lossも考慮している。 - メルケプストラム24次元を変換、基本周波数は線形変換。 - Network ArchitectureにはGated CNNを適用 ![screenshot from 2018-10-20 20-09-27](https://user-images.githubusercontent.com/32360147/47254993-8a035780-d4a4-11e8-8a8b-0f0fd18e9157.png) # Date 2017/11/30
process
parallel data free voice conversion using cycle consistent adversarial networks paper summary voice conversionにcycleganを適用、adversarial lossとcycle consistency lossに加え、identity mapping lossも考慮している。 、基本周波数は線形変換。 network architectureにはgated cnnを適用 date
1
14,053
16,857,301,031
IssuesEvent
2021-06-21 08:30:15
aiidateam/aiida-core
https://api.github.com/repos/aiidateam/aiida-core
closed
`ProcessBuilder` looses `AttributeDict` properties of nested namespaces when `_update` is used with nested dictionary
priority/important topic/processes type/bug
When a process builder is updated with a nested dictionary using the `_update` method, the nested namespaces are overwritten by plain dictionaries instead of `ProcessBuilderNamespace` instances, causing the nested namespaces to loose their functionality, such as validation and tab-completion. The solution is to properly recursively apply values of the dictionary in `_update` just setting the values in the namespace instead of overwriting it completely.
1.0
`ProcessBuilder` looses `AttributeDict` properties of nested namespaces when `_update` is used with nested dictionary - When a process builder is updated with a nested dictionary using the `_update` method, the nested namespaces are overwritten by plain dictionaries instead of `ProcessBuilderNamespace` instances, causing the nested namespaces to loose their functionality, such as validation and tab-completion. The solution is to properly recursively apply values of the dictionary in `_update` just setting the values in the namespace instead of overwriting it completely.
process
processbuilder looses attributedict properties of nested namespaces when update is used with nested dictionary when a process builder is updated with a nested dictionary using the update method the nested namespaces are overwritten by plain dictionaries instead of processbuildernamespace instances causing the nested namespaces to loose their functionality such as validation and tab completion the solution is to properly recursively apply values of the dictionary in update just setting the values in the namespace instead of overwriting it completely
1
16,148
20,507,379,119
IssuesEvent
2022-03-01 00:16:39
newrelic/docs-website
https://api.github.com/repos/newrelic/docs-website
closed
Process council: PM handbook updates
process
# Overview We are going to roll out a process council in the next week or so. In the spirit of such, we should follow our proposed council process for these changes. # Proposal The process for a change will be: 1. Someone has an idea for a process change or experiment. This can also just be identifying the issue without a solution. 2. They file and bring it up in Slack and file an issue. 3. Anyone interested in solving the issue can jump into the GH issue to collaborate. One person should be designated the Assignee (Driver). 4. Once the cohort is satisfied with the solution/process, the driver opens a PR against the PM handbook or style guide. This PR should explain, in detail, what the new process is. 5. The driver alerts the team in a Slack channel or team meeting about the proposed change. The team has 7 days to collaborate in the PR and give their feedback and thoughts. 6. After 7 days, the PR is either merged in and an official process, or closed to iterate on in the future. # Action items - [x] Collaborate on this issue to finalize a process - [x] Open a PR against the PM handbook/style guide - [x] #5121 - [ ] Merge in and inform the team.
1.0
Process council: PM handbook updates - # Overview We are going to roll out a process council in the next week or so. In the spirit of such, we should follow our proposed council process for these changes. # Proposal The process for a change will be: 1. Someone has an idea for a process change or experiment. This can also just be identifying the issue without a solution. 2. They file and bring it up in Slack and file an issue. 3. Anyone interested in solving the issue can jump into the GH issue to collaborate. One person should be designated the Assignee (Driver). 4. Once the cohort is satisfied with the solution/process, the driver opens a PR against the PM handbook or style guide. This PR should explain, in detail, what the new process is. 5. The driver alerts the team in a Slack channel or team meeting about the proposed change. The team has 7 days to collaborate in the PR and give their feedback and thoughts. 6. After 7 days, the PR is either merged in and an official process, or closed to iterate on in the future. # Action items - [x] Collaborate on this issue to finalize a process - [x] Open a PR against the PM handbook/style guide - [x] #5121 - [ ] Merge in and inform the team.
process
process council pm handbook updates overview we are going to roll out a process council in the next week or so in the spirit of such we should follow our proposed council process for these changes proposal the process for a change will be someone has an idea for a process change or experiment this can also just be identifying the issue without a solution they file and bring it up in slack and file an issue anyone interested in solving the issue can jump into the gh issue to collaborate one person should be designated the assignee driver once the cohort is satisfied with the solution process the driver opens a pr against the pm handbook or style guide this pr should explain in detail what the new process is the driver alerts the team in a slack channel or team meeting about the proposed change the team has days to collaborate in the pr and give their feedback and thoughts after days the pr is either merged in and an official process or closed to iterate on in the future action items collaborate on this issue to finalize a process open a pr against the pm handbook style guide merge in and inform the team
1
1,852
4,651,283,230
IssuesEvent
2016-10-03 09:30:37
openvstorage/volumedriver
https://api.github.com/repos/openvstorage/volumedriver
closed
Quickly removing and recreating a volume with the same name triggers a fault
process_wontfix
Disks of routeros virtual machines in OVC are being created with a specific naming convention. It happens that these routeros virtual machines are quickly recycled, and this triggers a race condition in the volumedriver. The result is that the "recreated" volume is left in a zombie state, that can only be solved by expert ops to clean it up.
1.0
Quickly removing and recreating a volume with the same name triggers a fault - Disks of routeros virtual machines in OVC are being created with a specific naming convention. It happens that these routeros virtual machines are quickly recycled, and this triggers a race condition in the volumedriver. The result is that the "recreated" volume is left in a zombie state, that can only be solved by expert ops to clean it up.
process
quickly removing and recreating a volume with the same name triggers a fault disks of routeros virtual machines in ovc are being created with a specific naming convention it happens that these routeros virtual machines are quickly recycled and this triggers a race condition in the volumedriver the result is that the recreated volume is left in a zombie state that can only be solved by expert ops to clean it up
1
22,578
31,805,302,396
IssuesEvent
2023-09-13 13:37:41
GSA/EDX
https://api.github.com/repos/GSA/EDX
opened
Update personal access token for GitHub Workflow (December 2023)
process
For the EDXPROJECT_TOKEN to automate the issue workflow (adding it to EDX's Inbox in its Kanban board) Instructions: Click on your user icon at the top right Click settings Scroll to bottom, click "Developer Settings" Under personal access tokens, click tokens classic You want to update the EDXPROJECT_TOKEN one with your updated API Key
1.0
Update personal access token for GitHub Workflow (December 2023) - For the EDXPROJECT_TOKEN to automate the issue workflow (adding it to EDX's Inbox in its Kanban board) Instructions: Click on your user icon at the top right Click settings Scroll to bottom, click "Developer Settings" Under personal access tokens, click tokens classic You want to update the EDXPROJECT_TOKEN one with your updated API Key
process
update personal access token for github workflow december for the edxproject token to automate the issue workflow adding it to edx s inbox in its kanban board instructions click on your user icon at the top right click settings scroll to bottom click developer settings under personal access tokens click tokens classic you want to update the edxproject token one with your updated api key
1
269,267
8,433,952,195
IssuesEvent
2018-10-17 08:51:54
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
m2.wyylde.com - see bug description
browser-firefox-mobile browser-focus-geckoview priority-normal
<!-- @browser: Firefox Mobile 62.0 --> <!-- @ua_header: Mozilla/5.0 (Android 8.1.0; Mobile; rv:62.0) Gecko/62.0 Firefox/62.0 --> <!-- @reported_with: --> <!-- @extra_labels: browser-focus-geckoview --> **URL**: https://m2.wyylde.com/ **Browser / Version**: Firefox Mobile 62.0 **Operating System**: Android 8.1.0 **Tested Another Browser**: Yes **Problem type**: Something else **Description**: The site won't showing up although I don't need it. **Steps to Reproduce**: _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
m2.wyylde.com - see bug description - <!-- @browser: Firefox Mobile 62.0 --> <!-- @ua_header: Mozilla/5.0 (Android 8.1.0; Mobile; rv:62.0) Gecko/62.0 Firefox/62.0 --> <!-- @reported_with: --> <!-- @extra_labels: browser-focus-geckoview --> **URL**: https://m2.wyylde.com/ **Browser / Version**: Firefox Mobile 62.0 **Operating System**: Android 8.1.0 **Tested Another Browser**: Yes **Problem type**: Something else **Description**: The site won't showing up although I don't need it. **Steps to Reproduce**: _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
wyylde com see bug description url browser version firefox mobile operating system android tested another browser yes problem type something else description the site won t showing up although i don t need it steps to reproduce from with ❤️
0
805,108
29,508,052,751
IssuesEvent
2023-06-03 15:08:22
jrsteensen/OpenHornet
https://api.github.com/repos/jrsteensen/OpenHornet
closed
[Bug]: UIP parts with same part number: OH1A1-28
Type: Bug/Obsolesce Category: MCAD Priority: Normal
### Discord Username Arribe ### Bug Summary On the UIP part number OH1A1-28 is used for both the UPPER UIP FACE MOUNT, RIGHT, and UIP RIGHT GLARESHIELD. ### Expected Results Parts should have their own part numbers. If possible please keep UPPER UIP FACE MOUNT, RIGHT associated with OH1A1-28 as that was used to generate the 3mf file. ### Actual Results Same part number used for 2 different parts. ### Screenshots/Images/Files ![image](https://github.com/jrsteensen/OpenHornet/assets/47125808/e0ca7d2f-8643-4da9-b776-ccd25800c271) ### Applicable Part Numbers OH1A1-28 ### Release Version 1.0.0-beta.1 ### Category Mechanical (Structure/Panels/Mechanisms) ### Applicable End Item(s) Upper Instrument Panel (UIP) ### Built to print? - [X] I built (or attempted to build) the part to the OpenHornet print without any deviations. - [ ] I am not building this part to the OH print. (List deviations in detail in the Miscellaneous Info text area below.) ### Miscellaneous Info Discovered during manufacturing file generation when comparing the release drawing to the list of exported files.
1.0
[Bug]: UIP parts with same part number: OH1A1-28 - ### Discord Username Arribe ### Bug Summary On the UIP part number OH1A1-28 is used for both the UPPER UIP FACE MOUNT, RIGHT, and UIP RIGHT GLARESHIELD. ### Expected Results Parts should have their own part numbers. If possible please keep UPPER UIP FACE MOUNT, RIGHT associated with OH1A1-28 as that was used to generate the 3mf file. ### Actual Results Same part number used for 2 different parts. ### Screenshots/Images/Files ![image](https://github.com/jrsteensen/OpenHornet/assets/47125808/e0ca7d2f-8643-4da9-b776-ccd25800c271) ### Applicable Part Numbers OH1A1-28 ### Release Version 1.0.0-beta.1 ### Category Mechanical (Structure/Panels/Mechanisms) ### Applicable End Item(s) Upper Instrument Panel (UIP) ### Built to print? - [X] I built (or attempted to build) the part to the OpenHornet print without any deviations. - [ ] I am not building this part to the OH print. (List deviations in detail in the Miscellaneous Info text area below.) ### Miscellaneous Info Discovered during manufacturing file generation when comparing the release drawing to the list of exported files.
non_process
uip parts with same part number discord username arribe bug summary on the uip part number is used for both the upper uip face mount right and uip right glareshield expected results parts should have their own part numbers if possible please keep upper uip face mount right associated with as that was used to generate the file actual results same part number used for different parts screenshots images files applicable part numbers release version beta category mechanical structure panels mechanisms applicable end item s upper instrument panel uip built to print i built or attempted to build the part to the openhornet print without any deviations i am not building this part to the oh print list deviations in detail in the miscellaneous info text area below miscellaneous info discovered during manufacturing file generation when comparing the release drawing to the list of exported files
0
16,484
21,443,566,330
IssuesEvent
2022-04-25 02:07:19
huutho77/CNPMNC_ThayAi
https://api.github.com/repos/huutho77/CNPMNC_ThayAi
opened
[Browser UI] Fearure Filter Products based on Category
processing dev/quocky2211 dev/haichao784 dev/phamtan
- Show category list - Get categoryID or categoryName - Filter related products based on Category - Show the products list from the result
1.0
[Browser UI] Fearure Filter Products based on Category - - Show category list - Get categoryID or categoryName - Filter related products based on Category - Show the products list from the result
process
fearure filter products based on category show category list get categoryid or categoryname filter related products based on category show the products list from the result
1
12,471
14,940,246,660
IssuesEvent
2021-01-25 18:00:36
eddieantonio/predictive-text-studio
https://api.github.com/repos/eddieantonio/predictive-text-studio
closed
Remove headers from uploaded wordlists via the Google Sheets API
data-backing data-processing good first issue 🔥 High priority
An extension of issue #199 > Remove headers such as "word" and "count" headers in the first row of all uploaded wordlists (either Google Sheets or Excel). > > This should be done automatically. Add functionality from #217 to the Google Sheets API. Code reuse is ideal.
1.0
Remove headers from uploaded wordlists via the Google Sheets API - An extension of issue #199 > Remove headers such as "word" and "count" headers in the first row of all uploaded wordlists (either Google Sheets or Excel). > > This should be done automatically. Add functionality from #217 to the Google Sheets API. Code reuse is ideal.
process
remove headers from uploaded wordlists via the google sheets api an extension of issue remove headers such as word and count headers in the first row of all uploaded wordlists either google sheets or excel this should be done automatically add functionality from to the google sheets api code reuse is ideal
1
50,204
3,006,246,791
IssuesEvent
2015-07-27 09:08:52
Itseez/opencv
https://api.github.com/repos/Itseez/opencv
opened
Add new create() method for Feature2D
auto-transferred category: features2d feature priority: normal
Transferred from http://code.opencv.org/issues/2333 ``` || Maria Dimashova on 2012-09-05 09:58 || Priority: Normal || Affected: None || Category: features2d || Tracker: Feature || Difficulty: None || PR: None || Platform: None / None ``` Add new create() method for Feature2D ----------- ``` with arguments Ptr<FeatureDetector> and Ptr<DescriptorExtractor>. ``` History ------- ##### Alexander Shishkov on 2012-09-07 13:32 ``` - Target version deleted (3.0) - Assignee deleted (Maria Dimashova) ```
1.0
Add new create() method for Feature2D - Transferred from http://code.opencv.org/issues/2333 ``` || Maria Dimashova on 2012-09-05 09:58 || Priority: Normal || Affected: None || Category: features2d || Tracker: Feature || Difficulty: None || PR: None || Platform: None / None ``` Add new create() method for Feature2D ----------- ``` with arguments Ptr<FeatureDetector> and Ptr<DescriptorExtractor>. ``` History ------- ##### Alexander Shishkov on 2012-09-07 13:32 ``` - Target version deleted (3.0) - Assignee deleted (Maria Dimashova) ```
non_process
add new create method for transferred from maria dimashova on priority normal affected none category tracker feature difficulty none pr none platform none none add new create method for with arguments ptr and ptr history alexander shishkov on target version deleted assignee deleted maria dimashova
0
59,973
6,668,407,844
IssuesEvent
2017-10-03 15:41:58
irvnet/cka-study-group
https://api.github.com/repos/irvnet/cka-study-group
opened
Write a storyboard for building an environment
test-domain:install-config-validate
The storyboard for the environment should describe the prep for building an environment to install K8s... locally or in the cloud. The chosen environment should be viable for single-node and multi-node clusters.
1.0
Write a storyboard for building an environment - The storyboard for the environment should describe the prep for building an environment to install K8s... locally or in the cloud. The chosen environment should be viable for single-node and multi-node clusters.
non_process
write a storyboard for building an environment the storyboard for the environment should describe the prep for building an environment to install locally or in the cloud the chosen environment should be viable for single node and multi node clusters
0
776
3,258,211,568
IssuesEvent
2015-10-20 21:15:10
hammerlab/pileup.js
https://api.github.com/repos/hammerlab/pileup.js
closed
"npm run watch" sometimes fails to update tests
process
I've had a hard time finding a reliable repro, but it's happened often enough that killing the task and re-running it is part of my usual debugging process. The ideal solution here might be to switch from browserify/jstransform/watchify to webpack/babel.
1.0
"npm run watch" sometimes fails to update tests - I've had a hard time finding a reliable repro, but it's happened often enough that killing the task and re-running it is part of my usual debugging process. The ideal solution here might be to switch from browserify/jstransform/watchify to webpack/babel.
process
npm run watch sometimes fails to update tests i ve had a hard time finding a reliable repro but it s happened often enough that killing the task and re running it is part of my usual debugging process the ideal solution here might be to switch from browserify jstransform watchify to webpack babel
1
229,003
17,497,071,348
IssuesEvent
2021-08-10 02:54:37
VikeLabs/courseup
https://api.github.com/repos/VikeLabs/courseup
closed
Community Contributions Documentation
documentation
Add documentation to the appropriate places (README etc.) about how to contribute code to this project.
1.0
Community Contributions Documentation - Add documentation to the appropriate places (README etc.) about how to contribute code to this project.
non_process
community contributions documentation add documentation to the appropriate places readme etc about how to contribute code to this project
0
22,533
31,682,208,403
IssuesEvent
2023-09-08 01:30:42
turt2live/matrix-media-repo
https://api.github.com/repos/turt2live/matrix-media-repo
closed
Refactoring checklist
enhancement media import release-blocker media export url previews multi-process datastores files antispam resource waste spec compliance performance transfer admin api gdpr
* [x] Update scripts: * [x] `compile_assets` * [x] `export_synapse_for_import` (related to gdpr_export) * [x] `gdpr_export` * [x] `gdpr_import` * [x] `import_synapse` * [x] `media_repo` (startup routine maintenance?) * [x] `plugin_antispam_ocr` * [x] `s3_consistency_check` * [x] Move URL previews to pipeline system * [x] Update "custom" endpoint handlers * [x] `datastores.go` * [x] `exports.go` * [x] `federation.go` * [x] `health.go` * [x] `imports.go` * [x] `media_attributes.go` * [x] `purge.go` * [x] `quarantine.go` * [x] `tasks.go` * [x] `usage.go` * [x] `version.go` * [x] Update "r0" endpoint handlers * [x] `download.go` * [x] `identicon.go` * [x] `logout.go` * [x] `preview_url.go` * [x] `public_config.go` * [x] `thumbnail.go` * [x] `upload_async.go` * [x] `upload_sync.go` * [x] Update "unstable" endpoint handlers * [x] `info.go` * [x] `local_copy.go` * [x] `public_usage.go` * [x] Update "v1" endpoint handlers * [x] `create.go` * [x] Make plugins work again * [x] Fix background and recurring tasks * [x] Validate auth works as expected (repo/local admin, per-user stuff) * [x] Delete dead code * [ ] Integration/end-to-end tests (all with workers) * [x] https://github.com/turt2live/matrix-media-repo/issues/41 * [ ] https://github.com/turt2live/matrix-media-repo/issues/430 * [ ] https://github.com/turt2live/matrix-media-repo/issues/431 * [ ] https://github.com/turt2live/matrix-media-repo/issues/432 * [ ] https://github.com/turt2live/matrix-media-repo/issues/433 * [ ] https://github.com/turt2live/matrix-media-repo/issues/434 * [ ] https://github.com/turt2live/matrix-media-repo/issues/435 * [ ] https://github.com/turt2live/matrix-media-repo/issues/436 * [ ] https://github.com/turt2live/matrix-media-repo/issues/437 * [ ] https://github.com/turt2live/matrix-media-repo/issues/438 * [x] https://github.com/turt2live/matrix-media-repo/issues/15 * [x] https://github.com/turt2live/matrix-media-repo/issues/426 * [x] https://github.com/turt2live/matrix-media-repo/issues/425 * [x] https://github.com/turt2live/matrix-media-repo/issues/439 * [x] https://github.com/turt2live/matrix-media-repo/issues/440 * [x] https://github.com/turt2live/matrix-media-repo/issues/441 * [x] https://github.com/turt2live/matrix-media-repo/issues/442 * [x] https://github.com/turt2live/matrix-media-repo/issues/443 * [x] https://github.com/turt2live/matrix-media-repo/issues/444 * [x] https://github.com/turt2live/matrix-media-repo/issues/445 * [x] https://github.com/turt2live/matrix-media-repo/issues/446 * [x] https://github.com/turt2live/matrix-media-repo/issues/447 * [ ] https://github.com/turt2live/matrix-media-repo/issues/448 * [x] https://github.com/turt2live/matrix-media-repo/issues/449 * [x] https://github.com/turt2live/matrix-media-repo/issues/450 * [x] https://github.com/turt2live/matrix-media-repo/issues/451 * [x] https://github.com/turt2live/matrix-media-repo/issues/457 * [x] Built-in hooks for https://github.com/turt2live/matrix-media-repo/issues/428 * [x] Documentation * [x] Remove warning from README After release: * [ ] https://github.com/turt2live/matrix-media-repo/issues/458
1.0
Refactoring checklist - * [x] Update scripts: * [x] `compile_assets` * [x] `export_synapse_for_import` (related to gdpr_export) * [x] `gdpr_export` * [x] `gdpr_import` * [x] `import_synapse` * [x] `media_repo` (startup routine maintenance?) * [x] `plugin_antispam_ocr` * [x] `s3_consistency_check` * [x] Move URL previews to pipeline system * [x] Update "custom" endpoint handlers * [x] `datastores.go` * [x] `exports.go` * [x] `federation.go` * [x] `health.go` * [x] `imports.go` * [x] `media_attributes.go` * [x] `purge.go` * [x] `quarantine.go` * [x] `tasks.go` * [x] `usage.go` * [x] `version.go` * [x] Update "r0" endpoint handlers * [x] `download.go` * [x] `identicon.go` * [x] `logout.go` * [x] `preview_url.go` * [x] `public_config.go` * [x] `thumbnail.go` * [x] `upload_async.go` * [x] `upload_sync.go` * [x] Update "unstable" endpoint handlers * [x] `info.go` * [x] `local_copy.go` * [x] `public_usage.go` * [x] Update "v1" endpoint handlers * [x] `create.go` * [x] Make plugins work again * [x] Fix background and recurring tasks * [x] Validate auth works as expected (repo/local admin, per-user stuff) * [x] Delete dead code * [ ] Integration/end-to-end tests (all with workers) * [x] https://github.com/turt2live/matrix-media-repo/issues/41 * [ ] https://github.com/turt2live/matrix-media-repo/issues/430 * [ ] https://github.com/turt2live/matrix-media-repo/issues/431 * [ ] https://github.com/turt2live/matrix-media-repo/issues/432 * [ ] https://github.com/turt2live/matrix-media-repo/issues/433 * [ ] https://github.com/turt2live/matrix-media-repo/issues/434 * [ ] https://github.com/turt2live/matrix-media-repo/issues/435 * [ ] https://github.com/turt2live/matrix-media-repo/issues/436 * [ ] https://github.com/turt2live/matrix-media-repo/issues/437 * [ ] https://github.com/turt2live/matrix-media-repo/issues/438 * [x] https://github.com/turt2live/matrix-media-repo/issues/15 * [x] https://github.com/turt2live/matrix-media-repo/issues/426 * [x] https://github.com/turt2live/matrix-media-repo/issues/425 * [x] https://github.com/turt2live/matrix-media-repo/issues/439 * [x] https://github.com/turt2live/matrix-media-repo/issues/440 * [x] https://github.com/turt2live/matrix-media-repo/issues/441 * [x] https://github.com/turt2live/matrix-media-repo/issues/442 * [x] https://github.com/turt2live/matrix-media-repo/issues/443 * [x] https://github.com/turt2live/matrix-media-repo/issues/444 * [x] https://github.com/turt2live/matrix-media-repo/issues/445 * [x] https://github.com/turt2live/matrix-media-repo/issues/446 * [x] https://github.com/turt2live/matrix-media-repo/issues/447 * [ ] https://github.com/turt2live/matrix-media-repo/issues/448 * [x] https://github.com/turt2live/matrix-media-repo/issues/449 * [x] https://github.com/turt2live/matrix-media-repo/issues/450 * [x] https://github.com/turt2live/matrix-media-repo/issues/451 * [x] https://github.com/turt2live/matrix-media-repo/issues/457 * [x] Built-in hooks for https://github.com/turt2live/matrix-media-repo/issues/428 * [x] Documentation * [x] Remove warning from README After release: * [ ] https://github.com/turt2live/matrix-media-repo/issues/458
process
refactoring checklist update scripts compile assets export synapse for import related to gdpr export gdpr export gdpr import import synapse media repo startup routine maintenance plugin antispam ocr consistency check move url previews to pipeline system update custom endpoint handlers datastores go exports go federation go health go imports go media attributes go purge go quarantine go tasks go usage go version go update endpoint handlers download go identicon go logout go preview url go public config go thumbnail go upload async go upload sync go update unstable endpoint handlers info go local copy go public usage go update endpoint handlers create go make plugins work again fix background and recurring tasks validate auth works as expected repo local admin per user stuff delete dead code integration end to end tests all with workers built in hooks for documentation remove warning from readme after release
1
13,977
16,748,434,930
IssuesEvent
2021-06-11 18:49:44
googleapis/repo-automation-bots
https://api.github.com/repos/googleapis/repo-automation-bots
closed
Warning: `Octokit` in `@octokit/rest@18.5.4` has incompatible type
priority: p2 type: process
If one of our bots pulls `@octokit/rest@18.5.4`, it will have incompatible `Octokit` type. The easy fix is to pin `@octokit/rest@18.5.3`, but it's not a long term solution. I filed a bug on upstream: https://github.com/octokit/rest.js/issues/54 We need to establish a long term solution.
1.0
Warning: `Octokit` in `@octokit/rest@18.5.4` has incompatible type - If one of our bots pulls `@octokit/rest@18.5.4`, it will have incompatible `Octokit` type. The easy fix is to pin `@octokit/rest@18.5.3`, but it's not a long term solution. I filed a bug on upstream: https://github.com/octokit/rest.js/issues/54 We need to establish a long term solution.
process
warning octokit in octokit rest has incompatible type if one of our bots pulls octokit rest it will have incompatible octokit type the easy fix is to pin octokit rest but it s not a long term solution i filed a bug on upstream we need to establish a long term solution
1
172,101
21,031,352,417
IssuesEvent
2022-03-31 01:22:37
LalithK90/covid19_data_collector_lkr
https://api.github.com/repos/LalithK90/covid19_data_collector_lkr
opened
CVE-2022-22950 (Medium) detected in spring-expression-5.3.9.jar
security vulnerability
## CVE-2022-22950 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-expression-5.3.9.jar</b></p></summary> <p>Spring Expression Language (SpEL)</p> <p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p> <p>Path to dependency file: /build.gradle</p> <p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.springframework/spring-expression/5.3.9/f5ca763cfb9d62d196efd5d25e8daca7d555ed75/spring-expression-5.3.9.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-security-2.5.3.jar (Root Library) - spring-security-config-5.5.1.jar - spring-security-core-5.5.1.jar - :x: **spring-expression-5.3.9.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Spring Framework versions 5.3.0 - 5.3.16 and older unsupported versions, it is possible for a user to provide a specially crafted SpEL expression that may cause a denial of service condition <p>Publish Date: 2022-01-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-22950>CVE-2022-22950</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://tanzu.vmware.com/security/cve-2022-22950">https://tanzu.vmware.com/security/cve-2022-22950</a></p> <p>Release Date: 2022-01-11</p> <p>Fix Resolution: org.springframework:spring-expression:5.3.17</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-22950 (Medium) detected in spring-expression-5.3.9.jar - ## CVE-2022-22950 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-expression-5.3.9.jar</b></p></summary> <p>Spring Expression Language (SpEL)</p> <p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p> <p>Path to dependency file: /build.gradle</p> <p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.springframework/spring-expression/5.3.9/f5ca763cfb9d62d196efd5d25e8daca7d555ed75/spring-expression-5.3.9.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-security-2.5.3.jar (Root Library) - spring-security-config-5.5.1.jar - spring-security-core-5.5.1.jar - :x: **spring-expression-5.3.9.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Spring Framework versions 5.3.0 - 5.3.16 and older unsupported versions, it is possible for a user to provide a specially crafted SpEL expression that may cause a denial of service condition <p>Publish Date: 2022-01-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-22950>CVE-2022-22950</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://tanzu.vmware.com/security/cve-2022-22950">https://tanzu.vmware.com/security/cve-2022-22950</a></p> <p>Release Date: 2022-01-11</p> <p>Fix Resolution: org.springframework:spring-expression:5.3.17</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in spring expression jar cve medium severity vulnerability vulnerable library spring expression jar spring expression language spel library home page a href path to dependency file build gradle path to vulnerable library home wss scanner gradle caches modules files org springframework spring expression spring expression jar dependency hierarchy spring boot starter security jar root library spring security config jar spring security core jar x spring expression jar vulnerable library found in base branch master vulnerability details in spring framework versions and older unsupported versions it is possible for a user to provide a specially crafted spel expression that may cause a denial of service condition publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org springframework spring expression step up your open source security game with whitesource
0
13,714
16,474,775,666
IssuesEvent
2021-05-24 02:41:45
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Drape (set Z value from mesh) not working
Bug Feedback Processing stale
**Describe the bug** The Drape tool under 'Vector geometry' isn't working. It doesn't change existing Z values nor does it create new Z values from the raster as it is supposed to. It apparently does nothing. **How to Reproduce** Here is how I am running it: ![image](https://user-images.githubusercontent.com/64212883/114286922-1d87c480-9a20-11eb-8456-1162996ee24a.png) The layer Lines contains line objects, which I believe are supposed to have nodes with X, Y and Z coordinates. After running this, I am using the info tool to inspect the temporary layer created: ![image](https://user-images.githubusercontent.com/64212883/114286988-91c26800-9a20-11eb-8690-bb8fa9277053.png) As you can see, the closest vertex Z value that it found was 0. Not correct value. **QGIS and OS versions** QGIS version 3.10.0-A Coruña QGIS code revision 6c816b4204 Compiled against Qt 5.11.2 Running against Qt 5.11.2 Compiled against GDAL/OGR 2.4.1 Running against GDAL/OGR 2.4.1 Compiled against GEOS 3.8.0-CAPI-1.13.1 Running against GEOS 3.8.0-CAPI-1.13.1 Compiled against SQLite 3.29.0 Running against SQLite 3.29.0 PostgreSQL Client Version 11.5 SpatiaLite Version 4.3.0 QWT Version 6.1.3 QScintilla2 Version 2.10.8 Compiled against PROJ 5.2.0 Running against PROJ Rel. 5.2.0, September 15th, 2018 OS Version Windows 10 (10.0) Active python plugins CanadianWebServices; go2streetview; quick_map_services; SemiAutomaticClassificationPlugin; db_manager; processing **Additional context** The raster dataset was a Geosoft GXF. The starting vector dataset was a shapefile containing polygon objects. I transformed these to line objects to use with Drape. If someone wants to try debugging with my data, I can send it to them.
1.0
Drape (set Z value from mesh) not working - **Describe the bug** The Drape tool under 'Vector geometry' isn't working. It doesn't change existing Z values nor does it create new Z values from the raster as it is supposed to. It apparently does nothing. **How to Reproduce** Here is how I am running it: ![image](https://user-images.githubusercontent.com/64212883/114286922-1d87c480-9a20-11eb-8456-1162996ee24a.png) The layer Lines contains line objects, which I believe are supposed to have nodes with X, Y and Z coordinates. After running this, I am using the info tool to inspect the temporary layer created: ![image](https://user-images.githubusercontent.com/64212883/114286988-91c26800-9a20-11eb-8690-bb8fa9277053.png) As you can see, the closest vertex Z value that it found was 0. Not correct value. **QGIS and OS versions** QGIS version 3.10.0-A Coruña QGIS code revision 6c816b4204 Compiled against Qt 5.11.2 Running against Qt 5.11.2 Compiled against GDAL/OGR 2.4.1 Running against GDAL/OGR 2.4.1 Compiled against GEOS 3.8.0-CAPI-1.13.1 Running against GEOS 3.8.0-CAPI-1.13.1 Compiled against SQLite 3.29.0 Running against SQLite 3.29.0 PostgreSQL Client Version 11.5 SpatiaLite Version 4.3.0 QWT Version 6.1.3 QScintilla2 Version 2.10.8 Compiled against PROJ 5.2.0 Running against PROJ Rel. 5.2.0, September 15th, 2018 OS Version Windows 10 (10.0) Active python plugins CanadianWebServices; go2streetview; quick_map_services; SemiAutomaticClassificationPlugin; db_manager; processing **Additional context** The raster dataset was a Geosoft GXF. The starting vector dataset was a shapefile containing polygon objects. I transformed these to line objects to use with Drape. If someone wants to try debugging with my data, I can send it to them.
process
drape set z value from mesh not working describe the bug the drape tool under vector geometry isn t working it doesn t change existing z values nor does it create new z values from the raster as it is supposed to it apparently does nothing how to reproduce here is how i am running it the layer lines contains line objects which i believe are supposed to have nodes with x y and z coordinates after running this i am using the info tool to inspect the temporary layer created as you can see the closest vertex z value that it found was not correct value qgis and os versions qgis version a coruã±a qgis code revision compiled against qt running against qt compiled against gdal ogr running against gdal ogr compiled against geos capi running against geos capi compiled against sqlite running against sqlite postgresql client version spatialite version qwt version version compiled against proj running against proj rel september os version windows active python plugins canadianwebservices quick map services semiautomaticclassificationplugin db manager processing additional context the raster dataset was a geosoft gxf the starting vector dataset was a shapefile containing polygon objects i transformed these to line objects to use with drape if someone wants to try debugging with my data i can send it to them
1
173,128
27,389,528,499
IssuesEvent
2023-02-28 15:25:15
microsoft/pyright
https://api.github.com/repos/microsoft/pyright
closed
Regression with default match pattern for Never type on 1.1.196
as designed
Take this simple example with `reportUnnecessaryComparison` set to `warning` ```python from typing import assert_never def get_foo() -> int | str: ... foo = get_foo() match foo: case int(): ... case str(): ... case _: # Pattern will never be matched for subject type "Never" (reportUnnecessaryComparison) assert_never(foo) ``` This was working fine on previous versions.
1.0
Regression with default match pattern for Never type on 1.1.196 - Take this simple example with `reportUnnecessaryComparison` set to `warning` ```python from typing import assert_never def get_foo() -> int | str: ... foo = get_foo() match foo: case int(): ... case str(): ... case _: # Pattern will never be matched for subject type "Never" (reportUnnecessaryComparison) assert_never(foo) ``` This was working fine on previous versions.
non_process
regression with default match pattern for never type on take this simple example with reportunnecessarycomparison set to warning python from typing import assert never def get foo int str foo get foo match foo case int case str case pattern will never be matched for subject type never reportunnecessarycomparison assert never foo this was working fine on previous versions
0
83,566
15,711,813,138
IssuesEvent
2021-03-27 09:10:42
emilykaldwin1827/java-goof
https://api.github.com/repos/emilykaldwin1827/java-goof
opened
CVE-2017-9787 (High) detected in xwork-core-2.3.20.jar, struts2-core-2.3.20.jar
security vulnerability
## CVE-2017-9787 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>xwork-core-2.3.20.jar</b>, <b>struts2-core-2.3.20.jar</b></p></summary> <p> <details><summary><b>xwork-core-2.3.20.jar</b></p></summary> <p>Apache Struts 2</p> <p>Library home page: <a href="http://struts.apache.org/">http://struts.apache.org/</a></p> <p>Path to dependency file: java-goof/todolist-web-struts/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/struts/xwork/xwork-core/2.3.20/xwork-core-2.3.20.jar,java-goof/todolist-web-struts/target/todolist/WEB-INF/lib/xwork-core-2.3.20.jar</p> <p> Dependency Hierarchy: - :x: **xwork-core-2.3.20.jar** (Vulnerable Library) </details> <details><summary><b>struts2-core-2.3.20.jar</b></p></summary> <p>Apache Struts 2</p> <p>Path to dependency file: java-goof/todolist-web-struts/pom.xml</p> <p>Path to vulnerable library: canner/.m2/repository/org/apache/struts/struts2-core/2.3.20/struts2-core-2.3.20.jar,java-goof/todolist-web-struts/target/todolist/WEB-INF/lib/struts2-core-2.3.20.jar</p> <p> Dependency Hierarchy: - :x: **struts2-core-2.3.20.jar** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/emilykaldwin1827/java-goof/commit/3a7de538bc7d10df3b6c2814dad942c8222f690e">3a7de538bc7d10df3b6c2814dad942c8222f690e</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> When using a Spring AOP functionality to secure Struts actions it is possible to perform a DoS attack. Solution is to upgrade to Apache Struts version 2.5.12 or 2.3.33. <p>Publish Date: 2017-07-13 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-9787>CVE-2017-9787</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/apache/struts/tree/STRUTS_2_3_33/">https://github.com/apache/struts/tree/STRUTS_2_3_33/</a></p> <p>Release Date: 2017-07-13</p> <p>Fix Resolution: org.apache.struts:struts2-core:2.3.33,org.apache.struts:struts2-core:2.5.12</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2017-9787 (High) detected in xwork-core-2.3.20.jar, struts2-core-2.3.20.jar - ## CVE-2017-9787 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>xwork-core-2.3.20.jar</b>, <b>struts2-core-2.3.20.jar</b></p></summary> <p> <details><summary><b>xwork-core-2.3.20.jar</b></p></summary> <p>Apache Struts 2</p> <p>Library home page: <a href="http://struts.apache.org/">http://struts.apache.org/</a></p> <p>Path to dependency file: java-goof/todolist-web-struts/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/struts/xwork/xwork-core/2.3.20/xwork-core-2.3.20.jar,java-goof/todolist-web-struts/target/todolist/WEB-INF/lib/xwork-core-2.3.20.jar</p> <p> Dependency Hierarchy: - :x: **xwork-core-2.3.20.jar** (Vulnerable Library) </details> <details><summary><b>struts2-core-2.3.20.jar</b></p></summary> <p>Apache Struts 2</p> <p>Path to dependency file: java-goof/todolist-web-struts/pom.xml</p> <p>Path to vulnerable library: canner/.m2/repository/org/apache/struts/struts2-core/2.3.20/struts2-core-2.3.20.jar,java-goof/todolist-web-struts/target/todolist/WEB-INF/lib/struts2-core-2.3.20.jar</p> <p> Dependency Hierarchy: - :x: **struts2-core-2.3.20.jar** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/emilykaldwin1827/java-goof/commit/3a7de538bc7d10df3b6c2814dad942c8222f690e">3a7de538bc7d10df3b6c2814dad942c8222f690e</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> When using a Spring AOP functionality to secure Struts actions it is possible to perform a DoS attack. Solution is to upgrade to Apache Struts version 2.5.12 or 2.3.33. <p>Publish Date: 2017-07-13 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-9787>CVE-2017-9787</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/apache/struts/tree/STRUTS_2_3_33/">https://github.com/apache/struts/tree/STRUTS_2_3_33/</a></p> <p>Release Date: 2017-07-13</p> <p>Fix Resolution: org.apache.struts:struts2-core:2.3.33,org.apache.struts:struts2-core:2.5.12</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in xwork core jar core jar cve high severity vulnerability vulnerable libraries xwork core jar core jar xwork core jar apache struts library home page a href path to dependency file java goof todolist web struts pom xml path to vulnerable library home wss scanner repository org apache struts xwork xwork core xwork core jar java goof todolist web struts target todolist web inf lib xwork core jar dependency hierarchy x xwork core jar vulnerable library core jar apache struts path to dependency file java goof todolist web struts pom xml path to vulnerable library canner repository org apache struts core core jar java goof todolist web struts target todolist web inf lib core jar dependency hierarchy x core jar vulnerable library found in head commit a href found in base branch master vulnerability details when using a spring aop functionality to secure struts actions it is possible to perform a dos attack solution is to upgrade to apache struts version or publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache struts core org apache struts core step up your open source security game with whitesource
0
330,938
10,057,920,913
IssuesEvent
2019-07-22 12:50:40
openshift/odo
https://api.github.com/repos/openshift/odo
closed
Binary component push fails with context flag
kind/bug points/3 priority/Critical state/Ready
[kind/bug] <!-- Welcome! - We kindly ask you to: 1. Fill out the issue template below 2. Use the chat and talk to us if you have a question rather than a bug or feature request. The chat room is at: https://chat.openshift.io/developers/channels/odo Thanks for understanding, and for contributing to the project! --> ## What versions of software are you using? - Operating System: I tested it on macOS - Output of `odo version`: master ## How did you run odo exactly? ``` $ odo create wildfly javaee-war-test --project myproject --binary tests/examples/binary/java/wildfly/ROOT.war --context tests/examples/binary/java/wildfly/ ✓ Validating component [112ms] Please use `odo push` command to create the component with source deployed $ odo push --context tests/examples/binary/java/wildfly/ Validation ✓ Validating component [86ms] ✓ Checking component [3ms] Configuration changes ✓ Initializing component ✓ Creating component [365ms] ✓ Applying configuration [0s] Pushing to component javaee-war-test of type binary ✗ open /Users/amit/go/src/github.com/openshift/odo/tests/examples/binary/java/wildfly/tests/examples/binary/java/wildfly: no such file or directory ``` ## Actual behavior Push fails with context flag ## Expected behavior Should push binary component ## Any logs, error output, etc? Check ```How did you run odo exactly?``` section
1.0
Binary component push fails with context flag - [kind/bug] <!-- Welcome! - We kindly ask you to: 1. Fill out the issue template below 2. Use the chat and talk to us if you have a question rather than a bug or feature request. The chat room is at: https://chat.openshift.io/developers/channels/odo Thanks for understanding, and for contributing to the project! --> ## What versions of software are you using? - Operating System: I tested it on macOS - Output of `odo version`: master ## How did you run odo exactly? ``` $ odo create wildfly javaee-war-test --project myproject --binary tests/examples/binary/java/wildfly/ROOT.war --context tests/examples/binary/java/wildfly/ ✓ Validating component [112ms] Please use `odo push` command to create the component with source deployed $ odo push --context tests/examples/binary/java/wildfly/ Validation ✓ Validating component [86ms] ✓ Checking component [3ms] Configuration changes ✓ Initializing component ✓ Creating component [365ms] ✓ Applying configuration [0s] Pushing to component javaee-war-test of type binary ✗ open /Users/amit/go/src/github.com/openshift/odo/tests/examples/binary/java/wildfly/tests/examples/binary/java/wildfly: no such file or directory ``` ## Actual behavior Push fails with context flag ## Expected behavior Should push binary component ## Any logs, error output, etc? Check ```How did you run odo exactly?``` section
non_process
binary component push fails with context flag welcome we kindly ask you to fill out the issue template below use the chat and talk to us if you have a question rather than a bug or feature request the chat room is at thanks for understanding and for contributing to the project what versions of software are you using operating system i tested it on macos output of odo version master how did you run odo exactly odo create wildfly javaee war test project myproject binary tests examples binary java wildfly root war context tests examples binary java wildfly ✓ validating component please use odo push command to create the component with source deployed odo push context tests examples binary java wildfly validation ✓ validating component ✓ checking component configuration changes ✓ initializing component ✓ creating component ✓ applying configuration pushing to component javaee war test of type binary ✗ open users amit go src github com openshift odo tests examples binary java wildfly tests examples binary java wildfly no such file or directory actual behavior push fails with context flag expected behavior should push binary component any logs error output etc check how did you run odo exactly section
0
335,330
30,023,212,157
IssuesEvent
2023-06-27 02:31:41
warriordog/ActivityPubSharp
https://api.github.com/repos/warriordog/ActivityPubSharp
closed
Scaffold unit tests
area:tests
We don't want to implement many at this stage, but it would be good to have the test project(s) created in the solution.
1.0
Scaffold unit tests - We don't want to implement many at this stage, but it would be good to have the test project(s) created in the solution.
non_process
scaffold unit tests we don t want to implement many at this stage but it would be good to have the test project s created in the solution
0
16,952
22,307,166,216
IssuesEvent
2022-06-13 13:58:06
sparc4-dev/astropop
https://api.github.com/repos/sparc4-dev/astropop
closed
Clip output for image registration
enhancement image-processing
Add a function to clip a list of images to the same only-valid pixels frame based on the shifts. This argument can be added too to `register_framedata_list`.
1.0
Clip output for image registration - Add a function to clip a list of images to the same only-valid pixels frame based on the shifts. This argument can be added too to `register_framedata_list`.
process
clip output for image registration add a function to clip a list of images to the same only valid pixels frame based on the shifts this argument can be added too to register framedata list
1
39,960
5,258,068,663
IssuesEvent
2017-02-02 22:15:51
Metaswitch/homestead
https://api.github.com/repos/Metaswitch/homestead
closed
REGISTERS fail with THRIFT_EAGAIN errors with 50ms latency between GR sites
bug cat:GR cat:system-test critical
<!-- This page is for reporting issues with Project Clearwater. If you have a question, rather than a bug report, the mailing list at clearwater@lists.projectclearwater.org is a better place for it. To give us the best chance of fixing the problem, we've suggested some information to give - please follow these guidelines if possible. Don't forget that you can attach logs and screenshots to Github issues - this may help us debug a problem. --> #### Symptoms <!-- What exactly went wrong? (If possible, copy and paste the exact output you saw) --> I set up a GR deployment with 50ms latency between the two sites (i.e a 100ms round-trip time). This is only half the supported value of 100ms latency. In this state, most REGISTER attempts fail: out of 100 REGISTERs only 27 succeeded. 17 failed with a 504 Server Timeout instead of the 200 OK for the REGISTER, and 56 failed with a 500 Internal Server Error. The SAS traces show the error: ``` Failed to connect to Cassandra Error text: Exception: THRIFT_EAGAIN (timed out) [2] ``` I've attached: * a SAS trace showing a 504 error * a SAS trace showing a 500 error * tcpdump from site1-homestead-1 (port 9160 or port 47471 or port 9042 or port 7000 or port 7199) while these REGISTERs were happening For confirmation of the topology/latency: ``` clearwater@gr-stress-site1-homestead-1:~$ sudo /usr/share/clearwater/bin/run-in-signaling-namespace nodetool status Datacenter: site1 ================= Status=Up/Down |/ State=Normal/Leaving/Joining/Moving -- Address Load Tokens Owns (effective) Host ID Rack UN 10.225.166.169 345.18 KB 256 49.6% 302e4165-91b1-4594-98ee-4e0ef26d6931 RAC1 UN 10.225.166.168 360.7 KB 256 48.1% 27c1995f-1d3a-4a33-bcbe-3c09e942b751 RAC1 Datacenter: site2 ================= Status=Up/Down |/ State=Normal/Leaving/Joining/Moving -- Address Load Tokens Owns (effective) Host ID Rack UN 10.225.166.173 166.73 KB 256 50.1% 028648f0-e700-4d72-8b92-63a218fe9c4a RAC1 UN 10.225.166.172 167.45 KB 256 52.1% 0f278b01-8deb-4122-b7f2-68712d9c36b8 RAC1 clearwater@gr-stress-site1-homestead-1:~$ sudo /usr/share/clearwater/bin/run-in-signaling-namespace ping -c 2 10.225.166.169 PING 10.225.166.169 (10.225.166.169) 56(84) bytes of data. 64 bytes from 10.225.166.169: icmp_seq=1 ttl=64 time=0.192 ms 64 bytes from 10.225.166.169: icmp_seq=2 ttl=64 time=0.375 ms --- 10.225.166.169 ping statistics --- 2 packets transmitted, 2 received, 0% packet loss, time 999ms rtt min/avg/max/mdev = 0.192/0.283/0.375/0.093 ms clearwater@gr-stress-site1-homestead-1:~$ sudo /usr/share/clearwater/bin/run-in-signaling-namespace ping -c 2 10.225.166.172 PING 10.225.166.172 (10.225.166.172) 56(84) bytes of data. 64 bytes from 10.225.166.172: icmp_seq=1 ttl=64 time=100 ms 64 bytes from 10.225.166.172: icmp_seq=2 ttl=64 time=100 ms --- 10.225.166.172 ping statistics --- 2 packets transmitted, 2 received, 0% packet loss, time 1001ms rtt min/avg/max/mdev = 100.363/100.401/100.440/0.319 ms clearwater@gr-stress-site1-homestead-1:~$ sudo /usr/share/clearwater/bin/run-in-signaling-namespace ping -c 2 10.225.166.173 PING 10.225.166.173 (10.225.166.173) 56(84) bytes of data. 64 bytes from 10.225.166.173: icmp_seq=1 ttl=64 time=100 ms 64 bytes from 10.225.166.173: icmp_seq=2 ttl=64 time=100 ms --- 10.225.166.173 ping statistics --- 2 packets transmitted, 2 received, 0% packet loss, time 1001ms rtt min/avg/max/mdev = 100.459/100.491/100.524/0.318 ms ``` #### Impact <!-- What is this preventing you from doing? Does this stop Clearwater working, or stop some calls being processed? --> Cannot provide basic call service in a moderate-latency GR deployment. #### Release and environment <!-- For example: the "Doom" release, whether you're on VMware or OpenStack, etc. --> 10.0.0-161216-~255.00.0-161216.020317 (this is a relatively old version - this is due to https://github.com/Metaswitch/clearwater-etcd/issues/390) #### Steps to reproduce Create a GR deployment. Set 50ms latency between GR sites. Run REGISTER traffic. (I have a deployment set up that can reproduce this quickly, e.g. to try out fixes on.)
1.0
REGISTERS fail with THRIFT_EAGAIN errors with 50ms latency between GR sites - <!-- This page is for reporting issues with Project Clearwater. If you have a question, rather than a bug report, the mailing list at clearwater@lists.projectclearwater.org is a better place for it. To give us the best chance of fixing the problem, we've suggested some information to give - please follow these guidelines if possible. Don't forget that you can attach logs and screenshots to Github issues - this may help us debug a problem. --> #### Symptoms <!-- What exactly went wrong? (If possible, copy and paste the exact output you saw) --> I set up a GR deployment with 50ms latency between the two sites (i.e a 100ms round-trip time). This is only half the supported value of 100ms latency. In this state, most REGISTER attempts fail: out of 100 REGISTERs only 27 succeeded. 17 failed with a 504 Server Timeout instead of the 200 OK for the REGISTER, and 56 failed with a 500 Internal Server Error. The SAS traces show the error: ``` Failed to connect to Cassandra Error text: Exception: THRIFT_EAGAIN (timed out) [2] ``` I've attached: * a SAS trace showing a 504 error * a SAS trace showing a 500 error * tcpdump from site1-homestead-1 (port 9160 or port 47471 or port 9042 or port 7000 or port 7199) while these REGISTERs were happening For confirmation of the topology/latency: ``` clearwater@gr-stress-site1-homestead-1:~$ sudo /usr/share/clearwater/bin/run-in-signaling-namespace nodetool status Datacenter: site1 ================= Status=Up/Down |/ State=Normal/Leaving/Joining/Moving -- Address Load Tokens Owns (effective) Host ID Rack UN 10.225.166.169 345.18 KB 256 49.6% 302e4165-91b1-4594-98ee-4e0ef26d6931 RAC1 UN 10.225.166.168 360.7 KB 256 48.1% 27c1995f-1d3a-4a33-bcbe-3c09e942b751 RAC1 Datacenter: site2 ================= Status=Up/Down |/ State=Normal/Leaving/Joining/Moving -- Address Load Tokens Owns (effective) Host ID Rack UN 10.225.166.173 166.73 KB 256 50.1% 028648f0-e700-4d72-8b92-63a218fe9c4a RAC1 UN 10.225.166.172 167.45 KB 256 52.1% 0f278b01-8deb-4122-b7f2-68712d9c36b8 RAC1 clearwater@gr-stress-site1-homestead-1:~$ sudo /usr/share/clearwater/bin/run-in-signaling-namespace ping -c 2 10.225.166.169 PING 10.225.166.169 (10.225.166.169) 56(84) bytes of data. 64 bytes from 10.225.166.169: icmp_seq=1 ttl=64 time=0.192 ms 64 bytes from 10.225.166.169: icmp_seq=2 ttl=64 time=0.375 ms --- 10.225.166.169 ping statistics --- 2 packets transmitted, 2 received, 0% packet loss, time 999ms rtt min/avg/max/mdev = 0.192/0.283/0.375/0.093 ms clearwater@gr-stress-site1-homestead-1:~$ sudo /usr/share/clearwater/bin/run-in-signaling-namespace ping -c 2 10.225.166.172 PING 10.225.166.172 (10.225.166.172) 56(84) bytes of data. 64 bytes from 10.225.166.172: icmp_seq=1 ttl=64 time=100 ms 64 bytes from 10.225.166.172: icmp_seq=2 ttl=64 time=100 ms --- 10.225.166.172 ping statistics --- 2 packets transmitted, 2 received, 0% packet loss, time 1001ms rtt min/avg/max/mdev = 100.363/100.401/100.440/0.319 ms clearwater@gr-stress-site1-homestead-1:~$ sudo /usr/share/clearwater/bin/run-in-signaling-namespace ping -c 2 10.225.166.173 PING 10.225.166.173 (10.225.166.173) 56(84) bytes of data. 64 bytes from 10.225.166.173: icmp_seq=1 ttl=64 time=100 ms 64 bytes from 10.225.166.173: icmp_seq=2 ttl=64 time=100 ms --- 10.225.166.173 ping statistics --- 2 packets transmitted, 2 received, 0% packet loss, time 1001ms rtt min/avg/max/mdev = 100.459/100.491/100.524/0.318 ms ``` #### Impact <!-- What is this preventing you from doing? Does this stop Clearwater working, or stop some calls being processed? --> Cannot provide basic call service in a moderate-latency GR deployment. #### Release and environment <!-- For example: the "Doom" release, whether you're on VMware or OpenStack, etc. --> 10.0.0-161216-~255.00.0-161216.020317 (this is a relatively old version - this is due to https://github.com/Metaswitch/clearwater-etcd/issues/390) #### Steps to reproduce Create a GR deployment. Set 50ms latency between GR sites. Run REGISTER traffic. (I have a deployment set up that can reproduce this quickly, e.g. to try out fixes on.)
non_process
registers fail with thrift eagain errors with latency between gr sites this page is for reporting issues with project clearwater if you have a question rather than a bug report the mailing list at clearwater lists projectclearwater org is a better place for it to give us the best chance of fixing the problem we ve suggested some information to give please follow these guidelines if possible don t forget that you can attach logs and screenshots to github issues this may help us debug a problem symptoms i set up a gr deployment with latency between the two sites i e a round trip time this is only half the supported value of latency in this state most register attempts fail out of registers only succeeded failed with a server timeout instead of the ok for the register and failed with a internal server error the sas traces show the error failed to connect to cassandra error text exception thrift eagain timed out i ve attached a sas trace showing a error a sas trace showing a error tcpdump from homestead port or port or port or port or port while these registers were happening for confirmation of the topology latency clearwater gr stress homestead sudo usr share clearwater bin run in signaling namespace nodetool status datacenter status up down state normal leaving joining moving address load tokens owns effective host id rack un kb un kb bcbe datacenter status up down state normal leaving joining moving address load tokens owns effective host id rack un kb un kb clearwater gr stress homestead sudo usr share clearwater bin run in signaling namespace ping c ping bytes of data bytes from icmp seq ttl time ms bytes from icmp seq ttl time ms ping statistics packets transmitted received packet loss time rtt min avg max mdev ms clearwater gr stress homestead sudo usr share clearwater bin run in signaling namespace ping c ping bytes of data bytes from icmp seq ttl time ms bytes from icmp seq ttl time ms ping statistics packets transmitted received packet loss time rtt min avg max mdev ms clearwater gr stress homestead sudo usr share clearwater bin run in signaling namespace ping c ping bytes of data bytes from icmp seq ttl time ms bytes from icmp seq ttl time ms ping statistics packets transmitted received packet loss time rtt min avg max mdev ms impact cannot provide basic call service in a moderate latency gr deployment release and environment this is a relatively old version this is due to steps to reproduce create a gr deployment set latency between gr sites run register traffic i have a deployment set up that can reproduce this quickly e g to try out fixes on
0
20,646
3,391,651,143
IssuesEvent
2015-11-30 16:20:07
dart-lang/sdk
https://api.github.com/repos/dart-lang/sdk
closed
Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in
area-analyzer priority-high started Type-Defect
I tried to reproduce the problem asked about at http://stackoverflow.com/questions/33881724 I modified the provide project a bit because it didn't comply with the pub package layout convention, but I'm not able to do much with it because the analyzer constantly throws. https://github.com/bwu-dart-playground/bug_repo/tree/master/DartRikuloTest I haven't done anything notable after opening the project in WebStorm. > 13:37:30 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:37:54 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:38:05 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:38:14 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:38:23 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:38:29 Unregistered VCS root detected The directory /home/myuser/dart/playground/bug_repo/DartRikuloTest is under Git, but is not registered in the Settings. Add root Configure Ignore 13:38:33 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:38:45 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:38:54 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:39:04 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:39:14 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:39:23 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:39:32 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:39:45 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:39:55 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:40:06 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:40:15 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:40:25 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:40:35 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:40:45 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:40:56 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:41:06 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:41:15 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:41:25 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:41:34 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:41:45 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:41:55 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:42:06 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:42:16 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:42:25 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:42:35 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:42:45 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:42:54 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:43:03 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:43:13 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:43:22 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:43:31 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon)
1.0
Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in - I tried to reproduce the problem asked about at http://stackoverflow.com/questions/33881724 I modified the provide project a bit because it didn't comply with the pub package layout convention, but I'm not able to do much with it because the analyzer constantly throws. https://github.com/bwu-dart-playground/bug_repo/tree/master/DartRikuloTest I haven't done anything notable after opening the project in WebStorm. > 13:37:30 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:37:54 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:38:05 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:38:14 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:38:23 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:38:29 Unregistered VCS root detected The directory /home/myuser/dart/playground/bug_repo/DartRikuloTest is under Git, but is not registered in the Settings. Add root Configure Ignore 13:38:33 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:38:45 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:38:54 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:39:04 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:39:14 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:39:23 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:39:32 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:39:45 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:39:55 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:40:06 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:40:15 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:40:25 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:40:35 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:40:45 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:40:56 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:41:06 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:41:15 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:41:25 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:41:34 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:41:45 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:41:55 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:42:06 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:42:16 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:42:25 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:42:35 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:42:45 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:42:54 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:43:03 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:43:13 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:43:22 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon) 13:43:31 Throwable: Dart analysis server, SDK version 1.14.0-edge.594183e0db633aebbe84429443958d725cc146b1, server version 1.12.0, error: Task failed: ComputePropagableVariableDependenciesTask for element Map _stdcolors in source /home/myuser/.pub-cache/hosted/pub.dartlang.org/rikulo_commons-2.0.6/lib/src/util/Color.dart NodeLocator failed to find a variable's declaration #0 ComputePropagableVariableDependenciesTask.internalPerform (package:analyzer/src/task/dart.dart:2054) #1 AnalysisTask._safelyPerform (package:analyzer/task/model.dart:310) #2 AnalysisTask.perform (package:analyzer/task/model.dart:210) #3 AnalysisDriver.performWorkItem (package:analyzer/src/task/driver.dart:272) #4 AnalysisDriver.performAnalysisTask (package:analyzer/src/task/driver.dart:247) #5 AnalysisContextImpl.performAnalysisTask. (package:analyzer/src/context/context.dart:1116) #6 _PerformanceTagImpl.makeCurrentWhile (package:analyzer/src/generated/utilities_gene... (show balloon)
non_process
task failed computepropagablevariabledependenciestask for element map stdcolors in i tried to reproduce the problem asked about at i modified the provide project a bit because it didn t comply with the pub package layout convention but i m not able to do much with it because the analyzer constantly throws i haven t done anything notable after opening the project in webstorm throwable dart analysis server sdk version edge server version error task failed computepropagablevariabledependenciestask for element map stdcolors in source home myuser pub cache hosted pub dartlang org rikulo commons lib src util color dart nodelocator failed to find a variable s declaration computepropagablevariabledependenciestask internalperform package analyzer src task dart dart analysistask safelyperform package analyzer task model dart analysistask perform package analyzer task model dart analysisdriver performworkitem package analyzer src task driver dart analysisdriver performanalysistask package analyzer src task driver dart analysiscontextimpl performanalysistask package analyzer src context context dart performancetagimpl makecurrentwhile package analyzer src generated utilities gene show balloon throwable dart analysis server sdk version edge server version error task failed computepropagablevariabledependenciestask for element map stdcolors in source home myuser pub cache hosted pub dartlang org rikulo commons lib src util color dart nodelocator failed to find a variable s declaration computepropagablevariabledependenciestask internalperform package analyzer src task dart dart analysistask safelyperform package analyzer task model dart analysistask perform package analyzer task model dart analysisdriver performworkitem package analyzer src task driver dart analysisdriver performanalysistask package analyzer src task driver dart analysiscontextimpl performanalysistask package analyzer src context context dart performancetagimpl makecurrentwhile package analyzer src generated utilities gene show balloon throwable dart analysis server sdk version edge server version error task failed computepropagablevariabledependenciestask for element map stdcolors in source home myuser pub cache hosted pub dartlang org rikulo commons lib src util color dart nodelocator failed to find a variable s declaration computepropagablevariabledependenciestask internalperform package analyzer src task dart dart analysistask safelyperform package analyzer task model dart analysistask perform package analyzer task model dart analysisdriver performworkitem package analyzer src task driver dart analysisdriver performanalysistask package analyzer src task driver dart analysiscontextimpl performanalysistask package analyzer src context context dart performancetagimpl makecurrentwhile package analyzer src generated utilities gene show balloon throwable dart analysis server sdk version edge server version error task failed computepropagablevariabledependenciestask for element map stdcolors in source home myuser pub cache hosted pub dartlang org rikulo commons lib src util color dart nodelocator failed to find a variable s declaration computepropagablevariabledependenciestask internalperform package analyzer src task dart dart analysistask safelyperform package analyzer task model dart analysistask perform package analyzer task model dart analysisdriver performworkitem package analyzer src task driver dart analysisdriver performanalysistask package analyzer src task driver dart analysiscontextimpl performanalysistask package analyzer src context context dart performancetagimpl makecurrentwhile package analyzer src generated utilities gene show balloon throwable dart analysis server sdk version edge server version error task failed computepropagablevariabledependenciestask for element map stdcolors in source home myuser pub cache hosted pub dartlang org rikulo commons lib src util color dart nodelocator failed to find a variable s declaration computepropagablevariabledependenciestask internalperform package analyzer src task dart dart analysistask safelyperform package analyzer task model dart analysistask perform package analyzer task model dart analysisdriver performworkitem package analyzer src task driver dart analysisdriver performanalysistask package analyzer src task driver dart analysiscontextimpl performanalysistask package analyzer src context context dart performancetagimpl makecurrentwhile package analyzer src generated utilities gene show balloon unregistered vcs root detected the directory home myuser dart playground bug repo dartrikulotest is under git but is not registered in the settings add root configure ignore throwable dart analysis server sdk version edge server version error task failed computepropagablevariabledependenciestask for element map stdcolors in source home myuser pub cache hosted pub dartlang org rikulo commons lib src util color dart nodelocator failed to find a variable s declaration computepropagablevariabledependenciestask internalperform package analyzer src task dart dart analysistask safelyperform package analyzer task model dart analysistask perform package analyzer task model dart analysisdriver performworkitem package analyzer src task driver dart analysisdriver performanalysistask package analyzer src task driver dart analysiscontextimpl performanalysistask package analyzer src context context dart performancetagimpl makecurrentwhile package analyzer src generated utilities gene show balloon throwable dart analysis server sdk version edge server version error task failed computepropagablevariabledependenciestask for element map stdcolors in source home myuser pub cache hosted pub dartlang org rikulo commons lib src util color dart nodelocator failed to find a variable s declaration computepropagablevariabledependenciestask internalperform package analyzer src task dart dart analysistask safelyperform package analyzer task model dart analysistask perform package analyzer task model dart analysisdriver performworkitem package analyzer src task driver dart analysisdriver performanalysistask package analyzer src task driver dart analysiscontextimpl performanalysistask package analyzer src context context dart performancetagimpl makecurrentwhile package analyzer src generated utilities gene show balloon throwable dart analysis server sdk version edge server version error task failed computepropagablevariabledependenciestask for element map stdcolors in source home myuser pub cache hosted pub dartlang org rikulo commons lib src util color dart nodelocator failed to find a variable s declaration computepropagablevariabledependenciestask internalperform package analyzer src task dart dart analysistask safelyperform package analyzer task model dart analysistask perform package analyzer task model dart analysisdriver performworkitem package analyzer src task driver dart analysisdriver performanalysistask package analyzer src task driver dart analysiscontextimpl performanalysistask package analyzer src context context dart performancetagimpl makecurrentwhile package analyzer src generated utilities gene show balloon throwable dart analysis server sdk version edge server version error task failed computepropagablevariabledependenciestask for element map stdcolors in source home myuser pub cache hosted pub dartlang org rikulo commons lib src util color dart nodelocator failed to find a variable s declaration computepropagablevariabledependenciestask internalperform package analyzer src task dart dart analysistask safelyperform package analyzer task model dart analysistask perform package analyzer task model dart analysisdriver performworkitem package analyzer src task driver dart analysisdriver performanalysistask package analyzer src task driver dart analysiscontextimpl performanalysistask package analyzer src context context dart performancetagimpl makecurrentwhile package analyzer src generated utilities gene show balloon throwable dart analysis server sdk version edge server version error task failed computepropagablevariabledependenciestask for element map stdcolors in source home myuser pub cache hosted pub dartlang org rikulo commons lib src util color dart nodelocator failed to find a variable s declaration computepropagablevariabledependenciestask internalperform package analyzer src task dart dart analysistask safelyperform package analyzer task model dart analysistask perform package analyzer task model dart analysisdriver performworkitem package analyzer src task driver dart analysisdriver performanalysistask package analyzer src task driver dart analysiscontextimpl performanalysistask package analyzer src context context dart performancetagimpl makecurrentwhile package analyzer src generated utilities gene show balloon throwable dart analysis server sdk version edge server version error task failed computepropagablevariabledependenciestask for element map stdcolors in source home myuser pub cache hosted pub dartlang org rikulo commons lib src util color dart nodelocator failed to find a variable s declaration computepropagablevariabledependenciestask internalperform package analyzer src task dart dart analysistask safelyperform package analyzer task model dart analysistask perform package analyzer task model dart analysisdriver performworkitem package analyzer src task driver dart analysisdriver performanalysistask package analyzer src task driver dart analysiscontextimpl performanalysistask package analyzer src context context dart performancetagimpl makecurrentwhile package analyzer src generated utilities gene show balloon throwable dart analysis server sdk version edge server version error task failed computepropagablevariabledependenciestask for element map stdcolors in source home myuser pub cache hosted pub dartlang org rikulo commons lib src util color dart nodelocator failed to find a variable s declaration computepropagablevariabledependenciestask internalperform package analyzer src task dart dart analysistask safelyperform package analyzer task model dart analysistask perform package analyzer task model dart analysisdriver performworkitem package analyzer src task driver dart analysisdriver performanalysistask package analyzer src task driver dart analysiscontextimpl performanalysistask package analyzer src context context dart performancetagimpl makecurrentwhile package analyzer src generated utilities gene show balloon throwable dart analysis server sdk version edge server version error task failed computepropagablevariabledependenciestask for element map stdcolors in source home myuser pub cache hosted pub dartlang org rikulo commons lib src util color dart nodelocator failed to find a variable s declaration computepropagablevariabledependenciestask internalperform package analyzer src task dart dart analysistask safelyperform package analyzer task model dart analysistask perform package analyzer task model dart analysisdriver performworkitem package analyzer src task driver dart analysisdriver performanalysistask package analyzer src task driver dart analysiscontextimpl performanalysistask package analyzer src context context dart performancetagimpl makecurrentwhile package analyzer src generated utilities gene show balloon throwable dart analysis server sdk version edge server version error task failed computepropagablevariabledependenciestask for element map stdcolors in source home myuser pub cache hosted pub dartlang org rikulo commons lib src util color dart nodelocator failed to find a variable s declaration computepropagablevariabledependenciestask internalperform package analyzer src task dart dart analysistask safelyperform package analyzer task model dart analysistask perform package analyzer task model dart analysisdriver performworkitem package analyzer src task driver dart analysisdriver performanalysistask package analyzer src task driver dart analysiscontextimpl performanalysistask package analyzer src context context dart performancetagimpl makecurrentwhile package analyzer src generated utilities gene show balloon throwable dart analysis server sdk version edge server version error task failed computepropagablevariabledependenciestask for element map stdcolors in source home myuser pub cache hosted pub dartlang org rikulo commons lib src util color dart nodelocator failed to find a variable s declaration computepropagablevariabledependenciestask internalperform package analyzer src task dart dart analysistask safelyperform package analyzer task model dart analysistask perform package analyzer task model dart analysisdriver performworkitem package analyzer src task driver dart analysisdriver performanalysistask package analyzer src task driver dart analysiscontextimpl performanalysistask package analyzer src context context dart performancetagimpl makecurrentwhile package analyzer src generated utilities gene show balloon throwable dart analysis server sdk version edge server version error task failed computepropagablevariabledependenciestask for element map stdcolors in source home myuser pub cache hosted pub dartlang org rikulo commons lib src util color dart nodelocator failed to find a variable s declaration computepropagablevariabledependenciestask internalperform package analyzer src task dart dart analysistask safelyperform package analyzer task model dart analysistask perform package analyzer task model dart analysisdriver performworkitem package analyzer src task driver dart analysisdriver performanalysistask package analyzer src task driver dart analysiscontextimpl performanalysistask package analyzer src context context dart performancetagimpl makecurrentwhile package analyzer src generated utilities gene show balloon throwable dart analysis server sdk version edge server version error task failed computepropagablevariabledependenciestask for element map stdcolors in source home myuser pub cache hosted pub dartlang org rikulo commons lib src util color dart nodelocator failed to find a variable s declaration computepropagablevariabledependenciestask internalperform package analyzer src task dart dart analysistask safelyperform package analyzer task model dart analysistask perform package analyzer task model dart analysisdriver performworkitem package analyzer src task driver dart analysisdriver performanalysistask package analyzer src task driver dart analysiscontextimpl performanalysistask package analyzer src context context dart performancetagimpl makecurrentwhile package analyzer src generated utilities gene show balloon throwable dart analysis server sdk version edge server version error task failed computepropagablevariabledependenciestask for element map stdcolors in source home myuser pub cache hosted pub dartlang org rikulo commons lib src util color dart nodelocator failed to find a variable s declaration computepropagablevariabledependenciestask internalperform package analyzer src task dart dart analysistask safelyperform package analyzer task model dart analysistask perform package analyzer task model dart analysisdriver performworkitem package analyzer src task driver dart analysisdriver performanalysistask package analyzer src task driver dart analysiscontextimpl performanalysistask package analyzer src context context dart performancetagimpl makecurrentwhile package analyzer src generated utilities gene show balloon throwable dart analysis server sdk version edge server version error task failed computepropagablevariabledependenciestask for element map stdcolors in source home myuser pub cache hosted pub dartlang org rikulo commons lib src util color dart nodelocator failed to find a variable s declaration computepropagablevariabledependenciestask internalperform package analyzer src task dart dart analysistask safelyperform package analyzer task model dart analysistask perform package analyzer task model dart analysisdriver performworkitem package analyzer src task driver dart analysisdriver performanalysistask package analyzer src task driver dart analysiscontextimpl performanalysistask package analyzer src context context dart performancetagimpl makecurrentwhile package analyzer src generated utilities gene show balloon throwable dart analysis server sdk version edge server version error task failed computepropagablevariabledependenciestask for element map stdcolors in source home myuser pub cache hosted pub dartlang org rikulo commons lib src util color dart nodelocator failed to find a variable s declaration computepropagablevariabledependenciestask internalperform package analyzer src task dart dart analysistask safelyperform package analyzer task model dart analysistask perform package analyzer task model dart analysisdriver performworkitem package analyzer src task driver dart analysisdriver performanalysistask package analyzer src task driver dart analysiscontextimpl performanalysistask package analyzer src context context dart performancetagimpl makecurrentwhile package analyzer src generated utilities gene show balloon throwable dart analysis server sdk version edge server version error task failed computepropagablevariabledependenciestask for element map stdcolors in source home myuser pub cache hosted pub dartlang org rikulo commons lib src util color dart nodelocator failed to find a variable s declaration computepropagablevariabledependenciestask internalperform package analyzer src task dart dart analysistask safelyperform package analyzer task model dart analysistask perform package analyzer task model dart analysisdriver performworkitem package analyzer src task driver dart analysisdriver performanalysistask package analyzer src task driver dart analysiscontextimpl performanalysistask package analyzer src context context dart performancetagimpl makecurrentwhile package analyzer src generated utilities gene show balloon throwable dart analysis server sdk version edge server version error task failed computepropagablevariabledependenciestask for element map stdcolors in source home myuser pub cache hosted pub dartlang org rikulo commons lib src util color dart nodelocator failed to find a variable s declaration computepropagablevariabledependenciestask internalperform package analyzer src task dart dart analysistask safelyperform package analyzer task model dart analysistask perform package analyzer task model dart analysisdriver performworkitem package analyzer src task driver dart analysisdriver performanalysistask package analyzer src task driver dart analysiscontextimpl performanalysistask package analyzer src context context dart performancetagimpl makecurrentwhile package analyzer src generated utilities gene show balloon throwable dart analysis server sdk version edge server version error task failed computepropagablevariabledependenciestask for element map stdcolors in source home myuser pub cache hosted pub dartlang org rikulo commons lib src util color dart nodelocator failed to find a variable s declaration computepropagablevariabledependenciestask internalperform package analyzer src task dart dart analysistask safelyperform package analyzer task model dart analysistask perform package analyzer task model dart analysisdriver performworkitem package analyzer src task driver dart analysisdriver performanalysistask package analyzer src task driver dart analysiscontextimpl performanalysistask package analyzer src context context dart performancetagimpl makecurrentwhile package analyzer src generated utilities gene show balloon throwable dart analysis server sdk version edge server version error task failed computepropagablevariabledependenciestask for element map stdcolors in source home myuser pub cache hosted pub dartlang org rikulo commons lib src util color dart nodelocator failed to find a variable s declaration computepropagablevariabledependenciestask internalperform package analyzer src task dart dart analysistask safelyperform package analyzer task model dart analysistask perform package analyzer task model dart analysisdriver performworkitem package analyzer src task driver dart analysisdriver performanalysistask package analyzer src task driver dart analysiscontextimpl performanalysistask package analyzer src context context dart performancetagimpl makecurrentwhile package analyzer src generated utilities gene show balloon throwable dart analysis server sdk version edge server version error task failed computepropagablevariabledependenciestask for element map stdcolors in source home myuser pub cache hosted pub dartlang org rikulo commons lib src util color dart nodelocator failed to find a variable s declaration computepropagablevariabledependenciestask internalperform package analyzer src task dart dart analysistask safelyperform package analyzer task model dart analysistask perform package analyzer task model dart analysisdriver performworkitem package analyzer src task driver dart analysisdriver performanalysistask package analyzer src task driver dart analysiscontextimpl performanalysistask package analyzer src context context dart performancetagimpl makecurrentwhile package analyzer src generated utilities gene show balloon throwable dart analysis server sdk version edge server version error task failed computepropagablevariabledependenciestask for element map stdcolors in source home myuser pub cache hosted pub dartlang org rikulo commons lib src util color dart nodelocator failed to find a variable s declaration computepropagablevariabledependenciestask internalperform package analyzer src task dart dart analysistask safelyperform package analyzer task model dart analysistask perform package analyzer task model dart analysisdriver performworkitem package analyzer src task driver dart analysisdriver performanalysistask package analyzer src task driver dart analysiscontextimpl performanalysistask package analyzer src context context dart performancetagimpl makecurrentwhile package analyzer src generated utilities gene show balloon throwable dart analysis server sdk version edge server version error task failed computepropagablevariabledependenciestask for element map stdcolors in source home myuser pub cache hosted pub dartlang org rikulo commons lib src util color dart nodelocator failed to find a variable s declaration computepropagablevariabledependenciestask internalperform package analyzer src task dart dart analysistask safelyperform package analyzer task model dart analysistask perform package analyzer task model dart analysisdriver performworkitem package analyzer src task driver dart analysisdriver performanalysistask package analyzer src task driver dart analysiscontextimpl performanalysistask package analyzer src context context dart performancetagimpl makecurrentwhile package analyzer src generated utilities gene show balloon throwable dart analysis server sdk version edge server version error task failed computepropagablevariabledependenciestask for element map stdcolors in source home myuser pub cache hosted pub dartlang org rikulo commons lib src util color dart nodelocator failed to find a variable s declaration computepropagablevariabledependenciestask internalperform package analyzer src task dart dart analysistask safelyperform package analyzer task model dart analysistask perform package analyzer task model dart analysisdriver performworkitem package analyzer src task driver dart analysisdriver performanalysistask package analyzer src task driver dart analysiscontextimpl performanalysistask package analyzer src context context dart performancetagimpl makecurrentwhile package analyzer src generated utilities gene show balloon throwable dart analysis server sdk version edge server version error task failed computepropagablevariabledependenciestask for element map stdcolors in source home myuser pub cache hosted pub dartlang org rikulo commons lib src util color dart nodelocator failed to find a variable s declaration computepropagablevariabledependenciestask internalperform package analyzer src task dart dart analysistask safelyperform package analyzer task model dart analysistask perform package analyzer task model dart analysisdriver performworkitem package analyzer src task driver dart analysisdriver performanalysistask package analyzer src task driver dart analysiscontextimpl performanalysistask package analyzer src context context dart performancetagimpl makecurrentwhile package analyzer src generated utilities gene show balloon throwable dart analysis server sdk version edge server version error task failed computepropagablevariabledependenciestask for element map stdcolors in source home myuser pub cache hosted pub dartlang org rikulo commons lib src util color dart nodelocator failed to find a variable s declaration computepropagablevariabledependenciestask internalperform package analyzer src task dart dart analysistask safelyperform package analyzer task model dart analysistask perform package analyzer task model dart analysisdriver performworkitem package analyzer src task driver dart analysisdriver performanalysistask package analyzer src task driver dart analysiscontextimpl performanalysistask package analyzer src context context dart performancetagimpl makecurrentwhile package analyzer src generated utilities gene show balloon throwable dart analysis server sdk version edge server version error task failed computepropagablevariabledependenciestask for element map stdcolors in source home myuser pub cache hosted pub dartlang org rikulo commons lib src util color dart nodelocator failed to find a variable s declaration computepropagablevariabledependenciestask internalperform package analyzer src task dart dart analysistask safelyperform package analyzer task model dart analysistask perform package analyzer task model dart analysisdriver performworkitem package analyzer src task driver dart analysisdriver performanalysistask package analyzer src task driver dart analysiscontextimpl performanalysistask package analyzer src context context dart performancetagimpl makecurrentwhile package analyzer src generated utilities gene show balloon throwable dart analysis server sdk version edge server version error task failed computepropagablevariabledependenciestask for element map stdcolors in source home myuser pub cache hosted pub dartlang org rikulo commons lib src util color dart nodelocator failed to find a variable s declaration computepropagablevariabledependenciestask internalperform package analyzer src task dart dart analysistask safelyperform package analyzer task model dart analysistask perform package analyzer task model dart analysisdriver performworkitem package analyzer src task driver dart analysisdriver performanalysistask package analyzer src task driver dart analysiscontextimpl performanalysistask package analyzer src context context dart performancetagimpl makecurrentwhile package analyzer src generated utilities gene show balloon throwable dart analysis server sdk version edge server version error task failed computepropagablevariabledependenciestask for element map stdcolors in source home myuser pub cache hosted pub dartlang org rikulo commons lib src util color dart nodelocator failed to find a variable s declaration computepropagablevariabledependenciestask internalperform package analyzer src task dart dart analysistask safelyperform package analyzer task model dart analysistask perform package analyzer task model dart analysisdriver performworkitem package analyzer src task driver dart analysisdriver performanalysistask package analyzer src task driver dart analysiscontextimpl performanalysistask package analyzer src context context dart performancetagimpl makecurrentwhile package analyzer src generated utilities gene show balloon throwable dart analysis server sdk version edge server version error task failed computepropagablevariabledependenciestask for element map stdcolors in source home myuser pub cache hosted pub dartlang org rikulo commons lib src util color dart nodelocator failed to find a variable s declaration computepropagablevariabledependenciestask internalperform package analyzer src task dart dart analysistask safelyperform package analyzer task model dart analysistask perform package analyzer task model dart analysisdriver performworkitem package analyzer src task driver dart analysisdriver performanalysistask package analyzer src task driver dart analysiscontextimpl performanalysistask package analyzer src context context dart performancetagimpl makecurrentwhile package analyzer src generated utilities gene show balloon throwable dart analysis server sdk version edge server version error task failed computepropagablevariabledependenciestask for element map stdcolors in source home myuser pub cache hosted pub dartlang org rikulo commons lib src util color dart nodelocator failed to find a variable s declaration computepropagablevariabledependenciestask internalperform package analyzer src task dart dart analysistask safelyperform package analyzer task model dart analysistask perform package analyzer task model dart analysisdriver performworkitem package analyzer src task driver dart analysisdriver performanalysistask package analyzer src task driver dart analysiscontextimpl performanalysistask package analyzer src context context dart performancetagimpl makecurrentwhile package analyzer src generated utilities gene show balloon throwable dart analysis server sdk version edge server version error task failed computepropagablevariabledependenciestask for element map stdcolors in source home myuser pub cache hosted pub dartlang org rikulo commons lib src util color dart nodelocator failed to find a variable s declaration computepropagablevariabledependenciestask internalperform package analyzer src task dart dart analysistask safelyperform package analyzer task model dart analysistask perform package analyzer task model dart analysisdriver performworkitem package analyzer src task driver dart analysisdriver performanalysistask package analyzer src task driver dart analysiscontextimpl performanalysistask package analyzer src context context dart performancetagimpl makecurrentwhile package analyzer src generated utilities gene show balloon
0
265,326
28,262,466,883
IssuesEvent
2023-04-07 01:24:44
turkdevops/landscapeapp
https://api.github.com/repos/turkdevops/landscapeapp
closed
CVE-2021-23440 (High) detected in set-value-2.0.1.tgz - autoclosed
Mend: dependency security vulnerability
## CVE-2021-23440 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>set-value-2.0.1.tgz</b></p></summary> <p>Create nested values and any intermediaries using dot notation (`'a.b.c'`) paths.</p> <p>Library home page: <a href="https://registry.npmjs.org/set-value/-/set-value-2.0.1.tgz">https://registry.npmjs.org/set-value/-/set-value-2.0.1.tgz</a></p> <p> Dependency Hierarchy: - git-branch-2.0.1.tgz (Root Library) - findup-sync-2.0.0.tgz - micromatch-3.1.10.tgz - snapdragon-0.8.2.tgz - base-0.11.2.tgz - cache-base-1.0.1.tgz - :x: **set-value-2.0.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/turkdevops/landscapeapp/commit/3657f85158253a9663b2210cdbe1dee4fc4d6249">3657f85158253a9663b2210cdbe1dee4fc4d6249</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> This affects the package set-value before <2.0.1, >=3.0.0 <4.0.1. A type confusion vulnerability can lead to a bypass of CVE-2019-10747 when the user-provided keys used in the path parameter are arrays. Mend Note: After conducting further research, Mend has determined that all versions of set-value up to version 4.0.0 are vulnerable to CVE-2021-23440. <p>Publish Date: 2021-09-12 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-23440>CVE-2021-23440</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2021-09-12</p> <p>Fix Resolution: set-value - 4.0.1 </p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-23440 (High) detected in set-value-2.0.1.tgz - autoclosed - ## CVE-2021-23440 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>set-value-2.0.1.tgz</b></p></summary> <p>Create nested values and any intermediaries using dot notation (`'a.b.c'`) paths.</p> <p>Library home page: <a href="https://registry.npmjs.org/set-value/-/set-value-2.0.1.tgz">https://registry.npmjs.org/set-value/-/set-value-2.0.1.tgz</a></p> <p> Dependency Hierarchy: - git-branch-2.0.1.tgz (Root Library) - findup-sync-2.0.0.tgz - micromatch-3.1.10.tgz - snapdragon-0.8.2.tgz - base-0.11.2.tgz - cache-base-1.0.1.tgz - :x: **set-value-2.0.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/turkdevops/landscapeapp/commit/3657f85158253a9663b2210cdbe1dee4fc4d6249">3657f85158253a9663b2210cdbe1dee4fc4d6249</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> This affects the package set-value before <2.0.1, >=3.0.0 <4.0.1. A type confusion vulnerability can lead to a bypass of CVE-2019-10747 when the user-provided keys used in the path parameter are arrays. Mend Note: After conducting further research, Mend has determined that all versions of set-value up to version 4.0.0 are vulnerable to CVE-2021-23440. <p>Publish Date: 2021-09-12 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-23440>CVE-2021-23440</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2021-09-12</p> <p>Fix Resolution: set-value - 4.0.1 </p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in set value tgz autoclosed cve high severity vulnerability vulnerable library set value tgz create nested values and any intermediaries using dot notation a b c paths library home page a href dependency hierarchy git branch tgz root library findup sync tgz micromatch tgz snapdragon tgz base tgz cache base tgz x set value tgz vulnerable library found in head commit a href found in base branch master vulnerability details this affects the package set value before a type confusion vulnerability can lead to a bypass of cve when the user provided keys used in the path parameter are arrays mend note after conducting further research mend has determined that all versions of set value up to version are vulnerable to cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution set value step up your open source security game with mend
0
104,470
22,676,667,439
IssuesEvent
2022-07-04 05:43:34
appsmithorg/appsmith
https://api.github.com/repos/appsmithorg/appsmith
closed
[Bug ]: Issues related to Google UQI
Bug Needs Triaging Deploy Preview UQI BE Coders Pod
### Is there an existing issue for this? - [X] I have searched the existing issues ### Description Below are the list of issues is Enhancement: 1. Highlight in red when the mapping with the field is wrong 2. Highlight in red when JS Syntax is wrong 3. Indicate the mandatory field 4. Highlight in red when the mapping with field is wrong 5. Highlight in red when JS Syntax is wrong 6. Filter to be left-aligned: If possible we can remove "Where" as well as we are displaying the condition as a separate field 7. Sort by field can have a drop-down 8. If possible display load time and size of Data 9. the search bar is not consistent 10. When the sheet is deleted the user must be displayed with an error message on the queries using the sheet 11. On "Read Only" permission the user must not display with CRUD option or on Trying to click on Generate page option an toast message must be displayed to user 12. ### Steps To Reproduce - [x] 1. When the user changes the name of the spreadsheet and the sheet name remains as previous it is observed that the results are still fetched from the sheet name and not from the spreadsheet.
 expectation: when the spreadsheet is changed we can clear other fields - [x] 2. On converting the spreadsheet field into JS it is observed that the URL does not get updated based on the name change pulled from the dropdown - [x] 3. Selecting Entity as Sheetrows in Fetch details is not functional - [x] 4. Indicate the mandatory field - [x] 5. Scroller is displayed to the user when the relationship pane is displayed to the user - [x] 6. Scroller is displayed in the dropdown of the filter by field commands - [x] 7. Uncaught promise exception https://www.loom.com/share/b64819cd914e4157a62707783fe64a7a - [x] 8. Does not work on a page that has decimal and multiple data types - [x] 9. Crud/Table does not get populated with different data types https://www.loom.com/share/36a4e4c973304ad49ae1a1483d44c224 - [x] 10. Negative number should not be supported in table heading row - [x] 11. Unable to import Apps. - [x] 12. CRUD : Insert and Update functionality is broken - [x] 13. Filter by and Sort By : On converting from the JS to regular field the field does not get updated based on the values written in the JS and vice versa (https://www.loom.com/share/c26b4f938b924b71b876b014bc58403e) - [ ] #14600 - [x] When user try to navigate to query params from entity section in property pane the query pane is displayed plan https://www.loom.com/share/5a0046bd40d84daaba6238ddcbc8e749 ### Public Sample App _No response_ ### Version Cloud
1.0
[Bug ]: Issues related to Google UQI - ### Is there an existing issue for this? - [X] I have searched the existing issues ### Description Below are the list of issues is Enhancement: 1. Highlight in red when the mapping with the field is wrong 2. Highlight in red when JS Syntax is wrong 3. Indicate the mandatory field 4. Highlight in red when the mapping with field is wrong 5. Highlight in red when JS Syntax is wrong 6. Filter to be left-aligned: If possible we can remove "Where" as well as we are displaying the condition as a separate field 7. Sort by field can have a drop-down 8. If possible display load time and size of Data 9. the search bar is not consistent 10. When the sheet is deleted the user must be displayed with an error message on the queries using the sheet 11. On "Read Only" permission the user must not display with CRUD option or on Trying to click on Generate page option an toast message must be displayed to user 12. ### Steps To Reproduce - [x] 1. When the user changes the name of the spreadsheet and the sheet name remains as previous it is observed that the results are still fetched from the sheet name and not from the spreadsheet.
 expectation: when the spreadsheet is changed we can clear other fields - [x] 2. On converting the spreadsheet field into JS it is observed that the URL does not get updated based on the name change pulled from the dropdown - [x] 3. Selecting Entity as Sheetrows in Fetch details is not functional - [x] 4. Indicate the mandatory field - [x] 5. Scroller is displayed to the user when the relationship pane is displayed to the user - [x] 6. Scroller is displayed in the dropdown of the filter by field commands - [x] 7. Uncaught promise exception https://www.loom.com/share/b64819cd914e4157a62707783fe64a7a - [x] 8. Does not work on a page that has decimal and multiple data types - [x] 9. Crud/Table does not get populated with different data types https://www.loom.com/share/36a4e4c973304ad49ae1a1483d44c224 - [x] 10. Negative number should not be supported in table heading row - [x] 11. Unable to import Apps. - [x] 12. CRUD : Insert and Update functionality is broken - [x] 13. Filter by and Sort By : On converting from the JS to regular field the field does not get updated based on the values written in the JS and vice versa (https://www.loom.com/share/c26b4f938b924b71b876b014bc58403e) - [ ] #14600 - [x] When user try to navigate to query params from entity section in property pane the query pane is displayed plan https://www.loom.com/share/5a0046bd40d84daaba6238ddcbc8e749 ### Public Sample App _No response_ ### Version Cloud
non_process
issues related to google uqi is there an existing issue for this i have searched the existing issues description below are the list of issues is enhancement highlight in red when the mapping with the field is wrong highlight in red when js syntax is wrong indicate the mandatory field highlight in red when the mapping with field is wrong highlight in red when js syntax is wrong filter to be left aligned if possible we can remove where as well as we are displaying the condition as a separate field sort by field can have a drop down if possible display load time and size of data the search bar is not consistent when the sheet is deleted the user must be displayed with an error message on the queries using the sheet on read only permission the user must not display with crud option or on trying to click on generate page option an toast message must be displayed to user steps to reproduce when the user changes the name of the spreadsheet and the sheet name remains as previous it is observed that the results are still fetched from the sheet name and not from the spreadsheet 
 expectation when the spreadsheet is changed we can clear other fields on converting the spreadsheet field into js it is observed that the url does not get updated based on the name change pulled from the dropdown selecting entity as sheetrows in fetch details is not functional indicate the mandatory field scroller is displayed to the user when the relationship pane is displayed to the user scroller is displayed in the dropdown of the filter by field commands uncaught promise exception does not work on a page that has decimal and multiple data types crud table does not get populated with different data types negative number should not be supported in table heading row unable to import apps crud insert and update functionality is broken filter by and sort by on converting from the js to regular field the field does not get updated based on the values written in the js and vice versa when user try to navigate to query params from entity section in property pane the query pane is displayed plan public sample app no response version cloud
0
8,963
12,069,463,017
IssuesEvent
2020-04-16 16:05:58
ORNL-AMO/AMO-Tools-Desktop
https://api.github.com/repos/ORNL-AMO/AMO-Tools-Desktop
closed
Explore Opps + Report
Fans Process Heating Pumps Steam enhancement
Need to report which “Assessment” modifications are checked, and echo the input parameters for each modification, i.e. “Install VFD”, “more efficient fan”, etc. - koa - @rmroot can we do this? what about assessments made in expert view? if we do it here we should do with with other assessments also Could we use something like the Select Scenario page + the checkboxes from Explore Opps? ![image.png](https://images.zenhubusercontent.com/5cd48a2af8cffa5a19122d27/5c27240a-97e2-4462-ae00-4e4ec7590e86)
1.0
Explore Opps + Report - Need to report which “Assessment” modifications are checked, and echo the input parameters for each modification, i.e. “Install VFD”, “more efficient fan”, etc. - koa - @rmroot can we do this? what about assessments made in expert view? if we do it here we should do with with other assessments also Could we use something like the Select Scenario page + the checkboxes from Explore Opps? ![image.png](https://images.zenhubusercontent.com/5cd48a2af8cffa5a19122d27/5c27240a-97e2-4462-ae00-4e4ec7590e86)
process
explore opps report need to report which “assessment” modifications are checked and echo the input parameters for each modification i e “install vfd” “more efficient fan” etc koa rmroot can we do this what about assessments made in expert view if we do it here we should do with with other assessments also could we use something like the select scenario page the checkboxes from explore opps
1
679,093
23,221,158,971
IssuesEvent
2022-08-02 18:21:21
pixiebrix/pixiebrix-extension
https://api.github.com/repos/pixiebrix/pixiebrix-extension
closed
Page Editor sometimes not appearing on Chrome 99
bug priority
- Chrome Bug Tracker: https://bugs.chromium.org/p/chromium/issues/detail?id=1304058 This is impacting all Chrome Extensions, not just ours - Team discussion: https://pixiebrix.slack.com/archives/C023KL47XV4/p1646751676312859
1.0
Page Editor sometimes not appearing on Chrome 99 - - Chrome Bug Tracker: https://bugs.chromium.org/p/chromium/issues/detail?id=1304058 This is impacting all Chrome Extensions, not just ours - Team discussion: https://pixiebrix.slack.com/archives/C023KL47XV4/p1646751676312859
non_process
page editor sometimes not appearing on chrome chrome bug tracker this is impacting all chrome extensions not just ours team discussion
0
390,140
11,525,361,432
IssuesEvent
2020-02-15 07:46:55
kubernetes/minikube
https://api.github.com/repos/kubernetes/minikube
closed
File copy to VM isn't working anymore on Windows
priority/important-soon
<!-- Please use this template while reporting an issue, providing as much information as possible. Failure to do so may result in a delayed response. Thank you! --> **The exact command to reproduce the issue**: minikube start **The full output of the command that failed**:<details> No error reported when minikube is starting. I didn't manage to display verbose/debug logs when adding --v=7. </details> **The output of the `minikube logs` command**:<details> * ==> Docker <== * -- Logs begin at Sun 2020-02-09 15:22:03 UTC, end at Sun 2020-02-09 16:22:02 UTC. -- * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.070778300Z" level=info msg="loading plugin "io.containerd.service.v1.namespaces-service"..." type=io.containerd.service.v1 * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.070785000Z" level=info msg="loading plugin "io.containerd.service.v1.snapshots-service"..." type=io.containerd.service.v1 * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.070791200Z" level=info msg="loading plugin "io.containerd.runtime.v1.linux"..." type=io.containerd.runtime.v1 * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.070890400Z" level=info msg="loading plugin "io.containerd.runtime.v2.task"..." type=io.containerd.runtime.v2 * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.070956400Z" level=info msg="loading plugin "io.containerd.monitor.v1.cgroups"..." type=io.containerd.monitor.v1 * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.071262000Z" level=info msg="loading plugin "io.containerd.service.v1.tasks-service"..." type=io.containerd.service.v1 * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.071297200Z" level=info msg="loading plugin "io.containerd.internal.v1.restart"..." type=io.containerd.internal.v1 * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.071341400Z" level=info msg="loading plugin "io.containerd.grpc.v1.containers"..." type=io.containerd.grpc.v1 * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.071366200Z" level=info msg="loading plugin "io.containerd.grpc.v1.content"..." type=io.containerd.grpc.v1 * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.071375700Z" level=info msg="loading plugin "io.containerd.grpc.v1.diff"..." type=io.containerd.grpc.v1 * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.071382500Z" level=info msg="loading plugin "io.containerd.grpc.v1.events"..." type=io.containerd.grpc.v1 * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.071388900Z" level=info msg="loading plugin "io.containerd.grpc.v1.healthcheck"..." type=io.containerd.grpc.v1 * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.071395700Z" level=info msg="loading plugin "io.containerd.grpc.v1.images"..." type=io.containerd.grpc.v1 * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.071402500Z" level=info msg="loading plugin "io.containerd.grpc.v1.leases"..." type=io.containerd.grpc.v1 * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.071409200Z" level=info msg="loading plugin "io.containerd.grpc.v1.namespaces"..." type=io.containerd.grpc.v1 * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.071415600Z" level=info msg="loading plugin "io.containerd.internal.v1.opt"..." type=io.containerd.internal.v1 * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.071447700Z" level=info msg="loading plugin "io.containerd.grpc.v1.snapshots"..." type=io.containerd.grpc.v1 * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.071456500Z" level=info msg="loading plugin "io.containerd.grpc.v1.tasks"..." type=io.containerd.grpc.v1 * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.071463500Z" level=info msg="loading plugin "io.containerd.grpc.v1.version"..." type=io.containerd.grpc.v1 * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.071471100Z" level=info msg="loading plugin "io.containerd.grpc.v1.introspection"..." type=io.containerd.grpc.v1 * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.071602300Z" level=info msg=serving... address="/var/run/docker/containerd/containerd-debug.sock" * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.071658800Z" level=info msg=serving... address="/var/run/docker/containerd/containerd.sock" * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.071691000Z" level=info msg="containerd successfully booted in 0.012932s" * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.075872100Z" level=info msg="parsed scheme: \"unix\"" module=grpc * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.075886200Z" level=info msg="scheme \"unix\" not registered, fallback to default scheme" module=grpc * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.075896500Z" level=info msg="ccResolverWrapper: sending update to cc: {[{unix:///var/run/docker/containerd/containerd.sock 0 <nil>}] <nil>}" module=grpc * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.075901900Z" level=info msg="ClientConn switching balancer to \"pick_first\"" module=grpc * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.076483100Z" level=info msg="parsed scheme: \"unix\"" module=grpc * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.076493700Z" level=info msg="scheme \"unix\" not registered, fallback to default scheme" module=grpc * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.076501500Z" level=info msg="ccResolverWrapper: sending update to cc: {[{unix:///var/run/docker/containerd/containerd.sock 0 <nil>}] <nil>}" module=grpc * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.076506600Z" level=info msg="ClientConn switching balancer to \"pick_first\"" module=grpc * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.102168000Z" level=warning msg="Your kernel does not support cgroup blkio weight" * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.102198300Z" level=warning msg="Your kernel does not support cgroup blkio weight_device" * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.102205500Z" level=warning msg="Your kernel does not support cgroup blkio throttle.read_bps_device" * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.102209600Z" level=warning msg="Your kernel does not support cgroup blkio throttle.write_bps_device" * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.102213300Z" level=warning msg="Your kernel does not support cgroup blkio throttle.read_iops_device" * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.102216800Z" level=warning msg="Your kernel does not support cgroup blkio throttle.write_iops_device" * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.102320900Z" level=info msg="Loading containers: start." * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.201520800Z" level=info msg="Loading containers: done." * Feb 09 15:22:14 minikube dockerd[2446]: time="2020-02-09T15:22:14.748699386Z" level=info msg="Docker daemon" commit=633a0ea838 graphdriver(s)=overlay2 version=19.03.5 * Feb 09 15:22:14 minikube dockerd[2446]: time="2020-02-09T15:22:14.748784786Z" level=info msg="Daemon has completed initialization" * Feb 09 15:22:14 minikube dockerd[2446]: time="2020-02-09T15:22:14.778764786Z" level=info msg="API listen on /var/run/docker.sock" * Feb 09 15:22:14 minikube systemd[1]: Started Docker Application Container Engine. * Feb 09 15:22:14 minikube dockerd[2446]: time="2020-02-09T15:22:14.778848486Z" level=info msg="API listen on [::]:2376" * Feb 09 15:23:18 minikube dockerd[2446]: time="2020-02-09T15:23:18.100724410Z" level=info msg="shim containerd-shim started" address="/containerd-shim/moby/bbc3d97848af3255e127c03812852285de598b42070b4c9beb5efd0a2f3d7bca/shim.sock" debug=false pid=4055 * Feb 09 15:23:18 minikube dockerd[2446]: time="2020-02-09T15:23:18.106189577Z" level=info msg="shim containerd-shim started" address="/containerd-shim/moby/bc046d7669fe473893cd4ff2642ce75d3b3b30927a286850e2b64782e5f37b3c/shim.sock" debug=false pid=4069 * Feb 09 15:23:18 minikube dockerd[2446]: time="2020-02-09T15:23:18.119607549Z" level=info msg="shim containerd-shim started" address="/containerd-shim/moby/5fa4b650ae8ebda642fa076f1d32460ad805b6edc21f8604182d3a99cbdee4c0/shim.sock" debug=false pid=4089 * Feb 09 15:23:18 minikube dockerd[2446]: time="2020-02-09T15:23:18.124728724Z" level=info msg="shim containerd-shim started" address="/containerd-shim/moby/16756ddc58412ac11ebe458c09767ebd2dde3d057d5f26bb7b207f4ff4713a3c/shim.sock" debug=false pid=4099 * Feb 09 15:23:18 minikube dockerd[2446]: time="2020-02-09T15:23:18.303385861Z" level=info msg="shim containerd-shim started" address="/containerd-shim/moby/22d8af4f8b0024e98da348de02d9989f95315d69f87e84394bb39e2386b905fb/shim.sock" debug=false pid=4334 * Feb 09 15:23:18 minikube dockerd[2446]: time="2020-02-09T15:23:18.329697819Z" level=info msg="shim containerd-shim started" address="/containerd-shim/moby/022bcbf4051107ac7cb0f3221b9863e6741b3c8a73fa66e3e212619f52c109db/shim.sock" debug=false pid=4367 * Feb 09 15:23:18 minikube dockerd[2446]: time="2020-02-09T15:23:18.332848142Z" level=info msg="shim containerd-shim started" address="/containerd-shim/moby/9c2e529b78d8735ce23e8d79379dea483a84f0fff8aabd371e1aecfa16c11504/shim.sock" debug=false pid=4376 * Feb 09 15:23:18 minikube dockerd[2446]: time="2020-02-09T15:23:18.340200062Z" level=info msg="shim containerd-shim started" address="/containerd-shim/moby/a1faeec091127bad2a535e24688a22b9ff1c308aea19ec85a007247b26220188/shim.sock" debug=false pid=4396 * Feb 09 15:23:44 minikube dockerd[2446]: time="2020-02-09T15:23:44.934584667Z" level=info msg="shim containerd-shim started" address="/containerd-shim/moby/c20ecdf378a05322fcbd85b850a842790c1b0a358281edaa2f64cfb70c44b03a/shim.sock" debug=false pid=5068 * Feb 09 15:23:45 minikube dockerd[2446]: time="2020-02-09T15:23:45.105277818Z" level=info msg="shim containerd-shim started" address="/containerd-shim/moby/02c3afd9bb285a2cb0af8c5211acd573f3130e345c597f10ec7242679d05fd6a/shim.sock" debug=false pid=5114 * Feb 09 15:23:46 minikube dockerd[2446]: time="2020-02-09T15:23:46.123084899Z" level=info msg="shim containerd-shim started" address="/containerd-shim/moby/da500db7aad4bfdc9a2f37c74df75383e2effa5f321ea97760fa1dfd73293715/shim.sock" debug=false pid=5242 * Feb 09 15:23:46 minikube dockerd[2446]: time="2020-02-09T15:23:46.281781063Z" level=info msg="shim containerd-shim started" address="/containerd-shim/moby/57946b280cca1e3a0ba89ff6d15e5fb0cacd00d846279341935ef9d18710266a/shim.sock" debug=false pid=5282 * Feb 09 15:23:48 minikube dockerd[2446]: time="2020-02-09T15:23:48.015340971Z" level=info msg="shim containerd-shim started" address="/containerd-shim/moby/e6614a19f9a159f6dca03fdf4690cb9bde4c11dc74c959aec58f7df37ddde3ac/shim.sock" debug=false pid=5355 * Feb 09 15:23:48 minikube dockerd[2446]: time="2020-02-09T15:23:48.019409757Z" level=info msg="shim containerd-shim started" address="/containerd-shim/moby/5f238a2dd6f24245006c1b87d3869ee3c972e82c573065907f8aac69fd3b15eb/shim.sock" debug=false pid=5364 * Feb 09 15:23:48 minikube dockerd[2446]: time="2020-02-09T15:23:48.279627640Z" level=info msg="shim containerd-shim started" address="/containerd-shim/moby/b694277a7d053b96e0cf21dfdc8e73c3c47902cf0082248271787803e093d97c/shim.sock" debug=false pid=5483 * Feb 09 15:23:48 minikube dockerd[2446]: time="2020-02-09T15:23:48.289742004Z" level=info msg="shim containerd-shim started" address="/containerd-shim/moby/3f923512a9f8332a90b628a1da8f6d9b30ee6cd082d4625c2ef9b9aef69de11d/shim.sock" debug=false pid=5502 * * ==> container status <== * CONTAINER IMAGE CREATED STATE NAME ATTEMPT POD ID * 3f923512a9f83 70f311871ae12 2 minutes ago Running coredns 0 e6614a19f9a15 * b694277a7d053 70f311871ae12 2 minutes ago Running coredns 0 5f238a2dd6f24 * 57946b280cca1 4689081edb103 2 minutes ago Running storage-provisioner 0 da500db7aad4b * 02c3afd9bb285 cba2a99699bdf 2 minutes ago Running kube-proxy 0 c20ecdf378a05 * a1faeec091127 303ce5db0e90d 3 minutes ago Running etcd 0 16756ddc58412 * 9c2e529b78d87 da5fd66c4068c 3 minutes ago Running kube-controller-manager 0 5fa4b650ae8eb * 022bcbf405110 41ef50a5f06a7 3 minutes ago Running kube-apiserver 0 bc046d7669fe4 * 22d8af4f8b002 f52d4c527ef2f 3 minutes ago Running kube-scheduler 0 bbc3d97848af3 * * ==> coredns ["3f923512a9f8"] <== * .:53 * [INFO] plugin/reload: Running configuration MD5 = 4e235fcc3696966e76816bcd9034ebc7 * CoreDNS-1.6.5 * linux/amd64, go1.13.4, c2fd1b2 * * ==> coredns ["b694277a7d05"] <== * .:53 * [INFO] plugin/reload: Running configuration MD5 = 4e235fcc3696966e76816bcd9034ebc7 * CoreDNS-1.6.5 * linux/amd64, go1.13.4, c2fd1b2 * * ==> dmesg <== * [Feb 9 15:21] You have booted with nomodeset. This means your GPU drivers are DISABLED * [ +0.000000] Any video related functionality will be severely degraded, and you may not even be able to suspend the system properly * [ +0.000000] Unless you actually understand what nomodeset does, you should reboot without enabling it * [ +0.046941] MDS CPU bug present and SMT on, data leak possible. See https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/mds.html for more details. * [ +0.000001] TAA CPU bug present and SMT on, data leak possible. See https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/tsx_async_abort.html for more details. * [ +0.000044] #2 #3 * [ +0.022317] acpi PNP0A03:00: fail to add MMCONFIG information, can't access extended PCI configuration space under this bridge. * [ +0.006675] * Found PM-Timer Bug on the chipset. Due to workarounds for a bug, * * this clock source is slow. Consider trying other clock sources * [Feb 9 15:22] Unstable clock detected, switching default tracing clock to "global" * If you want to keep using the local clock, then add: * "trace_clock=local" * on the kernel command line * [ +0.000039] platform regulatory.0: Direct firmware load for regulatory.db failed with error -2 * [ +0.408454] psmouse serio1: trackpoint: failed to get extended button data, assuming 3 buttons * [ +0.680971] systemd[1]: Failed to bump fs.file-max, ignoring: Invalid argument * [ +0.002245] systemd-fstab-generator[1249]: Ignoring "noauto" for root device * [ +0.002567] systemd[1]: File /usr/lib/systemd/system/systemd-journald.service:12 configures an IP firewall (IPAddressDeny=any), but the local system does not support BPF/cgroup based firewalling. * [ +0.000002] systemd[1]: Proceeding WITHOUT firewalling in effect! (This warning is only shown for the first loaded unit using IP firewalling.) * [ +1.472188] NFSD: the nfsdcld client tracking upcall will be removed in 3.10. Please transition to using nfsdcltrack. * [ +0.174629] vboxguest: loading out-of-tree module taints kernel. * [ +0.002674] vboxguest: PCI device not found, probably running on physical hardware. * [ +8.053422] systemd-fstab-generator[2390]: Ignoring "noauto" for root device * [ +0.981549] systemd-fstab-generator[2418]: Ignoring "noauto" for root device * [ +40.516646] systemd-fstab-generator[3337]: Ignoring "noauto" for root device * [ +0.877860] systemd-fstab-generator[3561]: Ignoring "noauto" for root device * [Feb 9 15:23] kauditd_printk_skb: 65 callbacks suppressed * [ +6.712206] systemd-fstab-generator[4798]: Ignoring "noauto" for root device * [ +21.325389] kauditd_printk_skb: 32 callbacks suppressed * [ +7.219938] kauditd_printk_skb: 44 callbacks suppressed * [Feb 9 15:24] NFSD: Unable to end grace period: -110 * * ==> kernel <== * 15:26:32 up 4 min, 0 users, load average: 0.09, 0.17, 0.08 * Linux minikube 4.19.88 #1 SMP Tue Feb 4 22:25:03 PST 2020 x86_64 GNU/Linux * PRETTY_NAME="Buildroot 2019.02.8" * * ==> kube-apiserver ["022bcbf40511"] <== * W0209 15:23:19.922216 1 genericapiserver.go:404] Skipping API discovery.k8s.io/v1alpha1 because it has no resources. * W0209 15:23:19.928317 1 genericapiserver.go:404] Skipping API node.k8s.io/v1alpha1 because it has no resources. * W0209 15:23:19.939588 1 genericapiserver.go:404] Skipping API rbac.authorization.k8s.io/v1alpha1 because it has no resources. * W0209 15:23:19.941700 1 genericapiserver.go:404] Skipping API scheduling.k8s.io/v1alpha1 because it has no resources. * W0209 15:23:19.949731 1 genericapiserver.go:404] Skipping API storage.k8s.io/v1alpha1 because it has no resources. * W0209 15:23:19.967312 1 genericapiserver.go:404] Skipping API apps/v1beta2 because it has no resources. * W0209 15:23:19.967330 1 genericapiserver.go:404] Skipping API apps/v1beta1 because it has no resources. * I0209 15:23:19.973790 1 plugins.go:158] Loaded 11 mutating admission controller(s) successfully in the following order: NamespaceLifecycle,LimitRanger,ServiceAccount,NodeRestriction,TaintNodesByCondition,Priority,DefaultTolerationSeconds,DefaultStorageClass,StorageObjectInUseProtection,MutatingAdmissionWebhook,RuntimeClass. * I0209 15:23:19.973815 1 plugins.go:161] Loaded 7 validating admission controller(s) successfully in the following order: LimitRanger,ServiceAccount,Priority,PersistentVolumeClaimResize,ValidatingAdmissionWebhook,RuntimeClass,ResourceQuota. * I0209 15:23:19.974828 1 client.go:361] parsed scheme: "endpoint" * I0209 15:23:19.974855 1 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://127.0.0.1:2379 0 <nil>}] * I0209 15:23:19.979591 1 client.go:361] parsed scheme: "endpoint" * I0209 15:23:19.979620 1 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://127.0.0.1:2379 0 <nil>}] * I0209 15:23:21.083733 1 dynamic_cafile_content.go:166] Starting request-header::/var/lib/minikube/certs/front-proxy-ca.crt * I0209 15:23:21.083770 1 dynamic_cafile_content.go:166] Starting client-ca-bundle::/var/lib/minikube/certs/ca.crt * I0209 15:23:21.083785 1 dynamic_serving_content.go:129] Starting serving-cert::/var/lib/minikube/certs/apiserver.crt::/var/lib/minikube/certs/apiserver.key * I0209 15:23:21.084157 1 secure_serving.go:178] Serving securely on [::]:8443 * I0209 15:23:21.084190 1 controller.go:81] Starting OpenAPI AggregationController * I0209 15:23:21.084219 1 tlsconfig.go:219] Starting DynamicServingCertificateController * I0209 15:23:21.084531 1 crd_finalizer.go:263] Starting CRDFinalizer * I0209 15:23:21.084627 1 cluster_authentication_trust_controller.go:440] Starting cluster_authentication_trust_controller controller * I0209 15:23:21.084669 1 apiapproval_controller.go:185] Starting KubernetesAPIApprovalPolicyConformantConditionController * I0209 15:23:21.084671 1 shared_informer.go:197] Waiting for caches to sync for cluster_authentication_trust_controller * I0209 15:23:21.084652 1 customresource_discovery_controller.go:208] Starting DiscoveryController * I0209 15:23:21.084656 1 naming_controller.go:288] Starting NamingConditionController * I0209 15:23:21.084660 1 establishing_controller.go:73] Starting EstablishingController * I0209 15:23:21.084665 1 nonstructuralschema_controller.go:191] Starting NonStructuralSchemaConditionController * I0209 15:23:21.084981 1 apiservice_controller.go:94] Starting APIServiceRegistrationController * I0209 15:23:21.085022 1 cache.go:32] Waiting for caches to sync for APIServiceRegistrationController controller * I0209 15:23:21.085089 1 available_controller.go:386] Starting AvailableConditionController * I0209 15:23:21.085121 1 cache.go:32] Waiting for caches to sync for AvailableConditionController controller * I0209 15:23:21.085171 1 dynamic_cafile_content.go:166] Starting client-ca-bundle::/var/lib/minikube/certs/ca.crt * I0209 15:23:21.085211 1 dynamic_cafile_content.go:166] Starting request-header::/var/lib/minikube/certs/front-proxy-ca.crt * I0209 15:23:21.084638 1 controller.go:85] Starting OpenAPI controller * I0209 15:23:21.086075 1 autoregister_controller.go:140] Starting autoregister controller * I0209 15:23:21.086084 1 cache.go:32] Waiting for caches to sync for autoregister controller * E0209 15:23:21.091691 1 controller.go:151] Unable to remove old endpoints from kubernetes service: StorageError: key not found, Code: 1, Key: /registry/masterleases/172.18.109.141, ResourceVersion: 0, AdditionalErrorMsg: * I0209 15:23:21.109117 1 crdregistration_controller.go:111] Starting crd-autoregister controller * I0209 15:23:21.109126 1 shared_informer.go:197] Waiting for caches to sync for crd-autoregister * I0209 15:23:21.185737 1 shared_informer.go:204] Caches are synced for cluster_authentication_trust_controller * I0209 15:23:21.186181 1 cache.go:39] Caches are synced for AvailableConditionController controller * I0209 15:23:21.186236 1 cache.go:39] Caches are synced for APIServiceRegistrationController controller * I0209 15:23:21.193358 1 cache.go:39] Caches are synced for autoregister controller * I0209 15:23:21.210070 1 shared_informer.go:204] Caches are synced for crd-autoregister * I0209 15:23:22.083884 1 controller.go:107] OpenAPI AggregationController: Processing item * I0209 15:23:22.083908 1 controller.go:130] OpenAPI AggregationController: action for item : Nothing (removed from the queue). * I0209 15:23:22.083968 1 controller.go:130] OpenAPI AggregationController: action for item k8s_internal_local_delegation_chain_0000000000: Nothing (removed from the queue). * I0209 15:23:22.086972 1 storage_scheduling.go:133] created PriorityClass system-node-critical with value 2000001000 * I0209 15:23:22.090176 1 storage_scheduling.go:133] created PriorityClass system-cluster-critical with value 2000000000 * I0209 15:23:22.090198 1 storage_scheduling.go:142] all system priority classes are created successfully or already exist. * I0209 15:23:22.349721 1 controller.go:606] quota admission added evaluator for: roles.rbac.authorization.k8s.io * I0209 15:23:22.377055 1 controller.go:606] quota admission added evaluator for: rolebindings.rbac.authorization.k8s.io * W0209 15:23:22.473430 1 lease.go:224] Resetting endpoints for master service "kubernetes" to [172.18.109.141] * I0209 15:23:22.473762 1 controller.go:606] quota admission added evaluator for: endpoints * I0209 15:23:23.225992 1 controller.go:606] quota admission added evaluator for: leases.coordination.k8s.io * I0209 15:23:23.958883 1 controller.go:606] quota admission added evaluator for: serviceaccounts * I0209 15:23:23.967803 1 controller.go:606] quota admission added evaluator for: deployments.apps * I0209 15:23:24.235451 1 controller.go:606] quota admission added evaluator for: daemonsets.apps * I0209 15:23:32.309375 1 controller.go:606] quota admission added evaluator for: replicasets.apps * I0209 15:23:32.348526 1 controller.go:606] quota admission added evaluator for: controllerrevisions.apps * * ==> kube-controller-manager ["9c2e529b78d8"] <== * I0209 15:23:30.499920 1 resource_quota_monitor.go:303] QuotaMonitor running * I0209 15:23:30.507894 1 controllermanager.go:533] Started "deployment" * W0209 15:23:30.507919 1 controllermanager.go:525] Skipping "root-ca-cert-publisher" * I0209 15:23:30.507949 1 deployment_controller.go:152] Starting deployment controller * I0209 15:23:30.507953 1 shared_informer.go:197] Waiting for caches to sync for deployment * I0209 15:23:31.199195 1 controllermanager.go:533] Started "horizontalpodautoscaling" * I0209 15:23:31.199279 1 horizontal.go:156] Starting HPA controller * I0209 15:23:31.199288 1 shared_informer.go:197] Waiting for caches to sync for HPA * I0209 15:23:31.449726 1 controllermanager.go:533] Started "ttl" * I0209 15:23:31.449808 1 ttl_controller.go:116] Starting TTL controller * I0209 15:23:31.449818 1 shared_informer.go:197] Waiting for caches to sync for TTL * I0209 15:23:31.698448 1 controllermanager.go:533] Started "bootstrapsigner" * I0209 15:23:31.698489 1 shared_informer.go:197] Waiting for caches to sync for bootstrap_signer * I0209 15:23:31.952827 1 controllermanager.go:533] Started "serviceaccount" * W0209 15:23:31.952953 1 controllermanager.go:525] Skipping "nodeipam" * I0209 15:23:31.953498 1 serviceaccounts_controller.go:116] Starting service account controller * I0209 15:23:31.953561 1 shared_informer.go:197] Waiting for caches to sync for service account * I0209 15:23:31.953685 1 shared_informer.go:197] Waiting for caches to sync for garbage collector * W0209 15:23:31.960056 1 actual_state_of_world.go:506] Failed to update statusUpdateNeeded field in actual state of world: Failed to set statusUpdateNeeded to needed true, because nodeName="minikube" does not exist * I0209 15:23:31.998629 1 shared_informer.go:204] Caches are synced for certificate-csrapproving * I0209 15:23:31.998633 1 shared_informer.go:204] Caches are synced for bootstrap_signer * I0209 15:23:31.999164 1 shared_informer.go:204] Caches are synced for expand * I0209 15:23:32.030788 1 shared_informer.go:204] Caches are synced for certificate-csrsigning * I0209 15:23:32.048893 1 shared_informer.go:204] Caches are synced for PV protection * I0209 15:23:32.050507 1 shared_informer.go:204] Caches are synced for TTL * I0209 15:23:32.053741 1 shared_informer.go:204] Caches are synced for ClusterRoleAggregator * I0209 15:23:32.252423 1 shared_informer.go:197] Waiting for caches to sync for resource quota * I0209 15:23:32.267774 1 shared_informer.go:204] Caches are synced for PVC protection * I0209 15:23:32.298870 1 shared_informer.go:204] Caches are synced for taint * I0209 15:23:32.298870 1 shared_informer.go:204] Caches are synced for ReplicaSet * I0209 15:23:32.298939 1 taint_manager.go:186] Starting NoExecuteTaintManager * I0209 15:23:32.298958 1 node_lifecycle_controller.go:1443] Initializing eviction metric for zone: * W0209 15:23:32.299011 1 node_lifecycle_controller.go:1058] Missing timestamp for Node minikube. Assuming now as a timestamp. * I0209 15:23:32.299015 1 shared_informer.go:204] Caches are synced for attach detach * I0209 15:23:32.299075 1 node_lifecycle_controller.go:1209] Controller detected that all Nodes are not-Ready. Entering master disruption mode. * I0209 15:23:32.299177 1 event.go:281] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"minikube", UID:"0bd76262-05b0-4350-be09-cb6fe95d5917", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'RegisteredNode' Node minikube event: Registered Node minikube in Controller * I0209 15:23:32.299255 1 shared_informer.go:204] Caches are synced for endpoint * I0209 15:23:32.301874 1 shared_informer.go:204] Caches are synced for disruption * I0209 15:23:32.301888 1 disruption.go:338] Sending events to api server. * I0209 15:23:32.303116 1 shared_informer.go:204] Caches are synced for ReplicationController * I0209 15:23:32.308185 1 shared_informer.go:204] Caches are synced for deployment * I0209 15:23:32.308215 1 shared_informer.go:204] Caches are synced for job * I0209 15:23:32.311175 1 event.go:281] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"kube-system", Name:"coredns", UID:"329e20c5-e656-4310-b781-1aba8093fb09", APIVersion:"apps/v1", ResourceVersion:"180", FieldPath:""}): type: 'Normal' reason: 'ScalingReplicaSet' Scaled up replica set coredns-6955765f44 to 2 * I0209 15:23:32.321513 1 event.go:281] Event(v1.ObjectReference{Kind:"ReplicaSet", Namespace:"kube-system", Name:"coredns-6955765f44", UID:"1c98f2f1-7906-4c98-8fd0-51f1ef943e02", APIVersion:"apps/v1", ResourceVersion:"312", FieldPath:""}): type: 'Normal' reason: 'SuccessfulCreate' Created pod: coredns-6955765f44-wprg2 * I0209 15:23:32.329100 1 event.go:281] Event(v1.ObjectReference{Kind:"ReplicaSet", Namespace:"kube-system", Name:"coredns-6955765f44", UID:"1c98f2f1-7906-4c98-8fd0-51f1ef943e02", APIVersion:"apps/v1", ResourceVersion:"312", FieldPath:""}): type: 'Normal' reason: 'SuccessfulCreate' Created pod: coredns-6955765f44-2cfq5 * I0209 15:23:32.346845 1 shared_informer.go:204] Caches are synced for daemon sets * I0209 15:23:32.348899 1 shared_informer.go:204] Caches are synced for stateful set * I0209 15:23:32.349483 1 shared_informer.go:204] Caches are synced for GC * I0209 15:23:32.351331 1 event.go:281] Event(v1.ObjectReference{Kind:"DaemonSet", Namespace:"kube-system", Name:"kube-proxy", UID:"7610c387-048f-471f-bb42-e89e44c143c0", APIVersion:"apps/v1", ResourceVersion:"185", FieldPath:""}): type: 'Normal' reason: 'SuccessfulCreate' Created pod: kube-proxy-tkmbk * I0209 15:23:32.357681 1 shared_informer.go:204] Caches are synced for persistent volume * E0209 15:23:32.364751 1 daemon_controller.go:290] kube-system/kube-proxy failed with : error storing status for daemon set &v1.DaemonSet{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"kube-proxy", GenerateName:"", Namespace:"kube-system", SelfLink:"/apis/apps/v1/namespaces/kube-system/daemonsets/kube-proxy", UID:"7610c387-048f-471f-bb42-e89e44c143c0", ResourceVersion:"185", Generation:1, CreationTimestamp:v1.Time{Time:time.Time{wall:0x0, ext:63716858604, loc:(*time.Location)(0x6b971e0)}}, DeletionTimestamp:(*v1.Time)(nil), DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-proxy"}, Annotations:map[string]string{"deprecated.daemonset.template.generation":"1"}, OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ClusterName:"", ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v1.DaemonSetSpec{Selector:(*v1.LabelSelector)(0xc00046c880), Template:v1.PodTemplateSpec{ObjectMeta:v1.ObjectMeta{Name:"", GenerateName:"", Namespace:"", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, DeletionTimestamp:(*v1.Time)(nil), DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-proxy"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ClusterName:"", ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v1.PodSpec{Volumes:[]v1.Volume{v1.Volume{Name:"kube-proxy", VolumeSource:v1.VolumeSource{HostPath:(*v1.HostPathVolumeSource)(nil), EmptyDir:(*v1.EmptyDirVolumeSource)(nil), GCEPersistentDisk:(*v1.GCEPersistentDiskVolumeSource)(nil), AWSElasticBlockStore:(*v1.AWSElasticBlockStoreVolumeSource)(nil), GitRepo:(*v1.GitRepoVolumeSource)(nil), Secret:(*v1.SecretVolumeSource)(nil), NFS:(*v1.NFSVolumeSource)(nil), ISCSI:(*v1.ISCSIVolumeSource)(nil), Glusterfs:(*v1.GlusterfsVolumeSource)(nil), PersistentVolumeClaim:(*v1.PersistentVolumeClaimVolumeSource)(nil), RBD:(*v1.RBDVolumeSource)(nil), FlexVolume:(*v1.FlexVolumeSource)(nil), Cinder:(*v1.CinderVolumeSource)(nil), CephFS:(*v1.CephFSVolumeSource)(nil), Flocker:(*v1.FlockerVolumeSource)(nil), DownwardAPI:(*v1.DownwardAPIVolumeSource)(nil), FC:(*v1.FCVolumeSource)(nil), AzureFile:(*v1.AzureFileVolumeSource)(nil), ConfigMap:(*v1.ConfigMapVolumeSource)(0xc00094f3c0), VsphereVolume:(*v1.VsphereVirtualDiskVolumeSource)(nil), Quobyte:(*v1.QuobyteVolumeSource)(nil), AzureDisk:(*v1.AzureDiskVolumeSource)(nil), PhotonPersistentDisk:(*v1.PhotonPersistentDiskVolumeSource)(nil), Projected:(*v1.ProjectedVolumeSource)(nil), PortworxVolume:(*v1.PortworxVolumeSource)(nil), ScaleIO:(*v1.ScaleIOVolumeSource)(nil), StorageOS:(*v1.StorageOSVolumeSource)(nil), CSI:(*v1.CSIVolumeSource)(nil)}}, v1.Volume{Name:"xtables-lock", VolumeSource:v1.VolumeSource{HostPath:(*v1.HostPathVolumeSource)(0xc00046ca00), EmptyDir:(*v1.EmptyDirVolumeSource)(nil), GCEPersistentDisk:(*v1.GCEPersistentDiskVolumeSource)(nil), AWSElasticBlockStore:(*v1.AWSElasticBlockStoreVolumeSource)(nil), GitRepo:(*v1.GitRepoVolumeSource)(nil), Secret:(*v1.SecretVolumeSource)(nil), NFS:(*v1.NFSVolumeSource)(nil), ISCSI:(*v1.ISCSIVolumeSource)(nil), Glusterfs:(*v1.GlusterfsVolumeSource)(nil), PersistentVolumeClaim:(*v1.PersistentVolumeClaimVolumeSource)(nil), RBD:(*v1.RBDVolumeSource)(nil), FlexVolume:(*v1.FlexVolumeSource)(nil), Cinder:(*v1.CinderVolumeSource)(nil), CephFS:(*v1.CephFSVolumeSource)(nil), Flocker:(*v1.FlockerVolumeSource)(nil), DownwardAPI:(*v1.DownwardAPIVolumeSource)(nil), FC:(*v1.FCVolumeSource)(nil), AzureFile:(*v1.AzureFileVolumeSource)(nil), ConfigMap:(*v1.ConfigMapVolumeSource)(nil), VsphereVolume:(*v1.VsphereVirtualDiskVolumeSource)(nil), Quobyte:(*v1.QuobyteVolumeSource)(nil), AzureDisk:(*v1.AzureDiskVolumeSource)(nil), PhotonPersistentDisk:(*v1.PhotonPersistentDiskVolumeSource)(nil), Projected:(*v1.ProjectedVolumeSource)(nil), PortworxVolume:(*v1.PortworxVolumeSource)(nil), ScaleIO:(*v1.ScaleIOVolumeSource)(nil), StorageOS:(*v1.StorageOSVolumeSource)(nil), CSI:(*v1.CSIVolumeSource)(nil)}}, v1.Volume{Name:"lib-modules", VolumeSource:v1.VolumeSource{HostPath:(*v1.HostPathVolumeSource)(0xc00046cbe0), EmptyDir:(*v1.EmptyDirVolumeSource)(nil), GCEPersistentDisk:(*v1.GCEPersistentDiskVolumeSource)(nil), AWSElasticBlockStore:(*v1.AWSElasticBlockStoreVolumeSource)(nil), GitRepo:(*v1.GitRepoVolumeSource)(nil), Secret:(*v1.SecretVolumeSource)(nil), NFS:(*v1.NFSVolumeSource)(nil), ISCSI:(*v1.ISCSIVolumeSource)(nil), Glusterfs:(*v1.GlusterfsVolumeSource)(nil), PersistentVolumeClaim:(*v1.PersistentVolumeClaimVolumeSource)(nil), RBD:(*v1.RBDVolumeSource)(nil), FlexVolume:(*v1.FlexVolumeSource)(nil), Cinder:(*v1.CinderVolumeSource)(nil), CephFS:(*v1.CephFSVolumeSource)(nil), Flocker:(*v1.FlockerVolumeSource)(nil), DownwardAPI:(*v1.DownwardAPIVolumeSource)(nil), FC:(*v1.FCVolumeSource)(nil), AzureFile:(*v1.AzureFileVolumeSource)(nil), ConfigMap:(*v1.ConfigMapVolumeSource)(nil), VsphereVolume:(*v1.VsphereVirtualDiskVolumeSource)(nil), Quobyte:(*v1.QuobyteVolumeSource)(nil), AzureDisk:(*v1.AzureDiskVolumeSource)(nil), PhotonPersistentDisk:(*v1.PhotonPersistentDiskVolumeSource)(nil), Projected:(*v1.ProjectedVolumeSource)(nil), PortworxVolume:(*v1.PortworxVolumeSource)(nil), ScaleIO:(*v1.ScaleIOVolumeSource)(nil), StorageOS:(*v1.StorageOSVolumeSource)(nil), CSI:(*v1.CSIVolumeSource)(nil)}}}, InitContainers:[]v1.Container(nil), Containers:[]v1.Container{v1.Container{Name:"kube-proxy", Image:"k8s.gcr.io/kube-proxy:v1.17.2", Command:[]string{"/usr/local/bin/kube-proxy", "--config=/var/lib/kube-proxy/config.conf", "--hostname-override=$(NODE_NAME)"}, Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar{v1.EnvVar{Name:"NODE_NAME", Value:"", ValueFrom:(*v1.EnvVarSource)(0xc00046d0c0)}}, Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList(nil)}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"kube-proxy", ReadOnly:false, MountPath:"/var/lib/kube-proxy", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}, v1.VolumeMount{Name:"xtables-lock", ReadOnly:false, MountPath:"/run/xtables.lock", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}, v1.VolumeMount{Name:"lib-modules", ReadOnly:true, MountPath:"/lib/modules", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), StartupProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(0xc000ba0730), Stdin:false, StdinOnce:false, TTY:false}}, EphemeralContainers:[]v1.EphemeralContainer(nil), RestartPolicy:"Always", TerminationGracePeriodSeconds:(*int64)(0xc000528dd8), ActiveDeadlineSeconds:(*int64)(nil), DNSPolicy:"ClusterFirst", NodeSelector:map[string]string{"beta.kubernetes.io/os":"linux"}, ServiceAccountName:"kube-proxy", DeprecatedServiceAccount:"kube-proxy", AutomountServiceAccountToken:(*bool)(nil), NodeName:"", HostNetwork:true, HostPID:false, HostIPC:false, ShareProcessNamespace:(*bool)(nil), SecurityContext:(*v1.PodSecurityContext)(0xc000efe960), ImagePullSecrets:[]v1.LocalObjectReference(nil), Hostname:"", Subdomain:"", Affinity:(*v1.Affinity)(nil), SchedulerName:"default-scheduler", Tolerations:[]v1.Toleration{v1.Toleration{Key:"CriticalAddonsOnly", Operator:"Exists", Value:"", Effect:"", TolerationSeconds:(*int64)(nil)}, v1.Toleration{Key:"", Operator:"Exists", Value:"", Effect:"", TolerationSeconds:(*int64)(nil)}}, HostAliases:[]v1.HostAlias(nil), PriorityClassName:"system-node-critical", Priority:(*int32)(nil), DNSConfig:(*v1.PodDNSConfig)(nil), ReadinessGates:[]v1.PodReadinessGate(nil), RuntimeClassName:(*string)(nil), EnableServiceLinks:(*bool)(nil), PreemptionPolicy:(*v1.PreemptionPolicy)(nil), Overhead:v1.ResourceList(nil), TopologySpreadConstraints:[]v1.TopologySpreadConstraint(nil)}}, UpdateStrategy:v1.DaemonSetUpdateStrategy{Type:"RollingUpdate", RollingUpdate:(*v1.RollingUpdateDaemonSet)(0xc00000e6c0)}, MinReadySeconds:0, RevisionHistoryLimit:(*int32)(0xc000528e58)}, Status:v1.DaemonSetStatus{CurrentNumberScheduled:0, NumberMisscheduled:0, DesiredNumberScheduled:0, NumberReady:0, ObservedGeneration:0, UpdatedNumberScheduled:0, NumberAvailable:0, NumberUnavailable:0, CollisionCount:(*int32)(nil), Conditions:[]v1.DaemonSetCondition(nil)}}: Operation cannot be fulfilled on daemonsets.apps "kube-proxy": the object has been modified; please apply your changes to the latest version and try again * I0209 15:23:32.399450 1 shared_informer.go:204] Caches are synced for HPA * I0209 15:23:32.552906 1 shared_informer.go:204] Caches are synced for resource quota * I0209 15:23:32.552931 1 shared_informer.go:204] Caches are synced for namespace * I0209 15:23:32.553917 1 shared_informer.go:204] Caches are synced for service account * I0209 15:23:32.553990 1 shared_informer.go:204] Caches are synced for garbage collector * I0209 15:23:32.597549 1 shared_informer.go:204] Caches are synced for garbage collector * I0209 15:23:32.597595 1 garbagecollector.go:138] Garbage collector: all resource monitors have synced. Proceeding to collect garbage * I0209 15:23:32.600066 1 shared_informer.go:204] Caches are synced for resource quota * I0209 15:23:47.300428 1 node_lifecycle_controller.go:1236] Controller detected that some Nodes are Ready. Exiting master disruption mode. * * ==> kube-proxy ["02c3afd9bb28"] <== * W0209 15:23:45.210525 1 server_others.go:323] Unknown proxy mode "", assuming iptables proxy * I0209 15:23:45.214518 1 node.go:135] Successfully retrieved node IP: 172.18.109.141 * I0209 15:23:45.214545 1 server_others.go:145] Using iptables Proxier. * W0209 15:23:45.214634 1 proxier.go:286] clusterCIDR not specified, unable to distinguish between internal and external traffic * I0209 15:23:45.214787 1 server.go:571] Version: v1.17.2 * I0209 15:23:45.214994 1 conntrack.go:100] Set sysctl 'net/netfilter/nf_conntrack_max' to 131072 * I0209 15:23:45.215013 1 conntrack.go:52] Setting nf_conntrack_max to 131072 * I0209 15:23:45.215440 1 conntrack.go:83] Setting conntrack hashsize to 32768 * I0209 15:23:45.221130 1 conntrack.go:100] Set sysctl 'net/netfilter/nf_conntrack_tcp_timeout_established' to 86400 * I0209 15:23:45.221190 1 conntrack.go:100] Set sysctl 'net/netfilter/nf_conntrack_tcp_timeout_close_wait' to 3600 * I0209 15:23:45.221287 1 config.go:313] Starting service config controller * I0209 15:23:45.221293 1 shared_informer.go:197] Waiting for caches to sync for service config * I0209 15:23:45.221314 1 config.go:131] Starting endpoints config controller * I0209 15:23:45.221335 1 shared_informer.go:197] Waiting for caches to sync for endpoints config * I0209 15:23:45.321702 1 shared_informer.go:204] Caches are synced for endpoints config * I0209 15:23:45.321705 1 shared_informer.go:204] Caches are synced for service config * * ==> kube-scheduler ["22d8af4f8b00"] <== * I0209 15:23:18.670592 1 serving.go:312] Generated self-signed cert in-memory * W0209 15:23:18.772671 1 configmap_cafile_content.go:102] unable to load initial CA bundle for: "client-ca::kube-system::extension-apiserver-authentication::client-ca-file" due to: configmap "extension-apiserver-authentication" not found * W0209 15:23:18.772931 1 configmap_cafile_content.go:102] unable to load initial CA bundle for: "client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" due to: configmap "extension-apiserver-authentication" not found * W0209 15:23:21.112588 1 authentication.go:348] Unable to get configmap/extension-apiserver-authentication in kube-system. Usually fixed by 'kubectl create rolebinding -n kube-system ROLEBINDING_NAME --role=extension-apiserver-authentication-reader --serviceaccount=YOUR_NS:YOUR_SA' * W0209 15:23:21.112729 1 authentication.go:296] Error looking up in-cluster authentication configuration: configmaps "extension-apiserver-authentication" is forbidden: User "system:kube-scheduler" cannot get resource "configmaps" in API group "" in the namespace "kube-system" * W0209 15:23:21.112783 1 authentication.go:297] Continuing without authentication configuration. This may treat all requests as anonymous. * W0209 15:23:21.112845 1 authentication.go:298] To require authentication configuration lookup to succeed, set --authentication-tolerate-lookup-failure=false * W0209 15:23:21.120507 1 authorization.go:47] Authorization is disabled * W0209 15:23:21.120550 1 authentication.go:92] Authentication is disabled * I0209 15:23:21.120569 1 deprecated_insecure_serving.go:51] Serving healthz insecurely on [::]:10251 * I0209 15:23:21.121504 1 configmap_cafile_content.go:205] Starting client-ca::kube-system::extension-apiserver-authentication::client-ca-file * I0209 15:23:21.121514 1 shared_informer.go:197] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file * I0209 15:23:21.121674 1 secure_serving.go:178] Serving securely on 127.0.0.1:10259 * I0209 15:23:21.121719 1 tlsconfig.go:219] Starting DynamicServingCertificateController * E0209 15:23:21.123202 1 reflector.go:153] k8s.io/client-go/informers/factory.go:135: Failed to list *v1beta1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User "system:kube-scheduler" cannot list resource "poddisruptionbudgets" in API group "policy" at the cluster scope * E0209 15:23:21.123205 1 reflector.go:153] k8s.io/apiserver/pkg/server/dynamiccertificates/configmap_cafile_content.go:209: Failed to list *v1.ConfigMap: configmaps "extension-apiserver-authentication" is forbidden: User "system:kube-scheduler" cannot list resource "configmaps" in API group "" in the namespace "kube-system" * E0209 15:23:21.123257 1 reflector.go:153] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.Service: services is forbidden: User "system:kube-scheduler" cannot list resource "services" in API group "" at the cluster scope * E0209 15:23:21.123279 1 reflector.go:153] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumes" in API group "" at the cluster scope * E0209 15:23:21.123292 1 reflector.go:153] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User "system:kube-scheduler" cannot list resource "replicationcontrollers" in API group "" at the cluster scope * E0209 15:23:21.123335 1 reflector.go:153] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "storageclasses" in API group "storage.k8s.io" at the cluster scope * E0209 15:23:21.123356 1 reflector.go:153] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.Node: nodes is forbidden: User "system:kube-scheduler" cannot list resource "nodes" in API group "" at the cluster scope * E0209 15:23:21.123406 1 reflector.go:153] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User "system:kube-scheduler" cannot list resource "statefulsets" in API group "apps" at the cluster scope * E0209 15:23:21.123412 1 reflector.go:153] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csinodes" in API group "storage.k8s.io" at the cluster scope * E0209 15:23:21.123445 1 reflector.go:153] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User "system:kube-scheduler" cannot list resource "replicasets" in API group "apps" at the cluster scope * E0209 15:23:21.123479 1 reflector.go:153] k8s.io/kubernetes/cmd/kube-scheduler/app/server.go:246: Failed to list *v1.Pod: pods is forbidden: User "system:kube-scheduler" cannot list resource "pods" in API group "" at the cluster scope * E0209 15:23:21.123508 1 reflector.go:153] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumeclaims" in API group "" at the cluster scope * E0209 15:23:22.123944 1 reflector.go:153] k8s.io/client-go/informers/factory.go:135: Failed to list *v1beta1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User "system:kube-scheduler" cannot list resource "poddisruptionbudgets" in API group "policy" at the cluster scope * E0209 15:23:22.125658 1 reflector.go:153] k8s.io/apiserver/pkg/server/dynamiccertificates/configmap_cafile_content.go:209: Failed to list *v1.ConfigMap: configmaps "extension-apiserver-authentication" is forbidden: User "system:kube-scheduler" cannot list resource "configmaps" in API group "" in the namespace "kube-system" * E0209 15:23:22.126717 1 reflector.go:153] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.Service: services is forbidden: User "system:kube-scheduler" cannot list resource "services" in API group "" at the cluster scope * E0209 15:23:22.127905 1 reflector.go:153] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumes" in API group "" at the cluster scope * E0209 15:23:22.129170 1 reflector.go:153] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User "system:kube-scheduler" cannot list resource "replicationcontrollers" in API group "" at the cluster scope * E0209 15:23:22.130373 1 reflector.go:153] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "storageclasses" in API group "storage.k8s.io" at the cluster scope * E0209 15:23:22.131346 1 reflector.go:153] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.Node: nodes is forbidden: User "system:kube-scheduler" cannot list resource "nodes" in API group "" at the cluster scope * E0209 15:23:22.132586 1 reflector.go:153] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User "system:kube-scheduler" cannot list resource "statefulsets" in API group "apps" at the cluster scope * E0209 15:23:22.133856 1 reflector.go:153] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csinodes" in API group "storage.k8s.io" at the cluster scope * E0209 15:23:22.134929 1 reflector.go:153] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User "system:kube-scheduler" cannot list resource "replicasets" in API group "apps" at the cluster scope * E0209 15:23:22.135943 1 reflector.go:153] k8s.io/kubernetes/cmd/kube-scheduler/app/server.go:246: Failed to list *v1.Pod: pods is forbidden: User "system:kube-scheduler" cannot list resource "pods" in API group "" at the cluster scope * E0209 15:23:22.136970 1 reflector.go:153] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumeclaims" in API group "" at the cluster scope * I0209 15:23:23.221808 1 shared_informer.go:204] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::client-ca-file * I0209 15:23:23.221910 1 leaderelection.go:242] attempting to acquire leader lease kube-system/kube-scheduler... * I0209 15:23:23.227348 1 leaderelection.go:252] successfully acquired lease kube-system/kube-scheduler * E0209 15:23:32.336961 1 factory.go:494] pod is already present in the activeQ * * ==> kubelet <== * -- Logs begin at Sun 2020-02-09 15:22:03 UTC, end at Sun 2020-02-09 16:22:02 UTC. -- * Feb 09 15:23:24 minikube kubelet[4807]: W0209 15:23:24.059376 4807 hostport_manager.go:69] The binary conntrack is not installed, this can cause failures in network connection cleanup. * Feb 09 15:23:24 minikube kubelet[4807]: I0209 15:23:24.060087 4807 docker_service.go:255] Docker cri networking managed by kubernetes.io/no-op * Feb 09 15:23:24 minikube kubelet[4807]: I0209 15:23:24.068436 4807 docker_service.go:260] Docker Info: &{ID:WGJF:QQZY:GQVU:4QER:IZUQ:5XPQ:D4JK:DXTO:LBGI:575C:U3YU:3SMB Containers:8 ContainersRunning:8 ContainersPaused:0 ContainersStopped:0 Images:10 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Native Overlay Diff true]] SystemStatus:[] Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:[] Log:[awslogs fluentd gcplogs gelf journald json-file local logentries splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:true KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6tables:true Debug:false NFd:66 OomKillDisable:true NGoroutines:75 SystemTime:2020-02-09T15:23:24.060737784Z LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:4.19.88 OperatingSystem:Buildroot 2019.02.8 OSType:linux Architecture:x86_64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:0xc00072e070 NCPU:4 MemTotal:4131684352 GenericResources:[] DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:minikube Labels:[provider=hyperv] ExperimentalBuild:false ServerVersion:19.03.5 ClusterStore: ClusterAdvertise: Runtimes:map[runc:{Path:runc Args:[]}] DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:[] Nodes:0 Managers:0 Cluster:<nil> Warnings:[]} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:b34a5c8af56e510852c35414db4c1f4fa6172339 Expected:b34a5c8af56e510852c35414db4c1f4fa6172339} RuncCommit:{ID:d736ef14f0288d6993a1845745d6756cfc9ddd5a Expected:d736ef14f0288d6993a1845745d6756cfc9ddd5a} InitCommit:{ID:fec3683 Expected:fec3683} SecurityOptions:[name=seccomp,profile=default] ProductLicense:Community Engine Warnings:[]} * Feb 09 15:23:24 minikube kubelet[4807]: I0209 15:23:24.068498 4807 docker_service.go:273] Setting cgroupDriver to cgroupfs * Feb 09 15:23:24 minikube kubelet[4807]: I0209 15:23:24.075778 4807 remote_runtime.go:59] parsed scheme: "" * Feb 09 15:23:24 minikube kubelet[4807]: I0209 15:23:24.075802 4807 remote_runtime.go:59] scheme "" not registered, fallback to default scheme * Feb 09 15:23:24 minikube kubelet[4807]: I0209 15:23:24.075818 4807 passthrough.go:48] ccResolverWrapper: sending update to cc: {[{/var/run/dockershim.sock 0 <nil>}] <nil>} * Feb 09 15:23:24 minikube kubelet[4807]: I0209 15:23:24.075823 4807 clientconn.go:577] ClientConn switching balancer to "pick_first" * Feb 09 15:23:24 minikube kubelet[4807]: I0209 15:23:24.075854 4807 remote_image.go:50] parsed scheme: "" * Feb 09 15:23:24 minikube kubelet[4807]: I0209 15:23:24.075860 4807 remote_image.go:50] scheme "" not registered, fallback to default scheme * Feb 09 15:23:24 minikube kubelet[4807]: I0209 15:23:24.075866 4807 passthrough.go:48] ccResolverWrapper: sending update to cc: {[{/var/run/dockershim.sock 0 <nil>}] <nil>} * Feb 09 15:23:24 minikube kubelet[4807]: I0209 15:23:24.075870 4807 clientconn.go:577] ClientConn switching balancer to "pick_first" * Feb 09 15:23:44 minikube kubelet[4807]: E0209 15:23:44.383561 4807 aws_credentials.go:77] while getting AWS credentials NoCredentialProviders: no valid providers in chain. Deprecated. * Feb 09 15:23:44 minikube kubelet[4807]: For verbose messaging see aws.Config.CredentialsChainVerboseErrors * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.392129 4807 kuberuntime_manager.go:211] Container runtime docker initialized, version: 19.03.5, apiVersion: 1.40.0 * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.400992 4807 server.go:1113] Started kubelet * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.401041 4807 server.go:143] Starting to listen on 0.0.0.0:10250 * Feb 09 15:23:44 minikube kubelet[4807]: E0209 15:23:44.401049 4807 kubelet.go:1302] Image garbage collection failed once. Stats initialization may not have completed yet: failed to get imageFs info: unable to find data in memory cache * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.401546 4807 server.go:354] Adding debug handlers to kubelet server. * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.401826 4807 fs_resource_analyzer.go:64] Starting FS ResourceAnalyzer * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.403047 4807 volume_manager.go:265] Starting Kubelet Volume Manager * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.403180 4807 desired_state_of_world_populator.go:138] Desired state populator starts to run * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.410733 4807 status_manager.go:157] Starting to sync pod status with apiserver * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.410760 4807 kubelet.go:1820] Starting kubelet main sync loop. * Feb 09 15:23:44 minikube kubelet[4807]: E0209 15:23:44.410816 4807 kubelet.go:1844] skipping pod synchronization - [container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful] * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.477297 4807 cpu_manager.go:173] [cpumanager] starting with none policy * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.477320 4807 cpu_manager.go:174] [cpumanager] reconciling every 10s * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.477329 4807 policy_none.go:43] [cpumanager] none policy: Start * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.478181 4807 plugin_manager.go:114] Starting Kubelet Plugin Manager * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.503276 4807 kubelet_node_status.go:294] Setting node annotation to enable volume controller attach/detach * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.527413 4807 kubelet_node_status.go:70] Attempting to register node minikube * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.704211 4807 reconciler.go:209] operationExecutor.VerifyControllerAttachedVolume started for volume "usr-share-ca-certificates" (UniqueName: "kubernetes.io/host-path/0ae6cf402f641e9b795a3aebca394220-usr-share-ca-certificates") pod "kube-controller-manager-minikube" (UID: "0ae6cf402f641e9b795a3aebca394220") * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.704446 4807 reconciler.go:209] operationExecutor.VerifyControllerAttachedVolume started for volume "xtables-lock" (UniqueName: "kubernetes.io/host-path/aec7f878-f624-400e-b3ac-7879ab47ec0c-xtables-lock") pod "kube-proxy-tkmbk" (UID: "aec7f878-f624-400e-b3ac-7879ab47ec0c") * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.704519 4807 reconciler.go:209] operationExecutor.VerifyControllerAttachedVolume started for volume "kube-proxy-token-92jtn" (UniqueName: "kubernetes.io/secret/aec7f878-f624-400e-b3ac-7879ab47ec0c-kube-proxy-token-92jtn") pod "kube-proxy-tkmbk" (UID: "aec7f878-f624-400e-b3ac-7879ab47ec0c") * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.704592 4807 reconciler.go:209] operationExecutor.VerifyControllerAttachedVolume started for volume "ca-certs" (UniqueName: "kubernetes.io/host-path/c1b171d918aa89531bd5657acb065f84-ca-certs") pod "kube-apiserver-minikube" (UID: "c1b171d918aa89531bd5657acb065f84") * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.704661 4807 reconciler.go:209] operationExecutor.VerifyControllerAttachedVolume started for volume "k8s-certs" (UniqueName: "kubernetes.io/host-path/c1b171d918aa89531bd5657acb065f84-k8s-certs") pod "kube-apiserver-minikube" (UID: "c1b171d918aa89531bd5657acb065f84") * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.704803 4807 reconciler.go:209] operationExecutor.VerifyControllerAttachedVolume started for volume "usr-share-ca-certificates" (UniqueName: "kubernetes.io/host-path/c1b171d918aa89531bd5657acb065f84-usr-share-ca-certificates") pod "kube-apiserver-minikube" (UID: "c1b171d918aa89531bd5657acb065f84") * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.705005 4807 reconciler.go:209] operationExecutor.VerifyControllerAttachedVolume started for volume "kubeconfig" (UniqueName: "kubernetes.io/host-path/0ae6cf402f641e9b795a3aebca394220-kubeconfig") pod "kube-controller-manager-minikube" (UID: "0ae6cf402f641e9b795a3aebca394220") * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.705132 4807 reconciler.go:209] operationExecutor.VerifyControllerAttachedVolume started for volume "kubeconfig" (UniqueName: "kubernetes.io/host-path/9c994ea62a2d8d6f1bb7498f10aa6fcf-kubeconfig") pod "kube-scheduler-minikube" (UID: "9c994ea62a2d8d6f1bb7498f10aa6fcf") * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.705208 4807 reconciler.go:209] operationExecutor.VerifyControllerAttachedVolume started for volume "kube-proxy" (UniqueName: "kubernetes.io/configmap/aec7f878-f624-400e-b3ac-7879ab47ec0c-kube-proxy") pod "kube-proxy-tkmbk" (UID: "aec7f878-f624-400e-b3ac-7879ab47ec0c") * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.705283 4807 reconciler.go:209] operationExecutor.VerifyControllerAttachedVolume started for volume "lib-modules" (UniqueName: "kubernetes.io/host-path/aec7f878-f624-400e-b3ac-7879ab47ec0c-lib-modules") pod "kube-proxy-tkmbk" (UID: "aec7f878-f624-400e-b3ac-7879ab47ec0c") * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.705340 4807 reconciler.go:209] operationExecutor.VerifyControllerAttachedVolume started for volume "etcd-certs" (UniqueName: "kubernetes.io/host-path/8df3b32a9a938b2ec4d0d2782b34f10e-etcd-certs") pod "etcd-minikube" (UID: "8df3b32a9a938b2ec4d0d2782b34f10e") * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.705395 4807 reconciler.go:209] operationExecutor.VerifyControllerAttachedVolume started for volume "ca-certs" (UniqueName: "kubernetes.io/host-path/0ae6cf402f641e9b795a3aebca394220-ca-certs") pod "kube-controller-manager-minikube" (UID: "0ae6cf402f641e9b795a3aebca394220") * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.705451 4807 reconciler.go:209] operationExecutor.VerifyControllerAttachedVolume started for volume "flexvolume-dir" (UniqueName: "kubernetes.io/host-path/0ae6cf402f641e9b795a3aebca394220-flexvolume-dir") pod "kube-controller-manager-minikube" (UID: "0ae6cf402f641e9b795a3aebca394220") * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.705665 4807 reconciler.go:209] operationExecutor.VerifyControllerAttachedVolume started for volume "etcd-data" (UniqueName: "kubernetes.io/host-path/8df3b32a9a938b2ec4d0d2782b34f10e-etcd-data") pod "etcd-minikube" (UID: "8df3b32a9a938b2ec4d0d2782b34f10e") * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.705942 4807 reconciler.go:209] operationExecutor.VerifyControllerAttachedVolume started for volume "k8s-certs" (UniqueName: "kubernetes.io/host-path/0ae6cf402f641e9b795a3aebca394220-k8s-certs") pod "kube-controller-manager-minikube" (UID: "0ae6cf402f641e9b795a3aebca394220") * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.706094 4807 reconciler.go:156] Reconciler: start to sync state * Feb 09 15:23:45 minikube kubelet[4807]: I0209 15:23:45.394000 4807 kubelet_node_status.go:112] Node minikube was previously registered * Feb 09 15:23:45 minikube kubelet[4807]: I0209 15:23:45.394079 4807 kubelet_node_status.go:73] Successfully registered node minikube * Feb 09 15:23:45 minikube kubelet[4807]: I0209 15:23:45.508553 4807 reconciler.go:209] operationExecutor.VerifyControllerAttachedVolume started for volume "storage-provisioner-token-dprnq" (UniqueName: "kubernetes.io/secret/20756ee1-7bca-4a20-b7bb-f46230f3c719-storage-provisioner-token-dprnq") pod "storage-provisioner" (UID: "20756ee1-7bca-4a20-b7bb-f46230f3c719") * Feb 09 15:23:45 minikube kubelet[4807]: I0209 15:23:45.508614 4807 reconciler.go:209] operationExecutor.VerifyControllerAttachedVolume started for volume "tmp" (UniqueName: "kubernetes.io/host-path/20756ee1-7bca-4a20-b7bb-f46230f3c719-tmp") pod "storage-provisioner" (UID: "20756ee1-7bca-4a20-b7bb-f46230f3c719") * Feb 09 15:23:46 minikube kubelet[4807]: E0209 15:23:46.008320 4807 kubelet.go:1662] Failed creating a mirror pod for "kube-scheduler-minikube_kube-system(9c994ea62a2d8d6f1bb7498f10aa6fcf)": pods "kube-scheduler-minikube" already exists * Feb 09 15:23:47 minikube kubelet[4807]: I0209 15:23:47.214091 4807 reconciler.go:209] operationExecutor.VerifyControllerAttachedVolume started for volume "coredns-token-gc9r4" (UniqueName: "kubernetes.io/secret/61df07f7-738e-4543-9398-a144fac001b5-coredns-token-gc9r4") pod "coredns-6955765f44-wprg2" (UID: "61df07f7-738e-4543-9398-a144fac001b5") * Feb 09 15:23:47 minikube kubelet[4807]: I0209 15:23:47.214457 4807 reconciler.go:209] operationExecutor.VerifyControllerAttachedVolume started for volume "config-volume" (UniqueName: "kubernetes.io/configmap/61df07f7-738e-4543-9398-a144fac001b5-config-volume") pod "coredns-6955765f44-wprg2" (UID: "61df07f7-738e-4543-9398-a144fac001b5") * Feb 09 15:23:47 minikube kubelet[4807]: I0209 15:23:47.214492 4807 reconciler.go:209] operationExecutor.VerifyControllerAttachedVolume started for volume "coredns-token-gc9r4" (UniqueName: "kubernetes.io/secret/c17abad8-a59d-46e3-9dd1-fbca191e2416-coredns-token-gc9r4") pod "coredns-6955765f44-2cfq5" (UID: "c17abad8-a59d-46e3-9dd1-fbca191e2416") * Feb 09 15:23:47 minikube kubelet[4807]: I0209 15:23:47.214509 4807 reconciler.go:209] operationExecutor.VerifyControllerAttachedVolume started for volume "config-volume" (UniqueName: "kubernetes.io/configmap/c17abad8-a59d-46e3-9dd1-fbca191e2416-config-volume") pod "coredns-6955765f44-2cfq5" (UID: "c17abad8-a59d-46e3-9dd1-fbca191e2416") * Feb 09 15:23:48 minikube kubelet[4807]: W0209 15:23:48.189840 4807 docker_sandbox.go:394] failed to read pod IP from plugin/docker: Couldn't find network status for kube-system/coredns-6955765f44-2cfq5 through plugin: invalid network status for * Feb 09 15:23:48 minikube kubelet[4807]: W0209 15:23:48.195778 4807 docker_sandbox.go:394] failed to read pod IP from plugin/docker: Couldn't find network status for kube-system/coredns-6955765f44-wprg2 through plugin: invalid network status for * Feb 09 15:23:48 minikube kubelet[4807]: W0209 15:23:48.447004 4807 docker_sandbox.go:394] failed to read pod IP from plugin/docker: Couldn't find network status for kube-system/coredns-6955765f44-2cfq5 through plugin: invalid network status for * Feb 09 15:23:48 minikube kubelet[4807]: W0209 15:23:48.450245 4807 docker_sandbox.go:394] failed to read pod IP from plugin/docker: Couldn't find network status for kube-system/coredns-6955765f44-wprg2 through plugin: invalid network status for * * ==> storage-provisioner ["57946b280cca"] <== </details> **The operating system version**: Windows 10 Pro (1909) with Hyper-V When using minikube version 1.5.2 or version 1.6.2, it is working fine: the file `C:\Users\Sebastien\.minikube\files\etc\hosts` is correctly copied in the VM as `/etc/hosts` after minikube has started. But it doesnt work anymore with minikube version 1.7.2. Two related questions: * How I can get logs when starting minikube? Adding the --v=7 shows nothing more. * I just would like to add an extra host in the VM /etc/hosts file: is-there a better/cleaner option to do that instead of copying a new file?
1.0
File copy to VM isn't working anymore on Windows - <!-- Please use this template while reporting an issue, providing as much information as possible. Failure to do so may result in a delayed response. Thank you! --> **The exact command to reproduce the issue**: minikube start **The full output of the command that failed**:<details> No error reported when minikube is starting. I didn't manage to display verbose/debug logs when adding --v=7. </details> **The output of the `minikube logs` command**:<details> * ==> Docker <== * -- Logs begin at Sun 2020-02-09 15:22:03 UTC, end at Sun 2020-02-09 16:22:02 UTC. -- * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.070778300Z" level=info msg="loading plugin "io.containerd.service.v1.namespaces-service"..." type=io.containerd.service.v1 * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.070785000Z" level=info msg="loading plugin "io.containerd.service.v1.snapshots-service"..." type=io.containerd.service.v1 * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.070791200Z" level=info msg="loading plugin "io.containerd.runtime.v1.linux"..." type=io.containerd.runtime.v1 * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.070890400Z" level=info msg="loading plugin "io.containerd.runtime.v2.task"..." type=io.containerd.runtime.v2 * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.070956400Z" level=info msg="loading plugin "io.containerd.monitor.v1.cgroups"..." type=io.containerd.monitor.v1 * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.071262000Z" level=info msg="loading plugin "io.containerd.service.v1.tasks-service"..." type=io.containerd.service.v1 * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.071297200Z" level=info msg="loading plugin "io.containerd.internal.v1.restart"..." type=io.containerd.internal.v1 * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.071341400Z" level=info msg="loading plugin "io.containerd.grpc.v1.containers"..." type=io.containerd.grpc.v1 * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.071366200Z" level=info msg="loading plugin "io.containerd.grpc.v1.content"..." type=io.containerd.grpc.v1 * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.071375700Z" level=info msg="loading plugin "io.containerd.grpc.v1.diff"..." type=io.containerd.grpc.v1 * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.071382500Z" level=info msg="loading plugin "io.containerd.grpc.v1.events"..." type=io.containerd.grpc.v1 * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.071388900Z" level=info msg="loading plugin "io.containerd.grpc.v1.healthcheck"..." type=io.containerd.grpc.v1 * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.071395700Z" level=info msg="loading plugin "io.containerd.grpc.v1.images"..." type=io.containerd.grpc.v1 * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.071402500Z" level=info msg="loading plugin "io.containerd.grpc.v1.leases"..." type=io.containerd.grpc.v1 * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.071409200Z" level=info msg="loading plugin "io.containerd.grpc.v1.namespaces"..." type=io.containerd.grpc.v1 * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.071415600Z" level=info msg="loading plugin "io.containerd.internal.v1.opt"..." type=io.containerd.internal.v1 * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.071447700Z" level=info msg="loading plugin "io.containerd.grpc.v1.snapshots"..." type=io.containerd.grpc.v1 * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.071456500Z" level=info msg="loading plugin "io.containerd.grpc.v1.tasks"..." type=io.containerd.grpc.v1 * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.071463500Z" level=info msg="loading plugin "io.containerd.grpc.v1.version"..." type=io.containerd.grpc.v1 * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.071471100Z" level=info msg="loading plugin "io.containerd.grpc.v1.introspection"..." type=io.containerd.grpc.v1 * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.071602300Z" level=info msg=serving... address="/var/run/docker/containerd/containerd-debug.sock" * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.071658800Z" level=info msg=serving... address="/var/run/docker/containerd/containerd.sock" * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.071691000Z" level=info msg="containerd successfully booted in 0.012932s" * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.075872100Z" level=info msg="parsed scheme: \"unix\"" module=grpc * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.075886200Z" level=info msg="scheme \"unix\" not registered, fallback to default scheme" module=grpc * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.075896500Z" level=info msg="ccResolverWrapper: sending update to cc: {[{unix:///var/run/docker/containerd/containerd.sock 0 <nil>}] <nil>}" module=grpc * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.075901900Z" level=info msg="ClientConn switching balancer to \"pick_first\"" module=grpc * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.076483100Z" level=info msg="parsed scheme: \"unix\"" module=grpc * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.076493700Z" level=info msg="scheme \"unix\" not registered, fallback to default scheme" module=grpc * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.076501500Z" level=info msg="ccResolverWrapper: sending update to cc: {[{unix:///var/run/docker/containerd/containerd.sock 0 <nil>}] <nil>}" module=grpc * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.076506600Z" level=info msg="ClientConn switching balancer to \"pick_first\"" module=grpc * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.102168000Z" level=warning msg="Your kernel does not support cgroup blkio weight" * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.102198300Z" level=warning msg="Your kernel does not support cgroup blkio weight_device" * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.102205500Z" level=warning msg="Your kernel does not support cgroup blkio throttle.read_bps_device" * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.102209600Z" level=warning msg="Your kernel does not support cgroup blkio throttle.write_bps_device" * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.102213300Z" level=warning msg="Your kernel does not support cgroup blkio throttle.read_iops_device" * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.102216800Z" level=warning msg="Your kernel does not support cgroup blkio throttle.write_iops_device" * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.102320900Z" level=info msg="Loading containers: start." * Feb 09 15:22:15 minikube dockerd[2446]: time="2020-02-09T15:22:15.201520800Z" level=info msg="Loading containers: done." * Feb 09 15:22:14 minikube dockerd[2446]: time="2020-02-09T15:22:14.748699386Z" level=info msg="Docker daemon" commit=633a0ea838 graphdriver(s)=overlay2 version=19.03.5 * Feb 09 15:22:14 minikube dockerd[2446]: time="2020-02-09T15:22:14.748784786Z" level=info msg="Daemon has completed initialization" * Feb 09 15:22:14 minikube dockerd[2446]: time="2020-02-09T15:22:14.778764786Z" level=info msg="API listen on /var/run/docker.sock" * Feb 09 15:22:14 minikube systemd[1]: Started Docker Application Container Engine. * Feb 09 15:22:14 minikube dockerd[2446]: time="2020-02-09T15:22:14.778848486Z" level=info msg="API listen on [::]:2376" * Feb 09 15:23:18 minikube dockerd[2446]: time="2020-02-09T15:23:18.100724410Z" level=info msg="shim containerd-shim started" address="/containerd-shim/moby/bbc3d97848af3255e127c03812852285de598b42070b4c9beb5efd0a2f3d7bca/shim.sock" debug=false pid=4055 * Feb 09 15:23:18 minikube dockerd[2446]: time="2020-02-09T15:23:18.106189577Z" level=info msg="shim containerd-shim started" address="/containerd-shim/moby/bc046d7669fe473893cd4ff2642ce75d3b3b30927a286850e2b64782e5f37b3c/shim.sock" debug=false pid=4069 * Feb 09 15:23:18 minikube dockerd[2446]: time="2020-02-09T15:23:18.119607549Z" level=info msg="shim containerd-shim started" address="/containerd-shim/moby/5fa4b650ae8ebda642fa076f1d32460ad805b6edc21f8604182d3a99cbdee4c0/shim.sock" debug=false pid=4089 * Feb 09 15:23:18 minikube dockerd[2446]: time="2020-02-09T15:23:18.124728724Z" level=info msg="shim containerd-shim started" address="/containerd-shim/moby/16756ddc58412ac11ebe458c09767ebd2dde3d057d5f26bb7b207f4ff4713a3c/shim.sock" debug=false pid=4099 * Feb 09 15:23:18 minikube dockerd[2446]: time="2020-02-09T15:23:18.303385861Z" level=info msg="shim containerd-shim started" address="/containerd-shim/moby/22d8af4f8b0024e98da348de02d9989f95315d69f87e84394bb39e2386b905fb/shim.sock" debug=false pid=4334 * Feb 09 15:23:18 minikube dockerd[2446]: time="2020-02-09T15:23:18.329697819Z" level=info msg="shim containerd-shim started" address="/containerd-shim/moby/022bcbf4051107ac7cb0f3221b9863e6741b3c8a73fa66e3e212619f52c109db/shim.sock" debug=false pid=4367 * Feb 09 15:23:18 minikube dockerd[2446]: time="2020-02-09T15:23:18.332848142Z" level=info msg="shim containerd-shim started" address="/containerd-shim/moby/9c2e529b78d8735ce23e8d79379dea483a84f0fff8aabd371e1aecfa16c11504/shim.sock" debug=false pid=4376 * Feb 09 15:23:18 minikube dockerd[2446]: time="2020-02-09T15:23:18.340200062Z" level=info msg="shim containerd-shim started" address="/containerd-shim/moby/a1faeec091127bad2a535e24688a22b9ff1c308aea19ec85a007247b26220188/shim.sock" debug=false pid=4396 * Feb 09 15:23:44 minikube dockerd[2446]: time="2020-02-09T15:23:44.934584667Z" level=info msg="shim containerd-shim started" address="/containerd-shim/moby/c20ecdf378a05322fcbd85b850a842790c1b0a358281edaa2f64cfb70c44b03a/shim.sock" debug=false pid=5068 * Feb 09 15:23:45 minikube dockerd[2446]: time="2020-02-09T15:23:45.105277818Z" level=info msg="shim containerd-shim started" address="/containerd-shim/moby/02c3afd9bb285a2cb0af8c5211acd573f3130e345c597f10ec7242679d05fd6a/shim.sock" debug=false pid=5114 * Feb 09 15:23:46 minikube dockerd[2446]: time="2020-02-09T15:23:46.123084899Z" level=info msg="shim containerd-shim started" address="/containerd-shim/moby/da500db7aad4bfdc9a2f37c74df75383e2effa5f321ea97760fa1dfd73293715/shim.sock" debug=false pid=5242 * Feb 09 15:23:46 minikube dockerd[2446]: time="2020-02-09T15:23:46.281781063Z" level=info msg="shim containerd-shim started" address="/containerd-shim/moby/57946b280cca1e3a0ba89ff6d15e5fb0cacd00d846279341935ef9d18710266a/shim.sock" debug=false pid=5282 * Feb 09 15:23:48 minikube dockerd[2446]: time="2020-02-09T15:23:48.015340971Z" level=info msg="shim containerd-shim started" address="/containerd-shim/moby/e6614a19f9a159f6dca03fdf4690cb9bde4c11dc74c959aec58f7df37ddde3ac/shim.sock" debug=false pid=5355 * Feb 09 15:23:48 minikube dockerd[2446]: time="2020-02-09T15:23:48.019409757Z" level=info msg="shim containerd-shim started" address="/containerd-shim/moby/5f238a2dd6f24245006c1b87d3869ee3c972e82c573065907f8aac69fd3b15eb/shim.sock" debug=false pid=5364 * Feb 09 15:23:48 minikube dockerd[2446]: time="2020-02-09T15:23:48.279627640Z" level=info msg="shim containerd-shim started" address="/containerd-shim/moby/b694277a7d053b96e0cf21dfdc8e73c3c47902cf0082248271787803e093d97c/shim.sock" debug=false pid=5483 * Feb 09 15:23:48 minikube dockerd[2446]: time="2020-02-09T15:23:48.289742004Z" level=info msg="shim containerd-shim started" address="/containerd-shim/moby/3f923512a9f8332a90b628a1da8f6d9b30ee6cd082d4625c2ef9b9aef69de11d/shim.sock" debug=false pid=5502 * * ==> container status <== * CONTAINER IMAGE CREATED STATE NAME ATTEMPT POD ID * 3f923512a9f83 70f311871ae12 2 minutes ago Running coredns 0 e6614a19f9a15 * b694277a7d053 70f311871ae12 2 minutes ago Running coredns 0 5f238a2dd6f24 * 57946b280cca1 4689081edb103 2 minutes ago Running storage-provisioner 0 da500db7aad4b * 02c3afd9bb285 cba2a99699bdf 2 minutes ago Running kube-proxy 0 c20ecdf378a05 * a1faeec091127 303ce5db0e90d 3 minutes ago Running etcd 0 16756ddc58412 * 9c2e529b78d87 da5fd66c4068c 3 minutes ago Running kube-controller-manager 0 5fa4b650ae8eb * 022bcbf405110 41ef50a5f06a7 3 minutes ago Running kube-apiserver 0 bc046d7669fe4 * 22d8af4f8b002 f52d4c527ef2f 3 minutes ago Running kube-scheduler 0 bbc3d97848af3 * * ==> coredns ["3f923512a9f8"] <== * .:53 * [INFO] plugin/reload: Running configuration MD5 = 4e235fcc3696966e76816bcd9034ebc7 * CoreDNS-1.6.5 * linux/amd64, go1.13.4, c2fd1b2 * * ==> coredns ["b694277a7d05"] <== * .:53 * [INFO] plugin/reload: Running configuration MD5 = 4e235fcc3696966e76816bcd9034ebc7 * CoreDNS-1.6.5 * linux/amd64, go1.13.4, c2fd1b2 * * ==> dmesg <== * [Feb 9 15:21] You have booted with nomodeset. This means your GPU drivers are DISABLED * [ +0.000000] Any video related functionality will be severely degraded, and you may not even be able to suspend the system properly * [ +0.000000] Unless you actually understand what nomodeset does, you should reboot without enabling it * [ +0.046941] MDS CPU bug present and SMT on, data leak possible. See https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/mds.html for more details. * [ +0.000001] TAA CPU bug present and SMT on, data leak possible. See https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/tsx_async_abort.html for more details. * [ +0.000044] #2 #3 * [ +0.022317] acpi PNP0A03:00: fail to add MMCONFIG information, can't access extended PCI configuration space under this bridge. * [ +0.006675] * Found PM-Timer Bug on the chipset. Due to workarounds for a bug, * * this clock source is slow. Consider trying other clock sources * [Feb 9 15:22] Unstable clock detected, switching default tracing clock to "global" * If you want to keep using the local clock, then add: * "trace_clock=local" * on the kernel command line * [ +0.000039] platform regulatory.0: Direct firmware load for regulatory.db failed with error -2 * [ +0.408454] psmouse serio1: trackpoint: failed to get extended button data, assuming 3 buttons * [ +0.680971] systemd[1]: Failed to bump fs.file-max, ignoring: Invalid argument * [ +0.002245] systemd-fstab-generator[1249]: Ignoring "noauto" for root device * [ +0.002567] systemd[1]: File /usr/lib/systemd/system/systemd-journald.service:12 configures an IP firewall (IPAddressDeny=any), but the local system does not support BPF/cgroup based firewalling. * [ +0.000002] systemd[1]: Proceeding WITHOUT firewalling in effect! (This warning is only shown for the first loaded unit using IP firewalling.) * [ +1.472188] NFSD: the nfsdcld client tracking upcall will be removed in 3.10. Please transition to using nfsdcltrack. * [ +0.174629] vboxguest: loading out-of-tree module taints kernel. * [ +0.002674] vboxguest: PCI device not found, probably running on physical hardware. * [ +8.053422] systemd-fstab-generator[2390]: Ignoring "noauto" for root device * [ +0.981549] systemd-fstab-generator[2418]: Ignoring "noauto" for root device * [ +40.516646] systemd-fstab-generator[3337]: Ignoring "noauto" for root device * [ +0.877860] systemd-fstab-generator[3561]: Ignoring "noauto" for root device * [Feb 9 15:23] kauditd_printk_skb: 65 callbacks suppressed * [ +6.712206] systemd-fstab-generator[4798]: Ignoring "noauto" for root device * [ +21.325389] kauditd_printk_skb: 32 callbacks suppressed * [ +7.219938] kauditd_printk_skb: 44 callbacks suppressed * [Feb 9 15:24] NFSD: Unable to end grace period: -110 * * ==> kernel <== * 15:26:32 up 4 min, 0 users, load average: 0.09, 0.17, 0.08 * Linux minikube 4.19.88 #1 SMP Tue Feb 4 22:25:03 PST 2020 x86_64 GNU/Linux * PRETTY_NAME="Buildroot 2019.02.8" * * ==> kube-apiserver ["022bcbf40511"] <== * W0209 15:23:19.922216 1 genericapiserver.go:404] Skipping API discovery.k8s.io/v1alpha1 because it has no resources. * W0209 15:23:19.928317 1 genericapiserver.go:404] Skipping API node.k8s.io/v1alpha1 because it has no resources. * W0209 15:23:19.939588 1 genericapiserver.go:404] Skipping API rbac.authorization.k8s.io/v1alpha1 because it has no resources. * W0209 15:23:19.941700 1 genericapiserver.go:404] Skipping API scheduling.k8s.io/v1alpha1 because it has no resources. * W0209 15:23:19.949731 1 genericapiserver.go:404] Skipping API storage.k8s.io/v1alpha1 because it has no resources. * W0209 15:23:19.967312 1 genericapiserver.go:404] Skipping API apps/v1beta2 because it has no resources. * W0209 15:23:19.967330 1 genericapiserver.go:404] Skipping API apps/v1beta1 because it has no resources. * I0209 15:23:19.973790 1 plugins.go:158] Loaded 11 mutating admission controller(s) successfully in the following order: NamespaceLifecycle,LimitRanger,ServiceAccount,NodeRestriction,TaintNodesByCondition,Priority,DefaultTolerationSeconds,DefaultStorageClass,StorageObjectInUseProtection,MutatingAdmissionWebhook,RuntimeClass. * I0209 15:23:19.973815 1 plugins.go:161] Loaded 7 validating admission controller(s) successfully in the following order: LimitRanger,ServiceAccount,Priority,PersistentVolumeClaimResize,ValidatingAdmissionWebhook,RuntimeClass,ResourceQuota. * I0209 15:23:19.974828 1 client.go:361] parsed scheme: "endpoint" * I0209 15:23:19.974855 1 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://127.0.0.1:2379 0 <nil>}] * I0209 15:23:19.979591 1 client.go:361] parsed scheme: "endpoint" * I0209 15:23:19.979620 1 endpoint.go:68] ccResolverWrapper: sending new addresses to cc: [{https://127.0.0.1:2379 0 <nil>}] * I0209 15:23:21.083733 1 dynamic_cafile_content.go:166] Starting request-header::/var/lib/minikube/certs/front-proxy-ca.crt * I0209 15:23:21.083770 1 dynamic_cafile_content.go:166] Starting client-ca-bundle::/var/lib/minikube/certs/ca.crt * I0209 15:23:21.083785 1 dynamic_serving_content.go:129] Starting serving-cert::/var/lib/minikube/certs/apiserver.crt::/var/lib/minikube/certs/apiserver.key * I0209 15:23:21.084157 1 secure_serving.go:178] Serving securely on [::]:8443 * I0209 15:23:21.084190 1 controller.go:81] Starting OpenAPI AggregationController * I0209 15:23:21.084219 1 tlsconfig.go:219] Starting DynamicServingCertificateController * I0209 15:23:21.084531 1 crd_finalizer.go:263] Starting CRDFinalizer * I0209 15:23:21.084627 1 cluster_authentication_trust_controller.go:440] Starting cluster_authentication_trust_controller controller * I0209 15:23:21.084669 1 apiapproval_controller.go:185] Starting KubernetesAPIApprovalPolicyConformantConditionController * I0209 15:23:21.084671 1 shared_informer.go:197] Waiting for caches to sync for cluster_authentication_trust_controller * I0209 15:23:21.084652 1 customresource_discovery_controller.go:208] Starting DiscoveryController * I0209 15:23:21.084656 1 naming_controller.go:288] Starting NamingConditionController * I0209 15:23:21.084660 1 establishing_controller.go:73] Starting EstablishingController * I0209 15:23:21.084665 1 nonstructuralschema_controller.go:191] Starting NonStructuralSchemaConditionController * I0209 15:23:21.084981 1 apiservice_controller.go:94] Starting APIServiceRegistrationController * I0209 15:23:21.085022 1 cache.go:32] Waiting for caches to sync for APIServiceRegistrationController controller * I0209 15:23:21.085089 1 available_controller.go:386] Starting AvailableConditionController * I0209 15:23:21.085121 1 cache.go:32] Waiting for caches to sync for AvailableConditionController controller * I0209 15:23:21.085171 1 dynamic_cafile_content.go:166] Starting client-ca-bundle::/var/lib/minikube/certs/ca.crt * I0209 15:23:21.085211 1 dynamic_cafile_content.go:166] Starting request-header::/var/lib/minikube/certs/front-proxy-ca.crt * I0209 15:23:21.084638 1 controller.go:85] Starting OpenAPI controller * I0209 15:23:21.086075 1 autoregister_controller.go:140] Starting autoregister controller * I0209 15:23:21.086084 1 cache.go:32] Waiting for caches to sync for autoregister controller * E0209 15:23:21.091691 1 controller.go:151] Unable to remove old endpoints from kubernetes service: StorageError: key not found, Code: 1, Key: /registry/masterleases/172.18.109.141, ResourceVersion: 0, AdditionalErrorMsg: * I0209 15:23:21.109117 1 crdregistration_controller.go:111] Starting crd-autoregister controller * I0209 15:23:21.109126 1 shared_informer.go:197] Waiting for caches to sync for crd-autoregister * I0209 15:23:21.185737 1 shared_informer.go:204] Caches are synced for cluster_authentication_trust_controller * I0209 15:23:21.186181 1 cache.go:39] Caches are synced for AvailableConditionController controller * I0209 15:23:21.186236 1 cache.go:39] Caches are synced for APIServiceRegistrationController controller * I0209 15:23:21.193358 1 cache.go:39] Caches are synced for autoregister controller * I0209 15:23:21.210070 1 shared_informer.go:204] Caches are synced for crd-autoregister * I0209 15:23:22.083884 1 controller.go:107] OpenAPI AggregationController: Processing item * I0209 15:23:22.083908 1 controller.go:130] OpenAPI AggregationController: action for item : Nothing (removed from the queue). * I0209 15:23:22.083968 1 controller.go:130] OpenAPI AggregationController: action for item k8s_internal_local_delegation_chain_0000000000: Nothing (removed from the queue). * I0209 15:23:22.086972 1 storage_scheduling.go:133] created PriorityClass system-node-critical with value 2000001000 * I0209 15:23:22.090176 1 storage_scheduling.go:133] created PriorityClass system-cluster-critical with value 2000000000 * I0209 15:23:22.090198 1 storage_scheduling.go:142] all system priority classes are created successfully or already exist. * I0209 15:23:22.349721 1 controller.go:606] quota admission added evaluator for: roles.rbac.authorization.k8s.io * I0209 15:23:22.377055 1 controller.go:606] quota admission added evaluator for: rolebindings.rbac.authorization.k8s.io * W0209 15:23:22.473430 1 lease.go:224] Resetting endpoints for master service "kubernetes" to [172.18.109.141] * I0209 15:23:22.473762 1 controller.go:606] quota admission added evaluator for: endpoints * I0209 15:23:23.225992 1 controller.go:606] quota admission added evaluator for: leases.coordination.k8s.io * I0209 15:23:23.958883 1 controller.go:606] quota admission added evaluator for: serviceaccounts * I0209 15:23:23.967803 1 controller.go:606] quota admission added evaluator for: deployments.apps * I0209 15:23:24.235451 1 controller.go:606] quota admission added evaluator for: daemonsets.apps * I0209 15:23:32.309375 1 controller.go:606] quota admission added evaluator for: replicasets.apps * I0209 15:23:32.348526 1 controller.go:606] quota admission added evaluator for: controllerrevisions.apps * * ==> kube-controller-manager ["9c2e529b78d8"] <== * I0209 15:23:30.499920 1 resource_quota_monitor.go:303] QuotaMonitor running * I0209 15:23:30.507894 1 controllermanager.go:533] Started "deployment" * W0209 15:23:30.507919 1 controllermanager.go:525] Skipping "root-ca-cert-publisher" * I0209 15:23:30.507949 1 deployment_controller.go:152] Starting deployment controller * I0209 15:23:30.507953 1 shared_informer.go:197] Waiting for caches to sync for deployment * I0209 15:23:31.199195 1 controllermanager.go:533] Started "horizontalpodautoscaling" * I0209 15:23:31.199279 1 horizontal.go:156] Starting HPA controller * I0209 15:23:31.199288 1 shared_informer.go:197] Waiting for caches to sync for HPA * I0209 15:23:31.449726 1 controllermanager.go:533] Started "ttl" * I0209 15:23:31.449808 1 ttl_controller.go:116] Starting TTL controller * I0209 15:23:31.449818 1 shared_informer.go:197] Waiting for caches to sync for TTL * I0209 15:23:31.698448 1 controllermanager.go:533] Started "bootstrapsigner" * I0209 15:23:31.698489 1 shared_informer.go:197] Waiting for caches to sync for bootstrap_signer * I0209 15:23:31.952827 1 controllermanager.go:533] Started "serviceaccount" * W0209 15:23:31.952953 1 controllermanager.go:525] Skipping "nodeipam" * I0209 15:23:31.953498 1 serviceaccounts_controller.go:116] Starting service account controller * I0209 15:23:31.953561 1 shared_informer.go:197] Waiting for caches to sync for service account * I0209 15:23:31.953685 1 shared_informer.go:197] Waiting for caches to sync for garbage collector * W0209 15:23:31.960056 1 actual_state_of_world.go:506] Failed to update statusUpdateNeeded field in actual state of world: Failed to set statusUpdateNeeded to needed true, because nodeName="minikube" does not exist * I0209 15:23:31.998629 1 shared_informer.go:204] Caches are synced for certificate-csrapproving * I0209 15:23:31.998633 1 shared_informer.go:204] Caches are synced for bootstrap_signer * I0209 15:23:31.999164 1 shared_informer.go:204] Caches are synced for expand * I0209 15:23:32.030788 1 shared_informer.go:204] Caches are synced for certificate-csrsigning * I0209 15:23:32.048893 1 shared_informer.go:204] Caches are synced for PV protection * I0209 15:23:32.050507 1 shared_informer.go:204] Caches are synced for TTL * I0209 15:23:32.053741 1 shared_informer.go:204] Caches are synced for ClusterRoleAggregator * I0209 15:23:32.252423 1 shared_informer.go:197] Waiting for caches to sync for resource quota * I0209 15:23:32.267774 1 shared_informer.go:204] Caches are synced for PVC protection * I0209 15:23:32.298870 1 shared_informer.go:204] Caches are synced for taint * I0209 15:23:32.298870 1 shared_informer.go:204] Caches are synced for ReplicaSet * I0209 15:23:32.298939 1 taint_manager.go:186] Starting NoExecuteTaintManager * I0209 15:23:32.298958 1 node_lifecycle_controller.go:1443] Initializing eviction metric for zone: * W0209 15:23:32.299011 1 node_lifecycle_controller.go:1058] Missing timestamp for Node minikube. Assuming now as a timestamp. * I0209 15:23:32.299015 1 shared_informer.go:204] Caches are synced for attach detach * I0209 15:23:32.299075 1 node_lifecycle_controller.go:1209] Controller detected that all Nodes are not-Ready. Entering master disruption mode. * I0209 15:23:32.299177 1 event.go:281] Event(v1.ObjectReference{Kind:"Node", Namespace:"", Name:"minikube", UID:"0bd76262-05b0-4350-be09-cb6fe95d5917", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'RegisteredNode' Node minikube event: Registered Node minikube in Controller * I0209 15:23:32.299255 1 shared_informer.go:204] Caches are synced for endpoint * I0209 15:23:32.301874 1 shared_informer.go:204] Caches are synced for disruption * I0209 15:23:32.301888 1 disruption.go:338] Sending events to api server. * I0209 15:23:32.303116 1 shared_informer.go:204] Caches are synced for ReplicationController * I0209 15:23:32.308185 1 shared_informer.go:204] Caches are synced for deployment * I0209 15:23:32.308215 1 shared_informer.go:204] Caches are synced for job * I0209 15:23:32.311175 1 event.go:281] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"kube-system", Name:"coredns", UID:"329e20c5-e656-4310-b781-1aba8093fb09", APIVersion:"apps/v1", ResourceVersion:"180", FieldPath:""}): type: 'Normal' reason: 'ScalingReplicaSet' Scaled up replica set coredns-6955765f44 to 2 * I0209 15:23:32.321513 1 event.go:281] Event(v1.ObjectReference{Kind:"ReplicaSet", Namespace:"kube-system", Name:"coredns-6955765f44", UID:"1c98f2f1-7906-4c98-8fd0-51f1ef943e02", APIVersion:"apps/v1", ResourceVersion:"312", FieldPath:""}): type: 'Normal' reason: 'SuccessfulCreate' Created pod: coredns-6955765f44-wprg2 * I0209 15:23:32.329100 1 event.go:281] Event(v1.ObjectReference{Kind:"ReplicaSet", Namespace:"kube-system", Name:"coredns-6955765f44", UID:"1c98f2f1-7906-4c98-8fd0-51f1ef943e02", APIVersion:"apps/v1", ResourceVersion:"312", FieldPath:""}): type: 'Normal' reason: 'SuccessfulCreate' Created pod: coredns-6955765f44-2cfq5 * I0209 15:23:32.346845 1 shared_informer.go:204] Caches are synced for daemon sets * I0209 15:23:32.348899 1 shared_informer.go:204] Caches are synced for stateful set * I0209 15:23:32.349483 1 shared_informer.go:204] Caches are synced for GC * I0209 15:23:32.351331 1 event.go:281] Event(v1.ObjectReference{Kind:"DaemonSet", Namespace:"kube-system", Name:"kube-proxy", UID:"7610c387-048f-471f-bb42-e89e44c143c0", APIVersion:"apps/v1", ResourceVersion:"185", FieldPath:""}): type: 'Normal' reason: 'SuccessfulCreate' Created pod: kube-proxy-tkmbk * I0209 15:23:32.357681 1 shared_informer.go:204] Caches are synced for persistent volume * E0209 15:23:32.364751 1 daemon_controller.go:290] kube-system/kube-proxy failed with : error storing status for daemon set &v1.DaemonSet{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"kube-proxy", GenerateName:"", Namespace:"kube-system", SelfLink:"/apis/apps/v1/namespaces/kube-system/daemonsets/kube-proxy", UID:"7610c387-048f-471f-bb42-e89e44c143c0", ResourceVersion:"185", Generation:1, CreationTimestamp:v1.Time{Time:time.Time{wall:0x0, ext:63716858604, loc:(*time.Location)(0x6b971e0)}}, DeletionTimestamp:(*v1.Time)(nil), DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-proxy"}, Annotations:map[string]string{"deprecated.daemonset.template.generation":"1"}, OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ClusterName:"", ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v1.DaemonSetSpec{Selector:(*v1.LabelSelector)(0xc00046c880), Template:v1.PodTemplateSpec{ObjectMeta:v1.ObjectMeta{Name:"", GenerateName:"", Namespace:"", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:v1.Time{Time:time.Time{wall:0x0, ext:0, loc:(*time.Location)(nil)}}, DeletionTimestamp:(*v1.Time)(nil), DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-proxy"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ClusterName:"", ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v1.PodSpec{Volumes:[]v1.Volume{v1.Volume{Name:"kube-proxy", VolumeSource:v1.VolumeSource{HostPath:(*v1.HostPathVolumeSource)(nil), EmptyDir:(*v1.EmptyDirVolumeSource)(nil), GCEPersistentDisk:(*v1.GCEPersistentDiskVolumeSource)(nil), AWSElasticBlockStore:(*v1.AWSElasticBlockStoreVolumeSource)(nil), GitRepo:(*v1.GitRepoVolumeSource)(nil), Secret:(*v1.SecretVolumeSource)(nil), NFS:(*v1.NFSVolumeSource)(nil), ISCSI:(*v1.ISCSIVolumeSource)(nil), Glusterfs:(*v1.GlusterfsVolumeSource)(nil), PersistentVolumeClaim:(*v1.PersistentVolumeClaimVolumeSource)(nil), RBD:(*v1.RBDVolumeSource)(nil), FlexVolume:(*v1.FlexVolumeSource)(nil), Cinder:(*v1.CinderVolumeSource)(nil), CephFS:(*v1.CephFSVolumeSource)(nil), Flocker:(*v1.FlockerVolumeSource)(nil), DownwardAPI:(*v1.DownwardAPIVolumeSource)(nil), FC:(*v1.FCVolumeSource)(nil), AzureFile:(*v1.AzureFileVolumeSource)(nil), ConfigMap:(*v1.ConfigMapVolumeSource)(0xc00094f3c0), VsphereVolume:(*v1.VsphereVirtualDiskVolumeSource)(nil), Quobyte:(*v1.QuobyteVolumeSource)(nil), AzureDisk:(*v1.AzureDiskVolumeSource)(nil), PhotonPersistentDisk:(*v1.PhotonPersistentDiskVolumeSource)(nil), Projected:(*v1.ProjectedVolumeSource)(nil), PortworxVolume:(*v1.PortworxVolumeSource)(nil), ScaleIO:(*v1.ScaleIOVolumeSource)(nil), StorageOS:(*v1.StorageOSVolumeSource)(nil), CSI:(*v1.CSIVolumeSource)(nil)}}, v1.Volume{Name:"xtables-lock", VolumeSource:v1.VolumeSource{HostPath:(*v1.HostPathVolumeSource)(0xc00046ca00), EmptyDir:(*v1.EmptyDirVolumeSource)(nil), GCEPersistentDisk:(*v1.GCEPersistentDiskVolumeSource)(nil), AWSElasticBlockStore:(*v1.AWSElasticBlockStoreVolumeSource)(nil), GitRepo:(*v1.GitRepoVolumeSource)(nil), Secret:(*v1.SecretVolumeSource)(nil), NFS:(*v1.NFSVolumeSource)(nil), ISCSI:(*v1.ISCSIVolumeSource)(nil), Glusterfs:(*v1.GlusterfsVolumeSource)(nil), PersistentVolumeClaim:(*v1.PersistentVolumeClaimVolumeSource)(nil), RBD:(*v1.RBDVolumeSource)(nil), FlexVolume:(*v1.FlexVolumeSource)(nil), Cinder:(*v1.CinderVolumeSource)(nil), CephFS:(*v1.CephFSVolumeSource)(nil), Flocker:(*v1.FlockerVolumeSource)(nil), DownwardAPI:(*v1.DownwardAPIVolumeSource)(nil), FC:(*v1.FCVolumeSource)(nil), AzureFile:(*v1.AzureFileVolumeSource)(nil), ConfigMap:(*v1.ConfigMapVolumeSource)(nil), VsphereVolume:(*v1.VsphereVirtualDiskVolumeSource)(nil), Quobyte:(*v1.QuobyteVolumeSource)(nil), AzureDisk:(*v1.AzureDiskVolumeSource)(nil), PhotonPersistentDisk:(*v1.PhotonPersistentDiskVolumeSource)(nil), Projected:(*v1.ProjectedVolumeSource)(nil), PortworxVolume:(*v1.PortworxVolumeSource)(nil), ScaleIO:(*v1.ScaleIOVolumeSource)(nil), StorageOS:(*v1.StorageOSVolumeSource)(nil), CSI:(*v1.CSIVolumeSource)(nil)}}, v1.Volume{Name:"lib-modules", VolumeSource:v1.VolumeSource{HostPath:(*v1.HostPathVolumeSource)(0xc00046cbe0), EmptyDir:(*v1.EmptyDirVolumeSource)(nil), GCEPersistentDisk:(*v1.GCEPersistentDiskVolumeSource)(nil), AWSElasticBlockStore:(*v1.AWSElasticBlockStoreVolumeSource)(nil), GitRepo:(*v1.GitRepoVolumeSource)(nil), Secret:(*v1.SecretVolumeSource)(nil), NFS:(*v1.NFSVolumeSource)(nil), ISCSI:(*v1.ISCSIVolumeSource)(nil), Glusterfs:(*v1.GlusterfsVolumeSource)(nil), PersistentVolumeClaim:(*v1.PersistentVolumeClaimVolumeSource)(nil), RBD:(*v1.RBDVolumeSource)(nil), FlexVolume:(*v1.FlexVolumeSource)(nil), Cinder:(*v1.CinderVolumeSource)(nil), CephFS:(*v1.CephFSVolumeSource)(nil), Flocker:(*v1.FlockerVolumeSource)(nil), DownwardAPI:(*v1.DownwardAPIVolumeSource)(nil), FC:(*v1.FCVolumeSource)(nil), AzureFile:(*v1.AzureFileVolumeSource)(nil), ConfigMap:(*v1.ConfigMapVolumeSource)(nil), VsphereVolume:(*v1.VsphereVirtualDiskVolumeSource)(nil), Quobyte:(*v1.QuobyteVolumeSource)(nil), AzureDisk:(*v1.AzureDiskVolumeSource)(nil), PhotonPersistentDisk:(*v1.PhotonPersistentDiskVolumeSource)(nil), Projected:(*v1.ProjectedVolumeSource)(nil), PortworxVolume:(*v1.PortworxVolumeSource)(nil), ScaleIO:(*v1.ScaleIOVolumeSource)(nil), StorageOS:(*v1.StorageOSVolumeSource)(nil), CSI:(*v1.CSIVolumeSource)(nil)}}}, InitContainers:[]v1.Container(nil), Containers:[]v1.Container{v1.Container{Name:"kube-proxy", Image:"k8s.gcr.io/kube-proxy:v1.17.2", Command:[]string{"/usr/local/bin/kube-proxy", "--config=/var/lib/kube-proxy/config.conf", "--hostname-override=$(NODE_NAME)"}, Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar{v1.EnvVar{Name:"NODE_NAME", Value:"", ValueFrom:(*v1.EnvVarSource)(0xc00046d0c0)}}, Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList(nil)}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"kube-proxy", ReadOnly:false, MountPath:"/var/lib/kube-proxy", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}, v1.VolumeMount{Name:"xtables-lock", ReadOnly:false, MountPath:"/run/xtables.lock", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}, v1.VolumeMount{Name:"lib-modules", ReadOnly:true, MountPath:"/lib/modules", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), StartupProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(0xc000ba0730), Stdin:false, StdinOnce:false, TTY:false}}, EphemeralContainers:[]v1.EphemeralContainer(nil), RestartPolicy:"Always", TerminationGracePeriodSeconds:(*int64)(0xc000528dd8), ActiveDeadlineSeconds:(*int64)(nil), DNSPolicy:"ClusterFirst", NodeSelector:map[string]string{"beta.kubernetes.io/os":"linux"}, ServiceAccountName:"kube-proxy", DeprecatedServiceAccount:"kube-proxy", AutomountServiceAccountToken:(*bool)(nil), NodeName:"", HostNetwork:true, HostPID:false, HostIPC:false, ShareProcessNamespace:(*bool)(nil), SecurityContext:(*v1.PodSecurityContext)(0xc000efe960), ImagePullSecrets:[]v1.LocalObjectReference(nil), Hostname:"", Subdomain:"", Affinity:(*v1.Affinity)(nil), SchedulerName:"default-scheduler", Tolerations:[]v1.Toleration{v1.Toleration{Key:"CriticalAddonsOnly", Operator:"Exists", Value:"", Effect:"", TolerationSeconds:(*int64)(nil)}, v1.Toleration{Key:"", Operator:"Exists", Value:"", Effect:"", TolerationSeconds:(*int64)(nil)}}, HostAliases:[]v1.HostAlias(nil), PriorityClassName:"system-node-critical", Priority:(*int32)(nil), DNSConfig:(*v1.PodDNSConfig)(nil), ReadinessGates:[]v1.PodReadinessGate(nil), RuntimeClassName:(*string)(nil), EnableServiceLinks:(*bool)(nil), PreemptionPolicy:(*v1.PreemptionPolicy)(nil), Overhead:v1.ResourceList(nil), TopologySpreadConstraints:[]v1.TopologySpreadConstraint(nil)}}, UpdateStrategy:v1.DaemonSetUpdateStrategy{Type:"RollingUpdate", RollingUpdate:(*v1.RollingUpdateDaemonSet)(0xc00000e6c0)}, MinReadySeconds:0, RevisionHistoryLimit:(*int32)(0xc000528e58)}, Status:v1.DaemonSetStatus{CurrentNumberScheduled:0, NumberMisscheduled:0, DesiredNumberScheduled:0, NumberReady:0, ObservedGeneration:0, UpdatedNumberScheduled:0, NumberAvailable:0, NumberUnavailable:0, CollisionCount:(*int32)(nil), Conditions:[]v1.DaemonSetCondition(nil)}}: Operation cannot be fulfilled on daemonsets.apps "kube-proxy": the object has been modified; please apply your changes to the latest version and try again * I0209 15:23:32.399450 1 shared_informer.go:204] Caches are synced for HPA * I0209 15:23:32.552906 1 shared_informer.go:204] Caches are synced for resource quota * I0209 15:23:32.552931 1 shared_informer.go:204] Caches are synced for namespace * I0209 15:23:32.553917 1 shared_informer.go:204] Caches are synced for service account * I0209 15:23:32.553990 1 shared_informer.go:204] Caches are synced for garbage collector * I0209 15:23:32.597549 1 shared_informer.go:204] Caches are synced for garbage collector * I0209 15:23:32.597595 1 garbagecollector.go:138] Garbage collector: all resource monitors have synced. Proceeding to collect garbage * I0209 15:23:32.600066 1 shared_informer.go:204] Caches are synced for resource quota * I0209 15:23:47.300428 1 node_lifecycle_controller.go:1236] Controller detected that some Nodes are Ready. Exiting master disruption mode. * * ==> kube-proxy ["02c3afd9bb28"] <== * W0209 15:23:45.210525 1 server_others.go:323] Unknown proxy mode "", assuming iptables proxy * I0209 15:23:45.214518 1 node.go:135] Successfully retrieved node IP: 172.18.109.141 * I0209 15:23:45.214545 1 server_others.go:145] Using iptables Proxier. * W0209 15:23:45.214634 1 proxier.go:286] clusterCIDR not specified, unable to distinguish between internal and external traffic * I0209 15:23:45.214787 1 server.go:571] Version: v1.17.2 * I0209 15:23:45.214994 1 conntrack.go:100] Set sysctl 'net/netfilter/nf_conntrack_max' to 131072 * I0209 15:23:45.215013 1 conntrack.go:52] Setting nf_conntrack_max to 131072 * I0209 15:23:45.215440 1 conntrack.go:83] Setting conntrack hashsize to 32768 * I0209 15:23:45.221130 1 conntrack.go:100] Set sysctl 'net/netfilter/nf_conntrack_tcp_timeout_established' to 86400 * I0209 15:23:45.221190 1 conntrack.go:100] Set sysctl 'net/netfilter/nf_conntrack_tcp_timeout_close_wait' to 3600 * I0209 15:23:45.221287 1 config.go:313] Starting service config controller * I0209 15:23:45.221293 1 shared_informer.go:197] Waiting for caches to sync for service config * I0209 15:23:45.221314 1 config.go:131] Starting endpoints config controller * I0209 15:23:45.221335 1 shared_informer.go:197] Waiting for caches to sync for endpoints config * I0209 15:23:45.321702 1 shared_informer.go:204] Caches are synced for endpoints config * I0209 15:23:45.321705 1 shared_informer.go:204] Caches are synced for service config * * ==> kube-scheduler ["22d8af4f8b00"] <== * I0209 15:23:18.670592 1 serving.go:312] Generated self-signed cert in-memory * W0209 15:23:18.772671 1 configmap_cafile_content.go:102] unable to load initial CA bundle for: "client-ca::kube-system::extension-apiserver-authentication::client-ca-file" due to: configmap "extension-apiserver-authentication" not found * W0209 15:23:18.772931 1 configmap_cafile_content.go:102] unable to load initial CA bundle for: "client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" due to: configmap "extension-apiserver-authentication" not found * W0209 15:23:21.112588 1 authentication.go:348] Unable to get configmap/extension-apiserver-authentication in kube-system. Usually fixed by 'kubectl create rolebinding -n kube-system ROLEBINDING_NAME --role=extension-apiserver-authentication-reader --serviceaccount=YOUR_NS:YOUR_SA' * W0209 15:23:21.112729 1 authentication.go:296] Error looking up in-cluster authentication configuration: configmaps "extension-apiserver-authentication" is forbidden: User "system:kube-scheduler" cannot get resource "configmaps" in API group "" in the namespace "kube-system" * W0209 15:23:21.112783 1 authentication.go:297] Continuing without authentication configuration. This may treat all requests as anonymous. * W0209 15:23:21.112845 1 authentication.go:298] To require authentication configuration lookup to succeed, set --authentication-tolerate-lookup-failure=false * W0209 15:23:21.120507 1 authorization.go:47] Authorization is disabled * W0209 15:23:21.120550 1 authentication.go:92] Authentication is disabled * I0209 15:23:21.120569 1 deprecated_insecure_serving.go:51] Serving healthz insecurely on [::]:10251 * I0209 15:23:21.121504 1 configmap_cafile_content.go:205] Starting client-ca::kube-system::extension-apiserver-authentication::client-ca-file * I0209 15:23:21.121514 1 shared_informer.go:197] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file * I0209 15:23:21.121674 1 secure_serving.go:178] Serving securely on 127.0.0.1:10259 * I0209 15:23:21.121719 1 tlsconfig.go:219] Starting DynamicServingCertificateController * E0209 15:23:21.123202 1 reflector.go:153] k8s.io/client-go/informers/factory.go:135: Failed to list *v1beta1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User "system:kube-scheduler" cannot list resource "poddisruptionbudgets" in API group "policy" at the cluster scope * E0209 15:23:21.123205 1 reflector.go:153] k8s.io/apiserver/pkg/server/dynamiccertificates/configmap_cafile_content.go:209: Failed to list *v1.ConfigMap: configmaps "extension-apiserver-authentication" is forbidden: User "system:kube-scheduler" cannot list resource "configmaps" in API group "" in the namespace "kube-system" * E0209 15:23:21.123257 1 reflector.go:153] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.Service: services is forbidden: User "system:kube-scheduler" cannot list resource "services" in API group "" at the cluster scope * E0209 15:23:21.123279 1 reflector.go:153] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumes" in API group "" at the cluster scope * E0209 15:23:21.123292 1 reflector.go:153] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User "system:kube-scheduler" cannot list resource "replicationcontrollers" in API group "" at the cluster scope * E0209 15:23:21.123335 1 reflector.go:153] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "storageclasses" in API group "storage.k8s.io" at the cluster scope * E0209 15:23:21.123356 1 reflector.go:153] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.Node: nodes is forbidden: User "system:kube-scheduler" cannot list resource "nodes" in API group "" at the cluster scope * E0209 15:23:21.123406 1 reflector.go:153] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User "system:kube-scheduler" cannot list resource "statefulsets" in API group "apps" at the cluster scope * E0209 15:23:21.123412 1 reflector.go:153] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csinodes" in API group "storage.k8s.io" at the cluster scope * E0209 15:23:21.123445 1 reflector.go:153] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User "system:kube-scheduler" cannot list resource "replicasets" in API group "apps" at the cluster scope * E0209 15:23:21.123479 1 reflector.go:153] k8s.io/kubernetes/cmd/kube-scheduler/app/server.go:246: Failed to list *v1.Pod: pods is forbidden: User "system:kube-scheduler" cannot list resource "pods" in API group "" at the cluster scope * E0209 15:23:21.123508 1 reflector.go:153] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumeclaims" in API group "" at the cluster scope * E0209 15:23:22.123944 1 reflector.go:153] k8s.io/client-go/informers/factory.go:135: Failed to list *v1beta1.PodDisruptionBudget: poddisruptionbudgets.policy is forbidden: User "system:kube-scheduler" cannot list resource "poddisruptionbudgets" in API group "policy" at the cluster scope * E0209 15:23:22.125658 1 reflector.go:153] k8s.io/apiserver/pkg/server/dynamiccertificates/configmap_cafile_content.go:209: Failed to list *v1.ConfigMap: configmaps "extension-apiserver-authentication" is forbidden: User "system:kube-scheduler" cannot list resource "configmaps" in API group "" in the namespace "kube-system" * E0209 15:23:22.126717 1 reflector.go:153] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.Service: services is forbidden: User "system:kube-scheduler" cannot list resource "services" in API group "" at the cluster scope * E0209 15:23:22.127905 1 reflector.go:153] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.PersistentVolume: persistentvolumes is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumes" in API group "" at the cluster scope * E0209 15:23:22.129170 1 reflector.go:153] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.ReplicationController: replicationcontrollers is forbidden: User "system:kube-scheduler" cannot list resource "replicationcontrollers" in API group "" at the cluster scope * E0209 15:23:22.130373 1 reflector.go:153] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.StorageClass: storageclasses.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "storageclasses" in API group "storage.k8s.io" at the cluster scope * E0209 15:23:22.131346 1 reflector.go:153] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.Node: nodes is forbidden: User "system:kube-scheduler" cannot list resource "nodes" in API group "" at the cluster scope * E0209 15:23:22.132586 1 reflector.go:153] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.StatefulSet: statefulsets.apps is forbidden: User "system:kube-scheduler" cannot list resource "statefulsets" in API group "apps" at the cluster scope * E0209 15:23:22.133856 1 reflector.go:153] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.CSINode: csinodes.storage.k8s.io is forbidden: User "system:kube-scheduler" cannot list resource "csinodes" in API group "storage.k8s.io" at the cluster scope * E0209 15:23:22.134929 1 reflector.go:153] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.ReplicaSet: replicasets.apps is forbidden: User "system:kube-scheduler" cannot list resource "replicasets" in API group "apps" at the cluster scope * E0209 15:23:22.135943 1 reflector.go:153] k8s.io/kubernetes/cmd/kube-scheduler/app/server.go:246: Failed to list *v1.Pod: pods is forbidden: User "system:kube-scheduler" cannot list resource "pods" in API group "" at the cluster scope * E0209 15:23:22.136970 1 reflector.go:153] k8s.io/client-go/informers/factory.go:135: Failed to list *v1.PersistentVolumeClaim: persistentvolumeclaims is forbidden: User "system:kube-scheduler" cannot list resource "persistentvolumeclaims" in API group "" at the cluster scope * I0209 15:23:23.221808 1 shared_informer.go:204] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::client-ca-file * I0209 15:23:23.221910 1 leaderelection.go:242] attempting to acquire leader lease kube-system/kube-scheduler... * I0209 15:23:23.227348 1 leaderelection.go:252] successfully acquired lease kube-system/kube-scheduler * E0209 15:23:32.336961 1 factory.go:494] pod is already present in the activeQ * * ==> kubelet <== * -- Logs begin at Sun 2020-02-09 15:22:03 UTC, end at Sun 2020-02-09 16:22:02 UTC. -- * Feb 09 15:23:24 minikube kubelet[4807]: W0209 15:23:24.059376 4807 hostport_manager.go:69] The binary conntrack is not installed, this can cause failures in network connection cleanup. * Feb 09 15:23:24 minikube kubelet[4807]: I0209 15:23:24.060087 4807 docker_service.go:255] Docker cri networking managed by kubernetes.io/no-op * Feb 09 15:23:24 minikube kubelet[4807]: I0209 15:23:24.068436 4807 docker_service.go:260] Docker Info: &{ID:WGJF:QQZY:GQVU:4QER:IZUQ:5XPQ:D4JK:DXTO:LBGI:575C:U3YU:3SMB Containers:8 ContainersRunning:8 ContainersPaused:0 ContainersStopped:0 Images:10 Driver:overlay2 DriverStatus:[[Backing Filesystem extfs] [Supports d_type true] [Native Overlay Diff true]] SystemStatus:[] Plugins:{Volume:[local] Network:[bridge host ipvlan macvlan null overlay] Authorization:[] Log:[awslogs fluentd gcplogs gelf journald json-file local logentries splunk syslog]} MemoryLimit:true SwapLimit:true KernelMemory:true KernelMemoryTCP:true CPUCfsPeriod:true CPUCfsQuota:true CPUShares:true CPUSet:true PidsLimit:true IPv4Forwarding:true BridgeNfIptables:true BridgeNfIP6tables:true Debug:false NFd:66 OomKillDisable:true NGoroutines:75 SystemTime:2020-02-09T15:23:24.060737784Z LoggingDriver:json-file CgroupDriver:cgroupfs NEventsListener:0 KernelVersion:4.19.88 OperatingSystem:Buildroot 2019.02.8 OSType:linux Architecture:x86_64 IndexServerAddress:https://index.docker.io/v1/ RegistryConfig:0xc00072e070 NCPU:4 MemTotal:4131684352 GenericResources:[] DockerRootDir:/var/lib/docker HTTPProxy: HTTPSProxy: NoProxy: Name:minikube Labels:[provider=hyperv] ExperimentalBuild:false ServerVersion:19.03.5 ClusterStore: ClusterAdvertise: Runtimes:map[runc:{Path:runc Args:[]}] DefaultRuntime:runc Swarm:{NodeID: NodeAddr: LocalNodeState:inactive ControlAvailable:false Error: RemoteManagers:[] Nodes:0 Managers:0 Cluster:<nil> Warnings:[]} LiveRestoreEnabled:false Isolation: InitBinary:docker-init ContainerdCommit:{ID:b34a5c8af56e510852c35414db4c1f4fa6172339 Expected:b34a5c8af56e510852c35414db4c1f4fa6172339} RuncCommit:{ID:d736ef14f0288d6993a1845745d6756cfc9ddd5a Expected:d736ef14f0288d6993a1845745d6756cfc9ddd5a} InitCommit:{ID:fec3683 Expected:fec3683} SecurityOptions:[name=seccomp,profile=default] ProductLicense:Community Engine Warnings:[]} * Feb 09 15:23:24 minikube kubelet[4807]: I0209 15:23:24.068498 4807 docker_service.go:273] Setting cgroupDriver to cgroupfs * Feb 09 15:23:24 minikube kubelet[4807]: I0209 15:23:24.075778 4807 remote_runtime.go:59] parsed scheme: "" * Feb 09 15:23:24 minikube kubelet[4807]: I0209 15:23:24.075802 4807 remote_runtime.go:59] scheme "" not registered, fallback to default scheme * Feb 09 15:23:24 minikube kubelet[4807]: I0209 15:23:24.075818 4807 passthrough.go:48] ccResolverWrapper: sending update to cc: {[{/var/run/dockershim.sock 0 <nil>}] <nil>} * Feb 09 15:23:24 minikube kubelet[4807]: I0209 15:23:24.075823 4807 clientconn.go:577] ClientConn switching balancer to "pick_first" * Feb 09 15:23:24 minikube kubelet[4807]: I0209 15:23:24.075854 4807 remote_image.go:50] parsed scheme: "" * Feb 09 15:23:24 minikube kubelet[4807]: I0209 15:23:24.075860 4807 remote_image.go:50] scheme "" not registered, fallback to default scheme * Feb 09 15:23:24 minikube kubelet[4807]: I0209 15:23:24.075866 4807 passthrough.go:48] ccResolverWrapper: sending update to cc: {[{/var/run/dockershim.sock 0 <nil>}] <nil>} * Feb 09 15:23:24 minikube kubelet[4807]: I0209 15:23:24.075870 4807 clientconn.go:577] ClientConn switching balancer to "pick_first" * Feb 09 15:23:44 minikube kubelet[4807]: E0209 15:23:44.383561 4807 aws_credentials.go:77] while getting AWS credentials NoCredentialProviders: no valid providers in chain. Deprecated. * Feb 09 15:23:44 minikube kubelet[4807]: For verbose messaging see aws.Config.CredentialsChainVerboseErrors * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.392129 4807 kuberuntime_manager.go:211] Container runtime docker initialized, version: 19.03.5, apiVersion: 1.40.0 * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.400992 4807 server.go:1113] Started kubelet * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.401041 4807 server.go:143] Starting to listen on 0.0.0.0:10250 * Feb 09 15:23:44 minikube kubelet[4807]: E0209 15:23:44.401049 4807 kubelet.go:1302] Image garbage collection failed once. Stats initialization may not have completed yet: failed to get imageFs info: unable to find data in memory cache * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.401546 4807 server.go:354] Adding debug handlers to kubelet server. * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.401826 4807 fs_resource_analyzer.go:64] Starting FS ResourceAnalyzer * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.403047 4807 volume_manager.go:265] Starting Kubelet Volume Manager * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.403180 4807 desired_state_of_world_populator.go:138] Desired state populator starts to run * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.410733 4807 status_manager.go:157] Starting to sync pod status with apiserver * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.410760 4807 kubelet.go:1820] Starting kubelet main sync loop. * Feb 09 15:23:44 minikube kubelet[4807]: E0209 15:23:44.410816 4807 kubelet.go:1844] skipping pod synchronization - [container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful] * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.477297 4807 cpu_manager.go:173] [cpumanager] starting with none policy * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.477320 4807 cpu_manager.go:174] [cpumanager] reconciling every 10s * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.477329 4807 policy_none.go:43] [cpumanager] none policy: Start * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.478181 4807 plugin_manager.go:114] Starting Kubelet Plugin Manager * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.503276 4807 kubelet_node_status.go:294] Setting node annotation to enable volume controller attach/detach * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.527413 4807 kubelet_node_status.go:70] Attempting to register node minikube * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.704211 4807 reconciler.go:209] operationExecutor.VerifyControllerAttachedVolume started for volume "usr-share-ca-certificates" (UniqueName: "kubernetes.io/host-path/0ae6cf402f641e9b795a3aebca394220-usr-share-ca-certificates") pod "kube-controller-manager-minikube" (UID: "0ae6cf402f641e9b795a3aebca394220") * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.704446 4807 reconciler.go:209] operationExecutor.VerifyControllerAttachedVolume started for volume "xtables-lock" (UniqueName: "kubernetes.io/host-path/aec7f878-f624-400e-b3ac-7879ab47ec0c-xtables-lock") pod "kube-proxy-tkmbk" (UID: "aec7f878-f624-400e-b3ac-7879ab47ec0c") * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.704519 4807 reconciler.go:209] operationExecutor.VerifyControllerAttachedVolume started for volume "kube-proxy-token-92jtn" (UniqueName: "kubernetes.io/secret/aec7f878-f624-400e-b3ac-7879ab47ec0c-kube-proxy-token-92jtn") pod "kube-proxy-tkmbk" (UID: "aec7f878-f624-400e-b3ac-7879ab47ec0c") * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.704592 4807 reconciler.go:209] operationExecutor.VerifyControllerAttachedVolume started for volume "ca-certs" (UniqueName: "kubernetes.io/host-path/c1b171d918aa89531bd5657acb065f84-ca-certs") pod "kube-apiserver-minikube" (UID: "c1b171d918aa89531bd5657acb065f84") * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.704661 4807 reconciler.go:209] operationExecutor.VerifyControllerAttachedVolume started for volume "k8s-certs" (UniqueName: "kubernetes.io/host-path/c1b171d918aa89531bd5657acb065f84-k8s-certs") pod "kube-apiserver-minikube" (UID: "c1b171d918aa89531bd5657acb065f84") * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.704803 4807 reconciler.go:209] operationExecutor.VerifyControllerAttachedVolume started for volume "usr-share-ca-certificates" (UniqueName: "kubernetes.io/host-path/c1b171d918aa89531bd5657acb065f84-usr-share-ca-certificates") pod "kube-apiserver-minikube" (UID: "c1b171d918aa89531bd5657acb065f84") * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.705005 4807 reconciler.go:209] operationExecutor.VerifyControllerAttachedVolume started for volume "kubeconfig" (UniqueName: "kubernetes.io/host-path/0ae6cf402f641e9b795a3aebca394220-kubeconfig") pod "kube-controller-manager-minikube" (UID: "0ae6cf402f641e9b795a3aebca394220") * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.705132 4807 reconciler.go:209] operationExecutor.VerifyControllerAttachedVolume started for volume "kubeconfig" (UniqueName: "kubernetes.io/host-path/9c994ea62a2d8d6f1bb7498f10aa6fcf-kubeconfig") pod "kube-scheduler-minikube" (UID: "9c994ea62a2d8d6f1bb7498f10aa6fcf") * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.705208 4807 reconciler.go:209] operationExecutor.VerifyControllerAttachedVolume started for volume "kube-proxy" (UniqueName: "kubernetes.io/configmap/aec7f878-f624-400e-b3ac-7879ab47ec0c-kube-proxy") pod "kube-proxy-tkmbk" (UID: "aec7f878-f624-400e-b3ac-7879ab47ec0c") * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.705283 4807 reconciler.go:209] operationExecutor.VerifyControllerAttachedVolume started for volume "lib-modules" (UniqueName: "kubernetes.io/host-path/aec7f878-f624-400e-b3ac-7879ab47ec0c-lib-modules") pod "kube-proxy-tkmbk" (UID: "aec7f878-f624-400e-b3ac-7879ab47ec0c") * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.705340 4807 reconciler.go:209] operationExecutor.VerifyControllerAttachedVolume started for volume "etcd-certs" (UniqueName: "kubernetes.io/host-path/8df3b32a9a938b2ec4d0d2782b34f10e-etcd-certs") pod "etcd-minikube" (UID: "8df3b32a9a938b2ec4d0d2782b34f10e") * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.705395 4807 reconciler.go:209] operationExecutor.VerifyControllerAttachedVolume started for volume "ca-certs" (UniqueName: "kubernetes.io/host-path/0ae6cf402f641e9b795a3aebca394220-ca-certs") pod "kube-controller-manager-minikube" (UID: "0ae6cf402f641e9b795a3aebca394220") * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.705451 4807 reconciler.go:209] operationExecutor.VerifyControllerAttachedVolume started for volume "flexvolume-dir" (UniqueName: "kubernetes.io/host-path/0ae6cf402f641e9b795a3aebca394220-flexvolume-dir") pod "kube-controller-manager-minikube" (UID: "0ae6cf402f641e9b795a3aebca394220") * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.705665 4807 reconciler.go:209] operationExecutor.VerifyControllerAttachedVolume started for volume "etcd-data" (UniqueName: "kubernetes.io/host-path/8df3b32a9a938b2ec4d0d2782b34f10e-etcd-data") pod "etcd-minikube" (UID: "8df3b32a9a938b2ec4d0d2782b34f10e") * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.705942 4807 reconciler.go:209] operationExecutor.VerifyControllerAttachedVolume started for volume "k8s-certs" (UniqueName: "kubernetes.io/host-path/0ae6cf402f641e9b795a3aebca394220-k8s-certs") pod "kube-controller-manager-minikube" (UID: "0ae6cf402f641e9b795a3aebca394220") * Feb 09 15:23:44 minikube kubelet[4807]: I0209 15:23:44.706094 4807 reconciler.go:156] Reconciler: start to sync state * Feb 09 15:23:45 minikube kubelet[4807]: I0209 15:23:45.394000 4807 kubelet_node_status.go:112] Node minikube was previously registered * Feb 09 15:23:45 minikube kubelet[4807]: I0209 15:23:45.394079 4807 kubelet_node_status.go:73] Successfully registered node minikube * Feb 09 15:23:45 minikube kubelet[4807]: I0209 15:23:45.508553 4807 reconciler.go:209] operationExecutor.VerifyControllerAttachedVolume started for volume "storage-provisioner-token-dprnq" (UniqueName: "kubernetes.io/secret/20756ee1-7bca-4a20-b7bb-f46230f3c719-storage-provisioner-token-dprnq") pod "storage-provisioner" (UID: "20756ee1-7bca-4a20-b7bb-f46230f3c719") * Feb 09 15:23:45 minikube kubelet[4807]: I0209 15:23:45.508614 4807 reconciler.go:209] operationExecutor.VerifyControllerAttachedVolume started for volume "tmp" (UniqueName: "kubernetes.io/host-path/20756ee1-7bca-4a20-b7bb-f46230f3c719-tmp") pod "storage-provisioner" (UID: "20756ee1-7bca-4a20-b7bb-f46230f3c719") * Feb 09 15:23:46 minikube kubelet[4807]: E0209 15:23:46.008320 4807 kubelet.go:1662] Failed creating a mirror pod for "kube-scheduler-minikube_kube-system(9c994ea62a2d8d6f1bb7498f10aa6fcf)": pods "kube-scheduler-minikube" already exists * Feb 09 15:23:47 minikube kubelet[4807]: I0209 15:23:47.214091 4807 reconciler.go:209] operationExecutor.VerifyControllerAttachedVolume started for volume "coredns-token-gc9r4" (UniqueName: "kubernetes.io/secret/61df07f7-738e-4543-9398-a144fac001b5-coredns-token-gc9r4") pod "coredns-6955765f44-wprg2" (UID: "61df07f7-738e-4543-9398-a144fac001b5") * Feb 09 15:23:47 minikube kubelet[4807]: I0209 15:23:47.214457 4807 reconciler.go:209] operationExecutor.VerifyControllerAttachedVolume started for volume "config-volume" (UniqueName: "kubernetes.io/configmap/61df07f7-738e-4543-9398-a144fac001b5-config-volume") pod "coredns-6955765f44-wprg2" (UID: "61df07f7-738e-4543-9398-a144fac001b5") * Feb 09 15:23:47 minikube kubelet[4807]: I0209 15:23:47.214492 4807 reconciler.go:209] operationExecutor.VerifyControllerAttachedVolume started for volume "coredns-token-gc9r4" (UniqueName: "kubernetes.io/secret/c17abad8-a59d-46e3-9dd1-fbca191e2416-coredns-token-gc9r4") pod "coredns-6955765f44-2cfq5" (UID: "c17abad8-a59d-46e3-9dd1-fbca191e2416") * Feb 09 15:23:47 minikube kubelet[4807]: I0209 15:23:47.214509 4807 reconciler.go:209] operationExecutor.VerifyControllerAttachedVolume started for volume "config-volume" (UniqueName: "kubernetes.io/configmap/c17abad8-a59d-46e3-9dd1-fbca191e2416-config-volume") pod "coredns-6955765f44-2cfq5" (UID: "c17abad8-a59d-46e3-9dd1-fbca191e2416") * Feb 09 15:23:48 minikube kubelet[4807]: W0209 15:23:48.189840 4807 docker_sandbox.go:394] failed to read pod IP from plugin/docker: Couldn't find network status for kube-system/coredns-6955765f44-2cfq5 through plugin: invalid network status for * Feb 09 15:23:48 minikube kubelet[4807]: W0209 15:23:48.195778 4807 docker_sandbox.go:394] failed to read pod IP from plugin/docker: Couldn't find network status for kube-system/coredns-6955765f44-wprg2 through plugin: invalid network status for * Feb 09 15:23:48 minikube kubelet[4807]: W0209 15:23:48.447004 4807 docker_sandbox.go:394] failed to read pod IP from plugin/docker: Couldn't find network status for kube-system/coredns-6955765f44-2cfq5 through plugin: invalid network status for * Feb 09 15:23:48 minikube kubelet[4807]: W0209 15:23:48.450245 4807 docker_sandbox.go:394] failed to read pod IP from plugin/docker: Couldn't find network status for kube-system/coredns-6955765f44-wprg2 through plugin: invalid network status for * * ==> storage-provisioner ["57946b280cca"] <== </details> **The operating system version**: Windows 10 Pro (1909) with Hyper-V When using minikube version 1.5.2 or version 1.6.2, it is working fine: the file `C:\Users\Sebastien\.minikube\files\etc\hosts` is correctly copied in the VM as `/etc/hosts` after minikube has started. But it doesnt work anymore with minikube version 1.7.2. Two related questions: * How I can get logs when starting minikube? Adding the --v=7 shows nothing more. * I just would like to add an extra host in the VM /etc/hosts file: is-there a better/cleaner option to do that instead of copying a new file?
non_process
file copy to vm isn t working anymore on windows the exact command to reproduce the issue minikube start the full output of the command that failed no error reported when minikube is starting i didn t manage to display verbose debug logs when adding v the output of the minikube logs command docker logs begin at sun utc end at sun utc feb minikube dockerd time level info msg loading plugin io containerd service namespaces service type io containerd service feb minikube dockerd time level info msg loading plugin io containerd service snapshots service type io containerd service feb minikube dockerd time level info msg loading plugin io containerd runtime linux type io containerd runtime feb minikube dockerd time level info msg loading plugin io containerd runtime task type io containerd runtime feb minikube dockerd time level info msg loading plugin io containerd monitor cgroups type io containerd monitor feb minikube dockerd time level info msg loading plugin io containerd service tasks service type io containerd service feb minikube dockerd time level info msg loading plugin io containerd internal restart type io containerd internal feb minikube dockerd time level info msg loading plugin io containerd grpc containers type io containerd grpc feb minikube dockerd time level info msg loading plugin io containerd grpc content type io containerd grpc feb minikube dockerd time level info msg loading plugin io containerd grpc diff type io containerd grpc feb minikube dockerd time level info msg loading plugin io containerd grpc events type io containerd grpc feb minikube dockerd time level info msg loading plugin io containerd grpc healthcheck type io containerd grpc feb minikube dockerd time level info msg loading plugin io containerd grpc images type io containerd grpc feb minikube dockerd time level info msg loading plugin io containerd grpc leases type io containerd grpc feb minikube dockerd time level info msg loading plugin io containerd grpc namespaces type io containerd grpc feb minikube dockerd time level info msg loading plugin io containerd internal opt type io containerd internal feb minikube dockerd time level info msg loading plugin io containerd grpc snapshots type io containerd grpc feb minikube dockerd time level info msg loading plugin io containerd grpc tasks type io containerd grpc feb minikube dockerd time level info msg loading plugin io containerd grpc version type io containerd grpc feb minikube dockerd time level info msg loading plugin io containerd grpc introspection type io containerd grpc feb minikube dockerd time level info msg serving address var run docker containerd containerd debug sock feb minikube dockerd time level info msg serving address var run docker containerd containerd sock feb minikube dockerd time level info msg containerd successfully booted in feb minikube dockerd time level info msg parsed scheme unix module grpc feb minikube dockerd time level info msg scheme unix not registered fallback to default scheme module grpc feb minikube dockerd time level info msg ccresolverwrapper sending update to cc module grpc feb minikube dockerd time level info msg clientconn switching balancer to pick first module grpc feb minikube dockerd time level info msg parsed scheme unix module grpc feb minikube dockerd time level info msg scheme unix not registered fallback to default scheme module grpc feb minikube dockerd time level info msg ccresolverwrapper sending update to cc module grpc feb minikube dockerd time level info msg clientconn switching balancer to pick first module grpc feb minikube dockerd time level warning msg your kernel does not support cgroup blkio weight feb minikube dockerd time level warning msg your kernel does not support cgroup blkio weight device feb minikube dockerd time level warning msg your kernel does not support cgroup blkio throttle read bps device feb minikube dockerd time level warning msg your kernel does not support cgroup blkio throttle write bps device feb minikube dockerd time level warning msg your kernel does not support cgroup blkio throttle read iops device feb minikube dockerd time level warning msg your kernel does not support cgroup blkio throttle write iops device feb minikube dockerd time level info msg loading containers start feb minikube dockerd time level info msg loading containers done feb minikube dockerd time level info msg docker daemon commit graphdriver s version feb minikube dockerd time level info msg daemon has completed initialization feb minikube dockerd time level info msg api listen on var run docker sock feb minikube systemd started docker application container engine feb minikube dockerd time level info msg api listen on feb minikube dockerd time level info msg shim containerd shim started address containerd shim moby shim sock debug false pid feb minikube dockerd time level info msg shim containerd shim started address containerd shim moby shim sock debug false pid feb minikube dockerd time level info msg shim containerd shim started address containerd shim moby shim sock debug false pid feb minikube dockerd time level info msg shim containerd shim started address containerd shim moby shim sock debug false pid feb minikube dockerd time level info msg shim containerd shim started address containerd shim moby shim sock debug false pid feb minikube dockerd time level info msg shim containerd shim started address containerd shim moby shim sock debug false pid feb minikube dockerd time level info msg shim containerd shim started address containerd shim moby shim sock debug false pid feb minikube dockerd time level info msg shim containerd shim started address containerd shim moby shim sock debug false pid feb minikube dockerd time level info msg shim containerd shim started address containerd shim moby shim sock debug false pid feb minikube dockerd time level info msg shim containerd shim started address containerd shim moby shim sock debug false pid feb minikube dockerd time level info msg shim containerd shim started address containerd shim moby shim sock debug false pid feb minikube dockerd time level info msg shim containerd shim started address containerd shim moby shim sock debug false pid feb minikube dockerd time level info msg shim containerd shim started address containerd shim moby shim sock debug false pid feb minikube dockerd time level info msg shim containerd shim started address containerd shim moby shim sock debug false pid feb minikube dockerd time level info msg shim containerd shim started address containerd shim moby shim sock debug false pid feb minikube dockerd time level info msg shim containerd shim started address containerd shim moby shim sock debug false pid container status container image created state name attempt pod id minutes ago running coredns minutes ago running coredns minutes ago running storage provisioner minutes ago running kube proxy minutes ago running etcd minutes ago running kube controller manager minutes ago running kube apiserver minutes ago running kube scheduler coredns plugin reload running configuration coredns linux coredns plugin reload running configuration coredns linux dmesg you have booted with nomodeset this means your gpu drivers are disabled any video related functionality will be severely degraded and you may not even be able to suspend the system properly unless you actually understand what nomodeset does you should reboot without enabling it mds cpu bug present and smt on data leak possible see for more details taa cpu bug present and smt on data leak possible see for more details acpi fail to add mmconfig information can t access extended pci configuration space under this bridge found pm timer bug on the chipset due to workarounds for a bug this clock source is slow consider trying other clock sources unstable clock detected switching default tracing clock to global if you want to keep using the local clock then add trace clock local on the kernel command line platform regulatory direct firmware load for regulatory db failed with error psmouse trackpoint failed to get extended button data assuming buttons systemd failed to bump fs file max ignoring invalid argument systemd fstab generator ignoring noauto for root device systemd file usr lib systemd system systemd journald service configures an ip firewall ipaddressdeny any but the local system does not support bpf cgroup based firewalling systemd proceeding without firewalling in effect this warning is only shown for the first loaded unit using ip firewalling nfsd the nfsdcld client tracking upcall will be removed in please transition to using nfsdcltrack vboxguest loading out of tree module taints kernel vboxguest pci device not found probably running on physical hardware systemd fstab generator ignoring noauto for root device systemd fstab generator ignoring noauto for root device systemd fstab generator ignoring noauto for root device systemd fstab generator ignoring noauto for root device kauditd printk skb callbacks suppressed systemd fstab generator ignoring noauto for root device kauditd printk skb callbacks suppressed kauditd printk skb callbacks suppressed nfsd unable to end grace period kernel up min users load average linux minikube smp tue feb pst gnu linux pretty name buildroot kube apiserver genericapiserver go skipping api discovery io because it has no resources genericapiserver go skipping api node io because it has no resources genericapiserver go skipping api rbac authorization io because it has no resources genericapiserver go skipping api scheduling io because it has no resources genericapiserver go skipping api storage io because it has no resources genericapiserver go skipping api apps because it has no resources genericapiserver go skipping api apps because it has no resources plugins go loaded mutating admission controller s successfully in the following order namespacelifecycle limitranger serviceaccount noderestriction taintnodesbycondition priority defaulttolerationseconds defaultstorageclass storageobjectinuseprotection mutatingadmissionwebhook runtimeclass plugins go loaded validating admission controller s successfully in the following order limitranger serviceaccount priority persistentvolumeclaimresize validatingadmissionwebhook runtimeclass resourcequota client go parsed scheme endpoint endpoint go ccresolverwrapper sending new addresses to cc client go parsed scheme endpoint endpoint go ccresolverwrapper sending new addresses to cc dynamic cafile content go starting request header var lib minikube certs front proxy ca crt dynamic cafile content go starting client ca bundle var lib minikube certs ca crt dynamic serving content go starting serving cert var lib minikube certs apiserver crt var lib minikube certs apiserver key secure serving go serving securely on controller go starting openapi aggregationcontroller tlsconfig go starting dynamicservingcertificatecontroller crd finalizer go starting crdfinalizer cluster authentication trust controller go starting cluster authentication trust controller controller apiapproval controller go starting kubernetesapiapprovalpolicyconformantconditioncontroller shared informer go waiting for caches to sync for cluster authentication trust controller customresource discovery controller go starting discoverycontroller naming controller go starting namingconditioncontroller establishing controller go starting establishingcontroller nonstructuralschema controller go starting nonstructuralschemaconditioncontroller apiservice controller go starting apiserviceregistrationcontroller cache go waiting for caches to sync for apiserviceregistrationcontroller controller available controller go starting availableconditioncontroller cache go waiting for caches to sync for availableconditioncontroller controller dynamic cafile content go starting client ca bundle var lib minikube certs ca crt dynamic cafile content go starting request header var lib minikube certs front proxy ca crt controller go starting openapi controller autoregister controller go starting autoregister controller cache go waiting for caches to sync for autoregister controller controller go unable to remove old endpoints from kubernetes service storageerror key not found code key registry masterleases resourceversion additionalerrormsg crdregistration controller go starting crd autoregister controller shared informer go waiting for caches to sync for crd autoregister shared informer go caches are synced for cluster authentication trust controller cache go caches are synced for availableconditioncontroller controller cache go caches are synced for apiserviceregistrationcontroller controller cache go caches are synced for autoregister controller shared informer go caches are synced for crd autoregister controller go openapi aggregationcontroller processing item controller go openapi aggregationcontroller action for item nothing removed from the queue controller go openapi aggregationcontroller action for item internal local delegation chain nothing removed from the queue storage scheduling go created priorityclass system node critical with value storage scheduling go created priorityclass system cluster critical with value storage scheduling go all system priority classes are created successfully or already exist controller go quota admission added evaluator for roles rbac authorization io controller go quota admission added evaluator for rolebindings rbac authorization io lease go resetting endpoints for master service kubernetes to controller go quota admission added evaluator for endpoints controller go quota admission added evaluator for leases coordination io controller go quota admission added evaluator for serviceaccounts controller go quota admission added evaluator for deployments apps controller go quota admission added evaluator for daemonsets apps controller go quota admission added evaluator for replicasets apps controller go quota admission added evaluator for controllerrevisions apps kube controller manager resource quota monitor go quotamonitor running controllermanager go started deployment controllermanager go skipping root ca cert publisher deployment controller go starting deployment controller shared informer go waiting for caches to sync for deployment controllermanager go started horizontalpodautoscaling horizontal go starting hpa controller shared informer go waiting for caches to sync for hpa controllermanager go started ttl ttl controller go starting ttl controller shared informer go waiting for caches to sync for ttl controllermanager go started bootstrapsigner shared informer go waiting for caches to sync for bootstrap signer controllermanager go started serviceaccount controllermanager go skipping nodeipam serviceaccounts controller go starting service account controller shared informer go waiting for caches to sync for service account shared informer go waiting for caches to sync for garbage collector actual state of world go failed to update statusupdateneeded field in actual state of world failed to set statusupdateneeded to needed true because nodename minikube does not exist shared informer go caches are synced for certificate csrapproving shared informer go caches are synced for bootstrap signer shared informer go caches are synced for expand shared informer go caches are synced for certificate csrsigning shared informer go caches are synced for pv protection shared informer go caches are synced for ttl shared informer go caches are synced for clusterroleaggregator shared informer go waiting for caches to sync for resource quota shared informer go caches are synced for pvc protection shared informer go caches are synced for taint shared informer go caches are synced for replicaset taint manager go starting noexecutetaintmanager node lifecycle controller go initializing eviction metric for zone node lifecycle controller go missing timestamp for node minikube assuming now as a timestamp shared informer go caches are synced for attach detach node lifecycle controller go controller detected that all nodes are not ready entering master disruption mode event go event objectreference kind node namespace name minikube uid apiversion resourceversion fieldpath type normal reason registerednode node minikube event registered node minikube in controller shared informer go caches are synced for endpoint shared informer go caches are synced for disruption disruption go sending events to api server shared informer go caches are synced for replicationcontroller shared informer go caches are synced for deployment shared informer go caches are synced for job event go event objectreference kind deployment namespace kube system name coredns uid apiversion apps resourceversion fieldpath type normal reason scalingreplicaset scaled up replica set coredns to event go event objectreference kind replicaset namespace kube system name coredns uid apiversion apps resourceversion fieldpath type normal reason successfulcreate created pod coredns event go event objectreference kind replicaset namespace kube system name coredns uid apiversion apps resourceversion fieldpath type normal reason successfulcreate created pod coredns shared informer go caches are synced for daemon sets shared informer go caches are synced for stateful set shared informer go caches are synced for gc event go event objectreference kind daemonset namespace kube system name kube proxy uid apiversion apps resourceversion fieldpath type normal reason successfulcreate created pod kube proxy tkmbk shared informer go caches are synced for persistent volume daemon controller go kube system kube proxy failed with error storing status for daemon set daemonset typemeta typemeta kind apiversion objectmeta objectmeta name kube proxy generatename namespace kube system selflink apis apps namespaces kube system daemonsets kube proxy uid resourceversion generation creationtimestamp time time time time wall ext loc time location deletiontimestamp time nil deletiongraceperiodseconds nil labels map string app kube proxy annotations map string deprecated daemonset template generation ownerreferences ownerreference nil finalizers string nil clustername managedfields managedfieldsentry nil spec daemonsetspec selector labelselector template podtemplatespec objectmeta objectmeta name generatename namespace selflink uid resourceversion generation creationtimestamp time time time time wall ext loc time location nil deletiontimestamp time nil deletiongraceperiodseconds nil labels map string app kube proxy annotations map string nil ownerreferences ownerreference nil finalizers string nil clustername managedfields managedfieldsentry nil spec podspec volumes volume volume name kube proxy volumesource volumesource hostpath hostpathvolumesource nil emptydir emptydirvolumesource nil gcepersistentdisk gcepersistentdiskvolumesource nil awselasticblockstore awselasticblockstorevolumesource nil gitrepo gitrepovolumesource nil secret secretvolumesource nil nfs nfsvolumesource nil iscsi iscsivolumesource nil glusterfs glusterfsvolumesource nil persistentvolumeclaim persistentvolumeclaimvolumesource nil rbd rbdvolumesource nil flexvolume flexvolumesource nil cinder cindervolumesource nil cephfs cephfsvolumesource nil flocker flockervolumesource nil downwardapi downwardapivolumesource nil fc fcvolumesource nil azurefile azurefilevolumesource nil configmap configmapvolumesource vspherevolume vspherevirtualdiskvolumesource nil quobyte quobytevolumesource nil azuredisk azurediskvolumesource nil photonpersistentdisk photonpersistentdiskvolumesource nil projected projectedvolumesource nil portworxvolume portworxvolumesource nil scaleio scaleiovolumesource nil storageos storageosvolumesource nil csi csivolumesource nil volume name xtables lock volumesource volumesource hostpath hostpathvolumesource emptydir emptydirvolumesource nil gcepersistentdisk gcepersistentdiskvolumesource nil awselasticblockstore awselasticblockstorevolumesource nil gitrepo gitrepovolumesource nil secret secretvolumesource nil nfs nfsvolumesource nil iscsi iscsivolumesource nil glusterfs glusterfsvolumesource nil persistentvolumeclaim persistentvolumeclaimvolumesource nil rbd rbdvolumesource nil flexvolume flexvolumesource nil cinder cindervolumesource nil cephfs cephfsvolumesource nil flocker flockervolumesource nil downwardapi downwardapivolumesource nil fc fcvolumesource nil azurefile azurefilevolumesource nil configmap configmapvolumesource nil vspherevolume vspherevirtualdiskvolumesource nil quobyte quobytevolumesource nil azuredisk azurediskvolumesource nil photonpersistentdisk photonpersistentdiskvolumesource nil projected projectedvolumesource nil portworxvolume portworxvolumesource nil scaleio scaleiovolumesource nil storageos storageosvolumesource nil csi csivolumesource nil volume name lib modules volumesource volumesource hostpath hostpathvolumesource emptydir emptydirvolumesource nil gcepersistentdisk gcepersistentdiskvolumesource nil awselasticblockstore awselasticblockstorevolumesource nil gitrepo gitrepovolumesource nil secret secretvolumesource nil nfs nfsvolumesource nil iscsi iscsivolumesource nil glusterfs glusterfsvolumesource nil persistentvolumeclaim persistentvolumeclaimvolumesource nil rbd rbdvolumesource nil flexvolume flexvolumesource nil cinder cindervolumesource nil cephfs cephfsvolumesource nil flocker flockervolumesource nil downwardapi downwardapivolumesource nil fc fcvolumesource nil azurefile azurefilevolumesource nil configmap configmapvolumesource nil vspherevolume vspherevirtualdiskvolumesource nil quobyte quobytevolumesource nil azuredisk azurediskvolumesource nil photonpersistentdisk photonpersistentdiskvolumesource nil projected projectedvolumesource nil portworxvolume portworxvolumesource nil scaleio scaleiovolumesource nil storageos storageosvolumesource nil csi csivolumesource nil initcontainers container nil containers container container name kube proxy image gcr io kube proxy command string usr local bin kube proxy config var lib kube proxy config conf hostname override node name args string nil workingdir ports containerport nil envfrom envfromsource nil env envvar envvar name node name value valuefrom envvarsource resources resourcerequirements limits resourcelist nil requests resourcelist nil volumemounts volumemount volumemount name kube proxy readonly false mountpath var lib kube proxy subpath mountpropagation mountpropagationmode nil subpathexpr volumemount name xtables lock readonly false mountpath run xtables lock subpath mountpropagation mountpropagationmode nil subpathexpr volumemount name lib modules readonly true mountpath lib modules subpath mountpropagation mountpropagationmode nil subpathexpr volumedevices volumedevice nil livenessprobe probe nil readinessprobe probe nil startupprobe probe nil lifecycle lifecycle nil terminationmessagepath dev termination log terminationmessagepolicy file imagepullpolicy ifnotpresent securitycontext securitycontext stdin false stdinonce false tty false ephemeralcontainers ephemeralcontainer nil restartpolicy always terminationgraceperiodseconds activedeadlineseconds nil dnspolicy clusterfirst nodeselector map string beta kubernetes io os linux serviceaccountname kube proxy deprecatedserviceaccount kube proxy automountserviceaccounttoken bool nil nodename hostnetwork true hostpid false hostipc false shareprocessnamespace bool nil securitycontext podsecuritycontext imagepullsecrets localobjectreference nil hostname subdomain affinity affinity nil schedulername default scheduler tolerations toleration toleration key criticaladdonsonly operator exists value effect tolerationseconds nil toleration key operator exists value effect tolerationseconds nil hostaliases hostalias nil priorityclassname system node critical priority nil dnsconfig poddnsconfig nil readinessgates podreadinessgate nil runtimeclassname string nil enableservicelinks bool nil preemptionpolicy preemptionpolicy nil overhead resourcelist nil topologyspreadconstraints topologyspreadconstraint nil updatestrategy daemonsetupdatestrategy type rollingupdate rollingupdate rollingupdatedaemonset minreadyseconds revisionhistorylimit status daemonsetstatus currentnumberscheduled numbermisscheduled desirednumberscheduled numberready observedgeneration updatednumberscheduled numberavailable numberunavailable collisioncount nil conditions daemonsetcondition nil operation cannot be fulfilled on daemonsets apps kube proxy the object has been modified please apply your changes to the latest version and try again shared informer go caches are synced for hpa shared informer go caches are synced for resource quota shared informer go caches are synced for namespace shared informer go caches are synced for service account shared informer go caches are synced for garbage collector shared informer go caches are synced for garbage collector garbagecollector go garbage collector all resource monitors have synced proceeding to collect garbage shared informer go caches are synced for resource quota node lifecycle controller go controller detected that some nodes are ready exiting master disruption mode kube proxy server others go unknown proxy mode assuming iptables proxy node go successfully retrieved node ip server others go using iptables proxier proxier go clustercidr not specified unable to distinguish between internal and external traffic server go version conntrack go set sysctl net netfilter nf conntrack max to conntrack go setting nf conntrack max to conntrack go setting conntrack hashsize to conntrack go set sysctl net netfilter nf conntrack tcp timeout established to conntrack go set sysctl net netfilter nf conntrack tcp timeout close wait to config go starting service config controller shared informer go waiting for caches to sync for service config config go starting endpoints config controller shared informer go waiting for caches to sync for endpoints config shared informer go caches are synced for endpoints config shared informer go caches are synced for service config kube scheduler serving go generated self signed cert in memory configmap cafile content go unable to load initial ca bundle for client ca kube system extension apiserver authentication client ca file due to configmap extension apiserver authentication not found configmap cafile content go unable to load initial ca bundle for client ca kube system extension apiserver authentication requestheader client ca file due to configmap extension apiserver authentication not found authentication go unable to get configmap extension apiserver authentication in kube system usually fixed by kubectl create rolebinding n kube system rolebinding name role extension apiserver authentication reader serviceaccount your ns your sa authentication go error looking up in cluster authentication configuration configmaps extension apiserver authentication is forbidden user system kube scheduler cannot get resource configmaps in api group in the namespace kube system authentication go continuing without authentication configuration this may treat all requests as anonymous authentication go to require authentication configuration lookup to succeed set authentication tolerate lookup failure false authorization go authorization is disabled authentication go authentication is disabled deprecated insecure serving go serving healthz insecurely on configmap cafile content go starting client ca kube system extension apiserver authentication client ca file shared informer go waiting for caches to sync for client ca kube system extension apiserver authentication client ca file secure serving go serving securely on tlsconfig go starting dynamicservingcertificatecontroller reflector go io client go informers factory go failed to list poddisruptionbudget poddisruptionbudgets policy is forbidden user system kube scheduler cannot list resource poddisruptionbudgets in api group policy at the cluster scope reflector go io apiserver pkg server dynamiccertificates configmap cafile content go failed to list configmap configmaps extension apiserver authentication is forbidden user system kube scheduler cannot list resource configmaps in api group in the namespace kube system reflector go io client go informers factory go failed to list service services is forbidden user system kube scheduler cannot list resource services in api group at the cluster scope reflector go io client go informers factory go failed to list persistentvolume persistentvolumes is forbidden user system kube scheduler cannot list resource persistentvolumes in api group at the cluster scope reflector go io client go informers factory go failed to list replicationcontroller replicationcontrollers is forbidden user system kube scheduler cannot list resource replicationcontrollers in api group at the cluster scope reflector go io client go informers factory go failed to list storageclass storageclasses storage io is forbidden user system kube scheduler cannot list resource storageclasses in api group storage io at the cluster scope reflector go io client go informers factory go failed to list node nodes is forbidden user system kube scheduler cannot list resource nodes in api group at the cluster scope reflector go io client go informers factory go failed to list statefulset statefulsets apps is forbidden user system kube scheduler cannot list resource statefulsets in api group apps at the cluster scope reflector go io client go informers factory go failed to list csinode csinodes storage io is forbidden user system kube scheduler cannot list resource csinodes in api group storage io at the cluster scope reflector go io client go informers factory go failed to list replicaset replicasets apps is forbidden user system kube scheduler cannot list resource replicasets in api group apps at the cluster scope reflector go io kubernetes cmd kube scheduler app server go failed to list pod pods is forbidden user system kube scheduler cannot list resource pods in api group at the cluster scope reflector go io client go informers factory go failed to list persistentvolumeclaim persistentvolumeclaims is forbidden user system kube scheduler cannot list resource persistentvolumeclaims in api group at the cluster scope reflector go io client go informers factory go failed to list poddisruptionbudget poddisruptionbudgets policy is forbidden user system kube scheduler cannot list resource poddisruptionbudgets in api group policy at the cluster scope reflector go io apiserver pkg server dynamiccertificates configmap cafile content go failed to list configmap configmaps extension apiserver authentication is forbidden user system kube scheduler cannot list resource configmaps in api group in the namespace kube system reflector go io client go informers factory go failed to list service services is forbidden user system kube scheduler cannot list resource services in api group at the cluster scope reflector go io client go informers factory go failed to list persistentvolume persistentvolumes is forbidden user system kube scheduler cannot list resource persistentvolumes in api group at the cluster scope reflector go io client go informers factory go failed to list replicationcontroller replicationcontrollers is forbidden user system kube scheduler cannot list resource replicationcontrollers in api group at the cluster scope reflector go io client go informers factory go failed to list storageclass storageclasses storage io is forbidden user system kube scheduler cannot list resource storageclasses in api group storage io at the cluster scope reflector go io client go informers factory go failed to list node nodes is forbidden user system kube scheduler cannot list resource nodes in api group at the cluster scope reflector go io client go informers factory go failed to list statefulset statefulsets apps is forbidden user system kube scheduler cannot list resource statefulsets in api group apps at the cluster scope reflector go io client go informers factory go failed to list csinode csinodes storage io is forbidden user system kube scheduler cannot list resource csinodes in api group storage io at the cluster scope reflector go io client go informers factory go failed to list replicaset replicasets apps is forbidden user system kube scheduler cannot list resource replicasets in api group apps at the cluster scope reflector go io kubernetes cmd kube scheduler app server go failed to list pod pods is forbidden user system kube scheduler cannot list resource pods in api group at the cluster scope reflector go io client go informers factory go failed to list persistentvolumeclaim persistentvolumeclaims is forbidden user system kube scheduler cannot list resource persistentvolumeclaims in api group at the cluster scope shared informer go caches are synced for client ca kube system extension apiserver authentication client ca file leaderelection go attempting to acquire leader lease kube system kube scheduler leaderelection go successfully acquired lease kube system kube scheduler factory go pod is already present in the activeq kubelet logs begin at sun utc end at sun utc feb minikube kubelet hostport manager go the binary conntrack is not installed this can cause failures in network connection cleanup feb minikube kubelet docker service go docker cri networking managed by kubernetes io no op feb minikube kubelet docker service go docker info id wgjf qqzy gqvu izuq dxto lbgi containers containersrunning containerspaused containersstopped images driver driverstatus systemstatus plugins volume network authorization log memorylimit true swaplimit true kernelmemory true kernelmemorytcp true cpucfsperiod true cpucfsquota true cpushares true cpuset true pidslimit true true bridgenfiptables true true debug false nfd oomkilldisable true ngoroutines systemtime loggingdriver json file cgroupdriver cgroupfs neventslistener kernelversion operatingsystem buildroot ostype linux architecture indexserveraddress registryconfig ncpu memtotal genericresources dockerrootdir var lib docker httpproxy httpsproxy noproxy name minikube labels experimentalbuild false serverversion clusterstore clusteradvertise runtimes map defaultruntime runc swarm nodeid nodeaddr localnodestate inactive controlavailable false error remotemanagers nodes managers cluster warnings liverestoreenabled false isolation initbinary docker init containerdcommit id expected runccommit id expected initcommit id expected securityoptions productlicense community engine warnings feb minikube kubelet docker service go setting cgroupdriver to cgroupfs feb minikube kubelet remote runtime go parsed scheme feb minikube kubelet remote runtime go scheme not registered fallback to default scheme feb minikube kubelet passthrough go ccresolverwrapper sending update to cc feb minikube kubelet clientconn go clientconn switching balancer to pick first feb minikube kubelet remote image go parsed scheme feb minikube kubelet remote image go scheme not registered fallback to default scheme feb minikube kubelet passthrough go ccresolverwrapper sending update to cc feb minikube kubelet clientconn go clientconn switching balancer to pick first feb minikube kubelet aws credentials go while getting aws credentials nocredentialproviders no valid providers in chain deprecated feb minikube kubelet for verbose messaging see aws config credentialschainverboseerrors feb minikube kubelet kuberuntime manager go container runtime docker initialized version apiversion feb minikube kubelet server go started kubelet feb minikube kubelet server go starting to listen on feb minikube kubelet kubelet go image garbage collection failed once stats initialization may not have completed yet failed to get imagefs info unable to find data in memory cache feb minikube kubelet server go adding debug handlers to kubelet server feb minikube kubelet fs resource analyzer go starting fs resourceanalyzer feb minikube kubelet volume manager go starting kubelet volume manager feb minikube kubelet desired state of world populator go desired state populator starts to run feb minikube kubelet status manager go starting to sync pod status with apiserver feb minikube kubelet kubelet go starting kubelet main sync loop feb minikube kubelet kubelet go skipping pod synchronization feb minikube kubelet cpu manager go starting with none policy feb minikube kubelet cpu manager go reconciling every feb minikube kubelet policy none go none policy start feb minikube kubelet plugin manager go starting kubelet plugin manager feb minikube kubelet kubelet node status go setting node annotation to enable volume controller attach detach feb minikube kubelet kubelet node status go attempting to register node minikube feb minikube kubelet reconciler go operationexecutor verifycontrollerattachedvolume started for volume usr share ca certificates uniquename kubernetes io host path usr share ca certificates pod kube controller manager minikube uid feb minikube kubelet reconciler go operationexecutor verifycontrollerattachedvolume started for volume xtables lock uniquename kubernetes io host path xtables lock pod kube proxy tkmbk uid feb minikube kubelet reconciler go operationexecutor verifycontrollerattachedvolume started for volume kube proxy token uniquename kubernetes io secret kube proxy token pod kube proxy tkmbk uid feb minikube kubelet reconciler go operationexecutor verifycontrollerattachedvolume started for volume ca certs uniquename kubernetes io host path ca certs pod kube apiserver minikube uid feb minikube kubelet reconciler go operationexecutor verifycontrollerattachedvolume started for volume certs uniquename kubernetes io host path certs pod kube apiserver minikube uid feb minikube kubelet reconciler go operationexecutor verifycontrollerattachedvolume started for volume usr share ca certificates uniquename kubernetes io host path usr share ca certificates pod kube apiserver minikube uid feb minikube kubelet reconciler go operationexecutor verifycontrollerattachedvolume started for volume kubeconfig uniquename kubernetes io host path kubeconfig pod kube controller manager minikube uid feb minikube kubelet reconciler go operationexecutor verifycontrollerattachedvolume started for volume kubeconfig uniquename kubernetes io host path kubeconfig pod kube scheduler minikube uid feb minikube kubelet reconciler go operationexecutor verifycontrollerattachedvolume started for volume kube proxy uniquename kubernetes io configmap kube proxy pod kube proxy tkmbk uid feb minikube kubelet reconciler go operationexecutor verifycontrollerattachedvolume started for volume lib modules uniquename kubernetes io host path lib modules pod kube proxy tkmbk uid feb minikube kubelet reconciler go operationexecutor verifycontrollerattachedvolume started for volume etcd certs uniquename kubernetes io host path etcd certs pod etcd minikube uid feb minikube kubelet reconciler go operationexecutor verifycontrollerattachedvolume started for volume ca certs uniquename kubernetes io host path ca certs pod kube controller manager minikube uid feb minikube kubelet reconciler go operationexecutor verifycontrollerattachedvolume started for volume flexvolume dir uniquename kubernetes io host path flexvolume dir pod kube controller manager minikube uid feb minikube kubelet reconciler go operationexecutor verifycontrollerattachedvolume started for volume etcd data uniquename kubernetes io host path etcd data pod etcd minikube uid feb minikube kubelet reconciler go operationexecutor verifycontrollerattachedvolume started for volume certs uniquename kubernetes io host path certs pod kube controller manager minikube uid feb minikube kubelet reconciler go reconciler start to sync state feb minikube kubelet kubelet node status go node minikube was previously registered feb minikube kubelet kubelet node status go successfully registered node minikube feb minikube kubelet reconciler go operationexecutor verifycontrollerattachedvolume started for volume storage provisioner token dprnq uniquename kubernetes io secret storage provisioner token dprnq pod storage provisioner uid feb minikube kubelet reconciler go operationexecutor verifycontrollerattachedvolume started for volume tmp uniquename kubernetes io host path tmp pod storage provisioner uid feb minikube kubelet kubelet go failed creating a mirror pod for kube scheduler minikube kube system pods kube scheduler minikube already exists feb minikube kubelet reconciler go operationexecutor verifycontrollerattachedvolume started for volume coredns token uniquename kubernetes io secret coredns token pod coredns uid feb minikube kubelet reconciler go operationexecutor verifycontrollerattachedvolume started for volume config volume uniquename kubernetes io configmap config volume pod coredns uid feb minikube kubelet reconciler go operationexecutor verifycontrollerattachedvolume started for volume coredns token uniquename kubernetes io secret coredns token pod coredns uid feb minikube kubelet reconciler go operationexecutor verifycontrollerattachedvolume started for volume config volume uniquename kubernetes io configmap config volume pod coredns uid feb minikube kubelet docker sandbox go failed to read pod ip from plugin docker couldn t find network status for kube system coredns through plugin invalid network status for feb minikube kubelet docker sandbox go failed to read pod ip from plugin docker couldn t find network status for kube system coredns through plugin invalid network status for feb minikube kubelet docker sandbox go failed to read pod ip from plugin docker couldn t find network status for kube system coredns through plugin invalid network status for feb minikube kubelet docker sandbox go failed to read pod ip from plugin docker couldn t find network status for kube system coredns through plugin invalid network status for storage provisioner the operating system version windows pro with hyper v when using minikube version or version it is working fine the file c users sebastien minikube files etc hosts is correctly copied in the vm as etc hosts after minikube has started but it doesnt work anymore with minikube version two related questions how i can get logs when starting minikube adding the v shows nothing more i just would like to add an extra host in the vm etc hosts file is there a better cleaner option to do that instead of copying a new file
0
196,733
22,514,208,428
IssuesEvent
2022-06-24 00:41:52
Baneeishaque/Android-Common-Utils17
https://api.github.com/repos/Baneeishaque/Android-Common-Utils17
closed
CVE-2021-35517 (High) detected in commons-compress-1.20.jar - autoclosed
security vulnerability
## CVE-2021-35517 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-compress-1.20.jar</b></p></summary> <p>Apache Commons Compress software defines an API for working with compression and archive formats. These include: bzip2, gzip, pack200, lzma, xz, Snappy, traditional Unix Compress, DEFLATE, DEFLATE64, LZ4, Brotli, Zstandard and ar, cpio, jar, tar, zip, dump, 7z, arj.</p> <p>Library home page: <a href="https://commons.apache.org/proper/commons-compress/">https://commons.apache.org/proper/commons-compress/</a></p> <p>Path to dependency file: /common17/build.gradle</p> <p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.apache.commons/commons-compress/1.20/b8df472b31e1f17c232d2ad78ceb1c84e00c641b/commons-compress-1.20.jar</p> <p> Dependency Hierarchy: - lint-gradle-30.0.0-beta02.jar (Root Library) - repository-30.0.0-beta02.jar - :x: **commons-compress-1.20.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Baneeishaque/Android-Common-Utils17/commit/ff8aaf8af8c3e9382fbf9ac635aa5fcb0569e668">ff8aaf8af8c3e9382fbf9ac635aa5fcb0569e668</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> When reading a specially crafted TAR archive, Compress can be made to allocate large amounts of memory that finally leads to an out of memory error even for very small inputs. This could be used to mount a denial of service attack against services that use Compress' tar package. <p>Publish Date: 2021-07-13 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-35517>CVE-2021-35517</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://commons.apache.org/proper/commons-compress/security-reports.html">https://commons.apache.org/proper/commons-compress/security-reports.html</a></p> <p>Release Date: 2021-07-13</p> <p>Fix Resolution: org.apache.commons:commons-compress:1.21</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-35517 (High) detected in commons-compress-1.20.jar - autoclosed - ## CVE-2021-35517 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-compress-1.20.jar</b></p></summary> <p>Apache Commons Compress software defines an API for working with compression and archive formats. These include: bzip2, gzip, pack200, lzma, xz, Snappy, traditional Unix Compress, DEFLATE, DEFLATE64, LZ4, Brotli, Zstandard and ar, cpio, jar, tar, zip, dump, 7z, arj.</p> <p>Library home page: <a href="https://commons.apache.org/proper/commons-compress/">https://commons.apache.org/proper/commons-compress/</a></p> <p>Path to dependency file: /common17/build.gradle</p> <p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.apache.commons/commons-compress/1.20/b8df472b31e1f17c232d2ad78ceb1c84e00c641b/commons-compress-1.20.jar</p> <p> Dependency Hierarchy: - lint-gradle-30.0.0-beta02.jar (Root Library) - repository-30.0.0-beta02.jar - :x: **commons-compress-1.20.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Baneeishaque/Android-Common-Utils17/commit/ff8aaf8af8c3e9382fbf9ac635aa5fcb0569e668">ff8aaf8af8c3e9382fbf9ac635aa5fcb0569e668</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> When reading a specially crafted TAR archive, Compress can be made to allocate large amounts of memory that finally leads to an out of memory error even for very small inputs. This could be used to mount a denial of service attack against services that use Compress' tar package. <p>Publish Date: 2021-07-13 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-35517>CVE-2021-35517</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://commons.apache.org/proper/commons-compress/security-reports.html">https://commons.apache.org/proper/commons-compress/security-reports.html</a></p> <p>Release Date: 2021-07-13</p> <p>Fix Resolution: org.apache.commons:commons-compress:1.21</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in commons compress jar autoclosed cve high severity vulnerability vulnerable library commons compress jar apache commons compress software defines an api for working with compression and archive formats these include gzip lzma xz snappy traditional unix compress deflate brotli zstandard and ar cpio jar tar zip dump arj library home page a href path to dependency file build gradle path to vulnerable library home wss scanner gradle caches modules files org apache commons commons compress commons compress jar dependency hierarchy lint gradle jar root library repository jar x commons compress jar vulnerable library found in head commit a href vulnerability details when reading a specially crafted tar archive compress can be made to allocate large amounts of memory that finally leads to an out of memory error even for very small inputs this could be used to mount a denial of service attack against services that use compress tar package publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache commons commons compress step up your open source security game with whitesource
0
22,730
32,048,933,930
IssuesEvent
2023-09-23 10:15:16
h4sh5/npm-auto-scanner
https://api.github.com/repos/h4sh5/npm-auto-scanner
opened
@dogu-tech/node 0.0.26 has 1 guarddog issues
npm-silent-process-execution
```{"npm-silent-process-execution":[{"code":" const child = (0, child_process_1.fork)(scriptPath, [pid.toString()], {\n stdio: 'ignore',\n detached: true,\n });","location":"package/build/src/tree-kill.js:63","message":"This package is silently executing another executable"}]}```
1.0
@dogu-tech/node 0.0.26 has 1 guarddog issues - ```{"npm-silent-process-execution":[{"code":" const child = (0, child_process_1.fork)(scriptPath, [pid.toString()], {\n stdio: 'ignore',\n detached: true,\n });","location":"package/build/src/tree-kill.js:63","message":"This package is silently executing another executable"}]}```
process
dogu tech node has guarddog issues npm silent process execution n stdio ignore n detached true n location package build src tree kill js message this package is silently executing another executable
1
197,589
6,961,688,424
IssuesEvent
2017-12-08 10:27:44
GFDRR/geonode
https://api.github.com/repos/GFDRR/geonode
closed
document thumbnail
High Priority
Document thumbnail should display title page of pdf documents. Currently shows some default text, no image. @vdeparday please provide the links to required item to install to enable this.
1.0
document thumbnail - Document thumbnail should display title page of pdf documents. Currently shows some default text, no image. @vdeparday please provide the links to required item to install to enable this.
non_process
document thumbnail document thumbnail should display title page of pdf documents currently shows some default text no image vdeparday please provide the links to required item to install to enable this
0