Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
22,307
30,860,277,855
IssuesEvent
2023-08-03 02:00:09
lizhihao6/get-daily-arxiv-noti
https://api.github.com/repos/lizhihao6/get-daily-arxiv-noti
opened
New submissions for Thu, 3 Aug 23
event camera white balance isp compression image signal processing image signal process raw raw image events camera color contrast events AWB
## Keyword: events There is no result ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB There is no result ## Keyword: ISP ### ELFNet: Evidential Local-global Fusion for Stereo Matching - **Authors:** Jieming Lou, Weide Liu, Zhuo Chen, Fayao Liu, Jun Cheng - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2308.00728 - **Pdf link:** https://arxiv.org/pdf/2308.00728 - **Abstract** Although existing stereo matching models have achieved continuous improvement, they often face issues related to trustworthiness due to the absence of uncertainty estimation. Additionally, effectively leveraging multi-scale and multi-view knowledge of stereo pairs remains unexplored. In this paper, we introduce the \textbf{E}vidential \textbf{L}ocal-global \textbf{F}usion (ELF) framework for stereo matching, which endows both uncertainty estimation and confidence-aware fusion with trustworthy heads. Instead of predicting the disparity map alone, our model estimates an evidential-based disparity considering both aleatoric and epistemic uncertainties. With the normal inverse-Gamma distribution as a bridge, the proposed framework realizes intra evidential fusion of multi-level predictions and inter evidential fusion between cost-volume-based and transformer-based stereo matching. Extensive experimental results show that the proposed framework exploits multi-view information effectively and achieves state-of-the-art overall performance both on accuracy and cross-domain generalization. The codes are available at https://github.com/jimmy19991222/ELFNet. ### Training-Free Instance Segmentation from Semantic Image Segmentation Masks - **Authors:** Yuchen Shen, Dong Zhang, Yuhui Zheng, Zechao Li, Liyong Fu, Qiaolin Ye - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2308.00949 - **Pdf link:** https://arxiv.org/pdf/2308.00949 - **Abstract** In recent years, the development of instance segmentation has garnered significant attention in a wide range of applications. However, the training of a fully-supervised instance segmentation model requires costly both instance-level and pixel-level annotations. In contrast, weakly-supervised instance segmentation methods (i.e., with image-level class labels or point labels) struggle to satisfy the accuracy and recall requirements of practical scenarios. In this paper, we propose a novel paradigm for instance segmentation called training-free instance segmentation (TFISeg), which achieves instance segmentation results from image masks predicted using off-the-shelf semantic segmentation models. TFISeg does not require training a semantic or/and instance segmentation model and avoids the need for instance-level image annotations. Therefore, it is highly efficient. Specifically, we first obtain a semantic segmentation mask of the input image via a trained semantic segmentation model. Then, we calculate a displacement field vector for each pixel based on the segmentation mask, which can indicate representations belonging to the same class but different instances, i.e., obtaining the instance-level object information. Finally, instance segmentation results are obtained after being refined by a learnable category-agnostic object boundary branch. Extensive experimental results on two challenging datasets and representative semantic segmentation baselines (including CNNs and Transformers) demonstrate that TFISeg can achieve competitive results compared to the state-of-the-art fully-supervised instance segmentation methods without the need for additional human resources or increased computational costs. The code is available at: TFISeg ### MDT3D: Multi-Dataset Training for LiDAR 3D Object Detection Generalization - **Authors:** Louis Soum-Fontez, Jean-Emmanuel Deschaud, François Goulette - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG) - **Arxiv link:** https://arxiv.org/abs/2308.01000 - **Pdf link:** https://arxiv.org/pdf/2308.01000 - **Abstract** Supervised 3D Object Detection models have been displaying increasingly better performance in single-domain cases where the training data comes from the same environment and sensor as the testing data. However, in real-world scenarios data from the target domain may not be available for finetuning or for domain adaptation methods. Indeed, 3D object detection models trained on a source dataset with a specific point distribution have shown difficulties in generalizing to unseen datasets. Therefore, we decided to leverage the information available from several annotated source datasets with our Multi-Dataset Training for 3D Object Detection (MDT3D) method to increase the robustness of 3D object detection models when tested in a new environment with a different sensor configuration. To tackle the labelling gap between datasets, we used a new label mapping based on coarse labels. Furthermore, we show how we managed the mix of datasets during training and finally introduce a new cross-dataset augmentation method: cross-dataset object injection. We demonstrate that this training paradigm shows improvements for different types of 3D object detection models. The source code and additional results for this research project will be publicly available on GitHub for interested parties to access and utilize: https://github.com/LouisSF/MDT3D ### WCCNet: Wavelet-integrated CNN with Crossmodal Rearranging Fusion for Fast Multispectral Pedestrian Detection - **Authors:** Xingjian Wang, Li Chai, Jiming Chen, Zhiguo Shi - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2308.01042 - **Pdf link:** https://arxiv.org/pdf/2308.01042 - **Abstract** Multispectral pedestrian detection achieves better visibility in challenging conditions and thus has a broad application in various tasks, for which both the accuracy and computational cost are of paramount importance. Most existing approaches treat RGB and infrared modalities equally, typically adopting two symmetrical CNN backbones for multimodal feature extraction, which ignores the substantial differences between modalities and brings great difficulty for the reduction of the computational cost as well as effective crossmodal fusion. In this work, we propose a novel and efficient framework named WCCNet that is able to differentially extract rich features of different spectra with lower computational complexity and semantically rearranges these features for effective crossmodal fusion. Specifically, the discrete wavelet transform (DWT) allowing fast inference and training speed is embedded to construct a dual-stream backbone for efficient feature extraction. The DWT layers of WCCNet extract frequency components for infrared modality, while the CNN layers extract spatial-domain features for RGB modality. This methodology not only significantly reduces the computational complexity, but also improves the extraction of infrared features to facilitate the subsequent crossmodal fusion. Based on the well extracted features, we elaborately design the crossmodal rearranging fusion module (CMRF), which can mitigate spatial misalignment and merge semantically complementary features of spatially-related local regions to amplify the crossmodal complementary information. We conduct comprehensive evaluations on KAIST and FLIR benchmarks, in which WCCNet outperforms state-of-the-art methods with considerable computational efficiency and competitive accuracy. We also perform the ablation study and analyze thoroughly the impact of different components on the performance of WCCNet. ### Interpretable End-to-End Driving Model for Implicit Scene Understanding - **Authors:** Yiyang Sun, Xiaonian Wang, Yangyang Zhang, Jiagui Tang, Xiaqiang Tang, Jing Yao - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Robotics (cs.RO) - **Arxiv link:** https://arxiv.org/abs/2308.01180 - **Pdf link:** https://arxiv.org/pdf/2308.01180 - **Abstract** Driving scene understanding is to obtain comprehensive scene information through the sensor data and provide a basis for downstream tasks, which is indispensable for the safety of self-driving vehicles. Specific perception tasks, such as object detection and scene graph generation, are commonly used. However, the results of these tasks are only equivalent to the characterization of sampling from high-dimensional scene features, which are not sufficient to represent the scenario. In addition, the goal of perception tasks is inconsistent with human driving that just focuses on what may affect the ego-trajectory. Therefore, we propose an end-to-end Interpretable Implicit Driving Scene Understanding (II-DSU) model to extract implicit high-dimensional scene features as scene understanding results guided by a planning module and to validate the plausibility of scene understanding using auxiliary perception tasks for visualization. Experimental results on CARLA benchmarks show that our approach achieves the new state-of-the-art and is able to obtain scene features that embody richer scene information relevant to driving, enabling superior performance of the downstream planning. ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression There is no result ## Keyword: RAW ### More Context, Less Distraction: Visual Classification by Inferring and Conditioning on Contextual Attributes - **Authors:** Bang An, Sicheng Zhu, Michael-Andrei Panaitescu-Liess, Chaithanya Kumar Mummadi, Furong Huang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Computation and Language (cs.CL); Machine Learning (cs.LG) - **Arxiv link:** https://arxiv.org/abs/2308.01313 - **Pdf link:** https://arxiv.org/pdf/2308.01313 - **Abstract** CLIP, as a foundational vision language model, is widely used in zero-shot image classification due to its ability to understand various visual concepts and natural language descriptions. However, how to fully leverage CLIP's unprecedented human-like understanding capabilities to achieve better zero-shot classification is still an open question. This paper draws inspiration from the human visual perception process: a modern neuroscience view suggests that in classifying an object, humans first infer its class-independent attributes (e.g., background and orientation) which help separate the foreground object from the background, and then make decisions based on this information. Inspired by this, we observe that providing CLIP with contextual attributes improves zero-shot classification and mitigates reliance on spurious features. We also observe that CLIP itself can reasonably infer the attributes from an image. With these observations, we propose a training-free, two-step zero-shot classification method named PerceptionCLIP. Given an image, it first infers contextual attributes (e.g., background) and then performs object classification conditioning on them. Our experiments show that PerceptionCLIP achieves better generalization, group robustness, and better interpretability. For example, PerceptionCLIP with ViT-L/14 improves the worst group accuracy by 16.5% on the Waterbirds dataset and by 3.5% on CelebA. ## Keyword: raw image There is no result
2.0
New submissions for Thu, 3 Aug 23 - ## Keyword: events There is no result ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB There is no result ## Keyword: ISP ### ELFNet: Evidential Local-global Fusion for Stereo Matching - **Authors:** Jieming Lou, Weide Liu, Zhuo Chen, Fayao Liu, Jun Cheng - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2308.00728 - **Pdf link:** https://arxiv.org/pdf/2308.00728 - **Abstract** Although existing stereo matching models have achieved continuous improvement, they often face issues related to trustworthiness due to the absence of uncertainty estimation. Additionally, effectively leveraging multi-scale and multi-view knowledge of stereo pairs remains unexplored. In this paper, we introduce the \textbf{E}vidential \textbf{L}ocal-global \textbf{F}usion (ELF) framework for stereo matching, which endows both uncertainty estimation and confidence-aware fusion with trustworthy heads. Instead of predicting the disparity map alone, our model estimates an evidential-based disparity considering both aleatoric and epistemic uncertainties. With the normal inverse-Gamma distribution as a bridge, the proposed framework realizes intra evidential fusion of multi-level predictions and inter evidential fusion between cost-volume-based and transformer-based stereo matching. Extensive experimental results show that the proposed framework exploits multi-view information effectively and achieves state-of-the-art overall performance both on accuracy and cross-domain generalization. The codes are available at https://github.com/jimmy19991222/ELFNet. ### Training-Free Instance Segmentation from Semantic Image Segmentation Masks - **Authors:** Yuchen Shen, Dong Zhang, Yuhui Zheng, Zechao Li, Liyong Fu, Qiaolin Ye - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2308.00949 - **Pdf link:** https://arxiv.org/pdf/2308.00949 - **Abstract** In recent years, the development of instance segmentation has garnered significant attention in a wide range of applications. However, the training of a fully-supervised instance segmentation model requires costly both instance-level and pixel-level annotations. In contrast, weakly-supervised instance segmentation methods (i.e., with image-level class labels or point labels) struggle to satisfy the accuracy and recall requirements of practical scenarios. In this paper, we propose a novel paradigm for instance segmentation called training-free instance segmentation (TFISeg), which achieves instance segmentation results from image masks predicted using off-the-shelf semantic segmentation models. TFISeg does not require training a semantic or/and instance segmentation model and avoids the need for instance-level image annotations. Therefore, it is highly efficient. Specifically, we first obtain a semantic segmentation mask of the input image via a trained semantic segmentation model. Then, we calculate a displacement field vector for each pixel based on the segmentation mask, which can indicate representations belonging to the same class but different instances, i.e., obtaining the instance-level object information. Finally, instance segmentation results are obtained after being refined by a learnable category-agnostic object boundary branch. Extensive experimental results on two challenging datasets and representative semantic segmentation baselines (including CNNs and Transformers) demonstrate that TFISeg can achieve competitive results compared to the state-of-the-art fully-supervised instance segmentation methods without the need for additional human resources or increased computational costs. The code is available at: TFISeg ### MDT3D: Multi-Dataset Training for LiDAR 3D Object Detection Generalization - **Authors:** Louis Soum-Fontez, Jean-Emmanuel Deschaud, François Goulette - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG) - **Arxiv link:** https://arxiv.org/abs/2308.01000 - **Pdf link:** https://arxiv.org/pdf/2308.01000 - **Abstract** Supervised 3D Object Detection models have been displaying increasingly better performance in single-domain cases where the training data comes from the same environment and sensor as the testing data. However, in real-world scenarios data from the target domain may not be available for finetuning or for domain adaptation methods. Indeed, 3D object detection models trained on a source dataset with a specific point distribution have shown difficulties in generalizing to unseen datasets. Therefore, we decided to leverage the information available from several annotated source datasets with our Multi-Dataset Training for 3D Object Detection (MDT3D) method to increase the robustness of 3D object detection models when tested in a new environment with a different sensor configuration. To tackle the labelling gap between datasets, we used a new label mapping based on coarse labels. Furthermore, we show how we managed the mix of datasets during training and finally introduce a new cross-dataset augmentation method: cross-dataset object injection. We demonstrate that this training paradigm shows improvements for different types of 3D object detection models. The source code and additional results for this research project will be publicly available on GitHub for interested parties to access and utilize: https://github.com/LouisSF/MDT3D ### WCCNet: Wavelet-integrated CNN with Crossmodal Rearranging Fusion for Fast Multispectral Pedestrian Detection - **Authors:** Xingjian Wang, Li Chai, Jiming Chen, Zhiguo Shi - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2308.01042 - **Pdf link:** https://arxiv.org/pdf/2308.01042 - **Abstract** Multispectral pedestrian detection achieves better visibility in challenging conditions and thus has a broad application in various tasks, for which both the accuracy and computational cost are of paramount importance. Most existing approaches treat RGB and infrared modalities equally, typically adopting two symmetrical CNN backbones for multimodal feature extraction, which ignores the substantial differences between modalities and brings great difficulty for the reduction of the computational cost as well as effective crossmodal fusion. In this work, we propose a novel and efficient framework named WCCNet that is able to differentially extract rich features of different spectra with lower computational complexity and semantically rearranges these features for effective crossmodal fusion. Specifically, the discrete wavelet transform (DWT) allowing fast inference and training speed is embedded to construct a dual-stream backbone for efficient feature extraction. The DWT layers of WCCNet extract frequency components for infrared modality, while the CNN layers extract spatial-domain features for RGB modality. This methodology not only significantly reduces the computational complexity, but also improves the extraction of infrared features to facilitate the subsequent crossmodal fusion. Based on the well extracted features, we elaborately design the crossmodal rearranging fusion module (CMRF), which can mitigate spatial misalignment and merge semantically complementary features of spatially-related local regions to amplify the crossmodal complementary information. We conduct comprehensive evaluations on KAIST and FLIR benchmarks, in which WCCNet outperforms state-of-the-art methods with considerable computational efficiency and competitive accuracy. We also perform the ablation study and analyze thoroughly the impact of different components on the performance of WCCNet. ### Interpretable End-to-End Driving Model for Implicit Scene Understanding - **Authors:** Yiyang Sun, Xiaonian Wang, Yangyang Zhang, Jiagui Tang, Xiaqiang Tang, Jing Yao - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Robotics (cs.RO) - **Arxiv link:** https://arxiv.org/abs/2308.01180 - **Pdf link:** https://arxiv.org/pdf/2308.01180 - **Abstract** Driving scene understanding is to obtain comprehensive scene information through the sensor data and provide a basis for downstream tasks, which is indispensable for the safety of self-driving vehicles. Specific perception tasks, such as object detection and scene graph generation, are commonly used. However, the results of these tasks are only equivalent to the characterization of sampling from high-dimensional scene features, which are not sufficient to represent the scenario. In addition, the goal of perception tasks is inconsistent with human driving that just focuses on what may affect the ego-trajectory. Therefore, we propose an end-to-end Interpretable Implicit Driving Scene Understanding (II-DSU) model to extract implicit high-dimensional scene features as scene understanding results guided by a planning module and to validate the plausibility of scene understanding using auxiliary perception tasks for visualization. Experimental results on CARLA benchmarks show that our approach achieves the new state-of-the-art and is able to obtain scene features that embody richer scene information relevant to driving, enabling superior performance of the downstream planning. ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression There is no result ## Keyword: RAW ### More Context, Less Distraction: Visual Classification by Inferring and Conditioning on Contextual Attributes - **Authors:** Bang An, Sicheng Zhu, Michael-Andrei Panaitescu-Liess, Chaithanya Kumar Mummadi, Furong Huang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Computation and Language (cs.CL); Machine Learning (cs.LG) - **Arxiv link:** https://arxiv.org/abs/2308.01313 - **Pdf link:** https://arxiv.org/pdf/2308.01313 - **Abstract** CLIP, as a foundational vision language model, is widely used in zero-shot image classification due to its ability to understand various visual concepts and natural language descriptions. However, how to fully leverage CLIP's unprecedented human-like understanding capabilities to achieve better zero-shot classification is still an open question. This paper draws inspiration from the human visual perception process: a modern neuroscience view suggests that in classifying an object, humans first infer its class-independent attributes (e.g., background and orientation) which help separate the foreground object from the background, and then make decisions based on this information. Inspired by this, we observe that providing CLIP with contextual attributes improves zero-shot classification and mitigates reliance on spurious features. We also observe that CLIP itself can reasonably infer the attributes from an image. With these observations, we propose a training-free, two-step zero-shot classification method named PerceptionCLIP. Given an image, it first infers contextual attributes (e.g., background) and then performs object classification conditioning on them. Our experiments show that PerceptionCLIP achieves better generalization, group robustness, and better interpretability. For example, PerceptionCLIP with ViT-L/14 improves the worst group accuracy by 16.5% on the Waterbirds dataset and by 3.5% on CelebA. ## Keyword: raw image There is no result
process
new submissions for thu aug keyword events there is no result keyword event camera there is no result keyword events camera there is no result keyword white balance there is no result keyword color contrast there is no result keyword awb there is no result keyword isp elfnet evidential local global fusion for stereo matching authors jieming lou weide liu zhuo chen fayao liu jun cheng subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract although existing stereo matching models have achieved continuous improvement they often face issues related to trustworthiness due to the absence of uncertainty estimation additionally effectively leveraging multi scale and multi view knowledge of stereo pairs remains unexplored in this paper we introduce the textbf e vidential textbf l ocal global textbf f usion elf framework for stereo matching which endows both uncertainty estimation and confidence aware fusion with trustworthy heads instead of predicting the disparity map alone our model estimates an evidential based disparity considering both aleatoric and epistemic uncertainties with the normal inverse gamma distribution as a bridge the proposed framework realizes intra evidential fusion of multi level predictions and inter evidential fusion between cost volume based and transformer based stereo matching extensive experimental results show that the proposed framework exploits multi view information effectively and achieves state of the art overall performance both on accuracy and cross domain generalization the codes are available at training free instance segmentation from semantic image segmentation masks authors yuchen shen dong zhang yuhui zheng zechao li liyong fu qiaolin ye subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract in recent years the development of instance segmentation has garnered significant attention in a wide range of applications however the training of a fully supervised instance segmentation model requires costly both instance level and pixel level annotations in contrast weakly supervised instance segmentation methods i e with image level class labels or point labels struggle to satisfy the accuracy and recall requirements of practical scenarios in this paper we propose a novel paradigm for instance segmentation called training free instance segmentation tfiseg which achieves instance segmentation results from image masks predicted using off the shelf semantic segmentation models tfiseg does not require training a semantic or and instance segmentation model and avoids the need for instance level image annotations therefore it is highly efficient specifically we first obtain a semantic segmentation mask of the input image via a trained semantic segmentation model then we calculate a displacement field vector for each pixel based on the segmentation mask which can indicate representations belonging to the same class but different instances i e obtaining the instance level object information finally instance segmentation results are obtained after being refined by a learnable category agnostic object boundary branch extensive experimental results on two challenging datasets and representative semantic segmentation baselines including cnns and transformers demonstrate that tfiseg can achieve competitive results compared to the state of the art fully supervised instance segmentation methods without the need for additional human resources or increased computational costs the code is available at tfiseg multi dataset training for lidar object detection generalization authors louis soum fontez jean emmanuel deschaud françois goulette subjects computer vision and pattern recognition cs cv machine learning cs lg arxiv link pdf link abstract supervised object detection models have been displaying increasingly better performance in single domain cases where the training data comes from the same environment and sensor as the testing data however in real world scenarios data from the target domain may not be available for finetuning or for domain adaptation methods indeed object detection models trained on a source dataset with a specific point distribution have shown difficulties in generalizing to unseen datasets therefore we decided to leverage the information available from several annotated source datasets with our multi dataset training for object detection method to increase the robustness of object detection models when tested in a new environment with a different sensor configuration to tackle the labelling gap between datasets we used a new label mapping based on coarse labels furthermore we show how we managed the mix of datasets during training and finally introduce a new cross dataset augmentation method cross dataset object injection we demonstrate that this training paradigm shows improvements for different types of object detection models the source code and additional results for this research project will be publicly available on github for interested parties to access and utilize wccnet wavelet integrated cnn with crossmodal rearranging fusion for fast multispectral pedestrian detection authors xingjian wang li chai jiming chen zhiguo shi subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract multispectral pedestrian detection achieves better visibility in challenging conditions and thus has a broad application in various tasks for which both the accuracy and computational cost are of paramount importance most existing approaches treat rgb and infrared modalities equally typically adopting two symmetrical cnn backbones for multimodal feature extraction which ignores the substantial differences between modalities and brings great difficulty for the reduction of the computational cost as well as effective crossmodal fusion in this work we propose a novel and efficient framework named wccnet that is able to differentially extract rich features of different spectra with lower computational complexity and semantically rearranges these features for effective crossmodal fusion specifically the discrete wavelet transform dwt allowing fast inference and training speed is embedded to construct a dual stream backbone for efficient feature extraction the dwt layers of wccnet extract frequency components for infrared modality while the cnn layers extract spatial domain features for rgb modality this methodology not only significantly reduces the computational complexity but also improves the extraction of infrared features to facilitate the subsequent crossmodal fusion based on the well extracted features we elaborately design the crossmodal rearranging fusion module cmrf which can mitigate spatial misalignment and merge semantically complementary features of spatially related local regions to amplify the crossmodal complementary information we conduct comprehensive evaluations on kaist and flir benchmarks in which wccnet outperforms state of the art methods with considerable computational efficiency and competitive accuracy we also perform the ablation study and analyze thoroughly the impact of different components on the performance of wccnet interpretable end to end driving model for implicit scene understanding authors yiyang sun xiaonian wang yangyang zhang jiagui tang xiaqiang tang jing yao subjects computer vision and pattern recognition cs cv robotics cs ro arxiv link pdf link abstract driving scene understanding is to obtain comprehensive scene information through the sensor data and provide a basis for downstream tasks which is indispensable for the safety of self driving vehicles specific perception tasks such as object detection and scene graph generation are commonly used however the results of these tasks are only equivalent to the characterization of sampling from high dimensional scene features which are not sufficient to represent the scenario in addition the goal of perception tasks is inconsistent with human driving that just focuses on what may affect the ego trajectory therefore we propose an end to end interpretable implicit driving scene understanding ii dsu model to extract implicit high dimensional scene features as scene understanding results guided by a planning module and to validate the plausibility of scene understanding using auxiliary perception tasks for visualization experimental results on carla benchmarks show that our approach achieves the new state of the art and is able to obtain scene features that embody richer scene information relevant to driving enabling superior performance of the downstream planning keyword image signal processing there is no result keyword image signal process there is no result keyword compression there is no result keyword raw more context less distraction visual classification by inferring and conditioning on contextual attributes authors bang an sicheng zhu michael andrei panaitescu liess chaithanya kumar mummadi furong huang subjects computer vision and pattern recognition cs cv artificial intelligence cs ai computation and language cs cl machine learning cs lg arxiv link pdf link abstract clip as a foundational vision language model is widely used in zero shot image classification due to its ability to understand various visual concepts and natural language descriptions however how to fully leverage clip s unprecedented human like understanding capabilities to achieve better zero shot classification is still an open question this paper draws inspiration from the human visual perception process a modern neuroscience view suggests that in classifying an object humans first infer its class independent attributes e g background and orientation which help separate the foreground object from the background and then make decisions based on this information inspired by this we observe that providing clip with contextual attributes improves zero shot classification and mitigates reliance on spurious features we also observe that clip itself can reasonably infer the attributes from an image with these observations we propose a training free two step zero shot classification method named perceptionclip given an image it first infers contextual attributes e g background and then performs object classification conditioning on them our experiments show that perceptionclip achieves better generalization group robustness and better interpretability for example perceptionclip with vit l improves the worst group accuracy by on the waterbirds dataset and by on celeba keyword raw image there is no result
1
13,956
16,738,157,852
IssuesEvent
2021-06-11 06:19:09
googleapis/repo-automation-bots
https://api.github.com/repos/googleapis/repo-automation-bots
closed
A canary is chirping
type: process
The dependencies and their versions are: {"dayjs":"^1.10.5","gcf-utils":"^8.0.2"} at 2021 06-10 15:15:11
1.0
A canary is chirping - The dependencies and their versions are: {"dayjs":"^1.10.5","gcf-utils":"^8.0.2"} at 2021 06-10 15:15:11
process
a canary is chirping the dependencies and their versions are dayjs gcf utils at
1
252,955
8,049,104,469
IssuesEvent
2018-08-01 09:04:26
biolab/orange3-bioinformatics
https://api.github.com/repos/biolab/orange3-bioinformatics
closed
Docs: bioinformatics modules
priority-medium task
to-do: - [X] **gene** - [X] **taxonomy** - [X] **resolwe** - [X] **geo** - [X] **go** - [X] **geo** - [X] **geneset** - [X] **dicty** - [X] **kegg**
1.0
Docs: bioinformatics modules - to-do: - [X] **gene** - [X] **taxonomy** - [X] **resolwe** - [X] **geo** - [X] **go** - [X] **geo** - [X] **geneset** - [X] **dicty** - [X] **kegg**
non_process
docs bioinformatics modules to do gene taxonomy resolwe geo go geo geneset dicty kegg
0
2,835
5,791,680,335
IssuesEvent
2017-05-02 07:01:25
gaocegege/maintainer
https://api.github.com/repos/gaocegege/maintainer
closed
Support more languages in contributing subcommand
priority/P2 process/wip type/enhancement
https://github.com/gaocegege/maintainer/blob/aa310a8938e6f767fdf10aee0d1e09c547949518/contributing/contributing.go#L117 Now contributing only supports golang. But it is easy to extend.
1.0
Support more languages in contributing subcommand - https://github.com/gaocegege/maintainer/blob/aa310a8938e6f767fdf10aee0d1e09c547949518/contributing/contributing.go#L117 Now contributing only supports golang. But it is easy to extend.
process
support more languages in contributing subcommand now contributing only supports golang but it is easy to extend
1
625,889
19,769,416,858
IssuesEvent
2022-01-17 08:26:48
aiidateam/aiida-core
https://api.github.com/repos/aiidateam/aiida-core
closed
Small changes to `List` and `Dict` data types API
priority/critical-blocking topic/data-types type/feature request
I'd like to discuss two small changes to the API of the `List` and `Dict` classes: ### Initialization #### Is your feature request related to a problem? Please describe When initializing a `List` or `Dict` node, you have to pass `dict` or `list` as a keyword argument: ```Python l = List(list=[1, 2, 3]) ``` If you try to instead just pass the list as a positional argument, you are faced with an error that doesn't clarify proper usage: ```Python l = List([1, 2, 3]) --------------------------------------------------------------------------- TypeError Traceback (most recent call last) <ipython-input-47-398676db57d5> in <module> ----> 1 List([1, 2, 3]) TypeError: __init__() takes 1 positional argument but 2 were given ``` and the docstring doesn't explain that you have to use the `list` keyword argument either: ```Python In [46]: List? Init signature: List(**kwargs) Docstring: `Data` sub class to represent a list. Init docstring: :param backend_entity: the backend model supporting this entity :type backend_entity: :class:`aiida.orm.implementation.entities.BackendEntity` File: ~/envs/aiida-docs/code/aiida-core/aiida/orm/nodes/data/list.py Type: AbstractNodeMeta Subclasses: ``` #### Describe the solution you'd like I think it would be more intuitive if the user could just pass the list/dict as the first positional argument. For example, we could change the current constructor: ```Python def __init__(self, **kwargs): data = kwargs.pop('list', list()) super().__init__(**kwargs) self.set_list(data) ``` to something like (thanks to @giovannipizzi for the suggestion to import the `builtins` module to avoid conflicts between the `list` argument and builtin, see [this discussion](https://github.com/aiidateam/aiida-core/pull/4469#discussion_r508263946)): ```Python def __init__(self, list=None, **kwargs): import builtins list = list or builtins.list() super().__init__(**kwargs) self.set_list(input_list) ``` Similarly for `Dict`. If for some reason we cannot make this change, we should at least add a docstring that clearly explains the API to the user. ### Using `.value` #### Is your feature request related to a problem? Please describe For mose data types in AiiDA derived from builtin Python types, you can use the `value` property to get the corresponding Python builtin. For `List` and `Dict` nodes, however, you have to use the `get_list()` and `get_dict()` methods. Adding the `value` property to these classes would make the API between the "builtin" AiiDA data types more consistent. #### Describe the solution you'd like Add the `value` property to both the `List` and `Dict` classes.
1.0
Small changes to `List` and `Dict` data types API - I'd like to discuss two small changes to the API of the `List` and `Dict` classes: ### Initialization #### Is your feature request related to a problem? Please describe When initializing a `List` or `Dict` node, you have to pass `dict` or `list` as a keyword argument: ```Python l = List(list=[1, 2, 3]) ``` If you try to instead just pass the list as a positional argument, you are faced with an error that doesn't clarify proper usage: ```Python l = List([1, 2, 3]) --------------------------------------------------------------------------- TypeError Traceback (most recent call last) <ipython-input-47-398676db57d5> in <module> ----> 1 List([1, 2, 3]) TypeError: __init__() takes 1 positional argument but 2 were given ``` and the docstring doesn't explain that you have to use the `list` keyword argument either: ```Python In [46]: List? Init signature: List(**kwargs) Docstring: `Data` sub class to represent a list. Init docstring: :param backend_entity: the backend model supporting this entity :type backend_entity: :class:`aiida.orm.implementation.entities.BackendEntity` File: ~/envs/aiida-docs/code/aiida-core/aiida/orm/nodes/data/list.py Type: AbstractNodeMeta Subclasses: ``` #### Describe the solution you'd like I think it would be more intuitive if the user could just pass the list/dict as the first positional argument. For example, we could change the current constructor: ```Python def __init__(self, **kwargs): data = kwargs.pop('list', list()) super().__init__(**kwargs) self.set_list(data) ``` to something like (thanks to @giovannipizzi for the suggestion to import the `builtins` module to avoid conflicts between the `list` argument and builtin, see [this discussion](https://github.com/aiidateam/aiida-core/pull/4469#discussion_r508263946)): ```Python def __init__(self, list=None, **kwargs): import builtins list = list or builtins.list() super().__init__(**kwargs) self.set_list(input_list) ``` Similarly for `Dict`. If for some reason we cannot make this change, we should at least add a docstring that clearly explains the API to the user. ### Using `.value` #### Is your feature request related to a problem? Please describe For mose data types in AiiDA derived from builtin Python types, you can use the `value` property to get the corresponding Python builtin. For `List` and `Dict` nodes, however, you have to use the `get_list()` and `get_dict()` methods. Adding the `value` property to these classes would make the API between the "builtin" AiiDA data types more consistent. #### Describe the solution you'd like Add the `value` property to both the `List` and `Dict` classes.
non_process
small changes to list and dict data types api i d like to discuss two small changes to the api of the list and dict classes initialization is your feature request related to a problem please describe when initializing a list or dict node you have to pass dict or list as a keyword argument python l list list if you try to instead just pass the list as a positional argument you are faced with an error that doesn t clarify proper usage python l list typeerror traceback most recent call last in list typeerror init takes positional argument but were given and the docstring doesn t explain that you have to use the list keyword argument either python in list init signature list kwargs docstring data sub class to represent a list init docstring param backend entity the backend model supporting this entity type backend entity class aiida orm implementation entities backendentity file envs aiida docs code aiida core aiida orm nodes data list py type abstractnodemeta subclasses describe the solution you d like i think it would be more intuitive if the user could just pass the list dict as the first positional argument for example we could change the current constructor python def init self kwargs data kwargs pop list list super init kwargs self set list data to something like thanks to giovannipizzi for the suggestion to import the builtins module to avoid conflicts between the list argument and builtin see python def init self list none kwargs import builtins list list or builtins list super init kwargs self set list input list similarly for dict if for some reason we cannot make this change we should at least add a docstring that clearly explains the api to the user using value is your feature request related to a problem please describe for mose data types in aiida derived from builtin python types you can use the value property to get the corresponding python builtin for list and dict nodes however you have to use the get list and get dict methods adding the value property to these classes would make the api between the builtin aiida data types more consistent describe the solution you d like add the value property to both the list and dict classes
0
11,392
14,228,767,222
IssuesEvent
2020-11-18 04:39:36
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Tessellation crushes with a triangle
Bug High Priority Processing Regression
**Describe the bug** tessellation crushes QGIS when one of the target features is a triangle. **How to Reproduce** 1. Create a new Project and create a shape file with a single feature of triangle 2. Apply tessellation 3. QGIS crushes **QGIS and OS versions** QGISバージョン | 3.16.0-Hannover | QGISコードリビジョン | 4af1cbfb97 -- | -- | -- | -- コンパイル時のQtバージョン | 5.14.2 | 実行中のQtバージョン | 5.14.2 コンパイル時のGDAL/OGRバージョン | 3.1.2 | 実行中のGDAL/OGRバージョン | 3.1.2 コンパイル時のGEOSバージョン | 3.8.1-CAPI-1.13.3 | 実行中のGEOSバージョン | 3.8.1-CAPI-1.13.3 コンパイル時のSQLiteバージョン | 3.31.1 | 実行中のSQLiteバージョン | 3.31.1 PostgreSQLクライアントバージョン | 12.3 | SpatiaLiteバージョン | 4.3.0a QWTバージョン | 6.1.4 | QScintilla2 バージョン | 2.11.4 コンパイル時のPROJバージョン | 6.3.2 | 実行中のPROJバージョン | Rel. 6.3.2, May 1st, 2020 OSのバージョン | macOS 10.15 Active pythonプラグイン | QNEAT3; contour; GroupStats; realcentroid; processing; db_manager; MetaSearch The Linux version also crushes. **Additional context** Square and Polygon with more vertices works fine.
1.0
Tessellation crushes with a triangle - **Describe the bug** tessellation crushes QGIS when one of the target features is a triangle. **How to Reproduce** 1. Create a new Project and create a shape file with a single feature of triangle 2. Apply tessellation 3. QGIS crushes **QGIS and OS versions** QGISバージョン | 3.16.0-Hannover | QGISコードリビジョン | 4af1cbfb97 -- | -- | -- | -- コンパイル時のQtバージョン | 5.14.2 | 実行中のQtバージョン | 5.14.2 コンパイル時のGDAL/OGRバージョン | 3.1.2 | 実行中のGDAL/OGRバージョン | 3.1.2 コンパイル時のGEOSバージョン | 3.8.1-CAPI-1.13.3 | 実行中のGEOSバージョン | 3.8.1-CAPI-1.13.3 コンパイル時のSQLiteバージョン | 3.31.1 | 実行中のSQLiteバージョン | 3.31.1 PostgreSQLクライアントバージョン | 12.3 | SpatiaLiteバージョン | 4.3.0a QWTバージョン | 6.1.4 | QScintilla2 バージョン | 2.11.4 コンパイル時のPROJバージョン | 6.3.2 | 実行中のPROJバージョン | Rel. 6.3.2, May 1st, 2020 OSのバージョン | macOS 10.15 Active pythonプラグイン | QNEAT3; contour; GroupStats; realcentroid; processing; db_manager; MetaSearch The Linux version also crushes. **Additional context** Square and Polygon with more vertices works fine.
process
tessellation crushes with a triangle describe the bug tessellation crushes qgis when one of the target features is a triangle how to reproduce create a new project and create a shape file with a single feature of triangle apply tessellation qgis crushes qgis and os versions qgisバージョン hannover qgisコードリビジョン コンパイル時のqtバージョン 実行中のqtバージョン コンパイル時のgdal ogrバージョン 実行中のgdal ogrバージョン コンパイル時のgeosバージョン capi 実行中のgeosバージョン capi コンパイル時のsqliteバージョン 実行中のsqliteバージョン postgresqlクライアントバージョン spatialiteバージョン qwtバージョン バージョン コンパイル時のprojバージョン 実行中のprojバージョン rel may osのバージョン macos active pythonプラグイン contour groupstats realcentroid processing db manager metasearch the linux version also crushes additional context square and polygon with more vertices works fine
1
22,546
31,722,316,345
IssuesEvent
2023-09-10 14:52:01
TableRise/tablerise-backend
https://api.github.com/repos/TableRise/tablerise-backend
closed
Cadastro Externo - Recuperar infos do usuário ( Google )
test-process
**Configurar a autorização OAuth2 com a API do Google no backend** Para tornar possível que o google seja utilizado como forma de cadastro é necessário recuperar informações do usuário para que futuramente utilizemos o passport para autenticação. Use este video como referência: https://www.youtube.com/watch?v=Q0a0594tOrc&t=10s Fluxo de cadastro com serviços externos ( google, facebook, discord e etc.. ) ![image](https://github.com/TableRise/tablerise-backend/assets/87550173/18dc3f8c-df79-4c3e-9bc3-0614f60754f5) - Campos não fornecidos pelo serviço externo devem passar como null; - O fluxo não será completado nesta task, em uma task futura as informações obtidas do google serão utilizadas para montar a entidade de usuário a ser salva no banco; - A informações devem ser passadas para um método helper que irá formatar os campos do objeto recuperado da API para um formato em comum entre todos os serviços. - Precisamos do email do usuário, do nome completo e da data de nascimento. - [ ] Desenvolver lógica e autorização para recuperar dados do usuário com a Google API
1.0
Cadastro Externo - Recuperar infos do usuário ( Google ) - **Configurar a autorização OAuth2 com a API do Google no backend** Para tornar possível que o google seja utilizado como forma de cadastro é necessário recuperar informações do usuário para que futuramente utilizemos o passport para autenticação. Use este video como referência: https://www.youtube.com/watch?v=Q0a0594tOrc&t=10s Fluxo de cadastro com serviços externos ( google, facebook, discord e etc.. ) ![image](https://github.com/TableRise/tablerise-backend/assets/87550173/18dc3f8c-df79-4c3e-9bc3-0614f60754f5) - Campos não fornecidos pelo serviço externo devem passar como null; - O fluxo não será completado nesta task, em uma task futura as informações obtidas do google serão utilizadas para montar a entidade de usuário a ser salva no banco; - A informações devem ser passadas para um método helper que irá formatar os campos do objeto recuperado da API para um formato em comum entre todos os serviços. - Precisamos do email do usuário, do nome completo e da data de nascimento. - [ ] Desenvolver lógica e autorização para recuperar dados do usuário com a Google API
process
cadastro externo recuperar infos do usuário google configurar a autorização com a api do google no backend para tornar possível que o google seja utilizado como forma de cadastro é necessário recuperar informações do usuário para que futuramente utilizemos o passport para autenticação use este video como referência fluxo de cadastro com serviços externos google facebook discord e etc campos não fornecidos pelo serviço externo devem passar como null o fluxo não será completado nesta task em uma task futura as informações obtidas do google serão utilizadas para montar a entidade de usuário a ser salva no banco a informações devem ser passadas para um método helper que irá formatar os campos do objeto recuperado da api para um formato em comum entre todos os serviços precisamos do email do usuário do nome completo e da data de nascimento desenvolver lógica e autorização para recuperar dados do usuário com a google api
1
276,956
8,614,853,537
IssuesEvent
2018-11-19 18:45:04
WordPress/gutenberg
https://api.github.com/repos/WordPress/gutenberg
closed
Media Library doesn't update after dragging image onto Image block
Drag and Drop Media Priority High [Status] In Progress [Type] Bug
## Issue Overview After inserting an Image block, I then dragged 'n dropped an image directly onto the block. I then inserted another image block and after clicking the _Add from Media Library_ button, and displaying the Media popup window, the previous image that I added to the page isn't displaying in the Media Library. New image that I dragged onto Image Block ![gutenberg-imgblockdragdrop](https://user-images.githubusercontent.com/2078245/35179551-875a8a7a-fdf0-11e7-854a-db49b7f40927.png) Newly added image not appearing in Media LIbrary ![gutenberg-imgnotinlibrary](https://user-images.githubusercontent.com/2078245/35179553-92390d22-fdf0-11e7-8329-7d32937daa1d.png) ## Steps to Reproduce (for bugs) 1. Insert Image block 2. Drag 'n drop image onto Image block 3. Insert new Image block 4. Click the _Add from Media Library_ button and try to add the same image. You can't, because it doesn't appear in the Media Library ## Expected Behavior If I drag'n drop an image onto an Image BLock, this should get added to my Media LIbraray, and should be displayed when I try to add another image by viewing the Media Library popup ## Current Behavior Images dragged onto Image Block aren't reflected in Media Library popup window _Firefox Quantum 57.0.4 (64-bit) Gutenberg 2.0.0 WordPress 4.9.2 No other plugins activated_
1.0
Media Library doesn't update after dragging image onto Image block - ## Issue Overview After inserting an Image block, I then dragged 'n dropped an image directly onto the block. I then inserted another image block and after clicking the _Add from Media Library_ button, and displaying the Media popup window, the previous image that I added to the page isn't displaying in the Media Library. New image that I dragged onto Image Block ![gutenberg-imgblockdragdrop](https://user-images.githubusercontent.com/2078245/35179551-875a8a7a-fdf0-11e7-854a-db49b7f40927.png) Newly added image not appearing in Media LIbrary ![gutenberg-imgnotinlibrary](https://user-images.githubusercontent.com/2078245/35179553-92390d22-fdf0-11e7-8329-7d32937daa1d.png) ## Steps to Reproduce (for bugs) 1. Insert Image block 2. Drag 'n drop image onto Image block 3. Insert new Image block 4. Click the _Add from Media Library_ button and try to add the same image. You can't, because it doesn't appear in the Media Library ## Expected Behavior If I drag'n drop an image onto an Image BLock, this should get added to my Media LIbraray, and should be displayed when I try to add another image by viewing the Media Library popup ## Current Behavior Images dragged onto Image Block aren't reflected in Media Library popup window _Firefox Quantum 57.0.4 (64-bit) Gutenberg 2.0.0 WordPress 4.9.2 No other plugins activated_
non_process
media library doesn t update after dragging image onto image block issue overview after inserting an image block i then dragged n dropped an image directly onto the block i then inserted another image block and after clicking the add from media library button and displaying the media popup window the previous image that i added to the page isn t displaying in the media library new image that i dragged onto image block newly added image not appearing in media library steps to reproduce for bugs insert image block drag n drop image onto image block insert new image block click the add from media library button and try to add the same image you can t because it doesn t appear in the media library expected behavior if i drag n drop an image onto an image block this should get added to my media libraray and should be displayed when i try to add another image by viewing the media library popup current behavior images dragged onto image block aren t reflected in media library popup window firefox quantum bit gutenberg wordpress no other plugins activated
0
7,702
10,790,396,590
IssuesEvent
2019-11-05 14:48:07
Addalin/cameranetwork
https://api.github.com/repos/Addalin/cameranetwork
closed
GUI - seek problem 2
Priority: High Type: Bug help wanted inverse pre-processing
**Describe the bug** seek button trouble relates to #28 **To Reproduce** Steps to reproduce the behavior: 1. seek tab 2. click on : update dates --> chose data --> fore read (True) --> Query --> chose time to show (pick HDR) --> push seek 2. Click on '....' 5. See error ``` Server cam_116L raised an error =============================== Calling the cmd handler caused an error: Traceback (most recent call last): File "C:\Users\addalin\Dropbox\cloud CT\cameranetwork\CameraNetwork\server.py", line 414, in _on_request answer = yield cb(*args, **kwds) File "C:\ProgramData\Anaconda2\envs\cn_client\lib\site-packages\tornado\gen.py", line 1055, in run value = future.result() File "C:\ProgramData\Anaconda2\envs\cn_client\lib\site-packages\tornado\concurrent.py", line 238, in result raise_exc_info(self._exc_info) File "C:\ProgramData\Anaconda2\envs\cn_client\lib\site-packages\tornado\gen.py", line 292, in wrapper result = func(*args, **kwargs) File "C:\Users\addalin\Dropbox\cloud CT\cameranetwork\CameraNetwork\server.py", line 992, in handle_seek timedelta_threshold=timedelta_threshold File "C:\Users\addalin\Dropbox\cloud CT\cameranetwork\CameraNetwork\controller.py", line 1381, in seekImageArray ignore_date_extrinsic File "C:\Users\addalin\Dropbox\cloud CT\cameranetwork\CameraNetwork\controller.py", line 1490, in preprocess_array img_array = self._vignetting.applyVignetting(img_array) AttributeError: 'Controller' object has no attribute '_vignetting' ``` **Expected behavior** Show images of the expected time **Screenshots** ![image](https://user-images.githubusercontent.com/9436237/66269662-a0d8c080-e853-11e9-8cc5-1160f3d61d28.png)
1.0
GUI - seek problem 2 - **Describe the bug** seek button trouble relates to #28 **To Reproduce** Steps to reproduce the behavior: 1. seek tab 2. click on : update dates --> chose data --> fore read (True) --> Query --> chose time to show (pick HDR) --> push seek 2. Click on '....' 5. See error ``` Server cam_116L raised an error =============================== Calling the cmd handler caused an error: Traceback (most recent call last): File "C:\Users\addalin\Dropbox\cloud CT\cameranetwork\CameraNetwork\server.py", line 414, in _on_request answer = yield cb(*args, **kwds) File "C:\ProgramData\Anaconda2\envs\cn_client\lib\site-packages\tornado\gen.py", line 1055, in run value = future.result() File "C:\ProgramData\Anaconda2\envs\cn_client\lib\site-packages\tornado\concurrent.py", line 238, in result raise_exc_info(self._exc_info) File "C:\ProgramData\Anaconda2\envs\cn_client\lib\site-packages\tornado\gen.py", line 292, in wrapper result = func(*args, **kwargs) File "C:\Users\addalin\Dropbox\cloud CT\cameranetwork\CameraNetwork\server.py", line 992, in handle_seek timedelta_threshold=timedelta_threshold File "C:\Users\addalin\Dropbox\cloud CT\cameranetwork\CameraNetwork\controller.py", line 1381, in seekImageArray ignore_date_extrinsic File "C:\Users\addalin\Dropbox\cloud CT\cameranetwork\CameraNetwork\controller.py", line 1490, in preprocess_array img_array = self._vignetting.applyVignetting(img_array) AttributeError: 'Controller' object has no attribute '_vignetting' ``` **Expected behavior** Show images of the expected time **Screenshots** ![image](https://user-images.githubusercontent.com/9436237/66269662-a0d8c080-e853-11e9-8cc5-1160f3d61d28.png)
process
gui seek problem describe the bug seek button trouble relates to to reproduce steps to reproduce the behavior seek tab click on update dates chose data fore read true query chose time to show pick hdr push seek click on see error server cam raised an error calling the cmd handler caused an error traceback most recent call last file c users addalin dropbox cloud ct cameranetwork cameranetwork server py line in on request answer yield cb args kwds file c programdata envs cn client lib site packages tornado gen py line in run value future result file c programdata envs cn client lib site packages tornado concurrent py line in result raise exc info self exc info file c programdata envs cn client lib site packages tornado gen py line in wrapper result func args kwargs file c users addalin dropbox cloud ct cameranetwork cameranetwork server py line in handle seek timedelta threshold timedelta threshold file c users addalin dropbox cloud ct cameranetwork cameranetwork controller py line in seekimagearray ignore date extrinsic file c users addalin dropbox cloud ct cameranetwork cameranetwork controller py line in preprocess array img array self vignetting applyvignetting img array attributeerror controller object has no attribute vignetting expected behavior show images of the expected time screenshots
1
8,388
11,562,202,464
IssuesEvent
2020-02-20 01:40:29
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
Process.Modules doesn't return the full module list of a Windows process when the architecture of the Process and the program don't match
area-System.Diagnostics.Process untriaged
Currently accessing the `Modules` property of a `Process` class doesn't return the full list of modules when the architecture of the `Process` doesn't match that of the compiled architecture of the program. As an example, if a WOW64 process is running and you create a program that is compiled under x64 and create a Process object around the WOW64 process, when you access `Process.Modules` you will only be able to see the x64 modules i.e. the x64 version of ntdll and the accompanying wow64 DLLs. If I remember correctly, the current runtime uses `EnumProcessModules` to resolve a module list of a remote process. This API call unfortunately doesn't parse both PEB's of a WOW64 process for a complete module list, hence why the module list is incomplete in the `Process` class. **The solution that I propose** Use `CreateToolhelp32Snapshot` with the `TH32CS_SNAPMODULE` and `TH32CS_SNAPMODULE32` flags accompanied by `Module32First` and `Module32Next` This will always get a full module list for the process regardless of if the architecture of the process and program match as opposed to `EnumProcessModules` and also doesn't require any manual parsing of the PEB(s)
1.0
Process.Modules doesn't return the full module list of a Windows process when the architecture of the Process and the program don't match - Currently accessing the `Modules` property of a `Process` class doesn't return the full list of modules when the architecture of the `Process` doesn't match that of the compiled architecture of the program. As an example, if a WOW64 process is running and you create a program that is compiled under x64 and create a Process object around the WOW64 process, when you access `Process.Modules` you will only be able to see the x64 modules i.e. the x64 version of ntdll and the accompanying wow64 DLLs. If I remember correctly, the current runtime uses `EnumProcessModules` to resolve a module list of a remote process. This API call unfortunately doesn't parse both PEB's of a WOW64 process for a complete module list, hence why the module list is incomplete in the `Process` class. **The solution that I propose** Use `CreateToolhelp32Snapshot` with the `TH32CS_SNAPMODULE` and `TH32CS_SNAPMODULE32` flags accompanied by `Module32First` and `Module32Next` This will always get a full module list for the process regardless of if the architecture of the process and program match as opposed to `EnumProcessModules` and also doesn't require any manual parsing of the PEB(s)
process
process modules doesn t return the full module list of a windows process when the architecture of the process and the program don t match currently accessing the modules property of a process class doesn t return the full list of modules when the architecture of the process doesn t match that of the compiled architecture of the program as an example if a process is running and you create a program that is compiled under and create a process object around the process when you access process modules you will only be able to see the modules i e the version of ntdll and the accompanying dlls if i remember correctly the current runtime uses enumprocessmodules to resolve a module list of a remote process this api call unfortunately doesn t parse both peb s of a process for a complete module list hence why the module list is incomplete in the process class the solution that i propose use with the snapmodule and flags accompanied by and this will always get a full module list for the process regardless of if the architecture of the process and program match as opposed to enumprocessmodules and also doesn t require any manual parsing of the peb s
1
9,828
12,827,744,732
IssuesEvent
2020-07-06 19:06:38
nodejs/node
https://api.github.com/repos/nodejs/node
closed
fork method, args parameter cannot be null
child_process doc help wanted
* **Version**: 10.5.0 (but I think this is not related) * **Platform**: Ubuntu 18.04.1 LTS (but I think this is not related) * **Subsystem**: child_process Hi, I've just realized that this code can't work because args parameter can't be null. But it's not documented and I think it can be disturbing. ``` childProcess.fork(app.js', null, { execPath: process.execPath }) ``` But the following will work : ``` childProcess.fork(app.js', [], { execPath: process.execPath }) ``` I think this line is related to the issue : https://github.com/nodejs/node/blob/2bea9cefbc10ed1dd497bbae61c07d971da287dd/lib/child_process.js#L71 I'm not sure if it's the intended behavior, or if it's a lack in the documentation ?
1.0
fork method, args parameter cannot be null - * **Version**: 10.5.0 (but I think this is not related) * **Platform**: Ubuntu 18.04.1 LTS (but I think this is not related) * **Subsystem**: child_process Hi, I've just realized that this code can't work because args parameter can't be null. But it's not documented and I think it can be disturbing. ``` childProcess.fork(app.js', null, { execPath: process.execPath }) ``` But the following will work : ``` childProcess.fork(app.js', [], { execPath: process.execPath }) ``` I think this line is related to the issue : https://github.com/nodejs/node/blob/2bea9cefbc10ed1dd497bbae61c07d971da287dd/lib/child_process.js#L71 I'm not sure if it's the intended behavior, or if it's a lack in the documentation ?
process
fork method args parameter cannot be null version but i think this is not related platform ubuntu lts but i think this is not related subsystem child process hi i ve just realized that this code can t work because args parameter can t be null but it s not documented and i think it can be disturbing childprocess fork app js null execpath process execpath but the following will work childprocess fork app js execpath process execpath i think this line is related to the issue i m not sure if it s the intended behavior or if it s a lack in the documentation
1
14,825
18,166,779,974
IssuesEvent
2021-09-27 15:20:53
googleapis/python-db-dtypes-pandas
https://api.github.com/repos/googleapis/python-db-dtypes-pandas
opened
unit test coverage missing `__arrow_array__` methods
type: process
``` db_dtypes/__init__.py 105 2 36 0 99% 100, 160 ``` https://github.com/googleapis/python-db-dtypes-pandas/blob/bc9f34d9930f0a394c4e3fd52bc61f71c597aebf/db_dtypes/__init__.py#L100 https://github.com/googleapis/python-db-dtypes-pandas/blob/bc9f34d9930f0a394c4e3fd52bc61f71c597aebf/db_dtypes/__init__.py#L160
1.0
unit test coverage missing `__arrow_array__` methods - ``` db_dtypes/__init__.py 105 2 36 0 99% 100, 160 ``` https://github.com/googleapis/python-db-dtypes-pandas/blob/bc9f34d9930f0a394c4e3fd52bc61f71c597aebf/db_dtypes/__init__.py#L100 https://github.com/googleapis/python-db-dtypes-pandas/blob/bc9f34d9930f0a394c4e3fd52bc61f71c597aebf/db_dtypes/__init__.py#L160
process
unit test coverage missing arrow array methods db dtypes init py
1
250,842
27,112,880,704
IssuesEvent
2023-02-15 16:27:29
jgeraigery/ice
https://api.github.com/repos/jgeraigery/ice
opened
jquery-1.9.1.min.js: 4 vulnerabilities (highest severity is: 6.1)
security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.9.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.9.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.9.1/jquery.min.js</a></p> <p>Path to vulnerable library: /web-app/js/jquery-1.9.1.min.js</p> <p> <p>Found in HEAD commit: <a href="https://github.com/jgeraigery/ice/commit/b7ed460aeeb9a688355805c92166c1120aefa51e">b7ed460aeeb9a688355805c92166c1120aefa51e</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (jquery version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2020-11023](https://www.mend.io/vulnerability-database/CVE-2020-11023) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | jquery-1.9.1.min.js | Direct | jquery - 3.5.0;jquery-rails - 4.4.0 | &#10060; | | [CVE-2020-11022](https://www.mend.io/vulnerability-database/CVE-2020-11022) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | jquery-1.9.1.min.js | Direct | jQuery - 3.5.0 | &#10060; | | [CVE-2015-9251](https://www.mend.io/vulnerability-database/CVE-2015-9251) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | jquery-1.9.1.min.js | Direct | jQuery - 3.0.0 | &#10060; | | [CVE-2019-11358](https://www.mend.io/vulnerability-database/CVE-2019-11358) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | jquery-1.9.1.min.js | Direct | jquery - 3.4.0 | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-11023</summary> ### Vulnerable Library - <b>jquery-1.9.1.min.js</b></p> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.9.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.9.1/jquery.min.js</a></p> <p>Path to vulnerable library: /web-app/js/jquery-1.9.1.min.js</p> <p> Dependency Hierarchy: - :x: **jquery-1.9.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jgeraigery/ice/commit/b7ed460aeeb9a688355805c92166c1120aefa51e">b7ed460aeeb9a688355805c92166c1120aefa51e</a></p> <p>Found in base branch: <b>twc-custom-master</b></p> </p> <p></p> ### Vulnerability Details <p> In jQuery versions greater than or equal to 1.0.3 and before 3.5.0, passing HTML containing <option> elements from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-11023>CVE-2020-11023</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/jquery/jquery/security/advisories/GHSA-jpcq-cgw6-v4j6,https://github.com/rails/jquery-rails/blob/master/CHANGELOG.md#440">https://github.com/jquery/jquery/security/advisories/GHSA-jpcq-cgw6-v4j6,https://github.com/rails/jquery-rails/blob/master/CHANGELOG.md#440</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jquery - 3.5.0;jquery-rails - 4.4.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-11022</summary> ### Vulnerable Library - <b>jquery-1.9.1.min.js</b></p> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.9.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.9.1/jquery.min.js</a></p> <p>Path to vulnerable library: /web-app/js/jquery-1.9.1.min.js</p> <p> Dependency Hierarchy: - :x: **jquery-1.9.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jgeraigery/ice/commit/b7ed460aeeb9a688355805c92166c1120aefa51e">b7ed460aeeb9a688355805c92166c1120aefa51e</a></p> <p>Found in base branch: <b>twc-custom-master</b></p> </p> <p></p> ### Vulnerability Details <p> In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-11022>CVE-2020-11022</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11022">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11022</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jQuery - 3.5.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2015-9251</summary> ### Vulnerable Library - <b>jquery-1.9.1.min.js</b></p> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.9.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.9.1/jquery.min.js</a></p> <p>Path to vulnerable library: /web-app/js/jquery-1.9.1.min.js</p> <p> Dependency Hierarchy: - :x: **jquery-1.9.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jgeraigery/ice/commit/b7ed460aeeb9a688355805c92166c1120aefa51e">b7ed460aeeb9a688355805c92166c1120aefa51e</a></p> <p>Found in base branch: <b>twc-custom-master</b></p> </p> <p></p> ### Vulnerability Details <p> jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed. <p>Publish Date: 2018-01-18 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2015-9251>CVE-2015-9251</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p> <p>Release Date: 2018-01-18</p> <p>Fix Resolution: jQuery - 3.0.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2019-11358</summary> ### Vulnerable Library - <b>jquery-1.9.1.min.js</b></p> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.9.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.9.1/jquery.min.js</a></p> <p>Path to vulnerable library: /web-app/js/jquery-1.9.1.min.js</p> <p> Dependency Hierarchy: - :x: **jquery-1.9.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jgeraigery/ice/commit/b7ed460aeeb9a688355805c92166c1120aefa51e">b7ed460aeeb9a688355805c92166c1120aefa51e</a></p> <p>Found in base branch: <b>twc-custom-master</b></p> </p> <p></p> ### Vulnerability Details <p> jQuery before 3.4.0, as used in Drupal, Backdrop CMS, and other products, mishandles jQuery.extend(true, {}, ...) because of Object.prototype pollution. If an unsanitized source object contained an enumerable __proto__ property, it could extend the native Object.prototype. <p>Publish Date: 2019-04-20 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-11358>CVE-2019-11358</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11358">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11358</a></p> <p>Release Date: 2019-04-20</p> <p>Fix Resolution: jquery - 3.4.0</p> </p> <p></p> </details>
True
jquery-1.9.1.min.js: 4 vulnerabilities (highest severity is: 6.1) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.9.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.9.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.9.1/jquery.min.js</a></p> <p>Path to vulnerable library: /web-app/js/jquery-1.9.1.min.js</p> <p> <p>Found in HEAD commit: <a href="https://github.com/jgeraigery/ice/commit/b7ed460aeeb9a688355805c92166c1120aefa51e">b7ed460aeeb9a688355805c92166c1120aefa51e</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (jquery version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2020-11023](https://www.mend.io/vulnerability-database/CVE-2020-11023) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | jquery-1.9.1.min.js | Direct | jquery - 3.5.0;jquery-rails - 4.4.0 | &#10060; | | [CVE-2020-11022](https://www.mend.io/vulnerability-database/CVE-2020-11022) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | jquery-1.9.1.min.js | Direct | jQuery - 3.5.0 | &#10060; | | [CVE-2015-9251](https://www.mend.io/vulnerability-database/CVE-2015-9251) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | jquery-1.9.1.min.js | Direct | jQuery - 3.0.0 | &#10060; | | [CVE-2019-11358](https://www.mend.io/vulnerability-database/CVE-2019-11358) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | jquery-1.9.1.min.js | Direct | jquery - 3.4.0 | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-11023</summary> ### Vulnerable Library - <b>jquery-1.9.1.min.js</b></p> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.9.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.9.1/jquery.min.js</a></p> <p>Path to vulnerable library: /web-app/js/jquery-1.9.1.min.js</p> <p> Dependency Hierarchy: - :x: **jquery-1.9.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jgeraigery/ice/commit/b7ed460aeeb9a688355805c92166c1120aefa51e">b7ed460aeeb9a688355805c92166c1120aefa51e</a></p> <p>Found in base branch: <b>twc-custom-master</b></p> </p> <p></p> ### Vulnerability Details <p> In jQuery versions greater than or equal to 1.0.3 and before 3.5.0, passing HTML containing <option> elements from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-11023>CVE-2020-11023</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/jquery/jquery/security/advisories/GHSA-jpcq-cgw6-v4j6,https://github.com/rails/jquery-rails/blob/master/CHANGELOG.md#440">https://github.com/jquery/jquery/security/advisories/GHSA-jpcq-cgw6-v4j6,https://github.com/rails/jquery-rails/blob/master/CHANGELOG.md#440</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jquery - 3.5.0;jquery-rails - 4.4.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-11022</summary> ### Vulnerable Library - <b>jquery-1.9.1.min.js</b></p> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.9.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.9.1/jquery.min.js</a></p> <p>Path to vulnerable library: /web-app/js/jquery-1.9.1.min.js</p> <p> Dependency Hierarchy: - :x: **jquery-1.9.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jgeraigery/ice/commit/b7ed460aeeb9a688355805c92166c1120aefa51e">b7ed460aeeb9a688355805c92166c1120aefa51e</a></p> <p>Found in base branch: <b>twc-custom-master</b></p> </p> <p></p> ### Vulnerability Details <p> In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-11022>CVE-2020-11022</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11022">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11022</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jQuery - 3.5.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2015-9251</summary> ### Vulnerable Library - <b>jquery-1.9.1.min.js</b></p> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.9.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.9.1/jquery.min.js</a></p> <p>Path to vulnerable library: /web-app/js/jquery-1.9.1.min.js</p> <p> Dependency Hierarchy: - :x: **jquery-1.9.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jgeraigery/ice/commit/b7ed460aeeb9a688355805c92166c1120aefa51e">b7ed460aeeb9a688355805c92166c1120aefa51e</a></p> <p>Found in base branch: <b>twc-custom-master</b></p> </p> <p></p> ### Vulnerability Details <p> jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed. <p>Publish Date: 2018-01-18 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2015-9251>CVE-2015-9251</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p> <p>Release Date: 2018-01-18</p> <p>Fix Resolution: jQuery - 3.0.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2019-11358</summary> ### Vulnerable Library - <b>jquery-1.9.1.min.js</b></p> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.9.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.9.1/jquery.min.js</a></p> <p>Path to vulnerable library: /web-app/js/jquery-1.9.1.min.js</p> <p> Dependency Hierarchy: - :x: **jquery-1.9.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jgeraigery/ice/commit/b7ed460aeeb9a688355805c92166c1120aefa51e">b7ed460aeeb9a688355805c92166c1120aefa51e</a></p> <p>Found in base branch: <b>twc-custom-master</b></p> </p> <p></p> ### Vulnerability Details <p> jQuery before 3.4.0, as used in Drupal, Backdrop CMS, and other products, mishandles jQuery.extend(true, {}, ...) because of Object.prototype pollution. If an unsanitized source object contained an enumerable __proto__ property, it could extend the native Object.prototype. <p>Publish Date: 2019-04-20 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-11358>CVE-2019-11358</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11358">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11358</a></p> <p>Release Date: 2019-04-20</p> <p>Fix Resolution: jquery - 3.4.0</p> </p> <p></p> </details>
non_process
jquery min js vulnerabilities highest severity is vulnerable library jquery min js javascript library for dom operations library home page a href path to vulnerable library web app js jquery min js found in head commit a href vulnerabilities cve severity cvss dependency type fixed in jquery version remediation available medium jquery min js direct jquery jquery rails medium jquery min js direct jquery medium jquery min js direct jquery medium jquery min js direct jquery details cve vulnerable library jquery min js javascript library for dom operations library home page a href path to vulnerable library web app js jquery min js dependency hierarchy x jquery min js vulnerable library found in head commit a href found in base branch twc custom master vulnerability details in jquery versions greater than or equal to and before passing html containing elements from untrusted sources even after sanitizing it to one of jquery s dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery jquery rails cve vulnerable library jquery min js javascript library for dom operations library home page a href path to vulnerable library web app js jquery min js dependency hierarchy x jquery min js vulnerable library found in head commit a href found in base branch twc custom master vulnerability details in jquery versions greater than or equal to and before passing html from untrusted sources even after sanitizing it to one of jquery s dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery cve vulnerable library jquery min js javascript library for dom operations library home page a href path to vulnerable library web app js jquery min js dependency hierarchy x jquery min js vulnerable library found in head commit a href found in base branch twc custom master vulnerability details jquery before is vulnerable to cross site scripting xss attacks when a cross domain ajax request is performed without the datatype option causing text javascript responses to be executed publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery cve vulnerable library jquery min js javascript library for dom operations library home page a href path to vulnerable library web app js jquery min js dependency hierarchy x jquery min js vulnerable library found in head commit a href found in base branch twc custom master vulnerability details jquery before as used in drupal backdrop cms and other products mishandles jquery extend true because of object prototype pollution if an unsanitized source object contained an enumerable proto property it could extend the native object prototype publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery
0
151,076
13,389,789,312
IssuesEvent
2020-09-02 19:27:39
SibylLab/Program-Wars
https://api.github.com/repos/SibylLab/Program-Wars
closed
Add proper user help back to the game
UI documentation project improvement
**Is your feature request related to a problem? Please describe.** The rules modal is now out of date for the v2.5 game. **Describe the solution you'd like** We need to re-write the rules so that they describe the components and actions a player can take well. We can include some GIFs and screenshots of different actions to make the drag and drop usage clear. The current modal for the rules can be a bit awkward. I would like to split it up into a set of pages that can be opened in a new tab. That way it does not interrupt the game in the same way, and the player can have the rules open and switch back and forth if they need to.
1.0
Add proper user help back to the game - **Is your feature request related to a problem? Please describe.** The rules modal is now out of date for the v2.5 game. **Describe the solution you'd like** We need to re-write the rules so that they describe the components and actions a player can take well. We can include some GIFs and screenshots of different actions to make the drag and drop usage clear. The current modal for the rules can be a bit awkward. I would like to split it up into a set of pages that can be opened in a new tab. That way it does not interrupt the game in the same way, and the player can have the rules open and switch back and forth if they need to.
non_process
add proper user help back to the game is your feature request related to a problem please describe the rules modal is now out of date for the game describe the solution you d like we need to re write the rules so that they describe the components and actions a player can take well we can include some gifs and screenshots of different actions to make the drag and drop usage clear the current modal for the rules can be a bit awkward i would like to split it up into a set of pages that can be opened in a new tab that way it does not interrupt the game in the same way and the player can have the rules open and switch back and forth if they need to
0
11,784
14,616,809,688
IssuesEvent
2020-12-22 13:50:59
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[participant manager] add proper padding when there are no records
Bug P1 Participant manager Process: Fixed Process: Tested dev UI
<img width="1305" alt="Screen Shot 2020-11-26 at 4 58 29 PM" src="https://user-images.githubusercontent.com/3300694/100393747-bcc79480-3008-11eb-8094-162b123adffb.png">
2.0
[participant manager] add proper padding when there are no records - <img width="1305" alt="Screen Shot 2020-11-26 at 4 58 29 PM" src="https://user-images.githubusercontent.com/3300694/100393747-bcc79480-3008-11eb-8094-162b123adffb.png">
process
add proper padding when there are no records img width alt screen shot at pm src
1
22,489
31,464,699,035
IssuesEvent
2023-08-30 00:23:39
h4sh5/npm-auto-scanner
https://api.github.com/repos/h4sh5/npm-auto-scanner
opened
@signageos/sdk 1.21.1 has 2 guarddog issues
npm-install-script npm-silent-process-execution
```{"npm-install-script":[{"code":" \"prepare\": \"npm run prebuild \u0026\u0026 npm run build\",","location":"package/package.json:19","message":"The package.json has a script automatically running when the package is installed"}],"npm-silent-process-execution":[{"code":" serverProcess = child_process.fork(serverPath, [appletUid, appletVersion, port.toString(), publicUrl], {\n detached: true,\n stdio: 'ignore',\n ... });","location":"package/dist/Development/Applet/Serve/AppletServeManagement.js:488","message":"This package is silently executing another executable"}]}```
1.0
@signageos/sdk 1.21.1 has 2 guarddog issues - ```{"npm-install-script":[{"code":" \"prepare\": \"npm run prebuild \u0026\u0026 npm run build\",","location":"package/package.json:19","message":"The package.json has a script automatically running when the package is installed"}],"npm-silent-process-execution":[{"code":" serverProcess = child_process.fork(serverPath, [appletUid, appletVersion, port.toString(), publicUrl], {\n detached: true,\n stdio: 'ignore',\n ... });","location":"package/dist/Development/Applet/Serve/AppletServeManagement.js:488","message":"This package is silently executing another executable"}]}```
process
signageos sdk has guarddog issues npm install script npm silent process execution n detached true n stdio ignore n location package dist development applet serve appletservemanagement js message this package is silently executing another executable
1
21,225
28,311,099,999
IssuesEvent
2023-04-10 15:27:50
cse442-at-ub/project_s23-cinco
https://api.github.com/repos/cse442-at-ub/project_s23-cinco
closed
Retrieve and load events from database to feed
Processing Task Sprint 3
Task Tests Test 1: 1. Go to https://www-student.cse.buffalo.edu/CSE442-542/2023-Spring/cse-442b/build 2. Verify you can see events that are different and unique. ![image.png](https://images.zenhubusercontent.com/63e2cfc50476ff4c250db692/d510fd3b-f0cc-4a70-b311-cae96283266c) 3. Click on an event and verify you can see the post info such as poster, description, thumbnail, image, etc. 4. Go to the database -> Posts https://www-student.cse.buffalo.edu/tools/db/phpmyadmin/sql.php?server=1&db=cse442_2023_spring_team_b_db&table=Posts&pos=0 and verify that you can see the thumbnail image name corresponding to the poster. ![image.png](https://images.zenhubusercontent.com/63e2cfc50476ff4c250db692/f6b9e06c-9b00-4161-82cd-7439ce75b4c4) 6. Access the directory within the cheshire server via WinSCP or CyberDuck 7. Go into the uploads folder and make sure you can see the exact file name in the folder. ![image.png](https://images.zenhubusercontent.com/63e2cfc50476ff4c250db692/1415761e-4665-495d-899f-a20af8636389) 8. Download it to make sure it is the exact same image.
1.0
Retrieve and load events from database to feed - Task Tests Test 1: 1. Go to https://www-student.cse.buffalo.edu/CSE442-542/2023-Spring/cse-442b/build 2. Verify you can see events that are different and unique. ![image.png](https://images.zenhubusercontent.com/63e2cfc50476ff4c250db692/d510fd3b-f0cc-4a70-b311-cae96283266c) 3. Click on an event and verify you can see the post info such as poster, description, thumbnail, image, etc. 4. Go to the database -> Posts https://www-student.cse.buffalo.edu/tools/db/phpmyadmin/sql.php?server=1&db=cse442_2023_spring_team_b_db&table=Posts&pos=0 and verify that you can see the thumbnail image name corresponding to the poster. ![image.png](https://images.zenhubusercontent.com/63e2cfc50476ff4c250db692/f6b9e06c-9b00-4161-82cd-7439ce75b4c4) 6. Access the directory within the cheshire server via WinSCP or CyberDuck 7. Go into the uploads folder and make sure you can see the exact file name in the folder. ![image.png](https://images.zenhubusercontent.com/63e2cfc50476ff4c250db692/1415761e-4665-495d-899f-a20af8636389) 8. Download it to make sure it is the exact same image.
process
retrieve and load events from database to feed task tests test go to verify you can see events that are different and unique click on an event and verify you can see the post info such as poster description thumbnail image etc go to the database posts and verify that you can see the thumbnail image name corresponding to the poster access the directory within the cheshire server via winscp or cyberduck go into the uploads folder and make sure you can see the exact file name in the folder download it to make sure it is the exact same image
1
17,300
23,116,575,867
IssuesEvent
2022-07-27 17:15:37
GoogleCloudPlatform/anthos-service-mesh-samples
https://api.github.com/repos/GoogleCloudPlatform/anthos-service-mesh-samples
reopened
Dependency Dashboard
type: process priority: p3
This issue lists Renovate updates and detected dependencies. Read the [Dependency Dashboard](https://docs.renovatebot.com/key-concepts/dashboard/) docs to learn more. ## Edited/Blocked These updates have been manually edited so Renovate will no longer make changes. To discard all commits and start over, click on a checkbox. - [ ] <!-- rebase-branch=renovate/google-beta-4.x -->chore(deps): update terraform google-beta to v4.30.0 ## Detected dependencies <details><summary>github-actions</summary> <blockquote> <details><summary>.github/workflows/tf-validate-ci.yml</summary> - `actions/checkout v3` </details> </blockquote> </details> <details><summary>terraform</summary> <blockquote> <details><summary>demos/asm-mcp-multi-clusters/terraform/main.tf</summary> - `terraform-google-modules/project-factory/google 13.0.0` - `undefined no version found` - `undefined no version found` - `undefined no version found` - `undefined no version found` - `undefined no version found` - `undefined no version found` - `undefined no version found` - `undefined no version found` - `undefined no version found` - `undefined no version found` - `undefined no version found` - `undefined no version found` </details> <details><summary>demos/asm-mcp-multi-clusters/terraform/providers.tf</summary> - `google 4.30.0` - `google-beta 4.26.0` - `google no version found` - `google-beta no version found` </details> <details><summary>docs/mtls-egress-ingress/terraform/main.tf</summary> - `hashicorp/terraform >= 0.13.5` - `google >= 3.52.0` - `google no version found` - `terraform-google-modules/network/google no version found` - `terraform-google-modules/kubernetes-engine/google no version found` - `undefined no version found` - `undefined no version found` - `undefined no version found` - `undefined no version found` - `undefined no version found` - `undefined no version found` - `undefined no version found` </details> <details><summary>docs/terraform-asm-mcp/gke.tf</summary> - `kubernetes no version found` - `terraform-google-modules/kubernetes-engine/google ~> 22.0` - `terraform-google-modules/project-factory/google 13.0.0` </details> <details><summary>docs/terraform-asm-mcp/main.tf</summary> - `undefined no version found` - `terraform-google-modules/kubernetes-engine/google ~> 22.0` </details> <details><summary>docs/terraform-asm-mcp/providers.tf</summary> - `google-beta 4.26.0` - `google-beta no version found` </details> </blockquote> </details> --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
1.0
Dependency Dashboard - This issue lists Renovate updates and detected dependencies. Read the [Dependency Dashboard](https://docs.renovatebot.com/key-concepts/dashboard/) docs to learn more. ## Edited/Blocked These updates have been manually edited so Renovate will no longer make changes. To discard all commits and start over, click on a checkbox. - [ ] <!-- rebase-branch=renovate/google-beta-4.x -->chore(deps): update terraform google-beta to v4.30.0 ## Detected dependencies <details><summary>github-actions</summary> <blockquote> <details><summary>.github/workflows/tf-validate-ci.yml</summary> - `actions/checkout v3` </details> </blockquote> </details> <details><summary>terraform</summary> <blockquote> <details><summary>demos/asm-mcp-multi-clusters/terraform/main.tf</summary> - `terraform-google-modules/project-factory/google 13.0.0` - `undefined no version found` - `undefined no version found` - `undefined no version found` - `undefined no version found` - `undefined no version found` - `undefined no version found` - `undefined no version found` - `undefined no version found` - `undefined no version found` - `undefined no version found` - `undefined no version found` - `undefined no version found` </details> <details><summary>demos/asm-mcp-multi-clusters/terraform/providers.tf</summary> - `google 4.30.0` - `google-beta 4.26.0` - `google no version found` - `google-beta no version found` </details> <details><summary>docs/mtls-egress-ingress/terraform/main.tf</summary> - `hashicorp/terraform >= 0.13.5` - `google >= 3.52.0` - `google no version found` - `terraform-google-modules/network/google no version found` - `terraform-google-modules/kubernetes-engine/google no version found` - `undefined no version found` - `undefined no version found` - `undefined no version found` - `undefined no version found` - `undefined no version found` - `undefined no version found` - `undefined no version found` </details> <details><summary>docs/terraform-asm-mcp/gke.tf</summary> - `kubernetes no version found` - `terraform-google-modules/kubernetes-engine/google ~> 22.0` - `terraform-google-modules/project-factory/google 13.0.0` </details> <details><summary>docs/terraform-asm-mcp/main.tf</summary> - `undefined no version found` - `terraform-google-modules/kubernetes-engine/google ~> 22.0` </details> <details><summary>docs/terraform-asm-mcp/providers.tf</summary> - `google-beta 4.26.0` - `google-beta no version found` </details> </blockquote> </details> --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
process
dependency dashboard this issue lists renovate updates and detected dependencies read the docs to learn more edited blocked these updates have been manually edited so renovate will no longer make changes to discard all commits and start over click on a checkbox chore deps update terraform google beta to detected dependencies github actions github workflows tf validate ci yml actions checkout terraform demos asm mcp multi clusters terraform main tf terraform google modules project factory google undefined no version found undefined no version found undefined no version found undefined no version found undefined no version found undefined no version found undefined no version found undefined no version found undefined no version found undefined no version found undefined no version found undefined no version found demos asm mcp multi clusters terraform providers tf google google beta google no version found google beta no version found docs mtls egress ingress terraform main tf hashicorp terraform google google no version found terraform google modules network google no version found terraform google modules kubernetes engine google no version found undefined no version found undefined no version found undefined no version found undefined no version found undefined no version found undefined no version found undefined no version found docs terraform asm mcp gke tf kubernetes no version found terraform google modules kubernetes engine google terraform google modules project factory google docs terraform asm mcp main tf undefined no version found terraform google modules kubernetes engine google docs terraform asm mcp providers tf google beta google beta no version found check this box to trigger a request for renovate to run again on this repository
1
20,708
27,400,436,626
IssuesEvent
2023-02-28 23:57:07
open-telemetry/opentelemetry-collector-contrib
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
closed
Update heroku mappings per the latest specification
enhancement processor/resourcedetection
### Component(s) processor/resourcedetection ### Is your feature request related to a problem? Please describe. Per https://github.com/open-telemetry/opentelemetry-specification/pull/3075, we have a new mapping from heroku environment variables to resource attributes. ### Describe the solution you'd like Apply this new mapping in the processor. ### Describe alternatives you've considered _No response_ ### Additional context _No response_
1.0
Update heroku mappings per the latest specification - ### Component(s) processor/resourcedetection ### Is your feature request related to a problem? Please describe. Per https://github.com/open-telemetry/opentelemetry-specification/pull/3075, we have a new mapping from heroku environment variables to resource attributes. ### Describe the solution you'd like Apply this new mapping in the processor. ### Describe alternatives you've considered _No response_ ### Additional context _No response_
process
update heroku mappings per the latest specification component s processor resourcedetection is your feature request related to a problem please describe per we have a new mapping from heroku environment variables to resource attributes describe the solution you d like apply this new mapping in the processor describe alternatives you ve considered no response additional context no response
1
298,071
9,195,556,261
IssuesEvent
2019-03-07 02:56:01
gw2efficiency/issues
https://api.github.com/repos/gw2efficiency/issues
closed
Remove MF upgrade recipes from crafting calculator
1-Type: Bug 2-Priority: B 3-Complexity: Low 4-Impact: Medium 5-Area: Crafting 9-Status: For next release 9-Status: Ready for Release
This mainly concerns loadstone upgrade recipes, which just cause the recipes consuming the loadstones to have the wrong prices. Spirit shards/mystic binding agents are not free.
1.0
Remove MF upgrade recipes from crafting calculator - This mainly concerns loadstone upgrade recipes, which just cause the recipes consuming the loadstones to have the wrong prices. Spirit shards/mystic binding agents are not free.
non_process
remove mf upgrade recipes from crafting calculator this mainly concerns loadstone upgrade recipes which just cause the recipes consuming the loadstones to have the wrong prices spirit shards mystic binding agents are not free
0
94
2,535,087,410
IssuesEvent
2015-01-25 17:42:06
chrisalexander/Learn-Chinese-app
https://api.github.com/repos/chrisalexander/Learn-Chinese-app
closed
Process completeness should be integer
LongRunningProcess
At the highest level, it should be exposed as an int Its change notification should only be triggered when it will produce a meaningful change to the UI
1.0
Process completeness should be integer - At the highest level, it should be exposed as an int Its change notification should only be triggered when it will produce a meaningful change to the UI
process
process completeness should be integer at the highest level it should be exposed as an int its change notification should only be triggered when it will produce a meaningful change to the ui
1
253,378
21,677,351,119
IssuesEvent
2022-05-08 23:01:46
tagbase/tagbase-server
https://api.github.com/repos/tagbase/tagbase-server
closed
Address ingestion issue(s) with 2F159903_2012_117464_eTUFF.txt
bug tests
``` postgres_1 | 2022-03-20 21:16:18.605 UTC [45] ERROR: invalid input syntax for type double precision: "A" at character 162 postgres_1 | 2022-03-20 21:16:18.605 UTC [45] STATEMENT: INSERT INTO proc_observations (date_time, variable_id, variable_value, submission_id, tag_id, final_value) VALUES ('2012-03-16T18:31:39+00:00'::timestamptz, 34, 'A', 5, '5', 'FALSE') tagbase_server_1 | ERROR:__main__:Exception on /v0.4.0/ingest [GET] tagbase_server_1 | Traceback (most recent call last): tagbase_server_1 | File "/usr/local/lib/python3.10/site-packages/flask/app.py", line 2073, in wsgi_app tagbase_server_1 | response = self.full_dispatch_request() tagbase_server_1 | File "/usr/local/lib/python3.10/site-packages/flask/app.py", line 1518, in full_dispatch_request tagbase_server_1 | rv = self.handle_user_exception(e) tagbase_server_1 | File "/usr/local/lib/python3.10/site-packages/flask/app.py", line 1516, in full_dispatch_request tagbase_server_1 | rv = self.dispatch_request() tagbase_server_1 | File "/usr/local/lib/python3.10/site-packages/flask/app.py", line 1502, in dispatch_request tagbase_server_1 | return self.ensure_sync(self.view_functions[rule.endpoint])(**req.view_args) tagbase_server_1 | File "/usr/local/lib/python3.10/site-packages/connexion/decorators/decorator.py", line 68, in wrapper tagbase_server_1 | response = function(request) tagbase_server_1 | File "/usr/local/lib/python3.10/site-packages/connexion/decorators/uri_parsing.py", line 149, in wrapper tagbase_server_1 | response = function(request) tagbase_server_1 | File "/usr/local/lib/python3.10/site-packages/connexion/decorators/validation.py", line 396, in wrapper tagbase_server_1 | return function(request) tagbase_server_1 | File "/usr/local/lib/python3.10/site-packages/connexion/decorators/parameter.py", line 116, in wrapper tagbase_server_1 | return function(**kwargs) tagbase_server_1 | File "/usr/src/app/tagbase_server/controllers/ingest_controller.py", line 201, in ingest_etuff_get tagbase_server_1 | cur.execute( tagbase_server_1 | psycopg2.errors.InvalidTextRepresentation: invalid input syntax for type double precision: "A" tagbase_server_1 | LINE 1: ...ES ('2012-03-16T18:31:39+00:00'::timestamptz, 34, 'A', 5, '5... tagbase_server_1 | ^ tagbase_server_1 | tagbase_server_1 | ERROR:tornado.access:500 GET /v0.4.0/ingest?file=file%3A%2F%2F%2Fusr%2Fsrc%2Fapp%2Fdata%2F159903_2012_117464_eTUFF.txt&type=etuff (172.24.0.1) 92.60ms ```
1.0
Address ingestion issue(s) with 2F159903_2012_117464_eTUFF.txt - ``` postgres_1 | 2022-03-20 21:16:18.605 UTC [45] ERROR: invalid input syntax for type double precision: "A" at character 162 postgres_1 | 2022-03-20 21:16:18.605 UTC [45] STATEMENT: INSERT INTO proc_observations (date_time, variable_id, variable_value, submission_id, tag_id, final_value) VALUES ('2012-03-16T18:31:39+00:00'::timestamptz, 34, 'A', 5, '5', 'FALSE') tagbase_server_1 | ERROR:__main__:Exception on /v0.4.0/ingest [GET] tagbase_server_1 | Traceback (most recent call last): tagbase_server_1 | File "/usr/local/lib/python3.10/site-packages/flask/app.py", line 2073, in wsgi_app tagbase_server_1 | response = self.full_dispatch_request() tagbase_server_1 | File "/usr/local/lib/python3.10/site-packages/flask/app.py", line 1518, in full_dispatch_request tagbase_server_1 | rv = self.handle_user_exception(e) tagbase_server_1 | File "/usr/local/lib/python3.10/site-packages/flask/app.py", line 1516, in full_dispatch_request tagbase_server_1 | rv = self.dispatch_request() tagbase_server_1 | File "/usr/local/lib/python3.10/site-packages/flask/app.py", line 1502, in dispatch_request tagbase_server_1 | return self.ensure_sync(self.view_functions[rule.endpoint])(**req.view_args) tagbase_server_1 | File "/usr/local/lib/python3.10/site-packages/connexion/decorators/decorator.py", line 68, in wrapper tagbase_server_1 | response = function(request) tagbase_server_1 | File "/usr/local/lib/python3.10/site-packages/connexion/decorators/uri_parsing.py", line 149, in wrapper tagbase_server_1 | response = function(request) tagbase_server_1 | File "/usr/local/lib/python3.10/site-packages/connexion/decorators/validation.py", line 396, in wrapper tagbase_server_1 | return function(request) tagbase_server_1 | File "/usr/local/lib/python3.10/site-packages/connexion/decorators/parameter.py", line 116, in wrapper tagbase_server_1 | return function(**kwargs) tagbase_server_1 | File "/usr/src/app/tagbase_server/controllers/ingest_controller.py", line 201, in ingest_etuff_get tagbase_server_1 | cur.execute( tagbase_server_1 | psycopg2.errors.InvalidTextRepresentation: invalid input syntax for type double precision: "A" tagbase_server_1 | LINE 1: ...ES ('2012-03-16T18:31:39+00:00'::timestamptz, 34, 'A', 5, '5... tagbase_server_1 | ^ tagbase_server_1 | tagbase_server_1 | ERROR:tornado.access:500 GET /v0.4.0/ingest?file=file%3A%2F%2F%2Fusr%2Fsrc%2Fapp%2Fdata%2F159903_2012_117464_eTUFF.txt&type=etuff (172.24.0.1) 92.60ms ```
non_process
address ingestion issue s with etuff txt postgres utc error invalid input syntax for type double precision a at character postgres utc statement insert into proc observations date time variable id variable value submission id tag id final value values timestamptz a false tagbase server error main exception on ingest tagbase server traceback most recent call last tagbase server file usr local lib site packages flask app py line in wsgi app tagbase server response self full dispatch request tagbase server file usr local lib site packages flask app py line in full dispatch request tagbase server rv self handle user exception e tagbase server file usr local lib site packages flask app py line in full dispatch request tagbase server rv self dispatch request tagbase server file usr local lib site packages flask app py line in dispatch request tagbase server return self ensure sync self view functions req view args tagbase server file usr local lib site packages connexion decorators decorator py line in wrapper tagbase server response function request tagbase server file usr local lib site packages connexion decorators uri parsing py line in wrapper tagbase server response function request tagbase server file usr local lib site packages connexion decorators validation py line in wrapper tagbase server return function request tagbase server file usr local lib site packages connexion decorators parameter py line in wrapper tagbase server return function kwargs tagbase server file usr src app tagbase server controllers ingest controller py line in ingest etuff get tagbase server cur execute tagbase server errors invalidtextrepresentation invalid input syntax for type double precision a tagbase server line es timestamptz a tagbase server tagbase server tagbase server error tornado access get ingest file file etuff txt type etuff
0
15,381
9,545,556,712
IssuesEvent
2019-05-01 17:24:28
uniquelyparticular/shopify-oauth
https://api.github.com/repos/uniquelyparticular/shopify-oauth
closed
WS-2019-0047 (Medium) detected in tar-2.2.1.tgz
security vulnerability
## WS-2019-0047 - Medium Severity Vulnerability <details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tar-2.2.1.tgz</b></p></summary> <p>tar for node</p> <p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-2.2.1.tgz">https://registry.npmjs.org/tar/-/tar-2.2.1.tgz</a></p> <p>Path to dependency file: /shopify-oauth/package.json</p> <p>Path to vulnerable library: /tmp/git/shopify-oauth/node_modules/npm/node_modules/node-gyp/node_modules/tar/package.json</p> <p> Dependency Hierarchy: - semantic-release-15.13.3.tgz (Root Library) - npm-5.1.7.tgz - npm-6.9.0.tgz - node-gyp-3.8.0.tgz - :x: **tar-2.2.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/uniquelyparticular/shopify-oauth/commit/7b1730d8525267e67007fdb48b9e406a0d86664c">7b1730d8525267e67007fdb48b9e406a0d86664c</a></p> </p> </details> <p></p> <details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Versions of node-tar prior to 4.4.2 are vulnerable to Arbitrary File Overwrite. Extracting tarballs containing a hardlink to a file that already exists in the system, and a file that matches the hardlink will overwrite the system's file with the contents of the extracted file. <p>Publish Date: 2019-04-05 <p>URL: <a href=https://github.com/npm/node-tar/commit/b0c58433c22f5e7fe8b1c76373f27e3f81dcd4c8>WS-2019-0047</a></p> </p> </details> <p></p> <details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/803">https://www.npmjs.com/advisories/803</a></p> <p>Release Date: 2019-04-05</p> <p>Fix Resolution: 4.4.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2019-0047 (Medium) detected in tar-2.2.1.tgz - ## WS-2019-0047 - Medium Severity Vulnerability <details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tar-2.2.1.tgz</b></p></summary> <p>tar for node</p> <p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-2.2.1.tgz">https://registry.npmjs.org/tar/-/tar-2.2.1.tgz</a></p> <p>Path to dependency file: /shopify-oauth/package.json</p> <p>Path to vulnerable library: /tmp/git/shopify-oauth/node_modules/npm/node_modules/node-gyp/node_modules/tar/package.json</p> <p> Dependency Hierarchy: - semantic-release-15.13.3.tgz (Root Library) - npm-5.1.7.tgz - npm-6.9.0.tgz - node-gyp-3.8.0.tgz - :x: **tar-2.2.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/uniquelyparticular/shopify-oauth/commit/7b1730d8525267e67007fdb48b9e406a0d86664c">7b1730d8525267e67007fdb48b9e406a0d86664c</a></p> </p> </details> <p></p> <details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Versions of node-tar prior to 4.4.2 are vulnerable to Arbitrary File Overwrite. Extracting tarballs containing a hardlink to a file that already exists in the system, and a file that matches the hardlink will overwrite the system's file with the contents of the extracted file. <p>Publish Date: 2019-04-05 <p>URL: <a href=https://github.com/npm/node-tar/commit/b0c58433c22f5e7fe8b1c76373f27e3f81dcd4c8>WS-2019-0047</a></p> </p> </details> <p></p> <details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/803">https://www.npmjs.com/advisories/803</a></p> <p>Release Date: 2019-04-05</p> <p>Fix Resolution: 4.4.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
ws medium detected in tar tgz ws medium severity vulnerability vulnerable library tar tgz tar for node library home page a href path to dependency file shopify oauth package json path to vulnerable library tmp git shopify oauth node modules npm node modules node gyp node modules tar package json dependency hierarchy semantic release tgz root library npm tgz npm tgz node gyp tgz x tar tgz vulnerable library found in head commit a href vulnerability details versions of node tar prior to are vulnerable to arbitrary file overwrite extracting tarballs containing a hardlink to a file that already exists in the system and a file that matches the hardlink will overwrite the system s file with the contents of the extracted file publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
19,022
25,027,900,864
IssuesEvent
2022-11-04 09:44:57
open-telemetry/opentelemetry-collector-contrib
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
closed
[processor/metricstransform] Add support for scaling histogram metrics
enhancement processor/metricstransform
### Is your feature request related to a problem? Please describe. We would like to convert latency histogram metrics we receive in milliseconds to seconds. ### Describe the solution you'd like Add support for scaling histogram datapoint sums and explicit bounds in the metricstransform processor. ### Describe alternatives you've considered _No response_ ### Additional context _No response_
1.0
[processor/metricstransform] Add support for scaling histogram metrics - ### Is your feature request related to a problem? Please describe. We would like to convert latency histogram metrics we receive in milliseconds to seconds. ### Describe the solution you'd like Add support for scaling histogram datapoint sums and explicit bounds in the metricstransform processor. ### Describe alternatives you've considered _No response_ ### Additional context _No response_
process
add support for scaling histogram metrics is your feature request related to a problem please describe we would like to convert latency histogram metrics we receive in milliseconds to seconds describe the solution you d like add support for scaling histogram datapoint sums and explicit bounds in the metricstransform processor describe alternatives you ve considered no response additional context no response
1
51,773
6,548,033,863
IssuesEvent
2017-09-04 18:19:59
Marri/glowfic
https://api.github.com/repos/Marri/glowfic
closed
Keyword dropdown display is blank if a reply's icon is not in the character's galleries
bug ui/design
If you edit a post where the icon used is not in the character used's gallery, the icon keyword dropdown displays a blank value. This is not the worst way for it to fallback, obviously, but might be unintended behavior? (Perhaps it should default to give you that keyword in the box?)
1.0
Keyword dropdown display is blank if a reply's icon is not in the character's galleries - If you edit a post where the icon used is not in the character used's gallery, the icon keyword dropdown displays a blank value. This is not the worst way for it to fallback, obviously, but might be unintended behavior? (Perhaps it should default to give you that keyword in the box?)
non_process
keyword dropdown display is blank if a reply s icon is not in the character s galleries if you edit a post where the icon used is not in the character used s gallery the icon keyword dropdown displays a blank value this is not the worst way for it to fallback obviously but might be unintended behavior perhaps it should default to give you that keyword in the box
0
238,944
18,255,961,957
IssuesEvent
2021-10-03 02:56:32
typeorm/typeorm
https://api.github.com/repos/typeorm/typeorm
opened
Repository Find Options: and with or
documentation requires triage
<!-- Please follow the template. If you don't, your issue may be closed. Have a question? This is the TypeORM issue tracker - and not the right place for general support or questions. Instead, check the "Support" Documentation on the best places to ask questions! https://github.com/typeorm/typeorm/blob/master/docs/support.md --> ## Documentation Issue ### What was unclear or otherwise insufficient? <!-- If relevant, Please be clear about the documentation file, as well as the location within the file. Link to the documentation in the repository. If the page does not exist, please be clear why a new documentation section is needed. --> ### Recommended Fix <!-- How should we fix this documentation issue? Should we add examples, clarify the language, or drop the page entirely? --> ### Additional Context <!-- Add any other context about the documentation issue here. --> ### Are you willing to resolve this issue by submitting a Pull Request? <!-- Put "✅" (:white_check_mark:) to one of these options, left "✖️" (:heavy_multiplication_x:) others: --> - ✅ Yes, I have the time, and I know how to start. - ✖️ Yes, I have the time, but I don't know how to start. I would need guidance. - ✖️ No, I don’t have the time, but I can support (using donations) development. - ✖️ No, I don’t have the time and I’m okay to wait for the community / maintainers to resolve this issue. <!-- Remember that first-time contributors are welcome! 🙌 👋 Have a great day and thank you for the documentation problem report! --> Querying a column from an embedded entity should be done with respect to the hierarchy in which it was defined. Example: ``` userRepository.find({ where: { project: { name: "TypeORM", initials: "TORM" }, } }); ``` Querying with OR operator: ``` userRepository.find({ where: [ { firstName: "Timber", lastName: "Saw" }, { firstName: "Stan", lastName: "Lee" }, ], }); ``` but and or? ``` userRepository.find({ where: { project: { name: "TypeORM", initials: "TORM" }, }, andWhere: [ { firstName: "Timber", lastName: "Saw" }, { firstName: "Stan", lastName: "Lee" }, ] }); ``` but andWhere not working. what should I do?
1.0
Repository Find Options: and with or - <!-- Please follow the template. If you don't, your issue may be closed. Have a question? This is the TypeORM issue tracker - and not the right place for general support or questions. Instead, check the "Support" Documentation on the best places to ask questions! https://github.com/typeorm/typeorm/blob/master/docs/support.md --> ## Documentation Issue ### What was unclear or otherwise insufficient? <!-- If relevant, Please be clear about the documentation file, as well as the location within the file. Link to the documentation in the repository. If the page does not exist, please be clear why a new documentation section is needed. --> ### Recommended Fix <!-- How should we fix this documentation issue? Should we add examples, clarify the language, or drop the page entirely? --> ### Additional Context <!-- Add any other context about the documentation issue here. --> ### Are you willing to resolve this issue by submitting a Pull Request? <!-- Put "✅" (:white_check_mark:) to one of these options, left "✖️" (:heavy_multiplication_x:) others: --> - ✅ Yes, I have the time, and I know how to start. - ✖️ Yes, I have the time, but I don't know how to start. I would need guidance. - ✖️ No, I don’t have the time, but I can support (using donations) development. - ✖️ No, I don’t have the time and I’m okay to wait for the community / maintainers to resolve this issue. <!-- Remember that first-time contributors are welcome! 🙌 👋 Have a great day and thank you for the documentation problem report! --> Querying a column from an embedded entity should be done with respect to the hierarchy in which it was defined. Example: ``` userRepository.find({ where: { project: { name: "TypeORM", initials: "TORM" }, } }); ``` Querying with OR operator: ``` userRepository.find({ where: [ { firstName: "Timber", lastName: "Saw" }, { firstName: "Stan", lastName: "Lee" }, ], }); ``` but and or? ``` userRepository.find({ where: { project: { name: "TypeORM", initials: "TORM" }, }, andWhere: [ { firstName: "Timber", lastName: "Saw" }, { firstName: "Stan", lastName: "Lee" }, ] }); ``` but andWhere not working. what should I do?
non_process
repository find options and with or please follow the template if you don t your issue may be closed have a question this is the typeorm issue tracker and not the right place for general support or questions instead check the support documentation on the best places to ask questions documentation issue what was unclear or otherwise insufficient if relevant please be clear about the documentation file as well as the location within the file link to the documentation in the repository if the page does not exist please be clear why a new documentation section is needed recommended fix how should we fix this documentation issue should we add examples clarify the language or drop the page entirely additional context add any other context about the documentation issue here are you willing to resolve this issue by submitting a pull request ✅ yes i have the time and i know how to start ✖️ yes i have the time but i don t know how to start i would need guidance ✖️ no i don’t have the time but i can support using donations development ✖️ no i don’t have the time and i’m okay to wait for the community maintainers to resolve this issue remember that first time contributors are welcome 🙌 👋 have a great day and thank you for the documentation problem report querying a column from an embedded entity should be done with respect to the hierarchy in which it was defined example userrepository find where project name typeorm initials torm querying with or operator userrepository find where firstname timber lastname saw firstname stan lastname lee but and or userrepository find where project name typeorm initials torm andwhere firstname timber lastname saw firstname stan lastname lee but andwhere not working what should i do?
0
825,391
31,388,221,570
IssuesEvent
2023-08-26 02:38:09
formkit/formkit
https://api.github.com/repos/formkit/formkit
closed
Missing types for slots in the Autocomplete input
🐛 bug 🚀 release-ready 🔴 priority-1 Pro
### Reproduction https://formkit.link/a8bc993b26501fd8a261d6d085faf48c ### Describe the bug Since 0.18.x the typings for the v-model seem to not match properly, even though the type is "number". In the source code it looks like the v-model type is derived from the value (which we don't want to set here) - but if I pass a number as value it does not work as well. ![image](https://github.com/formkit/formkit/assets/26108368/99587b82-a851-416b-b1fb-6afe3eb11f44) ```ts <FormKit v-model="finalPostalCode" name="zipCode" type="number" validation="required|number|length:5,5" :validation-messages="{ length: zipCodeValidationMessage, number: zipCodeValidationMessage, required: zipCodeValidationMessage, }" /> ``` ### Environment • OS: [e.g. iOS] • Browser [e.g. chrome, safari] • Version [e.g. 22]
1.0
Missing types for slots in the Autocomplete input - ### Reproduction https://formkit.link/a8bc993b26501fd8a261d6d085faf48c ### Describe the bug Since 0.18.x the typings for the v-model seem to not match properly, even though the type is "number". In the source code it looks like the v-model type is derived from the value (which we don't want to set here) - but if I pass a number as value it does not work as well. ![image](https://github.com/formkit/formkit/assets/26108368/99587b82-a851-416b-b1fb-6afe3eb11f44) ```ts <FormKit v-model="finalPostalCode" name="zipCode" type="number" validation="required|number|length:5,5" :validation-messages="{ length: zipCodeValidationMessage, number: zipCodeValidationMessage, required: zipCodeValidationMessage, }" /> ``` ### Environment • OS: [e.g. iOS] • Browser [e.g. chrome, safari] • Version [e.g. 22]
non_process
missing types for slots in the autocomplete input reproduction describe the bug since x the typings for the v model seem to not match properly even though the type is number in the source code it looks like the v model type is derived from the value which we don t want to set here but if i pass a number as value it does not work as well ts formkit v model finalpostalcode name zipcode type number validation required number length validation messages length zipcodevalidationmessage number zipcodevalidationmessage required zipcodevalidationmessage environment • os • browser • version
0
10,065
13,044,161,799
IssuesEvent
2020-07-29 03:47:26
tikv/tikv
https://api.github.com/repos/tikv/tikv
closed
UCP: Migrate scalar function `ExtractDatetime` from TiDB
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
## Description Port the scalar function `ExtractDatetime` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @lonng ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
2.0
UCP: Migrate scalar function `ExtractDatetime` from TiDB - ## Description Port the scalar function `ExtractDatetime` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @lonng ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
process
ucp migrate scalar function extractdatetime from tidb description port the scalar function extractdatetime from tidb to coprocessor score mentor s lonng recommended skills rust programming learning materials already implemented expressions ported from tidb
1
646,525
21,051,577,607
IssuesEvent
2022-03-31 21:02:20
uutils/coreutils
https://api.github.com/repos/uutils/coreutils
closed
"cargo test" fails on mageia linux v7 x86-64
I - Incorrect B - medium priority P - Linux U - du U - stat U - who
I am getting this: ``` thread 'test_who::test_short' panicked at 'assertion failed: `(left == right)` left: `"shlomif tty1 2019-02-06 17:03\nshlomif pts/0 2019-02-07 22:31 (:0)\nshlomif pts/2 2019-02-07 22:31 (:0)\nshlomif pts/11 2019-02-07 22:45 (:0)\nshlomif pts/5 2019-02-07 22:32 (:0)\nshlomif pts/9 2019-02-07 22:41 (:0)\nshlomif pts/12 2019-02-07 22:45 (:0)\nshlomif pts/10 2019-02-07 22:45 (:0)"`, right: `"shlomif tty1 2019-02-06 17:03\nshlomif pts/0 2019-02-07 22:31 (:0)\nshlomif pts/1 2019-02-06 17:05 (:0)\nshlomif pts/2 2019-02-07 22:31 (:0)\nshlomif pts/6 2019-02-06 17:22 (:0)\nshlomif pts/7 2019-02-06 17:22 (:0)\nshlomif pts/11 2019-02-07 22:45 (:0)\nshlomif pts/5 2019-02-07 22:32 (:0)\nshlomif pts/9 2019-02-07 22:41 (:0)\nshlomif pts/12 2019-02-07 22:45 (:0)\nshlomif pts/10 2019-02-07 22:45 (:0)"`', tests/common/util.rs:95:9 failures: test_du::test_du_basics_subdir test_du::test_du_d_flag test_du::test_du_hard_link test_du::test_du_soft_link test_who::test_all test_who::test_count test_who::test_heading test_who::test_short test result: FAILED. 573 passed; 8 failed; 12 ignored; 0 measured; 0 filtered out error: test failed, to rerun pass '--test tests' [shlomif@telaviv1 coreutils]$ rpm -q rust rust-1.32.0-1.mga7 [shlomif@telaviv1 coreutils]$ ```
1.0
"cargo test" fails on mageia linux v7 x86-64 - I am getting this: ``` thread 'test_who::test_short' panicked at 'assertion failed: `(left == right)` left: `"shlomif tty1 2019-02-06 17:03\nshlomif pts/0 2019-02-07 22:31 (:0)\nshlomif pts/2 2019-02-07 22:31 (:0)\nshlomif pts/11 2019-02-07 22:45 (:0)\nshlomif pts/5 2019-02-07 22:32 (:0)\nshlomif pts/9 2019-02-07 22:41 (:0)\nshlomif pts/12 2019-02-07 22:45 (:0)\nshlomif pts/10 2019-02-07 22:45 (:0)"`, right: `"shlomif tty1 2019-02-06 17:03\nshlomif pts/0 2019-02-07 22:31 (:0)\nshlomif pts/1 2019-02-06 17:05 (:0)\nshlomif pts/2 2019-02-07 22:31 (:0)\nshlomif pts/6 2019-02-06 17:22 (:0)\nshlomif pts/7 2019-02-06 17:22 (:0)\nshlomif pts/11 2019-02-07 22:45 (:0)\nshlomif pts/5 2019-02-07 22:32 (:0)\nshlomif pts/9 2019-02-07 22:41 (:0)\nshlomif pts/12 2019-02-07 22:45 (:0)\nshlomif pts/10 2019-02-07 22:45 (:0)"`', tests/common/util.rs:95:9 failures: test_du::test_du_basics_subdir test_du::test_du_d_flag test_du::test_du_hard_link test_du::test_du_soft_link test_who::test_all test_who::test_count test_who::test_heading test_who::test_short test result: FAILED. 573 passed; 8 failed; 12 ignored; 0 measured; 0 filtered out error: test failed, to rerun pass '--test tests' [shlomif@telaviv1 coreutils]$ rpm -q rust rust-1.32.0-1.mga7 [shlomif@telaviv1 coreutils]$ ```
non_process
cargo test fails on mageia linux i am getting this thread test who test short panicked at assertion failed left right left shlomif nshlomif pts nshlomif pts nshlomif pts nshlomif pts nshlomif pts nshlomif pts nshlomif pts right shlomif nshlomif pts nshlomif pts nshlomif pts nshlomif pts nshlomif pts nshlomif pts nshlomif pts nshlomif pts nshlomif pts nshlomif pts tests common util rs failures test du test du basics subdir test du test du d flag test du test du hard link test du test du soft link test who test all test who test count test who test heading test who test short test result failed passed failed ignored measured filtered out error test failed to rerun pass test tests rpm q rust rust
0
7,663
10,756,094,293
IssuesEvent
2019-10-31 10:29:01
opengeospatial/CityGML-3.0CM
https://api.github.com/repos/opengeospatial/CityGML-3.0CM
closed
Need a Profiling Mechanism
SWG Process
We seem to be reaching a consensus that the overall model has too many features for any specific application/use case. How do we allow implementers to choose only the features they need without forcing the recipients of a model to have code that can accept any combination of features from the universe described in the Conceptual Model. Are we smart enough to structure the modules of the model to achieve this goal? The ADE extendsion mechanism seems successful. Can we achieve the same with some kind of a clever "restriction" mechanism?
1.0
Need a Profiling Mechanism - We seem to be reaching a consensus that the overall model has too many features for any specific application/use case. How do we allow implementers to choose only the features they need without forcing the recipients of a model to have code that can accept any combination of features from the universe described in the Conceptual Model. Are we smart enough to structure the modules of the model to achieve this goal? The ADE extendsion mechanism seems successful. Can we achieve the same with some kind of a clever "restriction" mechanism?
process
need a profiling mechanism we seem to be reaching a consensus that the overall model has too many features for any specific application use case how do we allow implementers to choose only the features they need without forcing the recipients of a model to have code that can accept any combination of features from the universe described in the conceptual model are we smart enough to structure the modules of the model to achieve this goal the ade extendsion mechanism seems successful can we achieve the same with some kind of a clever restriction mechanism
1
351,606
32,012,737,926
IssuesEvent
2023-09-21 19:12:45
astropy/astropy
https://api.github.com/repos/astropy/astropy
opened
TST/BUG: votable.tree._parse_fits is not covered by tests
testing io.votable Bug
We reused much of the code for `_parse_parquet` in #15281 and windows seemed to be especially sensitive to `tmp_path` in the tests, the is pathlib.Path object rather than an absolute path. So that, and other relative paths may cause problems for the FITS case, too. (The main coverage link seems to be broken as it reports a very low number for astropy, as is, so I attach the local screenshot for a run of `pytest -P io.votable --cov astropy/io/votable/` instead). <img width="773" alt="Screenshot 2023-09-21 at 12 08 07" src="https://github.com/astropy/astropy/assets/6788290/d33ecc32-33a8-4e1b-830f-2975e8a3c216">
1.0
TST/BUG: votable.tree._parse_fits is not covered by tests - We reused much of the code for `_parse_parquet` in #15281 and windows seemed to be especially sensitive to `tmp_path` in the tests, the is pathlib.Path object rather than an absolute path. So that, and other relative paths may cause problems for the FITS case, too. (The main coverage link seems to be broken as it reports a very low number for astropy, as is, so I attach the local screenshot for a run of `pytest -P io.votable --cov astropy/io/votable/` instead). <img width="773" alt="Screenshot 2023-09-21 at 12 08 07" src="https://github.com/astropy/astropy/assets/6788290/d33ecc32-33a8-4e1b-830f-2975e8a3c216">
non_process
tst bug votable tree parse fits is not covered by tests we reused much of the code for parse parquet in and windows seemed to be especially sensitive to tmp path in the tests the is pathlib path object rather than an absolute path so that and other relative paths may cause problems for the fits case too the main coverage link seems to be broken as it reports a very low number for astropy as is so i attach the local screenshot for a run of pytest p io votable cov astropy io votable instead img width alt screenshot at src
0
7,034
10,194,458,130
IssuesEvent
2019-08-12 15:44:11
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Azure Datacenter IP address list is deprecated
Pri1 assigned-to-author automation/svc doc-enhancement process-automation/subsvc triaged
Hi MSFT, Can you possibly update the links to the "Azure Datacenter IP address list" as it is pointing to a deprecated list now. Thanks in advance. Kr, Bart van Berkel --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 3092ee3a-3c57-cc53-186b-b454e7d9d190 * Version Independent ID: 6f6b5a07-397a-98a2-6091-941244a77837 * Content: [Azure Automation Hybrid Runbook Worker](https://docs.microsoft.com/en-us/azure/automation/automation-hybrid-runbook-worker#) * Content Source: [articles/automation/automation-hybrid-runbook-worker.md](https://github.com/Microsoft/azure-docs/blob/master/articles/automation/automation-hybrid-runbook-worker.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @bobbytreed * Microsoft Alias: **robreed**
1.0
Azure Datacenter IP address list is deprecated - Hi MSFT, Can you possibly update the links to the "Azure Datacenter IP address list" as it is pointing to a deprecated list now. Thanks in advance. Kr, Bart van Berkel --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 3092ee3a-3c57-cc53-186b-b454e7d9d190 * Version Independent ID: 6f6b5a07-397a-98a2-6091-941244a77837 * Content: [Azure Automation Hybrid Runbook Worker](https://docs.microsoft.com/en-us/azure/automation/automation-hybrid-runbook-worker#) * Content Source: [articles/automation/automation-hybrid-runbook-worker.md](https://github.com/Microsoft/azure-docs/blob/master/articles/automation/automation-hybrid-runbook-worker.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @bobbytreed * Microsoft Alias: **robreed**
process
azure datacenter ip address list is deprecated hi msft can you possibly update the links to the azure datacenter ip address list as it is pointing to a deprecated list now thanks in advance kr bart van berkel document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login bobbytreed microsoft alias robreed
1
10,228
13,094,448,821
IssuesEvent
2020-08-03 12:25:39
opengeospatial/wps-rest-binding
https://api.github.com/repos/opengeospatial/wps-rest-binding
closed
Migration/support of Process Description using JSON schema
process description
During the Hackaton discussion, we have agreed migrating the JSON Process DEscription to a JSON schema is a great idea. My understanding of the replacement is the following: - Some JSON Schema templates/building blocks would define what the "Process Description" JSON Schema should comply to (simple profile). - In the OGC Processing API, the execute definition request becomes the provided JSON Schema Process Description. How do we define that in the OpenAPI spec ? (generic object type or something smarter ?) The JSON Schema also defines the outputs format, i.e. the jobs/result document I tried to express the process definition (inputs, outputs and metadata) using a JSON Schema definition instead of “process description” traditional document (see below my quick hackaton example). Some issues encountered when mapping Processing Description to JSON Schema: 1. Title and description can only be mapped to the single description property 2. No metadata can be provided. Can we extend JSON Schema ? This is a critical point ! Inputs may require either constraints (spatial, temporal), indication how to retrieve/find resources for this input (e.g. ows Context can provide this information), or client useful information. How can we provide such metadata in a JSON Schema ? If this cannot be achieved, I’m not in favour of adopting the JSON schema format. 3. For LiteralData types, allowed values cannot be provided. An alternative is the pattern regexp (e.g. pattern: '^\d{3}-\d{2}-\d{4}$') 4. Process Description spec includes for each input and output the supported mimeTypes (and default). OpenAPI only allows defining mime type definition for the whole request/response but not for the individual input/output (binary) files. How can we achieve provide this information in the schema itself ? Can we extend JSON schema ? This is a critical feature, and OGC Testbed 15 is exploring how to map compatible processes (matching input/output between processes). Do you have workaround for those issues ? ``` openapi: 3.0.0 info: title: Hackaton OpenAPI Process Description Example version: "Hackaton-processExamplle" description: 'GeomatysNDVIMultiSensor Process (example taken from Testbed 14)' contact: name: Open Geospatial Consortium email: standards@opengeospatial.org url: 'http://www.opengeospatial.org' license: name: CC-BY 4.0 license url: 'https://creativecommons.org/licenses/by/4.0/' components: schemas: execute: type: object required: - mode - response # indicates mandatory inputs - input1 properties: #We provide all the inputs directly as properties, some examples below #Case of an input array of files (binary or documents) input1: type: array description: this is possible to provide the abstract in here # cardinality can be expressed as follows (min/maxOccurs) minItems: 1 maxItems: 10 items: # Using the valueType allow to preserve that the prefered mimeType can be provided # and that the file can be provided raw or by reference (prefered) $ref: 'valueType.yaml' # output 2 and 3 are a simple params (literal data) input2: type: string input3: type: boolean # output 4 is date with temporal constraints # output 5 is integer with constraints input5: type: integer minimum: 1 maximum: 100 default: 20 # output 6 is string with default value and allowed values input6: type: string default: 'test' # How do we specify the required output mime types outputs: type: array items: $ref: '#/components/schemas/output' # If server supports the transmissionMode (sync, async), a conformance req is required mode: $ref: 'transmissionMode.yaml' # If server supports the response format (raw,doc) , a conformance req is required. response: $ref: 'responseMode.yaml' #to be improved result: type: object properties: # defines a single output which is an array of files output1: type: array items: # Using the referenceValue type allows to return the mime type information (raw or by ref) $ref: 'valueType.yaml' ```
1.0
Migration/support of Process Description using JSON schema - During the Hackaton discussion, we have agreed migrating the JSON Process DEscription to a JSON schema is a great idea. My understanding of the replacement is the following: - Some JSON Schema templates/building blocks would define what the "Process Description" JSON Schema should comply to (simple profile). - In the OGC Processing API, the execute definition request becomes the provided JSON Schema Process Description. How do we define that in the OpenAPI spec ? (generic object type or something smarter ?) The JSON Schema also defines the outputs format, i.e. the jobs/result document I tried to express the process definition (inputs, outputs and metadata) using a JSON Schema definition instead of “process description” traditional document (see below my quick hackaton example). Some issues encountered when mapping Processing Description to JSON Schema: 1. Title and description can only be mapped to the single description property 2. No metadata can be provided. Can we extend JSON Schema ? This is a critical point ! Inputs may require either constraints (spatial, temporal), indication how to retrieve/find resources for this input (e.g. ows Context can provide this information), or client useful information. How can we provide such metadata in a JSON Schema ? If this cannot be achieved, I’m not in favour of adopting the JSON schema format. 3. For LiteralData types, allowed values cannot be provided. An alternative is the pattern regexp (e.g. pattern: '^\d{3}-\d{2}-\d{4}$') 4. Process Description spec includes for each input and output the supported mimeTypes (and default). OpenAPI only allows defining mime type definition for the whole request/response but not for the individual input/output (binary) files. How can we achieve provide this information in the schema itself ? Can we extend JSON schema ? This is a critical feature, and OGC Testbed 15 is exploring how to map compatible processes (matching input/output between processes). Do you have workaround for those issues ? ``` openapi: 3.0.0 info: title: Hackaton OpenAPI Process Description Example version: "Hackaton-processExamplle" description: 'GeomatysNDVIMultiSensor Process (example taken from Testbed 14)' contact: name: Open Geospatial Consortium email: standards@opengeospatial.org url: 'http://www.opengeospatial.org' license: name: CC-BY 4.0 license url: 'https://creativecommons.org/licenses/by/4.0/' components: schemas: execute: type: object required: - mode - response # indicates mandatory inputs - input1 properties: #We provide all the inputs directly as properties, some examples below #Case of an input array of files (binary or documents) input1: type: array description: this is possible to provide the abstract in here # cardinality can be expressed as follows (min/maxOccurs) minItems: 1 maxItems: 10 items: # Using the valueType allow to preserve that the prefered mimeType can be provided # and that the file can be provided raw or by reference (prefered) $ref: 'valueType.yaml' # output 2 and 3 are a simple params (literal data) input2: type: string input3: type: boolean # output 4 is date with temporal constraints # output 5 is integer with constraints input5: type: integer minimum: 1 maximum: 100 default: 20 # output 6 is string with default value and allowed values input6: type: string default: 'test' # How do we specify the required output mime types outputs: type: array items: $ref: '#/components/schemas/output' # If server supports the transmissionMode (sync, async), a conformance req is required mode: $ref: 'transmissionMode.yaml' # If server supports the response format (raw,doc) , a conformance req is required. response: $ref: 'responseMode.yaml' #to be improved result: type: object properties: # defines a single output which is an array of files output1: type: array items: # Using the referenceValue type allows to return the mime type information (raw or by ref) $ref: 'valueType.yaml' ```
process
migration support of process description using json schema during the hackaton discussion we have agreed migrating the json process description to a json schema is a great idea my understanding of the replacement is the following some json schema templates building blocks would define what the process description json schema should comply to simple profile in the ogc processing api the execute definition request becomes the provided json schema process description how do we define that in the openapi spec generic object type or something smarter the json schema also defines the outputs format i e the jobs result document i tried to express the process definition inputs outputs and metadata using a json schema definition instead of “process description” traditional document see below my quick hackaton example some issues encountered when mapping processing description to json schema title and description can only be mapped to the single description property no metadata can be provided can we extend json schema this is a critical point inputs may require either constraints spatial temporal indication how to retrieve find resources for this input e g ows context can provide this information or client useful information how can we provide such metadata in a json schema if this cannot be achieved i’m not in favour of adopting the json schema format for literaldata types allowed values cannot be provided an alternative is the pattern regexp e g pattern d d d process description spec includes for each input and output the supported mimetypes and default openapi only allows defining mime type definition for the whole request response but not for the individual input output binary files how can we achieve provide this information in the schema itself can we extend json schema this is a critical feature and ogc testbed is exploring how to map compatible processes matching input output between processes do you have workaround for those issues openapi info title hackaton openapi process description example version hackaton processexamplle description geomatysndvimultisensor process example taken from testbed contact name open geospatial consortium email standards opengeospatial org url license name cc by license url components schemas execute type object required mode response indicates mandatory inputs properties we provide all the inputs directly as properties some examples below case of an input array of files binary or documents type array description this is possible to provide the abstract in here cardinality can be expressed as follows min maxoccurs minitems maxitems items using the valuetype allow to preserve that the prefered mimetype can be provided and that the file can be provided raw or by reference prefered ref valuetype yaml output and are a simple params literal data type string type boolean output is date with temporal constraints output is integer with constraints type integer minimum maximum default output is string with default value and allowed values type string default test how do we specify the required output mime types outputs type array items ref components schemas output if server supports the transmissionmode sync async a conformance req is required mode ref transmissionmode yaml if server supports the response format raw doc a conformance req is required response ref responsemode yaml to be improved result type object properties defines a single output which is an array of files type array items using the referencevalue type allows to return the mime type information raw or by ref ref valuetype yaml
1
45,902
5,988,871,289
IssuesEvent
2017-06-02 06:45:20
geetsisbac/IQ4LK23ROQOKHZC3VKUTGRNI
https://api.github.com/repos/geetsisbac/IQ4LK23ROQOKHZC3VKUTGRNI
reopened
C9sGJ03248TDDO9xOQ9vZLyWe5wnSz96xUm5eW3WTzntq+wrMTG0zUHMYOkUUAmHqw5sZKMGFhXNaH1/ZOIJKMS41rj1FvKovp3inj6/OvWzthT+HYFyUmbDWkQGS01wKlJoW8+W7ijkMLMHxaZ+e/689nLehhr4spCDgK4sqSM=
design
1k7+ImuR0ZyFQv8px5+FF1TWTogREtiwXT7QNu5dDStlcXjwEVAu0cN+j3+c91QN30ZpiEAdV9JR5Eea6iZWSFfpgEmUDBK4DJCF73dOczag2FyZzUKpBeJtpf5tV4q92r1bPK+31Un5em0qMMngg8HsioAl7lAlDNP68Xmgo0/JozqFbBDGiUMgvFISP3JWNMh6lvz18UXaoXAFypSxPPCqJLraj5M4sGOfjVtOjluPwxAMPFelTNyUqwZeD5eOuY/JCXT9sKiwEGadwSBQYkLFJ78lCRq97hVh66IFwHb06DPUGMU9sWN0q3OGgd53SMrLH4JlOUtxFvEG45/znmj7eADhDGbR9quNCHmNP5p1nFTI1vPDwvkcFrQitfWxaT0IBlYeR7uMGh4ZAU5LSipNzw1ODg0IzpLUFi32s+38+eXi8n4o5BsnY2HFL9FzeLjGI3UYH0bJtEZbrU0a/REmoCVBuF0E0PUVBejHITcjLzyj6SyubdwoFih7agrL6pnuhxJkTSfrai+1qONoSvAKBjT/AJEL1eZt2yqTzW/qCKcqYKVICvlnLi/tO7Yo6pnuhxJkTSfrai+1qONoSvAKBjT/AJEL1eZt2yqTzW/PP947HSBsVnT7uZMtB5/gYBfe7sg8dxCCSRqWPMI4w91HrpSWVKFbbZ1svSqztt4P+MmvsYcraz5btfGytOpZlU84e/W1sqmGlQyX0Y5zWX5k0CYg8M9/iUawlKJ/l5sxkLeBdzqqkbS7W7M3fKUQLP4Su62j5iNiHTE12MHl/VaunS/lbR+a5Uw9b/kSFLySkZUCKu591ngbj14hSpXqq1uM5j8DuoDRTH1mWDen/Q==
1.0
C9sGJ03248TDDO9xOQ9vZLyWe5wnSz96xUm5eW3WTzntq+wrMTG0zUHMYOkUUAmHqw5sZKMGFhXNaH1/ZOIJKMS41rj1FvKovp3inj6/OvWzthT+HYFyUmbDWkQGS01wKlJoW8+W7ijkMLMHxaZ+e/689nLehhr4spCDgK4sqSM= - 1k7+ImuR0ZyFQv8px5+FF1TWTogREtiwXT7QNu5dDStlcXjwEVAu0cN+j3+c91QN30ZpiEAdV9JR5Eea6iZWSFfpgEmUDBK4DJCF73dOczag2FyZzUKpBeJtpf5tV4q92r1bPK+31Un5em0qMMngg8HsioAl7lAlDNP68Xmgo0/JozqFbBDGiUMgvFISP3JWNMh6lvz18UXaoXAFypSxPPCqJLraj5M4sGOfjVtOjluPwxAMPFelTNyUqwZeD5eOuY/JCXT9sKiwEGadwSBQYkLFJ78lCRq97hVh66IFwHb06DPUGMU9sWN0q3OGgd53SMrLH4JlOUtxFvEG45/znmj7eADhDGbR9quNCHmNP5p1nFTI1vPDwvkcFrQitfWxaT0IBlYeR7uMGh4ZAU5LSipNzw1ODg0IzpLUFi32s+38+eXi8n4o5BsnY2HFL9FzeLjGI3UYH0bJtEZbrU0a/REmoCVBuF0E0PUVBejHITcjLzyj6SyubdwoFih7agrL6pnuhxJkTSfrai+1qONoSvAKBjT/AJEL1eZt2yqTzW/qCKcqYKVICvlnLi/tO7Yo6pnuhxJkTSfrai+1qONoSvAKBjT/AJEL1eZt2yqTzW/PP947HSBsVnT7uZMtB5/gYBfe7sg8dxCCSRqWPMI4w91HrpSWVKFbbZ1svSqztt4P+MmvsYcraz5btfGytOpZlU84e/W1sqmGlQyX0Y5zWX5k0CYg8M9/iUawlKJ/l5sxkLeBdzqqkbS7W7M3fKUQLP4Su62j5iNiHTE12MHl/VaunS/lbR+a5Uw9b/kSFLySkZUCKu591ngbj14hSpXqq1uM5j8DuoDRTH1mWDen/Q==
non_process
ovwztht e qckcqykvicvlnli iuawlkj vauns lbr q
0
6,846
9,986,794,532
IssuesEvent
2019-07-10 20:02:49
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
reopened
System.ServiceProcess.ServiceController tests fail in non-English locales
area-System.ServiceProcess test bug
Repro: 1. Install Czech language pack and select Czech as Windows display language 2. Run CoreFX System.ServiceProcess.ServiceController tests Result: ``` Starting: System.ServiceProcess.ServiceController.Tests (parallel test collections = on, max threads = 32) System.ServiceProcess.Tests.SafeServiceControllerTests.SetDisplayName_GetServiceName [FAIL] System.InvalidOperationException : Service 'Izolace klíce CNG' was not found on computer '.'. ---- System.ComponentModel.Win32Exception : Zadaná sluzba není nainstalovaná sluzba. Stack Trace: D:\corefx\src\System.ServiceProcess.ServiceController\src\System\ServiceProcess\ServiceController.cs(513,0): at System.ServiceProcess.ServiceController.GenerateNames() D:\corefx\src\System.ServiceProcess.ServiceController\src\System\ServiceProcess\ServiceController.cs(245,0): at System.ServiceProcess.ServiceController.get_ServiceName() D:\corefx\src\System.ServiceProcess.ServiceController\tests\SafeServiceControllerTests.cs(176,0): at System.ServiceProcess.Tests.SafeServiceControllerTests.SetDisplayName_GetServiceName() ----- Inner Stack Trace ----- System.ServiceProcess.Tests.SafeServiceControllerTests.InitializeDisplayName_GetServiceName [FAIL] System.InvalidOperationException : Service 'Izolace klíce CNG' was not found on computer '.'. ---- System.ComponentModel.Win32Exception : Zadaná sluzba není nainstalovaná sluzba. Stack Trace: D:\corefx\src\System.ServiceProcess.ServiceController\src\System\ServiceProcess\ServiceController.cs(513,0): at System.ServiceProcess.ServiceController.GenerateNames() D:\corefx\src\System.ServiceProcess.ServiceController\src\System\ServiceProcess\ServiceController.cs(245,0): at System.ServiceProcess.ServiceController.get_ServiceName() D:\corefx\src\System.ServiceProcess.ServiceController\tests\SafeServiceControllerTests.cs(193,0): at System.ServiceProcess.Tests.SafeServiceControllerTests.InitializeDisplayName_GetServiceName() ----- Inner Stack Trace ----- Finished: System.ServiceProcess.ServiceController.Tests === TEST EXECUTION SUMMARY === System.ServiceProcess.ServiceController.Tests Total: 19, Errors: 0, Failed: 2, Skipped: 0, Time: 2,818s ```
1.0
System.ServiceProcess.ServiceController tests fail in non-English locales - Repro: 1. Install Czech language pack and select Czech as Windows display language 2. Run CoreFX System.ServiceProcess.ServiceController tests Result: ``` Starting: System.ServiceProcess.ServiceController.Tests (parallel test collections = on, max threads = 32) System.ServiceProcess.Tests.SafeServiceControllerTests.SetDisplayName_GetServiceName [FAIL] System.InvalidOperationException : Service 'Izolace klíce CNG' was not found on computer '.'. ---- System.ComponentModel.Win32Exception : Zadaná sluzba není nainstalovaná sluzba. Stack Trace: D:\corefx\src\System.ServiceProcess.ServiceController\src\System\ServiceProcess\ServiceController.cs(513,0): at System.ServiceProcess.ServiceController.GenerateNames() D:\corefx\src\System.ServiceProcess.ServiceController\src\System\ServiceProcess\ServiceController.cs(245,0): at System.ServiceProcess.ServiceController.get_ServiceName() D:\corefx\src\System.ServiceProcess.ServiceController\tests\SafeServiceControllerTests.cs(176,0): at System.ServiceProcess.Tests.SafeServiceControllerTests.SetDisplayName_GetServiceName() ----- Inner Stack Trace ----- System.ServiceProcess.Tests.SafeServiceControllerTests.InitializeDisplayName_GetServiceName [FAIL] System.InvalidOperationException : Service 'Izolace klíce CNG' was not found on computer '.'. ---- System.ComponentModel.Win32Exception : Zadaná sluzba není nainstalovaná sluzba. Stack Trace: D:\corefx\src\System.ServiceProcess.ServiceController\src\System\ServiceProcess\ServiceController.cs(513,0): at System.ServiceProcess.ServiceController.GenerateNames() D:\corefx\src\System.ServiceProcess.ServiceController\src\System\ServiceProcess\ServiceController.cs(245,0): at System.ServiceProcess.ServiceController.get_ServiceName() D:\corefx\src\System.ServiceProcess.ServiceController\tests\SafeServiceControllerTests.cs(193,0): at System.ServiceProcess.Tests.SafeServiceControllerTests.InitializeDisplayName_GetServiceName() ----- Inner Stack Trace ----- Finished: System.ServiceProcess.ServiceController.Tests === TEST EXECUTION SUMMARY === System.ServiceProcess.ServiceController.Tests Total: 19, Errors: 0, Failed: 2, Skipped: 0, Time: 2,818s ```
process
system serviceprocess servicecontroller tests fail in non english locales repro install czech language pack and select czech as windows display language run corefx system serviceprocess servicecontroller tests result starting system serviceprocess servicecontroller tests parallel test collections on max threads system serviceprocess tests safeservicecontrollertests setdisplayname getservicename system invalidoperationexception service izolace klíce cng was not found on computer system componentmodel zadaná sluzba není nainstalovaná sluzba stack trace d corefx src system serviceprocess servicecontroller src system serviceprocess servicecontroller cs at system serviceprocess servicecontroller generatenames d corefx src system serviceprocess servicecontroller src system serviceprocess servicecontroller cs at system serviceprocess servicecontroller get servicename d corefx src system serviceprocess servicecontroller tests safeservicecontrollertests cs at system serviceprocess tests safeservicecontrollertests setdisplayname getservicename inner stack trace system serviceprocess tests safeservicecontrollertests initializedisplayname getservicename system invalidoperationexception service izolace klíce cng was not found on computer system componentmodel zadaná sluzba není nainstalovaná sluzba stack trace d corefx src system serviceprocess servicecontroller src system serviceprocess servicecontroller cs at system serviceprocess servicecontroller generatenames d corefx src system serviceprocess servicecontroller src system serviceprocess servicecontroller cs at system serviceprocess servicecontroller get servicename d corefx src system serviceprocess servicecontroller tests safeservicecontrollertests cs at system serviceprocess tests safeservicecontrollertests initializedisplayname getservicename inner stack trace finished system serviceprocess servicecontroller tests test execution summary system serviceprocess servicecontroller tests total errors failed skipped time
1
158,269
13,728,211,431
IssuesEvent
2020-10-04 10:31:56
appsmithorg/appsmith
https://api.github.com/repos/appsmithorg/appsmith
closed
[Docs] Change docker section steps number to correct values
Documentation
## Documentation Link https://docs.appsmith.com/quick-start#docker ## Describe the problem The steps mentioned in docker section shows step number as 1 for all steps mentioned in this section. ## Describe the improvement Changing the number of the step to 1,2,3 and 4 would be more readable and understandable.
1.0
[Docs] Change docker section steps number to correct values - ## Documentation Link https://docs.appsmith.com/quick-start#docker ## Describe the problem The steps mentioned in docker section shows step number as 1 for all steps mentioned in this section. ## Describe the improvement Changing the number of the step to 1,2,3 and 4 would be more readable and understandable.
non_process
change docker section steps number to correct values documentation link describe the problem the steps mentioned in docker section shows step number as for all steps mentioned in this section describe the improvement changing the number of the step to and would be more readable and understandable
0
10,701
13,495,889,234
IssuesEvent
2020-09-12 01:21:42
jayspur11/Arbeiterbiene
https://api.github.com/repos/jayspur11/Arbeiterbiene
closed
Subclass discord.Client rather than using a Bot
process
We aren't taking advantage of the capabilities introduced by the Bot class, so there's no sense in using it. We should subclass Client instead, which will give us a cleaner main file.
1.0
Subclass discord.Client rather than using a Bot - We aren't taking advantage of the capabilities introduced by the Bot class, so there's no sense in using it. We should subclass Client instead, which will give us a cleaner main file.
process
subclass discord client rather than using a bot we aren t taking advantage of the capabilities introduced by the bot class so there s no sense in using it we should subclass client instead which will give us a cleaner main file
1
2,486
5,265,988,340
IssuesEvent
2017-02-04 07:07:43
mitchellh/packer
https://api.github.com/repos/mitchellh/packer
closed
Packer crash on VMWare build of a Windows 2012R2 image
bug crash post-processor/compress windows-guest
Failed while compressing disks. [crash.log.txt](https://github.com/mitchellh/packer/files/32853/crash.log.txt) ``` json { "builders": [ { "type": "vmware-iso", "iso_url": "{{user `iso_url`}}", "iso_checksum_type": "{{user `iso_checksum_type`}}", "iso_checksum": "{{user `iso_checksum`}}", "headless": false, "boot_wait": "2m", "ssh_username": "vagrant", "ssh_password": "vagrant", "ssh_wait_timeout": "6h", "shutdown_command": "shutdown /s /t 10 /f /d p:4:1 /c \"Packer Shutdown\"", "guest_os_type": "windows8srv-64", "tools_upload_flavor": "windows", "disk_size": 25600, "vnc_port_min": 5900, "vnc_port_max": 5980, "floppy_files": [ "{{user `autounattend`}}", "./scripts/microsoft-updates.bat", "./scripts/win-updates.ps1", "./scripts/openssh.ps1" ], "vmx_data": { "RemoteDisplay.vnc.enabled": "false", "RemoteDisplay.vnc.port": "5900", "memsize": "2048", "numvcpus": "2", "scsi0.virtualDev": "lsisas1068" } }, { "type": "virtualbox-iso", "iso_url": "{{user `iso_url`}}", "iso_checksum_type": "{{user `iso_checksum_type`}}", "iso_checksum": "{{user `iso_checksum`}}", "headless": false, "boot_wait": "2m", "ssh_username": "vagrant", "ssh_password": "vagrant", "ssh_wait_timeout": "6h", "shutdown_command": "shutdown /s /t 10 /f /d p:4:1 /c \"Packer Shutdown\"", "guest_os_type": "Windows2012_64", "disk_size": 25600, "floppy_files": [ "{{user `autounattend`}}", "./scripts/microsoft-updates.bat", "./scripts/win-updates.ps1", "./scripts/openssh.ps1", "./scripts/oracle-cert.cer" ], "vboxmanage": [ [ "modifyvm", "{{.Name}}", "--memory", "2048" ], [ "modifyvm", "{{.Name}}", "--cpus", "2" ] ] } ], "provisioners": [ { "type": "shell", "remote_path": "/tmp/script.bat", "execute_command": "{{.Vars}} cmd /c C:/Windows/Temp/script.bat", "scripts": [ "./scripts/vm-guest-tools.bat", "./scripts/vagrant-ssh.bat", "./scripts/enable-rdp.bat", "./scripts/compile-dotnet-assemblies.bat", "./scripts/disable-auto-logon.bat", "./scripts/compact.bat" ] } ], "post-processors": [ { "type": "vagrant", "keep_input_artifact": false, "output": "windows_2012_r2_{{.Provider}}.box", "vagrantfile_template": "vagrantfile-windows_2012_r2.template" } ], "variables": { "iso_url": "iso/2012_r2_eval.iso", "iso_checksum_type": "md5", "iso_checksum": "458ff91f8abc21b75cb544744bf92e6a", "autounattend": "./answer_files/2012_r2/Autounattend.xml" } } ```
1.0
Packer crash on VMWare build of a Windows 2012R2 image - Failed while compressing disks. [crash.log.txt](https://github.com/mitchellh/packer/files/32853/crash.log.txt) ``` json { "builders": [ { "type": "vmware-iso", "iso_url": "{{user `iso_url`}}", "iso_checksum_type": "{{user `iso_checksum_type`}}", "iso_checksum": "{{user `iso_checksum`}}", "headless": false, "boot_wait": "2m", "ssh_username": "vagrant", "ssh_password": "vagrant", "ssh_wait_timeout": "6h", "shutdown_command": "shutdown /s /t 10 /f /d p:4:1 /c \"Packer Shutdown\"", "guest_os_type": "windows8srv-64", "tools_upload_flavor": "windows", "disk_size": 25600, "vnc_port_min": 5900, "vnc_port_max": 5980, "floppy_files": [ "{{user `autounattend`}}", "./scripts/microsoft-updates.bat", "./scripts/win-updates.ps1", "./scripts/openssh.ps1" ], "vmx_data": { "RemoteDisplay.vnc.enabled": "false", "RemoteDisplay.vnc.port": "5900", "memsize": "2048", "numvcpus": "2", "scsi0.virtualDev": "lsisas1068" } }, { "type": "virtualbox-iso", "iso_url": "{{user `iso_url`}}", "iso_checksum_type": "{{user `iso_checksum_type`}}", "iso_checksum": "{{user `iso_checksum`}}", "headless": false, "boot_wait": "2m", "ssh_username": "vagrant", "ssh_password": "vagrant", "ssh_wait_timeout": "6h", "shutdown_command": "shutdown /s /t 10 /f /d p:4:1 /c \"Packer Shutdown\"", "guest_os_type": "Windows2012_64", "disk_size": 25600, "floppy_files": [ "{{user `autounattend`}}", "./scripts/microsoft-updates.bat", "./scripts/win-updates.ps1", "./scripts/openssh.ps1", "./scripts/oracle-cert.cer" ], "vboxmanage": [ [ "modifyvm", "{{.Name}}", "--memory", "2048" ], [ "modifyvm", "{{.Name}}", "--cpus", "2" ] ] } ], "provisioners": [ { "type": "shell", "remote_path": "/tmp/script.bat", "execute_command": "{{.Vars}} cmd /c C:/Windows/Temp/script.bat", "scripts": [ "./scripts/vm-guest-tools.bat", "./scripts/vagrant-ssh.bat", "./scripts/enable-rdp.bat", "./scripts/compile-dotnet-assemblies.bat", "./scripts/disable-auto-logon.bat", "./scripts/compact.bat" ] } ], "post-processors": [ { "type": "vagrant", "keep_input_artifact": false, "output": "windows_2012_r2_{{.Provider}}.box", "vagrantfile_template": "vagrantfile-windows_2012_r2.template" } ], "variables": { "iso_url": "iso/2012_r2_eval.iso", "iso_checksum_type": "md5", "iso_checksum": "458ff91f8abc21b75cb544744bf92e6a", "autounattend": "./answer_files/2012_r2/Autounattend.xml" } } ```
process
packer crash on vmware build of a windows image failed while compressing disks json builders type vmware iso iso url user iso url iso checksum type user iso checksum type iso checksum user iso checksum headless false boot wait ssh username vagrant ssh password vagrant ssh wait timeout shutdown command shutdown s t f d p c packer shutdown guest os type tools upload flavor windows disk size vnc port min vnc port max floppy files user autounattend scripts microsoft updates bat scripts win updates scripts openssh vmx data remotedisplay vnc enabled false remotedisplay vnc port memsize numvcpus virtualdev type virtualbox iso iso url user iso url iso checksum type user iso checksum type iso checksum user iso checksum headless false boot wait ssh username vagrant ssh password vagrant ssh wait timeout shutdown command shutdown s t f d p c packer shutdown guest os type disk size floppy files user autounattend scripts microsoft updates bat scripts win updates scripts openssh scripts oracle cert cer vboxmanage modifyvm name memory modifyvm name cpus provisioners type shell remote path tmp script bat execute command vars cmd c c windows temp script bat scripts scripts vm guest tools bat scripts vagrant ssh bat scripts enable rdp bat scripts compile dotnet assemblies bat scripts disable auto logon bat scripts compact bat post processors type vagrant keep input artifact false output windows provider box vagrantfile template vagrantfile windows template variables iso url iso eval iso iso checksum type iso checksum autounattend answer files autounattend xml
1
293
2,732,218,935
IssuesEvent
2015-04-17 03:00:25
mitchellh/packer
https://api.github.com/repos/mitchellh/packer
closed
Atlas provisioner fails when using vmware and virtualbox as provisioners
bug post-processor/atlas post-processor/vagrant
```shell virtualbox-iso (vagrant-cloud): Box accessible and matches tag ==> virtualbox-iso (vagrant-cloud): Creating version: 0.1 ==> virtualbox-iso (vagrant-cloud): Creating provider: virtualbox ==> virtualbox-iso (vagrant-cloud): Preparing upload of box: centos7_64virtualbox.box virtualbox-iso (vagrant-cloud): Box upload prepared with token 31564a9c-b1fc-4b6e-a3f2-ea6fdcf2e56b ==> virtualbox-iso (vagrant-cloud): Uploading box: centos7_64virtualbox.box virtualbox-iso (vagrant-cloud): Depending on your internet connection and the size of the box, this may take some time vmware-iso (vagrant): Compressing: disk-s007.vmdk vmware-iso (vagrant): Compressing: disk-s008.vmdk virtualbox-iso (vagrant-cloud): Box succesfully uploaded ==> virtualbox-iso (vagrant-cloud): Verifying provider upload: virtualbox virtualbox-iso (vagrant-cloud): Waiting for upload token match virtualbox-iso (vagrant-cloud): Upload succesfully verified with token 31564a9c-b1fc-4b6e-a3f2-ea6fdcf2e56b ==> virtualbox-iso (vagrant-cloud): Releasing version: 0.1 virtualbox-iso (vagrant-cloud): Version successfully released and available Build 'virtualbox-iso' finished. vmware-iso (vagrant): Compressing: disk-s009.vmdk vmware-iso (vagrant): Compressing: disk-s010.vmdk vmware-iso (vagrant): Compressing: disk-s011.vmdk vmware-iso (vagrant): Compressing: disk.vmdk vmware-iso (vagrant): Compressing: metadata.json ==> vmware-iso: Running post-processor: vagrant-cloud ==> vmware-iso (vagrant-cloud): Verifying box is accessible: lmayorga1980/centos7-puppet vmware-iso (vagrant-cloud): Box accessible and matches tag ==> vmware-iso (vagrant-cloud): Creating version: 0.1 vmware-iso (vagrant-cloud): Version exists, skipping creation ==> vmware-iso (vagrant-cloud): Creating provider: vmware_desktop ==> vmware-iso (vagrant-cloud): Preparing upload of box: centos7_64vmware.box vmware-iso (vagrant-cloud): Box upload prepared with token 0739db2a-390a-4c27-a57e-0f57a03f1366 ==> vmware-iso (vagrant-cloud): Uploading box: centos7_64vmware.box vmware-iso (vagrant-cloud): Depending on your internet connection and the size of the box, this may take some time vmware-iso (vagrant-cloud): Box succesfully uploaded ==> vmware-iso (vagrant-cloud): Verifying provider upload: vmware_desktop vmware-iso (vagrant-cloud): Waiting for upload token match vmware-iso (vagrant-cloud): Upload succesfully verified with token 0739db2a-390a-4c27-a57e-0f57a03f1366 ==> vmware-iso (vagrant-cloud): Releasing version: 0.1 ==> vmware-iso (vagrant-cloud): Cleaning up provider vmware-iso (vagrant-cloud): Deleting provider: vmware_desktop vmware-iso (vagrant-cloud): Version was not created or previously existed, not deleting Build 'vmware-iso' errored: 1 error(s) occurred: * Post-processor failed: Error releasing version: base Version has already been released ```
2.0
Atlas provisioner fails when using vmware and virtualbox as provisioners - ```shell virtualbox-iso (vagrant-cloud): Box accessible and matches tag ==> virtualbox-iso (vagrant-cloud): Creating version: 0.1 ==> virtualbox-iso (vagrant-cloud): Creating provider: virtualbox ==> virtualbox-iso (vagrant-cloud): Preparing upload of box: centos7_64virtualbox.box virtualbox-iso (vagrant-cloud): Box upload prepared with token 31564a9c-b1fc-4b6e-a3f2-ea6fdcf2e56b ==> virtualbox-iso (vagrant-cloud): Uploading box: centos7_64virtualbox.box virtualbox-iso (vagrant-cloud): Depending on your internet connection and the size of the box, this may take some time vmware-iso (vagrant): Compressing: disk-s007.vmdk vmware-iso (vagrant): Compressing: disk-s008.vmdk virtualbox-iso (vagrant-cloud): Box succesfully uploaded ==> virtualbox-iso (vagrant-cloud): Verifying provider upload: virtualbox virtualbox-iso (vagrant-cloud): Waiting for upload token match virtualbox-iso (vagrant-cloud): Upload succesfully verified with token 31564a9c-b1fc-4b6e-a3f2-ea6fdcf2e56b ==> virtualbox-iso (vagrant-cloud): Releasing version: 0.1 virtualbox-iso (vagrant-cloud): Version successfully released and available Build 'virtualbox-iso' finished. vmware-iso (vagrant): Compressing: disk-s009.vmdk vmware-iso (vagrant): Compressing: disk-s010.vmdk vmware-iso (vagrant): Compressing: disk-s011.vmdk vmware-iso (vagrant): Compressing: disk.vmdk vmware-iso (vagrant): Compressing: metadata.json ==> vmware-iso: Running post-processor: vagrant-cloud ==> vmware-iso (vagrant-cloud): Verifying box is accessible: lmayorga1980/centos7-puppet vmware-iso (vagrant-cloud): Box accessible and matches tag ==> vmware-iso (vagrant-cloud): Creating version: 0.1 vmware-iso (vagrant-cloud): Version exists, skipping creation ==> vmware-iso (vagrant-cloud): Creating provider: vmware_desktop ==> vmware-iso (vagrant-cloud): Preparing upload of box: centos7_64vmware.box vmware-iso (vagrant-cloud): Box upload prepared with token 0739db2a-390a-4c27-a57e-0f57a03f1366 ==> vmware-iso (vagrant-cloud): Uploading box: centos7_64vmware.box vmware-iso (vagrant-cloud): Depending on your internet connection and the size of the box, this may take some time vmware-iso (vagrant-cloud): Box succesfully uploaded ==> vmware-iso (vagrant-cloud): Verifying provider upload: vmware_desktop vmware-iso (vagrant-cloud): Waiting for upload token match vmware-iso (vagrant-cloud): Upload succesfully verified with token 0739db2a-390a-4c27-a57e-0f57a03f1366 ==> vmware-iso (vagrant-cloud): Releasing version: 0.1 ==> vmware-iso (vagrant-cloud): Cleaning up provider vmware-iso (vagrant-cloud): Deleting provider: vmware_desktop vmware-iso (vagrant-cloud): Version was not created or previously existed, not deleting Build 'vmware-iso' errored: 1 error(s) occurred: * Post-processor failed: Error releasing version: base Version has already been released ```
process
atlas provisioner fails when using vmware and virtualbox as provisioners shell virtualbox iso vagrant cloud box accessible and matches tag virtualbox iso vagrant cloud creating version virtualbox iso vagrant cloud creating provider virtualbox virtualbox iso vagrant cloud preparing upload of box box virtualbox iso vagrant cloud box upload prepared with token virtualbox iso vagrant cloud uploading box box virtualbox iso vagrant cloud depending on your internet connection and the size of the box this may take some time vmware iso vagrant compressing disk vmdk vmware iso vagrant compressing disk vmdk virtualbox iso vagrant cloud box succesfully uploaded virtualbox iso vagrant cloud verifying provider upload virtualbox virtualbox iso vagrant cloud waiting for upload token match virtualbox iso vagrant cloud upload succesfully verified with token virtualbox iso vagrant cloud releasing version virtualbox iso vagrant cloud version successfully released and available build virtualbox iso finished vmware iso vagrant compressing disk vmdk vmware iso vagrant compressing disk vmdk vmware iso vagrant compressing disk vmdk vmware iso vagrant compressing disk vmdk vmware iso vagrant compressing metadata json vmware iso running post processor vagrant cloud vmware iso vagrant cloud verifying box is accessible puppet vmware iso vagrant cloud box accessible and matches tag vmware iso vagrant cloud creating version vmware iso vagrant cloud version exists skipping creation vmware iso vagrant cloud creating provider vmware desktop vmware iso vagrant cloud preparing upload of box box vmware iso vagrant cloud box upload prepared with token vmware iso vagrant cloud uploading box box vmware iso vagrant cloud depending on your internet connection and the size of the box this may take some time vmware iso vagrant cloud box succesfully uploaded vmware iso vagrant cloud verifying provider upload vmware desktop vmware iso vagrant cloud waiting for upload token match vmware iso vagrant cloud upload succesfully verified with token vmware iso vagrant cloud releasing version vmware iso vagrant cloud cleaning up provider vmware iso vagrant cloud deleting provider vmware desktop vmware iso vagrant cloud version was not created or previously existed not deleting build vmware iso errored error s occurred post processor failed error releasing version base version has already been released
1
15,152
18,908,319,466
IssuesEvent
2021-11-16 11:27:58
paul-buerkner/brms
https://api.github.com/repos/paul-buerkner/brms
closed
Unable to extract conditional smooths for factor by smooths
bug post-processing
It seems that it isn't possible to specify only factor by smooths in a GAM fitted with `brm()` using `conditional_smooths()` `smooths` argument. Here's a worked example: ```r # packages library('mgcv') library('brms') # simulate data set.seed(10) df <- gamSim(4, n = 400) # factor by model # fit model b <- brm(bf(y ~ fac + s(x2,by=fac)), data = df, chains = 4, cores = 4, iter = 3000, warmup = 1000, backend = "cmdstanr") # conditional smooths for selected smooth conditional_smooths(b, smooth = "s(x2)") # didn't expect this to work conditional_smooths(b, smooth = "s(x2):fac1") # mgcv notation - didn't expect this to work conditional_smooths(b, smooth = "s(x2, by = fac)") # expected this to work conditional_smooths(b, smooth = "s(x2,by=fac)") # expected this to work ``` Each of the four incantations for `conditional_smooths()` results in: ``` r$> conditional_smooths(b, smooth = "s(x2,by=fac)") Error: No valid smooth terms found in the model. ``` Expected behaviour: the selected smooth is identified and returned for **at least one of** the last two examples. I don't believe this can ever work given the current approach used in the code: 1. the formula or brmterms always contains a nicely spaced out formula for the smooth terms. I specifically used poor style in writing the `s(x2,by=fac))` term but this results in `~s(x2, by = fac)` in the `$formula` component. I can't change this by varying how I fit the model so this seems baked in somewhere, either in brms or in mgcv. 2. In `conditional_smooths.brmsfit` the user-provided `smooths` argument is processed smooths <- rm_wsp(as.character(smooths)) this removes the white space in `"s(x2, by = fac)"` resulting in `"s(x2,by=fac)"`. 3. In `conditional_smooths.btl` a comparison is done between the terms as known to the model and what the user passed in. This will always be comparing `"s(x2, by = fac)"` (expected, in `smterms` in the code) vs `"s(x2,by=fac)"` (the result of processing of the `smooths` arg passed by the user). Looking through the code, it seems that `tidy_smef()` calls `trim_wsp()` on the terms, while `conditional_smooths.brmsfit()` calls `rm_wsp()`. `trim_wsp()` leaves spaces in the code (replacing multiple whitespace instances with single spaces), while `rm_wsp()` removes all whitespace. ``` r$> brms:::trim_wsp("s(x, by = fac)") [1] "s(x, by = fac)" r$> brms:::rm_wsp("s(x, by = fac)") [1] "s(x,by=fac)" ``` Whether it would be that simple a fix to trim all whitespace in `tidy_smef()` I don't know :-) The above behaviour is observed with *brms* 2.16.1 and the following session info <details> <summary>session info</summary> ``` ─ Session info ────────────────────────────────────────────────────────────────────────────── hash: flag: St. Kitts & Nevis, fire, older person: medium-light skin tone setting value version R version 4.1.2 (2021-11-01) os Ubuntu 20.04.3 LTS system x86_64, linux-gnu ui X11 language en_GB:en collate en_GB.UTF-8 ctype en_GB.UTF-8 tz Europe/Copenhagen date 2021-11-16 pandoc 2.5 @ /usr/bin/pandoc ─ Packages ─────────────────────────────────────────────────────────────────────────────────── package * version date (UTC) lib source abind 1.4-5 2016-07-21 [1] CRAN (R 4.1.1) assertthat 0.2.1 2019-03-21 [1] CRAN (R 4.1.1) backports 1.3.0 2021-10-27 [1] CRAN (R 4.1.1) base64enc 0.1-3 2015-07-28 [1] CRAN (R 4.1.1) bayesplot 1.8.1 2021-06-14 [3] CRAN (R 4.1.0) boot 1.3-28 2021-05-03 [1] CRAN (R 4.1.1) bridgesampling 1.1-2 2021-04-16 [1] CRAN (R 4.1.1) brms * 2.16.1 2021-08-23 [1] CRAN (R 4.1.2) Brobdingnag 1.2-6 2018-08-13 [1] CRAN (R 4.1.1) callr 3.7.0 2021-04-20 [1] CRAN (R 4.1.1) checkmate 2.0.0 2020-02-06 [1] CRAN (R 4.1.1) cli 3.1.0.9000 2021-11-04 [1] Github (r-lib/cli@1af3d91) cmdstanr * 0.4.0 2021-11-16 [1] local coda 0.19-4 2020-09-30 [1] CRAN (R 4.1.1) codetools 0.2-18 2020-11-04 [1] CRAN (R 4.1.1) colorspace 2.0-2 2021-06-24 [3] CRAN (R 4.1.0) colourpicker 1.1.1 2021-10-04 [3] CRAN (R 4.1.1) crayon 1.4.2 2021-10-29 [1] CRAN (R 4.1.2) crosstalk 1.1.1 2021-01-12 [1] CRAN (R 4.1.1) curl 4.3.2 2021-06-23 [3] CRAN (R 4.1.0) data.table 1.14.2 2021-09-27 [3] CRAN (R 4.1.1) DBI 1.1.1 2021-01-15 [1] CRAN (R 4.1.1) digest 0.6.28 2021-09-23 [3] CRAN (R 4.1.1) distributional 0.2.2 2021-02-02 [1] CRAN (R 4.1.1) dplyr 1.0.7 2021-06-18 [1] CRAN (R 4.1.2) DT 0.19 2021-09-02 [3] CRAN (R 4.1.1) dygraphs 1.1.1.6 2018-07-11 [1] CRAN (R 4.1.1) ellipsis 0.3.2 2021-04-29 [1] CRAN (R 4.1.1) emmeans 1.7.0 2021-09-29 [3] CRAN (R 4.1.1) estimability 1.3 2018-02-11 [1] CRAN (R 4.1.1) fansi 0.5.0 2021-05-25 [3] CRAN (R 4.1.0) farver 2.1.0 2021-02-28 [1] CRAN (R 4.1.1) fastmap 1.1.0 2021-01-25 [1] CRAN (R 4.1.1) gamm4 0.2-6 2020-04-03 [1] CRAN (R 4.1.1) generics 0.1.1 2021-10-25 [1] CRAN (R 4.1.1) ggplot2 3.3.5 2021-06-25 [1] CRAN (R 4.1.2) ggridges 0.5.3 2021-01-08 [1] CRAN (R 4.1.1) glue 1.5.0 2021-11-07 [1] CRAN (R 4.1.2) gratia * 0.6.9300 2021-09-27 [1] local gridExtra 2.3 2017-09-09 [1] CRAN (R 4.1.1) gtable 0.3.0 2019-03-25 [1] CRAN (R 4.1.1) gtools 3.9.2 2021-06-06 [3] CRAN (R 4.1.0) htmltools 0.5.2 2021-08-25 [3] CRAN (R 4.1.1) htmlwidgets 1.5.4 2021-09-08 [3] CRAN (R 4.1.1) httpuv 1.6.3 2021-09-09 [3] CRAN (R 4.1.1) igraph 1.2.8 2021-11-07 [3] CRAN (R 4.1.2) inline 0.3.19 2021-05-31 [3] CRAN (R 4.1.0) jsonlite 1.7.2 2020-12-09 [1] CRAN (R 4.1.1) knitr 1.36 2021-09-29 [3] CRAN (R 4.1.1) later 1.3.0 2021-08-18 [3] CRAN (R 4.1.1) lattice 0.20-45 2021-09-22 [4] CRAN (R 4.1.1) lifecycle 1.0.1 2021-09-24 [1] CRAN (R 4.1.1) lme4 1.1-27.1 2021-06-22 [3] CRAN (R 4.1.0) loo 2.4.1 2020-12-09 [1] CRAN (R 4.1.1) magrittr 2.0.1 2020-11-17 [1] CRAN (R 4.1.1) markdown 1.1 2019-08-07 [1] CRAN (R 4.1.1) MASS 7.3-54 2021-05-03 [1] CRAN (R 4.1.1) Matrix 1.3-4 2021-06-01 [4] CRAN (R 4.1.0) matrixStats 0.61.0 2021-09-17 [3] CRAN (R 4.1.1) mgcv * 1.8-38 2021-10-06 [4] CRAN (R 4.1.1) mime 0.12 2021-09-28 [3] CRAN (R 4.1.1) miniUI 0.1.1.1 2018-05-18 [1] CRAN (R 4.1.1) minqa 1.2.4 2014-10-09 [1] CRAN (R 4.1.1) munsell 0.5.0 2018-06-12 [1] CRAN (R 4.1.1) mvnfast 0.2.7 2021-05-20 [3] CRAN (R 4.1.0) mvtnorm 1.1-3 2021-10-08 [3] CRAN (R 4.1.1) nlme * 3.1-153 2021-09-07 [1] CRAN (R 4.1.1) nloptr 1.2.2.3 2021-11-02 [1] CRAN (R 4.1.2) patchwork 1.1.1 2020-12-17 [1] CRAN (R 4.1.1) pillar 1.6.4 2021-10-18 [1] CRAN (R 4.1.1) pkgbuild 1.2.0 2020-12-15 [1] CRAN (R 4.1.1) pkgconfig 2.0.3 2019-09-22 [1] CRAN (R 4.1.1) plyr 1.8.6 2020-03-03 [1] CRAN (R 4.1.1) posterior 1.1.0 2021-09-09 [3] CRAN (R 4.1.1) prettyunits 1.1.1 2020-01-24 [1] CRAN (R 4.1.1) processx 3.5.2 2021-04-30 [1] CRAN (R 4.1.1) projpred 2.0.2 2020-10-28 [1] CRAN (R 4.1.1) promises 1.2.0.1 2021-02-11 [1] CRAN (R 4.1.1) ps 1.6.0 2021-02-28 [1] CRAN (R 4.1.1) purrr 0.3.4 2020-04-17 [1] CRAN (R 4.1.1) R6 2.5.1 2021-08-19 [1] CRAN (R 4.1.1) Rcpp * 1.0.7 2021-07-07 [1] CRAN (R 4.1.1) RcppParallel 5.1.4 2021-05-04 [1] CRAN (R 4.1.1) reshape2 1.4.4 2020-04-09 [1] CRAN (R 4.1.1) rlang 0.4.12 2021-10-18 [3] CRAN (R 4.1.1) rsconnect 0.8.24 2021-08-05 [3] CRAN (R 4.1.0) rstan 2.21.2 2020-07-27 [3] CRAN (R 4.1.0) rstantools 2.1.1 2020-07-06 [1] CRAN (R 4.1.1) scales 1.1.1 2020-05-11 [1] CRAN (R 4.1.1) sessioninfo 1.2.1 2021-11-02 [1] CRAN (R 4.1.2) shiny 1.7.1 2021-10-02 [3] CRAN (R 4.1.1) shinyjs 2.0.0 2020-09-09 [1] CRAN (R 4.1.1) shinystan 2.5.0 2018-05-01 [1] CRAN (R 4.1.1) shinythemes 1.2.0 2021-01-25 [1] CRAN (R 4.1.1) StanHeaders 2.21.0-7 2020-12-17 [1] CRAN (R 4.1.1) stringi 1.7.5 2021-10-04 [3] CRAN (R 4.1.1) stringr 1.4.0 2019-02-10 [1] CRAN (R 4.1.1) tensorA 0.36.2 2020-11-19 [1] CRAN (R 4.1.1) threejs 0.3.3 2020-01-21 [1] CRAN (R 4.1.1) tibble 3.1.6 2021-11-07 [3] CRAN (R 4.1.2) tidyr 1.1.4 2021-09-27 [1] CRAN (R 4.1.1) tidyselect 1.1.1 2021-04-30 [1] CRAN (R 4.1.1) utf8 1.2.2 2021-07-24 [3] CRAN (R 4.1.0) V8 3.5.0 2021-11-03 [1] CRAN (R 4.1.2) vctrs 0.3.8 2021-04-29 [1] CRAN (R 4.1.1) withr 2.4.2 2021-04-18 [1] CRAN (R 4.1.1) xfun 0.28 2021-11-04 [1] CRAN (R 4.1.2) xtable 1.8-4 2019-04-21 [1] CRAN (R 4.1.1) xts 0.12.1 2020-09-09 [1] CRAN (R 4.1.1) zoo 1.8-9 2021-03-09 [1] CRAN (R 4.1.1) [1] /home/au690221/R/x86_64-pc-linux-gnu-library/4.1 [2] /usr/local/lib/R/site-library [3] /usr/lib/R/site-library [4] /usr/lib/R/library ────────────────────────────────────────────────────────────────────────────────────────────── ``` </details>
1.0
Unable to extract conditional smooths for factor by smooths - It seems that it isn't possible to specify only factor by smooths in a GAM fitted with `brm()` using `conditional_smooths()` `smooths` argument. Here's a worked example: ```r # packages library('mgcv') library('brms') # simulate data set.seed(10) df <- gamSim(4, n = 400) # factor by model # fit model b <- brm(bf(y ~ fac + s(x2,by=fac)), data = df, chains = 4, cores = 4, iter = 3000, warmup = 1000, backend = "cmdstanr") # conditional smooths for selected smooth conditional_smooths(b, smooth = "s(x2)") # didn't expect this to work conditional_smooths(b, smooth = "s(x2):fac1") # mgcv notation - didn't expect this to work conditional_smooths(b, smooth = "s(x2, by = fac)") # expected this to work conditional_smooths(b, smooth = "s(x2,by=fac)") # expected this to work ``` Each of the four incantations for `conditional_smooths()` results in: ``` r$> conditional_smooths(b, smooth = "s(x2,by=fac)") Error: No valid smooth terms found in the model. ``` Expected behaviour: the selected smooth is identified and returned for **at least one of** the last two examples. I don't believe this can ever work given the current approach used in the code: 1. the formula or brmterms always contains a nicely spaced out formula for the smooth terms. I specifically used poor style in writing the `s(x2,by=fac))` term but this results in `~s(x2, by = fac)` in the `$formula` component. I can't change this by varying how I fit the model so this seems baked in somewhere, either in brms or in mgcv. 2. In `conditional_smooths.brmsfit` the user-provided `smooths` argument is processed smooths <- rm_wsp(as.character(smooths)) this removes the white space in `"s(x2, by = fac)"` resulting in `"s(x2,by=fac)"`. 3. In `conditional_smooths.btl` a comparison is done between the terms as known to the model and what the user passed in. This will always be comparing `"s(x2, by = fac)"` (expected, in `smterms` in the code) vs `"s(x2,by=fac)"` (the result of processing of the `smooths` arg passed by the user). Looking through the code, it seems that `tidy_smef()` calls `trim_wsp()` on the terms, while `conditional_smooths.brmsfit()` calls `rm_wsp()`. `trim_wsp()` leaves spaces in the code (replacing multiple whitespace instances with single spaces), while `rm_wsp()` removes all whitespace. ``` r$> brms:::trim_wsp("s(x, by = fac)") [1] "s(x, by = fac)" r$> brms:::rm_wsp("s(x, by = fac)") [1] "s(x,by=fac)" ``` Whether it would be that simple a fix to trim all whitespace in `tidy_smef()` I don't know :-) The above behaviour is observed with *brms* 2.16.1 and the following session info <details> <summary>session info</summary> ``` ─ Session info ────────────────────────────────────────────────────────────────────────────── hash: flag: St. Kitts & Nevis, fire, older person: medium-light skin tone setting value version R version 4.1.2 (2021-11-01) os Ubuntu 20.04.3 LTS system x86_64, linux-gnu ui X11 language en_GB:en collate en_GB.UTF-8 ctype en_GB.UTF-8 tz Europe/Copenhagen date 2021-11-16 pandoc 2.5 @ /usr/bin/pandoc ─ Packages ─────────────────────────────────────────────────────────────────────────────────── package * version date (UTC) lib source abind 1.4-5 2016-07-21 [1] CRAN (R 4.1.1) assertthat 0.2.1 2019-03-21 [1] CRAN (R 4.1.1) backports 1.3.0 2021-10-27 [1] CRAN (R 4.1.1) base64enc 0.1-3 2015-07-28 [1] CRAN (R 4.1.1) bayesplot 1.8.1 2021-06-14 [3] CRAN (R 4.1.0) boot 1.3-28 2021-05-03 [1] CRAN (R 4.1.1) bridgesampling 1.1-2 2021-04-16 [1] CRAN (R 4.1.1) brms * 2.16.1 2021-08-23 [1] CRAN (R 4.1.2) Brobdingnag 1.2-6 2018-08-13 [1] CRAN (R 4.1.1) callr 3.7.0 2021-04-20 [1] CRAN (R 4.1.1) checkmate 2.0.0 2020-02-06 [1] CRAN (R 4.1.1) cli 3.1.0.9000 2021-11-04 [1] Github (r-lib/cli@1af3d91) cmdstanr * 0.4.0 2021-11-16 [1] local coda 0.19-4 2020-09-30 [1] CRAN (R 4.1.1) codetools 0.2-18 2020-11-04 [1] CRAN (R 4.1.1) colorspace 2.0-2 2021-06-24 [3] CRAN (R 4.1.0) colourpicker 1.1.1 2021-10-04 [3] CRAN (R 4.1.1) crayon 1.4.2 2021-10-29 [1] CRAN (R 4.1.2) crosstalk 1.1.1 2021-01-12 [1] CRAN (R 4.1.1) curl 4.3.2 2021-06-23 [3] CRAN (R 4.1.0) data.table 1.14.2 2021-09-27 [3] CRAN (R 4.1.1) DBI 1.1.1 2021-01-15 [1] CRAN (R 4.1.1) digest 0.6.28 2021-09-23 [3] CRAN (R 4.1.1) distributional 0.2.2 2021-02-02 [1] CRAN (R 4.1.1) dplyr 1.0.7 2021-06-18 [1] CRAN (R 4.1.2) DT 0.19 2021-09-02 [3] CRAN (R 4.1.1) dygraphs 1.1.1.6 2018-07-11 [1] CRAN (R 4.1.1) ellipsis 0.3.2 2021-04-29 [1] CRAN (R 4.1.1) emmeans 1.7.0 2021-09-29 [3] CRAN (R 4.1.1) estimability 1.3 2018-02-11 [1] CRAN (R 4.1.1) fansi 0.5.0 2021-05-25 [3] CRAN (R 4.1.0) farver 2.1.0 2021-02-28 [1] CRAN (R 4.1.1) fastmap 1.1.0 2021-01-25 [1] CRAN (R 4.1.1) gamm4 0.2-6 2020-04-03 [1] CRAN (R 4.1.1) generics 0.1.1 2021-10-25 [1] CRAN (R 4.1.1) ggplot2 3.3.5 2021-06-25 [1] CRAN (R 4.1.2) ggridges 0.5.3 2021-01-08 [1] CRAN (R 4.1.1) glue 1.5.0 2021-11-07 [1] CRAN (R 4.1.2) gratia * 0.6.9300 2021-09-27 [1] local gridExtra 2.3 2017-09-09 [1] CRAN (R 4.1.1) gtable 0.3.0 2019-03-25 [1] CRAN (R 4.1.1) gtools 3.9.2 2021-06-06 [3] CRAN (R 4.1.0) htmltools 0.5.2 2021-08-25 [3] CRAN (R 4.1.1) htmlwidgets 1.5.4 2021-09-08 [3] CRAN (R 4.1.1) httpuv 1.6.3 2021-09-09 [3] CRAN (R 4.1.1) igraph 1.2.8 2021-11-07 [3] CRAN (R 4.1.2) inline 0.3.19 2021-05-31 [3] CRAN (R 4.1.0) jsonlite 1.7.2 2020-12-09 [1] CRAN (R 4.1.1) knitr 1.36 2021-09-29 [3] CRAN (R 4.1.1) later 1.3.0 2021-08-18 [3] CRAN (R 4.1.1) lattice 0.20-45 2021-09-22 [4] CRAN (R 4.1.1) lifecycle 1.0.1 2021-09-24 [1] CRAN (R 4.1.1) lme4 1.1-27.1 2021-06-22 [3] CRAN (R 4.1.0) loo 2.4.1 2020-12-09 [1] CRAN (R 4.1.1) magrittr 2.0.1 2020-11-17 [1] CRAN (R 4.1.1) markdown 1.1 2019-08-07 [1] CRAN (R 4.1.1) MASS 7.3-54 2021-05-03 [1] CRAN (R 4.1.1) Matrix 1.3-4 2021-06-01 [4] CRAN (R 4.1.0) matrixStats 0.61.0 2021-09-17 [3] CRAN (R 4.1.1) mgcv * 1.8-38 2021-10-06 [4] CRAN (R 4.1.1) mime 0.12 2021-09-28 [3] CRAN (R 4.1.1) miniUI 0.1.1.1 2018-05-18 [1] CRAN (R 4.1.1) minqa 1.2.4 2014-10-09 [1] CRAN (R 4.1.1) munsell 0.5.0 2018-06-12 [1] CRAN (R 4.1.1) mvnfast 0.2.7 2021-05-20 [3] CRAN (R 4.1.0) mvtnorm 1.1-3 2021-10-08 [3] CRAN (R 4.1.1) nlme * 3.1-153 2021-09-07 [1] CRAN (R 4.1.1) nloptr 1.2.2.3 2021-11-02 [1] CRAN (R 4.1.2) patchwork 1.1.1 2020-12-17 [1] CRAN (R 4.1.1) pillar 1.6.4 2021-10-18 [1] CRAN (R 4.1.1) pkgbuild 1.2.0 2020-12-15 [1] CRAN (R 4.1.1) pkgconfig 2.0.3 2019-09-22 [1] CRAN (R 4.1.1) plyr 1.8.6 2020-03-03 [1] CRAN (R 4.1.1) posterior 1.1.0 2021-09-09 [3] CRAN (R 4.1.1) prettyunits 1.1.1 2020-01-24 [1] CRAN (R 4.1.1) processx 3.5.2 2021-04-30 [1] CRAN (R 4.1.1) projpred 2.0.2 2020-10-28 [1] CRAN (R 4.1.1) promises 1.2.0.1 2021-02-11 [1] CRAN (R 4.1.1) ps 1.6.0 2021-02-28 [1] CRAN (R 4.1.1) purrr 0.3.4 2020-04-17 [1] CRAN (R 4.1.1) R6 2.5.1 2021-08-19 [1] CRAN (R 4.1.1) Rcpp * 1.0.7 2021-07-07 [1] CRAN (R 4.1.1) RcppParallel 5.1.4 2021-05-04 [1] CRAN (R 4.1.1) reshape2 1.4.4 2020-04-09 [1] CRAN (R 4.1.1) rlang 0.4.12 2021-10-18 [3] CRAN (R 4.1.1) rsconnect 0.8.24 2021-08-05 [3] CRAN (R 4.1.0) rstan 2.21.2 2020-07-27 [3] CRAN (R 4.1.0) rstantools 2.1.1 2020-07-06 [1] CRAN (R 4.1.1) scales 1.1.1 2020-05-11 [1] CRAN (R 4.1.1) sessioninfo 1.2.1 2021-11-02 [1] CRAN (R 4.1.2) shiny 1.7.1 2021-10-02 [3] CRAN (R 4.1.1) shinyjs 2.0.0 2020-09-09 [1] CRAN (R 4.1.1) shinystan 2.5.0 2018-05-01 [1] CRAN (R 4.1.1) shinythemes 1.2.0 2021-01-25 [1] CRAN (R 4.1.1) StanHeaders 2.21.0-7 2020-12-17 [1] CRAN (R 4.1.1) stringi 1.7.5 2021-10-04 [3] CRAN (R 4.1.1) stringr 1.4.0 2019-02-10 [1] CRAN (R 4.1.1) tensorA 0.36.2 2020-11-19 [1] CRAN (R 4.1.1) threejs 0.3.3 2020-01-21 [1] CRAN (R 4.1.1) tibble 3.1.6 2021-11-07 [3] CRAN (R 4.1.2) tidyr 1.1.4 2021-09-27 [1] CRAN (R 4.1.1) tidyselect 1.1.1 2021-04-30 [1] CRAN (R 4.1.1) utf8 1.2.2 2021-07-24 [3] CRAN (R 4.1.0) V8 3.5.0 2021-11-03 [1] CRAN (R 4.1.2) vctrs 0.3.8 2021-04-29 [1] CRAN (R 4.1.1) withr 2.4.2 2021-04-18 [1] CRAN (R 4.1.1) xfun 0.28 2021-11-04 [1] CRAN (R 4.1.2) xtable 1.8-4 2019-04-21 [1] CRAN (R 4.1.1) xts 0.12.1 2020-09-09 [1] CRAN (R 4.1.1) zoo 1.8-9 2021-03-09 [1] CRAN (R 4.1.1) [1] /home/au690221/R/x86_64-pc-linux-gnu-library/4.1 [2] /usr/local/lib/R/site-library [3] /usr/lib/R/site-library [4] /usr/lib/R/library ────────────────────────────────────────────────────────────────────────────────────────────── ``` </details>
process
unable to extract conditional smooths for factor by smooths it seems that it isn t possible to specify only factor by smooths in a gam fitted with brm using conditional smooths smooths argument here s a worked example r packages library mgcv library brms simulate data set seed df gamsim n factor by model fit model b brm bf y fac s by fac data df chains cores iter warmup backend cmdstanr conditional smooths for selected smooth conditional smooths b smooth s didn t expect this to work conditional smooths b smooth s mgcv notation didn t expect this to work conditional smooths b smooth s by fac expected this to work conditional smooths b smooth s by fac expected this to work each of the four incantations for conditional smooths results in r conditional smooths b smooth s by fac error no valid smooth terms found in the model expected behaviour the selected smooth is identified and returned for at least one of the last two examples i don t believe this can ever work given the current approach used in the code the formula or brmterms always contains a nicely spaced out formula for the smooth terms i specifically used poor style in writing the s by fac term but this results in s by fac in the formula component i can t change this by varying how i fit the model so this seems baked in somewhere either in brms or in mgcv in conditional smooths brmsfit the user provided smooths argument is processed smooths rm wsp as character smooths this removes the white space in s by fac resulting in s by fac in conditional smooths btl a comparison is done between the terms as known to the model and what the user passed in this will always be comparing s by fac expected in smterms in the code vs s by fac the result of processing of the smooths arg passed by the user looking through the code it seems that tidy smef calls trim wsp on the terms while conditional smooths brmsfit calls rm wsp trim wsp leaves spaces in the code replacing multiple whitespace instances with single spaces while rm wsp removes all whitespace r brms trim wsp s x by fac s x by fac r brms rm wsp s x by fac s x by fac whether it would be that simple a fix to trim all whitespace in tidy smef i don t know the above behaviour is observed with brms and the following session info session info ─ session info ────────────────────────────────────────────────────────────────────────────── hash flag st kitts nevis fire older person medium light skin tone setting value version r version os ubuntu lts system linux gnu ui language en gb en collate en gb utf ctype en gb utf tz europe copenhagen date pandoc usr bin pandoc ─ packages ─────────────────────────────────────────────────────────────────────────────────── package version date utc lib source abind cran r assertthat cran r backports cran r cran r bayesplot cran r boot cran r bridgesampling cran r brms cran r brobdingnag cran r callr cran r checkmate cran r cli github r lib cli cmdstanr local coda cran r codetools cran r colorspace cran r colourpicker cran r crayon cran r crosstalk cran r curl cran r data table cran r dbi cran r digest cran r distributional cran r dplyr cran r dt cran r dygraphs cran r ellipsis cran r emmeans cran r estimability cran r fansi cran r farver cran r fastmap cran r cran r generics cran r cran r ggridges cran r glue cran r gratia local gridextra cran r gtable cran r gtools cran r htmltools cran r htmlwidgets cran r httpuv cran r igraph cran r inline cran r jsonlite cran r knitr cran r later cran r lattice cran r lifecycle cran r cran r loo cran r magrittr cran r markdown cran r mass cran r matrix cran r matrixstats cran r mgcv cran r mime cran r miniui cran r minqa cran r munsell cran r mvnfast cran r mvtnorm cran r nlme cran r nloptr cran r patchwork cran r pillar cran r pkgbuild cran r pkgconfig cran r plyr cran r posterior cran r prettyunits cran r processx cran r projpred cran r promises cran r ps cran r purrr cran r cran r rcpp cran r rcppparallel cran r cran r rlang cran r rsconnect cran r rstan cran r rstantools cran r scales cran r sessioninfo cran r shiny cran r shinyjs cran r shinystan cran r shinythemes cran r stanheaders cran r stringi cran r stringr cran r tensora cran r threejs cran r tibble cran r tidyr cran r tidyselect cran r cran r cran r vctrs cran r withr cran r xfun cran r xtable cran r xts cran r zoo cran r home r pc linux gnu library usr local lib r site library usr lib r site library usr lib r library ──────────────────────────────────────────────────────────────────────────────────────────────
1
20,564
10,817,353,496
IssuesEvent
2019-11-08 09:33:20
music-stats/map
https://api.github.com/repos/music-stats/map
opened
Use an SVG sprite instead of embedding flag images into JS bundle
performance refactoring
:bicyclist: __Motivation__ Bundle size grows enormously because of those imports. <img width="800" alt="js" src="https://user-images.githubusercontent.com/2470363/68465603-10efb300-0213-11ea-97ec-72431a60dd5c.png"> <img width="800" alt="svg" src="https://user-images.githubusercontent.com/2470363/68465604-10efb300-0213-11ea-8048-2c4f6fa110b2.png"> :card_index: __Considerations__ The sprite could be either built once (containing all country flags) or re-built every time (to include only those countries that actually are represented in the playcount dataset). :clipboard: __Steps__ - [ ] Make the sprite (scriptable). - [ ] Update `<AreaList />` and its props to recognise corresponding offset in the sprite.
True
Use an SVG sprite instead of embedding flag images into JS bundle - :bicyclist: __Motivation__ Bundle size grows enormously because of those imports. <img width="800" alt="js" src="https://user-images.githubusercontent.com/2470363/68465603-10efb300-0213-11ea-97ec-72431a60dd5c.png"> <img width="800" alt="svg" src="https://user-images.githubusercontent.com/2470363/68465604-10efb300-0213-11ea-8048-2c4f6fa110b2.png"> :card_index: __Considerations__ The sprite could be either built once (containing all country flags) or re-built every time (to include only those countries that actually are represented in the playcount dataset). :clipboard: __Steps__ - [ ] Make the sprite (scriptable). - [ ] Update `<AreaList />` and its props to recognise corresponding offset in the sprite.
non_process
use an svg sprite instead of embedding flag images into js bundle bicyclist motivation bundle size grows enormously because of those imports img width alt js src img width alt svg src card index considerations the sprite could be either built once containing all country flags or re built every time to include only those countries that actually are represented in the playcount dataset clipboard steps make the sprite scriptable update and its props to recognise corresponding offset in the sprite
0
21,826
30,317,623,455
IssuesEvent
2023-07-10 16:41:11
tdwg/dwc
https://api.github.com/repos/tdwg/dwc
closed
Make Release 2023-06-28
Process - implement
Ratification is complete. Issues have been addressed. Publish changes to production. Check. Make Release. Create new versions of affected GBIF Extension XML files.
1.0
Make Release 2023-06-28 - Ratification is complete. Issues have been addressed. Publish changes to production. Check. Make Release. Create new versions of affected GBIF Extension XML files.
process
make release ratification is complete issues have been addressed publish changes to production check make release create new versions of affected gbif extension xml files
1
333,201
29,515,350,841
IssuesEvent
2023-06-04 12:30:31
3DAsset-eCommerce/3D-BE
https://api.github.com/repos/3DAsset-eCommerce/3D-BE
opened
[feat] 장바구니 담긴 개수 확인
Available Low Test
## Description >> 장바구니에 담긴 에셋의 개수를 보내주는 기능 ## Tasks - [ ] 사용자의 id와 조회를 요청한 id가 같은지 확인 - [ ] 사용자의 id로 cartRepository에서 count - [ ] Controller Test - [ ] Service Test - [ ] Repository Test - [ ] Integration Test
1.0
[feat] 장바구니 담긴 개수 확인 - ## Description >> 장바구니에 담긴 에셋의 개수를 보내주는 기능 ## Tasks - [ ] 사용자의 id와 조회를 요청한 id가 같은지 확인 - [ ] 사용자의 id로 cartRepository에서 count - [ ] Controller Test - [ ] Service Test - [ ] Repository Test - [ ] Integration Test
non_process
장바구니 담긴 개수 확인 description 장바구니에 담긴 에셋의 개수를 보내주는 기능 tasks 사용자의 id와 조회를 요청한 id가 같은지 확인 사용자의 id로 cartrepository에서 count controller test service test repository test integration test
0
64,059
12,414,969,739
IssuesEvent
2020-05-22 15:24:16
cython/cython
https://api.github.com/repos/cython/cython
opened
cpdef method override check fails for fused methods
Code Generation Cython Language Feature defect
Cython generates some code into cpdef methods that checks if the method has potentially been overridden when calling it directly through the C vtable, and if that's possible, it repeats the call through a Python method lookup. This works for non-heap types without instance dict, but fails otherwise, e.g. for Python subtypes and types created with type specs. For fused methods, this currently generates 1) duplicate code for the override check (probably from two `OverrideCheckNode`s where we should have only one) 2) a method lookup for the internal (mangled) fused function name, which is then not found. 3) a Python call from the specialised method to the generic method, which might then end up calling a different specialisation than what was chosen at compile time. Instead, it should 1) generate only one check 2) look up the normally named method 3a) check if that method is a fused function (if not, call it, since it has definitely been overwritten) 3b) if if is a fused function, look up the concrete specialised method from it (if that fails, call the fused function itself) 3c) then check if that specialised function points to the (running) C function (if not, call it) Along the way, it can use dict version checks for all involved dicts, i.e. the type's dict and the fused method's specialisation dict.
1.0
cpdef method override check fails for fused methods - Cython generates some code into cpdef methods that checks if the method has potentially been overridden when calling it directly through the C vtable, and if that's possible, it repeats the call through a Python method lookup. This works for non-heap types without instance dict, but fails otherwise, e.g. for Python subtypes and types created with type specs. For fused methods, this currently generates 1) duplicate code for the override check (probably from two `OverrideCheckNode`s where we should have only one) 2) a method lookup for the internal (mangled) fused function name, which is then not found. 3) a Python call from the specialised method to the generic method, which might then end up calling a different specialisation than what was chosen at compile time. Instead, it should 1) generate only one check 2) look up the normally named method 3a) check if that method is a fused function (if not, call it, since it has definitely been overwritten) 3b) if if is a fused function, look up the concrete specialised method from it (if that fails, call the fused function itself) 3c) then check if that specialised function points to the (running) C function (if not, call it) Along the way, it can use dict version checks for all involved dicts, i.e. the type's dict and the fused method's specialisation dict.
non_process
cpdef method override check fails for fused methods cython generates some code into cpdef methods that checks if the method has potentially been overridden when calling it directly through the c vtable and if that s possible it repeats the call through a python method lookup this works for non heap types without instance dict but fails otherwise e g for python subtypes and types created with type specs for fused methods this currently generates duplicate code for the override check probably from two overridechecknode s where we should have only one a method lookup for the internal mangled fused function name which is then not found a python call from the specialised method to the generic method which might then end up calling a different specialisation than what was chosen at compile time instead it should generate only one check look up the normally named method check if that method is a fused function if not call it since it has definitely been overwritten if if is a fused function look up the concrete specialised method from it if that fails call the fused function itself then check if that specialised function points to the running c function if not call it along the way it can use dict version checks for all involved dicts i e the type s dict and the fused method s specialisation dict
0
409,123
27,722,987,569
IssuesEvent
2023-03-14 22:27:17
HYF-Class19/RCP-Team
https://api.github.com/repos/HYF-Class19/RCP-Team
closed
01 - coding approach - v1
documentation
Created the first version of the coding approach.Will keep updating it as we will progress
1.0
01 - coding approach - v1 - Created the first version of the coding approach.Will keep updating it as we will progress
non_process
coding approach created the first version of the coding approach will keep updating it as we will progress
0
54,130
13,255,356,859
IssuesEvent
2020-08-20 10:47:37
googleapis/java-datalabeling
https://api.github.com/repos/googleapis/java-datalabeling
opened
datalabeling.v1beta1.it.ITSystemTest: listAnnotationSpecSetsTest failed
buildcop: issue priority: p1 type: bug
This test failed! To configure my behavior, see [the Build Cop Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/buildcop). If I'm commenting on this issue too often, add the `buildcop: quiet` label and I will stop commenting. --- commit: 67890f4f96e028e9b5e727f8465d912d3bf6c214 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/acbf92f7-83d6-416a-96b2-bdc0d69c9008), [Sponge](http://sponge2/acbf92f7-83d6-416a-96b2-bdc0d69c9008) status: failed <details><summary>Test output</summary><br><pre>com.google.api.gax.rpc.InternalException: io.grpc.StatusRuntimeException: INTERNAL: Failed to ListAnnotationSpecSets. at com.google.api.gax.rpc.ApiExceptionFactory.createException(ApiExceptionFactory.java:67) at com.google.api.gax.grpc.GrpcApiExceptionFactory.create(GrpcApiExceptionFactory.java:72) at com.google.api.gax.grpc.GrpcApiExceptionFactory.create(GrpcApiExceptionFactory.java:60) at com.google.api.gax.grpc.GrpcExceptionCallable$ExceptionTransformingFuture.onFailure(GrpcExceptionCallable.java:97) at com.google.api.core.ApiFutures$1.onFailure(ApiFutures.java:68) at com.google.common.util.concurrent.Futures$CallbackListener.run(Futures.java:1050) at com.google.common.util.concurrent.DirectExecutor.execute(DirectExecutor.java:30) at com.google.common.util.concurrent.AbstractFuture.executeListener(AbstractFuture.java:1176) at com.google.common.util.concurrent.AbstractFuture.complete(AbstractFuture.java:969) at com.google.common.util.concurrent.AbstractFuture.setException(AbstractFuture.java:760) at io.grpc.stub.ClientCalls$GrpcFuture.setException(ClientCalls.java:545) at io.grpc.stub.ClientCalls$UnaryStreamToFuture.onClose(ClientCalls.java:515) at io.grpc.internal.ClientCallImpl.closeObserver(ClientCallImpl.java:426) at io.grpc.internal.ClientCallImpl.access$500(ClientCallImpl.java:66) at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl.close(ClientCallImpl.java:689) at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl.access$900(ClientCallImpl.java:577) at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInternal(ClientCallImpl.java:751) at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInContext(ClientCallImpl.java:740) at io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37) at io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:123) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) Suppressed: com.google.api.gax.rpc.AsyncTaskException: Asynchronous task failed at com.google.api.gax.rpc.ApiExceptions.callAndTranslateApiException(ApiExceptions.java:57) at com.google.api.gax.rpc.UnaryCallable.call(UnaryCallable.java:112) at com.google.cloud.datalabeling.v1beta1.DataLabelingServiceClient.listAnnotationSpecSets(DataLabelingServiceClient.java:2432) at com.google.cloud.datalabeling.v1beta1.DataLabelingServiceClient.listAnnotationSpecSets(DataLabelingServiceClient.java:2406) at com.google.cloud.datalabeling.v1beta1.it.ITSystemTest.listAnnotationSpecSetsTest(ITSystemTest.java:137) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.apache.maven.surefire.junitcore.JUnitCore.run(JUnitCore.java:55) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.createRequestAndRun(JUnitCoreWrapper.java:137) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.executeEager(JUnitCoreWrapper.java:107) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:83) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:75) at org.apache.maven.surefire.junitcore.JUnitCoreProvider.invoke(JUnitCoreProvider.java:158) at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:377) at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:138) at org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:465) at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:451) Caused by: io.grpc.StatusRuntimeException: INTERNAL: Failed to ListAnnotationSpecSets. at io.grpc.Status.asRuntimeException(Status.java:533) ... 16 more </pre></details>
1.0
datalabeling.v1beta1.it.ITSystemTest: listAnnotationSpecSetsTest failed - This test failed! To configure my behavior, see [the Build Cop Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/buildcop). If I'm commenting on this issue too often, add the `buildcop: quiet` label and I will stop commenting. --- commit: 67890f4f96e028e9b5e727f8465d912d3bf6c214 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/acbf92f7-83d6-416a-96b2-bdc0d69c9008), [Sponge](http://sponge2/acbf92f7-83d6-416a-96b2-bdc0d69c9008) status: failed <details><summary>Test output</summary><br><pre>com.google.api.gax.rpc.InternalException: io.grpc.StatusRuntimeException: INTERNAL: Failed to ListAnnotationSpecSets. at com.google.api.gax.rpc.ApiExceptionFactory.createException(ApiExceptionFactory.java:67) at com.google.api.gax.grpc.GrpcApiExceptionFactory.create(GrpcApiExceptionFactory.java:72) at com.google.api.gax.grpc.GrpcApiExceptionFactory.create(GrpcApiExceptionFactory.java:60) at com.google.api.gax.grpc.GrpcExceptionCallable$ExceptionTransformingFuture.onFailure(GrpcExceptionCallable.java:97) at com.google.api.core.ApiFutures$1.onFailure(ApiFutures.java:68) at com.google.common.util.concurrent.Futures$CallbackListener.run(Futures.java:1050) at com.google.common.util.concurrent.DirectExecutor.execute(DirectExecutor.java:30) at com.google.common.util.concurrent.AbstractFuture.executeListener(AbstractFuture.java:1176) at com.google.common.util.concurrent.AbstractFuture.complete(AbstractFuture.java:969) at com.google.common.util.concurrent.AbstractFuture.setException(AbstractFuture.java:760) at io.grpc.stub.ClientCalls$GrpcFuture.setException(ClientCalls.java:545) at io.grpc.stub.ClientCalls$UnaryStreamToFuture.onClose(ClientCalls.java:515) at io.grpc.internal.ClientCallImpl.closeObserver(ClientCallImpl.java:426) at io.grpc.internal.ClientCallImpl.access$500(ClientCallImpl.java:66) at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl.close(ClientCallImpl.java:689) at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl.access$900(ClientCallImpl.java:577) at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInternal(ClientCallImpl.java:751) at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInContext(ClientCallImpl.java:740) at io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37) at io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:123) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) Suppressed: com.google.api.gax.rpc.AsyncTaskException: Asynchronous task failed at com.google.api.gax.rpc.ApiExceptions.callAndTranslateApiException(ApiExceptions.java:57) at com.google.api.gax.rpc.UnaryCallable.call(UnaryCallable.java:112) at com.google.cloud.datalabeling.v1beta1.DataLabelingServiceClient.listAnnotationSpecSets(DataLabelingServiceClient.java:2432) at com.google.cloud.datalabeling.v1beta1.DataLabelingServiceClient.listAnnotationSpecSets(DataLabelingServiceClient.java:2406) at com.google.cloud.datalabeling.v1beta1.it.ITSystemTest.listAnnotationSpecSetsTest(ITSystemTest.java:137) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.apache.maven.surefire.junitcore.JUnitCore.run(JUnitCore.java:55) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.createRequestAndRun(JUnitCoreWrapper.java:137) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.executeEager(JUnitCoreWrapper.java:107) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:83) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:75) at org.apache.maven.surefire.junitcore.JUnitCoreProvider.invoke(JUnitCoreProvider.java:158) at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:377) at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:138) at org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:465) at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:451) Caused by: io.grpc.StatusRuntimeException: INTERNAL: Failed to ListAnnotationSpecSets. at io.grpc.Status.asRuntimeException(Status.java:533) ... 16 more </pre></details>
non_process
datalabeling it itsystemtest listannotationspecsetstest failed this test failed to configure my behavior see if i m commenting on this issue too often add the buildcop quiet label and i will stop commenting commit buildurl status failed test output com google api gax rpc internalexception io grpc statusruntimeexception internal failed to listannotationspecsets at com google api gax rpc apiexceptionfactory createexception apiexceptionfactory java at com google api gax grpc grpcapiexceptionfactory create grpcapiexceptionfactory java at com google api gax grpc grpcapiexceptionfactory create grpcapiexceptionfactory java at com google api gax grpc grpcexceptioncallable exceptiontransformingfuture onfailure grpcexceptioncallable java at com google api core apifutures onfailure apifutures java at com google common util concurrent futures callbacklistener run futures java at com google common util concurrent directexecutor execute directexecutor java at com google common util concurrent abstractfuture executelistener abstractfuture java at com google common util concurrent abstractfuture complete abstractfuture java at com google common util concurrent abstractfuture setexception abstractfuture java at io grpc stub clientcalls grpcfuture setexception clientcalls java at io grpc stub clientcalls unarystreamtofuture onclose clientcalls java at io grpc internal clientcallimpl closeobserver clientcallimpl java at io grpc internal clientcallimpl access clientcallimpl java at io grpc internal clientcallimpl clientstreamlistenerimpl close clientcallimpl java at io grpc internal clientcallimpl clientstreamlistenerimpl access clientcallimpl java at io grpc internal clientcallimpl clientstreamlistenerimpl runinternal clientcallimpl java at io grpc internal clientcallimpl clientstreamlistenerimpl runincontext clientcallimpl java at io grpc internal contextrunnable run contextrunnable java at io grpc internal serializingexecutor run serializingexecutor java at java util concurrent executors runnableadapter call executors java at java util concurrent futuretask run futuretask java at java util concurrent scheduledthreadpoolexecutor scheduledfuturetask access scheduledthreadpoolexecutor java at java util concurrent scheduledthreadpoolexecutor scheduledfuturetask run scheduledthreadpoolexecutor java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java suppressed com google api gax rpc asynctaskexception asynchronous task failed at com google api gax rpc apiexceptions callandtranslateapiexception apiexceptions java at com google api gax rpc unarycallable call unarycallable java at com google cloud datalabeling datalabelingserviceclient listannotationspecsets datalabelingserviceclient java at com google cloud datalabeling datalabelingserviceclient listannotationspecsets datalabelingserviceclient java at com google cloud datalabeling it itsystemtest listannotationspecsetstest itsystemtest java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at org junit runners parentrunner evaluate parentrunner java at org junit runners evaluate java at org junit runners parentrunner runleaf parentrunner java at org junit runners runchild java at org junit runners runchild java at org junit runners parentrunner run parentrunner java at org junit runners parentrunner schedule parentrunner java at org junit runners parentrunner runchildren parentrunner java at org junit runners parentrunner access parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit internal runners statements runbefores evaluate runbefores java at org junit internal runners statements runafters evaluate runafters java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner run parentrunner java at org junit runners suite runchild suite java at org junit runners suite runchild suite java at org junit runners parentrunner run parentrunner java at org junit runners parentrunner schedule parentrunner java at org junit runners parentrunner runchildren parentrunner java at org junit runners parentrunner access parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner run parentrunner java at org apache maven surefire junitcore junitcore run junitcore java at org apache maven surefire junitcore junitcorewrapper createrequestandrun junitcorewrapper java at org apache maven surefire junitcore junitcorewrapper executeeager junitcorewrapper java at org apache maven surefire junitcore junitcorewrapper execute junitcorewrapper java at org apache maven surefire junitcore junitcorewrapper execute junitcorewrapper java at org apache maven surefire junitcore junitcoreprovider invoke junitcoreprovider java at org apache maven surefire booter forkedbooter runsuitesinprocess forkedbooter java at org apache maven surefire booter forkedbooter execute forkedbooter java at org apache maven surefire booter forkedbooter run forkedbooter java at org apache maven surefire booter forkedbooter main forkedbooter java caused by io grpc statusruntimeexception internal failed to listannotationspecsets at io grpc status asruntimeexception status java more
0
17,501
23,315,752,731
IssuesEvent
2022-08-08 12:22:46
prisma/prisma
https://api.github.com/repos/prisma/prisma
closed
Introspection crash, `libs\datamodel\connectors\dml\src\model.rs:494:29` (missing PK?)
bug/1-unconfirmed kind/bug process/candidate team/schema topic: prisma db pull topic: Your friendly prisma developers
Hi Prisma Team! Prisma Migrate just crashed. ## Command `db pull` ## Versions | Name | Version | |-------------|--------------------| | Platform | windows | | Node | v16.14.0 | | Prisma CLI | 4.1.0 | | Engine | 8d8414deb360336e4698a65aa45a1fbaf1ce13d8| ## Error ``` Error: [libs\datamodel\connectors\dml\src\model.rs:494:29] Hi there! We've been seeing this error in our error reporting backend, but cannot reproduce it in our own tests. The problem is that we have a primary key in the model `T_Clients` that uses the column `ConsigneeConfirmationDate` which we for some reason don't have in our internal representation. If you see this, could you please file an issue to https://github.com/prisma/prisma so we can discuss about fixing this. -- Your friendly prisma developers. ```
1.0
Introspection crash, `libs\datamodel\connectors\dml\src\model.rs:494:29` (missing PK?) - Hi Prisma Team! Prisma Migrate just crashed. ## Command `db pull` ## Versions | Name | Version | |-------------|--------------------| | Platform | windows | | Node | v16.14.0 | | Prisma CLI | 4.1.0 | | Engine | 8d8414deb360336e4698a65aa45a1fbaf1ce13d8| ## Error ``` Error: [libs\datamodel\connectors\dml\src\model.rs:494:29] Hi there! We've been seeing this error in our error reporting backend, but cannot reproduce it in our own tests. The problem is that we have a primary key in the model `T_Clients` that uses the column `ConsigneeConfirmationDate` which we for some reason don't have in our internal representation. If you see this, could you please file an issue to https://github.com/prisma/prisma so we can discuss about fixing this. -- Your friendly prisma developers. ```
process
introspection crash libs datamodel connectors dml src model rs missing pk hi prisma team prisma migrate just crashed command db pull versions name version platform windows node prisma cli engine error error hi there we ve been seeing this error in our error reporting backend but cannot reproduce it in our own tests the problem is that we have a primary key in the model t clients that uses the column consigneeconfirmationdate which we for some reason don t have in our internal representation if you see this could you please file an issue to so we can discuss about fixing this your friendly prisma developers
1
44,191
2,900,067,611
IssuesEvent
2015-06-17 14:46:26
icatproject/topcat
https://api.github.com/repos/icatproject/topcat
closed
Basic visualize or view data in TopCAT
Icat-TopCat invalid Priority-Medium Type-Enhancement
``` This might use jpowder or it might absorb code from eCat2. ``` Original issue reported on code.google.com by `dr.s.m.fisher` on 2013-04-11 15:51:13
1.0
Basic visualize or view data in TopCAT - ``` This might use jpowder or it might absorb code from eCat2. ``` Original issue reported on code.google.com by `dr.s.m.fisher` on 2013-04-11 15:51:13
non_process
basic visualize or view data in topcat this might use jpowder or it might absorb code from original issue reported on code google com by dr s m fisher on
0
43,105
23,119,423,073
IssuesEvent
2022-07-27 19:48:13
AzureAD/microsoft-authentication-library-for-dotnet
https://api.github.com/repos/AzureAD/microsoft-authentication-library-for-dotnet
closed
[Bug] Perf - MSAL stores authorities as strings and doing too many string operations. Try to use Uri instead.
performance
Refactor AuthorityInfo and related code. We should not be doing so much string operations.
True
[Bug] Perf - MSAL stores authorities as strings and doing too many string operations. Try to use Uri instead. - Refactor AuthorityInfo and related code. We should not be doing so much string operations.
non_process
perf msal stores authorities as strings and doing too many string operations try to use uri instead refactor authorityinfo and related code we should not be doing so much string operations
0
1,780
4,511,915,982
IssuesEvent
2016-09-03 09:41:39
sysown/proxysql
https://api.github.com/repos/sysown/proxysql
closed
Make ProxySQL aware of prepared statements
CONNECTION POOL cxx_pa development GLOBAL MYSQL PROTOCOL QUERY PROCESSOR ROUTING STATISTICS
While support for prepared statements is not planned any soon, ProxySQL should at least be aware of them and disable connection pool once it detects prepared statements,
1.0
Make ProxySQL aware of prepared statements - While support for prepared statements is not planned any soon, ProxySQL should at least be aware of them and disable connection pool once it detects prepared statements,
process
make proxysql aware of prepared statements while support for prepared statements is not planned any soon proxysql should at least be aware of them and disable connection pool once it detects prepared statements
1
10,638
13,446,121,839
IssuesEvent
2020-09-08 12:30:26
MHRA/products
https://api.github.com/repos/MHRA/products
closed
Do not allow duplicate license numbers
EPIC - PARs process
The form (and ideally the backend too) should validate that you haven't entered the same license number twice See point 3 in https://github.com/MHRA/products/issues/845
1.0
Do not allow duplicate license numbers - The form (and ideally the backend too) should validate that you haven't entered the same license number twice See point 3 in https://github.com/MHRA/products/issues/845
process
do not allow duplicate license numbers the form and ideally the backend too should validate that you haven t entered the same license number twice see point in
1
122,541
26,139,327,414
IssuesEvent
2022-12-29 16:08:07
dOpensource/dsiprouter
https://api.github.com/repos/dOpensource/dsiprouter
closed
0.641 Extra Bug Fixes
bug code committed 0.65 Release
Move bug fixes that didn't make it into 0.641 release into 0.642 release. Should be pulled from branch https://github.com/dOpensource/dsiprouter/tree/0.641-bugfixes
1.0
0.641 Extra Bug Fixes - Move bug fixes that didn't make it into 0.641 release into 0.642 release. Should be pulled from branch https://github.com/dOpensource/dsiprouter/tree/0.641-bugfixes
non_process
extra bug fixes move bug fixes that didn t make it into release into release should be pulled from branch
0
6,891
10,029,179,379
IssuesEvent
2019-07-17 13:25:23
habitat-sh/habitat
https://api.github.com/repos/habitat-sh/habitat
closed
Configurable kill signal and timeout
A-process-management A-supervisor C-feature E-less-easy L-rust V-sup X-feature
Right now, Habitat sends TERM to any process it's supervising on stop. Not all applications respond to TERM, some respond to INT or even USR1 for a custom, clean shutdown. I talked with @fnichol about this, and I think we agreed that it makes sense for plans to describe their termination schedule, perhaps like this: ```sh pkg_kill_signal=(INT) ``` Additionally, it would be nice if packages could customize the amount of time they want to wait for a clean shutdown like: ```sh pkg_kill_timeout=(30) ```
1.0
Configurable kill signal and timeout - Right now, Habitat sends TERM to any process it's supervising on stop. Not all applications respond to TERM, some respond to INT or even USR1 for a custom, clean shutdown. I talked with @fnichol about this, and I think we agreed that it makes sense for plans to describe their termination schedule, perhaps like this: ```sh pkg_kill_signal=(INT) ``` Additionally, it would be nice if packages could customize the amount of time they want to wait for a clean shutdown like: ```sh pkg_kill_timeout=(30) ```
process
configurable kill signal and timeout right now habitat sends term to any process it s supervising on stop not all applications respond to term some respond to int or even for a custom clean shutdown i talked with fnichol about this and i think we agreed that it makes sense for plans to describe their termination schedule perhaps like this sh pkg kill signal int additionally it would be nice if packages could customize the amount of time they want to wait for a clean shutdown like sh pkg kill timeout
1
2,872
5,831,189,239
IssuesEvent
2017-05-08 18:42:07
AllenFang/react-bootstrap-table
https://api.github.com/repos/AllenFang/react-bootstrap-table
closed
sizePerPageList
inprocess
@AllenFang When sizePerPageList has a custom option like ALL(100) . When i delete a row ,the size per pagelist dropdown is clearing off instead of showing ALL(99). Let me know if you did get my question.
1.0
sizePerPageList - @AllenFang When sizePerPageList has a custom option like ALL(100) . When i delete a row ,the size per pagelist dropdown is clearing off instead of showing ALL(99). Let me know if you did get my question.
process
sizeperpagelist allenfang when sizeperpagelist has a custom option like all when i delete a row the size per pagelist dropdown is clearing off instead of showing all let me know if you did get my question
1
12,190
14,742,286,595
IssuesEvent
2021-01-07 12:01:55
CGAL/cgal
https://api.github.com/repos/CGAL/cgal
closed
compile openGR and pointmatch wrapper
Pkg::Point_set_processing_3 question
Sorry to bother you. ## Issue Details I have installed CGAL 5.1.1 through **vcpkg** and compiled the example: CGAL-5.0.2-examples-Surface_mesh with Cmake and run it successfully. My aim is to use **OpenGR** and Pointmacher wrapper in CGAl to register point clouds gbobally. So I try to compile the example: Point_set_processing_3. But a compilation error raises. Selecting Windows SDK version 10.0.17763.0 to target Windows 10.0.18363. Visual Leak Detector (VLD) is not found. Using header-only CGAL Targetting Visual Studio 15 2017 Target build enviroment supports auto-linking Using VC toolset 141. Generator uses intermediate configuration directory: $(Configuration) Boost include dirs: E:/vcpkg/installed/x64-windows/include Boost libraries: USING DEBUG CXXFLAGS = '/DWIN32 /D_WINDOWS /GR /EHsc /Zi /Ob0 /Od /RTC1' USING DEBUG EXEFLAGS = '/machine:x64 /debug /INCREMENTAL' USING RELEASE CXXFLAGS = '/DWIN32 /D_WINDOWS /GR /EHsc /O2 /Ob2 /DNDEBUG' USING RELEASE EXEFLAGS = '/machine:x64 /INCREMENTAL:NO' NOTICE : the LAS reader test requires LASlib and will not be compiled. CMake Deprecation Warning at E:/vcpkg/installed/x64-windows/share/cgal/UseEigen3.cmake:11 (message): This file UseEigen.cmake is deprecated, and the imported target `CGAL::Eigen_support` from CGAL_Eigen_support.cmake should be used instead. Call Stack (most recent call first): CMakeLists.txt:76 (include) Configuring done I know this is because OpenGR and Eign are **Optional Third Party Libraries** so they are not installed when I install CGAL. The problem is, I dont quite understand the instructions on the web : **In order to use Eigen in CGAL programs, the executables should be linked with the CMake imported target CGAL::Eigen_support provided in CGAL_Eigen_support.cmake.** **In order to use OpenGR in CGAL programs, the executables should be linked with the CMake imported target CGAL::OpenGR_support provided in CGAL_OpenGR_support.cmake.** What does it mean? What should I do when compile the example in Cmake? ## Source Code error when compile the example: Point_set_processing_3. ## Environment * Operating system (Windows/64 bits): * Compiler: VS 2017, Cmake * CGAL version: 5.1.1 Thanks
1.0
compile openGR and pointmatch wrapper - Sorry to bother you. ## Issue Details I have installed CGAL 5.1.1 through **vcpkg** and compiled the example: CGAL-5.0.2-examples-Surface_mesh with Cmake and run it successfully. My aim is to use **OpenGR** and Pointmacher wrapper in CGAl to register point clouds gbobally. So I try to compile the example: Point_set_processing_3. But a compilation error raises. Selecting Windows SDK version 10.0.17763.0 to target Windows 10.0.18363. Visual Leak Detector (VLD) is not found. Using header-only CGAL Targetting Visual Studio 15 2017 Target build enviroment supports auto-linking Using VC toolset 141. Generator uses intermediate configuration directory: $(Configuration) Boost include dirs: E:/vcpkg/installed/x64-windows/include Boost libraries: USING DEBUG CXXFLAGS = '/DWIN32 /D_WINDOWS /GR /EHsc /Zi /Ob0 /Od /RTC1' USING DEBUG EXEFLAGS = '/machine:x64 /debug /INCREMENTAL' USING RELEASE CXXFLAGS = '/DWIN32 /D_WINDOWS /GR /EHsc /O2 /Ob2 /DNDEBUG' USING RELEASE EXEFLAGS = '/machine:x64 /INCREMENTAL:NO' NOTICE : the LAS reader test requires LASlib and will not be compiled. CMake Deprecation Warning at E:/vcpkg/installed/x64-windows/share/cgal/UseEigen3.cmake:11 (message): This file UseEigen.cmake is deprecated, and the imported target `CGAL::Eigen_support` from CGAL_Eigen_support.cmake should be used instead. Call Stack (most recent call first): CMakeLists.txt:76 (include) Configuring done I know this is because OpenGR and Eign are **Optional Third Party Libraries** so they are not installed when I install CGAL. The problem is, I dont quite understand the instructions on the web : **In order to use Eigen in CGAL programs, the executables should be linked with the CMake imported target CGAL::Eigen_support provided in CGAL_Eigen_support.cmake.** **In order to use OpenGR in CGAL programs, the executables should be linked with the CMake imported target CGAL::OpenGR_support provided in CGAL_OpenGR_support.cmake.** What does it mean? What should I do when compile the example in Cmake? ## Source Code error when compile the example: Point_set_processing_3. ## Environment * Operating system (Windows/64 bits): * Compiler: VS 2017, Cmake * CGAL version: 5.1.1 Thanks
process
compile opengr and pointmatch wrapper sorry to bother you issue details i have installed cgal through vcpkg and compiled the example cgal examples surface mesh with cmake and run it successfully my aim is to use opengr and pointmacher wrapper in cgal to register point clouds gbobally so i try to compile the example point set processing but a compilation error raises selecting windows sdk version to target windows visual leak detector vld is not found using header only cgal targetting visual studio target build enviroment supports auto linking using vc toolset generator uses intermediate configuration directory configuration boost include dirs e vcpkg installed windows include boost libraries using debug cxxflags d windows gr ehsc zi od using debug exeflags machine debug incremental using release cxxflags d windows gr ehsc dndebug using release exeflags machine incremental no notice the las reader test requires laslib and will not be compiled cmake deprecation warning at e vcpkg installed windows share cgal cmake message this file useeigen cmake is deprecated and the imported target cgal eigen support from cgal eigen support cmake should be used instead call stack most recent call first cmakelists txt include configuring done i know this is because opengr and eign are optional third party libraries so they are not installed when i install cgal the problem is i dont quite understand the instructions on the web in order to use eigen in cgal programs the executables should be linked with the cmake imported target cgal eigen support provided in cgal eigen support cmake in order to use opengr in cgal programs the executables should be linked with the cmake imported target cgal opengr support provided in cgal opengr support cmake what does it mean what should i do when compile the example in cmake source code error when compile the example point set processing environment operating system windows bits compiler vs cmake cgal version thanks
1
690,776
23,671,929,029
IssuesEvent
2022-08-27 13:45:38
RED7Studios/RED7Community
https://api.github.com/repos/RED7Studios/RED7Community
opened
Add the ability to add an icon to a clan
Enhancement Medium Priority
## Description Add a setting in the manage tab for clans to have an icon.
1.0
Add the ability to add an icon to a clan - ## Description Add a setting in the manage tab for clans to have an icon.
non_process
add the ability to add an icon to a clan description add a setting in the manage tab for clans to have an icon
0
20,337
26,994,181,476
IssuesEvent
2023-02-09 22:46:47
googleapis/sloth
https://api.github.com/repos/googleapis/sloth
closed
delete repos.json
type: process
With both `repos.json` and DRIFT responsible for tracking GitHub repositories, it means there are two sources of truth. We should make DRIFT the single source of truth. Let's use this issue to track down all usages of `repos.json` that need to be rationalized. - [x] [autosynth](https://github.com/googleapis/synthtool/issues/645) - [x] [google-cloud-node](https://github.com/googleapis/google-cloud-node/issues/3012) - [x] [github-repo-automation](https://github.com/googleapis/github-repo-automation/issues/417) - [x] [repo-automation-bots](https://github.com/googleapis/repo-automation-bots/issues/623) - [ ] This repository!
1.0
delete repos.json - With both `repos.json` and DRIFT responsible for tracking GitHub repositories, it means there are two sources of truth. We should make DRIFT the single source of truth. Let's use this issue to track down all usages of `repos.json` that need to be rationalized. - [x] [autosynth](https://github.com/googleapis/synthtool/issues/645) - [x] [google-cloud-node](https://github.com/googleapis/google-cloud-node/issues/3012) - [x] [github-repo-automation](https://github.com/googleapis/github-repo-automation/issues/417) - [x] [repo-automation-bots](https://github.com/googleapis/repo-automation-bots/issues/623) - [ ] This repository!
process
delete repos json with both repos json and drift responsible for tracking github repositories it means there are two sources of truth we should make drift the single source of truth let s use this issue to track down all usages of repos json that need to be rationalized this repository
1
256,416
19,412,287,857
IssuesEvent
2021-12-20 10:56:18
dbwebb-se/databas
https://api.github.com/repos/dbwebb-se/databas
closed
Mikaels setup för att spela in föreläsningar
kursutveckling documentation
Jag skrev ner lite korta kommentarer till mina kära kollegor om hur miljön för inspelning ser ut. * En greenscreen på väggen, ca 270 bred, ca 50cm utrymme till där du står (eller större om det går). * Sen skrivbord med dator o skärmar. * Utrymme ungefär som vi har i våra arbetsrum på kontoret räcker bra (bredden alltså) ca 2m men gärna mer. * Ett högre bord där man kan ställa sin streamdeck så man kan klicka (byta scener, byta slide). * 3 lampor räcker för mig, samma som ni köpt på jobbet (modell enkla/billiga). Vissa rekommenderar 5 lampor om man inte får rätt belysning på green-srceenen. * Jag körde på min Yeti mikrofon men den är lite för långt bort (burkigt ljud), måste fixa en arm, schockmount och klickfilter. * Ska även försöka ljudisolera lite med en tyngre matta och textilier på väggarna. * Ljudet har för tillfället tre filter via OBS, ska fippla lite med dem och lägga till en equaliser addon (skall ta bort lite eko). * En gammal webcam Logitech 920 på stativ (tog mitt kamerastativ). * Camen körs via ett OBS greenscreenfilter (funkade klockrent) och ett crop-filter. * Två skärmar och OBS aktivt på skärmen framför mig så jag ser mig själv i den aktiva scenen. * Plockade hem den nya pc-datorn från jobbet och kör på windows. Bild på "studion" https://photos.app.goo.gl/nYpnYwcttwJmjRB2A
1.0
Mikaels setup för att spela in föreläsningar - Jag skrev ner lite korta kommentarer till mina kära kollegor om hur miljön för inspelning ser ut. * En greenscreen på väggen, ca 270 bred, ca 50cm utrymme till där du står (eller större om det går). * Sen skrivbord med dator o skärmar. * Utrymme ungefär som vi har i våra arbetsrum på kontoret räcker bra (bredden alltså) ca 2m men gärna mer. * Ett högre bord där man kan ställa sin streamdeck så man kan klicka (byta scener, byta slide). * 3 lampor räcker för mig, samma som ni köpt på jobbet (modell enkla/billiga). Vissa rekommenderar 5 lampor om man inte får rätt belysning på green-srceenen. * Jag körde på min Yeti mikrofon men den är lite för långt bort (burkigt ljud), måste fixa en arm, schockmount och klickfilter. * Ska även försöka ljudisolera lite med en tyngre matta och textilier på väggarna. * Ljudet har för tillfället tre filter via OBS, ska fippla lite med dem och lägga till en equaliser addon (skall ta bort lite eko). * En gammal webcam Logitech 920 på stativ (tog mitt kamerastativ). * Camen körs via ett OBS greenscreenfilter (funkade klockrent) och ett crop-filter. * Två skärmar och OBS aktivt på skärmen framför mig så jag ser mig själv i den aktiva scenen. * Plockade hem den nya pc-datorn från jobbet och kör på windows. Bild på "studion" https://photos.app.goo.gl/nYpnYwcttwJmjRB2A
non_process
mikaels setup för att spela in föreläsningar jag skrev ner lite korta kommentarer till mina kära kollegor om hur miljön för inspelning ser ut en greenscreen på väggen ca bred ca utrymme till där du står eller större om det går sen skrivbord med dator o skärmar utrymme ungefär som vi har i våra arbetsrum på kontoret räcker bra bredden alltså ca men gärna mer ett högre bord där man kan ställa sin streamdeck så man kan klicka byta scener byta slide lampor räcker för mig samma som ni köpt på jobbet modell enkla billiga vissa rekommenderar lampor om man inte får rätt belysning på green srceenen jag körde på min yeti mikrofon men den är lite för långt bort burkigt ljud måste fixa en arm schockmount och klickfilter ska även försöka ljudisolera lite med en tyngre matta och textilier på väggarna ljudet har för tillfället tre filter via obs ska fippla lite med dem och lägga till en equaliser addon skall ta bort lite eko en gammal webcam logitech på stativ tog mitt kamerastativ camen körs via ett obs greenscreenfilter funkade klockrent och ett crop filter två skärmar och obs aktivt på skärmen framför mig så jag ser mig själv i den aktiva scenen plockade hem den nya pc datorn från jobbet och kör på windows bild på studion
0
91,629
15,856,540,912
IssuesEvent
2021-04-08 02:35:14
arohablue/skill-india-frontend
https://api.github.com/repos/arohablue/skill-india-frontend
opened
CVE-2020-7729 (High) detected in grunt-0.4.5.tgz
security vulnerability
## CVE-2020-7729 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>grunt-0.4.5.tgz</b></p></summary> <p>The JavaScript Task Runner</p> <p>Library home page: <a href="https://registry.npmjs.org/grunt/-/grunt-0.4.5.tgz">https://registry.npmjs.org/grunt/-/grunt-0.4.5.tgz</a></p> <p>Path to dependency file: /skill-india-frontend/src/assets/Bootstrap/lib/jquery-ui/package.json</p> <p>Path to vulnerable library: skill-india-frontend/src/app/DisplayCourses/Bootstrap/lib/jquery-ui/node_modules/grunt/package.json,skill-india-frontend/src/app/DisplayCourses/Bootstrap/lib/jquery-ui/node_modules/grunt/package.json</p> <p> Dependency Hierarchy: - :x: **grunt-0.4.5.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package grunt before 1.3.0 are vulnerable to Arbitrary Code Execution due to the default usage of the function load() instead of its secure replacement safeLoad() of the package js-yaml inside grunt.file.readYAML. <p>Publish Date: 2020-09-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7729>CVE-2020-7729</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: Low - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7729">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7729</a></p> <p>Release Date: 2020-07-21</p> <p>Fix Resolution: 1.3.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-7729 (High) detected in grunt-0.4.5.tgz - ## CVE-2020-7729 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>grunt-0.4.5.tgz</b></p></summary> <p>The JavaScript Task Runner</p> <p>Library home page: <a href="https://registry.npmjs.org/grunt/-/grunt-0.4.5.tgz">https://registry.npmjs.org/grunt/-/grunt-0.4.5.tgz</a></p> <p>Path to dependency file: /skill-india-frontend/src/assets/Bootstrap/lib/jquery-ui/package.json</p> <p>Path to vulnerable library: skill-india-frontend/src/app/DisplayCourses/Bootstrap/lib/jquery-ui/node_modules/grunt/package.json,skill-india-frontend/src/app/DisplayCourses/Bootstrap/lib/jquery-ui/node_modules/grunt/package.json</p> <p> Dependency Hierarchy: - :x: **grunt-0.4.5.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package grunt before 1.3.0 are vulnerable to Arbitrary Code Execution due to the default usage of the function load() instead of its secure replacement safeLoad() of the package js-yaml inside grunt.file.readYAML. <p>Publish Date: 2020-09-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7729>CVE-2020-7729</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: Low - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7729">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7729</a></p> <p>Release Date: 2020-07-21</p> <p>Fix Resolution: 1.3.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in grunt tgz cve high severity vulnerability vulnerable library grunt tgz the javascript task runner library home page a href path to dependency file skill india frontend src assets bootstrap lib jquery ui package json path to vulnerable library skill india frontend src app displaycourses bootstrap lib jquery ui node modules grunt package json skill india frontend src app displaycourses bootstrap lib jquery ui node modules grunt package json dependency hierarchy x grunt tgz vulnerable library vulnerability details the package grunt before are vulnerable to arbitrary code execution due to the default usage of the function load instead of its secure replacement safeload of the package js yaml inside grunt file readyaml publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required low user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
33,609
9,196,691,432
IssuesEvent
2019-03-07 07:58:03
Microsoft/WindowsTemplateStudio
https://api.github.com/repos/Microsoft/WindowsTemplateStudio
closed
Build dev.templates.tests.full_20190306.5 failed
bug vsts-build
## Build dev.templates.tests.full_20190306.5 - **Build result:** `failed` - **Build queued:** 3/6/2019 10:53:38 PM - **Build duration:** 48.41 minutes ### Details Build [dev.templates.tests.full_20190306.5](https://winappstudio.visualstudio.com/web/build.aspx?pcguid=a4ef43be-68ce-4195-a619-079b4d9834c2&builduri=vstfs%3a%2f%2f%2fBuild%2fBuild%2f27205) failed + xunit.console.exe : GenerateAllPagesAndFeaturesAndCheckWithSonarLintAsync(projectType: "Blank", framework: SupportedFramework { Name = "MVVMBasic", Type = FrontEnd }, platform: "Uwp") [FAIL] At pbatch:27 char:27 + + CategoryInfo : NotSpecified: ( GenerateAll...: "Uwp") [FAIL]:String) [], RemoteException + FullyQualifiedErrorId : NativeCommandError + PSComputerName : [localhost] GenerateAllPagesAndFeaturesAndCheckWithSonarLintAsync(projectType: "Blank", framework: SupportedFramework { Name = "MVVMLight", Type = FrontEnd }, platform: "Uwp") [FAIL] GenerateAllPagesAndFeaturesAndCheckWithSonarLintAsync(projectType: "Blank", framework: SupportedFramework { Name = "CodeBehind", Type = FrontEnd }, platform: "Uwp") [FAIL] GenerateAllPagesAndFeaturesAndCheckWithSonarLintAsync(projectType: "SplitView", framework: SupportedFramework { Name = "MVVMBasic", Type = FrontEnd }, platform: "Uwp") [FAIL] GenerateAllPagesAndFeaturesAndCheckWithSonarLintAsync(projectType: "SplitView", framework: SupportedFramework { Name = "MVVMLight", Type = FrontEnd }, platform: "Uwp") [FAIL] GenerateAllPagesAndFeaturesAndCheckWithSonarLintAsync(projectType: "SplitView", framework: SupportedFramework { Name = "CodeBehind", Type = FrontEnd }, platform: "Uwp") [FAIL] GenerateAllPagesAndFeaturesAndCheckWithSonarLintAsync(projectType: "TabbedNav", framework: SupportedFramework { Name = "MVVMBasic", Type = FrontEnd }, platform: "Uwp") [FAIL] GenerateAllPagesAndFeaturesAndCheckWithSonarLintAsync(projectType: "TabbedNav", framework: SupportedFramework { Name = "MVVMLight", Type = FrontEnd }, platform: "Uwp") [FAIL] GenerateAllPagesAndFeaturesAndCheckWithSonarLintAsync(projectType: "TabbedNav", framework: SupportedFramework { Name = "CodeBehind", Type = FrontEnd }, platform: "Uwp") [FAIL] GenerateAllPagesAndFeaturesAndCheckWithVBStyleAsync(projectType: "Blank", framework: SupportedFramework { Name = "MVVMBasic", Type = FrontEnd }, platform: "Uwp") [FAIL] GenerateAllPagesAndFeaturesAndCheckWithVBStyleAsync(projectType: "Blank", framework: SupportedFramework { Name = "MVVMLight", Type = FrontEnd }, platform: "Uwp") [FAIL] GenerateAllPagesAndFeaturesAndCheckWithVBStyleAsync(projectType: "Blank", framework: SupportedFramework { Name = "CodeBehind", Type = FrontEnd }, platform: "Uwp") [FAIL] GenerateAllPagesAndFeaturesAndCheckWithVBStyleAsync(projectType: "SplitView", framework: SupportedFramework { Name = "MVVMBasic", Type = FrontEnd }, platform: "Uwp") [FAIL] GenerateAllPagesAndFeaturesAndCheckWithVBStyleAsync(projectType: "SplitView", framework: SupportedFramework { Name = "MVVMLight", Type = FrontEnd }, platform: "Uwp") [FAIL] GenerateAllPagesAndFeaturesAndCheckWithVBStyleAsync(projectType: "SplitView", framework: SupportedFramework { Name = "CodeBehind", Type = FrontEnd }, platform: "Uwp") [FAIL] GenerateAllPagesAndFeaturesAndCheckWithVBStyleAsync(projectType: "TabbedNav", framework: SupportedFramework { Name = "MVVMBasic", Type = FrontEnd }, platform: "Uwp") [FAIL] GenerateAllPagesAndFeaturesAndCheckWithVBStyleAsync(projectType: "TabbedNav", framework: SupportedFramework { Name = "MVVMLight", Type = FrontEnd }, platform: "Uwp") [FAIL] GenerateAllPagesAndFeaturesAndCheckWithVBStyleAsync(projectType: "TabbedNav", framework: SupportedFramework { Name = "CodeBehind", Type = FrontEnd }, platform: "Uwp") [FAIL] + Process completed with exit code 18 and had 1 error(s) written to the error stream. Find detailed information in the [build log files](https://uwpctdiags.blob.core.windows.net/buildlogs/dev.templates.tests.full_20190306.5_logs.zip)
1.0
Build dev.templates.tests.full_20190306.5 failed - ## Build dev.templates.tests.full_20190306.5 - **Build result:** `failed` - **Build queued:** 3/6/2019 10:53:38 PM - **Build duration:** 48.41 minutes ### Details Build [dev.templates.tests.full_20190306.5](https://winappstudio.visualstudio.com/web/build.aspx?pcguid=a4ef43be-68ce-4195-a619-079b4d9834c2&builduri=vstfs%3a%2f%2f%2fBuild%2fBuild%2f27205) failed + xunit.console.exe : GenerateAllPagesAndFeaturesAndCheckWithSonarLintAsync(projectType: "Blank", framework: SupportedFramework { Name = "MVVMBasic", Type = FrontEnd }, platform: "Uwp") [FAIL] At pbatch:27 char:27 + + CategoryInfo : NotSpecified: ( GenerateAll...: "Uwp") [FAIL]:String) [], RemoteException + FullyQualifiedErrorId : NativeCommandError + PSComputerName : [localhost] GenerateAllPagesAndFeaturesAndCheckWithSonarLintAsync(projectType: "Blank", framework: SupportedFramework { Name = "MVVMLight", Type = FrontEnd }, platform: "Uwp") [FAIL] GenerateAllPagesAndFeaturesAndCheckWithSonarLintAsync(projectType: "Blank", framework: SupportedFramework { Name = "CodeBehind", Type = FrontEnd }, platform: "Uwp") [FAIL] GenerateAllPagesAndFeaturesAndCheckWithSonarLintAsync(projectType: "SplitView", framework: SupportedFramework { Name = "MVVMBasic", Type = FrontEnd }, platform: "Uwp") [FAIL] GenerateAllPagesAndFeaturesAndCheckWithSonarLintAsync(projectType: "SplitView", framework: SupportedFramework { Name = "MVVMLight", Type = FrontEnd }, platform: "Uwp") [FAIL] GenerateAllPagesAndFeaturesAndCheckWithSonarLintAsync(projectType: "SplitView", framework: SupportedFramework { Name = "CodeBehind", Type = FrontEnd }, platform: "Uwp") [FAIL] GenerateAllPagesAndFeaturesAndCheckWithSonarLintAsync(projectType: "TabbedNav", framework: SupportedFramework { Name = "MVVMBasic", Type = FrontEnd }, platform: "Uwp") [FAIL] GenerateAllPagesAndFeaturesAndCheckWithSonarLintAsync(projectType: "TabbedNav", framework: SupportedFramework { Name = "MVVMLight", Type = FrontEnd }, platform: "Uwp") [FAIL] GenerateAllPagesAndFeaturesAndCheckWithSonarLintAsync(projectType: "TabbedNav", framework: SupportedFramework { Name = "CodeBehind", Type = FrontEnd }, platform: "Uwp") [FAIL] GenerateAllPagesAndFeaturesAndCheckWithVBStyleAsync(projectType: "Blank", framework: SupportedFramework { Name = "MVVMBasic", Type = FrontEnd }, platform: "Uwp") [FAIL] GenerateAllPagesAndFeaturesAndCheckWithVBStyleAsync(projectType: "Blank", framework: SupportedFramework { Name = "MVVMLight", Type = FrontEnd }, platform: "Uwp") [FAIL] GenerateAllPagesAndFeaturesAndCheckWithVBStyleAsync(projectType: "Blank", framework: SupportedFramework { Name = "CodeBehind", Type = FrontEnd }, platform: "Uwp") [FAIL] GenerateAllPagesAndFeaturesAndCheckWithVBStyleAsync(projectType: "SplitView", framework: SupportedFramework { Name = "MVVMBasic", Type = FrontEnd }, platform: "Uwp") [FAIL] GenerateAllPagesAndFeaturesAndCheckWithVBStyleAsync(projectType: "SplitView", framework: SupportedFramework { Name = "MVVMLight", Type = FrontEnd }, platform: "Uwp") [FAIL] GenerateAllPagesAndFeaturesAndCheckWithVBStyleAsync(projectType: "SplitView", framework: SupportedFramework { Name = "CodeBehind", Type = FrontEnd }, platform: "Uwp") [FAIL] GenerateAllPagesAndFeaturesAndCheckWithVBStyleAsync(projectType: "TabbedNav", framework: SupportedFramework { Name = "MVVMBasic", Type = FrontEnd }, platform: "Uwp") [FAIL] GenerateAllPagesAndFeaturesAndCheckWithVBStyleAsync(projectType: "TabbedNav", framework: SupportedFramework { Name = "MVVMLight", Type = FrontEnd }, platform: "Uwp") [FAIL] GenerateAllPagesAndFeaturesAndCheckWithVBStyleAsync(projectType: "TabbedNav", framework: SupportedFramework { Name = "CodeBehind", Type = FrontEnd }, platform: "Uwp") [FAIL] + Process completed with exit code 18 and had 1 error(s) written to the error stream. Find detailed information in the [build log files](https://uwpctdiags.blob.core.windows.net/buildlogs/dev.templates.tests.full_20190306.5_logs.zip)
non_process
build dev templates tests full failed build dev templates tests full build result failed build queued pm build duration minutes details build failed xunit console exe generateallpagesandfeaturesandcheckwithsonarlintasync projecttype blank framework supportedframework name mvvmbasic type frontend platform uwp at pbatch char categoryinfo notspecified generateall uwp string remoteexception fullyqualifiederrorid nativecommanderror pscomputername generateallpagesandfeaturesandcheckwithsonarlintasync projecttype blank framework supportedframework name mvvmlight type frontend platform uwp generateallpagesandfeaturesandcheckwithsonarlintasync projecttype blank framework supportedframework name codebehind type frontend platform uwp generateallpagesandfeaturesandcheckwithsonarlintasync projecttype splitview framework supportedframework name mvvmbasic type frontend platform uwp generateallpagesandfeaturesandcheckwithsonarlintasync projecttype splitview framework supportedframework name mvvmlight type frontend platform uwp generateallpagesandfeaturesandcheckwithsonarlintasync projecttype splitview framework supportedframework name codebehind type frontend platform uwp generateallpagesandfeaturesandcheckwithsonarlintasync projecttype tabbednav framework supportedframework name mvvmbasic type frontend platform uwp generateallpagesandfeaturesandcheckwithsonarlintasync projecttype tabbednav framework supportedframework name mvvmlight type frontend platform uwp generateallpagesandfeaturesandcheckwithsonarlintasync projecttype tabbednav framework supportedframework name codebehind type frontend platform uwp generateallpagesandfeaturesandcheckwithvbstyleasync projecttype blank framework supportedframework name mvvmbasic type frontend platform uwp generateallpagesandfeaturesandcheckwithvbstyleasync projecttype blank framework supportedframework name mvvmlight type frontend platform uwp generateallpagesandfeaturesandcheckwithvbstyleasync projecttype blank framework supportedframework name codebehind type frontend platform uwp generateallpagesandfeaturesandcheckwithvbstyleasync projecttype splitview framework supportedframework name mvvmbasic type frontend platform uwp generateallpagesandfeaturesandcheckwithvbstyleasync projecttype splitview framework supportedframework name mvvmlight type frontend platform uwp generateallpagesandfeaturesandcheckwithvbstyleasync projecttype splitview framework supportedframework name codebehind type frontend platform uwp generateallpagesandfeaturesandcheckwithvbstyleasync projecttype tabbednav framework supportedframework name mvvmbasic type frontend platform uwp generateallpagesandfeaturesandcheckwithvbstyleasync projecttype tabbednav framework supportedframework name mvvmlight type frontend platform uwp generateallpagesandfeaturesandcheckwithvbstyleasync projecttype tabbednav framework supportedframework name codebehind type frontend platform uwp process completed with exit code and had error s written to the error stream find detailed information in the
0
102,398
21,960,003,121
IssuesEvent
2022-05-24 15:04:12
Onelinerhub/onelinerhub
https://api.github.com/repos/Onelinerhub/onelinerhub
opened
Short solution needed: "How to update plot dynamically" (python-matplotlib)
help wanted good first issue code python-matplotlib
Please help us write most modern and shortest code solution for this issue: **How to update plot dynamically** (technology: [python-matplotlib](https://onelinerhub.com/python-matplotlib)) ### Fast way Just write the code solution in the comments. ### Prefered way 1. Create [pull request](https://github.com/Onelinerhub/onelinerhub/blob/main/how-to-contribute.md) with a new code file inside [inbox folder](https://github.com/Onelinerhub/onelinerhub/tree/main/inbox). 2. Don't forget to [use comments](https://github.com/Onelinerhub/onelinerhub/blob/main/how-to-contribute.md#code-file-md-format) explain solution. 3. Link to this issue in comments of pull request.
1.0
Short solution needed: "How to update plot dynamically" (python-matplotlib) - Please help us write most modern and shortest code solution for this issue: **How to update plot dynamically** (technology: [python-matplotlib](https://onelinerhub.com/python-matplotlib)) ### Fast way Just write the code solution in the comments. ### Prefered way 1. Create [pull request](https://github.com/Onelinerhub/onelinerhub/blob/main/how-to-contribute.md) with a new code file inside [inbox folder](https://github.com/Onelinerhub/onelinerhub/tree/main/inbox). 2. Don't forget to [use comments](https://github.com/Onelinerhub/onelinerhub/blob/main/how-to-contribute.md#code-file-md-format) explain solution. 3. Link to this issue in comments of pull request.
non_process
short solution needed how to update plot dynamically python matplotlib please help us write most modern and shortest code solution for this issue how to update plot dynamically technology fast way just write the code solution in the comments prefered way create with a new code file inside don t forget to explain solution link to this issue in comments of pull request
0
2,498
5,271,982,541
IssuesEvent
2017-02-06 11:22:54
opentrials/opentrials
https://api.github.com/repos/opentrials/opentrials
closed
Pin our dependencies versions
4. Ready for Review Collectors Processors
We already pin them on the API and Explorer, but don't on collectors/processors. This means that our builds aren't reproducible (as the dependencies can change in between) and that changes in our dependencies can break our code (as they just did with `dataset==0.8.0`, see https://travis-ci.org/opentrials/processors/builds/196530268).
1.0
Pin our dependencies versions - We already pin them on the API and Explorer, but don't on collectors/processors. This means that our builds aren't reproducible (as the dependencies can change in between) and that changes in our dependencies can break our code (as they just did with `dataset==0.8.0`, see https://travis-ci.org/opentrials/processors/builds/196530268).
process
pin our dependencies versions we already pin them on the api and explorer but don t on collectors processors this means that our builds aren t reproducible as the dependencies can change in between and that changes in our dependencies can break our code as they just did with dataset see
1
2,389
5,187,642,530
IssuesEvent
2017-01-20 17:24:53
Alfresco/alfresco-ng2-components
https://api.github.com/repos/Alfresco/alfresco-ng2-components
closed
activiti-processlist component uses TaskQueryRequestRepresentationModel to build request
bug comp: activiti-processList
**Type of issue:** (check with "[x]") ``` - [ ] New feature request - [x] Bug - [ ] Support request ``` **Current behavior:** Currently it works, however when I look at the source code it doesn't make sense to use a TaskQuery representation for process instance query! Wondering if this needs to be fixed, raising an issue anyway! **Expected behavior:** Process Instance query request to be built from a process query representation model. **Component name and version:** ng2-activiti-processlist v1.0.0 **Browser and version:** all
1.0
activiti-processlist component uses TaskQueryRequestRepresentationModel to build request - **Type of issue:** (check with "[x]") ``` - [ ] New feature request - [x] Bug - [ ] Support request ``` **Current behavior:** Currently it works, however when I look at the source code it doesn't make sense to use a TaskQuery representation for process instance query! Wondering if this needs to be fixed, raising an issue anyway! **Expected behavior:** Process Instance query request to be built from a process query representation model. **Component name and version:** ng2-activiti-processlist v1.0.0 **Browser and version:** all
process
activiti processlist component uses taskqueryrequestrepresentationmodel to build request type of issue check with new feature request bug support request current behavior currently it works however when i look at the source code it doesn t make sense to use a taskquery representation for process instance query wondering if this needs to be fixed raising an issue anyway expected behavior process instance query request to be built from a process query representation model component name and version activiti processlist browser and version all
1
15,967
20,177,342,152
IssuesEvent
2022-02-10 15:30:21
ossf/tac
https://api.github.com/repos/ossf/tac
closed
TAC Election Process: Term limits
ElectionProcess
**Proposal:** - 2 year staggered terms. Even year, GB vote. Odd year, TI vote.
1.0
TAC Election Process: Term limits - **Proposal:** - 2 year staggered terms. Even year, GB vote. Odd year, TI vote.
process
tac election process term limits proposal year staggered terms even year gb vote odd year ti vote
1
2,655
5,430,471,148
IssuesEvent
2017-03-03 21:20:37
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
closed
InnerLoop_Fedora24_Debug: Test "System.Diagnostics.Tests.ProcessStreamReadTests.TestSyncErrorStream" failed with "Xunit.Sdk.EqualException"
area-System.Diagnostics.Process blocking-clean-ci test-run-core
Failed test: System.Diagnostics.Tests.ProcessStreamReadTests.TestSyncErrorStream Detail: https://ci.dot.net/job/dotnet_corefx/job/master/job/fedora24_debug/1140/consoleText Configuration: InnerLoop_Fedora24_Debug ([build#1140](https://ci.dot.net/job/dotnet_corefx/job/master/job/fedora24_debug/1140/testReport/)) Message: ~~~ System.Diagnostics.Tests.ProcessStreamReadTests.TestSyncErrorStream [FAIL] Assert.Equal() Failure ↓ (pos 0) Expected: RemoteExecutorConsoleApp.exe started erro··· Actual: Did you mean to run dotnet SDK commands? ··· ↑ (pos 0) ~~~ Stack Trace: ~~~ /mnt/resource/j/workspace/dotnet_corefx/master/fedora24_debug/src/System.Diagnostics.Process/tests/ProcessStreamReadTests.cs(23,0): at System.Diagnostics.Tests.ProcessStreamReadTests.TestSyncErrorStream() ~~~
1.0
InnerLoop_Fedora24_Debug: Test "System.Diagnostics.Tests.ProcessStreamReadTests.TestSyncErrorStream" failed with "Xunit.Sdk.EqualException" - Failed test: System.Diagnostics.Tests.ProcessStreamReadTests.TestSyncErrorStream Detail: https://ci.dot.net/job/dotnet_corefx/job/master/job/fedora24_debug/1140/consoleText Configuration: InnerLoop_Fedora24_Debug ([build#1140](https://ci.dot.net/job/dotnet_corefx/job/master/job/fedora24_debug/1140/testReport/)) Message: ~~~ System.Diagnostics.Tests.ProcessStreamReadTests.TestSyncErrorStream [FAIL] Assert.Equal() Failure ↓ (pos 0) Expected: RemoteExecutorConsoleApp.exe started erro··· Actual: Did you mean to run dotnet SDK commands? ··· ↑ (pos 0) ~~~ Stack Trace: ~~~ /mnt/resource/j/workspace/dotnet_corefx/master/fedora24_debug/src/System.Diagnostics.Process/tests/ProcessStreamReadTests.cs(23,0): at System.Diagnostics.Tests.ProcessStreamReadTests.TestSyncErrorStream() ~~~
process
innerloop debug test system diagnostics tests processstreamreadtests testsyncerrorstream failed with xunit sdk equalexception failed test system diagnostics tests processstreamreadtests testsyncerrorstream detail configuration innerloop debug message system diagnostics tests processstreamreadtests testsyncerrorstream assert equal failure ↓ pos expected remoteexecutorconsoleapp exe started erro··· actual did you mean to run dotnet sdk commands ··· ↑ pos stack trace mnt resource j workspace dotnet corefx master debug src system diagnostics process tests processstreamreadtests cs at system diagnostics tests processstreamreadtests testsyncerrorstream
1
365,079
25,519,516,531
IssuesEvent
2022-11-28 19:11:10
lieion/OpenSource_Final_Project
https://api.github.com/repos/lieion/OpenSource_Final_Project
opened
현재 진행상황 Update
documentation
# + **Term Server 관련 개발 정보** 현재 개발 완료된 정보 ~~1. 각 카페 별 json file~~ json 파일이 필요하기 때문에 디렉토리 구조를 public 내부에 json 파일 폴더를 추가하였습니다.(./public/json) ~~2. Blue Port 가격 계산~~ 가격 계산 구조가 Post로 주문한 음료의 개수를 각각 보내고 서버에서 가격을 계산한 뒤에 응답으로 그 가격을 보내면 출력해주는 방식으로 구현하였습니다. ~~3. Blue Port 주문~~ ~~4. myPage에서 Blue Port 주문정보 출력~~ 개발 예정 정보 **1. 카페 관리자** 카페 관리하는 입장에서의 페이지 제작, 신고 추가 (~29) + **Suggestion** 다른 카페 부분도 구조가 비슷해서 변수만 수정하면 구현하는데 문제가 없을 것 같습니다. + **Additional Info** 개발할 시 주의해야 할 사항, 다른 file과의 dependency 등 추가적으로 알려줘야 할 내용 jquery를 사용하기 때문에 충돌이 나는 경우 `<script src="https://code.jquery.com/jquery-1.12.4.js"></script>` `<script src="https://code.jquery.com/ui/1.12.1/jquery-ui.js"></script>` head에 불러오지 않아서 일어나는 일이 잦았습니다. #
1.0
현재 진행상황 Update - # + **Term Server 관련 개발 정보** 현재 개발 완료된 정보 ~~1. 각 카페 별 json file~~ json 파일이 필요하기 때문에 디렉토리 구조를 public 내부에 json 파일 폴더를 추가하였습니다.(./public/json) ~~2. Blue Port 가격 계산~~ 가격 계산 구조가 Post로 주문한 음료의 개수를 각각 보내고 서버에서 가격을 계산한 뒤에 응답으로 그 가격을 보내면 출력해주는 방식으로 구현하였습니다. ~~3. Blue Port 주문~~ ~~4. myPage에서 Blue Port 주문정보 출력~~ 개발 예정 정보 **1. 카페 관리자** 카페 관리하는 입장에서의 페이지 제작, 신고 추가 (~29) + **Suggestion** 다른 카페 부분도 구조가 비슷해서 변수만 수정하면 구현하는데 문제가 없을 것 같습니다. + **Additional Info** 개발할 시 주의해야 할 사항, 다른 file과의 dependency 등 추가적으로 알려줘야 할 내용 jquery를 사용하기 때문에 충돌이 나는 경우 `<script src="https://code.jquery.com/jquery-1.12.4.js"></script>` `<script src="https://code.jquery.com/ui/1.12.1/jquery-ui.js"></script>` head에 불러오지 않아서 일어나는 일이 잦았습니다. #
non_process
현재 진행상황 update term server 관련 개발 정보 현재 개발 완료된 정보 각 카페 별 json file json 파일이 필요하기 때문에 디렉토리 구조를 public 내부에 json 파일 폴더를 추가하였습니다 public json blue port 가격 계산 가격 계산 구조가 post로 주문한 음료의 개수를 각각 보내고 서버에서 가격을 계산한 뒤에 응답으로 그 가격을 보내면 출력해주는 방식으로 구현하였습니다 blue port 주문 mypage에서 blue port 주문정보 출력 개발 예정 정보 카페 관리자 카페 관리하는 입장에서의 페이지 제작 신고 추가 suggestion 다른 카페 부분도 구조가 비슷해서 변수만 수정하면 구현하는데 문제가 없을 것 같습니다 additional info 개발할 시 주의해야 할 사항 다른 file과의 dependency 등 추가적으로 알려줘야 할 내용 jquery를 사용하기 때문에 충돌이 나는 경우 script src script src head에 불러오지 않아서 일어나는 일이 잦았습니다
0
646,938
21,082,530,944
IssuesEvent
2022-04-03 05:21:15
Appbird/kss-recorders
https://api.github.com/repos/Appbird/kss-recorders
opened
KSSRsでモード制を廃止する
enhancement high priority
(書きかけ) # 定義 - **モード制**とは、ユーザーに閲覧したいゲームモードを選択させて、ページにそのゲームモードに関連する情報/記録を表示するクライアント側のシステムのことである。 # 動機 - モード制を導入したせいで、記録を投稿する、記録を閲覧するまでの道のりが気楽なものでなくなっている。 - モード制は、ユーザーの関心のないゲームモードの情報を遮断するために導入されたものだった。 - モード制でゲームモードの選択を強制していたので... - 記録の検索において、表示される条件の選択肢を限定していた。 - つまり、関心のない別のゲームモードの条件を表示しないようにできていた。 - しかし、#37 統一検索ボックスの導入によりこのようなシステムが不要になった。 - 記録の申請において、ユーザーが記録のデータを入力する際に選択肢をそのゲームモードのものだけに限定できていた。 - これによりデータの入力での混乱を減らすことが期待されていた。 - しかし、そもそもデータの入力を管理委員会側で行うことになった(あとで追記)のでこのような仕組みが不要になった。 # 実装 - #37
1.0
KSSRsでモード制を廃止する - (書きかけ) # 定義 - **モード制**とは、ユーザーに閲覧したいゲームモードを選択させて、ページにそのゲームモードに関連する情報/記録を表示するクライアント側のシステムのことである。 # 動機 - モード制を導入したせいで、記録を投稿する、記録を閲覧するまでの道のりが気楽なものでなくなっている。 - モード制は、ユーザーの関心のないゲームモードの情報を遮断するために導入されたものだった。 - モード制でゲームモードの選択を強制していたので... - 記録の検索において、表示される条件の選択肢を限定していた。 - つまり、関心のない別のゲームモードの条件を表示しないようにできていた。 - しかし、#37 統一検索ボックスの導入によりこのようなシステムが不要になった。 - 記録の申請において、ユーザーが記録のデータを入力する際に選択肢をそのゲームモードのものだけに限定できていた。 - これによりデータの入力での混乱を減らすことが期待されていた。 - しかし、そもそもデータの入力を管理委員会側で行うことになった(あとで追記)のでこのような仕組みが不要になった。 # 実装 - #37
non_process
kssrsでモード制を廃止する 書きかけ 定義 モード制 とは、ユーザーに閲覧したいゲームモードを選択させて、ページにそのゲームモードに関連する情報 記録を表示するクライアント側のシステムのことである。 動機 モード制を導入したせいで、記録を投稿する、記録を閲覧するまでの道のりが気楽なものでなくなっている。 モード制は、ユーザーの関心のないゲームモードの情報を遮断するために導入されたものだった。 モード制でゲームモードの選択を強制していたので 記録の検索において、表示される条件の選択肢を限定していた。 つまり、関心のない別のゲームモードの条件を表示しないようにできていた。 しかし、 統一検索ボックスの導入によりこのようなシステムが不要になった。 記録の申請において、ユーザーが記録のデータを入力する際に選択肢をそのゲームモードのものだけに限定できていた。 これによりデータの入力での混乱を減らすことが期待されていた。 しかし、そもそもデータの入力を管理委員会側で行うことになった あとで追記 のでこのような仕組みが不要になった。 実装
0
8,277
11,431,936,494
IssuesEvent
2020-02-04 13:11:10
openopps/openopps-platform
https://api.github.com/repos/openopps/openopps-platform
closed
Add applicant status pills to student landing page
Apply Process Approved Landing page Requirements Ready State Dept.
Who: Student applicants What: Applicant status pills appear on landing page Why: To provide visual status Acceptance Criteria: - Add student applicant status pills to student landing page. Currently only the words appear ![image.png](https://images.zenhubusercontent.com/59ee08f1a468affe6df7cd6f/c8a7e11c-9e45-4e90-9a3b-9d4418d2a644) Related tickets: 4411 - Create applicant status pills 4412 - Display applicant status pills on student landing page 4329 - Applicant dashboard enhancements 4414 - Add applicant status pills to non student landing page
1.0
Add applicant status pills to student landing page - Who: Student applicants What: Applicant status pills appear on landing page Why: To provide visual status Acceptance Criteria: - Add student applicant status pills to student landing page. Currently only the words appear ![image.png](https://images.zenhubusercontent.com/59ee08f1a468affe6df7cd6f/c8a7e11c-9e45-4e90-9a3b-9d4418d2a644) Related tickets: 4411 - Create applicant status pills 4412 - Display applicant status pills on student landing page 4329 - Applicant dashboard enhancements 4414 - Add applicant status pills to non student landing page
process
add applicant status pills to student landing page who student applicants what applicant status pills appear on landing page why to provide visual status acceptance criteria add student applicant status pills to student landing page currently only the words appear related tickets create applicant status pills display applicant status pills on student landing page applicant dashboard enhancements add applicant status pills to non student landing page
1
66,748
12,821,810,972
IssuesEvent
2020-07-06 08:45:10
joomla/joomla-cms
https://api.github.com/repos/joomla/joomla-cms
closed
[4.0] error untranslated on registration page
No Code Attached Yet
### Steps to reproduce the issue Install Joomla 4.0-dev with sample data Enable User Registration Enable Lang debug (for testing) On frontend click "Create an account" Dont complete the form Click Register button See red error message Note that "error" is lowercase and not translatable ### Expected result `**Error**` ### Actual result error <img width="731" alt="Screenshot 2020-07-05 at 21 51 09" src="https://user-images.githubusercontent.com/400092/86542037-e4a74600-bf09-11ea-8ed4-5ee2e6cb1e3a.png"> ### System information (as much as possible) safari on mac ### Additional comments
1.0
[4.0] error untranslated on registration page - ### Steps to reproduce the issue Install Joomla 4.0-dev with sample data Enable User Registration Enable Lang debug (for testing) On frontend click "Create an account" Dont complete the form Click Register button See red error message Note that "error" is lowercase and not translatable ### Expected result `**Error**` ### Actual result error <img width="731" alt="Screenshot 2020-07-05 at 21 51 09" src="https://user-images.githubusercontent.com/400092/86542037-e4a74600-bf09-11ea-8ed4-5ee2e6cb1e3a.png"> ### System information (as much as possible) safari on mac ### Additional comments
non_process
error untranslated on registration page steps to reproduce the issue install joomla dev with sample data enable user registration enable lang debug for testing on frontend click create an account dont complete the form click register button see red error message note that error is lowercase and not translatable expected result error actual result error img width alt screenshot at src system information as much as possible safari on mac additional comments
0
6,666
23,682,970,319
IssuesEvent
2022-08-29 01:35:25
tm24fan8/Home-Assistant-Configs
https://api.github.com/repos/tm24fan8/Home-Assistant-Configs
closed
Fix living room media scenes
bug monitoring lighting multimedia automation
Need to investigate...used to work, now doesn't...I've changed nothing in between.
1.0
Fix living room media scenes - Need to investigate...used to work, now doesn't...I've changed nothing in between.
non_process
fix living room media scenes need to investigate used to work now doesn t i ve changed nothing in between
0
16,891
22,193,074,231
IssuesEvent
2022-06-07 02:32:46
streamnative/flink
https://api.github.com/repos/streamnative/flink
closed
[Feature][FLINK-27399] SourceEvent protocol for limiting the Consumer.seek operation
compute/data-processing type/feature
The Pulsar source needs to seek the desired consuming position when bootstrapping. The seeking action couldn’t be executed concurrently. We have designed a [new mechanism](https://github.com/apache/flink/pull/17119#pullrequestreview-746035072) that should be accomplished in the Flink 1.14.1 release.
1.0
[Feature][FLINK-27399] SourceEvent protocol for limiting the Consumer.seek operation - The Pulsar source needs to seek the desired consuming position when bootstrapping. The seeking action couldn’t be executed concurrently. We have designed a [new mechanism](https://github.com/apache/flink/pull/17119#pullrequestreview-746035072) that should be accomplished in the Flink 1.14.1 release.
process
sourceevent protocol for limiting the consumer seek operation the pulsar source needs to seek the desired consuming position when bootstrapping the seeking action couldn’t be executed concurrently we have designed a that should be accomplished in the flink release
1
13,587
16,162,946,843
IssuesEvent
2021-05-01 01:25:59
tdwg/chrono
https://api.github.com/repos/tdwg/chrono
closed
Change term - maximumChronometricAge, maximumChronometricAge
Process - prepare for Executive review Term - change
## Change term * Submitter: John Wieczorek (from public commentary) * Justification (why is this change necessary?): clarity * Proponents (who needs this change): Everyone **Originally proposed new attributes of the term**: * **Organized in Class**: ChronometricAge * **Term names**: maximumChronometricAge, minimumChronometricAge * **Definition of maximumChronometricAge**: Upper limit for the age (the earliest possible age) of a specimen as determined by a dating method. * **Definition of minimumChronometricAge**: Lower limit for the age (the latest possible age) of a specimen as determined by a dating method. <br> **Updated (current) proposed new attributes of the terms**: * **Organized in Class**: ChronometricAge * **Term name**: earliestChronometricAge * **Label**: Earliest Chronometric Age * **Definition**: The maximum/earliest/oldest possible age of a specimen as determined by a dating method. * **Comments**: The expected unit for this field is years. This field, if populated, must have an associated earliestChronometricAgeReferenceSystem. * **Example**: `100` <br> * **Term name**: latestChronometricAge * **Definition**: The minimum/latest/youngest possible age of a specimen as determined by a dating method. * **Comments**: The expected unit for this field is years. This field, if populated, must have an associated latestChronometricAgeReferenceSystem. * **Example**: `27` <br> The change of term names also requires changes to the corresponding terms for age reference systems. * **Term name**: earliestChronometricAgeReferenceSystem * **Definition**: The reference system associated with the earliestChronometricAge. * **Comments**: Recommended best practice is to use a controlled vocabulary. * **Examples**: `kya`,`mya`,`BP`,`AD`,`BCE`,`ka`,`Ma`,`Ga` <br> * **Term name**: latestChronometricAgeReferenceSystem * **Definition**: The reference system associated with the latestChronometricAge. * **Comments**: Recommended best practice is to use a controlled vocabulary. * **Examples**: `kya`,`mya`,`BP`,`AD`,`BCE`,`ka`,`Ma`,`Ga`
1.0
Change term - maximumChronometricAge, maximumChronometricAge - ## Change term * Submitter: John Wieczorek (from public commentary) * Justification (why is this change necessary?): clarity * Proponents (who needs this change): Everyone **Originally proposed new attributes of the term**: * **Organized in Class**: ChronometricAge * **Term names**: maximumChronometricAge, minimumChronometricAge * **Definition of maximumChronometricAge**: Upper limit for the age (the earliest possible age) of a specimen as determined by a dating method. * **Definition of minimumChronometricAge**: Lower limit for the age (the latest possible age) of a specimen as determined by a dating method. <br> **Updated (current) proposed new attributes of the terms**: * **Organized in Class**: ChronometricAge * **Term name**: earliestChronometricAge * **Label**: Earliest Chronometric Age * **Definition**: The maximum/earliest/oldest possible age of a specimen as determined by a dating method. * **Comments**: The expected unit for this field is years. This field, if populated, must have an associated earliestChronometricAgeReferenceSystem. * **Example**: `100` <br> * **Term name**: latestChronometricAge * **Definition**: The minimum/latest/youngest possible age of a specimen as determined by a dating method. * **Comments**: The expected unit for this field is years. This field, if populated, must have an associated latestChronometricAgeReferenceSystem. * **Example**: `27` <br> The change of term names also requires changes to the corresponding terms for age reference systems. * **Term name**: earliestChronometricAgeReferenceSystem * **Definition**: The reference system associated with the earliestChronometricAge. * **Comments**: Recommended best practice is to use a controlled vocabulary. * **Examples**: `kya`,`mya`,`BP`,`AD`,`BCE`,`ka`,`Ma`,`Ga` <br> * **Term name**: latestChronometricAgeReferenceSystem * **Definition**: The reference system associated with the latestChronometricAge. * **Comments**: Recommended best practice is to use a controlled vocabulary. * **Examples**: `kya`,`mya`,`BP`,`AD`,`BCE`,`ka`,`Ma`,`Ga`
process
change term maximumchronometricage maximumchronometricage change term submitter john wieczorek from public commentary justification why is this change necessary clarity proponents who needs this change everyone originally proposed new attributes of the term organized in class chronometricage term names maximumchronometricage minimumchronometricage definition of maximumchronometricage upper limit for the age the earliest possible age of a specimen as determined by a dating method definition of minimumchronometricage lower limit for the age the latest possible age of a specimen as determined by a dating method updated current proposed new attributes of the terms organized in class chronometricage term name earliestchronometricage label earliest chronometric age definition the maximum earliest oldest possible age of a specimen as determined by a dating method comments the expected unit for this field is years this field if populated must have an associated earliestchronometricagereferencesystem example term name latestchronometricage definition the minimum latest youngest possible age of a specimen as determined by a dating method comments the expected unit for this field is years this field if populated must have an associated latestchronometricagereferencesystem example the change of term names also requires changes to the corresponding terms for age reference systems term name earliestchronometricagereferencesystem definition the reference system associated with the earliestchronometricage comments recommended best practice is to use a controlled vocabulary examples kya mya bp ad bce ka ma ga term name latestchronometricagereferencesystem definition the reference system associated with the latestchronometricage comments recommended best practice is to use a controlled vocabulary examples kya mya bp ad bce ka ma ga
1
1,085
3,548,582,793
IssuesEvent
2016-01-20 15:01:00
LOVDnl/LOVD3
https://api.github.com/repos/LOVDnl/LOVD3
opened
Mesage: Please select a valid entry from the 'refseq_genomic' selection box should not be displayed.
import process
In object_genes the variable $aSelectRefseqGenomic is not set properly. It is only filled with the first NC that is found, hereafter the caching in the getForm function (if (!empty($this->aFormData)) { return parent::getForm(); }) cause that this variable is not further completed.
1.0
Mesage: Please select a valid entry from the 'refseq_genomic' selection box should not be displayed. - In object_genes the variable $aSelectRefseqGenomic is not set properly. It is only filled with the first NC that is found, hereafter the caching in the getForm function (if (!empty($this->aFormData)) { return parent::getForm(); }) cause that this variable is not further completed.
process
mesage please select a valid entry from the refseq genomic selection box should not be displayed in object genes the variable aselectrefseqgenomic is not set properly it is only filled with the first nc that is found hereafter the caching in the getform function if empty this aformdata return parent getform cause that this variable is not further completed
1
16,342
21,001,498,671
IssuesEvent
2022-03-29 17:55:00
googleapis/python-bigquery
https://api.github.com/repos/googleapis/python-bigquery
closed
sync v3 branch with main and cut another prerelease
api: bigquery type: process
It has been a while since the last sync and there have been several fixes that have gone into main. We should sync again and cut a prerelease for those folks that are helping us test with the new package. May want to wait until after https://github.com/googleapis/python-bigquery/pull/1125 as refactorings are the most likely to cause trouble?
1.0
sync v3 branch with main and cut another prerelease - It has been a while since the last sync and there have been several fixes that have gone into main. We should sync again and cut a prerelease for those folks that are helping us test with the new package. May want to wait until after https://github.com/googleapis/python-bigquery/pull/1125 as refactorings are the most likely to cause trouble?
process
sync branch with main and cut another prerelease it has been a while since the last sync and there have been several fixes that have gone into main we should sync again and cut a prerelease for those folks that are helping us test with the new package may want to wait until after as refactorings are the most likely to cause trouble
1
560,021
16,583,224,503
IssuesEvent
2021-05-31 14:37:30
huridocs/uwazi
https://api.github.com/repos/huridocs/uwazi
closed
Some detected labels in the ui are not translatable. (2 pts)
Priority: High Sprint
We detected some labels that are not translatable, this is the list we want to make them translatable via the inline translation system, probably there are more but these are the priority. - The submit label of the submit button in the public form. - The labels "Label" and "URL" of the link field.
1.0
Some detected labels in the ui are not translatable. (2 pts) - We detected some labels that are not translatable, this is the list we want to make them translatable via the inline translation system, probably there are more but these are the priority. - The submit label of the submit button in the public form. - The labels "Label" and "URL" of the link field.
non_process
some detected labels in the ui are not translatable pts we detected some labels that are not translatable this is the list we want to make them translatable via the inline translation system probably there are more but these are the priority the submit label of the submit button in the public form the labels label and url of the link field
0
138
2,575,026,064
IssuesEvent
2015-02-11 20:25:40
pyne/pyne
https://api.github.com/repos/pyne/pyne
closed
AHOT Integration
feature in process
The AHOT has been kindly contributed by Yousry Azmy. Now it needs integration. See more details on the mailing list [1]. 1. https://groups.google.com/d/topic/pyne-dev/rtAYzER_7_w/discussion
1.0
AHOT Integration - The AHOT has been kindly contributed by Yousry Azmy. Now it needs integration. See more details on the mailing list [1]. 1. https://groups.google.com/d/topic/pyne-dev/rtAYzER_7_w/discussion
process
ahot integration the ahot has been kindly contributed by yousry azmy now it needs integration see more details on the mailing list
1
91,568
11,520,827,548
IssuesEvent
2020-02-14 15:31:11
gsoft-inc/sg-orbit
https://api.github.com/repos/gsoft-inc/sg-orbit
opened
Focus states are the same as active and hover
bug report theming validate with design
This is problematic in situation where you click on a button that acts as a toggle, unselecting it ... This is also problematic for accessibility reasons. Explorations have been made with designers although nothing final. ![eHewCuCCVH](https://user-images.githubusercontent.com/361632/74544398-109f3c00-4f15-11ea-976c-efc686dac1b1.gif) ### Steps to reproduce 1. Hover a button 2. Click it 3. There is no difference between active and hover ### Expected behavior An active element has a different style.
1.0
Focus states are the same as active and hover - This is problematic in situation where you click on a button that acts as a toggle, unselecting it ... This is also problematic for accessibility reasons. Explorations have been made with designers although nothing final. ![eHewCuCCVH](https://user-images.githubusercontent.com/361632/74544398-109f3c00-4f15-11ea-976c-efc686dac1b1.gif) ### Steps to reproduce 1. Hover a button 2. Click it 3. There is no difference between active and hover ### Expected behavior An active element has a different style.
non_process
focus states are the same as active and hover this is problematic in situation where you click on a button that acts as a toggle unselecting it this is also problematic for accessibility reasons explorations have been made with designers although nothing final steps to reproduce hover a button click it there is no difference between active and hover expected behavior an active element has a different style
0
471,750
13,609,635,666
IssuesEvent
2020-09-23 05:47:15
open-telemetry/opentelemetry-java
https://api.github.com/repos/open-telemetry/opentelemetry-java
opened
Create a combined B3 Propagator that handle both single and multi-header cases
help wanted priority:p2 release:required-for-ga
See the recently updated spec: https://github.com/open-telemetry/opentelemetry-specification/blob/master/specification/context/api-propagators.md#b3-requirements
1.0
Create a combined B3 Propagator that handle both single and multi-header cases - See the recently updated spec: https://github.com/open-telemetry/opentelemetry-specification/blob/master/specification/context/api-propagators.md#b3-requirements
non_process
create a combined propagator that handle both single and multi header cases see the recently updated spec
0
17,278
23,069,689,825
IssuesEvent
2022-07-25 16:48:58
retaildevcrews/pib-cli
https://api.github.com/repos/retaildevcrews/pib-cli
closed
Sprint 1 Goals:
Process
Customer Support: - [x] WMT - ~KiC HoL #280 (slipped - customer)~ - [x] KFC - [x] Failover design #590 - [x] McD/AD In-Person Support/Demo: - [x] McD EBC #281 - [x] Modify .net w/KV #535 - [x] Arc - GO ext #491 - [x] Arc KV ext #489 - [x] Arc Demo #213 ------------------------------------------- PG Transition: - [x] Grafana Sub/PG #204 - [x] Groups GH/MS #486 - [x] Demo Env #205 - [x] Align trackers #573 - [x] create pib-dev #575 ------------------------------------------- EngFun: - [x] Delete Repo #1 (INDIA) - [x] Repo Cleanup/Plan #438 - [x] Finalize MVP scope #488 ------------------------------------------- PiB MVP Technical Debt: - [x] gitops-template #540 - [x] latest docker image #507, #508 - [x] Consolidate pib-webv #501 - [x] Consolidate VOE fleet #485 Bugs: - [x] GitOps Auto #519 CLI: - [x] CLI #487 DevExp: - ~App-new #444 (pushed to S2)~ - [x] voe-app GH #478
1.0
Sprint 1 Goals: - Customer Support: - [x] WMT - ~KiC HoL #280 (slipped - customer)~ - [x] KFC - [x] Failover design #590 - [x] McD/AD In-Person Support/Demo: - [x] McD EBC #281 - [x] Modify .net w/KV #535 - [x] Arc - GO ext #491 - [x] Arc KV ext #489 - [x] Arc Demo #213 ------------------------------------------- PG Transition: - [x] Grafana Sub/PG #204 - [x] Groups GH/MS #486 - [x] Demo Env #205 - [x] Align trackers #573 - [x] create pib-dev #575 ------------------------------------------- EngFun: - [x] Delete Repo #1 (INDIA) - [x] Repo Cleanup/Plan #438 - [x] Finalize MVP scope #488 ------------------------------------------- PiB MVP Technical Debt: - [x] gitops-template #540 - [x] latest docker image #507, #508 - [x] Consolidate pib-webv #501 - [x] Consolidate VOE fleet #485 Bugs: - [x] GitOps Auto #519 CLI: - [x] CLI #487 DevExp: - ~App-new #444 (pushed to S2)~ - [x] voe-app GH #478
process
sprint goals customer support wmt kic hol slipped customer kfc failover design mcd ad in person support demo mcd ebc modify net w kv arc go ext arc kv ext arc demo pg transition grafana sub pg groups gh ms demo env align trackers create pib dev engfun delete repo india repo cleanup plan finalize mvp scope pib mvp technical debt gitops template latest docker image consolidate pib webv consolidate voe fleet bugs gitops auto cli cli devexp app new pushed to voe app gh
1
20,046
26,534,900,823
IssuesEvent
2023-01-19 15:01:06
apache/arrow-datafusion
https://api.github.com/repos/apache/arrow-datafusion
opened
Changelog generator not working for patch releases
bug development-process
**Describe the bug** I could not get the changelog script to generate an accurate changelog for the maint-16.x branch. It showed changes from the master branch that are not in maint-16.x **To Reproduce** Try doing the above **Expected behavior** Generate correct changelog **Additional context**
1.0
Changelog generator not working for patch releases - **Describe the bug** I could not get the changelog script to generate an accurate changelog for the maint-16.x branch. It showed changes from the master branch that are not in maint-16.x **To Reproduce** Try doing the above **Expected behavior** Generate correct changelog **Additional context**
process
changelog generator not working for patch releases describe the bug i could not get the changelog script to generate an accurate changelog for the maint x branch it showed changes from the master branch that are not in maint x to reproduce try doing the above expected behavior generate correct changelog additional context
1
11,314
14,120,967,081
IssuesEvent
2020-11-09 00:19:21
tommy-josepovic/smarthome-simulator-team-2
https://api.github.com/repos/tommy-josepovic/smarthome-simulator-team-2
closed
D2 - Architecture design documentation - Architecture
software process
The architectural design should be reflected in the implementation of well-separated modules and/or folders. Implement at least one new creational and one new structural design pattern and clearly motivate your design decision
1.0
D2 - Architecture design documentation - Architecture - The architectural design should be reflected in the implementation of well-separated modules and/or folders. Implement at least one new creational and one new structural design pattern and clearly motivate your design decision
process
architecture design documentation architecture the architectural design should be reflected in the implementation of well separated modules and or folders implement at least one new creational and one new structural design pattern and clearly motivate your design decision
1
13,355
15,817,792,862
IssuesEvent
2021-04-05 15:06:03
retaildevcrews/ngsa
https://api.github.com/repos/retaildevcrews/ngsa
opened
Engineering Fundamentals Checklist
Pri1 Process
# Tech Lead's Engineering Fundamentals Checklist This checklist helps to ensure that our projects meet our Engineering Fundamentals. ## Source Control - [ ] The main branch is locked. - [ ] Merges are done through PRs. - [ ] PRs reference related work items. - [ ] Commit history is consistent and commit messages are informative (what, why). - [ ] Secrets are not part of the commit history or made public. (see [Credential scanning](https://github.com/microsoft/code-with-engineering-playbook/tree/master/continuous-integration/credential-scanning/readme.md)) - [ ] Public repositories follow the [OSS guidelines](https://github.com/microsoft/code-with-engineering-playbook/tree/master/source-control/readme.md#creating-a-new-repository), see `Required files in default branch for public repositories`. More details on [Source Control](https://github.com/microsoft/code-with-engineering-playbook/tree/master/source-control/readme.md) ## Work Item Tracking - [ ] All items are tracked in AzDevOps (or similar). - [ ] The board is organized (swim lanes, feature tags, technology tags). ## Testing - [ ] Unit tests cover the majority of all components (>90% if possible). - [ ] Integration tests run to test the solution e2e. More details on [Unit Testing](https://github.com/microsoft/code-with-engineering-playbook/tree/master/automated-testing/unit-testing/readme.md) ## CI/CD - [ ] Project runs CI with automated build and test on each PR. - [ ] Project uses CD to manage deployments to a replica environment before PRs are merged. - [ ] Main branch is always shippable. ## Security - TO DO - [ ] Access control. - [ ] Separation of concerns. - [ ] Robust treatment of secrets. - [ ] Encryption for data in transit (and if necessary at rest) and password hashing. ## Observability - [ ] Significant business and functional events are tracked and related metrics collected. - [ ] Application faults and errors are logged. - [ ] Health of the system is monitored. - [ ] The client and server side observability data can be differentiated. - [ ] Logging configuration can be modified without code changes (eg: verbose mode). - [ ] [Incoming tracing context](https://github.com/microsoft/code-with-engineering-playbook/tree/master/observability/correlation-id.md) is propagated to allow for production issue debugging purposes. - [ ] GDPR compliance is ensured regarding PII (Personally Identifiable Information). ## Agile/Scrum - [ ] Process Lead (fixed/rotating) to run standup daily. - [ ] Agile process clearly defined within team. - [ ] Tech Lead (+ PO/Others) have responsibility for backlog management and grooming. - [ ] Working agreement between members of team and customer. ## Design Reviews - [ ] Process for conducting design reviews is included in the [Working Agreement](https://github.com/microsoft/code-with-engineering-playbook/tree/master/agile-development/team-agreements/working-agreements/readme.md) - [ ] Design reviews for each major component of the solution are carried out and documented, including alternatives. - [ ] Stories and/or PRs link to the design document. - [ ] Each user story includes a task for design review by default, which is assigned or removed during sprint planning. - [ ] Project advisors are invited to design reviews or asked to give feedback to the design decisions captured in documentation. - [ ] Discover all the reviews that the customer's processes require and plan for them. ## Code Reviews - [ ] Clear agreement in the team as to function of code reviews. - [ ] Code review checklist or established process. - [ ] A minimum number of reviewers (usually 2) for a PR merge is enforced by policy. - [ ] Linters/Code Analyzers, unit tests and successful builds for PR merges are set up. - [ ] Process to enforce a quick review turnaround. More details on [Code Reviews](https://github.com/microsoft/code-with-engineering-playbook/tree/master/code-reviews/README.md) ## Retrospectives - [ ] Set time for retrospectives each week/at the end of each sprint. - [ ] 1-3 proposed experiments to be tried each week/sprint to improve the process. - [ ] Experiments have owners and are added to project backlog. - [ ] Longer retrospective for Milestones and project completion. More details on [Retrospectives](https://github.com/microsoft/code-with-engineering-playbook/tree/master/agile-development/retrospectives/readme.md) ## Engineering Feedback - [ ] Submit business and technical blockers that prevent project success - [ ] Add suggestions for improvements to leveraged services and components - [ ] Ensure feedback is detailed and repeatable More details on [Engineering Feedback](https://github.com/microsoft/code-with-engineering-playbook/tree/master/engineering-feedback/readme.md)
1.0
Engineering Fundamentals Checklist - # Tech Lead's Engineering Fundamentals Checklist This checklist helps to ensure that our projects meet our Engineering Fundamentals. ## Source Control - [ ] The main branch is locked. - [ ] Merges are done through PRs. - [ ] PRs reference related work items. - [ ] Commit history is consistent and commit messages are informative (what, why). - [ ] Secrets are not part of the commit history or made public. (see [Credential scanning](https://github.com/microsoft/code-with-engineering-playbook/tree/master/continuous-integration/credential-scanning/readme.md)) - [ ] Public repositories follow the [OSS guidelines](https://github.com/microsoft/code-with-engineering-playbook/tree/master/source-control/readme.md#creating-a-new-repository), see `Required files in default branch for public repositories`. More details on [Source Control](https://github.com/microsoft/code-with-engineering-playbook/tree/master/source-control/readme.md) ## Work Item Tracking - [ ] All items are tracked in AzDevOps (or similar). - [ ] The board is organized (swim lanes, feature tags, technology tags). ## Testing - [ ] Unit tests cover the majority of all components (>90% if possible). - [ ] Integration tests run to test the solution e2e. More details on [Unit Testing](https://github.com/microsoft/code-with-engineering-playbook/tree/master/automated-testing/unit-testing/readme.md) ## CI/CD - [ ] Project runs CI with automated build and test on each PR. - [ ] Project uses CD to manage deployments to a replica environment before PRs are merged. - [ ] Main branch is always shippable. ## Security - TO DO - [ ] Access control. - [ ] Separation of concerns. - [ ] Robust treatment of secrets. - [ ] Encryption for data in transit (and if necessary at rest) and password hashing. ## Observability - [ ] Significant business and functional events are tracked and related metrics collected. - [ ] Application faults and errors are logged. - [ ] Health of the system is monitored. - [ ] The client and server side observability data can be differentiated. - [ ] Logging configuration can be modified without code changes (eg: verbose mode). - [ ] [Incoming tracing context](https://github.com/microsoft/code-with-engineering-playbook/tree/master/observability/correlation-id.md) is propagated to allow for production issue debugging purposes. - [ ] GDPR compliance is ensured regarding PII (Personally Identifiable Information). ## Agile/Scrum - [ ] Process Lead (fixed/rotating) to run standup daily. - [ ] Agile process clearly defined within team. - [ ] Tech Lead (+ PO/Others) have responsibility for backlog management and grooming. - [ ] Working agreement between members of team and customer. ## Design Reviews - [ ] Process for conducting design reviews is included in the [Working Agreement](https://github.com/microsoft/code-with-engineering-playbook/tree/master/agile-development/team-agreements/working-agreements/readme.md) - [ ] Design reviews for each major component of the solution are carried out and documented, including alternatives. - [ ] Stories and/or PRs link to the design document. - [ ] Each user story includes a task for design review by default, which is assigned or removed during sprint planning. - [ ] Project advisors are invited to design reviews or asked to give feedback to the design decisions captured in documentation. - [ ] Discover all the reviews that the customer's processes require and plan for them. ## Code Reviews - [ ] Clear agreement in the team as to function of code reviews. - [ ] Code review checklist or established process. - [ ] A minimum number of reviewers (usually 2) for a PR merge is enforced by policy. - [ ] Linters/Code Analyzers, unit tests and successful builds for PR merges are set up. - [ ] Process to enforce a quick review turnaround. More details on [Code Reviews](https://github.com/microsoft/code-with-engineering-playbook/tree/master/code-reviews/README.md) ## Retrospectives - [ ] Set time for retrospectives each week/at the end of each sprint. - [ ] 1-3 proposed experiments to be tried each week/sprint to improve the process. - [ ] Experiments have owners and are added to project backlog. - [ ] Longer retrospective for Milestones and project completion. More details on [Retrospectives](https://github.com/microsoft/code-with-engineering-playbook/tree/master/agile-development/retrospectives/readme.md) ## Engineering Feedback - [ ] Submit business and technical blockers that prevent project success - [ ] Add suggestions for improvements to leveraged services and components - [ ] Ensure feedback is detailed and repeatable More details on [Engineering Feedback](https://github.com/microsoft/code-with-engineering-playbook/tree/master/engineering-feedback/readme.md)
process
engineering fundamentals checklist tech lead s engineering fundamentals checklist this checklist helps to ensure that our projects meet our engineering fundamentals source control the main branch is locked merges are done through prs prs reference related work items commit history is consistent and commit messages are informative what why secrets are not part of the commit history or made public see public repositories follow the see required files in default branch for public repositories more details on work item tracking all items are tracked in azdevops or similar the board is organized swim lanes feature tags technology tags testing unit tests cover the majority of all components if possible integration tests run to test the solution more details on ci cd project runs ci with automated build and test on each pr project uses cd to manage deployments to a replica environment before prs are merged main branch is always shippable security to do access control separation of concerns robust treatment of secrets encryption for data in transit and if necessary at rest and password hashing observability significant business and functional events are tracked and related metrics collected application faults and errors are logged health of the system is monitored the client and server side observability data can be differentiated logging configuration can be modified without code changes eg verbose mode is propagated to allow for production issue debugging purposes gdpr compliance is ensured regarding pii personally identifiable information agile scrum process lead fixed rotating to run standup daily agile process clearly defined within team tech lead po others have responsibility for backlog management and grooming working agreement between members of team and customer design reviews process for conducting design reviews is included in the design reviews for each major component of the solution are carried out and documented including alternatives stories and or prs link to the design document each user story includes a task for design review by default which is assigned or removed during sprint planning project advisors are invited to design reviews or asked to give feedback to the design decisions captured in documentation discover all the reviews that the customer s processes require and plan for them code reviews clear agreement in the team as to function of code reviews code review checklist or established process a minimum number of reviewers usually for a pr merge is enforced by policy linters code analyzers unit tests and successful builds for pr merges are set up process to enforce a quick review turnaround more details on retrospectives set time for retrospectives each week at the end of each sprint proposed experiments to be tried each week sprint to improve the process experiments have owners and are added to project backlog longer retrospective for milestones and project completion more details on engineering feedback submit business and technical blockers that prevent project success add suggestions for improvements to leveraged services and components ensure feedback is detailed and repeatable more details on
1
5,660
8,529,379,530
IssuesEvent
2018-11-03 11:34:00
pelias/schema
https://api.github.com/repos/pelias/schema
closed
Explore using not_analyzed for source
experiment low priority processed
The source field currently uses the "keyword" analyzer, which basically keeps the full string as a single token with no changes. According to [the keyword analyzer](https://www.elastic.co/guide/en/elasticsearch/reference/current/analysis-keyword-analyzer.html) docs, it might make more sense to use the "not_analyzed" setting, briefly touched on [here](https://www.elastic.co/guide/en/elasticsearch/reference/1.7/mapping-core-types.html) in the docs. It seems like it might do the same thing while somehow being faster.
1.0
Explore using not_analyzed for source - The source field currently uses the "keyword" analyzer, which basically keeps the full string as a single token with no changes. According to [the keyword analyzer](https://www.elastic.co/guide/en/elasticsearch/reference/current/analysis-keyword-analyzer.html) docs, it might make more sense to use the "not_analyzed" setting, briefly touched on [here](https://www.elastic.co/guide/en/elasticsearch/reference/1.7/mapping-core-types.html) in the docs. It seems like it might do the same thing while somehow being faster.
process
explore using not analyzed for source the source field currently uses the keyword analyzer which basically keeps the full string as a single token with no changes according to docs it might make more sense to use the not analyzed setting briefly touched on in the docs it seems like it might do the same thing while somehow being faster
1
31,493
5,955,560,000
IssuesEvent
2017-05-28 07:49:03
sinonjs/sinon
https://api.github.com/repos/sinonjs/sinon
closed
Behavior of sandbox creation inconsistent with documentation.
Documentation
The provided [documentation](https://github.com/cjohansen/sinon-web/blob/master/docs/index.html#L1482-L1502) seems to indicate that a call to `sinon.sandbox.create` will create a sandbox with a default configuration (namely, enabled fake timers and fake server). However, this is inconsistent with the actual behavior. See [example fiddle](http://jsfiddle.net/8k0zv6tp/). Instead, an unconfigured instance is returned with `sinon.sandbox` in the prototype chain ([link](https://github.com/cjohansen/Sinon.JS/blob/master/lib/sinon/sandbox.js#L110-L112)). It doesn't appear as if there is any default configuration behavior for sandboxes, so this section of the documentation is misleading at best. Either the documentation or the behavior should be updated to be consistent with the other. I'm happy to submit a PR for either of these cases, if you can point me in the preferred direction.
1.0
Behavior of sandbox creation inconsistent with documentation. - The provided [documentation](https://github.com/cjohansen/sinon-web/blob/master/docs/index.html#L1482-L1502) seems to indicate that a call to `sinon.sandbox.create` will create a sandbox with a default configuration (namely, enabled fake timers and fake server). However, this is inconsistent with the actual behavior. See [example fiddle](http://jsfiddle.net/8k0zv6tp/). Instead, an unconfigured instance is returned with `sinon.sandbox` in the prototype chain ([link](https://github.com/cjohansen/Sinon.JS/blob/master/lib/sinon/sandbox.js#L110-L112)). It doesn't appear as if there is any default configuration behavior for sandboxes, so this section of the documentation is misleading at best. Either the documentation or the behavior should be updated to be consistent with the other. I'm happy to submit a PR for either of these cases, if you can point me in the preferred direction.
non_process
behavior of sandbox creation inconsistent with documentation the provided seems to indicate that a call to sinon sandbox create will create a sandbox with a default configuration namely enabled fake timers and fake server however this is inconsistent with the actual behavior see instead an unconfigured instance is returned with sinon sandbox in the prototype chain it doesn t appear as if there is any default configuration behavior for sandboxes so this section of the documentation is misleading at best either the documentation or the behavior should be updated to be consistent with the other i m happy to submit a pr for either of these cases if you can point me in the preferred direction
0
326,746
24,099,535,767
IssuesEvent
2022-09-19 22:22:31
atlanticwave-sdx/sdx-controller-client
https://api.github.com/repos/atlanticwave-sdx/sdx-controller-client
opened
Update README
documentation
Our README is missing a few things from what's expected of such a document: - A basic introduction - A tested and up-to-date installation method - Maybe a description of the Swagger definition - The link to https://virtserver.swaggerhub.com/SDX-Controller/1.0.0 is broken
1.0
Update README - Our README is missing a few things from what's expected of such a document: - A basic introduction - A tested and up-to-date installation method - Maybe a description of the Swagger definition - The link to https://virtserver.swaggerhub.com/SDX-Controller/1.0.0 is broken
non_process
update readme our readme is missing a few things from what s expected of such a document a basic introduction a tested and up to date installation method maybe a description of the swagger definition the link to is broken
0
16,103
20,329,409,046
IssuesEvent
2022-02-18 09:15:18
googlefonts/noto-fonts
https://api.github.com/repos/googlefonts/noto-fonts
closed
Question: Is it possible to have the whole Noto Fonts collection (CJK + non-CJK) as a .TTC file?
Noto-Process-Issue
I'm using Python Pillow to generate images containing text in multiple languages and scripts. Python Pillow doesn't allow fallback fonts—the single font file must contain all the glyphs that will be used in the text. Right now I'm using Adobe's Source Han Sans, which allows me to use CJK scripts and Latin/Cyrillic/Greek scripts. The font collection is packaged in a TTC file. I noticed that Noto-CJK fonts are also available as a single TTC file. Would it be possible to compile the Noto font collection, with CJK and non-CJK scripts, as a single TTC file? I understand that there would be some choices to be made (`*-UI` fonts for Arabic scripts, for example), but apart from that, does a TTC file have the same glyphs number limit as a TTF file? Thank you!
1.0
Question: Is it possible to have the whole Noto Fonts collection (CJK + non-CJK) as a .TTC file? - I'm using Python Pillow to generate images containing text in multiple languages and scripts. Python Pillow doesn't allow fallback fonts—the single font file must contain all the glyphs that will be used in the text. Right now I'm using Adobe's Source Han Sans, which allows me to use CJK scripts and Latin/Cyrillic/Greek scripts. The font collection is packaged in a TTC file. I noticed that Noto-CJK fonts are also available as a single TTC file. Would it be possible to compile the Noto font collection, with CJK and non-CJK scripts, as a single TTC file? I understand that there would be some choices to be made (`*-UI` fonts for Arabic scripts, for example), but apart from that, does a TTC file have the same glyphs number limit as a TTF file? Thank you!
process
question is it possible to have the whole noto fonts collection cjk non cjk as a ttc file i m using python pillow to generate images containing text in multiple languages and scripts python pillow doesn t allow fallback fonts—the single font file must contain all the glyphs that will be used in the text right now i m using adobe s source han sans which allows me to use cjk scripts and latin cyrillic greek scripts the font collection is packaged in a ttc file i noticed that noto cjk fonts are also available as a single ttc file would it be possible to compile the noto font collection with cjk and non cjk scripts as a single ttc file i understand that there would be some choices to be made ui fonts for arabic scripts for example but apart from that does a ttc file have the same glyphs number limit as a ttf file thank you
1
62,543
26,030,049,932
IssuesEvent
2022-12-21 20:11:29
aws/amazon-chime-sdk-js
https://api.github.com/repos/aws/amazon-chime-sdk-js
closed
Update Media capture pipeline to record constant frame rate to avoid audio sync issues
feature-request capture-service
# Community Note * Please vote for this issue by adding a 👍 reaction to the issue. This will help the community and maintainers from the Amazon Chime SDK team to prioritize this request. * Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request. * Please go through the existing [issues](https://github.com/aws/amazon-chime-sdk-js/issues) to see if your request has already been recommended. * Please do not use this template for bugs or support. Each individual repository has a PR template, documentation and resources for [troubleshooting](https://github.com/aws/amazon-chime-sdk-js#troubleshooting-and-support). * If you are interested in working on this issue or have submitted a pull request, please leave a comment. # Tell us about your request ## What do you want us to build? Which Amazon Chime SDK or feature area is this request for? Such as: [amazon-chime-sdk-js](https://github.com/aws/amazon-chime-sdk-js), [amazon-chime-sdk-ios](https://github.com/aws/amazon-chime-sdk-ios), [amazon-chime-sdk-android](https://github.com/aws/amazon-chime-sdk-android), [amazon-chime-sdk-component-library-react](https://github.com/aws/amazon-chime-sdk-component-library-react), a new feature idea. Mediapipeline recording audio goes out of sync as time goes when you are trying to use a video editor. This is caused by the recording format being VFR which doesn't have wide support with editors. ## Tell us about the problem you are trying to solve and why is it hard? The more details you provide us, the better we will be able to help you. Currently the process to use this backup recording is: 1. Download all segments 2. Generate a list of the video files and concatenate to a single mp4 file (FFmpeg) 3. Run it through handbrake to convert it from VFR to CFR https://handbrake.fr/ Edit/cut (Unsure if this was also the issue with the ffmpeg recordings) ## How are you currently solving a problem? Are there any workarounds or mitigations you have used? Why is it not working for you? ## Additional context Anything else we should know?
1.0
Update Media capture pipeline to record constant frame rate to avoid audio sync issues - # Community Note * Please vote for this issue by adding a 👍 reaction to the issue. This will help the community and maintainers from the Amazon Chime SDK team to prioritize this request. * Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request. * Please go through the existing [issues](https://github.com/aws/amazon-chime-sdk-js/issues) to see if your request has already been recommended. * Please do not use this template for bugs or support. Each individual repository has a PR template, documentation and resources for [troubleshooting](https://github.com/aws/amazon-chime-sdk-js#troubleshooting-and-support). * If you are interested in working on this issue or have submitted a pull request, please leave a comment. # Tell us about your request ## What do you want us to build? Which Amazon Chime SDK or feature area is this request for? Such as: [amazon-chime-sdk-js](https://github.com/aws/amazon-chime-sdk-js), [amazon-chime-sdk-ios](https://github.com/aws/amazon-chime-sdk-ios), [amazon-chime-sdk-android](https://github.com/aws/amazon-chime-sdk-android), [amazon-chime-sdk-component-library-react](https://github.com/aws/amazon-chime-sdk-component-library-react), a new feature idea. Mediapipeline recording audio goes out of sync as time goes when you are trying to use a video editor. This is caused by the recording format being VFR which doesn't have wide support with editors. ## Tell us about the problem you are trying to solve and why is it hard? The more details you provide us, the better we will be able to help you. Currently the process to use this backup recording is: 1. Download all segments 2. Generate a list of the video files and concatenate to a single mp4 file (FFmpeg) 3. Run it through handbrake to convert it from VFR to CFR https://handbrake.fr/ Edit/cut (Unsure if this was also the issue with the ffmpeg recordings) ## How are you currently solving a problem? Are there any workarounds or mitigations you have used? Why is it not working for you? ## Additional context Anything else we should know?
non_process
update media capture pipeline to record constant frame rate to avoid audio sync issues community note please vote for this issue by adding a 👍 reaction to the issue this will help the community and maintainers from the amazon chime sdk team to prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request please go through the existing to see if your request has already been recommended please do not use this template for bugs or support each individual repository has a pr template documentation and resources for if you are interested in working on this issue or have submitted a pull request please leave a comment tell us about your request what do you want us to build which amazon chime sdk or feature area is this request for such as a new feature idea mediapipeline recording audio goes out of sync as time goes when you are trying to use a video editor this is caused by the recording format being vfr which doesn t have wide support with editors tell us about the problem you are trying to solve and why is it hard the more details you provide us the better we will be able to help you currently the process to use this backup recording is download all segments generate a list of the video files and concatenate to a single file ffmpeg run it through handbrake to convert it from vfr to cfr edit cut unsure if this was also the issue with the ffmpeg recordings how are you currently solving a problem are there any workarounds or mitigations you have used why is it not working for you additional context anything else we should know
0
23,702
22,613,456,956
IssuesEvent
2022-06-29 19:25:03
forcedotcom/salesforcedx-vscode
https://api.github.com/repos/forcedotcom/salesforcedx-vscode
closed
Error message when clicking on Apex error : 'The editor could not be opened because the file was not found.'
type:bug area:apex area:usability
<!-- NOTICE: While GitHub is the preferred channel for reporting issues/feedback, this is not a mechanism for receiving support under any agreement or SLA. If you require immediate assistance, please use official support channels. --> <!-- FOR BUGS RELATED TO THE SALEFORCE CLI, please use this repository: https://github.com/forcedotcom/cli --> ### Summary _Short summary of what is going on or to provide context_. ### Steps To Reproduce: 1. Create a project with manifest. 2. Authorize and Org 3. Write an Apex code that has a few errors in it and save it. 4. Right click on the file and select SFDX: Deploy Source to Org 5. Click on any of the issues that show in the VS Code PROBLEMS tab. ### Expected result It should have opened the file with the error and taken me to the row number that had the error. ### Actual result Error message pops up on a different tab showing the message : 'The editor could not be opened because the file was not found.' ### Additional information ![VS Code error](https://user-images.githubusercontent.com/23403403/175095631-6f5b2005-411e-4e72-a6e4-b4743811341f.jpg)free to attach a screenshot_. **Salesforce Extension Version in VS Code**: **SFDX CLI Version**: sfdx-cli/7.155.1 win32-x86 node-v16.15.1 **OS and version**: OS Name: Microsoft Windows 11 Home OS Version: 10.0.22000 N/A Build 22000
True
Error message when clicking on Apex error : 'The editor could not be opened because the file was not found.' - <!-- NOTICE: While GitHub is the preferred channel for reporting issues/feedback, this is not a mechanism for receiving support under any agreement or SLA. If you require immediate assistance, please use official support channels. --> <!-- FOR BUGS RELATED TO THE SALEFORCE CLI, please use this repository: https://github.com/forcedotcom/cli --> ### Summary _Short summary of what is going on or to provide context_. ### Steps To Reproduce: 1. Create a project with manifest. 2. Authorize and Org 3. Write an Apex code that has a few errors in it and save it. 4. Right click on the file and select SFDX: Deploy Source to Org 5. Click on any of the issues that show in the VS Code PROBLEMS tab. ### Expected result It should have opened the file with the error and taken me to the row number that had the error. ### Actual result Error message pops up on a different tab showing the message : 'The editor could not be opened because the file was not found.' ### Additional information ![VS Code error](https://user-images.githubusercontent.com/23403403/175095631-6f5b2005-411e-4e72-a6e4-b4743811341f.jpg)free to attach a screenshot_. **Salesforce Extension Version in VS Code**: **SFDX CLI Version**: sfdx-cli/7.155.1 win32-x86 node-v16.15.1 **OS and version**: OS Name: Microsoft Windows 11 Home OS Version: 10.0.22000 N/A Build 22000
non_process
error message when clicking on apex error the editor could not be opened because the file was not found notice while github is the preferred channel for reporting issues feedback this is not a mechanism for receiving support under any agreement or sla if you require immediate assistance please use official support channels for bugs related to the saleforce cli please use this repository summary short summary of what is going on or to provide context steps to reproduce create a project with manifest authorize and org write an apex code that has a few errors in it and save it right click on the file and select sfdx deploy source to org click on any of the issues that show in the vs code problems tab expected result it should have opened the file with the error and taken me to the row number that had the error actual result error message pops up on a different tab showing the message the editor could not be opened because the file was not found additional information to attach a screenshot salesforce extension version in vs code sfdx cli version sfdx cli node os and version os name microsoft windows home os version n a build
0
17,138
22,677,601,190
IssuesEvent
2022-07-04 06:53:05
quark-engine/quark-engine
https://api.github.com/repos/quark-engine/quark-engine
opened
Update README for the recently released features.
issue-processing-state-06
Recently, the team added many features to Quark (Rule Viewer, Web Report, and RadioContrast API). To illustrate the power of these features, we documented them in the README with examples. However, it also causes some problems with the file. 1. Outdated information. - For example, the command introduced in the Detail Report section is wrong. 2. Lengthy content. - Users need to scroll eight times to find the installation steps. 3. Unclear layout. - The file does not have an overview of the Quark features. Hence, we may need a complete update on this file, including **cleaning up outdated information**, **simplifying the content**, and **adjusting the layout**.
1.0
Update README for the recently released features. - Recently, the team added many features to Quark (Rule Viewer, Web Report, and RadioContrast API). To illustrate the power of these features, we documented them in the README with examples. However, it also causes some problems with the file. 1. Outdated information. - For example, the command introduced in the Detail Report section is wrong. 2. Lengthy content. - Users need to scroll eight times to find the installation steps. 3. Unclear layout. - The file does not have an overview of the Quark features. Hence, we may need a complete update on this file, including **cleaning up outdated information**, **simplifying the content**, and **adjusting the layout**.
process
update readme for the recently released features recently the team added many features to quark rule viewer web report and radiocontrast api to illustrate the power of these features we documented them in the readme with examples however it also causes some problems with the file outdated information for example the command introduced in the detail report section is wrong lengthy content users need to scroll eight times to find the installation steps unclear layout the file does not have an overview of the quark features hence we may need a complete update on this file including cleaning up outdated information simplifying the content and adjusting the layout
1
13,129
15,529,216,061
IssuesEvent
2021-03-13 14:20:35
bow-simulation/virtualbow
https://api.github.com/repos/bow-simulation/virtualbow
opened
Some of the icons don't work as svg on macOS
area: software process platform: macos type: bug
* background.svg * run-statics.svg * run-dynamics.svg * view-*.svg They don't show up in the application, only when they're converted to png beforehand. Only the background image produces the output `Coould not create pixmap from background.svg`, the others fail silently.
1.0
Some of the icons don't work as svg on macOS - * background.svg * run-statics.svg * run-dynamics.svg * view-*.svg They don't show up in the application, only when they're converted to png beforehand. Only the background image produces the output `Coould not create pixmap from background.svg`, the others fail silently.
process
some of the icons don t work as svg on macos background svg run statics svg run dynamics svg view svg they don t show up in the application only when they re converted to png beforehand only the background image produces the output coould not create pixmap from background svg the others fail silently
1
14,431
17,480,988,377
IssuesEvent
2021-08-09 02:12:42
Leviatan-Analytics/LA-data-processing
https://api.github.com/repos/Leviatan-Analytics/LA-data-processing
closed
Adapt text recognition model output to match defined schema [2]
Data Processing Week 1 Sprint 3
Covert model detection output to defined json data schema to export.
1.0
Adapt text recognition model output to match defined schema [2] - Covert model detection output to defined json data schema to export.
process
adapt text recognition model output to match defined schema covert model detection output to defined json data schema to export
1
475,154
13,687,941,672
IssuesEvent
2020-09-30 10:54:26
openxpki/openxpki
https://api.github.com/repos/openxpki/openxpki
closed
Support non-ASCII characters (i.e. UTF8) in cert search
bug high priority
Currently, search fails (program error) if e.g. umlauts are entered.
1.0
Support non-ASCII characters (i.e. UTF8) in cert search - Currently, search fails (program error) if e.g. umlauts are entered.
non_process
support non ascii characters i e in cert search currently search fails program error if e g umlauts are entered
0
8,642
11,788,318,004
IssuesEvent
2020-03-17 15:23:24
digitalmethodsinitiative/4cat
https://api.github.com/repos/digitalmethodsinitiative/4cat
closed
Make 4CAT available as a general PENELOPE workbench-like interface
big meta processors
Some random initial thoughts on this - Post-processors and data sources are uploaded to a central repository, separate from 4cat code, and decoupled from it as much as possible - The repository is essentially a set of links to a Github repository along with some metadata for each of these links such as a description, name, maybe a logo, et cetera - Such post-processors and data sources are expected to implement a particular API that is developed and maintained as part of PENELOPE - To mesh with existing OpenAPI-based components, a generic OpenAPI post-processor can be made available that allows interacting with any of these components through a generic interface, though realistically this would sometimes be at the cost of decreased user-friendliness in which case more tailor-made post-processors can be offered. - 4CAT gets a feature that allows it to use any post-processor and data source implementing this common API as part of its existing query-and-analyse pipeline and interface - Other interfaces can be built on top of the same code and people can install their own versions of e.g. 4CAT and only install the data sources and post-processors relevant to them Organisation-wise: - We fix outstanding 1.0 issues first - After this, for 2.0, we create a new branch
1.0
Make 4CAT available as a general PENELOPE workbench-like interface - Some random initial thoughts on this - Post-processors and data sources are uploaded to a central repository, separate from 4cat code, and decoupled from it as much as possible - The repository is essentially a set of links to a Github repository along with some metadata for each of these links such as a description, name, maybe a logo, et cetera - Such post-processors and data sources are expected to implement a particular API that is developed and maintained as part of PENELOPE - To mesh with existing OpenAPI-based components, a generic OpenAPI post-processor can be made available that allows interacting with any of these components through a generic interface, though realistically this would sometimes be at the cost of decreased user-friendliness in which case more tailor-made post-processors can be offered. - 4CAT gets a feature that allows it to use any post-processor and data source implementing this common API as part of its existing query-and-analyse pipeline and interface - Other interfaces can be built on top of the same code and people can install their own versions of e.g. 4CAT and only install the data sources and post-processors relevant to them Organisation-wise: - We fix outstanding 1.0 issues first - After this, for 2.0, we create a new branch
process
make available as a general penelope workbench like interface some random initial thoughts on this post processors and data sources are uploaded to a central repository separate from code and decoupled from it as much as possible the repository is essentially a set of links to a github repository along with some metadata for each of these links such as a description name maybe a logo et cetera such post processors and data sources are expected to implement a particular api that is developed and maintained as part of penelope to mesh with existing openapi based components a generic openapi post processor can be made available that allows interacting with any of these components through a generic interface though realistically this would sometimes be at the cost of decreased user friendliness in which case more tailor made post processors can be offered gets a feature that allows it to use any post processor and data source implementing this common api as part of its existing query and analyse pipeline and interface other interfaces can be built on top of the same code and people can install their own versions of e g and only install the data sources and post processors relevant to them organisation wise we fix outstanding issues first after this for we create a new branch
1
7,284
10,434,736,019
IssuesEvent
2019-09-17 15:47:08
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
python error when using the QGIS "assign projection" tool in batch mode and output format is GPKG
Bug Processing
Author Name: **Giovanni Manghi** (@gioman) Original Redmine Issue: [21498](https://issues.qgis.org/issues/21498) Affected QGIS version: 3.4.5 Redmine category:processing/qgis --- Input layers have already a "fid" field. Does not happen not using batch mode. Possibly affects other tools? error is like Algorithm Assign projection starting… Input parameters: {'CRS': <qgis._core.QgsCoordinateReferenceSystem object at 0x0000000010CABCA8>, 'INPUT': 'input1', 'OUTPUT': <QgsProcessingOutputLayerDefinition {'sink':C:/Users/qgis/test1.gpkg, 'createOptions': {}}>} Could not create layer C:/Users/qgis/test1.gpkg: Creation of field fid failed (OGR error: Wrong field type for fid)
1.0
python error when using the QGIS "assign projection" tool in batch mode and output format is GPKG - Author Name: **Giovanni Manghi** (@gioman) Original Redmine Issue: [21498](https://issues.qgis.org/issues/21498) Affected QGIS version: 3.4.5 Redmine category:processing/qgis --- Input layers have already a "fid" field. Does not happen not using batch mode. Possibly affects other tools? error is like Algorithm Assign projection starting… Input parameters: {'CRS': <qgis._core.QgsCoordinateReferenceSystem object at 0x0000000010CABCA8>, 'INPUT': 'input1', 'OUTPUT': <QgsProcessingOutputLayerDefinition {'sink':C:/Users/qgis/test1.gpkg, 'createOptions': {}}>} Could not create layer C:/Users/qgis/test1.gpkg: Creation of field fid failed (OGR error: Wrong field type for fid)
process
python error when using the qgis assign projection tool in batch mode and output format is gpkg author name giovanni manghi gioman original redmine issue affected qgis version redmine category processing qgis input layers have already a fid field does not happen not using batch mode possibly affects other tools error is like algorithm assign projection starting… input parameters crs input output could not create layer c users qgis gpkg creation of field fid failed ogr error wrong field type for fid
1
2,129
4,969,774,063
IssuesEvent
2016-12-05 14:24:04
IMA-WorldHealth/bhima-2.X
https://api.github.com/repos/IMA-WorldHealth/bhima-2.X
closed
(chore) setting up real and common inventory list
lukonga test data processes
This issue is about to setting up a real list of common inventory list, in the current state of BHIMA we have only test data for inventory list; We must have common inventory list often meet in hospital sectors already defined in BHIMA
1.0
(chore) setting up real and common inventory list - This issue is about to setting up a real list of common inventory list, in the current state of BHIMA we have only test data for inventory list; We must have common inventory list often meet in hospital sectors already defined in BHIMA
process
chore setting up real and common inventory list this issue is about to setting up a real list of common inventory list in the current state of bhima we have only test data for inventory list we must have common inventory list often meet in hospital sectors already defined in bhima
1
10,214
4,021,481,799
IssuesEvent
2016-05-16 22:08:08
buckte/buckte
https://api.github.com/repos/buckte/buckte
closed
Fix "Insecure Dependency" issue in Gemfile.lock
codeclimate
Possible Input Validation Circumvention in Active Model https://codeclimate.com/github/buckte/buckte/Gemfile.lock#issue_573a287e108f0100019dab49
1.0
Fix "Insecure Dependency" issue in Gemfile.lock - Possible Input Validation Circumvention in Active Model https://codeclimate.com/github/buckte/buckte/Gemfile.lock#issue_573a287e108f0100019dab49
non_process
fix insecure dependency issue in gemfile lock possible input validation circumvention in active model
0
54,635
6,398,751,673
IssuesEvent
2017-08-04 21:32:00
PulpQE/pulp-smash
https://api.github.com/repos/PulpQE/pulp-smash
closed
Test os-tree copy repo
Issue Type: Test Case
https://pulp.plan.io/issues/2544 Create as OSTree Repo Sync the repo Create a 2nd ostree repo Copy repo1 to repo2 with filter Assert that copy is done with filter
1.0
Test os-tree copy repo - https://pulp.plan.io/issues/2544 Create as OSTree Repo Sync the repo Create a 2nd ostree repo Copy repo1 to repo2 with filter Assert that copy is done with filter
non_process
test os tree copy repo create as ostree repo sync the repo create a ostree repo copy to with filter assert that copy is done with filter
0
355,374
25,175,916,212
IssuesEvent
2022-11-11 09:14:54
jeromepui/pe
https://api.github.com/repos/jeromepui/pe
opened
[Developer Guide] Missing important terms in section 6.5 glossary
severity.Medium type.DocumentationBug
These terms are insufficient for the amount of complexity you have in the app. Terms such as professor roles: coordinator, lecturer, tabs, assignments. Some of the terms I listed may not be very important but there could be more important ones to provide for the user. ![Screenshot 2022-11-11 at 5.13.16 PM.png](https://raw.githubusercontent.com/jeromepui/pe/main/files/b4bf76ff-b885-46e6-b076-6f6b66af4fce.png) <!--session: 1668153057787-9e9283b8-9304-4dc3-8744-f7c54d9ba55b--> <!--Version: Web v3.4.4-->
1.0
[Developer Guide] Missing important terms in section 6.5 glossary - These terms are insufficient for the amount of complexity you have in the app. Terms such as professor roles: coordinator, lecturer, tabs, assignments. Some of the terms I listed may not be very important but there could be more important ones to provide for the user. ![Screenshot 2022-11-11 at 5.13.16 PM.png](https://raw.githubusercontent.com/jeromepui/pe/main/files/b4bf76ff-b885-46e6-b076-6f6b66af4fce.png) <!--session: 1668153057787-9e9283b8-9304-4dc3-8744-f7c54d9ba55b--> <!--Version: Web v3.4.4-->
non_process
missing important terms in section glossary these terms are insufficient for the amount of complexity you have in the app terms such as professor roles coordinator lecturer tabs assignments some of the terms i listed may not be very important but there could be more important ones to provide for the user
0
5,021
7,845,576,981
IssuesEvent
2018-06-19 13:20:44
openvstorage/framework-alba-plugin
https://api.github.com/repos/openvstorage/framework-alba-plugin
closed
Backend gui not loading due to namespace with wrong preset name
process_wontfix
# Problem With the manually fixes of Romain/Ops to reattach a volume back to the ABM cluster, it could be that a namespace is reattached with a wrong preset or a preset that doesn't exists. for example: ``` 018-06-04 14:37:48 39300 +0200 - NY2SRV0011 - 11813/140460485877584 - api/middleware - 386 - ERROR - An unhandled exception occurred: 'global-no-encrypt' Traceback (most recent call last): File "/usr/lib/python2.7/dist-packages/django/core/handlers/base.py", line 132, in get_response response = wrapped_callback(request, *callback_args, **callback_kwargs) File "/usr/lib/python2.7/dist-packages/rest_framework/viewsets.py", line 78, in view return self.dispatch(request, *args, **kwargs) File "/usr/lib/python2.7/dist-packages/django/views/decorators/csrf.py", line 58, in wrapped_view return view_func(*args, **kwargs) File "/usr/lib/python2.7/dist-packages/rest_framework/views.py", line 399, in dispatch response = self.handle_exception(exc) File "/usr/lib/python2.7/dist-packages/rest_framework/views.py", line 396, in dispatch response = handler(request, *args, **kwargs) File "/opt/OpenvStorage/webapps/api/backend/decorators.py", line 589, in new_function return_value = f(*args, **kwargs) File "/opt/OpenvStorage/webapps/api/backend/decorators.py", line 86, in new_function result = f(*args, **kw) File "/opt/OpenvStorage/webapps/api/backend/decorators.py", line 411, in new_function data = FullSerializer(object_type, contents=contents, instance=obj).data File "/usr/lib/python2.7/dist-packages/rest_framework/serializers.py", line 573, in data self._data = self.to_native(obj) File "/usr/lib/python2.7/dist-packages/rest_framework/serializers.py", line 349, in to_native value = field.field_to_native(obj, field_name) File "/usr/lib/python2.7/dist-packages/rest_framework/fields.py", line 198, in field_to_native value = get_component(value, component) File "/usr/lib/python2.7/dist-packages/rest_framework/fields.py", line 56, in get_component val = getattr(obj, attr_name) File "/opt/OpenvStorage/ovs/dal/dataobject.py", line 348, in <lambda> fget = lambda s: s._get_dynamic_property(dynamic) File "/opt/OpenvStorage/ovs/dal/dataobject.py", line 417, in _get_dynamic_property return self._backend_property(data_loader, dynamic) File "/opt/OpenvStorage/ovs/dal/dataobject.py", line 1030, in _backend_property dynamic_data = fct() File "/opt/OpenvStorage/ovs/dal/hybrids/albabackend.py", line 293, in _presets preset = preset_dict[namespace['namespace']['preset_name']] KeyError: 'global-no-encrypt' ``` Not sure if the gui should not load if a namespace have a wrong preset name. Maybe just throw an error on the gui instead of an empty page. ![image](https://user-images.githubusercontent.com/2900612/40919072-ac267f58-6808-11e8-909d-fb2058520861.png)
1.0
Backend gui not loading due to namespace with wrong preset name - # Problem With the manually fixes of Romain/Ops to reattach a volume back to the ABM cluster, it could be that a namespace is reattached with a wrong preset or a preset that doesn't exists. for example: ``` 018-06-04 14:37:48 39300 +0200 - NY2SRV0011 - 11813/140460485877584 - api/middleware - 386 - ERROR - An unhandled exception occurred: 'global-no-encrypt' Traceback (most recent call last): File "/usr/lib/python2.7/dist-packages/django/core/handlers/base.py", line 132, in get_response response = wrapped_callback(request, *callback_args, **callback_kwargs) File "/usr/lib/python2.7/dist-packages/rest_framework/viewsets.py", line 78, in view return self.dispatch(request, *args, **kwargs) File "/usr/lib/python2.7/dist-packages/django/views/decorators/csrf.py", line 58, in wrapped_view return view_func(*args, **kwargs) File "/usr/lib/python2.7/dist-packages/rest_framework/views.py", line 399, in dispatch response = self.handle_exception(exc) File "/usr/lib/python2.7/dist-packages/rest_framework/views.py", line 396, in dispatch response = handler(request, *args, **kwargs) File "/opt/OpenvStorage/webapps/api/backend/decorators.py", line 589, in new_function return_value = f(*args, **kwargs) File "/opt/OpenvStorage/webapps/api/backend/decorators.py", line 86, in new_function result = f(*args, **kw) File "/opt/OpenvStorage/webapps/api/backend/decorators.py", line 411, in new_function data = FullSerializer(object_type, contents=contents, instance=obj).data File "/usr/lib/python2.7/dist-packages/rest_framework/serializers.py", line 573, in data self._data = self.to_native(obj) File "/usr/lib/python2.7/dist-packages/rest_framework/serializers.py", line 349, in to_native value = field.field_to_native(obj, field_name) File "/usr/lib/python2.7/dist-packages/rest_framework/fields.py", line 198, in field_to_native value = get_component(value, component) File "/usr/lib/python2.7/dist-packages/rest_framework/fields.py", line 56, in get_component val = getattr(obj, attr_name) File "/opt/OpenvStorage/ovs/dal/dataobject.py", line 348, in <lambda> fget = lambda s: s._get_dynamic_property(dynamic) File "/opt/OpenvStorage/ovs/dal/dataobject.py", line 417, in _get_dynamic_property return self._backend_property(data_loader, dynamic) File "/opt/OpenvStorage/ovs/dal/dataobject.py", line 1030, in _backend_property dynamic_data = fct() File "/opt/OpenvStorage/ovs/dal/hybrids/albabackend.py", line 293, in _presets preset = preset_dict[namespace['namespace']['preset_name']] KeyError: 'global-no-encrypt' ``` Not sure if the gui should not load if a namespace have a wrong preset name. Maybe just throw an error on the gui instead of an empty page. ![image](https://user-images.githubusercontent.com/2900612/40919072-ac267f58-6808-11e8-909d-fb2058520861.png)
process
backend gui not loading due to namespace with wrong preset name problem with the manually fixes of romain ops to reattach a volume back to the abm cluster it could be that a namespace is reattached with a wrong preset or a preset that doesn t exists for example api middleware error an unhandled exception occurred global no encrypt traceback most recent call last file usr lib dist packages django core handlers base py line in get response response wrapped callback request callback args callback kwargs file usr lib dist packages rest framework viewsets py line in view return self dispatch request args kwargs file usr lib dist packages django views decorators csrf py line in wrapped view return view func args kwargs file usr lib dist packages rest framework views py line in dispatch response self handle exception exc file usr lib dist packages rest framework views py line in dispatch response handler request args kwargs file opt openvstorage webapps api backend decorators py line in new function return value f args kwargs file opt openvstorage webapps api backend decorators py line in new function result f args kw file opt openvstorage webapps api backend decorators py line in new function data fullserializer object type contents contents instance obj data file usr lib dist packages rest framework serializers py line in data self data self to native obj file usr lib dist packages rest framework serializers py line in to native value field field to native obj field name file usr lib dist packages rest framework fields py line in field to native value get component value component file usr lib dist packages rest framework fields py line in get component val getattr obj attr name file opt openvstorage ovs dal dataobject py line in fget lambda s s get dynamic property dynamic file opt openvstorage ovs dal dataobject py line in get dynamic property return self backend property data loader dynamic file opt openvstorage ovs dal dataobject py line in backend property dynamic data fct file opt openvstorage ovs dal hybrids albabackend py line in presets preset preset dict keyerror global no encrypt not sure if the gui should not load if a namespace have a wrong preset name maybe just throw an error on the gui instead of an empty page
1
21,020
27,967,915,653
IssuesEvent
2023-03-24 21:32:15
googleapis/gapic-generator-java
https://api.github.com/repos/googleapis/gapic-generator-java
opened
CI Workflow failures from javax.net.ssl.SSLHandshakeException: PKIX path validation failed
type: process priority: p2
Started noticing CI failures from https://github.com/googleapis/gapic-generator-java/pull/1557 presubmits that do not look related to the changes made. Failing workflow run: https://github.com/googleapis/gapic-generator-java/actions/runs/4515219935 Latest successful workflow run: https://github.com/googleapis/gapic-generator-java/actions/runs/4514393539 ``` [INFO] --- download-maven-plugin:1.6.8:wget (download-metadata-proto) @ gapic-generator-java --- Warning: Could not get content javax.net.ssl.SSLHandshakeException: PKIX path validation failed: java.security.cert.CertPathValidatorException: validity check failed at sun.security.ssl.Alert.createSSLException (Alert.java:131) at sun.security.ssl.TransportContext.fatal (TransportContext.java:353) ... Error: Failed to execute goal com.googlecode.maven-download-plugin:download-maven-plugin:1.6.8:wget (download-metadata-proto) on project gapic-generator-java: IO Error: Could not get content -> [Help 1] ```
1.0
CI Workflow failures from javax.net.ssl.SSLHandshakeException: PKIX path validation failed - Started noticing CI failures from https://github.com/googleapis/gapic-generator-java/pull/1557 presubmits that do not look related to the changes made. Failing workflow run: https://github.com/googleapis/gapic-generator-java/actions/runs/4515219935 Latest successful workflow run: https://github.com/googleapis/gapic-generator-java/actions/runs/4514393539 ``` [INFO] --- download-maven-plugin:1.6.8:wget (download-metadata-proto) @ gapic-generator-java --- Warning: Could not get content javax.net.ssl.SSLHandshakeException: PKIX path validation failed: java.security.cert.CertPathValidatorException: validity check failed at sun.security.ssl.Alert.createSSLException (Alert.java:131) at sun.security.ssl.TransportContext.fatal (TransportContext.java:353) ... Error: Failed to execute goal com.googlecode.maven-download-plugin:download-maven-plugin:1.6.8:wget (download-metadata-proto) on project gapic-generator-java: IO Error: Could not get content -> [Help 1] ```
process
ci workflow failures from javax net ssl sslhandshakeexception pkix path validation failed started noticing ci failures from presubmits that do not look related to the changes made failing workflow run latest successful workflow run download maven plugin wget download metadata proto gapic generator java warning could not get content javax net ssl sslhandshakeexception pkix path validation failed java security cert certpathvalidatorexception validity check failed at sun security ssl alert createsslexception alert java at sun security ssl transportcontext fatal transportcontext java error failed to execute goal com googlecode maven download plugin download maven plugin wget download metadata proto on project gapic generator java io error could not get content
1