Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
2,691
2,702,777,021
IssuesEvent
2015-04-06 12:23:02
twosigma/beaker-notebook
https://api.github.com/repos/twosigma/beaker-notebook
opened
Note tutorial - there's no way to interrupt the execution
bug Documentation
and when you change page you get an exception in the JS console
1.0
Note tutorial - there's no way to interrupt the execution - and when you change page you get an exception in the JS console
non_process
note tutorial there s no way to interrupt the execution and when you change page you get an exception in the js console
0
6,968
10,119,727,586
IssuesEvent
2019-07-31 12:13:32
linnovate/root
https://api.github.com/repos/linnovate/root
opened
plus button task from project (from discussion) opens task from discussion instead
Process bug Projects
go to meetings create a new meeting and name it go to projects click on the fitting discussion from the projects from discussion tab create a project from the discussion from the plus button create a new task instead of a task from project being opened, a task from discussion is created ![image](https://user-images.githubusercontent.com/38312178/62210778-a1616080-b3a5-11e9-8ba5-95a007e78257.png) ![image](https://user-images.githubusercontent.com/38312178/62210806-ae7e4f80-b3a5-11e9-970f-a7c091e86937.png)
1.0
plus button task from project (from discussion) opens task from discussion instead - go to meetings create a new meeting and name it go to projects click on the fitting discussion from the projects from discussion tab create a project from the discussion from the plus button create a new task instead of a task from project being opened, a task from discussion is created ![image](https://user-images.githubusercontent.com/38312178/62210778-a1616080-b3a5-11e9-8ba5-95a007e78257.png) ![image](https://user-images.githubusercontent.com/38312178/62210806-ae7e4f80-b3a5-11e9-970f-a7c091e86937.png)
process
plus button task from project from discussion opens task from discussion instead go to meetings create a new meeting and name it go to projects click on the fitting discussion from the projects from discussion tab create a project from the discussion from the plus button create a new task instead of a task from project being opened a task from discussion is created
1
67,367
14,862,163,875
IssuesEvent
2021-01-19 01:05:23
fufunoyu/example-maven-travis
https://api.github.com/repos/fufunoyu/example-maven-travis
opened
CVE-2020-36189 (High) detected in jackson-databind-2.9.10.7.jar
security vulnerability
## CVE-2020-36189 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.10.7.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: example-maven-travis/pom.xml</p> <p>Path to vulnerable library: canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.10.7/jackson-databind-2.9.10.7.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.9.10.7.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/fufunoyu/example-maven-travis/commit/de1c84bba50d30975d47b84e2be0ec5feb00419d">de1c84bba50d30975d47b84e2be0ec5feb00419d</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to com.newrelic.agent.deps.ch.qos.logback.core.db.DriverManagerConnectionSource. <p>Publish Date: 2021-01-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36189>CVE-2020-36189</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2996">https://github.com/FasterXML/jackson-databind/issues/2996</a></p> <p>Release Date: 2021-01-06</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p> </p> </details> <p></p>
True
CVE-2020-36189 (High) detected in jackson-databind-2.9.10.7.jar - ## CVE-2020-36189 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.10.7.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: example-maven-travis/pom.xml</p> <p>Path to vulnerable library: canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.10.7/jackson-databind-2.9.10.7.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.9.10.7.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/fufunoyu/example-maven-travis/commit/de1c84bba50d30975d47b84e2be0ec5feb00419d">de1c84bba50d30975d47b84e2be0ec5feb00419d</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to com.newrelic.agent.deps.ch.qos.logback.core.db.DriverManagerConnectionSource. <p>Publish Date: 2021-01-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36189>CVE-2020-36189</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2996">https://github.com/FasterXML/jackson-databind/issues/2996</a></p> <p>Release Date: 2021-01-06</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p> </p> </details> <p></p>
non_process
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file example maven travis pom xml path to vulnerable library canner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to com newrelic agent deps ch qos logback core db drivermanagerconnectionsource publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind
0
22,333
30,921,890,134
IssuesEvent
2023-08-06 02:00:09
lizhihao6/get-daily-arxiv-noti
https://api.github.com/repos/lizhihao6/get-daily-arxiv-noti
opened
New submissions for Fri, 4 Aug 23
event camera white balance isp compression image signal processing image signal process raw raw image events camera color contrast events AWB
## Keyword: events ### Reconstructing Three-Dimensional Models of Interacting Humans - **Authors:** Mihai Fieraru, Mihai Zanfir, Elisabeta Oneata, Alin-Ionut Popa, Vlad Olaru, Cristian Sminchisescu - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2308.01854 - **Pdf link:** https://arxiv.org/pdf/2308.01854 - **Abstract** Understanding 3d human interactions is fundamental for fine-grained scene analysis and behavioural modeling. However, most of the existing models predict incorrect, lifeless 3d estimates, that miss the subtle human contact aspects--the essence of the event--and are of little use for detailed behavioral understanding. This paper addresses such issues with several contributions: (1) we introduce models for interaction signature estimation (ISP) encompassing contact detection, segmentation, and 3d contact signature prediction; (2) we show how such components can be leveraged to ensure contact consistency during 3d reconstruction; (3) we construct several large datasets for learning and evaluating 3d contact prediction and reconstruction methods; specifically, we introduce CHI3D, a lab-based accurate 3d motion capture dataset with 631 sequences containing $2,525$ contact events, $728,664$ ground truth 3d poses, as well as FlickrCI3D, a dataset of $11,216$ images, with $14,081$ processed pairs of people, and $81,233$ facet-level surface correspondences. Finally, (4) we propose methodology for recovering the ground-truth pose and shape of interacting people in a controlled setup and (5) annotate all 3d interaction motions in CHI3D with textual descriptions. Motion data in multiple formats (GHUM and SMPLX parameters, Human3.6m 3d joints) is made available for research purposes at \url{https://ci3d.imar.ro}, together with an evaluation server and a public benchmark. ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB ### Multimodal Adaptation of CLIP for Few-Shot Action Recognition - **Authors:** Jiazheng Xing, Mengmeng Wang, Xiaojun Hou, Guang Dai, Jingdong Wang, Yong Liu - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2308.01532 - **Pdf link:** https://arxiv.org/pdf/2308.01532 - **Abstract** Applying large-scale pre-trained visual models like CLIP to few-shot action recognition tasks can benefit performance and efficiency. Utilizing the "pre-training, fine-tuning" paradigm makes it possible to avoid training a network from scratch, which can be time-consuming and resource-intensive. However, this method has two drawbacks. First, limited labeled samples for few-shot action recognition necessitate minimizing the number of tunable parameters to mitigate over-fitting, also leading to inadequate fine-tuning that increases resource consumption and may disrupt the generalized representation of models. Second, the video's extra-temporal dimension challenges few-shot recognition's effective temporal modeling, while pre-trained visual models are usually image models. This paper proposes a novel method called Multimodal Adaptation of CLIP (MA-CLIP) to address these issues. It adapts CLIP for few-shot action recognition by adding lightweight adapters, which can minimize the number of learnable parameters and enable the model to transfer across different tasks quickly. The adapters we design can combine information from video-text multimodal sources for task-oriented spatiotemporal modeling, which is fast, efficient, and has low training costs. Additionally, based on the attention mechanism, we design a text-guided prototype construction module that can fully utilize video-text information to enhance the representation of video prototypes. Our MA-CLIP is plug-and-play, which can be used in any different few-shot action recognition temporal alignment metric. ## Keyword: ISP ### Reconstructing Three-Dimensional Models of Interacting Humans - **Authors:** Mihai Fieraru, Mihai Zanfir, Elisabeta Oneata, Alin-Ionut Popa, Vlad Olaru, Cristian Sminchisescu - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2308.01854 - **Pdf link:** https://arxiv.org/pdf/2308.01854 - **Abstract** Understanding 3d human interactions is fundamental for fine-grained scene analysis and behavioural modeling. However, most of the existing models predict incorrect, lifeless 3d estimates, that miss the subtle human contact aspects--the essence of the event--and are of little use for detailed behavioral understanding. This paper addresses such issues with several contributions: (1) we introduce models for interaction signature estimation (ISP) encompassing contact detection, segmentation, and 3d contact signature prediction; (2) we show how such components can be leveraged to ensure contact consistency during 3d reconstruction; (3) we construct several large datasets for learning and evaluating 3d contact prediction and reconstruction methods; specifically, we introduce CHI3D, a lab-based accurate 3d motion capture dataset with 631 sequences containing $2,525$ contact events, $728,664$ ground truth 3d poses, as well as FlickrCI3D, a dataset of $11,216$ images, with $14,081$ processed pairs of people, and $81,233$ facet-level surface correspondences. Finally, (4) we propose methodology for recovering the ground-truth pose and shape of interacting people in a controlled setup and (5) annotate all 3d interaction motions in CHI3D with textual descriptions. Motion data in multiple formats (GHUM and SMPLX parameters, Human3.6m 3d joints) is made available for research purposes at \url{https://ci3d.imar.ro}, together with an evaluation server and a public benchmark. ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression ### TDMD: A Database for Dynamic Color Mesh Subjective and Objective Quality Explorations - **Authors:** Qi Yang, Joel Jung, Timon Deschamps, Xiaozhong Xu, Shan Liu - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2308.01499 - **Pdf link:** https://arxiv.org/pdf/2308.01499 - **Abstract** Dynamic colored meshes (DCM) are widely used in various applications; however, these meshes may undergo different processes, such as compression or transmission, which can distort them and degrade their quality. To facilitate the development of objective metrics for DCMs and study the influence of typical distortions on their perception, we create the Tencent - dynamic colored mesh database (TDMD) containing eight reference DCM objects with six typical distortions. Using processed video sequences (PVS) derived from the DCM, we have conducted a large-scale subjective experiment that resulted in 303 distorted DCM samples with mean opinion scores, making the TDMD the largest available DCM database to our knowledge. This database enabled us to study the impact of different types of distortion on human perception and offer recommendations for DCM compression and related tasks. Additionally, we have evaluated three types of state-of-the-art objective metrics on the TDMD, including image-based, point-based, and video-based metrics, on the TDMD. Our experimental results highlight the strengths and weaknesses of each metric, and we provide suggestions about the selection of metrics in practical DCM applications. The TDMD will be made publicly available at the following location: https://multimedia.tencent.com/resources/tdmd. ### MVFlow: Deep Optical Flow Estimation of Compressed Videos with Motion Vector Prior - **Authors:** Shili Zhou, Xuhao Jiang, Weimin Tan, Ruian He, Bo Yan - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2308.01568 - **Pdf link:** https://arxiv.org/pdf/2308.01568 - **Abstract** In recent years, many deep learning-based methods have been proposed to tackle the problem of optical flow estimation and achieved promising results. However, they hardly consider that most videos are compressed and thus ignore the pre-computed information in compressed video streams. Motion vectors, one of the compression information, record the motion of the video frames. They can be directly extracted from the compression code stream without computational cost and serve as a solid prior for optical flow estimation. Therefore, we propose an optical flow model, MVFlow, which uses motion vectors to improve the speed and accuracy of optical flow estimation for compressed videos. In detail, MVFlow includes a key Motion-Vector Converting Module, which ensures that the motion vectors can be transformed into the same domain of optical flow and then be utilized fully by the flow estimation module. Meanwhile, we construct four optical flow datasets for compressed videos containing frames and motion vectors in pairs. The experimental results demonstrate the superiority of our proposed MVFlow, which can reduce the AEPE by 1.09 compared to existing models or save 52% time to achieve similar accuracy to existing models. ### A Novel Tensor Decomposition of arbitrary order based on Block Convolution with Reflective Boundary Conditions for Multi-Dimensional Data Analysis - **Authors:** Mahdi Molavi, Mansoor Rezghi, Tayyebeh Saeedi - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2308.01768 - **Pdf link:** https://arxiv.org/pdf/2308.01768 - **Abstract** Tensor decompositions are powerful tools for analyzing multi-dimensional data in their original format. Besides tensor decompositions like Tucker and CP, Tensor SVD (t-SVD) which is based on the t-product of tensors is another extension of SVD to tensors that recently developed and has found numerous applications in analyzing high dimensional data. This paper offers a new insight into the t-Product and shows that this product is a block convolution of two tensors with periodic boundary conditions. Based on this viewpoint, we propose a new tensor-tensor product called the $\star_c{}\text{-Product}$ based on Block convolution with reflective boundary conditions. Using a tensor framework, this product can be easily extended to tensors of arbitrary order. Additionally, we introduce a tensor decomposition based on our $\star_c{}\text{-Product}$ for arbitrary order tensors. Compared to t-SVD, our new decomposition has lower complexity, and experiments show that it yields higher-quality results in applications such as classification and compression. ## Keyword: RAW ### PPI-NET: End-to-End Parametric Primitive Inference - **Authors:** Liang Wang, Xiaogang Wang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2308.01521 - **Pdf link:** https://arxiv.org/pdf/2308.01521 - **Abstract** In engineering applications, line, circle, arc, and point are collectively referred to as primitives, and they play a crucial role in path planning, simulation analysis, and manufacturing. When designing CAD models, engineers typically start by sketching the model's orthographic view on paper or a whiteboard and then translate the design intent into a CAD program. Although this design method is powerful, it often involves challenging and repetitive tasks, requiring engineers to perform numerous similar operations in each design. To address this conversion process, we propose an efficient and accurate end-to-end method that avoids the inefficiency and error accumulation issues associated with using auto-regressive models to infer parametric primitives from hand-drawn sketch images. Since our model samples match the representation format of standard CAD software, they can be imported into CAD software for solving, editing, and applied to downstream design tasks. ### Data Augmentation for Human Behavior Analysis in Multi-Person Conversations - **Authors:** Kun Li, Dan Guo, Guoliang Chen, Feiyang Liu, Meng Wang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2308.01526 - **Pdf link:** https://arxiv.org/pdf/2308.01526 - **Abstract** In this paper, we present the solution of our team HFUT-VUT for the MultiMediate Grand Challenge 2023 at ACM Multimedia 2023. The solution covers three sub-challenges: bodily behavior recognition, eye contact detection, and next speaker prediction. We select Swin Transformer as the baseline and exploit data augmentation strategies to address the above three tasks. Specifically, we crop the raw video to remove the noise from other parts. At the same time, we utilize data augmentation to improve the generalization of the model. As a result, our solution achieves the best results of 0.6262 for bodily behavior recognition in terms of mean average precision and the accuracy of 0.7771 for eye contact detection on the corresponding test set. In addition, our approach also achieves comparable results of 0.5281 for the next speaker prediction in terms of unweighted average recall. ### Multimodal Adaptation of CLIP for Few-Shot Action Recognition - **Authors:** Jiazheng Xing, Mengmeng Wang, Xiaojun Hou, Guang Dai, Jingdong Wang, Yong Liu - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2308.01532 - **Pdf link:** https://arxiv.org/pdf/2308.01532 - **Abstract** Applying large-scale pre-trained visual models like CLIP to few-shot action recognition tasks can benefit performance and efficiency. Utilizing the "pre-training, fine-tuning" paradigm makes it possible to avoid training a network from scratch, which can be time-consuming and resource-intensive. However, this method has two drawbacks. First, limited labeled samples for few-shot action recognition necessitate minimizing the number of tunable parameters to mitigate over-fitting, also leading to inadequate fine-tuning that increases resource consumption and may disrupt the generalized representation of models. Second, the video's extra-temporal dimension challenges few-shot recognition's effective temporal modeling, while pre-trained visual models are usually image models. This paper proposes a novel method called Multimodal Adaptation of CLIP (MA-CLIP) to address these issues. It adapts CLIP for few-shot action recognition by adding lightweight adapters, which can minimize the number of learnable parameters and enable the model to transfer across different tasks quickly. The adapters we design can combine information from video-text multimodal sources for task-oriented spatiotemporal modeling, which is fast, efficient, and has low training costs. Additionally, based on the attention mechanism, we design a text-guided prototype construction module that can fully utilize video-text information to enhance the representation of video prototypes. Our MA-CLIP is plug-and-play, which can be used in any different few-shot action recognition temporal alignment metric. ### A Novel Convolutional Neural Network Architecture with a Continuous Symmetry - **Authors:** Yao Liu, Hang Shao, Bing Bai - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG); Neural and Evolutionary Computing (cs.NE) - **Arxiv link:** https://arxiv.org/abs/2308.01621 - **Pdf link:** https://arxiv.org/pdf/2308.01621 - **Abstract** This paper introduces a new Convolutional Neural Network (ConvNet) architecture inspired by a class of partial differential equations (PDEs) called quasi-linear hyperbolic systems. With comparable performance on image classification task, it allows for the modification of the weights via a continuous group of symmetry. This is a significant shift from traditional models where the architecture and weights are essentially fixed. We wish to promote the (internal) symmetry as a new desirable property for a neural network, and to draw attention to the PDE perspective in analyzing and interpreting ConvNets in the broader Deep Learning community. ### Weakly Supervised 3D Instance Segmentation without Instance-level Annotations - **Authors:** Shichao Dong, Guosheng Lin - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2308.01721 - **Pdf link:** https://arxiv.org/pdf/2308.01721 - **Abstract** 3D semantic scene understanding tasks have achieved great success with the emergence of deep learning, but often require a huge amount of manually annotated training data. To alleviate the annotation cost, we propose the first weakly-supervised 3D instance segmentation method that only requires categorical semantic labels as supervision, and we do not need instance-level labels. The required semantic annotations can be either dense or extreme sparse (e.g. 0.02% of total points). Even without having any instance-related ground-truth, we design an approach to break point clouds into raw fragments and find the most confident samples for learning instance centroids. Furthermore, we construct a recomposed dataset using pseudo instances, which is used to learn our defined multilevel shape-aware objectness signal. An asymmetrical object inference algorithm is followed to process core points and boundary points with different strategies, and generate high-quality pseudo instance labels to guide iterative training. Experiments demonstrate that our method can achieve comparable results with recent fully supervised methods. By generating pseudo instance labels from categorical semantic labels, our designed approach can also assist existing methods for learning 3D instance segmentation at reduced annotation cost. ## Keyword: raw image There is no result
2.0
New submissions for Fri, 4 Aug 23 - ## Keyword: events ### Reconstructing Three-Dimensional Models of Interacting Humans - **Authors:** Mihai Fieraru, Mihai Zanfir, Elisabeta Oneata, Alin-Ionut Popa, Vlad Olaru, Cristian Sminchisescu - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2308.01854 - **Pdf link:** https://arxiv.org/pdf/2308.01854 - **Abstract** Understanding 3d human interactions is fundamental for fine-grained scene analysis and behavioural modeling. However, most of the existing models predict incorrect, lifeless 3d estimates, that miss the subtle human contact aspects--the essence of the event--and are of little use for detailed behavioral understanding. This paper addresses such issues with several contributions: (1) we introduce models for interaction signature estimation (ISP) encompassing contact detection, segmentation, and 3d contact signature prediction; (2) we show how such components can be leveraged to ensure contact consistency during 3d reconstruction; (3) we construct several large datasets for learning and evaluating 3d contact prediction and reconstruction methods; specifically, we introduce CHI3D, a lab-based accurate 3d motion capture dataset with 631 sequences containing $2,525$ contact events, $728,664$ ground truth 3d poses, as well as FlickrCI3D, a dataset of $11,216$ images, with $14,081$ processed pairs of people, and $81,233$ facet-level surface correspondences. Finally, (4) we propose methodology for recovering the ground-truth pose and shape of interacting people in a controlled setup and (5) annotate all 3d interaction motions in CHI3D with textual descriptions. Motion data in multiple formats (GHUM and SMPLX parameters, Human3.6m 3d joints) is made available for research purposes at \url{https://ci3d.imar.ro}, together with an evaluation server and a public benchmark. ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB ### Multimodal Adaptation of CLIP for Few-Shot Action Recognition - **Authors:** Jiazheng Xing, Mengmeng Wang, Xiaojun Hou, Guang Dai, Jingdong Wang, Yong Liu - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2308.01532 - **Pdf link:** https://arxiv.org/pdf/2308.01532 - **Abstract** Applying large-scale pre-trained visual models like CLIP to few-shot action recognition tasks can benefit performance and efficiency. Utilizing the "pre-training, fine-tuning" paradigm makes it possible to avoid training a network from scratch, which can be time-consuming and resource-intensive. However, this method has two drawbacks. First, limited labeled samples for few-shot action recognition necessitate minimizing the number of tunable parameters to mitigate over-fitting, also leading to inadequate fine-tuning that increases resource consumption and may disrupt the generalized representation of models. Second, the video's extra-temporal dimension challenges few-shot recognition's effective temporal modeling, while pre-trained visual models are usually image models. This paper proposes a novel method called Multimodal Adaptation of CLIP (MA-CLIP) to address these issues. It adapts CLIP for few-shot action recognition by adding lightweight adapters, which can minimize the number of learnable parameters and enable the model to transfer across different tasks quickly. The adapters we design can combine information from video-text multimodal sources for task-oriented spatiotemporal modeling, which is fast, efficient, and has low training costs. Additionally, based on the attention mechanism, we design a text-guided prototype construction module that can fully utilize video-text information to enhance the representation of video prototypes. Our MA-CLIP is plug-and-play, which can be used in any different few-shot action recognition temporal alignment metric. ## Keyword: ISP ### Reconstructing Three-Dimensional Models of Interacting Humans - **Authors:** Mihai Fieraru, Mihai Zanfir, Elisabeta Oneata, Alin-Ionut Popa, Vlad Olaru, Cristian Sminchisescu - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2308.01854 - **Pdf link:** https://arxiv.org/pdf/2308.01854 - **Abstract** Understanding 3d human interactions is fundamental for fine-grained scene analysis and behavioural modeling. However, most of the existing models predict incorrect, lifeless 3d estimates, that miss the subtle human contact aspects--the essence of the event--and are of little use for detailed behavioral understanding. This paper addresses such issues with several contributions: (1) we introduce models for interaction signature estimation (ISP) encompassing contact detection, segmentation, and 3d contact signature prediction; (2) we show how such components can be leveraged to ensure contact consistency during 3d reconstruction; (3) we construct several large datasets for learning and evaluating 3d contact prediction and reconstruction methods; specifically, we introduce CHI3D, a lab-based accurate 3d motion capture dataset with 631 sequences containing $2,525$ contact events, $728,664$ ground truth 3d poses, as well as FlickrCI3D, a dataset of $11,216$ images, with $14,081$ processed pairs of people, and $81,233$ facet-level surface correspondences. Finally, (4) we propose methodology for recovering the ground-truth pose and shape of interacting people in a controlled setup and (5) annotate all 3d interaction motions in CHI3D with textual descriptions. Motion data in multiple formats (GHUM and SMPLX parameters, Human3.6m 3d joints) is made available for research purposes at \url{https://ci3d.imar.ro}, together with an evaluation server and a public benchmark. ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression ### TDMD: A Database for Dynamic Color Mesh Subjective and Objective Quality Explorations - **Authors:** Qi Yang, Joel Jung, Timon Deschamps, Xiaozhong Xu, Shan Liu - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2308.01499 - **Pdf link:** https://arxiv.org/pdf/2308.01499 - **Abstract** Dynamic colored meshes (DCM) are widely used in various applications; however, these meshes may undergo different processes, such as compression or transmission, which can distort them and degrade their quality. To facilitate the development of objective metrics for DCMs and study the influence of typical distortions on their perception, we create the Tencent - dynamic colored mesh database (TDMD) containing eight reference DCM objects with six typical distortions. Using processed video sequences (PVS) derived from the DCM, we have conducted a large-scale subjective experiment that resulted in 303 distorted DCM samples with mean opinion scores, making the TDMD the largest available DCM database to our knowledge. This database enabled us to study the impact of different types of distortion on human perception and offer recommendations for DCM compression and related tasks. Additionally, we have evaluated three types of state-of-the-art objective metrics on the TDMD, including image-based, point-based, and video-based metrics, on the TDMD. Our experimental results highlight the strengths and weaknesses of each metric, and we provide suggestions about the selection of metrics in practical DCM applications. The TDMD will be made publicly available at the following location: https://multimedia.tencent.com/resources/tdmd. ### MVFlow: Deep Optical Flow Estimation of Compressed Videos with Motion Vector Prior - **Authors:** Shili Zhou, Xuhao Jiang, Weimin Tan, Ruian He, Bo Yan - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2308.01568 - **Pdf link:** https://arxiv.org/pdf/2308.01568 - **Abstract** In recent years, many deep learning-based methods have been proposed to tackle the problem of optical flow estimation and achieved promising results. However, they hardly consider that most videos are compressed and thus ignore the pre-computed information in compressed video streams. Motion vectors, one of the compression information, record the motion of the video frames. They can be directly extracted from the compression code stream without computational cost and serve as a solid prior for optical flow estimation. Therefore, we propose an optical flow model, MVFlow, which uses motion vectors to improve the speed and accuracy of optical flow estimation for compressed videos. In detail, MVFlow includes a key Motion-Vector Converting Module, which ensures that the motion vectors can be transformed into the same domain of optical flow and then be utilized fully by the flow estimation module. Meanwhile, we construct four optical flow datasets for compressed videos containing frames and motion vectors in pairs. The experimental results demonstrate the superiority of our proposed MVFlow, which can reduce the AEPE by 1.09 compared to existing models or save 52% time to achieve similar accuracy to existing models. ### A Novel Tensor Decomposition of arbitrary order based on Block Convolution with Reflective Boundary Conditions for Multi-Dimensional Data Analysis - **Authors:** Mahdi Molavi, Mansoor Rezghi, Tayyebeh Saeedi - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2308.01768 - **Pdf link:** https://arxiv.org/pdf/2308.01768 - **Abstract** Tensor decompositions are powerful tools for analyzing multi-dimensional data in their original format. Besides tensor decompositions like Tucker and CP, Tensor SVD (t-SVD) which is based on the t-product of tensors is another extension of SVD to tensors that recently developed and has found numerous applications in analyzing high dimensional data. This paper offers a new insight into the t-Product and shows that this product is a block convolution of two tensors with periodic boundary conditions. Based on this viewpoint, we propose a new tensor-tensor product called the $\star_c{}\text{-Product}$ based on Block convolution with reflective boundary conditions. Using a tensor framework, this product can be easily extended to tensors of arbitrary order. Additionally, we introduce a tensor decomposition based on our $\star_c{}\text{-Product}$ for arbitrary order tensors. Compared to t-SVD, our new decomposition has lower complexity, and experiments show that it yields higher-quality results in applications such as classification and compression. ## Keyword: RAW ### PPI-NET: End-to-End Parametric Primitive Inference - **Authors:** Liang Wang, Xiaogang Wang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2308.01521 - **Pdf link:** https://arxiv.org/pdf/2308.01521 - **Abstract** In engineering applications, line, circle, arc, and point are collectively referred to as primitives, and they play a crucial role in path planning, simulation analysis, and manufacturing. When designing CAD models, engineers typically start by sketching the model's orthographic view on paper or a whiteboard and then translate the design intent into a CAD program. Although this design method is powerful, it often involves challenging and repetitive tasks, requiring engineers to perform numerous similar operations in each design. To address this conversion process, we propose an efficient and accurate end-to-end method that avoids the inefficiency and error accumulation issues associated with using auto-regressive models to infer parametric primitives from hand-drawn sketch images. Since our model samples match the representation format of standard CAD software, they can be imported into CAD software for solving, editing, and applied to downstream design tasks. ### Data Augmentation for Human Behavior Analysis in Multi-Person Conversations - **Authors:** Kun Li, Dan Guo, Guoliang Chen, Feiyang Liu, Meng Wang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2308.01526 - **Pdf link:** https://arxiv.org/pdf/2308.01526 - **Abstract** In this paper, we present the solution of our team HFUT-VUT for the MultiMediate Grand Challenge 2023 at ACM Multimedia 2023. The solution covers three sub-challenges: bodily behavior recognition, eye contact detection, and next speaker prediction. We select Swin Transformer as the baseline and exploit data augmentation strategies to address the above three tasks. Specifically, we crop the raw video to remove the noise from other parts. At the same time, we utilize data augmentation to improve the generalization of the model. As a result, our solution achieves the best results of 0.6262 for bodily behavior recognition in terms of mean average precision and the accuracy of 0.7771 for eye contact detection on the corresponding test set. In addition, our approach also achieves comparable results of 0.5281 for the next speaker prediction in terms of unweighted average recall. ### Multimodal Adaptation of CLIP for Few-Shot Action Recognition - **Authors:** Jiazheng Xing, Mengmeng Wang, Xiaojun Hou, Guang Dai, Jingdong Wang, Yong Liu - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2308.01532 - **Pdf link:** https://arxiv.org/pdf/2308.01532 - **Abstract** Applying large-scale pre-trained visual models like CLIP to few-shot action recognition tasks can benefit performance and efficiency. Utilizing the "pre-training, fine-tuning" paradigm makes it possible to avoid training a network from scratch, which can be time-consuming and resource-intensive. However, this method has two drawbacks. First, limited labeled samples for few-shot action recognition necessitate minimizing the number of tunable parameters to mitigate over-fitting, also leading to inadequate fine-tuning that increases resource consumption and may disrupt the generalized representation of models. Second, the video's extra-temporal dimension challenges few-shot recognition's effective temporal modeling, while pre-trained visual models are usually image models. This paper proposes a novel method called Multimodal Adaptation of CLIP (MA-CLIP) to address these issues. It adapts CLIP for few-shot action recognition by adding lightweight adapters, which can minimize the number of learnable parameters and enable the model to transfer across different tasks quickly. The adapters we design can combine information from video-text multimodal sources for task-oriented spatiotemporal modeling, which is fast, efficient, and has low training costs. Additionally, based on the attention mechanism, we design a text-guided prototype construction module that can fully utilize video-text information to enhance the representation of video prototypes. Our MA-CLIP is plug-and-play, which can be used in any different few-shot action recognition temporal alignment metric. ### A Novel Convolutional Neural Network Architecture with a Continuous Symmetry - **Authors:** Yao Liu, Hang Shao, Bing Bai - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG); Neural and Evolutionary Computing (cs.NE) - **Arxiv link:** https://arxiv.org/abs/2308.01621 - **Pdf link:** https://arxiv.org/pdf/2308.01621 - **Abstract** This paper introduces a new Convolutional Neural Network (ConvNet) architecture inspired by a class of partial differential equations (PDEs) called quasi-linear hyperbolic systems. With comparable performance on image classification task, it allows for the modification of the weights via a continuous group of symmetry. This is a significant shift from traditional models where the architecture and weights are essentially fixed. We wish to promote the (internal) symmetry as a new desirable property for a neural network, and to draw attention to the PDE perspective in analyzing and interpreting ConvNets in the broader Deep Learning community. ### Weakly Supervised 3D Instance Segmentation without Instance-level Annotations - **Authors:** Shichao Dong, Guosheng Lin - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2308.01721 - **Pdf link:** https://arxiv.org/pdf/2308.01721 - **Abstract** 3D semantic scene understanding tasks have achieved great success with the emergence of deep learning, but often require a huge amount of manually annotated training data. To alleviate the annotation cost, we propose the first weakly-supervised 3D instance segmentation method that only requires categorical semantic labels as supervision, and we do not need instance-level labels. The required semantic annotations can be either dense or extreme sparse (e.g. 0.02% of total points). Even without having any instance-related ground-truth, we design an approach to break point clouds into raw fragments and find the most confident samples for learning instance centroids. Furthermore, we construct a recomposed dataset using pseudo instances, which is used to learn our defined multilevel shape-aware objectness signal. An asymmetrical object inference algorithm is followed to process core points and boundary points with different strategies, and generate high-quality pseudo instance labels to guide iterative training. Experiments demonstrate that our method can achieve comparable results with recent fully supervised methods. By generating pseudo instance labels from categorical semantic labels, our designed approach can also assist existing methods for learning 3D instance segmentation at reduced annotation cost. ## Keyword: raw image There is no result
process
new submissions for fri aug keyword events reconstructing three dimensional models of interacting humans authors mihai fieraru mihai zanfir elisabeta oneata alin ionut popa vlad olaru cristian sminchisescu subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract understanding human interactions is fundamental for fine grained scene analysis and behavioural modeling however most of the existing models predict incorrect lifeless estimates that miss the subtle human contact aspects the essence of the event and are of little use for detailed behavioral understanding this paper addresses such issues with several contributions we introduce models for interaction signature estimation isp encompassing contact detection segmentation and contact signature prediction we show how such components can be leveraged to ensure contact consistency during reconstruction we construct several large datasets for learning and evaluating contact prediction and reconstruction methods specifically we introduce a lab based accurate motion capture dataset with sequences containing contact events ground truth poses as well as a dataset of images with processed pairs of people and facet level surface correspondences finally we propose methodology for recovering the ground truth pose and shape of interacting people in a controlled setup and annotate all interaction motions in with textual descriptions motion data in multiple formats ghum and smplx parameters joints is made available for research purposes at url together with an evaluation server and a public benchmark keyword event camera there is no result keyword events camera there is no result keyword white balance there is no result keyword color contrast there is no result keyword awb multimodal adaptation of clip for few shot action recognition authors jiazheng xing mengmeng wang xiaojun hou guang dai jingdong wang yong liu subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract applying large scale pre trained visual models like clip to few shot action recognition tasks can benefit performance and efficiency utilizing the pre training fine tuning paradigm makes it possible to avoid training a network from scratch which can be time consuming and resource intensive however this method has two drawbacks first limited labeled samples for few shot action recognition necessitate minimizing the number of tunable parameters to mitigate over fitting also leading to inadequate fine tuning that increases resource consumption and may disrupt the generalized representation of models second the video s extra temporal dimension challenges few shot recognition s effective temporal modeling while pre trained visual models are usually image models this paper proposes a novel method called multimodal adaptation of clip ma clip to address these issues it adapts clip for few shot action recognition by adding lightweight adapters which can minimize the number of learnable parameters and enable the model to transfer across different tasks quickly the adapters we design can combine information from video text multimodal sources for task oriented spatiotemporal modeling which is fast efficient and has low training costs additionally based on the attention mechanism we design a text guided prototype construction module that can fully utilize video text information to enhance the representation of video prototypes our ma clip is plug and play which can be used in any different few shot action recognition temporal alignment metric keyword isp reconstructing three dimensional models of interacting humans authors mihai fieraru mihai zanfir elisabeta oneata alin ionut popa vlad olaru cristian sminchisescu subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract understanding human interactions is fundamental for fine grained scene analysis and behavioural modeling however most of the existing models predict incorrect lifeless estimates that miss the subtle human contact aspects the essence of the event and are of little use for detailed behavioral understanding this paper addresses such issues with several contributions we introduce models for interaction signature estimation isp encompassing contact detection segmentation and contact signature prediction we show how such components can be leveraged to ensure contact consistency during reconstruction we construct several large datasets for learning and evaluating contact prediction and reconstruction methods specifically we introduce a lab based accurate motion capture dataset with sequences containing contact events ground truth poses as well as a dataset of images with processed pairs of people and facet level surface correspondences finally we propose methodology for recovering the ground truth pose and shape of interacting people in a controlled setup and annotate all interaction motions in with textual descriptions motion data in multiple formats ghum and smplx parameters joints is made available for research purposes at url together with an evaluation server and a public benchmark keyword image signal processing there is no result keyword image signal process there is no result keyword compression tdmd a database for dynamic color mesh subjective and objective quality explorations authors qi yang joel jung timon deschamps xiaozhong xu shan liu subjects computer vision and pattern recognition cs cv image and video processing eess iv arxiv link pdf link abstract dynamic colored meshes dcm are widely used in various applications however these meshes may undergo different processes such as compression or transmission which can distort them and degrade their quality to facilitate the development of objective metrics for dcms and study the influence of typical distortions on their perception we create the tencent dynamic colored mesh database tdmd containing eight reference dcm objects with six typical distortions using processed video sequences pvs derived from the dcm we have conducted a large scale subjective experiment that resulted in distorted dcm samples with mean opinion scores making the tdmd the largest available dcm database to our knowledge this database enabled us to study the impact of different types of distortion on human perception and offer recommendations for dcm compression and related tasks additionally we have evaluated three types of state of the art objective metrics on the tdmd including image based point based and video based metrics on the tdmd our experimental results highlight the strengths and weaknesses of each metric and we provide suggestions about the selection of metrics in practical dcm applications the tdmd will be made publicly available at the following location mvflow deep optical flow estimation of compressed videos with motion vector prior authors shili zhou xuhao jiang weimin tan ruian he bo yan subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract in recent years many deep learning based methods have been proposed to tackle the problem of optical flow estimation and achieved promising results however they hardly consider that most videos are compressed and thus ignore the pre computed information in compressed video streams motion vectors one of the compression information record the motion of the video frames they can be directly extracted from the compression code stream without computational cost and serve as a solid prior for optical flow estimation therefore we propose an optical flow model mvflow which uses motion vectors to improve the speed and accuracy of optical flow estimation for compressed videos in detail mvflow includes a key motion vector converting module which ensures that the motion vectors can be transformed into the same domain of optical flow and then be utilized fully by the flow estimation module meanwhile we construct four optical flow datasets for compressed videos containing frames and motion vectors in pairs the experimental results demonstrate the superiority of our proposed mvflow which can reduce the aepe by compared to existing models or save time to achieve similar accuracy to existing models a novel tensor decomposition of arbitrary order based on block convolution with reflective boundary conditions for multi dimensional data analysis authors mahdi molavi mansoor rezghi tayyebeh saeedi subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract tensor decompositions are powerful tools for analyzing multi dimensional data in their original format besides tensor decompositions like tucker and cp tensor svd t svd which is based on the t product of tensors is another extension of svd to tensors that recently developed and has found numerous applications in analyzing high dimensional data this paper offers a new insight into the t product and shows that this product is a block convolution of two tensors with periodic boundary conditions based on this viewpoint we propose a new tensor tensor product called the star c text product based on block convolution with reflective boundary conditions using a tensor framework this product can be easily extended to tensors of arbitrary order additionally we introduce a tensor decomposition based on our star c text product for arbitrary order tensors compared to t svd our new decomposition has lower complexity and experiments show that it yields higher quality results in applications such as classification and compression keyword raw ppi net end to end parametric primitive inference authors liang wang xiaogang wang subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract in engineering applications line circle arc and point are collectively referred to as primitives and they play a crucial role in path planning simulation analysis and manufacturing when designing cad models engineers typically start by sketching the model s orthographic view on paper or a whiteboard and then translate the design intent into a cad program although this design method is powerful it often involves challenging and repetitive tasks requiring engineers to perform numerous similar operations in each design to address this conversion process we propose an efficient and accurate end to end method that avoids the inefficiency and error accumulation issues associated with using auto regressive models to infer parametric primitives from hand drawn sketch images since our model samples match the representation format of standard cad software they can be imported into cad software for solving editing and applied to downstream design tasks data augmentation for human behavior analysis in multi person conversations authors kun li dan guo guoliang chen feiyang liu meng wang subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract in this paper we present the solution of our team hfut vut for the multimediate grand challenge at acm multimedia the solution covers three sub challenges bodily behavior recognition eye contact detection and next speaker prediction we select swin transformer as the baseline and exploit data augmentation strategies to address the above three tasks specifically we crop the raw video to remove the noise from other parts at the same time we utilize data augmentation to improve the generalization of the model as a result our solution achieves the best results of for bodily behavior recognition in terms of mean average precision and the accuracy of for eye contact detection on the corresponding test set in addition our approach also achieves comparable results of for the next speaker prediction in terms of unweighted average recall multimodal adaptation of clip for few shot action recognition authors jiazheng xing mengmeng wang xiaojun hou guang dai jingdong wang yong liu subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract applying large scale pre trained visual models like clip to few shot action recognition tasks can benefit performance and efficiency utilizing the pre training fine tuning paradigm makes it possible to avoid training a network from scratch which can be time consuming and resource intensive however this method has two drawbacks first limited labeled samples for few shot action recognition necessitate minimizing the number of tunable parameters to mitigate over fitting also leading to inadequate fine tuning that increases resource consumption and may disrupt the generalized representation of models second the video s extra temporal dimension challenges few shot recognition s effective temporal modeling while pre trained visual models are usually image models this paper proposes a novel method called multimodal adaptation of clip ma clip to address these issues it adapts clip for few shot action recognition by adding lightweight adapters which can minimize the number of learnable parameters and enable the model to transfer across different tasks quickly the adapters we design can combine information from video text multimodal sources for task oriented spatiotemporal modeling which is fast efficient and has low training costs additionally based on the attention mechanism we design a text guided prototype construction module that can fully utilize video text information to enhance the representation of video prototypes our ma clip is plug and play which can be used in any different few shot action recognition temporal alignment metric a novel convolutional neural network architecture with a continuous symmetry authors yao liu hang shao bing bai subjects computer vision and pattern recognition cs cv machine learning cs lg neural and evolutionary computing cs ne arxiv link pdf link abstract this paper introduces a new convolutional neural network convnet architecture inspired by a class of partial differential equations pdes called quasi linear hyperbolic systems with comparable performance on image classification task it allows for the modification of the weights via a continuous group of symmetry this is a significant shift from traditional models where the architecture and weights are essentially fixed we wish to promote the internal symmetry as a new desirable property for a neural network and to draw attention to the pde perspective in analyzing and interpreting convnets in the broader deep learning community weakly supervised instance segmentation without instance level annotations authors shichao dong guosheng lin subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract semantic scene understanding tasks have achieved great success with the emergence of deep learning but often require a huge amount of manually annotated training data to alleviate the annotation cost we propose the first weakly supervised instance segmentation method that only requires categorical semantic labels as supervision and we do not need instance level labels the required semantic annotations can be either dense or extreme sparse e g of total points even without having any instance related ground truth we design an approach to break point clouds into raw fragments and find the most confident samples for learning instance centroids furthermore we construct a recomposed dataset using pseudo instances which is used to learn our defined multilevel shape aware objectness signal an asymmetrical object inference algorithm is followed to process core points and boundary points with different strategies and generate high quality pseudo instance labels to guide iterative training experiments demonstrate that our method can achieve comparable results with recent fully supervised methods by generating pseudo instance labels from categorical semantic labels our designed approach can also assist existing methods for learning instance segmentation at reduced annotation cost keyword raw image there is no result
1
10,921
13,724,347,743
IssuesEvent
2020-10-03 13:51:53
MicrosoftDocs/azure-devops-docs
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
closed
IP ranges to access non-AKS cluster
Pri2 devops-cicd-process/tech devops/prod
When adding a Kubernetes resource which is not AKS, the resource details are retrieved from IPs not found on any public list of service tag for whitelisting (example: 40.74.28.3 for West Europe). Are there any fixed IPs that can be added to the firewall so that access to the Kubernetes API can be limited to Azure DevOps only? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 7730ae4d-4101-9c83-1823-4ff43ff161ce * Version Independent ID: 20a7e263-4819-783e-c984-c4f3b459e22f * Content: [Environment - Kubernetes resource - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/environments-kubernetes?view=azure-devops) * Content Source: [docs/pipelines/process/environments-kubernetes.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/environments-kubernetes.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
1.0
IP ranges to access non-AKS cluster - When adding a Kubernetes resource which is not AKS, the resource details are retrieved from IPs not found on any public list of service tag for whitelisting (example: 40.74.28.3 for West Europe). Are there any fixed IPs that can be added to the firewall so that access to the Kubernetes API can be limited to Azure DevOps only? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 7730ae4d-4101-9c83-1823-4ff43ff161ce * Version Independent ID: 20a7e263-4819-783e-c984-c4f3b459e22f * Content: [Environment - Kubernetes resource - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/environments-kubernetes?view=azure-devops) * Content Source: [docs/pipelines/process/environments-kubernetes.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/environments-kubernetes.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
process
ip ranges to access non aks cluster when adding a kubernetes resource which is not aks the resource details are retrieved from ips not found on any public list of service tag for whitelisting example for west europe are there any fixed ips that can be added to the firewall so that access to the kubernetes api can be limited to azure devops only document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
1
79,251
10,114,876,801
IssuesEvent
2019-07-30 20:17:48
uiowa/uiowa01
https://api.github.com/repos/uiowa/uiowa01
opened
Add documentation for removing feature branch created by blt uis
documentation
`git checkout master` `git branch -D initialize-sitename` `git push origin :initialize-sitename`
1.0
Add documentation for removing feature branch created by blt uis - `git checkout master` `git branch -D initialize-sitename` `git push origin :initialize-sitename`
non_process
add documentation for removing feature branch created by blt uis git checkout master git branch d initialize sitename git push origin initialize sitename
0
11,200
13,957,702,949
IssuesEvent
2020-10-24 08:13:46
alexanderkotsev/geoportal
https://api.github.com/repos/alexanderkotsev/geoportal
opened
MT: Harvest
Geoportal Harvesting process MT - Malta
Good morning Angelo, Can you kindly perform a harvest on the Maltese CSW as we made some changes and would like to check the results. Thanks in advance for your support. Regards, Rene
1.0
MT: Harvest - Good morning Angelo, Can you kindly perform a harvest on the Maltese CSW as we made some changes and would like to check the results. Thanks in advance for your support. Regards, Rene
process
mt harvest good morning angelo can you kindly perform a harvest on the maltese csw as we made some changes and would like to check the results thanks in advance for your support regards rene
1
15,646
19,846,247,944
IssuesEvent
2022-01-21 06:49:45
ooi-data/CE04OSSM-RID26-06-PHSEND000-recovered_inst-phsen_abcdef_instrument
https://api.github.com/repos/ooi-data/CE04OSSM-RID26-06-PHSEND000-recovered_inst-phsen_abcdef_instrument
opened
🛑 Processing failed: ValueError
process
## Overview `ValueError` found in `processing_task` task during run ended on 2022-01-21T06:49:44.709173. ## Details Flow name: `CE04OSSM-RID26-06-PHSEND000-recovered_inst-phsen_abcdef_instrument` Task name: `processing_task` Error type: `ValueError` Error message: not enough values to unpack (expected 3, got 0) <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing final_path = finalize_data_stream( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream append_to_zarr(mod_ds, final_store, enc, logger=logger) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr _append_zarr(store, mod_ds) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr existing_arr.append(var_data.values) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values return _as_array_or_item(self._data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item data = np.asarray(data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__ x = self.compute() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute (result,) = compute(self, traverse=False, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute results = schedule(dsk, keys, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get results = get_async( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async raise_exception(exc, tb) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise raise exc File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task result = _execute_task(task, data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task return func(*(_execute_task(a, cache) for a in args)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter c = np.asarray(c) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__ self._ensure_cached() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached self.array = NumpyIndexingAdapter(np.asarray(self.array)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__ return array[key.tuple] File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__ return self.get_basic_selection(selection, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection return self._get_basic_selection_nd(selection=selection, out=out, File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd return self._get_selection(indexer=indexer, out=out, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection lchunk_coords, lchunk_selection, lout_selection = zip(*indexer) ValueError: not enough values to unpack (expected 3, got 0) ``` </details>
1.0
🛑 Processing failed: ValueError - ## Overview `ValueError` found in `processing_task` task during run ended on 2022-01-21T06:49:44.709173. ## Details Flow name: `CE04OSSM-RID26-06-PHSEND000-recovered_inst-phsen_abcdef_instrument` Task name: `processing_task` Error type: `ValueError` Error message: not enough values to unpack (expected 3, got 0) <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing final_path = finalize_data_stream( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream append_to_zarr(mod_ds, final_store, enc, logger=logger) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr _append_zarr(store, mod_ds) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr existing_arr.append(var_data.values) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values return _as_array_or_item(self._data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item data = np.asarray(data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__ x = self.compute() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute (result,) = compute(self, traverse=False, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute results = schedule(dsk, keys, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get results = get_async( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async raise_exception(exc, tb) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise raise exc File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task result = _execute_task(task, data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task return func(*(_execute_task(a, cache) for a in args)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter c = np.asarray(c) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__ self._ensure_cached() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached self.array = NumpyIndexingAdapter(np.asarray(self.array)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__ return array[key.tuple] File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__ return self.get_basic_selection(selection, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection return self._get_basic_selection_nd(selection=selection, out=out, File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd return self._get_selection(indexer=indexer, out=out, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection lchunk_coords, lchunk_selection, lout_selection = zip(*indexer) ValueError: not enough values to unpack (expected 3, got 0) ``` </details>
process
🛑 processing failed valueerror overview valueerror found in processing task task during run ended on details flow name recovered inst phsen abcdef instrument task name processing task error type valueerror error message not enough values to unpack expected got traceback traceback most recent call last file srv conda envs notebook lib site packages ooi harvester processor pipeline py line in processing final path finalize data stream file srv conda envs notebook lib site packages ooi harvester processor init py line in finalize data stream append to zarr mod ds final store enc logger logger file srv conda envs notebook lib site packages ooi harvester processor init py line in append to zarr append zarr store mod ds file srv conda envs notebook lib site packages ooi harvester processor utils py line in append zarr existing arr append var data values file srv conda envs notebook lib site packages xarray core variable py line in values return as array or item self data file srv conda envs notebook lib site packages xarray core variable py line in as array or item data np asarray data file srv conda envs notebook lib site packages dask array core py line in array x self compute file srv conda envs notebook lib site packages dask base py line in compute result compute self traverse false kwargs file srv conda envs notebook lib site packages dask base py line in compute results schedule dsk keys kwargs file srv conda envs notebook lib site packages dask threaded py line in get results get async file srv conda envs notebook lib site packages dask local py line in get async raise exception exc tb file srv conda envs notebook lib site packages dask local py line in reraise raise exc file srv conda envs notebook lib site packages dask local py line in execute task result execute task task data file srv conda envs notebook lib site packages dask core py line in execute task return func execute task a cache for a in args file srv conda envs notebook lib site packages dask array core py line in getter c np asarray c file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array self ensure cached file srv conda envs notebook lib site packages xarray core indexing py line in ensure cached self array numpyindexingadapter np asarray self array file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray backends zarr py line in getitem return array file srv conda envs notebook lib site packages zarr core py line in getitem return self get basic selection selection fields fields file srv conda envs notebook lib site packages zarr core py line in get basic selection return self get basic selection nd selection selection out out file srv conda envs notebook lib site packages zarr core py line in get basic selection nd return self get selection indexer indexer out out fields fields file srv conda envs notebook lib site packages zarr core py line in get selection lchunk coords lchunk selection lout selection zip indexer valueerror not enough values to unpack expected got
1
22,563
31,789,564,147
IssuesEvent
2023-09-13 01:26:00
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
`bazel build @local_config_cc//...` failed on macOS Big Sur
P3 platform: apple type: support / not a bug (process) team-Rules-CPP stale
<!-- ATTENTION! Please read and follow: - if this is a _question_ about how to build / test / query / deploy using Bazel, or a _discussion starter_, send it to bazel-discuss@googlegroups.com - if this is a _bug_ or _feature request_, fill the form below as best as you can. --> ### Description of the problem : `bazel build @local_config_cc//...` failed on macOS Big Sur. Error: `no such target '@local_config_cc_toolchains//:osx_archs.bzl': target 'osx_archs.bzl' not declared in package '';`. The complete error output: ```bash bazel build @local_config_cc//... --sandbox_debug Starting local Bazel server and connecting to it... ERROR: /private/var/tmp/_bazel_.../8f777a299a24713562d3cc77c8ffd989/external/local_config_cc/BUILD:12:6: no such target '@local_config_cc_toolchains//:osx_archs.bzl': target 'osx_archs.bzl' not declared in package ''; however, a source file of this name exists. (Perhaps add 'exports_files(["osx_archs.bzl"])' to /BUILD?) defined by /private/var/tmp/_bazel_.../8f777a299a24713562d3cc77c8ffd989/external/local_config_cc_toolchains/BUILD and referenced by '@local_config_cc//:osx_archs.bzl' ERROR: Analysis of target '@local_config_cc//:osx_archs.bzl' failed; build aborted: Analysis failed INFO: Elapsed time: 3.041s INFO: 0 processes. FAILED: Build did NOT complete successfully (4 packages loaded, 0 targets configured) ``` ### Bugs: what's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible. run `bazel build @local_config_cc//...` on macOS Big Sur. ### What operating system are you running Bazel on? macOS Big Sur 11.6 ### What's the output of `bazel info release`? release 4.2.2 ### Have you found anything relevant by searching the web? No. <!-- Places to look: - StackOverflow: http://stackoverflow.com/questions/tagged/bazel - GitHub issues: https://github.com/bazelbuild/bazel/issues - email threads on https://groups.google.com/forum/#!forum/bazel-discuss --> ### Any other information, logs, or outputs that you want to share? Command output: ```bash bazel build @local_config_cc//... --sandbox_debug Starting local Bazel server and connecting to it... ERROR: /private/var/tmp/_bazel_.../8f777a299a24713562d3cc77c8ffd989/external/local_config_cc/BUILD:12:6: no such target '@local_config_cc_toolchains//:osx_archs.bzl': target 'osx_archs.bzl' not declared in package ''; however, a source file of this name exists. (Perhaps add 'exports_files(["osx_archs.bzl"])' to /BUILD?) defined by /private/var/tmp/_bazel_.../8f777a299a24713562d3cc77c8ffd989/external/local_config_cc_toolchains/BUILD and referenced by '@local_config_cc//:osx_archs.bzl' ERROR: Analysis of target '@local_config_cc//:osx_archs.bzl' failed; build aborted: Analysis failed INFO: Elapsed time: 3.041s INFO: 0 processes. FAILED: Build did NOT complete successfully (4 packages loaded, 0 targets configured) ``` <!-- If the files are large, upload as attachment or provide link. -->
1.0
`bazel build @local_config_cc//...` failed on macOS Big Sur - <!-- ATTENTION! Please read and follow: - if this is a _question_ about how to build / test / query / deploy using Bazel, or a _discussion starter_, send it to bazel-discuss@googlegroups.com - if this is a _bug_ or _feature request_, fill the form below as best as you can. --> ### Description of the problem : `bazel build @local_config_cc//...` failed on macOS Big Sur. Error: `no such target '@local_config_cc_toolchains//:osx_archs.bzl': target 'osx_archs.bzl' not declared in package '';`. The complete error output: ```bash bazel build @local_config_cc//... --sandbox_debug Starting local Bazel server and connecting to it... ERROR: /private/var/tmp/_bazel_.../8f777a299a24713562d3cc77c8ffd989/external/local_config_cc/BUILD:12:6: no such target '@local_config_cc_toolchains//:osx_archs.bzl': target 'osx_archs.bzl' not declared in package ''; however, a source file of this name exists. (Perhaps add 'exports_files(["osx_archs.bzl"])' to /BUILD?) defined by /private/var/tmp/_bazel_.../8f777a299a24713562d3cc77c8ffd989/external/local_config_cc_toolchains/BUILD and referenced by '@local_config_cc//:osx_archs.bzl' ERROR: Analysis of target '@local_config_cc//:osx_archs.bzl' failed; build aborted: Analysis failed INFO: Elapsed time: 3.041s INFO: 0 processes. FAILED: Build did NOT complete successfully (4 packages loaded, 0 targets configured) ``` ### Bugs: what's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible. run `bazel build @local_config_cc//...` on macOS Big Sur. ### What operating system are you running Bazel on? macOS Big Sur 11.6 ### What's the output of `bazel info release`? release 4.2.2 ### Have you found anything relevant by searching the web? No. <!-- Places to look: - StackOverflow: http://stackoverflow.com/questions/tagged/bazel - GitHub issues: https://github.com/bazelbuild/bazel/issues - email threads on https://groups.google.com/forum/#!forum/bazel-discuss --> ### Any other information, logs, or outputs that you want to share? Command output: ```bash bazel build @local_config_cc//... --sandbox_debug Starting local Bazel server and connecting to it... ERROR: /private/var/tmp/_bazel_.../8f777a299a24713562d3cc77c8ffd989/external/local_config_cc/BUILD:12:6: no such target '@local_config_cc_toolchains//:osx_archs.bzl': target 'osx_archs.bzl' not declared in package ''; however, a source file of this name exists. (Perhaps add 'exports_files(["osx_archs.bzl"])' to /BUILD?) defined by /private/var/tmp/_bazel_.../8f777a299a24713562d3cc77c8ffd989/external/local_config_cc_toolchains/BUILD and referenced by '@local_config_cc//:osx_archs.bzl' ERROR: Analysis of target '@local_config_cc//:osx_archs.bzl' failed; build aborted: Analysis failed INFO: Elapsed time: 3.041s INFO: 0 processes. FAILED: Build did NOT complete successfully (4 packages loaded, 0 targets configured) ``` <!-- If the files are large, upload as attachment or provide link. -->
process
bazel build local config cc failed on macos big sur attention please read and follow if this is a question about how to build test query deploy using bazel or a discussion starter send it to bazel discuss googlegroups com if this is a bug or feature request fill the form below as best as you can description of the problem bazel build local config cc failed on macos big sur error no such target local config cc toolchains osx archs bzl target osx archs bzl not declared in package the complete error output bash bazel build local config cc sandbox debug starting local bazel server and connecting to it error private var tmp bazel external local config cc build no such target local config cc toolchains osx archs bzl target osx archs bzl not declared in package however a source file of this name exists perhaps add exports files to build defined by private var tmp bazel external local config cc toolchains build and referenced by local config cc osx archs bzl error analysis of target local config cc osx archs bzl failed build aborted analysis failed info elapsed time info processes failed build did not complete successfully packages loaded targets configured bugs what s the simplest easiest way to reproduce this bug please provide a minimal example if possible run bazel build local config cc on macos big sur what operating system are you running bazel on macos big sur what s the output of bazel info release release have you found anything relevant by searching the web no places to look stackoverflow github issues email threads on any other information logs or outputs that you want to share command output bash bazel build local config cc sandbox debug starting local bazel server and connecting to it error private var tmp bazel external local config cc build no such target local config cc toolchains osx archs bzl target osx archs bzl not declared in package however a source file of this name exists perhaps add exports files to build defined by private var tmp bazel external local config cc toolchains build and referenced by local config cc osx archs bzl error analysis of target local config cc osx archs bzl failed build aborted analysis failed info elapsed time info processes failed build did not complete successfully packages loaded targets configured
1
810,859
30,264,933,359
IssuesEvent
2023-07-07 11:03:58
mozilla/addons-server
https://api.github.com/repos/mozilla/addons-server
closed
optimize MLBF generation so it can actually finish
priority:p1
`fetch_blocked_from_db `executes a query that returns `BlockVersion` instances, but then accesses fks to `Block` and `Version` for each instance - effectively executing an extra 2 queries for every blocked guid. On dev and stage there were sufficiently few blocks that we didn't notice; on prod not so much. The generation cron usually takes 2-3 minutes (according to statsd pings recorded in grafana); last night it took over an hour. Considering we run this cron 4 times a day this is way too slow and puts an unacceptable strain on the database - if it doesn't entirely timeout.
1.0
optimize MLBF generation so it can actually finish - `fetch_blocked_from_db `executes a query that returns `BlockVersion` instances, but then accesses fks to `Block` and `Version` for each instance - effectively executing an extra 2 queries for every blocked guid. On dev and stage there were sufficiently few blocks that we didn't notice; on prod not so much. The generation cron usually takes 2-3 minutes (according to statsd pings recorded in grafana); last night it took over an hour. Considering we run this cron 4 times a day this is way too slow and puts an unacceptable strain on the database - if it doesn't entirely timeout.
non_process
optimize mlbf generation so it can actually finish fetch blocked from db executes a query that returns blockversion instances but then accesses fks to block and version for each instance effectively executing an extra queries for every blocked guid on dev and stage there were sufficiently few blocks that we didn t notice on prod not so much the generation cron usually takes minutes according to statsd pings recorded in grafana last night it took over an hour considering we run this cron times a day this is way too slow and puts an unacceptable strain on the database if it doesn t entirely timeout
0
4,579
7,410,055,139
IssuesEvent
2018-03-21 00:00:13
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Reader role can't view functions within a function app
active-directory assigned-to-author doc-enhancement in-process triaged
This document doesn't mention it, but users with reader level access also can't see the functions within a function app. In left navigation, you see: Functions (No access) Proxies (No access) --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 8b3ea2e8-963b-cc01-3d5c-4fecd6a3cec1 * Version Independent ID: 198149eb-9b33-ec3f-527f-ae64fb96d65c * Content: [Troubleshoot role-based access control Azure RBAC | Microsoft Docs](https://docs.microsoft.com/en-us/azure/active-directory/role-based-access-control-troubleshooting) * Content Source: [articles/active-directory/role-based-access-control-troubleshooting.md](https://github.com/Microsoft/azure-docs/blob/master/articles/active-directory/role-based-access-control-troubleshooting.md) * Service: **active-directory** * GitHub Login: @rolyon * Microsoft Alias: **rolyon**
1.0
Reader role can't view functions within a function app - This document doesn't mention it, but users with reader level access also can't see the functions within a function app. In left navigation, you see: Functions (No access) Proxies (No access) --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 8b3ea2e8-963b-cc01-3d5c-4fecd6a3cec1 * Version Independent ID: 198149eb-9b33-ec3f-527f-ae64fb96d65c * Content: [Troubleshoot role-based access control Azure RBAC | Microsoft Docs](https://docs.microsoft.com/en-us/azure/active-directory/role-based-access-control-troubleshooting) * Content Source: [articles/active-directory/role-based-access-control-troubleshooting.md](https://github.com/Microsoft/azure-docs/blob/master/articles/active-directory/role-based-access-control-troubleshooting.md) * Service: **active-directory** * GitHub Login: @rolyon * Microsoft Alias: **rolyon**
process
reader role can t view functions within a function app this document doesn t mention it but users with reader level access also can t see the functions within a function app in left navigation you see functions no access proxies no access document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service active directory github login rolyon microsoft alias rolyon
1
9,819
12,826,993,332
IssuesEvent
2020-07-06 17:36:08
MicrosoftDocs/azure-devops-docs
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
closed
Deployment Pipeline not trigger when build pipeline complete
Pri1 devops-cicd-process/tech devops/prod doc-bug
Both my build and deployment pipelines are yaml based. I want the deployment pipeline to trigger when the build pipeline has completed. I have added the following block to my deployment pipeline yaml file ```yaml ### DISABLE RUN ON Commit trigger: none ### ENABLE RUN WHEN THE BUILD PIPELINE RunS resources: pipelines: - pipeline: buildPipelnie # Name of the pipeline resource source: 'Name of my Build Pipeline' # Name of the triggering pipeline trigger: branches: - prefix-* ``` I am not sure I am doing it right or not, but I am following the doc. I have also tried disabling the `trigger: none` ```yaml ### DISABLE RUN ON Commit ##trigger: none ### ENABLE RUN WHEN THE BUILD PIPELINE Run resources: pipelines: - pipeline: buildPipelnie # Name of the pipeline resource source: 'Name of my Build Pipeline' # Name of the triggering pipeline trigger: branches: - prefix-* ``` And didn't work. I have also tried the `include/exclude` syntax of the `branches.prefix` block. And it didn't work either. Any suggestions would be greatly appreciated --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 86285f72-9e28-da97-59bb-c29eb60f627d * Version Independent ID: 18d5a591-a7d3-c261-6bff-8808ae433f54 * Content: [Configure pipeline triggers - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/pipeline-triggers?view=azure-devops&tabs=yaml) * Content Source: [docs/pipelines/process/pipeline-triggers.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/pipeline-triggers.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @ashkir * Microsoft Alias: **ashkir**
1.0
Deployment Pipeline not trigger when build pipeline complete - Both my build and deployment pipelines are yaml based. I want the deployment pipeline to trigger when the build pipeline has completed. I have added the following block to my deployment pipeline yaml file ```yaml ### DISABLE RUN ON Commit trigger: none ### ENABLE RUN WHEN THE BUILD PIPELINE RunS resources: pipelines: - pipeline: buildPipelnie # Name of the pipeline resource source: 'Name of my Build Pipeline' # Name of the triggering pipeline trigger: branches: - prefix-* ``` I am not sure I am doing it right or not, but I am following the doc. I have also tried disabling the `trigger: none` ```yaml ### DISABLE RUN ON Commit ##trigger: none ### ENABLE RUN WHEN THE BUILD PIPELINE Run resources: pipelines: - pipeline: buildPipelnie # Name of the pipeline resource source: 'Name of my Build Pipeline' # Name of the triggering pipeline trigger: branches: - prefix-* ``` And didn't work. I have also tried the `include/exclude` syntax of the `branches.prefix` block. And it didn't work either. Any suggestions would be greatly appreciated --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 86285f72-9e28-da97-59bb-c29eb60f627d * Version Independent ID: 18d5a591-a7d3-c261-6bff-8808ae433f54 * Content: [Configure pipeline triggers - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/pipeline-triggers?view=azure-devops&tabs=yaml) * Content Source: [docs/pipelines/process/pipeline-triggers.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/pipeline-triggers.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @ashkir * Microsoft Alias: **ashkir**
process
deployment pipeline not trigger when build pipeline complete both my build and deployment pipelines are yaml based i want the deployment pipeline to trigger when the build pipeline has completed i have added the following block to my deployment pipeline yaml file yaml disable run on commit trigger none enable run when the build pipeline runs resources pipelines pipeline buildpipelnie name of the pipeline resource source name of my build pipeline name of the triggering pipeline trigger branches prefix i am not sure i am doing it right or not but i am following the doc i have also tried disabling the trigger none yaml disable run on commit trigger none enable run when the build pipeline run resources pipelines pipeline buildpipelnie name of the pipeline resource source name of my build pipeline name of the triggering pipeline trigger branches prefix and didn t work i have also tried the include exclude syntax of the branches prefix block and it didn t work either any suggestions would be greatly appreciated document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login ashkir microsoft alias ashkir
1
7,878
11,046,697,201
IssuesEvent
2019-12-09 17:21:47
prisma/quaint
https://api.github.com/repos/prisma/quaint
closed
Change default postgres settings
process/candidate
Write-up of the issue: https://github.com/prisma/specs/issues/325 We discussed it with Matt and Jan, and decided that for now, we should align on what all other clients are doing, i.e. accept invalid (=not trusted) ssl certificates by default.
1.0
Change default postgres settings - Write-up of the issue: https://github.com/prisma/specs/issues/325 We discussed it with Matt and Jan, and decided that for now, we should align on what all other clients are doing, i.e. accept invalid (=not trusted) ssl certificates by default.
process
change default postgres settings write up of the issue we discussed it with matt and jan and decided that for now we should align on what all other clients are doing i e accept invalid not trusted ssl certificates by default
1
66,380
20,163,563,405
IssuesEvent
2022-02-10 00:33:17
SeleniumHQ/selenium
https://api.github.com/repos/SeleniumHQ/selenium
closed
[🐛 Bug]: CLOSED
I-defect needs-triaging
### What happened? If i try to locate the the element <span> rzrmhxbwuiteam.testinator.com</span> in the below html code i get error: OpenQA.Selenium.StaleElementReferenceException: "stale element reference: element is not attached to the page document (Session info: chrome=98.0.4758.82)" But my code is right with inspector i can locate the element by search with //div[@class='description-block']/div/p[2]/span and how you see there is NO iframe in the code. So why i cant get locate the element. I think there is something with that site that block it. I realy need to get acces to this part of the HTML and need a solution maybe there is something in selenium i dont know to trick it. I Use Chromedriver as follow: program.options.AddArguments("disable-blink-features=AutomationControlled"); program.driver3 = new ChromeDriver(program.options); This is the full html code of the site <html lang="en"><head> <meta charset="utf-8"> <meta name="viewport" content="width=device-width,initial-scale=1,shrink-to-fit=no"> <link rel="stylesheet" href="../css/bootstrap.min.css"> <link rel="stylesheet" href="../../v4/css/style.bundle.css"> <title>Mailinator</title> <script src="../js/jquery-3.5.1.min.js"></script> <script src="../js/bootstrap.min.js"></script> <script type="text/javascript" src="js/m8rauth.js"></script> <script src="../js/jquery.OverlayScrollbars.js"></script> <style> a { color: darkblue; } </style></head> <body><header class="header"> <div class="primary-header secondary-header"> <div class="d-flex flex-wrap justify-content-between"> <div class="order-0 p-y-15"> <div class="wrapper-nav-brand d-flex align-items-center h-100"> <div id="menuToggle" class="d-flex flex-column"> <button type="button" class="hamburger-menu" data-toggle="modal" data-target="#nav-top-header" aria-label="Menu button"><img src="../img/homepage/menu-icon.svg" alt="hamburger menu"> </button> </div> <div class="modal fade nav-top-header" id="nav-top-header" tabindex="-1" role="dialog" aria-labelledby="nav-top-header"> <div class="modal-dialog h-100" role="document"> <div class="modal-content h-100"> <div class="modal-header align-items-center"> <div class="wrapper-modal-title d-flex align-items-center justify-content-center"> <div class="modal-logo"> <img src="https://www.mailinator.com/site/wp-content/uploads/2021/05/Mailinator-Logo.svg" width="80" height="30" alt="modal logo"> </div> </div> <button type="button" class="close" data-dismiss="modal" aria-label="Close"> <img src="../img/homepage/close-button.svg" alt="close button"> </button> </div> <!------------- TOP NAV - MOBILE -----------> <div class="modal-body"> <div id="modal-top" class="wrapper-modal d-flex flex-column"> <a href="/index.jsp" class="modal-item" aria-label="Home"> Home </a> <a href="/v4/accessibility.html" class="modal-item" aria-label="Terms"> Accessibility</a> <a href="/v4/terms.html" class="modal-item" aria-label="Terms"> Terms</a> <a href="/v4/privacy.html" class="modal-item" aria-label="Privacy Policy"> Privacy Policy </a> <a href="#" class="modal-item copyright" aria-label="© Copyright 2022 Manybrain, LLC. All Rights Reserved"> Copyright 2022 Manybrain, LLC. All Rights Reserved </a> </div> </div> <div class="modal-footer border-0"> <a href="/v4/logout.jsp" style="color:#000" class="modal-login d-flex justify-content-center align-items-center w-100" aria-label="Login"> Logout </a> </div> </div> </div> </div> <div class="nav-logo"> <a href="/"> <img src="https://www.mailinator.com/site/wp-content/uploads/2021/05/Mailinator-Logo.svg" width="200" height="50" alt="modal logo"> </a> </div> </div> </div> <div class="order-2 order-xl-1 d-flex lg-center align-items-center fixed-header"> <div class="wrapper-nav-inputs"> <div class="input-group"> <input class="form-control input-text h-auto" placeholder="Enter Inbox Name" aria-label="Enter Inbox Name" aria-describedby="go-to-public" id="addOverlay"> <div class="input-group-append"> <button class="btn btn-go-public h-auto" type="button" id="go-to-public" aria-label="Go to public"> GO TO PUBLIC MAILINATOR </button> </div> </div> </div> </div> <!------------- TOP NAV - DESKTOP -----------> <div class="order-1 order-xl-2 d-flex align-items-center justify-content-end"> <div class="wrapper-nav-items d-flex align-items-center h-100"> <a href="/v4/private/inboxes.jsp" class="nav-item" aria-label="Email"> Email </a> <a href="https://www.mailinator.com/site/pricing/" class="nav-item" aria-label="Pricing"> Pricing </a> <a href="https://www.mailinator.com/docs/index.html" class="nav-item hidden-link-header" aria-label="Documentation"> Documentation </a> <a href="https://www.mailinator.com/site/mailinator-faqs/" class="nav-item" aria-label="FAQ"> FAQ </a> <a href="/v4/private/profile.jsp" class="nav-item d-flex align-items-center justify-content-center" style="color:#FFF;font-size: .8em;" aria-label="User Email"> 1621917affiliate123@it-ce.de <svg width="20" height="20" viewBox="0 0 20 20" fill="none" xmlns="http://www.w3.org/2000/svg"> <path fill-rule="evenodd" clip-rule="evenodd" d="M5.29289 7.29289C5.68342 6.90237 6.31658 6.90237 6.70711 7.29289L10 10.5858L13.2929 7.29289C13.6834 6.90237 14.3166 6.90237 14.7071 7.29289C15.0976 7.68342 15.0976 8.31658 14.7071 8.70711L10.7071 12.7071C10.3166 13.0976 9.68342 13.0976 9.29289 12.7071L5.29289 8.70711C4.90237 8.31658 4.90237 7.68342 5.29289 7.29289Z" fill="#ffffff"></path> </svg> </a> <button class="button-primary" onclick="window.location='/v4/logout.jsp';" style="color: white;background-color: #000;padding:5px;font-size: 9px;"> LOGOUT </button> </div> </div> </div> </div> </header> <div class="default-class main d-flex flex-grow-1"> <!-------------------------------------- LEFT NAV --------------------------------------------> <aside class="sidebar d-flex flex-column"> <div class="d-flex flex-column primary-line m-b-50"> <a class="sidebar-links m-b-25 d-flex align-items-center text-decoration-none" href="/v4/private/inboxes.jsp"> <svg class="sidebar-svg m-r-15" width="20" height="16" viewBox="0 0 20 16" fill="none" xmlns="http://www.w3.org/2000/svg"> <path d="M18 0H2C0.9 0 0.00999999 0.9 0.00999999 2L0 14C0 15.1 0.9 16 2 16H18C19.1 16 20 15.1 20 14V2C20 0.9 19.1 0 18 0ZM18 4L10 9L2 4V2L10 7L18 2V4Z" fill="#60719B"></path> </svg> Private Team Inbox </a> <a class="sidebar-links m-b-25 d-flex align-items-center text-decoration-none" href="/v4/private/team_settings.jsp"> <svg class="sidebar-svg m-r-15" width="20" height="21" viewBox="0 0 20 21" fill="none" xmlns="http://www.w3.org/2000/svg"> <path d="M17.6502 11.5237C17.6888 11.1956 17.7145 10.8544 17.7145 10.5C17.7145 10.1456 17.6888 9.80438 17.6373 9.47625L19.8102 7.74375C20.0031 7.58625 20.0545 7.2975 19.9388 7.07437L17.8816 3.43875C17.753 3.2025 17.483 3.12375 17.2516 3.2025L14.693 4.2525C14.153 3.8325 13.5872 3.49125 12.9572 3.22875L12.5715 0.44625C12.5329 0.18375 12.3143 0 12.0572 0H7.94281C7.68566 0 7.47994 0.18375 7.44137 0.44625L7.05565 3.22875C6.42563 3.49125 5.84705 3.84562 5.31989 4.2525L2.76126 3.2025C2.52982 3.11063 2.25982 3.2025 2.13124 3.43875L0.0740529 7.07437C-0.0545216 7.31062 -0.00309159 7.58625 0.202628 7.74375L2.37554 9.47625C2.32411 9.80438 2.28553 10.1587 2.28553 10.5C2.28553 10.8413 2.31125 11.1956 2.36268 11.5237L0.18977 13.2563C-0.00309189 13.4138 -0.0545215 13.7025 0.0611955 13.9256L2.11839 17.5613C2.24696 17.7975 2.51697 17.8763 2.7484 17.7975L5.30703 16.7475C5.84705 17.1675 6.41277 17.5087 7.04279 17.7712L7.42851 20.5537C7.47994 20.8162 7.68566 21 7.94281 21H12.0572C12.3143 21 12.5329 20.8162 12.5586 20.5537L12.9444 17.7712C13.5744 17.5087 14.153 17.1544 14.6801 16.7475L17.2387 17.7975C17.4702 17.8894 17.7402 17.7975 17.8688 17.5613L19.9259 13.9256C20.0545 13.6894 20.0031 13.4138 19.7974 13.2563L17.6502 11.5237ZM10 14.4375C7.87852 14.4375 6.14277 12.6656 6.14277 10.5C6.14277 8.33438 7.87852 6.5625 10 6.5625C12.1215 6.5625 13.8572 8.33438 13.8572 10.5C13.8572 12.6656 12.1215 14.4375 10 14.4375Z" fill="#60719B"></path> </svg> Team Settings </a> <a class="sidebar-links m-b-25 d-flex align-items-center text-decoration-none" href="/v4/private/team_management.jsp"> <svg class="sidebar-svg m-r-15" width="20" height="20" viewBox="0 0 20 20" fill="none" <defs=""> <style>.cls-1 { fill: #fff; opacity: 0; } .cls-2 { fill: #60719B }</style> <title>people</title> <g id="Layer_2" data-name="Layer 2"> <g id="people"> <g id="people-2" data-name="people"> <rect class="cls-1" width="24" height="24" fill="#60719B"></rect> <path class="cls-2" d="M9,11A4,4,0,1,0,5,7,4,4,0,0,0,9,11Z" fill="#60719B"></path> <path class="cls-2" d="M17,13a3,3,0,1,0-3-3A3,3,0,0,0,17,13Z" fill="#60719B"></path> <path class="cls-2" d="M21,20a1,1,0,0,0,1-1,5,5,0,0,0-8.06-3.95A7,7,0,0,0,2,20a1,1,0,0,0,1,1H15a1,1,0,0,0,1-1" fill="#60719B"></path> </g> </g> </g> </svg> Team Management </a> <a class="sidebar-links m-b-30 d-flex align-items-center text-decoration-none" href="/v4/private/stats.jsp"> <svg class="sidebar-svg m-r-15" width="20" height="16" viewBox="0 0 20 16" fill="none" xmlns="http://www.w3.org/2000/svg"> <g class="bars"> <rect fill="#60719B" width="35%" height="25"></rect>; <rect fill="#60719B" width="45%" height="20"></rect> </g> </svg> Statistics </a> <a class="sidebar-links m-b-25 d-flex align-items-center text-decoration-none" href="/v4/private/message_rules.jsp"> <svg class="sidebar-svg m-r-15" width="22" height="21" viewBox="0 0 22 21" fill="none" xmlns="http://www.w3.org/2000/svg"> <path d="M13.0005 13.5001V11.0001H9.00055V14.0001H7.00055V10.0001C7.00055 9.73484 7.1059 9.48049 7.29344 9.29295C7.48098 9.10542 7.73533 9.00006 8.00055 9.00006H13.0005V6.50006L16.5005 10.0001L13.0005 13.5001ZM20.7105 10.2901L11.7105 1.29006H11.7005C11.3105 0.900059 10.6805 0.900059 10.2905 1.29006L1.29055 10.2901C0.900547 10.6801 0.900547 11.3201 1.29055 11.7101L10.2905 20.7101C10.6805 21.0901 11.3105 21.1001 11.7105 20.7101L20.7105 11.7101C21.1005 11.3201 21.1005 10.6801 20.7105 10.2901Z" fill="#60719B"></path> </svg> Message Rules </a> <a class="sidebar-links p-b-25 primary-border-b d-flex align-items-center text-decoration-none" href="/v4/private/quick_start.jsp"> <svg class="sidebar-svg m-r-15" width="16" height="20" viewBox="0 0 16 20" fill="none" xmlns="http://www.w3.org/2000/svg"> <path d="M9 7H14.5L9 1.5V7ZM2 0H10L16 6V18C16 18.5304 15.7893 19.0391 15.4142 19.4142C15.0391 19.7893 14.5304 20 14 20H2C0.89 20 0 19.1 0 18V2C0 0.89 0.89 0 2 0ZM10 16V14H3V16H10ZM13 12V10H3V12H13Z" fill="#60719B"></path> </svg> Quick Start Docs </a> </div> <div class="d-flex flex-column primary-line m-b-50"> <a class="sidebar-links m-b-30 d-flex align-items-center text-decoration-none" href="/v4/public/inboxes.jsp"> <svg class="sidebar-svg m-r-15" width="20" height="16" viewBox="0 0 20 16" fill="none" xmlns="http://www.w3.org/2000/svg"> <path d="M18 0H2C0.9 0 0.00999999 0.9 0.00999999 2L0 14C0 15.1 0.9 16 2 16H18C19.1 16 20 15.1 20 14V2C20 0.9 19.1 0 18 0ZM18 4L10 9L2 4V2L10 7L18 2V4Z" fill="#60719B"></path> </svg> Public Inboxes </a> <div style="text-align: center;font-weight: bold;"> Your Trial Ends in <br> <span style="color:red;"> 13 </span>days. <br> <a href="https://www.mailinator.com/site/pricing/">Subscribe Now!</a> </div> </div> <div class="d-flex flex-column justify-content-end flex-grow-1 primary-line"> <a style="color:#000;" href="https://status.mailinator.com">System Status</a> <br> <a style="color:#000;" href="https://www.mailinator.com/site/terms-of-use/">Terms</a> <br> <a style="color:#000;" href="https://www.mailinator.com/site/privacy-policy/">Privacy</a> <br> <a style="color:#000;" href="https://www.mailinator.com/site/accessibility/">Accessibility</a> <br> <a style="color:#000;" href="https://www.mailinator.com/site/security-details/">Security</a> </div> <div class="p-r-50 p-t-25"> <p style="font-size:.7em;">Copyright 2022 Manybrain, LLC. All Rights Reserved</p> </div> </aside> <!-- dont close the initial div here --> <main class="main-content"> <div class="main-container-fluid"> <div class="quick-start d-flex justify-content-between"> <h4 class="fw-700">Mailinator Quick Start</h4> <a href="https://www.mailinator.com/docs/index.html" class="link d-flex align-items-center text-decoration-none"> <svg class="m-r-10" width="16" height="16" viewBox="0 0 16 16" fill="none" xmlns="http://www.w3.org/2000/svg"> <path d="M9.77778 12.4444H3.55556V10.6667H9.77778V12.4444ZM12.4444 8.88889H3.55556V7.11111H12.4444V8.88889ZM12.4444 5.33333H3.55556V3.55556H12.4444V5.33333ZM14.2222 0H1.77778C0.791111 0 0 0.791111 0 1.77778V14.2222C0 14.6937 0.187301 15.1459 0.520699 15.4793C0.854097 15.8127 1.30628 16 1.77778 16H14.2222C14.6937 16 15.1459 15.8127 15.4793 15.4793C15.8127 15.1459 16 14.6937 16 14.2222V1.77778C16 0.791111 15.2 0 14.2222 0Z" fill="#2979FF"></path> </svg> Go to Full Documentation</a> </div> <div class="description-block"> <div class="item" style="font-size: 15px;"> <div class="m-b-20 fz-20 fw-700">Your Own Private Domain</div> <p class="description"> You now have your own Private Domain! All email in that domain stays Private to you, sticks around until you delete it (or you fill up your storage), and doesn't have rate limits !</p> <p class="description"> Your Private Domain is: <span> rzrmhxbwuiteam.testinator.com </span> </p> <p class="description"> (You can change it on the <a class="text-decoration-none" href="team_settings.jsp"> Team Settings </a> page) </p> <p class="description"> All email sent to email addresses* <b>@rzrmhxbwuiteam.testinator.com </b> will arrive in your Private Team Inbox. <br> </p> <p class="description"> You can try it right now! Send an email to: </p> <p class="description"> <span> test1@rzrmhxbwuiteam.testinator.com</span></p> <p class="description"> Then check your Private Team Inbox. </p> </div> <div class="item" style="font-size: 15px;"> <div class="m-b-20 fz-20 fw-700">API</div> <p class="description"> You also have access to all Mailinator email (public and private) through our API. You'll find your token on the Team Settings page. Use that token in a Authorization Header to authorize your api calls (see the API documentation for more information). </p> <p class="description"> We also have Mailinator API SDKs for many popular programming languages including <a href="https://github.com/manybrain/mailinator-java-client" rel="nofollow">Java</a>, <a href="https://github.com/manybrain/mailinator-csharp-client" rel="nofollow">C#</a>, <a href="https://www.npmjs.com/package/mailinator-client" rel="nofollow">Javascript</a>, <a href="https://github.com/manybrain/mailinator-ruby-client" rel="nofollow">Ruby</a>, <a href="https://github.com/manybrain/mailinator-go-client" rel="nofollow">Go</a>, and <a href="https://github.com/manybrain/mailinator-python-client" rel="nofollow">Python</a>. </p> <p class="description"> You can of course use the API to read your Private Domain email too. Consult the <a class="text-decoration-none" href="https://www.mailinator.com/docs/index.html"> API documentation </a> for specifics! </p> <p class="description">Try it now - click this link to test the API:</p> <p class="description"> <span> <a href="https://www.mailinator.com/api/v2/domains/public/inboxes/test"> www.mailinator.com/api/v2/domains/public/inboxes/test </a> </span> </p> <p class="fz-18">(Note JSON is often not well-formatted in web browsers but you get the idea).</p> </div> <div class="item" style="font-size: 15px;"> <div class="m-b-20 fz-20 fw-700">Message Routing Rules</div> <p class="description">Mailinator Message Rules allow you to tell the system to act on messages the moment they arrive.</p> <p class="description">For example, you could create rules like:</p> <div> <p class="rules m-b-10"> <span style="font-family: 'Courier New';color:#000;background-color: #fff;"> IF inbox == bob THEN <br> &nbsp; &nbsp; &nbsp; &nbsp; webhook: yourwebsite.com/rest/webhook </span> </p> <p class="rules m-b-10"> <span style="font-family: 'Courier New';color:#000;background-color: #fff;"> IF inbox == joe THEN CLICK-ALL-LINKS </span> </p> </div> <p class="description">The first rule will "push" emails, as they arrive, to a REST endpoint you setup. The second rule will automatically click all links in any email that arrives in your Private Domain's "joe" inbox.</p> <p class="description">Check out the "Message Rules" page to set them up and Documentation for more info.</p> </div> <div class="item" style="font-size: 15px;"> <div class="m-b-20 fz-20 fw-700">Team Invites</div> <p class="description">If you're a Team Admin, you can invite your co-workers into your Mailinator account. See the Team Management page for details.</p> <p class="description">We hope you enjoy Mailinator. If you have any questions, email us at <b>support@manybrain.com</b> </p> </div> </div> </div> <div style="margin-left:30px;font-size: .8em"><p> * A small number of inbox names are restricted when using a Private Subdomain (i.e. xyz.testinator.com) for security reasons. Don't fret however, you still have Trillions to choose from. <br>There are no restrictions on inbox names for Private Domains where you own the domain.</p> </div> </main> </div> <footer class="footer hide"> <div class="footer-container d-flex justify-content-between align-items-center"> <div class="wrapper-nav-footer d-flex ff-futura-book"> <div> <ul class="d-flex"> <li class="item"><a href="/" aria-label="Home"> Home </a> </li> <li class="item"><a href="/v4/public/inboxes.jsp" aria-label="Email"> Email </a> </li> <li class="item"> <div class="dropdown d-flex"> <a href="/site/pricing/" class="modal-item dropdown-toggle d-flex align-items-center" type="" id="dropdownFooter" data-toggle="dropdown" aria-haspopup="true" aria-expanded="false"> Pricing </a> </div> </li> </ul> </div> </div> <div> <ul class="d-flex"> <li class="item"> <a href="https://www.mailinator.com/site/mailinator-faqs/" aria-label="FAQ"> FAQ </a> </li> </ul> </div> <div> <ul class="d-flex"> <li class="item"> <a href="https://www.mailinator.com/site/terms-of-use/" aria-label="Terms"> Terms </a> </li> <li class="item"> <a href="https://www.mailinator.com/site/privacy-policy/" aria-label="Privacy Policy"> Privacy </a> </li> </ul> </div> </div> <div class="wrapper-copyright"> Copyright 2022 Manybrain, Inc. All Rights Reserved </div> </footer> <!-- unneeded ? --> <script src="../../v4/js/bundle.js"></script> </body></html> ### How can we reproduce the issue? ```shell Goto Mailnator and login its the first page after login. ``` ### Relevant log output ```shell No ``` ### Operating System Windows 10 ### Selenium version 4.0.1 ### What are the browser(s) and version(s) where you see this issue? Chrome 98.0.4758.82 ### What are the browser driver(s) and version(s) where you see this issue? Chrome Driver 98.0.4758.82 ### Are you using Selenium Grid? No
1.0
[🐛 Bug]: CLOSED - ### What happened? If i try to locate the the element <span> rzrmhxbwuiteam.testinator.com</span> in the below html code i get error: OpenQA.Selenium.StaleElementReferenceException: "stale element reference: element is not attached to the page document (Session info: chrome=98.0.4758.82)" But my code is right with inspector i can locate the element by search with //div[@class='description-block']/div/p[2]/span and how you see there is NO iframe in the code. So why i cant get locate the element. I think there is something with that site that block it. I realy need to get acces to this part of the HTML and need a solution maybe there is something in selenium i dont know to trick it. I Use Chromedriver as follow: program.options.AddArguments("disable-blink-features=AutomationControlled"); program.driver3 = new ChromeDriver(program.options); This is the full html code of the site <html lang="en"><head> <meta charset="utf-8"> <meta name="viewport" content="width=device-width,initial-scale=1,shrink-to-fit=no"> <link rel="stylesheet" href="../css/bootstrap.min.css"> <link rel="stylesheet" href="../../v4/css/style.bundle.css"> <title>Mailinator</title> <script src="../js/jquery-3.5.1.min.js"></script> <script src="../js/bootstrap.min.js"></script> <script type="text/javascript" src="js/m8rauth.js"></script> <script src="../js/jquery.OverlayScrollbars.js"></script> <style> a { color: darkblue; } </style></head> <body><header class="header"> <div class="primary-header secondary-header"> <div class="d-flex flex-wrap justify-content-between"> <div class="order-0 p-y-15"> <div class="wrapper-nav-brand d-flex align-items-center h-100"> <div id="menuToggle" class="d-flex flex-column"> <button type="button" class="hamburger-menu" data-toggle="modal" data-target="#nav-top-header" aria-label="Menu button"><img src="../img/homepage/menu-icon.svg" alt="hamburger menu"> </button> </div> <div class="modal fade nav-top-header" id="nav-top-header" tabindex="-1" role="dialog" aria-labelledby="nav-top-header"> <div class="modal-dialog h-100" role="document"> <div class="modal-content h-100"> <div class="modal-header align-items-center"> <div class="wrapper-modal-title d-flex align-items-center justify-content-center"> <div class="modal-logo"> <img src="https://www.mailinator.com/site/wp-content/uploads/2021/05/Mailinator-Logo.svg" width="80" height="30" alt="modal logo"> </div> </div> <button type="button" class="close" data-dismiss="modal" aria-label="Close"> <img src="../img/homepage/close-button.svg" alt="close button"> </button> </div> <!------------- TOP NAV - MOBILE -----------> <div class="modal-body"> <div id="modal-top" class="wrapper-modal d-flex flex-column"> <a href="/index.jsp" class="modal-item" aria-label="Home"> Home </a> <a href="/v4/accessibility.html" class="modal-item" aria-label="Terms"> Accessibility</a> <a href="/v4/terms.html" class="modal-item" aria-label="Terms"> Terms</a> <a href="/v4/privacy.html" class="modal-item" aria-label="Privacy Policy"> Privacy Policy </a> <a href="#" class="modal-item copyright" aria-label="© Copyright 2022 Manybrain, LLC. All Rights Reserved"> Copyright 2022 Manybrain, LLC. All Rights Reserved </a> </div> </div> <div class="modal-footer border-0"> <a href="/v4/logout.jsp" style="color:#000" class="modal-login d-flex justify-content-center align-items-center w-100" aria-label="Login"> Logout </a> </div> </div> </div> </div> <div class="nav-logo"> <a href="/"> <img src="https://www.mailinator.com/site/wp-content/uploads/2021/05/Mailinator-Logo.svg" width="200" height="50" alt="modal logo"> </a> </div> </div> </div> <div class="order-2 order-xl-1 d-flex lg-center align-items-center fixed-header"> <div class="wrapper-nav-inputs"> <div class="input-group"> <input class="form-control input-text h-auto" placeholder="Enter Inbox Name" aria-label="Enter Inbox Name" aria-describedby="go-to-public" id="addOverlay"> <div class="input-group-append"> <button class="btn btn-go-public h-auto" type="button" id="go-to-public" aria-label="Go to public"> GO TO PUBLIC MAILINATOR </button> </div> </div> </div> </div> <!------------- TOP NAV - DESKTOP -----------> <div class="order-1 order-xl-2 d-flex align-items-center justify-content-end"> <div class="wrapper-nav-items d-flex align-items-center h-100"> <a href="/v4/private/inboxes.jsp" class="nav-item" aria-label="Email"> Email </a> <a href="https://www.mailinator.com/site/pricing/" class="nav-item" aria-label="Pricing"> Pricing </a> <a href="https://www.mailinator.com/docs/index.html" class="nav-item hidden-link-header" aria-label="Documentation"> Documentation </a> <a href="https://www.mailinator.com/site/mailinator-faqs/" class="nav-item" aria-label="FAQ"> FAQ </a> <a href="/v4/private/profile.jsp" class="nav-item d-flex align-items-center justify-content-center" style="color:#FFF;font-size: .8em;" aria-label="User Email"> 1621917affiliate123@it-ce.de <svg width="20" height="20" viewBox="0 0 20 20" fill="none" xmlns="http://www.w3.org/2000/svg"> <path fill-rule="evenodd" clip-rule="evenodd" d="M5.29289 7.29289C5.68342 6.90237 6.31658 6.90237 6.70711 7.29289L10 10.5858L13.2929 7.29289C13.6834 6.90237 14.3166 6.90237 14.7071 7.29289C15.0976 7.68342 15.0976 8.31658 14.7071 8.70711L10.7071 12.7071C10.3166 13.0976 9.68342 13.0976 9.29289 12.7071L5.29289 8.70711C4.90237 8.31658 4.90237 7.68342 5.29289 7.29289Z" fill="#ffffff"></path> </svg> </a> <button class="button-primary" onclick="window.location='/v4/logout.jsp';" style="color: white;background-color: #000;padding:5px;font-size: 9px;"> LOGOUT </button> </div> </div> </div> </div> </header> <div class="default-class main d-flex flex-grow-1"> <!-------------------------------------- LEFT NAV --------------------------------------------> <aside class="sidebar d-flex flex-column"> <div class="d-flex flex-column primary-line m-b-50"> <a class="sidebar-links m-b-25 d-flex align-items-center text-decoration-none" href="/v4/private/inboxes.jsp"> <svg class="sidebar-svg m-r-15" width="20" height="16" viewBox="0 0 20 16" fill="none" xmlns="http://www.w3.org/2000/svg"> <path d="M18 0H2C0.9 0 0.00999999 0.9 0.00999999 2L0 14C0 15.1 0.9 16 2 16H18C19.1 16 20 15.1 20 14V2C20 0.9 19.1 0 18 0ZM18 4L10 9L2 4V2L10 7L18 2V4Z" fill="#60719B"></path> </svg> Private Team Inbox </a> <a class="sidebar-links m-b-25 d-flex align-items-center text-decoration-none" href="/v4/private/team_settings.jsp"> <svg class="sidebar-svg m-r-15" width="20" height="21" viewBox="0 0 20 21" fill="none" xmlns="http://www.w3.org/2000/svg"> <path d="M17.6502 11.5237C17.6888 11.1956 17.7145 10.8544 17.7145 10.5C17.7145 10.1456 17.6888 9.80438 17.6373 9.47625L19.8102 7.74375C20.0031 7.58625 20.0545 7.2975 19.9388 7.07437L17.8816 3.43875C17.753 3.2025 17.483 3.12375 17.2516 3.2025L14.693 4.2525C14.153 3.8325 13.5872 3.49125 12.9572 3.22875L12.5715 0.44625C12.5329 0.18375 12.3143 0 12.0572 0H7.94281C7.68566 0 7.47994 0.18375 7.44137 0.44625L7.05565 3.22875C6.42563 3.49125 5.84705 3.84562 5.31989 4.2525L2.76126 3.2025C2.52982 3.11063 2.25982 3.2025 2.13124 3.43875L0.0740529 7.07437C-0.0545216 7.31062 -0.00309159 7.58625 0.202628 7.74375L2.37554 9.47625C2.32411 9.80438 2.28553 10.1587 2.28553 10.5C2.28553 10.8413 2.31125 11.1956 2.36268 11.5237L0.18977 13.2563C-0.00309189 13.4138 -0.0545215 13.7025 0.0611955 13.9256L2.11839 17.5613C2.24696 17.7975 2.51697 17.8763 2.7484 17.7975L5.30703 16.7475C5.84705 17.1675 6.41277 17.5087 7.04279 17.7712L7.42851 20.5537C7.47994 20.8162 7.68566 21 7.94281 21H12.0572C12.3143 21 12.5329 20.8162 12.5586 20.5537L12.9444 17.7712C13.5744 17.5087 14.153 17.1544 14.6801 16.7475L17.2387 17.7975C17.4702 17.8894 17.7402 17.7975 17.8688 17.5613L19.9259 13.9256C20.0545 13.6894 20.0031 13.4138 19.7974 13.2563L17.6502 11.5237ZM10 14.4375C7.87852 14.4375 6.14277 12.6656 6.14277 10.5C6.14277 8.33438 7.87852 6.5625 10 6.5625C12.1215 6.5625 13.8572 8.33438 13.8572 10.5C13.8572 12.6656 12.1215 14.4375 10 14.4375Z" fill="#60719B"></path> </svg> Team Settings </a> <a class="sidebar-links m-b-25 d-flex align-items-center text-decoration-none" href="/v4/private/team_management.jsp"> <svg class="sidebar-svg m-r-15" width="20" height="20" viewBox="0 0 20 20" fill="none" <defs=""> <style>.cls-1 { fill: #fff; opacity: 0; } .cls-2 { fill: #60719B }</style> <title>people</title> <g id="Layer_2" data-name="Layer 2"> <g id="people"> <g id="people-2" data-name="people"> <rect class="cls-1" width="24" height="24" fill="#60719B"></rect> <path class="cls-2" d="M9,11A4,4,0,1,0,5,7,4,4,0,0,0,9,11Z" fill="#60719B"></path> <path class="cls-2" d="M17,13a3,3,0,1,0-3-3A3,3,0,0,0,17,13Z" fill="#60719B"></path> <path class="cls-2" d="M21,20a1,1,0,0,0,1-1,5,5,0,0,0-8.06-3.95A7,7,0,0,0,2,20a1,1,0,0,0,1,1H15a1,1,0,0,0,1-1" fill="#60719B"></path> </g> </g> </g> </svg> Team Management </a> <a class="sidebar-links m-b-30 d-flex align-items-center text-decoration-none" href="/v4/private/stats.jsp"> <svg class="sidebar-svg m-r-15" width="20" height="16" viewBox="0 0 20 16" fill="none" xmlns="http://www.w3.org/2000/svg"> <g class="bars"> <rect fill="#60719B" width="35%" height="25"></rect>; <rect fill="#60719B" width="45%" height="20"></rect> </g> </svg> Statistics </a> <a class="sidebar-links m-b-25 d-flex align-items-center text-decoration-none" href="/v4/private/message_rules.jsp"> <svg class="sidebar-svg m-r-15" width="22" height="21" viewBox="0 0 22 21" fill="none" xmlns="http://www.w3.org/2000/svg"> <path d="M13.0005 13.5001V11.0001H9.00055V14.0001H7.00055V10.0001C7.00055 9.73484 7.1059 9.48049 7.29344 9.29295C7.48098 9.10542 7.73533 9.00006 8.00055 9.00006H13.0005V6.50006L16.5005 10.0001L13.0005 13.5001ZM20.7105 10.2901L11.7105 1.29006H11.7005C11.3105 0.900059 10.6805 0.900059 10.2905 1.29006L1.29055 10.2901C0.900547 10.6801 0.900547 11.3201 1.29055 11.7101L10.2905 20.7101C10.6805 21.0901 11.3105 21.1001 11.7105 20.7101L20.7105 11.7101C21.1005 11.3201 21.1005 10.6801 20.7105 10.2901Z" fill="#60719B"></path> </svg> Message Rules </a> <a class="sidebar-links p-b-25 primary-border-b d-flex align-items-center text-decoration-none" href="/v4/private/quick_start.jsp"> <svg class="sidebar-svg m-r-15" width="16" height="20" viewBox="0 0 16 20" fill="none" xmlns="http://www.w3.org/2000/svg"> <path d="M9 7H14.5L9 1.5V7ZM2 0H10L16 6V18C16 18.5304 15.7893 19.0391 15.4142 19.4142C15.0391 19.7893 14.5304 20 14 20H2C0.89 20 0 19.1 0 18V2C0 0.89 0.89 0 2 0ZM10 16V14H3V16H10ZM13 12V10H3V12H13Z" fill="#60719B"></path> </svg> Quick Start Docs </a> </div> <div class="d-flex flex-column primary-line m-b-50"> <a class="sidebar-links m-b-30 d-flex align-items-center text-decoration-none" href="/v4/public/inboxes.jsp"> <svg class="sidebar-svg m-r-15" width="20" height="16" viewBox="0 0 20 16" fill="none" xmlns="http://www.w3.org/2000/svg"> <path d="M18 0H2C0.9 0 0.00999999 0.9 0.00999999 2L0 14C0 15.1 0.9 16 2 16H18C19.1 16 20 15.1 20 14V2C20 0.9 19.1 0 18 0ZM18 4L10 9L2 4V2L10 7L18 2V4Z" fill="#60719B"></path> </svg> Public Inboxes </a> <div style="text-align: center;font-weight: bold;"> Your Trial Ends in <br> <span style="color:red;"> 13 </span>days. <br> <a href="https://www.mailinator.com/site/pricing/">Subscribe Now!</a> </div> </div> <div class="d-flex flex-column justify-content-end flex-grow-1 primary-line"> <a style="color:#000;" href="https://status.mailinator.com">System Status</a> <br> <a style="color:#000;" href="https://www.mailinator.com/site/terms-of-use/">Terms</a> <br> <a style="color:#000;" href="https://www.mailinator.com/site/privacy-policy/">Privacy</a> <br> <a style="color:#000;" href="https://www.mailinator.com/site/accessibility/">Accessibility</a> <br> <a style="color:#000;" href="https://www.mailinator.com/site/security-details/">Security</a> </div> <div class="p-r-50 p-t-25"> <p style="font-size:.7em;">Copyright 2022 Manybrain, LLC. All Rights Reserved</p> </div> </aside> <!-- dont close the initial div here --> <main class="main-content"> <div class="main-container-fluid"> <div class="quick-start d-flex justify-content-between"> <h4 class="fw-700">Mailinator Quick Start</h4> <a href="https://www.mailinator.com/docs/index.html" class="link d-flex align-items-center text-decoration-none"> <svg class="m-r-10" width="16" height="16" viewBox="0 0 16 16" fill="none" xmlns="http://www.w3.org/2000/svg"> <path d="M9.77778 12.4444H3.55556V10.6667H9.77778V12.4444ZM12.4444 8.88889H3.55556V7.11111H12.4444V8.88889ZM12.4444 5.33333H3.55556V3.55556H12.4444V5.33333ZM14.2222 0H1.77778C0.791111 0 0 0.791111 0 1.77778V14.2222C0 14.6937 0.187301 15.1459 0.520699 15.4793C0.854097 15.8127 1.30628 16 1.77778 16H14.2222C14.6937 16 15.1459 15.8127 15.4793 15.4793C15.8127 15.1459 16 14.6937 16 14.2222V1.77778C16 0.791111 15.2 0 14.2222 0Z" fill="#2979FF"></path> </svg> Go to Full Documentation</a> </div> <div class="description-block"> <div class="item" style="font-size: 15px;"> <div class="m-b-20 fz-20 fw-700">Your Own Private Domain</div> <p class="description"> You now have your own Private Domain! All email in that domain stays Private to you, sticks around until you delete it (or you fill up your storage), and doesn't have rate limits !</p> <p class="description"> Your Private Domain is: <span> rzrmhxbwuiteam.testinator.com </span> </p> <p class="description"> (You can change it on the <a class="text-decoration-none" href="team_settings.jsp"> Team Settings </a> page) </p> <p class="description"> All email sent to email addresses* <b>@rzrmhxbwuiteam.testinator.com </b> will arrive in your Private Team Inbox. <br> </p> <p class="description"> You can try it right now! Send an email to: </p> <p class="description"> <span> test1@rzrmhxbwuiteam.testinator.com</span></p> <p class="description"> Then check your Private Team Inbox. </p> </div> <div class="item" style="font-size: 15px;"> <div class="m-b-20 fz-20 fw-700">API</div> <p class="description"> You also have access to all Mailinator email (public and private) through our API. You'll find your token on the Team Settings page. Use that token in a Authorization Header to authorize your api calls (see the API documentation for more information). </p> <p class="description"> We also have Mailinator API SDKs for many popular programming languages including <a href="https://github.com/manybrain/mailinator-java-client" rel="nofollow">Java</a>, <a href="https://github.com/manybrain/mailinator-csharp-client" rel="nofollow">C#</a>, <a href="https://www.npmjs.com/package/mailinator-client" rel="nofollow">Javascript</a>, <a href="https://github.com/manybrain/mailinator-ruby-client" rel="nofollow">Ruby</a>, <a href="https://github.com/manybrain/mailinator-go-client" rel="nofollow">Go</a>, and <a href="https://github.com/manybrain/mailinator-python-client" rel="nofollow">Python</a>. </p> <p class="description"> You can of course use the API to read your Private Domain email too. Consult the <a class="text-decoration-none" href="https://www.mailinator.com/docs/index.html"> API documentation </a> for specifics! </p> <p class="description">Try it now - click this link to test the API:</p> <p class="description"> <span> <a href="https://www.mailinator.com/api/v2/domains/public/inboxes/test"> www.mailinator.com/api/v2/domains/public/inboxes/test </a> </span> </p> <p class="fz-18">(Note JSON is often not well-formatted in web browsers but you get the idea).</p> </div> <div class="item" style="font-size: 15px;"> <div class="m-b-20 fz-20 fw-700">Message Routing Rules</div> <p class="description">Mailinator Message Rules allow you to tell the system to act on messages the moment they arrive.</p> <p class="description">For example, you could create rules like:</p> <div> <p class="rules m-b-10"> <span style="font-family: 'Courier New';color:#000;background-color: #fff;"> IF inbox == bob THEN <br> &nbsp; &nbsp; &nbsp; &nbsp; webhook: yourwebsite.com/rest/webhook </span> </p> <p class="rules m-b-10"> <span style="font-family: 'Courier New';color:#000;background-color: #fff;"> IF inbox == joe THEN CLICK-ALL-LINKS </span> </p> </div> <p class="description">The first rule will "push" emails, as they arrive, to a REST endpoint you setup. The second rule will automatically click all links in any email that arrives in your Private Domain's "joe" inbox.</p> <p class="description">Check out the "Message Rules" page to set them up and Documentation for more info.</p> </div> <div class="item" style="font-size: 15px;"> <div class="m-b-20 fz-20 fw-700">Team Invites</div> <p class="description">If you're a Team Admin, you can invite your co-workers into your Mailinator account. See the Team Management page for details.</p> <p class="description">We hope you enjoy Mailinator. If you have any questions, email us at <b>support@manybrain.com</b> </p> </div> </div> </div> <div style="margin-left:30px;font-size: .8em"><p> * A small number of inbox names are restricted when using a Private Subdomain (i.e. xyz.testinator.com) for security reasons. Don't fret however, you still have Trillions to choose from. <br>There are no restrictions on inbox names for Private Domains where you own the domain.</p> </div> </main> </div> <footer class="footer hide"> <div class="footer-container d-flex justify-content-between align-items-center"> <div class="wrapper-nav-footer d-flex ff-futura-book"> <div> <ul class="d-flex"> <li class="item"><a href="/" aria-label="Home"> Home </a> </li> <li class="item"><a href="/v4/public/inboxes.jsp" aria-label="Email"> Email </a> </li> <li class="item"> <div class="dropdown d-flex"> <a href="/site/pricing/" class="modal-item dropdown-toggle d-flex align-items-center" type="" id="dropdownFooter" data-toggle="dropdown" aria-haspopup="true" aria-expanded="false"> Pricing </a> </div> </li> </ul> </div> </div> <div> <ul class="d-flex"> <li class="item"> <a href="https://www.mailinator.com/site/mailinator-faqs/" aria-label="FAQ"> FAQ </a> </li> </ul> </div> <div> <ul class="d-flex"> <li class="item"> <a href="https://www.mailinator.com/site/terms-of-use/" aria-label="Terms"> Terms </a> </li> <li class="item"> <a href="https://www.mailinator.com/site/privacy-policy/" aria-label="Privacy Policy"> Privacy </a> </li> </ul> </div> </div> <div class="wrapper-copyright"> Copyright 2022 Manybrain, Inc. All Rights Reserved </div> </footer> <!-- unneeded ? --> <script src="../../v4/js/bundle.js"></script> </body></html> ### How can we reproduce the issue? ```shell Goto Mailnator and login its the first page after login. ``` ### Relevant log output ```shell No ``` ### Operating System Windows 10 ### Selenium version 4.0.1 ### What are the browser(s) and version(s) where you see this issue? Chrome 98.0.4758.82 ### What are the browser driver(s) and version(s) where you see this issue? Chrome Driver 98.0.4758.82 ### Are you using Selenium Grid? No
non_process
closed what happened if i try to locate the the element rzrmhxbwuiteam testinator com in the below html code i get error openqa selenium staleelementreferenceexception stale element reference element is not attached to the page document session info chrome but my code is right with inspector i can locate the element by search with div div p span and how you see there is no iframe in the code so why i cant get locate the element i think there is something with that site that block it i realy need to get acces to this part of the html and need a solution maybe there is something in selenium i dont know to trick it i use chromedriver as follow program options addarguments disable blink features automationcontrolled program new chromedriver program options this is the full html code of the site mailinator a color darkblue home accessibility terms privacy policy copyright manybrain llc all rights reserved logout go to public mailinator email pricing documentation faq it ce de svg width height viewbox fill none xmlns logout svg class sidebar svg m r width height viewbox fill none xmlns private team inbox svg class sidebar svg m r width height viewbox fill none xmlns team settings cls fill fff opacity cls fill people team management svg class sidebar svg m r width height viewbox fill none xmlns statistics svg class sidebar svg m r width height viewbox fill none xmlns message rules svg class sidebar svg m r width height viewbox fill none xmlns quick start docs svg class sidebar svg m r width height viewbox fill none xmlns public inboxes your trial ends in days a style color href a style color href a style color href a style color href copyright manybrain llc all rights reserved mailinator quick start svg class m r width height viewbox fill none xmlns go to full documentation your own private domain you now have your own private domain all email in that domain stays private to you sticks around until you delete it or you fill up your storage and doesn t have rate limits your private domain is rzrmhxbwuiteam testinator com you can change it on the team settings page all email sent to email addresses rzrmhxbwuiteam testinator com will arrive in your private team inbox you can try it right now send an email to rzrmhxbwuiteam testinator com then check your private team inbox api you also have access to all mailinator email public and private through our api you ll find your token on the team settings page use that token in a authorization header to authorize your api calls see the api documentation for more information we also have mailinator api sdks for many popular programming languages including java c javascript ruby go and python you can of course use the api to read your private domain email too consult the a class text decoration none href api documentation for specifics try it now click this link to test the api a href note json is often not well formatted in web browsers but you get the idea message routing rules mailinator message rules allow you to tell the system to act on messages the moment they arrive for example you could create rules like if inbox bob then nbsp nbsp nbsp nbsp webhook yourwebsite com rest webhook if inbox joe then click all links the first rule will push emails as they arrive to a rest endpoint you setup the second rule will automatically click all links in any email that arrives in your private domain s joe inbox check out the message rules page to set them up and documentation for more info team invites if you re a team admin you can invite your co workers into your mailinator account see the team management page for details we hope you enjoy mailinator if you have any questions email us at support manybrain com a small number of inbox names are restricted when using a private subdomain i e xyz testinator com for security reasons don t fret however you still have trillions to choose from there are no restrictions on inbox names for private domains where you own the domain home email pricing faq terms privacy copyright manybrain inc all rights reserved how can we reproduce the issue shell goto mailnator and login its the first page after login relevant log output shell no operating system windows selenium version what are the browser s and version s where you see this issue chrome what are the browser driver s and version s where you see this issue chrome driver are you using selenium grid no
0
16,264
4,031,161,159
IssuesEvent
2016-05-18 16:16:26
flutter/flutter
https://api.github.com/repos/flutter/flutter
closed
Update FAQ re: interop
⚠ clear documentation
Current answer: https://flutter.io/faq/#can-i-interop-with-my-mobile-platforms-default-programming-language We might want to mention something about the message pipe and FlutterView, at least so developers know that we're working on options other than just Mojo.
1.0
Update FAQ re: interop - Current answer: https://flutter.io/faq/#can-i-interop-with-my-mobile-platforms-default-programming-language We might want to mention something about the message pipe and FlutterView, at least so developers know that we're working on options other than just Mojo.
non_process
update faq re interop current answer we might want to mention something about the message pipe and flutterview at least so developers know that we re working on options other than just mojo
0
5,095
7,878,848,329
IssuesEvent
2018-06-26 11:37:38
qgis/QGIS-Documentation
https://api.github.com/repos/qgis/QGIS-Documentation
opened
Improve algorithm parameters description using tables
Guidelines Processing help enhancement
## Description This is a place holder to discuss the possible use of tables to describe the processing algorithms parameters instead of headings and bold syntax. As discussed in Madeira Hackfest, tables are hard to maintain in Sphinx. Meanwhile, while discussing #2738, I found some nice directives that may make it easier: https://rest-sphinx-memo.readthedocs.io/en/latest/ReST.html#csv-table I will try to come up with a structure for using them with the parameters. And see if it works. ## Checklist > This is the issue/bug trackers for [QGIS Documentation](https://docs.qgis.org). Cleaning the queue is a process done by project maintainers, mostly on a volunteer basis. We try to keep the overhead as small as possible and appreciate if you help us to do so by completing the following items. - [x] I'm aware this repository is about QGIS Documentation and the issue I'm reporting is related to its usage. If it's related to QGIS application or website, please refer to http://qgis.org/en/site/getinvolved/development/bugreporting.html for the right issue tracker or to the [support channels](http://qgis.org/en/site/forusers/support.html) - [ ] I have added a link to the page concerned by the issue
1.0
Improve algorithm parameters description using tables - ## Description This is a place holder to discuss the possible use of tables to describe the processing algorithms parameters instead of headings and bold syntax. As discussed in Madeira Hackfest, tables are hard to maintain in Sphinx. Meanwhile, while discussing #2738, I found some nice directives that may make it easier: https://rest-sphinx-memo.readthedocs.io/en/latest/ReST.html#csv-table I will try to come up with a structure for using them with the parameters. And see if it works. ## Checklist > This is the issue/bug trackers for [QGIS Documentation](https://docs.qgis.org). Cleaning the queue is a process done by project maintainers, mostly on a volunteer basis. We try to keep the overhead as small as possible and appreciate if you help us to do so by completing the following items. - [x] I'm aware this repository is about QGIS Documentation and the issue I'm reporting is related to its usage. If it's related to QGIS application or website, please refer to http://qgis.org/en/site/getinvolved/development/bugreporting.html for the right issue tracker or to the [support channels](http://qgis.org/en/site/forusers/support.html) - [ ] I have added a link to the page concerned by the issue
process
improve algorithm parameters description using tables description this is a place holder to discuss the possible use of tables to describe the processing algorithms parameters instead of headings and bold syntax as discussed in madeira hackfest tables are hard to maintain in sphinx meanwhile while discussing i found some nice directives that may make it easier i will try to come up with a structure for using them with the parameters and see if it works checklist this is the issue bug trackers for cleaning the queue is a process done by project maintainers mostly on a volunteer basis we try to keep the overhead as small as possible and appreciate if you help us to do so by completing the following items i m aware this repository is about qgis documentation and the issue i m reporting is related to its usage if it s related to qgis application or website please refer to for the right issue tracker or to the i have added a link to the page concerned by the issue
1
6,070
8,909,023,362
IssuesEvent
2019-01-18 03:49:07
googleapis/google-cloud-node
https://api.github.com/repos/googleapis/google-cloud-node
closed
Automatically generate key.json.enc
type: process
It would be great have a tool that automatically creates the key.json.enc file needed for system and samples tests for a repo run on CircleCI so that the file is always generated consistently, reliably, and securely for all repos.
1.0
Automatically generate key.json.enc - It would be great have a tool that automatically creates the key.json.enc file needed for system and samples tests for a repo run on CircleCI so that the file is always generated consistently, reliably, and securely for all repos.
process
automatically generate key json enc it would be great have a tool that automatically creates the key json enc file needed for system and samples tests for a repo run on circleci so that the file is always generated consistently reliably and securely for all repos
1
11,343
14,165,813,601
IssuesEvent
2020-11-12 07:55:46
tikv/tikv
https://api.github.com/repos/tikv/tikv
closed
cases::test_coprocessor::test_deadline_2 not work
priority/low severity/Minor sig/coprocessor type/bug
## Bug Report <!-- Thanks for your bug report! Don't worry if you can't fill out all the sections. --> ### What version of TiKV are you using? master ### Steps to reproduce - add `println!("list {:?}", fail::list());` at `fail_point!("rockskv_async_snapshot",` - `cargo test --package tests --test failpoints -- cases::test_coprocessor::test_deadline_2 --exact --nocapture` ### What did you expect? fail::list prints failpoint configuration
1.0
cases::test_coprocessor::test_deadline_2 not work - ## Bug Report <!-- Thanks for your bug report! Don't worry if you can't fill out all the sections. --> ### What version of TiKV are you using? master ### Steps to reproduce - add `println!("list {:?}", fail::list());` at `fail_point!("rockskv_async_snapshot",` - `cargo test --package tests --test failpoints -- cases::test_coprocessor::test_deadline_2 --exact --nocapture` ### What did you expect? fail::list prints failpoint configuration
process
cases test coprocessor test deadline not work bug report what version of tikv are you using master steps to reproduce add println list fail list at fail point rockskv async snapshot cargo test package tests test failpoints cases test coprocessor test deadline exact nocapture what did you expect fail list prints failpoint configuration
1
4,790
2,754,264,374
IssuesEvent
2015-04-25 14:11:05
statsmodels/statsmodels
https://api.github.com/repos/statsmodels/statsmodels
opened
Design: modifiers for loglike, score, hessian
comp-base design type-enh
Specific case: Where and how do we add penalization terms to the loglike, score and hessian? Another case that is already in use is calling a `transform_params` to separate parameters to be optimized from internal representation. `transform_params` are currently mainly used for internal reparameterization (inequality constraints with interior solution) and yet not much for imposing equality constraints. The former are only relevant in optimization, the latter are also relevant for inference and other results. The general task is that we need a wrapper that modifies the loglike and others on entry and exit. It would be a typical case for a decorator, except we don't want to hardcode a decorator in there. The main purpose is for optimization with the standard scipy optimizers, but we need it also for inference, e.g. calculating the sandwiches for cov_params. Optimization that uses other optimization methods can add a new fit method, but that still leaves robust cov_params as open question. possible designs - use subclasses and Mixin, e.g. ` class PoissonPenalized(PenalizedMixin, Poisson)` - (+) doesn't require changes to existing models - (-) needs a new class for every existing class - add optimizer options - current discrete fit_regularized adds separate optimizer that includes penalty, specific return classes - weaker option: outsource optimizers (PR) and allow overwriting functions/methods that are used in optimization, e.g. the lambda functions used in the optimizers based on loglike, score and hessian - (-) does not provide access to penalties/reparameterization for inference. - build it into all models - check in each method, loglike, score, hessian, whether a "modifier" should be applied, similar to some of the current `trans_params` - call overhead: just one additional `if` check in the methods - needs a way to define the option to use it. current internal flags is not "nice" and doesn't generalize. Use keyword options instead. For that we need `**kwds` to be used by optimizers. - (-) needs to replicate code in every method - possible: refactor current loglike to leading underscore and add new loglike as wrapper. (similar to scipy.stats.distributions) - combinations of the above. other possibilities - "prefix" class with delegation, `mod_m = PenalizedMEstimator(modpr, smpen.SCADSmoothed(0.1, c0=0.0001))` - problem It's difficult to add all the necessary methods of the model class to the the prefix class
1.0
Design: modifiers for loglike, score, hessian - Specific case: Where and how do we add penalization terms to the loglike, score and hessian? Another case that is already in use is calling a `transform_params` to separate parameters to be optimized from internal representation. `transform_params` are currently mainly used for internal reparameterization (inequality constraints with interior solution) and yet not much for imposing equality constraints. The former are only relevant in optimization, the latter are also relevant for inference and other results. The general task is that we need a wrapper that modifies the loglike and others on entry and exit. It would be a typical case for a decorator, except we don't want to hardcode a decorator in there. The main purpose is for optimization with the standard scipy optimizers, but we need it also for inference, e.g. calculating the sandwiches for cov_params. Optimization that uses other optimization methods can add a new fit method, but that still leaves robust cov_params as open question. possible designs - use subclasses and Mixin, e.g. ` class PoissonPenalized(PenalizedMixin, Poisson)` - (+) doesn't require changes to existing models - (-) needs a new class for every existing class - add optimizer options - current discrete fit_regularized adds separate optimizer that includes penalty, specific return classes - weaker option: outsource optimizers (PR) and allow overwriting functions/methods that are used in optimization, e.g. the lambda functions used in the optimizers based on loglike, score and hessian - (-) does not provide access to penalties/reparameterization for inference. - build it into all models - check in each method, loglike, score, hessian, whether a "modifier" should be applied, similar to some of the current `trans_params` - call overhead: just one additional `if` check in the methods - needs a way to define the option to use it. current internal flags is not "nice" and doesn't generalize. Use keyword options instead. For that we need `**kwds` to be used by optimizers. - (-) needs to replicate code in every method - possible: refactor current loglike to leading underscore and add new loglike as wrapper. (similar to scipy.stats.distributions) - combinations of the above. other possibilities - "prefix" class with delegation, `mod_m = PenalizedMEstimator(modpr, smpen.SCADSmoothed(0.1, c0=0.0001))` - problem It's difficult to add all the necessary methods of the model class to the the prefix class
non_process
design modifiers for loglike score hessian specific case where and how do we add penalization terms to the loglike score and hessian another case that is already in use is calling a transform params to separate parameters to be optimized from internal representation transform params are currently mainly used for internal reparameterization inequality constraints with interior solution and yet not much for imposing equality constraints the former are only relevant in optimization the latter are also relevant for inference and other results the general task is that we need a wrapper that modifies the loglike and others on entry and exit it would be a typical case for a decorator except we don t want to hardcode a decorator in there the main purpose is for optimization with the standard scipy optimizers but we need it also for inference e g calculating the sandwiches for cov params optimization that uses other optimization methods can add a new fit method but that still leaves robust cov params as open question possible designs use subclasses and mixin e g class poissonpenalized penalizedmixin poisson doesn t require changes to existing models needs a new class for every existing class add optimizer options current discrete fit regularized adds separate optimizer that includes penalty specific return classes weaker option outsource optimizers pr and allow overwriting functions methods that are used in optimization e g the lambda functions used in the optimizers based on loglike score and hessian does not provide access to penalties reparameterization for inference build it into all models check in each method loglike score hessian whether a modifier should be applied similar to some of the current trans params call overhead just one additional if check in the methods needs a way to define the option to use it current internal flags is not nice and doesn t generalize use keyword options instead for that we need kwds to be used by optimizers needs to replicate code in every method possible refactor current loglike to leading underscore and add new loglike as wrapper similar to scipy stats distributions combinations of the above other possibilities prefix class with delegation mod m penalizedmestimator modpr smpen scadsmoothed problem it s difficult to add all the necessary methods of the model class to the the prefix class
0
11,822
14,644,782,000
IssuesEvent
2020-12-26 02:29:13
lishu/vscode-svg2
https://api.github.com/repos/lishu/vscode-svg2
closed
SVG preview doesn't update from CSS style settings (only after closing/reopening the editor)
In process
**Issue:** _When specifying a css stylesheet, changes to that stylesheet don't get reflected in the preview._ Repro steps: 1. Create a svg file with the following code: ``` <svg> <defs> <link rel="stylesheet" href="styles.css" /> </defs> <rect class="testrect" x="0" y="0" width="100" height="100%" /> </svg> ``` 2. Create css file with the following code ``` .testrect { fill: red } ``` 3. Make sure you select the css file in the preview panel ![image](https://user-images.githubusercontent.com/15620434/103142973-85610c00-46ca-11eb-9231-57e638b696ec.png) 4. Note that the style will be reflected in the preview 5. Change the fill color in the css file to blue. Note that the preview doesn't change 6. Edit something in the .svg file. Note that the preview still don't change. 7. Close and reopen the preview. Note that now the style change is reflected. Obs: Sometimes 6 works on the first time, but not on the second time.
1.0
SVG preview doesn't update from CSS style settings (only after closing/reopening the editor) - **Issue:** _When specifying a css stylesheet, changes to that stylesheet don't get reflected in the preview._ Repro steps: 1. Create a svg file with the following code: ``` <svg> <defs> <link rel="stylesheet" href="styles.css" /> </defs> <rect class="testrect" x="0" y="0" width="100" height="100%" /> </svg> ``` 2. Create css file with the following code ``` .testrect { fill: red } ``` 3. Make sure you select the css file in the preview panel ![image](https://user-images.githubusercontent.com/15620434/103142973-85610c00-46ca-11eb-9231-57e638b696ec.png) 4. Note that the style will be reflected in the preview 5. Change the fill color in the css file to blue. Note that the preview doesn't change 6. Edit something in the .svg file. Note that the preview still don't change. 7. Close and reopen the preview. Note that now the style change is reflected. Obs: Sometimes 6 works on the first time, but not on the second time.
process
svg preview doesn t update from css style settings only after closing reopening the editor issue when specifying a css stylesheet changes to that stylesheet don t get reflected in the preview repro steps create a svg file with the following code create css file with the following code testrect fill red make sure you select the css file in the preview panel note that the style will be reflected in the preview change the fill color in the css file to blue note that the preview doesn t change edit something in the svg file note that the preview still don t change close and reopen the preview note that now the style change is reflected obs sometimes works on the first time but not on the second time
1
19,276
25,464,404,315
IssuesEvent
2022-11-25 01:29:59
openxla/stablehlo
https://api.github.com/repos/openxla/stablehlo
opened
Link Verification and Type Inference tickets in the spec
Process
So far, whenever we encountered misalignments between the spec and verifiers / type inference implemented in the StableHLO dialect, we opened tickets and marked Verification and Type Inference columns in status.md as "revisit". This is useful, but in practice this loses some important information. When seeing those revisits in status.md, I'm having a hard time remembering what was the reason for them. Similarly, when reading the spec, I sometimes struggle to remember the status of the StableHLO dialect with respect to it. As discussed earlier, we should: 1. Go through all Verification and Type Inference tickets. 2. Identify the tickets which describe misalignments with the spec (not new features, but bugs!). 3. Mention these tickets in specs, right next to relevant parts of Inputs/Outputs/Constraints.
1.0
Link Verification and Type Inference tickets in the spec - So far, whenever we encountered misalignments between the spec and verifiers / type inference implemented in the StableHLO dialect, we opened tickets and marked Verification and Type Inference columns in status.md as "revisit". This is useful, but in practice this loses some important information. When seeing those revisits in status.md, I'm having a hard time remembering what was the reason for them. Similarly, when reading the spec, I sometimes struggle to remember the status of the StableHLO dialect with respect to it. As discussed earlier, we should: 1. Go through all Verification and Type Inference tickets. 2. Identify the tickets which describe misalignments with the spec (not new features, but bugs!). 3. Mention these tickets in specs, right next to relevant parts of Inputs/Outputs/Constraints.
process
link verification and type inference tickets in the spec so far whenever we encountered misalignments between the spec and verifiers type inference implemented in the stablehlo dialect we opened tickets and marked verification and type inference columns in status md as revisit this is useful but in practice this loses some important information when seeing those revisits in status md i m having a hard time remembering what was the reason for them similarly when reading the spec i sometimes struggle to remember the status of the stablehlo dialect with respect to it as discussed earlier we should go through all verification and type inference tickets identify the tickets which describe misalignments with the spec not new features but bugs mention these tickets in specs right next to relevant parts of inputs outputs constraints
1
4,112
7,057,873,199
IssuesEvent
2018-01-04 18:03:12
pwittchen/swipe
https://api.github.com/repos/pwittchen/swipe
closed
Release 0.2.0
release process
**Initial release notes**: - migrated library to RxJava2.x as a separate artifact on a separate Git branch - removed `master` branch from the repo - updated project dependencies - updated Gradle to 3.x - added Retrolambda to sample Java app **Things to do**: - [x] `RxJava1.x` branch - [x] bump library version to 0.2.0 - [x] upload Archives to Maven Central Repository - [x] close and release artifact on Nexus - [x] update JavaDoc on gh-pages - [x] update `CHANGELOG.md` after Maven Sync - [x] update download section in `README.md` after Maven Sync - [x] create new GitHub release - [x] `RxJava2.x` branch - [x] bump library version to 0.2.0 - [x] upload Archives to Maven Central Repository - [x] close and release artifact on Nexus - [x] update JavaDoc on gh-pages - [x] update `CHANGELOG.md` after Maven Sync - [x] update download section in `README.md` after Maven Sync - [x] create new GitHub release
1.0
Release 0.2.0 - **Initial release notes**: - migrated library to RxJava2.x as a separate artifact on a separate Git branch - removed `master` branch from the repo - updated project dependencies - updated Gradle to 3.x - added Retrolambda to sample Java app **Things to do**: - [x] `RxJava1.x` branch - [x] bump library version to 0.2.0 - [x] upload Archives to Maven Central Repository - [x] close and release artifact on Nexus - [x] update JavaDoc on gh-pages - [x] update `CHANGELOG.md` after Maven Sync - [x] update download section in `README.md` after Maven Sync - [x] create new GitHub release - [x] `RxJava2.x` branch - [x] bump library version to 0.2.0 - [x] upload Archives to Maven Central Repository - [x] close and release artifact on Nexus - [x] update JavaDoc on gh-pages - [x] update `CHANGELOG.md` after Maven Sync - [x] update download section in `README.md` after Maven Sync - [x] create new GitHub release
process
release initial release notes migrated library to x as a separate artifact on a separate git branch removed master branch from the repo updated project dependencies updated gradle to x added retrolambda to sample java app things to do x branch bump library version to upload archives to maven central repository close and release artifact on nexus update javadoc on gh pages update changelog md after maven sync update download section in readme md after maven sync create new github release x branch bump library version to upload archives to maven central repository close and release artifact on nexus update javadoc on gh pages update changelog md after maven sync update download section in readme md after maven sync create new github release
1
14,717
17,926,858,416
IssuesEvent
2021-09-10 01:32:22
Leviatan-Analytics/LA-data-processing
https://api.github.com/repos/Leviatan-Analytics/LA-data-processing
closed
Implement filter text recognition output to get only players info [2]
Data Processing Week 1 Sprint 4
Use the researched techniques to implement the filter of only players position from the output.
1.0
Implement filter text recognition output to get only players info [2] - Use the researched techniques to implement the filter of only players position from the output.
process
implement filter text recognition output to get only players info use the researched techniques to implement the filter of only players position from the output
1
19,808
26,195,774,609
IssuesEvent
2023-01-03 13:19:41
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
"strict transitive dependency mode" not work ?
P4 type: support / not a bug (process) team-Rules-CPP
### Description of the bug: Suppose there is the following dependency a->b->c, **a contains the header file of c, but does not use anything**, then a's build file depends on b, but not on c. Then this forms an **implicit dependency,** and bazel will not report a error! ### What's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible. _No response_ ### Which operating system are you running Bazel on? ubuntu 18.04 ### What is the output of `bazel info release`? (11:37:20) INFO: Invocation ID: b93b0d49-edac-4215-9d9a-366bf997e477 release 3.7.1 ### If `bazel info release` returns `development version` or `(@non-git)`, tell us how you built Bazel. _No response_ ### What's the output of `git remote get-url origin; git rev-parse master; git rev-parse HEAD` ? _No response_ ### Have you found anything relevant by searching the web? _No response_ ### Any other information, logs, or outputs that you want to share? _No response_
1.0
"strict transitive dependency mode" not work ? - ### Description of the bug: Suppose there is the following dependency a->b->c, **a contains the header file of c, but does not use anything**, then a's build file depends on b, but not on c. Then this forms an **implicit dependency,** and bazel will not report a error! ### What's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible. _No response_ ### Which operating system are you running Bazel on? ubuntu 18.04 ### What is the output of `bazel info release`? (11:37:20) INFO: Invocation ID: b93b0d49-edac-4215-9d9a-366bf997e477 release 3.7.1 ### If `bazel info release` returns `development version` or `(@non-git)`, tell us how you built Bazel. _No response_ ### What's the output of `git remote get-url origin; git rev-parse master; git rev-parse HEAD` ? _No response_ ### Have you found anything relevant by searching the web? _No response_ ### Any other information, logs, or outputs that you want to share? _No response_
process
strict transitive dependency mode not work description of the bug suppose there is the following dependency a b c a contains the header file of c but does not use anything then a s build file depends on b but not on c then this forms an implicit dependency and bazel will not report a error what s the simplest easiest way to reproduce this bug please provide a minimal example if possible no response which operating system are you running bazel on ubuntu what is the output of bazel info release info invocation id edac release if bazel info release returns development version or non git tell us how you built bazel no response what s the output of git remote get url origin git rev parse master git rev parse head no response have you found anything relevant by searching the web no response any other information logs or outputs that you want to share no response
1
20,277
26,909,094,107
IssuesEvent
2023-02-06 21:46:10
metabase/metabase
https://api.github.com/repos/metabase/metabase
closed
MongoDB support Custom Column and Custom Expression
Database/Mongo Querying/Processor Type:New Feature .Completeness Querying/Notebook/Custom Column Querying/Notebook/Custom Expression
### Feature requests and proposals It seems that MongoDB "Custom" questions do not offer support for "Custom Expressions". Here's what I see when I use my PostgreSQL database: ![screenshot from 2018-03-08 18-52-51](https://user-images.githubusercontent.com/3029017/37178713-12e05648-2302-11e8-90c8-1e6740297cbd.png) And these are the options for MongoDB: ![screenshot from 2018-03-08 18-53-14](https://user-images.githubusercontent.com/3029017/37178708-0f977ffc-2302-11e8-8d93-993b96f5dce7.png) Since MongoDB's aggregation pipeline is essentially a JavaScript function, I believe it should be possible to include custom expressions for this database as well. Maybe even allowing arbitrary expressions instead of just `+`, `-`, `*`, `/`. :arrow_down: Please click the :+1: reaction instead of leaving a `+1` or `update?` comment
1.0
MongoDB support Custom Column and Custom Expression - ### Feature requests and proposals It seems that MongoDB "Custom" questions do not offer support for "Custom Expressions". Here's what I see when I use my PostgreSQL database: ![screenshot from 2018-03-08 18-52-51](https://user-images.githubusercontent.com/3029017/37178713-12e05648-2302-11e8-90c8-1e6740297cbd.png) And these are the options for MongoDB: ![screenshot from 2018-03-08 18-53-14](https://user-images.githubusercontent.com/3029017/37178708-0f977ffc-2302-11e8-8d93-993b96f5dce7.png) Since MongoDB's aggregation pipeline is essentially a JavaScript function, I believe it should be possible to include custom expressions for this database as well. Maybe even allowing arbitrary expressions instead of just `+`, `-`, `*`, `/`. :arrow_down: Please click the :+1: reaction instead of leaving a `+1` or `update?` comment
process
mongodb support custom column and custom expression feature requests and proposals it seems that mongodb custom questions do not offer support for custom expressions here s what i see when i use my postgresql database and these are the options for mongodb since mongodb s aggregation pipeline is essentially a javascript function i believe it should be possible to include custom expressions for this database as well maybe even allowing arbitrary expressions instead of just arrow down please click the reaction instead of leaving a or update comment
1
15,745
19,910,886,557
IssuesEvent
2022-01-25 17:02:41
googleapis/java-storage-transfer
https://api.github.com/repos/googleapis/java-storage-transfer
closed
Your .repo-metadata.json file has a problem 🤒
type: process api: storagetransfer repo-metadata: lint
You have a problem with your .repo-metadata.json file: Result of scan 📈: * api_shortname 'storage-transfer' invalid in .repo-metadata.json ☝️ Once you address these problems, you can close this issue. ### Need help? * [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field. * [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**. * Reach out to **go/github-automation** if you have any questions.
1.0
Your .repo-metadata.json file has a problem 🤒 - You have a problem with your .repo-metadata.json file: Result of scan 📈: * api_shortname 'storage-transfer' invalid in .repo-metadata.json ☝️ Once you address these problems, you can close this issue. ### Need help? * [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field. * [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**. * Reach out to **go/github-automation** if you have any questions.
process
your repo metadata json file has a problem 🤒 you have a problem with your repo metadata json file result of scan 📈 api shortname storage transfer invalid in repo metadata json ☝️ once you address these problems you can close this issue need help lists valid options for each field for grpc libraries api shortname should match the subdomain of an api s hostname reach out to go github automation if you have any questions
1
10,010
13,043,869,717
IssuesEvent
2020-07-29 02:54:15
tikv/tikv
https://api.github.com/repos/tikv/tikv
closed
UCP: Migrate scalar function `Ord` from TiDB
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
## Description Port the scalar function `Ord` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @iosmanthus ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
2.0
UCP: Migrate scalar function `Ord` from TiDB - ## Description Port the scalar function `Ord` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @iosmanthus ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
process
ucp migrate scalar function ord from tidb description port the scalar function ord from tidb to coprocessor score mentor s iosmanthus recommended skills rust programming learning materials already implemented expressions ported from tidb
1
18,276
24,354,462,019
IssuesEvent
2022-10-03 05:44:48
prometheus-community/windows_exporter
https://api.github.com/repos/prometheus-community/windows_exporter
closed
Failed to get process level information
collector/process
Hi, I want to track the CPU, Memory and total count of processes running on the server. I am using the below command to initiate the windows exporter - **c:\PROGRA~1\windows_exporter\windows_exporter.exe --collector.process.whitelist="chrome.+"** But, I am getting only services-level information, nothing related to the process. C:\WINDOWS\system32>"C:\Program Files\windows_exporter\windows_exporter.exe" --log.format logger:eventlog?name=windows_exporter --telemetry.addr :9182 --collectors.enabled "[defaults],memory,process,tcp" --collector.process.whitelist="chrome.+" time="2022-09-27T12:18:45+05:30" level=warning msg="No where-clause specified for service collector. This will generate a very large number of metrics!" source="service.go:48" time="2022-09-27T12:18:45+05:30" level=info msg="Enabled collectors: net, tcp, cpu, os, cs, logical_disk, service, system, textfile, memory, process" source="exporter.go:348" time="2022-09-27T12:18:45+05:30" level=info msg="Starting windows_exporter (version=0.17.1, branch=heads/tags/v0.17.1, revision=d9f4264fc4ee183462fc23032ce9b60444411b87)" source="exporter.go:400" time="2022-09-27T12:18:45+05:30" level=info msg="Build context (go=go1.17.5, user=runneradmin@fv-az177-480, date=20220102-09:24:12)" source="exporter.go:401" time="2022-09-27T12:18:45+05:30" level=info msg="Starting server on :9182" source="exporter.go:404" time="2022-09-27T12:18:46+05:30" level=info msg="TLS is disabled." source="gokit_adapter.go:38"
1.0
Failed to get process level information - Hi, I want to track the CPU, Memory and total count of processes running on the server. I am using the below command to initiate the windows exporter - **c:\PROGRA~1\windows_exporter\windows_exporter.exe --collector.process.whitelist="chrome.+"** But, I am getting only services-level information, nothing related to the process. C:\WINDOWS\system32>"C:\Program Files\windows_exporter\windows_exporter.exe" --log.format logger:eventlog?name=windows_exporter --telemetry.addr :9182 --collectors.enabled "[defaults],memory,process,tcp" --collector.process.whitelist="chrome.+" time="2022-09-27T12:18:45+05:30" level=warning msg="No where-clause specified for service collector. This will generate a very large number of metrics!" source="service.go:48" time="2022-09-27T12:18:45+05:30" level=info msg="Enabled collectors: net, tcp, cpu, os, cs, logical_disk, service, system, textfile, memory, process" source="exporter.go:348" time="2022-09-27T12:18:45+05:30" level=info msg="Starting windows_exporter (version=0.17.1, branch=heads/tags/v0.17.1, revision=d9f4264fc4ee183462fc23032ce9b60444411b87)" source="exporter.go:400" time="2022-09-27T12:18:45+05:30" level=info msg="Build context (go=go1.17.5, user=runneradmin@fv-az177-480, date=20220102-09:24:12)" source="exporter.go:401" time="2022-09-27T12:18:45+05:30" level=info msg="Starting server on :9182" source="exporter.go:404" time="2022-09-27T12:18:46+05:30" level=info msg="TLS is disabled." source="gokit_adapter.go:38"
process
failed to get process level information hi i want to track the cpu memory and total count of processes running on the server i am using the below command to initiate the windows exporter c progra windows exporter windows exporter exe collector process whitelist chrome but i am getting only services level information nothing related to the process c windows c program files windows exporter windows exporter exe log format logger eventlog name windows exporter telemetry addr collectors enabled memory process tcp collector process whitelist chrome time level warning msg no where clause specified for service collector this will generate a very large number of metrics source service go time level info msg enabled collectors net tcp cpu os cs logical disk service system textfile memory process source exporter go time level info msg starting windows exporter version branch heads tags revision source exporter go time level info msg build context go user runneradmin fv date source exporter go time level info msg starting server on source exporter go time level info msg tls is disabled source gokit adapter go
1
113,738
11,813,159,482
IssuesEvent
2020-03-19 21:43:05
carla-simulator/carla
https://api.github.com/repos/carla-simulator/carla
closed
A simple tutorial for Carla
backlog documentation
I am Frank from China and very interested in Carla Simulator. Thanks for Carla team hard work. I want to do some self-driving car simulation in Chinese traffic scenes based on Carla. From my side, a simple tutorial is appreciated for a new user to know: 1, how to set start points 2, how to set goal points 3, how to use map data 4, a simple straight path planning using existing API I spent some time on documents and python client code but still have no idea to integrate my modular pipeline self-driving car algorithm into Carla. These tasks may be very basic but really important for a beginner. This simple tutorial will benefit for a beginner a lot to know how to use Carla soon. Please consider this suggestion. Thanks a lot.
1.0
A simple tutorial for Carla - I am Frank from China and very interested in Carla Simulator. Thanks for Carla team hard work. I want to do some self-driving car simulation in Chinese traffic scenes based on Carla. From my side, a simple tutorial is appreciated for a new user to know: 1, how to set start points 2, how to set goal points 3, how to use map data 4, a simple straight path planning using existing API I spent some time on documents and python client code but still have no idea to integrate my modular pipeline self-driving car algorithm into Carla. These tasks may be very basic but really important for a beginner. This simple tutorial will benefit for a beginner a lot to know how to use Carla soon. Please consider this suggestion. Thanks a lot.
non_process
a simple tutorial for carla i am frank from china and very interested in carla simulator thanks for carla team hard work i want to do some self driving car simulation in chinese traffic scenes based on carla from my side a simple tutorial is appreciated for a new user to know how to set start points how to set goal points how to use map data a simple straight path planning using existing api i spent some time on documents and python client code but still have no idea to integrate my modular pipeline self driving car algorithm into carla these tasks may be very basic but really important for a beginner this simple tutorial will benefit for a beginner a lot to know how to use carla soon please consider this suggestion thanks a lot
0
22,385
31,142,284,751
IssuesEvent
2023-08-16 01:44:09
cypress-io/cypress
https://api.github.com/repos/cypress-io/cypress
closed
Flaky test: We encountered an unexpected error talking to our servers. StatusCodeError: 409 - "error code: 1023"
process: flaky test topic: flake ❄️ stage: flake stale
### Link to dashboard or CircleCI failure https://app.circleci.com/pipelines/github/cypress-io/cypress/41870/workflows/432e2428-682e-4fe8-9b60-bb05e0cda202/jobs/1736315 ### Link to failing test in GitHub Not specific to a failing test ### Analysis We're seeing [status code 409](https://developer.mozilla.org/en-US/docs/Web/HTTP/Status/409) here: > The HTTP 409 Conflict response status code indicates a request conflict with the current state of the target resource. Conflicts are most likely to occur in response to a [PUT](https://developer.mozilla.org/en-US/docs/Web/HTTP/Methods/PUT) request. For example, you may get a 409 response when uploading a file that is older than the existing one on the server, resulting in a version control conflict. I wonder if there's some config we can set to avoid this <img width="1128" alt="Screen Shot 2022-08-17 at 5 56 15 PM" src="https://user-images.githubusercontent.com/26726429/185268932-c578d2e5-4768-48dc-b3a2-226a65318c85.png"> I'm not sure if this error is coming from CloudFlare, but if it is, [here's their definition of the 1023 error code](https://support.cloudflare.com/hc/en-us/articles/360029779472-Troubleshooting-Cloudflare-1XXX-errors#error1023): <img width="1174" alt="Screen Shot 2022-08-17 at 5 59 55 PM" src="https://user-images.githubusercontent.com/26726429/185269271-308df975-b43c-4865-97c7-8b7ad14ed9dc.png"> ### Cypress Version 10.6.0 ### Other _No response_
1.0
Flaky test: We encountered an unexpected error talking to our servers. StatusCodeError: 409 - "error code: 1023" - ### Link to dashboard or CircleCI failure https://app.circleci.com/pipelines/github/cypress-io/cypress/41870/workflows/432e2428-682e-4fe8-9b60-bb05e0cda202/jobs/1736315 ### Link to failing test in GitHub Not specific to a failing test ### Analysis We're seeing [status code 409](https://developer.mozilla.org/en-US/docs/Web/HTTP/Status/409) here: > The HTTP 409 Conflict response status code indicates a request conflict with the current state of the target resource. Conflicts are most likely to occur in response to a [PUT](https://developer.mozilla.org/en-US/docs/Web/HTTP/Methods/PUT) request. For example, you may get a 409 response when uploading a file that is older than the existing one on the server, resulting in a version control conflict. I wonder if there's some config we can set to avoid this <img width="1128" alt="Screen Shot 2022-08-17 at 5 56 15 PM" src="https://user-images.githubusercontent.com/26726429/185268932-c578d2e5-4768-48dc-b3a2-226a65318c85.png"> I'm not sure if this error is coming from CloudFlare, but if it is, [here's their definition of the 1023 error code](https://support.cloudflare.com/hc/en-us/articles/360029779472-Troubleshooting-Cloudflare-1XXX-errors#error1023): <img width="1174" alt="Screen Shot 2022-08-17 at 5 59 55 PM" src="https://user-images.githubusercontent.com/26726429/185269271-308df975-b43c-4865-97c7-8b7ad14ed9dc.png"> ### Cypress Version 10.6.0 ### Other _No response_
process
flaky test we encountered an unexpected error talking to our servers statuscodeerror error code link to dashboard or circleci failure link to failing test in github not specific to a failing test analysis we re seeing here the http conflict response status code indicates a request conflict with the current state of the target resource conflicts are most likely to occur in response to a request for example you may get a response when uploading a file that is older than the existing one on the server resulting in a version control conflict i wonder if there s some config we can set to avoid this img width alt screen shot at pm src i m not sure if this error is coming from cloudflare but if it is img width alt screen shot at pm src cypress version other no response
1
2,088
4,923,848,700
IssuesEvent
2016-11-25 11:55:19
matz-e/lobster
https://api.github.com/repos/matz-e/lobster
opened
Problems running with CMSSW > 8_0_15
bug processing
I was going to write the stuff from below, and am wondering if the way we create our release setup on the worker is to blame? I'd like to align that with how WMCore does things (basically unpacking the sandbox over a previously created release). > I'm trying to run our simple example again, using `CMSSW_8_0_21`. Tasks start, but `cmsRun` never generates any output. It looks like `cmsRun` just hangs with this output: > > %MSG-i ThreadSetup: (NoModuleName) 25-Nov-2016 06:12:06 EST pre-events > setting # threads 1 > %MSG > > I've bisected this before to happen starting with the release `CMSSW_8_0_16`, see also the > [Changelog for `CMSSW_8_0_16`](https://github.com/cms-sw/cmssw/releases/tag/CMSSW_8_0_16) > > Note that `jemalloc` got upgraded, and was at version `3.x.y` before, IIRC. I have this config patched together, which works on `crcfe01`: > ```shell > dir=$(mktemp -d) > cd $dir > > export HTTP_PROXY="http://eddie.crc.nd.edu:3128" > > cat <<EOB >script > set -x > source /cvmfs/cms.cern.ch/cmsset_default.sh > scramv1 project CMSSW CMSSW_8_0_20 > cd CMSSW_8_0_20/src > eval \$(scramv1 runtime -sh) > cmsDriver.py MinBias_13TeV_pythia8_TuneCUETP8M1_cfi \ > --python_filename sim.py \ > --conditions auto:run2_mc --fast -n 100 \ > --era Run2_2016 --eventcontent FASTPU --relval 100000,1000 \ > -s GEN,SIM,RECOBEFMIX --datatier GEN-SIM-RECO --beamspot Realistic50ns13TeVCollision > cmsRun sim.py > EOB > > # parrot_run -d all bash script > parrot_run bash script > > echo Please remove $dir > ```
1.0
Problems running with CMSSW > 8_0_15 - I was going to write the stuff from below, and am wondering if the way we create our release setup on the worker is to blame? I'd like to align that with how WMCore does things (basically unpacking the sandbox over a previously created release). > I'm trying to run our simple example again, using `CMSSW_8_0_21`. Tasks start, but `cmsRun` never generates any output. It looks like `cmsRun` just hangs with this output: > > %MSG-i ThreadSetup: (NoModuleName) 25-Nov-2016 06:12:06 EST pre-events > setting # threads 1 > %MSG > > I've bisected this before to happen starting with the release `CMSSW_8_0_16`, see also the > [Changelog for `CMSSW_8_0_16`](https://github.com/cms-sw/cmssw/releases/tag/CMSSW_8_0_16) > > Note that `jemalloc` got upgraded, and was at version `3.x.y` before, IIRC. I have this config patched together, which works on `crcfe01`: > ```shell > dir=$(mktemp -d) > cd $dir > > export HTTP_PROXY="http://eddie.crc.nd.edu:3128" > > cat <<EOB >script > set -x > source /cvmfs/cms.cern.ch/cmsset_default.sh > scramv1 project CMSSW CMSSW_8_0_20 > cd CMSSW_8_0_20/src > eval \$(scramv1 runtime -sh) > cmsDriver.py MinBias_13TeV_pythia8_TuneCUETP8M1_cfi \ > --python_filename sim.py \ > --conditions auto:run2_mc --fast -n 100 \ > --era Run2_2016 --eventcontent FASTPU --relval 100000,1000 \ > -s GEN,SIM,RECOBEFMIX --datatier GEN-SIM-RECO --beamspot Realistic50ns13TeVCollision > cmsRun sim.py > EOB > > # parrot_run -d all bash script > parrot_run bash script > > echo Please remove $dir > ```
process
problems running with cmssw i was going to write the stuff from below and am wondering if the way we create our release setup on the worker is to blame i d like to align that with how wmcore does things basically unpacking the sandbox over a previously created release i m trying to run our simple example again using cmssw tasks start but cmsrun never generates any output it looks like cmsrun just hangs with this output msg i threadsetup nomodulename nov est pre events setting threads msg i ve bisected this before to happen starting with the release cmssw see also the note that jemalloc got upgraded and was at version x y before iirc i have this config patched together which works on shell dir mktemp d cd dir export http proxy cat script set x source cvmfs cms cern ch cmsset default sh project cmssw cmssw cd cmssw src eval runtime sh cmsdriver py minbias cfi python filename sim py conditions auto mc fast n era eventcontent fastpu relval s gen sim recobefmix datatier gen sim reco beamspot cmsrun sim py eob parrot run d all bash script parrot run bash script echo please remove dir
1
17,179
22,758,342,206
IssuesEvent
2022-07-07 18:38:29
GoogleCloudPlatform/nodejs-getting-started
https://api.github.com/repos/GoogleCloudPlatform/nodejs-getting-started
closed
run tests in parallel
priority: p2 type: process
Currently tests are executed serially. If we run the tests in parallel, it will speed up the build time.
1.0
run tests in parallel - Currently tests are executed serially. If we run the tests in parallel, it will speed up the build time.
process
run tests in parallel currently tests are executed serially if we run the tests in parallel it will speed up the build time
1
418,172
28,113,836,829
IssuesEvent
2023-03-31 09:14:25
L-K-Chng/ped
https://api.github.com/repos/L-K-Chng/ped
opened
UserGuide misses out important details
type.DocumentationBug severity.Medium
The User Guide does not include FAQ for common questions which the user may have. As such, they may continue to have their doubts when using the product, affecting their overall user experience. In addition, a Command Summary, preferably in table form, should be provided as a quick guide to the possible commands that the user can input so that they can learn to use the program quickly without having to scroll through the entire User Guide for information. ![image.png](https://raw.githubusercontent.com/L-K-Chng/ped/main/files/6e318eb2-2337-4e79-8996-eab80b9d149d.png) <!--session: 1680252522620-8e3cd6e2-3bb8-485a-a535-eb72e2609aa7--> <!--Version: Web v3.4.7-->
1.0
UserGuide misses out important details - The User Guide does not include FAQ for common questions which the user may have. As such, they may continue to have their doubts when using the product, affecting their overall user experience. In addition, a Command Summary, preferably in table form, should be provided as a quick guide to the possible commands that the user can input so that they can learn to use the program quickly without having to scroll through the entire User Guide for information. ![image.png](https://raw.githubusercontent.com/L-K-Chng/ped/main/files/6e318eb2-2337-4e79-8996-eab80b9d149d.png) <!--session: 1680252522620-8e3cd6e2-3bb8-485a-a535-eb72e2609aa7--> <!--Version: Web v3.4.7-->
non_process
userguide misses out important details the user guide does not include faq for common questions which the user may have as such they may continue to have their doubts when using the product affecting their overall user experience in addition a command summary preferably in table form should be provided as a quick guide to the possible commands that the user can input so that they can learn to use the program quickly without having to scroll through the entire user guide for information
0
25,011
7,611,594,535
IssuesEvent
2018-05-01 14:31:50
ORNL-CEES/DataTransferKit
https://api.github.com/repos/ORNL-CEES/DataTransferKit
closed
PGI 17.10
Build Medium Priority Testing
PGI 17.10 was released yesterday. It is supported by Kokkos so we should make it work too.
1.0
PGI 17.10 - PGI 17.10 was released yesterday. It is supported by Kokkos so we should make it work too.
non_process
pgi pgi was released yesterday it is supported by kokkos so we should make it work too
0
8,472
11,642,557,695
IssuesEvent
2020-02-29 07:50:06
Ultimate-Hosts-Blacklist/whitelist
https://api.github.com/repos/Ultimate-Hosts-Blacklist/whitelist
closed
Whitelist doi.org
whitelisting process
*@Somebodyisnobody commented on Jan 31, 2020, 8:33 PM UTC:* doi.org is not a phishing site. It's a Digital Object Identifier service. *This issue was moved by [funilrys](https://github.com/funilrys) from [Ultimate-Hosts-Blacklist/Phishing.Database#1](https://github.com/Ultimate-Hosts-Blacklist/Phishing.Database/issues/1).*
1.0
Whitelist doi.org - *@Somebodyisnobody commented on Jan 31, 2020, 8:33 PM UTC:* doi.org is not a phishing site. It's a Digital Object Identifier service. *This issue was moved by [funilrys](https://github.com/funilrys) from [Ultimate-Hosts-Blacklist/Phishing.Database#1](https://github.com/Ultimate-Hosts-Blacklist/Phishing.Database/issues/1).*
process
whitelist doi org somebodyisnobody commented on jan pm utc doi org is not a phishing site it s a digital object identifier service this issue was moved by from
1
19,962
26,442,024,687
IssuesEvent
2023-01-16 01:55:22
vivianafu/dt-ui
https://api.github.com/repos/vivianafu/dt-ui
closed
Tooltip
processing
Usage: ```tsx <Tooltip label={<TooltipUI />}> {({ close, isOpen }) => <span>Trigger</span>} </Tooltip> or <Tooltip label={<TooltipUI />}> <span>Trigger</span> </Tooltip> ``` Props: TODO
1.0
Tooltip - Usage: ```tsx <Tooltip label={<TooltipUI />}> {({ close, isOpen }) => <span>Trigger</span>} </Tooltip> or <Tooltip label={<TooltipUI />}> <span>Trigger</span> </Tooltip> ``` Props: TODO
process
tooltip usage tsx close isopen trigger or trigger props todo
1
13,242
15,709,742,903
IssuesEvent
2021-03-26 23:18:56
isabel-lombardi/be_cat_photo
https://api.github.com/repos/isabel-lombardi/be_cat_photo
closed
[UPLOAD] save image for classification
processing img upload
saves the images in a dictionary for use by the classification script
1.0
[UPLOAD] save image for classification - saves the images in a dictionary for use by the classification script
process
save image for classification saves the images in a dictionary for use by the classification script
1
21,667
30,111,205,444
IssuesEvent
2023-06-30 07:55:00
0xPolygonMiden/miden-vm
https://api.github.com/repos/0xPolygonMiden/miden-vm
opened
Update range-checker to use log derivative-based multiset checks
processor v0.7
With updates done in #949, we simplified the range checker to use 3 main trace columns and 2 auxiliary trace columns. We should be able to simplify it even more to use just 2 trace columns and 1 auxiliary column. To get rid of 1 auxiliary column we'd need to figure out how to handle the fact that range check requests could come form the stack and memory modules. This is covered by #299. In my mind this is not a high priority issue as total number of aux column is 8 right now, and reducing it down to 7 would have marginal benefit. However, we could get rid of 1 main trace column by using something similar to [Logarithmic derivatives](https://eprint.iacr.org/2022/1530) for multiset checks. The high level idea is that we'd have 2 main trace columns which look as follows: - $v$ - column containing the values to be range-checked. - $n$ - column containing the number this value is to be range-checked. We will then have a running sum column (rather than a running product column) which sums up the following terms for each row: $$ \frac{n}{\alpha - v} $$ Where $\alpha$ is the verifier randomness. I think a similar approach could work on the "other side" of the running sum (the consumer side of the bus) - but this is still something that needs to be thought through.
1.0
Update range-checker to use log derivative-based multiset checks - With updates done in #949, we simplified the range checker to use 3 main trace columns and 2 auxiliary trace columns. We should be able to simplify it even more to use just 2 trace columns and 1 auxiliary column. To get rid of 1 auxiliary column we'd need to figure out how to handle the fact that range check requests could come form the stack and memory modules. This is covered by #299. In my mind this is not a high priority issue as total number of aux column is 8 right now, and reducing it down to 7 would have marginal benefit. However, we could get rid of 1 main trace column by using something similar to [Logarithmic derivatives](https://eprint.iacr.org/2022/1530) for multiset checks. The high level idea is that we'd have 2 main trace columns which look as follows: - $v$ - column containing the values to be range-checked. - $n$ - column containing the number this value is to be range-checked. We will then have a running sum column (rather than a running product column) which sums up the following terms for each row: $$ \frac{n}{\alpha - v} $$ Where $\alpha$ is the verifier randomness. I think a similar approach could work on the "other side" of the running sum (the consumer side of the bus) - but this is still something that needs to be thought through.
process
update range checker to use log derivative based multiset checks with updates done in we simplified the range checker to use main trace columns and auxiliary trace columns we should be able to simplify it even more to use just trace columns and auxiliary column to get rid of auxiliary column we d need to figure out how to handle the fact that range check requests could come form the stack and memory modules this is covered by in my mind this is not a high priority issue as total number of aux column is right now and reducing it down to would have marginal benefit however we could get rid of main trace column by using something similar to for multiset checks the high level idea is that we d have main trace columns which look as follows v column containing the values to be range checked n column containing the number this value is to be range checked we will then have a running sum column rather than a running product column which sums up the following terms for each row frac n alpha v where alpha is the verifier randomness i think a similar approach could work on the other side of the running sum the consumer side of the bus but this is still something that needs to be thought through
1
3,957
6,893,594,034
IssuesEvent
2017-11-23 05:15:12
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
closed
System.Diagnostics.Tests.ProcessThreadTests.TestThreadStateProperty failed in CI
area-System.Diagnostics.Process test-run-core
Failed test: System.Diagnostics.Tests.ProcessThreadTests.TestThreadStateProperty Configuration: osx10.12_debug Detail: https://ci.dot.net/job/dotnet_corefx/job/master/job/osx10.12_debug/3296/testReport/System.Diagnostics.Tests/ProcessThreadTests/TestThreadStateProperty/ MESSAGE: ~~~ System.ArgumentOutOfRangeException : Index was out of range. Must be non-negative and less than the size of the collection.\nParameter name: index ~~~ STACK TRACE: ~~~ at System.Collections.ArrayList.get_Item(Int32 index) at System.Diagnostics.ProcessThreadCollection.get_Item(Int32 index) in /Users/dotnet-bot/j/workspace/dotnet_corefx/master/osx10.12_debug/src/System.Diagnostics.Process/src/System/Diagnostics/ProcessThreadCollection.cs:line 35 at System.Diagnostics.Tests.ProcessThreadTests.TestThreadStateProperty() in /Users/dotnet-bot/j/workspace/dotnet_corefx/master/osx10.12_debug/src/System.Diagnostics.Process/tests/ProcessThreadTests.cs:line 199 ~~~
1.0
System.Diagnostics.Tests.ProcessThreadTests.TestThreadStateProperty failed in CI - Failed test: System.Diagnostics.Tests.ProcessThreadTests.TestThreadStateProperty Configuration: osx10.12_debug Detail: https://ci.dot.net/job/dotnet_corefx/job/master/job/osx10.12_debug/3296/testReport/System.Diagnostics.Tests/ProcessThreadTests/TestThreadStateProperty/ MESSAGE: ~~~ System.ArgumentOutOfRangeException : Index was out of range. Must be non-negative and less than the size of the collection.\nParameter name: index ~~~ STACK TRACE: ~~~ at System.Collections.ArrayList.get_Item(Int32 index) at System.Diagnostics.ProcessThreadCollection.get_Item(Int32 index) in /Users/dotnet-bot/j/workspace/dotnet_corefx/master/osx10.12_debug/src/System.Diagnostics.Process/src/System/Diagnostics/ProcessThreadCollection.cs:line 35 at System.Diagnostics.Tests.ProcessThreadTests.TestThreadStateProperty() in /Users/dotnet-bot/j/workspace/dotnet_corefx/master/osx10.12_debug/src/System.Diagnostics.Process/tests/ProcessThreadTests.cs:line 199 ~~~
process
system diagnostics tests processthreadtests testthreadstateproperty failed in ci failed test system diagnostics tests processthreadtests testthreadstateproperty configuration debug detail message system argumentoutofrangeexception index was out of range must be non negative and less than the size of the collection nparameter name index stack trace at system collections arraylist get item index at system diagnostics processthreadcollection get item index in users dotnet bot j workspace dotnet corefx master debug src system diagnostics process src system diagnostics processthreadcollection cs line at system diagnostics tests processthreadtests testthreadstateproperty in users dotnet bot j workspace dotnet corefx master debug src system diagnostics process tests processthreadtests cs line
1
13,076
15,418,915,311
IssuesEvent
2021-03-05 09:27:25
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
closed
checkpoints should be linked to cell cycle transition terms via the "checkpoint signalling" term
PomBase cell cycle and DNA processes
from https://github.com/geneontology/go-ontology/issues/15582 like: ![20180418_092327_resized](https://user-images.githubusercontent.com/7359272/38920773-fba68f6a-42eb-11e8-9988-f50cc8e5a5ad.jpg)
1.0
checkpoints should be linked to cell cycle transition terms via the "checkpoint signalling" term - from https://github.com/geneontology/go-ontology/issues/15582 like: ![20180418_092327_resized](https://user-images.githubusercontent.com/7359272/38920773-fba68f6a-42eb-11e8-9988-f50cc8e5a5ad.jpg)
process
checkpoints should be linked to cell cycle transition terms via the checkpoint signalling term from like
1
369,252
25,835,368,231
IssuesEvent
2022-12-12 19:11:39
unikraft/docs
https://api.github.com/repos/unikraft/docs
closed
Improve Contributing: Split Coding Style + Add references under Overview
documentation enhancement
I noticed by looking up our commit message format documentation that we should implement some structural improvements to the `Contributing` page: 1) One would expect under the page `Coding style` only suggestions regarding the source code style and maybe (I am not so sure if this is the best idea) how to do documentation in the code. `Commit Message Format` and `Developer’s Certificate of Origin` should get moved and be integrated to `Submitting Changes`. 2) A reference to `Submitting changes` should be added under `Overview`/`Pull requests` 3) `Developer Certificate of Origin` (DCO) must stay and explains the actual meaning of the `Signed-Off` tag. This is not really pointed out under `Commits` and `Adressing Multipe Authors`. The signed-off is **required** from each author of a patch **because** this way they certify that the submission is published under the DCO. 4) The `Example Commit mesage` should show the **full** firstname in the signed-off tag. It is an example and people should see here a best practice example that contains full firstname and full lastname. The mail address can be whatever that person has, obviously.
1.0
Improve Contributing: Split Coding Style + Add references under Overview - I noticed by looking up our commit message format documentation that we should implement some structural improvements to the `Contributing` page: 1) One would expect under the page `Coding style` only suggestions regarding the source code style and maybe (I am not so sure if this is the best idea) how to do documentation in the code. `Commit Message Format` and `Developer’s Certificate of Origin` should get moved and be integrated to `Submitting Changes`. 2) A reference to `Submitting changes` should be added under `Overview`/`Pull requests` 3) `Developer Certificate of Origin` (DCO) must stay and explains the actual meaning of the `Signed-Off` tag. This is not really pointed out under `Commits` and `Adressing Multipe Authors`. The signed-off is **required** from each author of a patch **because** this way they certify that the submission is published under the DCO. 4) The `Example Commit mesage` should show the **full** firstname in the signed-off tag. It is an example and people should see here a best practice example that contains full firstname and full lastname. The mail address can be whatever that person has, obviously.
non_process
improve contributing split coding style add references under overview i noticed by looking up our commit message format documentation that we should implement some structural improvements to the contributing page one would expect under the page coding style only suggestions regarding the source code style and maybe i am not so sure if this is the best idea how to do documentation in the code commit message format and developer’s certificate of origin should get moved and be integrated to submitting changes a reference to submitting changes should be added under overview pull requests developer certificate of origin dco must stay and explains the actual meaning of the signed off tag this is not really pointed out under commits and adressing multipe authors the signed off is required from each author of a patch because this way they certify that the submission is published under the dco the example commit mesage should show the full firstname in the signed off tag it is an example and people should see here a best practice example that contains full firstname and full lastname the mail address can be whatever that person has obviously
0
13,253
15,725,718,680
IssuesEvent
2021-03-29 10:20:10
hashicorp/packer-plugin-amazon
https://api.github.com/repos/hashicorp/packer-plugin-amazon
opened
amazon-import post-processor handles multiple disks improperly
bug post-processor/amazon-import
_This issue was originally opened by @benjamb as hashicorp/packer#8590. It was migrated here as a result of the [Packer plugin split](https://github.com/hashicorp/packer/issues/8610#issuecomment-770034737). The original body of the issue is below._ <hr> #### Overview of the Issue In my case, I'm using the hyperv-iso builder and have entries within `disk_additional_size`, when this is passed on to the amazon-import post-processor it ends up finding the wrong disk. The code in question isn't particularly smart. It iterates over artefacts from a builder and returns the first artefact it finds that has the specified suffix (in my case vhdx). The issue with this, on Windows at least, is that `vm-0.vhdx` is found before the core image without the index suffix, i.e. `vm.vhdx` when iterating over the artefacts. for _, path := range artifact.Files() { if strings.HasSuffix(path, "."+p.config.Format) { source = path break } } It would make sense to perhaps specify which image you would like to be imported, rather than attempting to guess which one. As an additional improvement, Amazon's VM Import supports multiple disk images so this could be implemented (it would still make sense to add an optional config to specify disk images, perhaps using an explicit list or regex/glob). ### Packer version 1.5.1 (master) ### Operating system and Environment details Windows 10 Pro amd64
1.0
amazon-import post-processor handles multiple disks improperly - _This issue was originally opened by @benjamb as hashicorp/packer#8590. It was migrated here as a result of the [Packer plugin split](https://github.com/hashicorp/packer/issues/8610#issuecomment-770034737). The original body of the issue is below._ <hr> #### Overview of the Issue In my case, I'm using the hyperv-iso builder and have entries within `disk_additional_size`, when this is passed on to the amazon-import post-processor it ends up finding the wrong disk. The code in question isn't particularly smart. It iterates over artefacts from a builder and returns the first artefact it finds that has the specified suffix (in my case vhdx). The issue with this, on Windows at least, is that `vm-0.vhdx` is found before the core image without the index suffix, i.e. `vm.vhdx` when iterating over the artefacts. for _, path := range artifact.Files() { if strings.HasSuffix(path, "."+p.config.Format) { source = path break } } It would make sense to perhaps specify which image you would like to be imported, rather than attempting to guess which one. As an additional improvement, Amazon's VM Import supports multiple disk images so this could be implemented (it would still make sense to add an optional config to specify disk images, perhaps using an explicit list or regex/glob). ### Packer version 1.5.1 (master) ### Operating system and Environment details Windows 10 Pro amd64
process
amazon import post processor handles multiple disks improperly this issue was originally opened by benjamb as hashicorp packer it was migrated here as a result of the the original body of the issue is below overview of the issue in my case i m using the hyperv iso builder and have entries within disk additional size when this is passed on to the amazon import post processor it ends up finding the wrong disk the code in question isn t particularly smart it iterates over artefacts from a builder and returns the first artefact it finds that has the specified suffix in my case vhdx the issue with this on windows at least is that vm vhdx is found before the core image without the index suffix i e vm vhdx when iterating over the artefacts for path range artifact files if strings hassuffix path p config format source path break it would make sense to perhaps specify which image you would like to be imported rather than attempting to guess which one as an additional improvement amazon s vm import supports multiple disk images so this could be implemented it would still make sense to add an optional config to specify disk images perhaps using an explicit list or regex glob packer version master operating system and environment details windows pro
1
11,500
14,379,913,442
IssuesEvent
2020-12-02 01:30:41
tdwg/dwc
https://api.github.com/repos/tdwg/dwc
opened
Change term - coordinateUncertaintyInMeters
Class - Location Process - implement Term - change
## Change term * Submitter: https://github.com/RicardoOrtizG * Justification (why is this change necessary?): For completeness * Proponents (who needs this change): Everyone Proposed new attributes of the term: * Term name (in lowerCamelCase): coordinateUncertaintyInMeters * Organized in Class (e.g. Location, Taxon): Location * Examples: `30` (reasonable lower limit on or after 2020-05-01 of a GPS reading under good conditions if the actual precision was not recorded at the time). `100` (reasonable lower limit before 2020-05-01 of a GPS reading under good conditions if the actual precision was not recorded at the time). `71` (uncertainty for a UTM coordinate having 100 meter precision and a known spatial reference system). Submitted via https://github.com/gbif/doc-georeferencing-quick-reference-guide/issues/7
1.0
Change term - coordinateUncertaintyInMeters - ## Change term * Submitter: https://github.com/RicardoOrtizG * Justification (why is this change necessary?): For completeness * Proponents (who needs this change): Everyone Proposed new attributes of the term: * Term name (in lowerCamelCase): coordinateUncertaintyInMeters * Organized in Class (e.g. Location, Taxon): Location * Examples: `30` (reasonable lower limit on or after 2020-05-01 of a GPS reading under good conditions if the actual precision was not recorded at the time). `100` (reasonable lower limit before 2020-05-01 of a GPS reading under good conditions if the actual precision was not recorded at the time). `71` (uncertainty for a UTM coordinate having 100 meter precision and a known spatial reference system). Submitted via https://github.com/gbif/doc-georeferencing-quick-reference-guide/issues/7
process
change term coordinateuncertaintyinmeters change term submitter justification why is this change necessary for completeness proponents who needs this change everyone proposed new attributes of the term term name in lowercamelcase coordinateuncertaintyinmeters organized in class e g location taxon location examples reasonable lower limit on or after of a gps reading under good conditions if the actual precision was not recorded at the time reasonable lower limit before of a gps reading under good conditions if the actual precision was not recorded at the time uncertainty for a utm coordinate having meter precision and a known spatial reference system submitted via
1
3,035
6,038,038,146
IssuesEvent
2017-06-09 20:18:49
metabase/metabase
https://api.github.com/repos/metabase/metabase
opened
Give duplicate columns nicer aliases
Proposal Query Processor UX
From a Slack discussion between me and @senior. Right now if we have two columns with the same identifier, e.g. `name`, the Clojure JDBC library "helpfully" suffixes duplicates so we'll have `name` and `name_2`. It would be better if we just handled this ourselves where possible so we could give them more informative aliases, e.g. `venue_name` and `category_name`. This would be 100% under-the-hood so no UI changes. However we need to consider UX implementations. We might get complaints about us changing MB’s behavior, e.g. > why are the columns in my CSV different in 25 from 24?
1.0
Give duplicate columns nicer aliases - From a Slack discussion between me and @senior. Right now if we have two columns with the same identifier, e.g. `name`, the Clojure JDBC library "helpfully" suffixes duplicates so we'll have `name` and `name_2`. It would be better if we just handled this ourselves where possible so we could give them more informative aliases, e.g. `venue_name` and `category_name`. This would be 100% under-the-hood so no UI changes. However we need to consider UX implementations. We might get complaints about us changing MB’s behavior, e.g. > why are the columns in my CSV different in 25 from 24?
process
give duplicate columns nicer aliases from a slack discussion between me and senior right now if we have two columns with the same identifier e g name the clojure jdbc library helpfully suffixes duplicates so we ll have name and name it would be better if we just handled this ourselves where possible so we could give them more informative aliases e g venue name and category name this would be under the hood so no ui changes however we need to consider ux implementations we might get complaints about us changing mb’s behavior e g why are the columns in my csv different in from
1
7,823
10,996,990,856
IssuesEvent
2019-12-03 08:10:22
Altinn/altinn-studio
https://api.github.com/repos/Altinn/altinn-studio
closed
isArchived is not set to True when an instance is sent in
area/message-box area/process kind/bug solution/sbl team/nusse
## Describe the bug isArchived is not set to True when an instance is sent in by an user from SBL ## To Reproduce Steps to reproduce the behavior: 1. Login to AT21 and start a instance of a deployed app 2. Fill in data and send inn the instance 3. The receipt page is presented to the user 4. Navigate back to inbox and see that the instance is still available in the Inbox 5. Open the instance and click on 'Gå til skjema' and user is presented with a receipt ## Expected behavior When an instance is submitted/sent in, isArchived on the instance should be set to true and the instance should be moved to the archived. ## Additional info App: ttd/pdp-test Env: AT21 Browser: Chrome 78
1.0
isArchived is not set to True when an instance is sent in - ## Describe the bug isArchived is not set to True when an instance is sent in by an user from SBL ## To Reproduce Steps to reproduce the behavior: 1. Login to AT21 and start a instance of a deployed app 2. Fill in data and send inn the instance 3. The receipt page is presented to the user 4. Navigate back to inbox and see that the instance is still available in the Inbox 5. Open the instance and click on 'Gå til skjema' and user is presented with a receipt ## Expected behavior When an instance is submitted/sent in, isArchived on the instance should be set to true and the instance should be moved to the archived. ## Additional info App: ttd/pdp-test Env: AT21 Browser: Chrome 78
process
isarchived is not set to true when an instance is sent in describe the bug isarchived is not set to true when an instance is sent in by an user from sbl to reproduce steps to reproduce the behavior login to and start a instance of a deployed app fill in data and send inn the instance the receipt page is presented to the user navigate back to inbox and see that the instance is still available in the inbox open the instance and click on gå til skjema and user is presented with a receipt expected behavior when an instance is submitted sent in isarchived on the instance should be set to true and the instance should be moved to the archived additional info app ttd pdp test env browser chrome
1
12,725
15,095,056,931
IssuesEvent
2021-02-07 09:26:44
e4exp/paper_manager_abstract
https://api.github.com/repos/e4exp/paper_manager_abstract
opened
DeBERTa: Decoding-enhanced BERT with Disentangled Attention
2020 BERT Natural Language Processing Transformer _read_later
* https://arxiv.org/abs/2006.03654 * 2020 最近の事前学習型ニューラル言語モデルの進歩により、多くの自然言語処理(NLP)タスクの性能が大幅に向上した。 本論文では、2つの新しい技術を用いてBERTとRoBERTaモデルを改善する新しいモデルアーキテクチャDeBERTa(Decoding-enhanced BERT with disentangled attention)を提案する。 第一に、各単語はそれぞれその内容と位置を符号化する2つのベクトルを用いて表現され、単語間の注目度重みは、それぞれその内容と相対位置の分散行列を用いて計算される。 第二に、モデルの事前学習において、マスクされたトークンを予測するために、強化されたマスクデコーダを用いて、デコード層に絶対位置を組み込む。 さらに、モデルの一般化を向上させるための微調整に、新しい仮想敵対訓練法を用いた。 これらの手法により、モデルの事前学習の効率が大幅に改善され、自然言語理解(NLU)と自然言語生成(NLG)の両方の下流タスクの性能が向上することを示した。 RoBERTa-Large と比較して、訓練データの半分で訓練された DeBERTa モデルは、幅広い NLP タスクで一貫して優れた性能を発揮し、MNLI で +0.9% (90.2% 対 91.1%)、SQuAD v2.0 で +2.3% (88.4% 対 90.7%)、RACE で +3.6% (83.2% 対 86.8%) の改善を達成した。 特筆すべきは、15億個のパラメータを持つ48の変形層からなるより大きなバージョンを訓練することで、DeBERTaをスケールアップすることである。 この大幅な性能向上により、単一のDeBERTaモデルは、マクロ平均スコア(89.9対89.8)の点で初めてSuperGLUEベンチマーク(Wangら、2019a)の人間性能を上回り、アンサンブルDeBERTaモデルは、2021年1月6日時点でSuperGLUEリーダーボードのトップに位置し、人間のベースライン(90.3対89.8)をかなりの差で上回っている。
1.0
DeBERTa: Decoding-enhanced BERT with Disentangled Attention - * https://arxiv.org/abs/2006.03654 * 2020 最近の事前学習型ニューラル言語モデルの進歩により、多くの自然言語処理(NLP)タスクの性能が大幅に向上した。 本論文では、2つの新しい技術を用いてBERTとRoBERTaモデルを改善する新しいモデルアーキテクチャDeBERTa(Decoding-enhanced BERT with disentangled attention)を提案する。 第一に、各単語はそれぞれその内容と位置を符号化する2つのベクトルを用いて表現され、単語間の注目度重みは、それぞれその内容と相対位置の分散行列を用いて計算される。 第二に、モデルの事前学習において、マスクされたトークンを予測するために、強化されたマスクデコーダを用いて、デコード層に絶対位置を組み込む。 さらに、モデルの一般化を向上させるための微調整に、新しい仮想敵対訓練法を用いた。 これらの手法により、モデルの事前学習の効率が大幅に改善され、自然言語理解(NLU)と自然言語生成(NLG)の両方の下流タスクの性能が向上することを示した。 RoBERTa-Large と比較して、訓練データの半分で訓練された DeBERTa モデルは、幅広い NLP タスクで一貫して優れた性能を発揮し、MNLI で +0.9% (90.2% 対 91.1%)、SQuAD v2.0 で +2.3% (88.4% 対 90.7%)、RACE で +3.6% (83.2% 対 86.8%) の改善を達成した。 特筆すべきは、15億個のパラメータを持つ48の変形層からなるより大きなバージョンを訓練することで、DeBERTaをスケールアップすることである。 この大幅な性能向上により、単一のDeBERTaモデルは、マクロ平均スコア(89.9対89.8)の点で初めてSuperGLUEベンチマーク(Wangら、2019a)の人間性能を上回り、アンサンブルDeBERTaモデルは、2021年1月6日時点でSuperGLUEリーダーボードのトップに位置し、人間のベースライン(90.3対89.8)をかなりの差で上回っている。
process
deberta decoding enhanced bert with disentangled attention 最近の事前学習型ニューラル言語モデルの進歩により、多くの自然言語処理 nlp タスクの性能が大幅に向上した。 本論文では、 (decoding enhanced bert with disentangled attention)を提案する。 第一に、 、単語間の注目度重みは、それぞれその内容と相対位置の分散行列を用いて計算される。 第二に、モデルの事前学習において、マスクされたトークンを予測するために、強化されたマスクデコーダを用いて、デコード層に絶対位置を組み込む。 さらに、モデルの一般化を向上させるための微調整に、新しい仮想敵対訓練法を用いた。 これらの手法により、モデルの事前学習の効率が大幅に改善され、自然言語理解(nlu)と自然言語生成(nlg)の両方の下流タスクの性能が向上することを示した。 roberta large と比較して、訓練データの半分で訓練された deberta モデルは、幅広い nlp タスクで一貫して優れた性能を発揮し、mnli で 対 、squad で 対 、race で 対 の改善を達成した。 特筆すべきは、 、debertaをスケールアップすることである。 この大幅な性能向上により、単一のdebertaモデルは、マクロ平均スコア( )の点で初めてsuperglueベンチマーク(wangら、 )の人間性能を上回り、アンサンブルdebertaモデルは、 、人間のベースライン( )をかなりの差で上回っている。
1
9,919
12,958,827,021
IssuesEvent
2020-07-20 12:02:55
scikit-learn/scikit-learn
https://api.github.com/repos/scikit-learn/scikit-learn
closed
Warning when optimized value for GP's kernel hits bound
module:gaussian_process
Hi, in `sklearn.gaussian_process`, the parameters `alpha` and `length_scale` of the given kernel are optimized given a specific range of values. It can happen that the best parameter is outside this range. The fact that the returned parameter matches the upper or lower bound of the range can be an indicator of such an issue, and therefore I was wondering if it could be of interest to raise a Warning in such a case to recommend the user to broaden the range of values and re-do a fit. I could try to do it if you think it's worth it.
1.0
Warning when optimized value for GP's kernel hits bound - Hi, in `sklearn.gaussian_process`, the parameters `alpha` and `length_scale` of the given kernel are optimized given a specific range of values. It can happen that the best parameter is outside this range. The fact that the returned parameter matches the upper or lower bound of the range can be an indicator of such an issue, and therefore I was wondering if it could be of interest to raise a Warning in such a case to recommend the user to broaden the range of values and re-do a fit. I could try to do it if you think it's worth it.
process
warning when optimized value for gp s kernel hits bound hi in sklearn gaussian process the parameters alpha and length scale of the given kernel are optimized given a specific range of values it can happen that the best parameter is outside this range the fact that the returned parameter matches the upper or lower bound of the range can be an indicator of such an issue and therefore i was wondering if it could be of interest to raise a warning in such a case to recommend the user to broaden the range of values and re do a fit i could try to do it if you think it s worth it
1
61,596
12,192,963,494
IssuesEvent
2020-04-29 13:44:36
foss-responders/fossresponders.com
https://api.github.com/repos/foss-responders/fossresponders.com
closed
Change wording around cancelled events
code good first issue help wanted
Change from, "We are aggregating all of the cancelled, postponed, or virtual-only conferences affected by coronavirus. Take a look at the events here." to: "We’re contributing to a database of all of the cancelled, postponed, or virtual-only conferences affected by COVID-19. Take a look at the events here." Since we're contributing, not curating, it would be great to call out that we're also contributing to the community effort, and that we're not behind the actual airtable. From a comms standpoint, we should make sure that we are using COVID-19 vs "coronavirus."
1.0
Change wording around cancelled events - Change from, "We are aggregating all of the cancelled, postponed, or virtual-only conferences affected by coronavirus. Take a look at the events here." to: "We’re contributing to a database of all of the cancelled, postponed, or virtual-only conferences affected by COVID-19. Take a look at the events here." Since we're contributing, not curating, it would be great to call out that we're also contributing to the community effort, and that we're not behind the actual airtable. From a comms standpoint, we should make sure that we are using COVID-19 vs "coronavirus."
non_process
change wording around cancelled events change from we are aggregating all of the cancelled postponed or virtual only conferences affected by coronavirus take a look at the events here to we’re contributing to a database of all of the cancelled postponed or virtual only conferences affected by covid take a look at the events here since we re contributing not curating it would be great to call out that we re also contributing to the community effort and that we re not behind the actual airtable from a comms standpoint we should make sure that we are using covid vs coronavirus
0
95,985
16,113,016,315
IssuesEvent
2021-04-28 01:22:53
idonthaveafifaaddiction/react
https://api.github.com/repos/idonthaveafifaaddiction/react
opened
CVE-2021-23369 (High) detected in multiple libraries
security vulnerability
## CVE-2021-23369 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>handlebars-4.0.11.tgz</b>, <b>handlebars-4.0.6.tgz</b>, <b>handlebars-4.0.10.tgz</b>, <b>handlebars-4.0.5.tgz</b></p></summary> <p> <details><summary><b>handlebars-4.0.11.tgz</b></p></summary> <p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p> <p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.0.11.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.0.11.tgz</a></p> <p>Path to dependency file: react/fixtures/expiration/node_modules/handlebars/package.json</p> <p>Path to vulnerable library: react/fixtures/expiration/node_modules/handlebars/package.json</p> <p> Dependency Hierarchy: - react-scripts-1.0.17.tgz (Root Library) - jest-20.0.4.tgz - jest-cli-20.0.4.tgz - istanbul-api-1.2.1.tgz - istanbul-reports-1.1.3.tgz - :x: **handlebars-4.0.11.tgz** (Vulnerable Library) </details> <details><summary><b>handlebars-4.0.6.tgz</b></p></summary> <p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p> <p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.0.6.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.0.6.tgz</a></p> <p>Path to dependency file: react/fixtures/fiber-debugger/node_modules/handlebars/package.json</p> <p>Path to vulnerable library: react/fixtures/fiber-debugger/node_modules/handlebars/package.json,react/fixtures/ssr/node_modules/handlebars/package.json,react/fixtures/dom/node_modules/handlebars/package.json</p> <p> Dependency Hierarchy: - react-scripts-1.0.11.tgz (Root Library) - jest-20.0.4.tgz - jest-cli-20.0.4.tgz - istanbul-api-1.1.11.tgz - istanbul-reports-1.1.1.tgz - :x: **handlebars-4.0.6.tgz** (Vulnerable Library) </details> <details><summary><b>handlebars-4.0.10.tgz</b></p></summary> <p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p> <p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.0.10.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.0.10.tgz</a></p> <p>Path to dependency file: react/fixtures/attribute-behavior/node_modules/handlebars/package.json</p> <p>Path to vulnerable library: react/fixtures/attribute-behavior/node_modules/handlebars/package.json</p> <p> Dependency Hierarchy: - react-scripts-1.0.11.tgz (Root Library) - jest-20.0.4.tgz - jest-cli-20.0.4.tgz - istanbul-api-1.1.12.tgz - istanbul-reports-1.1.1.tgz - :x: **handlebars-4.0.10.tgz** (Vulnerable Library) </details> <details><summary><b>handlebars-4.0.5.tgz</b></p></summary> <p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p> <p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.0.5.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.0.5.tgz</a></p> <p>Path to dependency file: react/node_modules/handlebars/package.json</p> <p>Path to vulnerable library: react/node_modules/handlebars/package.json,react/scripts/bench/node_modules/handlebars/package.json</p> <p> Dependency Hierarchy: - lighthouse-2.0.0.tgz (Root Library) - :x: **handlebars-4.0.5.tgz** (Vulnerable Library) </details> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package handlebars before 4.7.7 are vulnerable to Remote Code Execution (RCE) when selecting certain compiling options to compile templates coming from an untrusted source. <p>Publish Date: 2021-04-12 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23369>CVE-2021-23369</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23369">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23369</a></p> <p>Release Date: 2021-04-12</p> <p>Fix Resolution: handlebars - 4.7.7</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"handlebars","packageVersion":"4.0.11","packageFilePaths":["/fixtures/expiration/node_modules/handlebars/package.json"],"isTransitiveDependency":true,"dependencyTree":"react-scripts:1.0.17;jest:20.0.4;jest-cli:20.0.4;istanbul-api:1.2.1;istanbul-reports:1.1.3;handlebars:4.0.11","isMinimumFixVersionAvailable":true,"minimumFixVersion":"handlebars - 4.7.7"},{"packageType":"javascript/Node.js","packageName":"handlebars","packageVersion":"4.0.6","packageFilePaths":["/fixtures/fiber-debugger/node_modules/handlebars/package.json","/fixtures/ssr/node_modules/handlebars/package.json","/fixtures/dom/node_modules/handlebars/package.json"],"isTransitiveDependency":true,"dependencyTree":"react-scripts:1.0.11;jest:20.0.4;jest-cli:20.0.4;istanbul-api:1.1.11;istanbul-reports:1.1.1;handlebars:4.0.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"handlebars - 4.7.7"},{"packageType":"javascript/Node.js","packageName":"handlebars","packageVersion":"4.0.10","packageFilePaths":["/fixtures/attribute-behavior/node_modules/handlebars/package.json"],"isTransitiveDependency":true,"dependencyTree":"react-scripts:1.0.11;jest:20.0.4;jest-cli:20.0.4;istanbul-api:1.1.12;istanbul-reports:1.1.1;handlebars:4.0.10","isMinimumFixVersionAvailable":true,"minimumFixVersion":"handlebars - 4.7.7"},{"packageType":"javascript/Node.js","packageName":"handlebars","packageVersion":"4.0.5","packageFilePaths":["/node_modules/handlebars/package.json","/scripts/bench/node_modules/handlebars/package.json"],"isTransitiveDependency":true,"dependencyTree":"lighthouse:2.0.0;handlebars:4.0.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"handlebars - 4.7.7"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-23369","vulnerabilityDetails":"The package handlebars before 4.7.7 are vulnerable to Remote Code Execution (RCE) when selecting certain compiling options to compile templates coming from an untrusted source.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23369","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2021-23369 (High) detected in multiple libraries - ## CVE-2021-23369 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>handlebars-4.0.11.tgz</b>, <b>handlebars-4.0.6.tgz</b>, <b>handlebars-4.0.10.tgz</b>, <b>handlebars-4.0.5.tgz</b></p></summary> <p> <details><summary><b>handlebars-4.0.11.tgz</b></p></summary> <p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p> <p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.0.11.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.0.11.tgz</a></p> <p>Path to dependency file: react/fixtures/expiration/node_modules/handlebars/package.json</p> <p>Path to vulnerable library: react/fixtures/expiration/node_modules/handlebars/package.json</p> <p> Dependency Hierarchy: - react-scripts-1.0.17.tgz (Root Library) - jest-20.0.4.tgz - jest-cli-20.0.4.tgz - istanbul-api-1.2.1.tgz - istanbul-reports-1.1.3.tgz - :x: **handlebars-4.0.11.tgz** (Vulnerable Library) </details> <details><summary><b>handlebars-4.0.6.tgz</b></p></summary> <p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p> <p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.0.6.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.0.6.tgz</a></p> <p>Path to dependency file: react/fixtures/fiber-debugger/node_modules/handlebars/package.json</p> <p>Path to vulnerable library: react/fixtures/fiber-debugger/node_modules/handlebars/package.json,react/fixtures/ssr/node_modules/handlebars/package.json,react/fixtures/dom/node_modules/handlebars/package.json</p> <p> Dependency Hierarchy: - react-scripts-1.0.11.tgz (Root Library) - jest-20.0.4.tgz - jest-cli-20.0.4.tgz - istanbul-api-1.1.11.tgz - istanbul-reports-1.1.1.tgz - :x: **handlebars-4.0.6.tgz** (Vulnerable Library) </details> <details><summary><b>handlebars-4.0.10.tgz</b></p></summary> <p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p> <p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.0.10.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.0.10.tgz</a></p> <p>Path to dependency file: react/fixtures/attribute-behavior/node_modules/handlebars/package.json</p> <p>Path to vulnerable library: react/fixtures/attribute-behavior/node_modules/handlebars/package.json</p> <p> Dependency Hierarchy: - react-scripts-1.0.11.tgz (Root Library) - jest-20.0.4.tgz - jest-cli-20.0.4.tgz - istanbul-api-1.1.12.tgz - istanbul-reports-1.1.1.tgz - :x: **handlebars-4.0.10.tgz** (Vulnerable Library) </details> <details><summary><b>handlebars-4.0.5.tgz</b></p></summary> <p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p> <p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.0.5.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.0.5.tgz</a></p> <p>Path to dependency file: react/node_modules/handlebars/package.json</p> <p>Path to vulnerable library: react/node_modules/handlebars/package.json,react/scripts/bench/node_modules/handlebars/package.json</p> <p> Dependency Hierarchy: - lighthouse-2.0.0.tgz (Root Library) - :x: **handlebars-4.0.5.tgz** (Vulnerable Library) </details> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package handlebars before 4.7.7 are vulnerable to Remote Code Execution (RCE) when selecting certain compiling options to compile templates coming from an untrusted source. <p>Publish Date: 2021-04-12 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23369>CVE-2021-23369</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23369">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23369</a></p> <p>Release Date: 2021-04-12</p> <p>Fix Resolution: handlebars - 4.7.7</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"handlebars","packageVersion":"4.0.11","packageFilePaths":["/fixtures/expiration/node_modules/handlebars/package.json"],"isTransitiveDependency":true,"dependencyTree":"react-scripts:1.0.17;jest:20.0.4;jest-cli:20.0.4;istanbul-api:1.2.1;istanbul-reports:1.1.3;handlebars:4.0.11","isMinimumFixVersionAvailable":true,"minimumFixVersion":"handlebars - 4.7.7"},{"packageType":"javascript/Node.js","packageName":"handlebars","packageVersion":"4.0.6","packageFilePaths":["/fixtures/fiber-debugger/node_modules/handlebars/package.json","/fixtures/ssr/node_modules/handlebars/package.json","/fixtures/dom/node_modules/handlebars/package.json"],"isTransitiveDependency":true,"dependencyTree":"react-scripts:1.0.11;jest:20.0.4;jest-cli:20.0.4;istanbul-api:1.1.11;istanbul-reports:1.1.1;handlebars:4.0.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"handlebars - 4.7.7"},{"packageType":"javascript/Node.js","packageName":"handlebars","packageVersion":"4.0.10","packageFilePaths":["/fixtures/attribute-behavior/node_modules/handlebars/package.json"],"isTransitiveDependency":true,"dependencyTree":"react-scripts:1.0.11;jest:20.0.4;jest-cli:20.0.4;istanbul-api:1.1.12;istanbul-reports:1.1.1;handlebars:4.0.10","isMinimumFixVersionAvailable":true,"minimumFixVersion":"handlebars - 4.7.7"},{"packageType":"javascript/Node.js","packageName":"handlebars","packageVersion":"4.0.5","packageFilePaths":["/node_modules/handlebars/package.json","/scripts/bench/node_modules/handlebars/package.json"],"isTransitiveDependency":true,"dependencyTree":"lighthouse:2.0.0;handlebars:4.0.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"handlebars - 4.7.7"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-23369","vulnerabilityDetails":"The package handlebars before 4.7.7 are vulnerable to Remote Code Execution (RCE) when selecting certain compiling options to compile templates coming from an untrusted source.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23369","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_process
cve high detected in multiple libraries cve high severity vulnerability vulnerable libraries handlebars tgz handlebars tgz handlebars tgz handlebars tgz handlebars tgz handlebars provides the power necessary to let you build semantic templates effectively with no frustration library home page a href path to dependency file react fixtures expiration node modules handlebars package json path to vulnerable library react fixtures expiration node modules handlebars package json dependency hierarchy react scripts tgz root library jest tgz jest cli tgz istanbul api tgz istanbul reports tgz x handlebars tgz vulnerable library handlebars tgz handlebars provides the power necessary to let you build semantic templates effectively with no frustration library home page a href path to dependency file react fixtures fiber debugger node modules handlebars package json path to vulnerable library react fixtures fiber debugger node modules handlebars package json react fixtures ssr node modules handlebars package json react fixtures dom node modules handlebars package json dependency hierarchy react scripts tgz root library jest tgz jest cli tgz istanbul api tgz istanbul reports tgz x handlebars tgz vulnerable library handlebars tgz handlebars provides the power necessary to let you build semantic templates effectively with no frustration library home page a href path to dependency file react fixtures attribute behavior node modules handlebars package json path to vulnerable library react fixtures attribute behavior node modules handlebars package json dependency hierarchy react scripts tgz root library jest tgz jest cli tgz istanbul api tgz istanbul reports tgz x handlebars tgz vulnerable library handlebars tgz handlebars provides the power necessary to let you build semantic templates effectively with no frustration library home page a href path to dependency file react node modules handlebars package json path to vulnerable library react node modules handlebars package json react scripts bench node modules handlebars package json dependency hierarchy lighthouse tgz root library x handlebars tgz vulnerable library found in base branch master vulnerability details the package handlebars before are vulnerable to remote code execution rce when selecting certain compiling options to compile templates coming from an untrusted source publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution handlebars isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree react scripts jest jest cli istanbul api istanbul reports handlebars isminimumfixversionavailable true minimumfixversion handlebars packagetype javascript node js packagename handlebars packageversion packagefilepaths istransitivedependency true dependencytree react scripts jest jest cli istanbul api istanbul reports handlebars isminimumfixversionavailable true minimumfixversion handlebars packagetype javascript node js packagename handlebars packageversion packagefilepaths istransitivedependency true dependencytree react scripts jest jest cli istanbul api istanbul reports handlebars isminimumfixversionavailable true minimumfixversion handlebars packagetype javascript node js packagename handlebars packageversion packagefilepaths istransitivedependency true dependencytree lighthouse handlebars isminimumfixversionavailable true minimumfixversion handlebars basebranches vulnerabilityidentifier cve vulnerabilitydetails the package handlebars before are vulnerable to remote code execution rce when selecting certain compiling options to compile templates coming from an untrusted source vulnerabilityurl
0
13,210
15,679,492,010
IssuesEvent
2021-03-25 00:37:46
nion-software/nionswift
https://api.github.com/repos/nion-software/nionswift
opened
Expand data item organize data in arbitrary structures (array, table, structure, etc.)
f - organization f - processing feature level - difficult stage - planning type - enhancement
This is a placeholder issue ... more to follow.
1.0
Expand data item organize data in arbitrary structures (array, table, structure, etc.) - This is a placeholder issue ... more to follow.
process
expand data item organize data in arbitrary structures array table structure etc this is a placeholder issue more to follow
1
47,494
7,329,285,811
IssuesEvent
2018-03-05 03:52:10
JavaEden/Orchid
https://api.github.com/repos/JavaEden/Orchid
closed
Improve Documentation
documentation
The documentation must be complete before Orchid is ready for public release. Specifically: - [x] Finish User Manual wiki entries - [x] Set each page to a Template (extends a base template, and overrides all blocks) - [x] Ensure each page is fully written and content style is consistent across all pages - [x] Ensure each page has examples and screenshots, if necessary - [ ] (maybe) Add a set of video tutorials on Youtube, demonstrating basic usage and customization - [ ] One video for using Orchid from the Starter Repo. Include getting set up with Netlify button and the Netlify CMS, and also running locally - [ ] One video for integrating Orchid into an existing Gradle project (use Clog as the example) - [ ] One video showing how to build custom versions for each of the following Orchid parts: Components, Tags, Functions, Menus, Generators - [x] Finish the sub-documentation sites for every official plugin - [x] Make sure each one has a unique color, name, and short description - [x] Ensure each plugins site clearly demonstrates basic usage of that plugin - [x] Ensure each site shows its custom classes' options, and that each option has a description - [x] Finish the sub-documentation sites for every official theme - [x] Give each Theme site identical content, and differ as little as possible in configuration - [x] Ensure each site shows its theme's options, and that each option has a description
1.0
Improve Documentation - The documentation must be complete before Orchid is ready for public release. Specifically: - [x] Finish User Manual wiki entries - [x] Set each page to a Template (extends a base template, and overrides all blocks) - [x] Ensure each page is fully written and content style is consistent across all pages - [x] Ensure each page has examples and screenshots, if necessary - [ ] (maybe) Add a set of video tutorials on Youtube, demonstrating basic usage and customization - [ ] One video for using Orchid from the Starter Repo. Include getting set up with Netlify button and the Netlify CMS, and also running locally - [ ] One video for integrating Orchid into an existing Gradle project (use Clog as the example) - [ ] One video showing how to build custom versions for each of the following Orchid parts: Components, Tags, Functions, Menus, Generators - [x] Finish the sub-documentation sites for every official plugin - [x] Make sure each one has a unique color, name, and short description - [x] Ensure each plugins site clearly demonstrates basic usage of that plugin - [x] Ensure each site shows its custom classes' options, and that each option has a description - [x] Finish the sub-documentation sites for every official theme - [x] Give each Theme site identical content, and differ as little as possible in configuration - [x] Ensure each site shows its theme's options, and that each option has a description
non_process
improve documentation the documentation must be complete before orchid is ready for public release specifically finish user manual wiki entries set each page to a template extends a base template and overrides all blocks ensure each page is fully written and content style is consistent across all pages ensure each page has examples and screenshots if necessary maybe add a set of video tutorials on youtube demonstrating basic usage and customization one video for using orchid from the starter repo include getting set up with netlify button and the netlify cms and also running locally one video for integrating orchid into an existing gradle project use clog as the example one video showing how to build custom versions for each of the following orchid parts components tags functions menus generators finish the sub documentation sites for every official plugin make sure each one has a unique color name and short description ensure each plugins site clearly demonstrates basic usage of that plugin ensure each site shows its custom classes options and that each option has a description finish the sub documentation sites for every official theme give each theme site identical content and differ as little as possible in configuration ensure each site shows its theme s options and that each option has a description
0
18,477
3,067,270,280
IssuesEvent
2015-08-18 09:22:51
contao/core
https://api.github.com/repos/contao/core
closed
Cannot unset string offsets in Controller.php on line 1478
defect
In [Controller.php:1478](https://github.com/contao/core/blob/ac68761904694febb7636efacf34c30575e720a0/system/modules/core/library/Contao/Controller.php#L1478) we try to unset the third array item, but `$size` isn’t always an array and this leads to the fatal error “Cannot unset string offsets”. I think we should change to a `deserialize($arrItem['size'], true)` in [Controller.php:1432](https://github.com/contao/core/blob/ac68761904694febb7636efacf34c30575e720a0/system/modules/core/library/Contao/Controller.php#L1432). Related issue: #7875
1.0
Cannot unset string offsets in Controller.php on line 1478 - In [Controller.php:1478](https://github.com/contao/core/blob/ac68761904694febb7636efacf34c30575e720a0/system/modules/core/library/Contao/Controller.php#L1478) we try to unset the third array item, but `$size` isn’t always an array and this leads to the fatal error “Cannot unset string offsets”. I think we should change to a `deserialize($arrItem['size'], true)` in [Controller.php:1432](https://github.com/contao/core/blob/ac68761904694febb7636efacf34c30575e720a0/system/modules/core/library/Contao/Controller.php#L1432). Related issue: #7875
non_process
cannot unset string offsets in controller php on line in we try to unset the third array item but size isn’t always an array and this leads to the fatal error “cannot unset string offsets” i think we should change to a deserialize arritem true in related issue
0
15,940
20,161,271,700
IssuesEvent
2022-02-09 21:51:11
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
opened
DISABLED test_terminate_exit (__main__.SpawnTest)
module: multiprocessing triaged module: flaky-tests skipped
Platforms%3A%20linux%0A%0A%20%20%20%20This%20test%20was%20disabled%20because%20it%20is%20failing%20on%20trunk.%20See%20%5Brecent%20examples%5D(http%3A%2F%2Ftorch-ci.com%2Ffailure%2Ftest_terminate_exit%252C%2520SpawnTest)%20and%20the%20most%20recent%0A%20%20%20%20%5Bworkflow%20logs%5D(https%3A%2F%2Fgithub.com%2Fpytorch%2Fpytorch%2Factions%2Fruns%2F1817341053).%0A%0A%20%20%20%20Over%20the%20past%206%20hours%2C%20it%20has%20been%20determined%20flaky%20in%201%20workflows%20with%0A%20%20%20%201%20red%20and%203%20green.
1.0
DISABLED test_terminate_exit (__main__.SpawnTest) - Platforms%3A%20linux%0A%0A%20%20%20%20This%20test%20was%20disabled%20because%20it%20is%20failing%20on%20trunk.%20See%20%5Brecent%20examples%5D(http%3A%2F%2Ftorch-ci.com%2Ffailure%2Ftest_terminate_exit%252C%2520SpawnTest)%20and%20the%20most%20recent%0A%20%20%20%20%5Bworkflow%20logs%5D(https%3A%2F%2Fgithub.com%2Fpytorch%2Fpytorch%2Factions%2Fruns%2F1817341053).%0A%0A%20%20%20%20Over%20the%20past%206%20hours%2C%20it%20has%20been%20determined%20flaky%20in%201%20workflows%20with%0A%20%20%20%201%20red%20and%203%20green.
process
disabled test terminate exit main spawntest platforms http ci com terminate exit https com
1
56,040
8,042,612,266
IssuesEvent
2018-07-31 08:43:43
fossasia/susi_linux
https://api.github.com/repos/fossasia/susi_linux
closed
To Update Documentation of the missing components
documentation
**Is your feature request related to a problem? Please describe.** Currently documentation not present regarding 1> Update Daemon : no of times updated, update requirements, etc. 2> How the SUSI Linux work in general : priority between the online/offline server, how the online and offline TTS and SST transition . 3> Where should SUSI be cloned for the complete process to work perfectly 4> The architecture of the code **Describe the solution you'd like** To update the documentation **Additional context** I would like to work on it
1.0
To Update Documentation of the missing components - **Is your feature request related to a problem? Please describe.** Currently documentation not present regarding 1> Update Daemon : no of times updated, update requirements, etc. 2> How the SUSI Linux work in general : priority between the online/offline server, how the online and offline TTS and SST transition . 3> Where should SUSI be cloned for the complete process to work perfectly 4> The architecture of the code **Describe the solution you'd like** To update the documentation **Additional context** I would like to work on it
non_process
to update documentation of the missing components is your feature request related to a problem please describe currently documentation not present regarding update daemon no of times updated update requirements etc how the susi linux work in general priority between the online offline server how the online and offline tts and sst transition where should susi be cloned for the complete process to work perfectly the architecture of the code describe the solution you d like to update the documentation additional context i would like to work on it
0
86,841
10,825,017,089
IssuesEvent
2019-11-09 13:18:18
InfiniteFlightAirportEditing/Airports
https://api.github.com/repos/InfiniteFlightAirportEditing/Airports
closed
LIMS-Piacenza-San Damiano Air Base-EMILIA-ROMAGNA-ITALY
Being Redone Design Issue Low Priority
# Airport Name Piacenza-San Damiano Air Base # Country? Italy # Improvements that need to be made? scratch # Are you working on this airport? yes # Airport Priority? (IF Event, 10000ft+ Runway, World/US Capital, Low) mo
1.0
LIMS-Piacenza-San Damiano Air Base-EMILIA-ROMAGNA-ITALY - # Airport Name Piacenza-San Damiano Air Base # Country? Italy # Improvements that need to be made? scratch # Are you working on this airport? yes # Airport Priority? (IF Event, 10000ft+ Runway, World/US Capital, Low) mo
non_process
lims piacenza san damiano air base emilia romagna italy airport name piacenza san damiano air base country italy improvements that need to be made scratch are you working on this airport yes airport priority if event runway world us capital low mo
0
146,783
19,471,082,454
IssuesEvent
2021-12-24 01:17:31
panasalap/frameworks_native
https://api.github.com/repos/panasalap/frameworks_native
opened
CVE-2021-0919 (Medium) detected in nativeandroid-10.0.0_r39
security vulnerability
## CVE-2021-0919 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>nativeandroid-10.0.0_r39</b></p></summary> <p> <p>Library home page: <a href=https://android.googlesource.com/platform/frameworks/native>https://android.googlesource.com/platform/frameworks/native</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/libs/binder/IServiceManager.cpp</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In getService of IServiceManager.cpp, there is a possible unhandled exception due to an integer overflow. This could lead to local denial of service making the lockscreen unusable with no additional execution privileges needed. User interaction is needed for exploitation.Product: AndroidVersions: Android-10 Android-11 Android-9Android ID: A-197336441 <p>Publish Date: 2021-12-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-0919>CVE-2021-0919</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.0</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-0919 (Medium) detected in nativeandroid-10.0.0_r39 - ## CVE-2021-0919 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>nativeandroid-10.0.0_r39</b></p></summary> <p> <p>Library home page: <a href=https://android.googlesource.com/platform/frameworks/native>https://android.googlesource.com/platform/frameworks/native</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/libs/binder/IServiceManager.cpp</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In getService of IServiceManager.cpp, there is a possible unhandled exception due to an integer overflow. This could lead to local denial of service making the lockscreen unusable with no additional execution privileges needed. User interaction is needed for exploitation.Product: AndroidVersions: Android-10 Android-11 Android-9Android ID: A-197336441 <p>Publish Date: 2021-12-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-0919>CVE-2021-0919</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.0</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in nativeandroid cve medium severity vulnerability vulnerable library nativeandroid library home page a href found in base branch master vulnerable source files libs binder iservicemanager cpp vulnerability details in getservice of iservicemanager cpp there is a possible unhandled exception due to an integer overflow this could lead to local denial of service making the lockscreen unusable with no additional execution privileges needed user interaction is needed for exploitation product androidversions android android android id a publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href step up your open source security game with whitesource
0
262,089
22,793,781,481
IssuesEvent
2022-07-10 12:03:35
realthunder/FreeCAD
https://api.github.com/repos/realthunder/FreeCAD
closed
Path workbench won't load.
testers needed
As soon as I got the message there is a new release I downloaded it. Found out that path workbench didn't work. `ModuleNotFoundError: No module named 'pivy'` `OS: Windows 10 (10.0) Word size of FreeCAD: 64-bit Version: 2022.704.26244 +4971 (Git) Build type: Release Branch: LinkDaily Hash: a87320cc64699a8e4226bad2e9b37ac3a1296ab9 Python version: 3.9.13 Qt version: 5.12.9 Coin version: 4.0.1 OCC version: 7.5.3 Locale: Dutch/Netherlands (nl_NL) `
1.0
Path workbench won't load. - As soon as I got the message there is a new release I downloaded it. Found out that path workbench didn't work. `ModuleNotFoundError: No module named 'pivy'` `OS: Windows 10 (10.0) Word size of FreeCAD: 64-bit Version: 2022.704.26244 +4971 (Git) Build type: Release Branch: LinkDaily Hash: a87320cc64699a8e4226bad2e9b37ac3a1296ab9 Python version: 3.9.13 Qt version: 5.12.9 Coin version: 4.0.1 OCC version: 7.5.3 Locale: Dutch/Netherlands (nl_NL) `
non_process
path workbench won t load as soon as i got the message there is a new release i downloaded it found out that path workbench didn t work modulenotfounderror no module named pivy os windows word size of freecad bit version git build type release branch linkdaily hash python version qt version coin version occ version locale dutch netherlands nl nl
0
8,580
11,747,506,811
IssuesEvent
2020-03-12 13:45:26
pwittchen/ReactiveNetwork
https://api.github.com/repos/pwittchen/ReactiveNetwork
opened
release 3.0.7
release process
**release notes**: - updated project dependencies - fixed bug #379 (changed default host protocol to https in `InternetObservingSettings) **things to do:** - [ ] update javadocs - [ ] bump version - [ ] release library - [ ] update changelog - [ ] create github release
1.0
release 3.0.7 - **release notes**: - updated project dependencies - fixed bug #379 (changed default host protocol to https in `InternetObservingSettings) **things to do:** - [ ] update javadocs - [ ] bump version - [ ] release library - [ ] update changelog - [ ] create github release
process
release release notes updated project dependencies fixed bug changed default host protocol to https in internetobservingsettings things to do update javadocs bump version release library update changelog create github release
1
344,727
30,756,007,630
IssuesEvent
2023-07-29 03:57:56
goalapa/spring-cacamuca
https://api.github.com/repos/goalapa/spring-cacamuca
closed
[test] 4주차 - [신고] 테스트 코드 검증
🎯test
✏️Description - 신고, 블랙리스트의 테스트 코드 검증 ✅TODO - - [x] Repository Service Tests - [x] Report Check Null Service Tests - [x] BlackList Service Tests 🐾ETC -
1.0
[test] 4주차 - [신고] 테스트 코드 검증 - ✏️Description - 신고, 블랙리스트의 테스트 코드 검증 ✅TODO - - [x] Repository Service Tests - [x] Report Check Null Service Tests - [x] BlackList Service Tests 🐾ETC -
non_process
테스트 코드 검증 ✏️description 신고 블랙리스트의 테스트 코드 검증 ✅todo repository service tests report check null service tests blacklist service tests 🐾etc
0
298,429
9,200,104,536
IssuesEvent
2019-03-07 16:19:47
canmet-energy/btap_tasks
https://api.github.com/repos/canmet-energy/btap_tasks
closed
NECB bug: LargeHotel now has > 3000 unmet hours
In Progress NECB 2011 NECB 2015 NECB 2017 Priority High Standards
### Description The LargeHotel archetype now shows more than 3000 unmet hours. This is the only archetype with that many unmet hours. ### Approach The unmet hours for this archetype jumped significantly with commit number 420f530 on July 24, 2018 by ckirney to nrcan branch. The problem appears to be that the 'Water Use Equipment Definitions' for thermal zones with multipliers includes the multiplier rather than being just for an individual thermal zone. For example 'Room_3_mult9_flr_6 Service Water Use Def' has a "Peak Flow Rate" of 2.9 x 10^-4 rather than 3 x 10^-5. This is roughly 9 times (which is the thermal zone multiplier) the amount it should be (for one individual thermal zone). Fix is to adjust the 'Water Use Equipment Definitions' to remove the multiplier. ### Testing Plan Use existing tests for LargeHotel. ### Waiting On Nothing. ### Repositories Involved https://github.com/NREL/openstudio-standards/tree/nrcan ### Select the Project Code you will be charging to 2018-2019 | Client |WSBE |Name |Manager |---------- |--- |--- |--- |PERD |P-002518.001.03 |Building Archetype |Phylroy Lopez
1.0
NECB bug: LargeHotel now has > 3000 unmet hours - ### Description The LargeHotel archetype now shows more than 3000 unmet hours. This is the only archetype with that many unmet hours. ### Approach The unmet hours for this archetype jumped significantly with commit number 420f530 on July 24, 2018 by ckirney to nrcan branch. The problem appears to be that the 'Water Use Equipment Definitions' for thermal zones with multipliers includes the multiplier rather than being just for an individual thermal zone. For example 'Room_3_mult9_flr_6 Service Water Use Def' has a "Peak Flow Rate" of 2.9 x 10^-4 rather than 3 x 10^-5. This is roughly 9 times (which is the thermal zone multiplier) the amount it should be (for one individual thermal zone). Fix is to adjust the 'Water Use Equipment Definitions' to remove the multiplier. ### Testing Plan Use existing tests for LargeHotel. ### Waiting On Nothing. ### Repositories Involved https://github.com/NREL/openstudio-standards/tree/nrcan ### Select the Project Code you will be charging to 2018-2019 | Client |WSBE |Name |Manager |---------- |--- |--- |--- |PERD |P-002518.001.03 |Building Archetype |Phylroy Lopez
non_process
necb bug largehotel now has unmet hours description the largehotel archetype now shows more than unmet hours this is the only archetype with that many unmet hours approach the unmet hours for this archetype jumped significantly with commit number on july by ckirney to nrcan branch the problem appears to be that the water use equipment definitions for thermal zones with multipliers includes the multiplier rather than being just for an individual thermal zone for example room flr service water use def has a peak flow rate of x rather than x this is roughly times which is the thermal zone multiplier the amount it should be for one individual thermal zone fix is to adjust the water use equipment definitions to remove the multiplier testing plan use existing tests for largehotel waiting on nothing repositories involved select the project code you will be charging to client wsbe name manager perd p building archetype phylroy lopez
0
11,423
14,248,086,846
IssuesEvent
2020-11-19 12:26:06
tikv/tikv
https://api.github.com/repos/tikv/tikv
closed
coprocessor/expression: push down scalar functions
difficulty/easy sig/coprocessor status/help-wanted
# Update The content in this issue is outdated. Please refer to https://github.com/tikv/tikv/issues/5751 for a latest list. <details> <summary>Click to expand the original content</summary> ## Feature Request In coprocessor, for reading operations, some functions have been pushed down to tikv to make the computation distribute into each region. For example, when TiDB receives a SQL query like ```sql select sum(col1+col2) from table1 ``` TiDB would push down the computation `sum(col1+col2)` to each region (TiKV) of the table. Now there are lots of functions need to been pushed down into TiKV, some have been implemented already while the rest may need your help. Here is the list of functions to been pushed down, you may pick one to make a pull request: ### Casting - [X] CastIntAsInt - [X] CastIntAsReal - [X] CastIntAsString - [X] CastIntAsDecimal - [X] CastIntAsTime - [X] CastIntAsDuration - [X] CastIntAsJson - [X] CastRealAsInt - [X] CastRealAsReal - [X] CastRealAsString - [X] CastRealAsDecimal - [X] CastRealAsTime - [X] CastRealAsDuration - [X] CastRealAsJson - [X] CastDecimalAsInt - [X] CastDecimalAsReal - [X] CastDecimalAsString - [X] CastDecimalAsDecimal - [X] CastDecimalAsTime - [X] CastDecimalAsDuration - [X] CastDecimalAsJson - [X] CastStringAsInt - [X] CastStringAsReal - [X] CastStringAsString - [X] CastStringAsDecimal - [X] CastStringAsTime - [X] CastStringAsDuration - [X] CastStringAsJson - [X] CastTimeAsInt - [X] CastTimeAsReal - [X] CastTimeAsString - [X] CastTimeAsDecimal - [X] CastTimeAsTime - [X] CastTimeAsDuration - [X] CastTimeAsJson - [X] CastDurationAsInt - [X] CastDurationAsReal - [X] CastDurationAsString - [X] CastDurationAsDecimal - [X] CastDurationAsTime - [X] CastDurationAsDuration - [X] CastDurationAsJson - [X] CastJsonAsInt - [X] CastJsonAsReal - [X] CastJsonAsString - [X] CastJsonAsDecimal - [X] CastJsonAsTime - [X] CastJsonAsDuration - [X] CastJsonAsJson ### Compare - [X] CoalesceInt - [X] CoalesceReal - [X] CoalesceDecimal - [X] CoalesceString - [X] CoalesceTime - [X] CoalesceDuration - [X] CoalesceJson - [X] LTInt - [X] LTReal - [X] LTDecimal - [X] LTString - [X] LTTime - [X] LTDuration - [X] LTJson - [X] LEInt - [X] LEReal - [X] LEDecimal - [X] LEString - [X] LETime - [X] LEDuration - [X] LEJson - [X] GTInt - [X] GTReal - [X] GTDecimal - [X] GTString - [X] GTTime - [X] GTDuration - [X] GTJson - [X] GreatestInt @bb7133 #3113 - [X] GreatestReal @bb7133 #3113 - [X] GreatestDecimal @bb7133 #3113 - [X] GreatestString @bb7133 #3113 - [X] GreatestTime @bb7133 #3113 - [X] LeastInt @bb7133 #3113 - [X] LeastReal @bb7133 #3113 - [X] LeastDecimal @bb7133 #3113 - [X] LeastString @bb7133 #3113 - [X] LeastTime @bb7133 #3113 - [X] IntervalInt @bb7133 #3330 - [X] IntervalReal @bb7133 #3330 - [X] GEInt - [X] GEReal - [X] GEDecimal - [X] GEString - [X] GETime - [X] GEDuration - [X] GEJson - [X] EQInt - [X] EQReal - [X] EQDecimal - [X] EQString - [X] EQTime - [X] EQDuration - [X] EQJson - [X] NEInt - [X] NEReal - [X] NEDecimal - [X] NEString - [X] NETime - [X] NEDuration - [X] NEJson - [X] NullEQInt - [X] NullEQReal - [X] NullEQDecimal - [X] NullEQString - [X] NullEQTime - [X] NullEQDuration - [X] NullEQJson ### Arithmetic - [X] PlusReal - [X] PlusDecimal - [X] PlusInt - [X] MinusReal - [X] MinusDecimal - [X] MinusInt - [X] MultiplyReal - [X] MultiplyDecimal - [X] MultiplyInt - [X] MultiplyIntUnsigned - [X] DivideReal - [X] DivideDecimal - [X] IntDivideInt @bb7133 #3030 - [X] IntDivideDecimal @bb7133 #3030 - [X] ModReal @bb7133 #3030 - [X] ModDecimal @bb7133 #3030 - [X] ModInt @bb7133 #3030 ### Math - [X] AbsInt - [X] AbsUInt - [X] AbsReal - [X] AbsDecimal - [X] CeilIntToDec - [X] CeilIntToInt - [X] CeilDecToInt - [X] CeilDecToDec - [X] CeilReal - [X] FloorIntToDec - [X] FloorIntToInt - [X] FloorDecToInt - [X] FloorDecToDec - [X] FloorReal - [X] RoundReal @colinback #3621 - [X] RoundInt @intellild #3395 - [X] RoundDec @colinback #3621 - [X] RoundWithFracReal @colinback #3621 - [X] RoundWithFracInt @colinback #3621 - [X] RoundWithFracDec @colinback #3621 - [X] Log1Arg @sllt #3603 - [X] Log2Args @sllt #3603 - [X] Log2 @sllt #3379 - [X] Log10 @sllt #3379 - [X] Rand @xiangyuf #3415 - [X] RandWithSeed @xiangyuf #3415 - [X] Pow @smallyard #3475 - [X] Conv @niedhui #3691 - [X] CRC32 @TennyZhuang #3374 - [X] Sign @Observer42 #3518 - [X] Sqrt @xiangyuf #3476 - [X] Acos @malc0lm #3482 - [X] Asin @malc0lm #3482 - [X] Atan1Arg @Observer42 #3520 - [X] Atan2Args @Observer42 #3520 - [X] Cos @vkorenev #3410 - [X] Cot @mtunique #3543 - [X] Degrees @mtunique #3543 - [X] Exp @niedhui #3686 - [X] PI @sweetIan #3382 - [X] Radians @niedhui #3683 - [X] Sin @liufuyang #3406 - [X] Tan @arosspope #3456 - [X] TruncateInt @niedhui #3532 - [X] TruncateReal @niedhui #3633 - [X] TruncateDecimal @niedhui #3637 ### Operator - [X] LogicalAnd - [X] LogicalOr - [X] LogicalXor - [X] UnaryNot - [X] UnaryMinusInt - [X] UnaryMinusReal - [X] UnaryMinusDecimal - [X] DecimalIsNull - [X] DurationIsNull - [X] RealIsNull - [X] StringIsNull - [X] TimeIsNull - [X] IntIsNull - [X] JsonIsNull - [X] BitAndSig - [X] BitOrSig - [X] BitXorSig - [X] BitNegSig - [X] IntIsTrue - [X] RealIsTrue - [X] DecimalIsTrue - [X] IntIsFalse - [X] RealIsFalse - [X] DecimalIsFalse - [X] LeftShift @spongedu #3391 - [X] RightShift @spongedu #3391 ### Control - [X] IfNullInt - [X] IfNullReal - [X] IfNullDecimal - [X] IfNullString - [X] IfNullTime - [X] IfNullDuration - [X] IfInt - [X] IfReal - [X] IfDecimal - [X] IfString - [X] IfTime - [X] IfDuration - [X] IfNullJson - [X] IfJson - [X] CaseWhenInt - [X] CaseWhenReal - [X] CaseWhenDecimal - [X] CaseWhenString - [X] CaseWhenTime - [X] CaseWhenDuration - [X] CaseWhenJson ### Encryption - [ ] AesDecrypt - [ ] AesEncrypt - [X] Compress @niedhui #3856 - [X] MD5 @Hijiao #3554 - [ ] Password - [ ] RandomBytes - [X] SHA1 @haoxiang47 #3612 - [X] SHA2 @spongedu #3649 - [X] Uncompress @niedhui #3856 - [X] UncompressedLength @niedhui #3856 ### Info - [ ] ~Database~ - [ ] ~FoundRows~ - [ ] ~CurrentUser~ - [ ] ~User~ - [ ] ~ConnectionID~ - [ ] ~LastInsertID~ - [ ] ~LastInsertIDWithID~ - [ ] ~Version~ - [ ] ~TiDBVersion~ - [ ] ~RowCount~ ### Miscellaneous - [ ] Sleep - [ ] Lock - [ ] ReleaseLock - [ ] DecimalAnyValue - [ ] DurationAnyValue - [ ] IntAnyValue - [ ] JSONAnyValue - [ ] RealAnyValue - [ ] StringAnyValue - [ ] TimeAnyValue - [X] InetAton @rleungx #3659 - [X] InetNtoa @rleungx #3659 - [X] Inet6Aton @sweetIan #3480 - [X] Inet6Ntoa @sweetIan #3519 - [X] IsIPv4 @opensourcegeek #3460 - [ ] IsIPv4Compat - [ ] IsIPv4Mapped - [X] IsIPv6 @opensourcegeek #3479 - [ ] UUID ### Like - [X] LikeSig - [X] RegexpBinarySig @bb7133 #3196 - [X] RegexpSig @bb7133 #3196 ### JSON - [X] JsonExtractSig - [X] JsonUnquoteSig - [X] JsonTypeSig - [X] JsonSetSig - [X] JsonInsertSig - [X] JsonReplaceSig - [X] JsonRemoveSig - [X] JsonMergeSig - [X] JsonObjectSig - [X] JsonArraySig ### Time - [X] DateFormatSig - [ ] DateLiteral - [x] DateDiff @edwardpku #3937 - [ ] NullTimeDiff - [ ] TimeStringTimeDiff - [ ] DurationStringTimeDiff - [ ] DurationDurationTimeDiff - [ ] StringTimeTimeDiff - [ ] StringDurationTimeDiff - [ ] StringStringTimeDiff - [ ] TimeTimeTimeDiff - [X] Date @hawkingrei #3428 - [X] Hour @koushiro #3753 - [X] Minute @koushiro #3753 - [X] Second @koushiro #3753 - [X] MicroSecond @koushiro #3753 - [X] Month @chux0519 #3569 - [X] MonthName @koushiro #3735 - [ ] NowWithArg - [ ] NowWithoutArg - [X] DayName @koushiro #3774 - [X] DayOfMonth @koushiro #3774 - [X] DayOfWeek @koushiro #3774 - [X] DayOfYear @koushiro #3774 - [X] WeekWithMode @AbnerZheng #3857 - [X] WeekWithoutMode @AbnerZheng #3861 - [X] WeekDay @koushiro #3871 - [X] WeekOfYear @koushiro #3871 - [X] Year @Kingwl #3622 - [x] YearWeekWithMode @AbnerDBFan #3876 - [x] YearWeekWithoutMode @AbnerDBFan #3876 - [ ] GetFormat - [ ] SysDateWithFsp - [ ] SysDateWithoutFsp - [ ] ~CurrentDate~ - [ ] CurrentTime0Arg - [ ] CurrentTime1Arg - [ ] Time - [ ] TimeLiteral - [ ] UTCDate - [ ] UTCTimestampWithArg - [ ] UTCTimestampWithoutArg - [x] AddDatetimeAndDuration @koushiro #3899 - [x] AddDatetimeAndString @koushiro #3899 - [x] AddTimeDateTimeNull @koushiro #4063 - [ ] AddStringAndDuration - [ ] AddStringAndString - [ ] AddTimeStringNull - [x] AddDurationAndDuration @GinYM #3984 - [x] AddDurationAndString @DCjanus #4010 - [x] AddTimeDurationNull @koushiro #4063 - [ ] AddDateAndDuration - [ ] AddDateAndString - [ ] SubDatetimeAndDuration - [ ] SubDatetimeAndString - [ ] SubTimeDateTimeNull - [ ] SubStringAndDuration - [ ] SubStringAndString - [ ] SubTimeStringNull - [ ] SubDurationAndDuration - [ ] SubDurationAndString - [ ] SubTimeDurationNull - [ ] SubDateAndDuration - [ ] SubDateAndString - [ ] UnixTimestampCurrent - [ ] UnixTimestampInt - [ ] UnixTimestampDec - [ ] ConvertTz - [ ] MakeDate - [ ] MakeTime - [ ] PeriodAdd - [ ] PeriodDiff - [ ] Quarter - [ ] SecToTime - [ ] TimeToSec - [ ] TimestampAdd - [x] ToDays @GinYM #3978 - [ ] ToSeconds - [ ] UTCTimeWithArg - [ ] UTCTimeWithoutArg - [ ] Timestamp1Arg - [ ] Timestamp2Args - [ ] TimestampLiteral - [X] LastDay @WPH95 #3556 - [ ] StrToDateDate - [ ] StrToDateDatetime - [ ] StrToDateDuration - [ ] FromUnixTime1Arg - [ ] FromUnixTime2Arg - [ ] ExtractDatetime - [ ] ExtractDuration - [ ] AddDateStringString - [ ] AddDateStringInt - [ ] AddDateStringDecimal - [ ] AddDateIntString - [ ] AddDateIntInt - [ ] AddDateDatetimeString - [ ] AddDateDatetimeInt - [ ] SubDateStringString - [ ] SubDateStringInt - [ ] SubDateStringDecimal - [ ] SubDateIntString - [ ] SubDateIntInt - [ ] SubDateDatetimeString - [ ] SubDateDatetimeInt - [ ] FromDays - [ ] TimeFormat - [ ] TimestampDiff ### String functions - [X] BitLength @spongedu #3376 - [X] Bin @spongedu #3397 - [X] ASCII @spongedu #3436 - [ ] Char - [X] CharLength @spongedu #3461 - [X] Concat @crazycs520 #3654 - [x] ConcatWS @kg88 #3818 - [ ] Convert - [X] Elt @spongedu #3555 - [ ] ExportSet3Arg - [ ] ExportSet4Arg - [ ] ExportSet5Arg - [x] FieldInt @manifoldQAQ #4007 - [x] FieldReal @manifoldQAQ #4007 - [x] FieldString @manifoldQAQ #4007 - [ ] FindInSet - [ ] Format - [ ] FormatWithLocale - [X] FromBase64 @niedhui #3716 - [X] HexIntArg @sweetIan #3478 - [X] HexStrArg @sweetIan #3478 - [ ] Insert - [ ] InsertBinary - [ ] Instr - [ ] InstrBinary - [x] LTrim @spongedu #3400 - [X] Left @spongedu #3413 - [x] LeftBinary - [X] Length @spongedu #3376 - [x] Locate2Args @gaodayue #4016 - [x] Locate3Args @gaodayue #4016 - [x] LocateBinary2Args @gaodayue #4016 - [x] LocateBinary3Args @gaodayue #4016 - [X] Lower @spongedu #3433 - [x] Lpad @niedhui #3943 - [x] LpadBinary @niedhui #3943 - [ ] MakeSet - [ ] OctInt @yjhmelody #3605 - [ ] OctString - [ ] Ord - [ ] Quote - [X] RTrim @spongedu #3400 - [ ] Repeat - [x] Replace @lerencao #4360 - [X] Reverse @spongedu #3435 - [X] ReverseBinary @spongedu #3435 - [X] Right @rleungx #3653 - [x] RightBinary @niedhui #3982 - [x] Rpad @niedhui #3914 - [x] RpadBinary @niedhui #3914 - [X] Space @niedhui #3841 - [X] Strcmp @niedhui #3879 - [X] Substring2Args @niedhui #3472 - [X] Substring3Args @niedhui #3472 - [X] SubstringBinary2Args @niedhui #3813 - [X] SubstringBinary3Args @niedhui #3813 - [X] SubstringIndex @niedhui #3717 - [X] ToBase64 @niedhui #3716 - [X] Trim1Arg @niedhui #3698 - [X] Trim2Args @niedhui #3698 - [X] Trim3Args @niedhui #3698 - [X] UnHex @sweetIan #3469 - [X] Upper @spongedu #3433 ### Other - [X] BitCount @spongedu #3394 - [ ] GetParamString - [ ] GetVar - [ ] RowSig - [ ] SetVar - [ ] ValuesDecimal - [ ] ValuesDuration - [ ] ValuesInt - [ ] ValuesJSON - [ ] ValuesReal - [ ] ValuesString - [ ] ValuesTime - [X] InInt @winoros #2411 - [X] InReal @winoros #2411 - [X] InDecimal @winoros #2411 - [X] InString @winoros #2411 - [X] InTime @winoros #2411 - [X] InDuration @winoros #2411 - [X] InJson @winoros #2411 </details>
1.0
coprocessor/expression: push down scalar functions - # Update The content in this issue is outdated. Please refer to https://github.com/tikv/tikv/issues/5751 for a latest list. <details> <summary>Click to expand the original content</summary> ## Feature Request In coprocessor, for reading operations, some functions have been pushed down to tikv to make the computation distribute into each region. For example, when TiDB receives a SQL query like ```sql select sum(col1+col2) from table1 ``` TiDB would push down the computation `sum(col1+col2)` to each region (TiKV) of the table. Now there are lots of functions need to been pushed down into TiKV, some have been implemented already while the rest may need your help. Here is the list of functions to been pushed down, you may pick one to make a pull request: ### Casting - [X] CastIntAsInt - [X] CastIntAsReal - [X] CastIntAsString - [X] CastIntAsDecimal - [X] CastIntAsTime - [X] CastIntAsDuration - [X] CastIntAsJson - [X] CastRealAsInt - [X] CastRealAsReal - [X] CastRealAsString - [X] CastRealAsDecimal - [X] CastRealAsTime - [X] CastRealAsDuration - [X] CastRealAsJson - [X] CastDecimalAsInt - [X] CastDecimalAsReal - [X] CastDecimalAsString - [X] CastDecimalAsDecimal - [X] CastDecimalAsTime - [X] CastDecimalAsDuration - [X] CastDecimalAsJson - [X] CastStringAsInt - [X] CastStringAsReal - [X] CastStringAsString - [X] CastStringAsDecimal - [X] CastStringAsTime - [X] CastStringAsDuration - [X] CastStringAsJson - [X] CastTimeAsInt - [X] CastTimeAsReal - [X] CastTimeAsString - [X] CastTimeAsDecimal - [X] CastTimeAsTime - [X] CastTimeAsDuration - [X] CastTimeAsJson - [X] CastDurationAsInt - [X] CastDurationAsReal - [X] CastDurationAsString - [X] CastDurationAsDecimal - [X] CastDurationAsTime - [X] CastDurationAsDuration - [X] CastDurationAsJson - [X] CastJsonAsInt - [X] CastJsonAsReal - [X] CastJsonAsString - [X] CastJsonAsDecimal - [X] CastJsonAsTime - [X] CastJsonAsDuration - [X] CastJsonAsJson ### Compare - [X] CoalesceInt - [X] CoalesceReal - [X] CoalesceDecimal - [X] CoalesceString - [X] CoalesceTime - [X] CoalesceDuration - [X] CoalesceJson - [X] LTInt - [X] LTReal - [X] LTDecimal - [X] LTString - [X] LTTime - [X] LTDuration - [X] LTJson - [X] LEInt - [X] LEReal - [X] LEDecimal - [X] LEString - [X] LETime - [X] LEDuration - [X] LEJson - [X] GTInt - [X] GTReal - [X] GTDecimal - [X] GTString - [X] GTTime - [X] GTDuration - [X] GTJson - [X] GreatestInt @bb7133 #3113 - [X] GreatestReal @bb7133 #3113 - [X] GreatestDecimal @bb7133 #3113 - [X] GreatestString @bb7133 #3113 - [X] GreatestTime @bb7133 #3113 - [X] LeastInt @bb7133 #3113 - [X] LeastReal @bb7133 #3113 - [X] LeastDecimal @bb7133 #3113 - [X] LeastString @bb7133 #3113 - [X] LeastTime @bb7133 #3113 - [X] IntervalInt @bb7133 #3330 - [X] IntervalReal @bb7133 #3330 - [X] GEInt - [X] GEReal - [X] GEDecimal - [X] GEString - [X] GETime - [X] GEDuration - [X] GEJson - [X] EQInt - [X] EQReal - [X] EQDecimal - [X] EQString - [X] EQTime - [X] EQDuration - [X] EQJson - [X] NEInt - [X] NEReal - [X] NEDecimal - [X] NEString - [X] NETime - [X] NEDuration - [X] NEJson - [X] NullEQInt - [X] NullEQReal - [X] NullEQDecimal - [X] NullEQString - [X] NullEQTime - [X] NullEQDuration - [X] NullEQJson ### Arithmetic - [X] PlusReal - [X] PlusDecimal - [X] PlusInt - [X] MinusReal - [X] MinusDecimal - [X] MinusInt - [X] MultiplyReal - [X] MultiplyDecimal - [X] MultiplyInt - [X] MultiplyIntUnsigned - [X] DivideReal - [X] DivideDecimal - [X] IntDivideInt @bb7133 #3030 - [X] IntDivideDecimal @bb7133 #3030 - [X] ModReal @bb7133 #3030 - [X] ModDecimal @bb7133 #3030 - [X] ModInt @bb7133 #3030 ### Math - [X] AbsInt - [X] AbsUInt - [X] AbsReal - [X] AbsDecimal - [X] CeilIntToDec - [X] CeilIntToInt - [X] CeilDecToInt - [X] CeilDecToDec - [X] CeilReal - [X] FloorIntToDec - [X] FloorIntToInt - [X] FloorDecToInt - [X] FloorDecToDec - [X] FloorReal - [X] RoundReal @colinback #3621 - [X] RoundInt @intellild #3395 - [X] RoundDec @colinback #3621 - [X] RoundWithFracReal @colinback #3621 - [X] RoundWithFracInt @colinback #3621 - [X] RoundWithFracDec @colinback #3621 - [X] Log1Arg @sllt #3603 - [X] Log2Args @sllt #3603 - [X] Log2 @sllt #3379 - [X] Log10 @sllt #3379 - [X] Rand @xiangyuf #3415 - [X] RandWithSeed @xiangyuf #3415 - [X] Pow @smallyard #3475 - [X] Conv @niedhui #3691 - [X] CRC32 @TennyZhuang #3374 - [X] Sign @Observer42 #3518 - [X] Sqrt @xiangyuf #3476 - [X] Acos @malc0lm #3482 - [X] Asin @malc0lm #3482 - [X] Atan1Arg @Observer42 #3520 - [X] Atan2Args @Observer42 #3520 - [X] Cos @vkorenev #3410 - [X] Cot @mtunique #3543 - [X] Degrees @mtunique #3543 - [X] Exp @niedhui #3686 - [X] PI @sweetIan #3382 - [X] Radians @niedhui #3683 - [X] Sin @liufuyang #3406 - [X] Tan @arosspope #3456 - [X] TruncateInt @niedhui #3532 - [X] TruncateReal @niedhui #3633 - [X] TruncateDecimal @niedhui #3637 ### Operator - [X] LogicalAnd - [X] LogicalOr - [X] LogicalXor - [X] UnaryNot - [X] UnaryMinusInt - [X] UnaryMinusReal - [X] UnaryMinusDecimal - [X] DecimalIsNull - [X] DurationIsNull - [X] RealIsNull - [X] StringIsNull - [X] TimeIsNull - [X] IntIsNull - [X] JsonIsNull - [X] BitAndSig - [X] BitOrSig - [X] BitXorSig - [X] BitNegSig - [X] IntIsTrue - [X] RealIsTrue - [X] DecimalIsTrue - [X] IntIsFalse - [X] RealIsFalse - [X] DecimalIsFalse - [X] LeftShift @spongedu #3391 - [X] RightShift @spongedu #3391 ### Control - [X] IfNullInt - [X] IfNullReal - [X] IfNullDecimal - [X] IfNullString - [X] IfNullTime - [X] IfNullDuration - [X] IfInt - [X] IfReal - [X] IfDecimal - [X] IfString - [X] IfTime - [X] IfDuration - [X] IfNullJson - [X] IfJson - [X] CaseWhenInt - [X] CaseWhenReal - [X] CaseWhenDecimal - [X] CaseWhenString - [X] CaseWhenTime - [X] CaseWhenDuration - [X] CaseWhenJson ### Encryption - [ ] AesDecrypt - [ ] AesEncrypt - [X] Compress @niedhui #3856 - [X] MD5 @Hijiao #3554 - [ ] Password - [ ] RandomBytes - [X] SHA1 @haoxiang47 #3612 - [X] SHA2 @spongedu #3649 - [X] Uncompress @niedhui #3856 - [X] UncompressedLength @niedhui #3856 ### Info - [ ] ~Database~ - [ ] ~FoundRows~ - [ ] ~CurrentUser~ - [ ] ~User~ - [ ] ~ConnectionID~ - [ ] ~LastInsertID~ - [ ] ~LastInsertIDWithID~ - [ ] ~Version~ - [ ] ~TiDBVersion~ - [ ] ~RowCount~ ### Miscellaneous - [ ] Sleep - [ ] Lock - [ ] ReleaseLock - [ ] DecimalAnyValue - [ ] DurationAnyValue - [ ] IntAnyValue - [ ] JSONAnyValue - [ ] RealAnyValue - [ ] StringAnyValue - [ ] TimeAnyValue - [X] InetAton @rleungx #3659 - [X] InetNtoa @rleungx #3659 - [X] Inet6Aton @sweetIan #3480 - [X] Inet6Ntoa @sweetIan #3519 - [X] IsIPv4 @opensourcegeek #3460 - [ ] IsIPv4Compat - [ ] IsIPv4Mapped - [X] IsIPv6 @opensourcegeek #3479 - [ ] UUID ### Like - [X] LikeSig - [X] RegexpBinarySig @bb7133 #3196 - [X] RegexpSig @bb7133 #3196 ### JSON - [X] JsonExtractSig - [X] JsonUnquoteSig - [X] JsonTypeSig - [X] JsonSetSig - [X] JsonInsertSig - [X] JsonReplaceSig - [X] JsonRemoveSig - [X] JsonMergeSig - [X] JsonObjectSig - [X] JsonArraySig ### Time - [X] DateFormatSig - [ ] DateLiteral - [x] DateDiff @edwardpku #3937 - [ ] NullTimeDiff - [ ] TimeStringTimeDiff - [ ] DurationStringTimeDiff - [ ] DurationDurationTimeDiff - [ ] StringTimeTimeDiff - [ ] StringDurationTimeDiff - [ ] StringStringTimeDiff - [ ] TimeTimeTimeDiff - [X] Date @hawkingrei #3428 - [X] Hour @koushiro #3753 - [X] Minute @koushiro #3753 - [X] Second @koushiro #3753 - [X] MicroSecond @koushiro #3753 - [X] Month @chux0519 #3569 - [X] MonthName @koushiro #3735 - [ ] NowWithArg - [ ] NowWithoutArg - [X] DayName @koushiro #3774 - [X] DayOfMonth @koushiro #3774 - [X] DayOfWeek @koushiro #3774 - [X] DayOfYear @koushiro #3774 - [X] WeekWithMode @AbnerZheng #3857 - [X] WeekWithoutMode @AbnerZheng #3861 - [X] WeekDay @koushiro #3871 - [X] WeekOfYear @koushiro #3871 - [X] Year @Kingwl #3622 - [x] YearWeekWithMode @AbnerDBFan #3876 - [x] YearWeekWithoutMode @AbnerDBFan #3876 - [ ] GetFormat - [ ] SysDateWithFsp - [ ] SysDateWithoutFsp - [ ] ~CurrentDate~ - [ ] CurrentTime0Arg - [ ] CurrentTime1Arg - [ ] Time - [ ] TimeLiteral - [ ] UTCDate - [ ] UTCTimestampWithArg - [ ] UTCTimestampWithoutArg - [x] AddDatetimeAndDuration @koushiro #3899 - [x] AddDatetimeAndString @koushiro #3899 - [x] AddTimeDateTimeNull @koushiro #4063 - [ ] AddStringAndDuration - [ ] AddStringAndString - [ ] AddTimeStringNull - [x] AddDurationAndDuration @GinYM #3984 - [x] AddDurationAndString @DCjanus #4010 - [x] AddTimeDurationNull @koushiro #4063 - [ ] AddDateAndDuration - [ ] AddDateAndString - [ ] SubDatetimeAndDuration - [ ] SubDatetimeAndString - [ ] SubTimeDateTimeNull - [ ] SubStringAndDuration - [ ] SubStringAndString - [ ] SubTimeStringNull - [ ] SubDurationAndDuration - [ ] SubDurationAndString - [ ] SubTimeDurationNull - [ ] SubDateAndDuration - [ ] SubDateAndString - [ ] UnixTimestampCurrent - [ ] UnixTimestampInt - [ ] UnixTimestampDec - [ ] ConvertTz - [ ] MakeDate - [ ] MakeTime - [ ] PeriodAdd - [ ] PeriodDiff - [ ] Quarter - [ ] SecToTime - [ ] TimeToSec - [ ] TimestampAdd - [x] ToDays @GinYM #3978 - [ ] ToSeconds - [ ] UTCTimeWithArg - [ ] UTCTimeWithoutArg - [ ] Timestamp1Arg - [ ] Timestamp2Args - [ ] TimestampLiteral - [X] LastDay @WPH95 #3556 - [ ] StrToDateDate - [ ] StrToDateDatetime - [ ] StrToDateDuration - [ ] FromUnixTime1Arg - [ ] FromUnixTime2Arg - [ ] ExtractDatetime - [ ] ExtractDuration - [ ] AddDateStringString - [ ] AddDateStringInt - [ ] AddDateStringDecimal - [ ] AddDateIntString - [ ] AddDateIntInt - [ ] AddDateDatetimeString - [ ] AddDateDatetimeInt - [ ] SubDateStringString - [ ] SubDateStringInt - [ ] SubDateStringDecimal - [ ] SubDateIntString - [ ] SubDateIntInt - [ ] SubDateDatetimeString - [ ] SubDateDatetimeInt - [ ] FromDays - [ ] TimeFormat - [ ] TimestampDiff ### String functions - [X] BitLength @spongedu #3376 - [X] Bin @spongedu #3397 - [X] ASCII @spongedu #3436 - [ ] Char - [X] CharLength @spongedu #3461 - [X] Concat @crazycs520 #3654 - [x] ConcatWS @kg88 #3818 - [ ] Convert - [X] Elt @spongedu #3555 - [ ] ExportSet3Arg - [ ] ExportSet4Arg - [ ] ExportSet5Arg - [x] FieldInt @manifoldQAQ #4007 - [x] FieldReal @manifoldQAQ #4007 - [x] FieldString @manifoldQAQ #4007 - [ ] FindInSet - [ ] Format - [ ] FormatWithLocale - [X] FromBase64 @niedhui #3716 - [X] HexIntArg @sweetIan #3478 - [X] HexStrArg @sweetIan #3478 - [ ] Insert - [ ] InsertBinary - [ ] Instr - [ ] InstrBinary - [x] LTrim @spongedu #3400 - [X] Left @spongedu #3413 - [x] LeftBinary - [X] Length @spongedu #3376 - [x] Locate2Args @gaodayue #4016 - [x] Locate3Args @gaodayue #4016 - [x] LocateBinary2Args @gaodayue #4016 - [x] LocateBinary3Args @gaodayue #4016 - [X] Lower @spongedu #3433 - [x] Lpad @niedhui #3943 - [x] LpadBinary @niedhui #3943 - [ ] MakeSet - [ ] OctInt @yjhmelody #3605 - [ ] OctString - [ ] Ord - [ ] Quote - [X] RTrim @spongedu #3400 - [ ] Repeat - [x] Replace @lerencao #4360 - [X] Reverse @spongedu #3435 - [X] ReverseBinary @spongedu #3435 - [X] Right @rleungx #3653 - [x] RightBinary @niedhui #3982 - [x] Rpad @niedhui #3914 - [x] RpadBinary @niedhui #3914 - [X] Space @niedhui #3841 - [X] Strcmp @niedhui #3879 - [X] Substring2Args @niedhui #3472 - [X] Substring3Args @niedhui #3472 - [X] SubstringBinary2Args @niedhui #3813 - [X] SubstringBinary3Args @niedhui #3813 - [X] SubstringIndex @niedhui #3717 - [X] ToBase64 @niedhui #3716 - [X] Trim1Arg @niedhui #3698 - [X] Trim2Args @niedhui #3698 - [X] Trim3Args @niedhui #3698 - [X] UnHex @sweetIan #3469 - [X] Upper @spongedu #3433 ### Other - [X] BitCount @spongedu #3394 - [ ] GetParamString - [ ] GetVar - [ ] RowSig - [ ] SetVar - [ ] ValuesDecimal - [ ] ValuesDuration - [ ] ValuesInt - [ ] ValuesJSON - [ ] ValuesReal - [ ] ValuesString - [ ] ValuesTime - [X] InInt @winoros #2411 - [X] InReal @winoros #2411 - [X] InDecimal @winoros #2411 - [X] InString @winoros #2411 - [X] InTime @winoros #2411 - [X] InDuration @winoros #2411 - [X] InJson @winoros #2411 </details>
process
coprocessor expression push down scalar functions update the content in this issue is outdated please refer to for a latest list click to expand the original content feature request in coprocessor for reading operations some functions have been pushed down to tikv to make the computation distribute into each region for example when tidb receives a sql query like sql select sum from tidb would push down the computation sum to each region tikv of the table now there are lots of functions need to been pushed down into tikv some have been implemented already while the rest may need your help here is the list of functions to been pushed down you may pick one to make a pull request casting castintasint castintasreal castintasstring castintasdecimal castintastime castintasduration castintasjson castrealasint castrealasreal castrealasstring castrealasdecimal castrealastime castrealasduration castrealasjson castdecimalasint castdecimalasreal castdecimalasstring castdecimalasdecimal castdecimalastime castdecimalasduration castdecimalasjson caststringasint caststringasreal caststringasstring caststringasdecimal caststringastime caststringasduration caststringasjson casttimeasint casttimeasreal casttimeasstring casttimeasdecimal casttimeastime casttimeasduration casttimeasjson castdurationasint castdurationasreal castdurationasstring castdurationasdecimal castdurationastime castdurationasduration castdurationasjson castjsonasint castjsonasreal castjsonasstring castjsonasdecimal castjsonastime castjsonasduration castjsonasjson compare coalesceint coalescereal coalescedecimal coalescestring coalescetime coalesceduration coalescejson ltint ltreal ltdecimal ltstring lttime ltduration ltjson leint lereal ledecimal lestring letime leduration lejson gtint gtreal gtdecimal gtstring gttime gtduration gtjson greatestint greatestreal greatestdecimal greateststring greatesttime leastint leastreal leastdecimal leaststring leasttime intervalint intervalreal geint gereal gedecimal gestring getime geduration gejson eqint eqreal eqdecimal eqstring eqtime eqduration eqjson neint nereal nedecimal nestring netime neduration nejson nulleqint nulleqreal nulleqdecimal nulleqstring nulleqtime nulleqduration nulleqjson arithmetic plusreal plusdecimal plusint minusreal minusdecimal minusint multiplyreal multiplydecimal multiplyint multiplyintunsigned dividereal dividedecimal intdivideint intdividedecimal modreal moddecimal modint math absint absuint absreal absdecimal ceilinttodec ceilinttoint ceildectoint ceildectodec ceilreal floorinttodec floorinttoint floordectoint floordectodec floorreal roundreal colinback roundint intellild rounddec colinback roundwithfracreal colinback roundwithfracint colinback roundwithfracdec colinback sllt sllt sllt sllt rand xiangyuf randwithseed xiangyuf pow smallyard conv niedhui tennyzhuang sign sqrt xiangyuf acos asin cos vkorenev cot mtunique degrees mtunique exp niedhui pi sweetian radians niedhui sin liufuyang tan arosspope truncateint niedhui truncatereal niedhui truncatedecimal niedhui operator logicaland logicalor logicalxor unarynot unaryminusint unaryminusreal unaryminusdecimal decimalisnull durationisnull realisnull stringisnull timeisnull intisnull jsonisnull bitandsig bitorsig bitxorsig bitnegsig intistrue realistrue decimalistrue intisfalse realisfalse decimalisfalse leftshift spongedu rightshift spongedu control ifnullint ifnullreal ifnulldecimal ifnullstring ifnulltime ifnullduration ifint ifreal ifdecimal ifstring iftime ifduration ifnulljson ifjson casewhenint casewhenreal casewhendecimal casewhenstring casewhentime casewhenduration casewhenjson encryption aesdecrypt aesencrypt compress niedhui hijiao password randombytes spongedu uncompress niedhui uncompressedlength niedhui info database foundrows currentuser user connectionid lastinsertid lastinsertidwithid version tidbversion rowcount miscellaneous sleep lock releaselock decimalanyvalue durationanyvalue intanyvalue jsonanyvalue realanyvalue stringanyvalue timeanyvalue inetaton rleungx inetntoa rleungx sweetian sweetian opensourcegeek opensourcegeek uuid like likesig regexpbinarysig regexpsig json jsonextractsig jsonunquotesig jsontypesig jsonsetsig jsoninsertsig jsonreplacesig jsonremovesig jsonmergesig jsonobjectsig jsonarraysig time dateformatsig dateliteral datediff edwardpku nulltimediff timestringtimediff durationstringtimediff durationdurationtimediff stringtimetimediff stringdurationtimediff stringstringtimediff timetimetimediff date hawkingrei hour koushiro minute koushiro second koushiro microsecond koushiro month monthname koushiro nowwitharg nowwithoutarg dayname koushiro dayofmonth koushiro dayofweek koushiro dayofyear koushiro weekwithmode abnerzheng weekwithoutmode abnerzheng weekday koushiro weekofyear koushiro year kingwl yearweekwithmode abnerdbfan yearweekwithoutmode abnerdbfan getformat sysdatewithfsp sysdatewithoutfsp currentdate time timeliteral utcdate utctimestampwitharg utctimestampwithoutarg adddatetimeandduration koushiro adddatetimeandstring koushiro addtimedatetimenull koushiro addstringandduration addstringandstring addtimestringnull adddurationandduration ginym adddurationandstring dcjanus addtimedurationnull koushiro adddateandduration adddateandstring subdatetimeandduration subdatetimeandstring subtimedatetimenull substringandduration substringandstring subtimestringnull subdurationandduration subdurationandstring subtimedurationnull subdateandduration subdateandstring unixtimestampcurrent unixtimestampint unixtimestampdec converttz makedate maketime periodadd perioddiff quarter sectotime timetosec timestampadd todays ginym toseconds utctimewitharg utctimewithoutarg timestampliteral lastday strtodatedate strtodatedatetime strtodateduration extractdatetime extractduration adddatestringstring adddatestringint adddatestringdecimal adddateintstring adddateintint adddatedatetimestring adddatedatetimeint subdatestringstring subdatestringint subdatestringdecimal subdateintstring subdateintint subdatedatetimestring subdatedatetimeint fromdays timeformat timestampdiff string functions bitlength spongedu bin spongedu ascii spongedu char charlength spongedu concat concatws convert elt spongedu fieldint manifoldqaq fieldreal manifoldqaq fieldstring manifoldqaq findinset format formatwithlocale niedhui hexintarg sweetian hexstrarg sweetian insert insertbinary instr instrbinary ltrim spongedu left spongedu leftbinary length spongedu gaodayue gaodayue gaodayue gaodayue lower spongedu lpad niedhui lpadbinary niedhui makeset octint yjhmelody octstring ord quote rtrim spongedu repeat replace lerencao reverse spongedu reversebinary spongedu right rleungx rightbinary niedhui rpad niedhui rpadbinary niedhui space niedhui strcmp niedhui niedhui niedhui niedhui niedhui substringindex niedhui niedhui niedhui niedhui niedhui unhex sweetian upper spongedu other bitcount spongedu getparamstring getvar rowsig setvar valuesdecimal valuesduration valuesint valuesjson valuesreal valuesstring valuestime inint winoros inreal winoros indecimal winoros instring winoros intime winoros induration winoros injson winoros
1
89,901
18,047,453,814
IssuesEvent
2021-09-19 06:06:45
wxcapture/wxcapture
https://api.github.com/repos/wxcapture/wxcapture
closed
Meteor image processing / enhancements
enhancement Pi code / config Priority Medium
The current Meteor images are impacted by: - thick black lines due to the buffer overflow on the satellite (M-2) - random noise lines where 1/2/3 of the sensors (Red / Green / Blue) does not return data [Red could be visible red or one of the 3 infra-red options] - images can be dark, especially when they are close to / include dawn / dusk Code to improve images to address some / all the above is required. Note there is a simple PoC - wxcutils_pi.fix_image which implements a very basic version of this solution which is already hooked into receive_meteor.py. This code should be replaced by any updated solution.
1.0
Meteor image processing / enhancements - The current Meteor images are impacted by: - thick black lines due to the buffer overflow on the satellite (M-2) - random noise lines where 1/2/3 of the sensors (Red / Green / Blue) does not return data [Red could be visible red or one of the 3 infra-red options] - images can be dark, especially when they are close to / include dawn / dusk Code to improve images to address some / all the above is required. Note there is a simple PoC - wxcutils_pi.fix_image which implements a very basic version of this solution which is already hooked into receive_meteor.py. This code should be replaced by any updated solution.
non_process
meteor image processing enhancements the current meteor images are impacted by thick black lines due to the buffer overflow on the satellite m random noise lines where of the sensors red green blue does not return data images can be dark especially when they are close to include dawn dusk code to improve images to address some all the above is required note there is a simple poc wxcutils pi fix image which implements a very basic version of this solution which is already hooked into receive meteor py this code should be replaced by any updated solution
0
22,353
31,030,404,887
IssuesEvent
2023-08-10 12:06:19
metabase/metabase
https://api.github.com/repos/metabase/metabase
closed
"$regex has to be a string" when using text filters on dashboards on MongoDB
Type:Bug Priority:P1 Database/Mongo Reporting/Dashboards Querying/Parameters & Variables .Team/QueryProcessor :hammer_and_wrench:
**Describe the bug** Normal match queries work without problem, but when using the $regex operators in dashboards, the query fails (misteriously it works on filters in the questions) **Logs** <details> ``` metabase_mongo | 2023-04-15 16:24:36,388 ERROR middleware.catch-exceptions :: Error processing query: Command failed with error 2 (BadValue): '$regex has to be a string' on server mongo_4_4:27017. The full response is {"ok": 0.0, "errmsg": "$regex has to be a string", "code": 2, "codeName": "BadValue"} metabase_mongo | {:database_id 3, metabase_mongo | :started_at #t "2023-04-15T16:24:36.284606Z[GMT]", metabase_mongo | :json_query metabase_mongo | {:constraints {:max-results 10000, :max-results-bare-rows 2000}, metabase_mongo | :type :native, metabase_mongo | :middleware {:js-int-to-string? true, :ignore-cached-results? false}, metabase_mongo | :native metabase_mongo | {:template-tags metabase_mongo | {"source" {:id "78da1237-0268-6e91-b3c4-32973fefae8a", :name "source", :display-name "Source", :type :text}}, metabase_mongo | :query metabase_mongo | "[\n {\"$match\": \n {\n \"source\": {\n \"$regex\": {{source}},\n \"$options\": \"i\"\n }\n }\n },\n {\n \"$project\": {\n \"_id\": \"$_id\",\n \"address\": \"$address\",\n \"email\": \"$email\",\n \"password\": \"$password\",\n \"name\": \"$name\",\n \"city\": \"$city\",\n \"longitude\": \"$longitude\",\n \"state\": \"$state\",\n \"source\": \"$source\",\n \"birth_date\": \"$birth_date\",\n \"zip\": \"$zip\",\n \"id\": \"$id\",\n \"latitude\": \"$latitude\",\n \"created_at\": \"$created_at\"\n }\n }\n]", metabase_mongo | :collection "people"}, metabase_mongo | :database 3, metabase_mongo | :parameters [{:type :string/=, :value ["facebook"], :id "9ff9b8ee", :target [:variable [:template-tag "source"]]}], metabase_mongo | :async? true, metabase_mongo | :cache-ttl nil}, metabase_mongo | :status :failed, metabase_mongo | :class com.mongodb.MongoCommandException, metabase_mongo | :stacktrace metabase_mongo | ["com.mongodb.internal.connection.ProtocolHelper.getCommandFailureException(ProtocolHelper.java:175)" metabase_mongo | "com.mongodb.internal.connection.InternalStreamConnection.receiveCommandMessageResponse(InternalStreamConnection.java:302)" metabase_mongo | "com.mongodb.internal.connection.InternalStreamConnection.sendAndReceive(InternalStreamConnection.java:258)" metabase_mongo | "com.mongodb.internal.connection.UsageTrackingInternalConnection.sendAndReceive(UsageTrackingInternalConnection.java:99)" metabase_mongo | "com.mongodb.internal.connection.DefaultConnectionPool$PooledConnection.sendAndReceive(DefaultConnectionPool.java:450)" metabase_mongo | "com.mongodb.internal.connection.CommandProtocolImpl.execute(CommandProtocolImpl.java:72)" metabase_mongo | "com.mongodb.internal.connection.DefaultServer$DefaultServerProtocolExecutor.execute(DefaultServer.java:226)" metabase_mongo | "com.mongodb.internal.connection.DefaultServerConnection.executeProtocol(DefaultServerConnection.java:269)" metabase_mongo | "com.mongodb.internal.connection.DefaultServerConnection.command(DefaultServerConnection.java:131)" metabase_mongo | "com.mongodb.internal.connection.DefaultServerConnection.command(DefaultServerConnection.java:123)" metabase_mongo | "com.mongodb.operation.CommandOperationHelper.executeCommand(CommandOperationHelper.java:343)" metabase_mongo | "com.mongodb.operation.CommandOperationHelper.executeCommand(CommandOperationHelper.java:334)" metabase_mongo | "com.mongodb.operation.CommandOperationHelper.executeCommandWithConnection(CommandOperationHelper.java:220)" metabase_mongo | "com.mongodb.operation.CommandOperationHelper$5.call(CommandOperationHelper.java:206)" metabase_mongo | "com.mongodb.operation.OperationHelper.withReadConnectionSource(OperationHelper.java:463)" metabase_mongo | "com.mongodb.operation.CommandOperationHelper.executeCommand(CommandOperationHelper.java:203)" metabase_mongo | "com.mongodb.operation.AggregateOperationImpl.execute(AggregateOperationImpl.java:200)" metabase_mongo | "com.mongodb.operation.AggregateOperation.execute(AggregateOperation.java:333)" metabase_mongo | "com.mongodb.operation.AggregateOperation.execute(AggregateOperation.java:42)" metabase_mongo | "com.mongodb.client.internal.MongoClientDelegate$DelegateOperationExecutor.execute(MongoClientDelegate.java:199)" metabase_mongo | "com.mongodb.client.internal.MongoClientDelegate$DelegateOperationExecutor.execute(MongoClientDelegate.java:180)" metabase_mongo | "com.mongodb.DBCollection.aggregate(DBCollection.java:1530)" metabase_mongo | "com.mongodb.DBCollection.aggregate(DBCollection.java:1488)" metabase_mongo | "com.mongodb.DBCollection.aggregate(DBCollection.java:1473)" metabase_mongo | "--> driver.mongo.execute$aggregate.invokeStatic(execute.clj:142)" metabase_mongo | "driver.mongo.execute$aggregate.invoke(execute.clj:136)" metabase_mongo | "driver.mongo.execute$execute_reducible_query.invokeStatic(execute.clj:179)" metabase_mongo | "driver.mongo.execute$execute_reducible_query.invoke(execute.clj:173)" metabase_mongo | "driver.mongo$fn__120589$f__119411__auto____120590.invoke(mongo.clj:308)" metabase_mongo | "driver.mongo.util$do_with_mongo_connection$fn__119399.invoke(util.clj:230)" metabase_mongo | "util.ssh$do_with_ssh_tunnel.invokeStatic(ssh.clj:141)" metabase_mongo | "util.ssh$do_with_ssh_tunnel.invoke(ssh.clj:130)" metabase_mongo | "driver.mongo.util$do_with_mongo_connection.invokeStatic(util.clj:224)" metabase_mongo | "driver.mongo.util$do_with_mongo_connection.invoke(util.clj:219)" metabase_mongo | "driver.mongo$fn__120589.invokeStatic(mongo.clj:307)" metabase_mongo | "driver.mongo$fn__120589.invoke(mongo.clj:305)" metabase_mongo | "query_processor.context$executef.invokeStatic(context.clj:60)" metabase_mongo | "query_processor.context$executef.invoke(context.clj:49)" metabase_mongo | "query_processor.context.default$default_runf.invokeStatic(default.clj:68)" metabase_mongo | "query_processor.context.default$default_runf.invoke(default.clj:66)" metabase_mongo | "query_processor.context$runf.invokeStatic(context.clj:46)" metabase_mongo | "query_processor.context$runf.invoke(context.clj:40)" metabase_mongo | "query_processor.reducible$identity_qp.invokeStatic(reducible.clj:15)" metabase_mongo | "query_processor.reducible$identity_qp.invoke(reducible.clj:12)" metabase_mongo | "query_processor.middleware.cache$maybe_return_cached_results$maybe_return_cached_results_STAR___69171.invoke(cache.clj:224)" metabase_mongo | "query_processor.middleware.permissions$check_query_permissions$fn__64643.invoke(permissions.clj:126)" metabase_mongo | "metabase_enterprise.advanced_permissions.query_processor.middleware.permissions$check_download_permissions$fn__105770.invoke(permissions.clj:125)" metabase_mongo | "metabase_enterprise.sandbox.query_processor.middleware.column_level_perms_check$maybe_apply_column_level_perms_check$fn__107204.invoke(column_level_perms_check.clj:31)" metabase_mongo | "query_processor.middleware.mbql_to_native$mbql__GT_native$fn__68062.invoke(mbql_to_native.clj:24)" metabase_mongo | "query_processor$fn__70690$combined_post_process__70695$combined_post_process_STAR___70696.invoke(query_processor.clj:243)" metabase_mongo | "query_processor$fn__70690$combined_pre_process__70691$combined_pre_process_STAR___70692.invoke(query_processor.clj:240)" metabase_mongo | "query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__69082$fn__69087.invoke(resolve_database_and_driver.clj:36)" metabase_mongo | "driver$do_with_driver.invokeStatic(driver.clj:90)" metabase_mongo | "driver$do_with_driver.invoke(driver.clj:86)" metabase_mongo | "query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__69082.invoke(resolve_database_and_driver.clj:35)" metabase_mongo | "query_processor.middleware.fetch_source_query$resolve_card_id_source_tables$fn__64952.invoke(fetch_source_query.clj:310)" metabase_mongo | "query_processor.middleware.store$initialize_store$fn__65130$fn__65131.invoke(store.clj:12)" metabase_mongo | "query_processor.store$do_with_store.invokeStatic(store.clj:47)" metabase_mongo | "query_processor.store$do_with_store.invoke(store.clj:41)" metabase_mongo | "query_processor.middleware.store$initialize_store$fn__65130.invoke(store.clj:11)" metabase_mongo | "query_processor.middleware.normalize_query$normalize$fn__69371.invoke(normalize_query.clj:25)" metabase_mongo | "metabase_enterprise.audit_app.query_processor.middleware.handle_audit_queries$handle_internal_queries$fn__75976.invoke(handle_audit_queries.clj:131)" metabase_mongo | "query_processor.middleware.constraints$add_default_userland_constraints$fn__66308.invoke(constraints.clj:54)" metabase_mongo | "query_processor.middleware.process_userland_query$process_userland_query$fn__69307.invoke(process_userland_query.clj:150)" metabase_mongo | "query_processor.middleware.catch_exceptions$catch_exceptions$fn__69684.invoke(catch_exceptions.clj:171)" metabase_mongo | "query_processor.reducible$async_qp$qp_STAR___59455$thunk__59457.invoke(reducible.clj:103)" metabase_mongo | "query_processor.reducible$async_qp$qp_STAR___59455$fn__59459.invoke(reducible.clj:108)"], metabase_mongo | :card_id 1, metabase_mongo | :context :dashboard, metabase_mongo | :error metabase_mongo | "Command failed with error 2 (BadValue): '$regex has to be a string' on server mongo_4_4:27017. The full response is {\"ok\": 0.0, \"errmsg\": \"$regex has to be a string\", \"code\": 2, \"codeName\": \"BadValue\"}", metabase_mongo | :row_count 0, metabase_mongo | :running_time 0, metabase_mongo | :data {:rows [], :cols []}} metabase_mongo | ``` </details> **To Reproduce** 1) create a native question in MongoDB like ``` [ {"$match": { "source": { "$regex": {{source}}, "$options": "i" } } }, { "$project": { "_id": "$_id", "address": "$address", "email": "$email", "password": "$password", "name": "$name", "city": "$city", "longitude": "$longitude", "state": "$state", "source": "$source", "birth_date": "$birth_date", "zip": "$zip", "id": "$id", "latitude": "$latitude", "created_at": "$created_at" } } ] ``` 2) Add the question to a dashboard 3) add a text filter to that dashboard 4) connect the text filter to the question and insert "Facebook" ... hit the issue Misteriously when you use the filter in the question it works **Expected behavior** Filters should work on every screen **Screenshots/videos** NA **Information about your Metabase Installation:** - Metabase version: 1.46.1 **Severity** P1 **Additional context** NA
1.0
"$regex has to be a string" when using text filters on dashboards on MongoDB - **Describe the bug** Normal match queries work without problem, but when using the $regex operators in dashboards, the query fails (misteriously it works on filters in the questions) **Logs** <details> ``` metabase_mongo | 2023-04-15 16:24:36,388 ERROR middleware.catch-exceptions :: Error processing query: Command failed with error 2 (BadValue): '$regex has to be a string' on server mongo_4_4:27017. The full response is {"ok": 0.0, "errmsg": "$regex has to be a string", "code": 2, "codeName": "BadValue"} metabase_mongo | {:database_id 3, metabase_mongo | :started_at #t "2023-04-15T16:24:36.284606Z[GMT]", metabase_mongo | :json_query metabase_mongo | {:constraints {:max-results 10000, :max-results-bare-rows 2000}, metabase_mongo | :type :native, metabase_mongo | :middleware {:js-int-to-string? true, :ignore-cached-results? false}, metabase_mongo | :native metabase_mongo | {:template-tags metabase_mongo | {"source" {:id "78da1237-0268-6e91-b3c4-32973fefae8a", :name "source", :display-name "Source", :type :text}}, metabase_mongo | :query metabase_mongo | "[\n {\"$match\": \n {\n \"source\": {\n \"$regex\": {{source}},\n \"$options\": \"i\"\n }\n }\n },\n {\n \"$project\": {\n \"_id\": \"$_id\",\n \"address\": \"$address\",\n \"email\": \"$email\",\n \"password\": \"$password\",\n \"name\": \"$name\",\n \"city\": \"$city\",\n \"longitude\": \"$longitude\",\n \"state\": \"$state\",\n \"source\": \"$source\",\n \"birth_date\": \"$birth_date\",\n \"zip\": \"$zip\",\n \"id\": \"$id\",\n \"latitude\": \"$latitude\",\n \"created_at\": \"$created_at\"\n }\n }\n]", metabase_mongo | :collection "people"}, metabase_mongo | :database 3, metabase_mongo | :parameters [{:type :string/=, :value ["facebook"], :id "9ff9b8ee", :target [:variable [:template-tag "source"]]}], metabase_mongo | :async? true, metabase_mongo | :cache-ttl nil}, metabase_mongo | :status :failed, metabase_mongo | :class com.mongodb.MongoCommandException, metabase_mongo | :stacktrace metabase_mongo | ["com.mongodb.internal.connection.ProtocolHelper.getCommandFailureException(ProtocolHelper.java:175)" metabase_mongo | "com.mongodb.internal.connection.InternalStreamConnection.receiveCommandMessageResponse(InternalStreamConnection.java:302)" metabase_mongo | "com.mongodb.internal.connection.InternalStreamConnection.sendAndReceive(InternalStreamConnection.java:258)" metabase_mongo | "com.mongodb.internal.connection.UsageTrackingInternalConnection.sendAndReceive(UsageTrackingInternalConnection.java:99)" metabase_mongo | "com.mongodb.internal.connection.DefaultConnectionPool$PooledConnection.sendAndReceive(DefaultConnectionPool.java:450)" metabase_mongo | "com.mongodb.internal.connection.CommandProtocolImpl.execute(CommandProtocolImpl.java:72)" metabase_mongo | "com.mongodb.internal.connection.DefaultServer$DefaultServerProtocolExecutor.execute(DefaultServer.java:226)" metabase_mongo | "com.mongodb.internal.connection.DefaultServerConnection.executeProtocol(DefaultServerConnection.java:269)" metabase_mongo | "com.mongodb.internal.connection.DefaultServerConnection.command(DefaultServerConnection.java:131)" metabase_mongo | "com.mongodb.internal.connection.DefaultServerConnection.command(DefaultServerConnection.java:123)" metabase_mongo | "com.mongodb.operation.CommandOperationHelper.executeCommand(CommandOperationHelper.java:343)" metabase_mongo | "com.mongodb.operation.CommandOperationHelper.executeCommand(CommandOperationHelper.java:334)" metabase_mongo | "com.mongodb.operation.CommandOperationHelper.executeCommandWithConnection(CommandOperationHelper.java:220)" metabase_mongo | "com.mongodb.operation.CommandOperationHelper$5.call(CommandOperationHelper.java:206)" metabase_mongo | "com.mongodb.operation.OperationHelper.withReadConnectionSource(OperationHelper.java:463)" metabase_mongo | "com.mongodb.operation.CommandOperationHelper.executeCommand(CommandOperationHelper.java:203)" metabase_mongo | "com.mongodb.operation.AggregateOperationImpl.execute(AggregateOperationImpl.java:200)" metabase_mongo | "com.mongodb.operation.AggregateOperation.execute(AggregateOperation.java:333)" metabase_mongo | "com.mongodb.operation.AggregateOperation.execute(AggregateOperation.java:42)" metabase_mongo | "com.mongodb.client.internal.MongoClientDelegate$DelegateOperationExecutor.execute(MongoClientDelegate.java:199)" metabase_mongo | "com.mongodb.client.internal.MongoClientDelegate$DelegateOperationExecutor.execute(MongoClientDelegate.java:180)" metabase_mongo | "com.mongodb.DBCollection.aggregate(DBCollection.java:1530)" metabase_mongo | "com.mongodb.DBCollection.aggregate(DBCollection.java:1488)" metabase_mongo | "com.mongodb.DBCollection.aggregate(DBCollection.java:1473)" metabase_mongo | "--> driver.mongo.execute$aggregate.invokeStatic(execute.clj:142)" metabase_mongo | "driver.mongo.execute$aggregate.invoke(execute.clj:136)" metabase_mongo | "driver.mongo.execute$execute_reducible_query.invokeStatic(execute.clj:179)" metabase_mongo | "driver.mongo.execute$execute_reducible_query.invoke(execute.clj:173)" metabase_mongo | "driver.mongo$fn__120589$f__119411__auto____120590.invoke(mongo.clj:308)" metabase_mongo | "driver.mongo.util$do_with_mongo_connection$fn__119399.invoke(util.clj:230)" metabase_mongo | "util.ssh$do_with_ssh_tunnel.invokeStatic(ssh.clj:141)" metabase_mongo | "util.ssh$do_with_ssh_tunnel.invoke(ssh.clj:130)" metabase_mongo | "driver.mongo.util$do_with_mongo_connection.invokeStatic(util.clj:224)" metabase_mongo | "driver.mongo.util$do_with_mongo_connection.invoke(util.clj:219)" metabase_mongo | "driver.mongo$fn__120589.invokeStatic(mongo.clj:307)" metabase_mongo | "driver.mongo$fn__120589.invoke(mongo.clj:305)" metabase_mongo | "query_processor.context$executef.invokeStatic(context.clj:60)" metabase_mongo | "query_processor.context$executef.invoke(context.clj:49)" metabase_mongo | "query_processor.context.default$default_runf.invokeStatic(default.clj:68)" metabase_mongo | "query_processor.context.default$default_runf.invoke(default.clj:66)" metabase_mongo | "query_processor.context$runf.invokeStatic(context.clj:46)" metabase_mongo | "query_processor.context$runf.invoke(context.clj:40)" metabase_mongo | "query_processor.reducible$identity_qp.invokeStatic(reducible.clj:15)" metabase_mongo | "query_processor.reducible$identity_qp.invoke(reducible.clj:12)" metabase_mongo | "query_processor.middleware.cache$maybe_return_cached_results$maybe_return_cached_results_STAR___69171.invoke(cache.clj:224)" metabase_mongo | "query_processor.middleware.permissions$check_query_permissions$fn__64643.invoke(permissions.clj:126)" metabase_mongo | "metabase_enterprise.advanced_permissions.query_processor.middleware.permissions$check_download_permissions$fn__105770.invoke(permissions.clj:125)" metabase_mongo | "metabase_enterprise.sandbox.query_processor.middleware.column_level_perms_check$maybe_apply_column_level_perms_check$fn__107204.invoke(column_level_perms_check.clj:31)" metabase_mongo | "query_processor.middleware.mbql_to_native$mbql__GT_native$fn__68062.invoke(mbql_to_native.clj:24)" metabase_mongo | "query_processor$fn__70690$combined_post_process__70695$combined_post_process_STAR___70696.invoke(query_processor.clj:243)" metabase_mongo | "query_processor$fn__70690$combined_pre_process__70691$combined_pre_process_STAR___70692.invoke(query_processor.clj:240)" metabase_mongo | "query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__69082$fn__69087.invoke(resolve_database_and_driver.clj:36)" metabase_mongo | "driver$do_with_driver.invokeStatic(driver.clj:90)" metabase_mongo | "driver$do_with_driver.invoke(driver.clj:86)" metabase_mongo | "query_processor.middleware.resolve_database_and_driver$resolve_database_and_driver$fn__69082.invoke(resolve_database_and_driver.clj:35)" metabase_mongo | "query_processor.middleware.fetch_source_query$resolve_card_id_source_tables$fn__64952.invoke(fetch_source_query.clj:310)" metabase_mongo | "query_processor.middleware.store$initialize_store$fn__65130$fn__65131.invoke(store.clj:12)" metabase_mongo | "query_processor.store$do_with_store.invokeStatic(store.clj:47)" metabase_mongo | "query_processor.store$do_with_store.invoke(store.clj:41)" metabase_mongo | "query_processor.middleware.store$initialize_store$fn__65130.invoke(store.clj:11)" metabase_mongo | "query_processor.middleware.normalize_query$normalize$fn__69371.invoke(normalize_query.clj:25)" metabase_mongo | "metabase_enterprise.audit_app.query_processor.middleware.handle_audit_queries$handle_internal_queries$fn__75976.invoke(handle_audit_queries.clj:131)" metabase_mongo | "query_processor.middleware.constraints$add_default_userland_constraints$fn__66308.invoke(constraints.clj:54)" metabase_mongo | "query_processor.middleware.process_userland_query$process_userland_query$fn__69307.invoke(process_userland_query.clj:150)" metabase_mongo | "query_processor.middleware.catch_exceptions$catch_exceptions$fn__69684.invoke(catch_exceptions.clj:171)" metabase_mongo | "query_processor.reducible$async_qp$qp_STAR___59455$thunk__59457.invoke(reducible.clj:103)" metabase_mongo | "query_processor.reducible$async_qp$qp_STAR___59455$fn__59459.invoke(reducible.clj:108)"], metabase_mongo | :card_id 1, metabase_mongo | :context :dashboard, metabase_mongo | :error metabase_mongo | "Command failed with error 2 (BadValue): '$regex has to be a string' on server mongo_4_4:27017. The full response is {\"ok\": 0.0, \"errmsg\": \"$regex has to be a string\", \"code\": 2, \"codeName\": \"BadValue\"}", metabase_mongo | :row_count 0, metabase_mongo | :running_time 0, metabase_mongo | :data {:rows [], :cols []}} metabase_mongo | ``` </details> **To Reproduce** 1) create a native question in MongoDB like ``` [ {"$match": { "source": { "$regex": {{source}}, "$options": "i" } } }, { "$project": { "_id": "$_id", "address": "$address", "email": "$email", "password": "$password", "name": "$name", "city": "$city", "longitude": "$longitude", "state": "$state", "source": "$source", "birth_date": "$birth_date", "zip": "$zip", "id": "$id", "latitude": "$latitude", "created_at": "$created_at" } } ] ``` 2) Add the question to a dashboard 3) add a text filter to that dashboard 4) connect the text filter to the question and insert "Facebook" ... hit the issue Misteriously when you use the filter in the question it works **Expected behavior** Filters should work on every screen **Screenshots/videos** NA **Information about your Metabase Installation:** - Metabase version: 1.46.1 **Severity** P1 **Additional context** NA
process
regex has to be a string when using text filters on dashboards on mongodb describe the bug normal match queries work without problem but when using the regex operators in dashboards the query fails misteriously it works on filters in the questions logs metabase mongo error middleware catch exceptions error processing query command failed with error badvalue regex has to be a string on server mongo the full response is ok errmsg regex has to be a string code codename badvalue metabase mongo database id metabase mongo started at t metabase mongo json query metabase mongo constraints max results max results bare rows metabase mongo type native metabase mongo middleware js int to string true ignore cached results false metabase mongo native metabase mongo template tags metabase mongo source id name source display name source type text metabase mongo query metabase mongo metabase mongo collection people metabase mongo database metabase mongo parameters id target metabase mongo async true metabase mongo cache ttl nil metabase mongo status failed metabase mongo class com mongodb mongocommandexception metabase mongo stacktrace metabase mongo com mongodb internal connection protocolhelper getcommandfailureexception protocolhelper java metabase mongo com mongodb internal connection internalstreamconnection receivecommandmessageresponse internalstreamconnection java metabase mongo com mongodb internal connection internalstreamconnection sendandreceive internalstreamconnection java metabase mongo com mongodb internal connection usagetrackinginternalconnection sendandreceive usagetrackinginternalconnection java metabase mongo com mongodb internal connection defaultconnectionpool pooledconnection sendandreceive defaultconnectionpool java metabase mongo com mongodb internal connection commandprotocolimpl execute commandprotocolimpl java metabase mongo com mongodb internal connection defaultserver defaultserverprotocolexecutor execute defaultserver java metabase mongo com mongodb internal connection defaultserverconnection executeprotocol defaultserverconnection java metabase mongo com mongodb internal connection defaultserverconnection command defaultserverconnection java metabase mongo com mongodb internal connection defaultserverconnection command defaultserverconnection java metabase mongo com mongodb operation commandoperationhelper executecommand commandoperationhelper java metabase mongo com mongodb operation commandoperationhelper executecommand commandoperationhelper java metabase mongo com mongodb operation commandoperationhelper executecommandwithconnection commandoperationhelper java metabase mongo com mongodb operation commandoperationhelper call commandoperationhelper java metabase mongo com mongodb operation operationhelper withreadconnectionsource operationhelper java metabase mongo com mongodb operation commandoperationhelper executecommand commandoperationhelper java metabase mongo com mongodb operation aggregateoperationimpl execute aggregateoperationimpl java metabase mongo com mongodb operation aggregateoperation execute aggregateoperation java metabase mongo com mongodb operation aggregateoperation execute aggregateoperation java metabase mongo com mongodb client internal mongoclientdelegate delegateoperationexecutor execute mongoclientdelegate java metabase mongo com mongodb client internal mongoclientdelegate delegateoperationexecutor execute mongoclientdelegate java metabase mongo com mongodb dbcollection aggregate dbcollection java metabase mongo com mongodb dbcollection aggregate dbcollection java metabase mongo com mongodb dbcollection aggregate dbcollection java metabase mongo driver mongo execute aggregate invokestatic execute clj metabase mongo driver mongo execute aggregate invoke execute clj metabase mongo driver mongo execute execute reducible query invokestatic execute clj metabase mongo driver mongo execute execute reducible query invoke execute clj metabase mongo driver mongo fn f auto invoke mongo clj metabase mongo driver mongo util do with mongo connection fn invoke util clj metabase mongo util ssh do with ssh tunnel invokestatic ssh clj metabase mongo util ssh do with ssh tunnel invoke ssh clj metabase mongo driver mongo util do with mongo connection invokestatic util clj metabase mongo driver mongo util do with mongo connection invoke util clj metabase mongo driver mongo fn invokestatic mongo clj metabase mongo driver mongo fn invoke mongo clj metabase mongo query processor context executef invokestatic context clj metabase mongo query processor context executef invoke context clj metabase mongo query processor context default default runf invokestatic default clj metabase mongo query processor context default default runf invoke default clj metabase mongo query processor context runf invokestatic context clj metabase mongo query processor context runf invoke context clj metabase mongo query processor reducible identity qp invokestatic reducible clj metabase mongo query processor reducible identity qp invoke reducible clj metabase mongo query processor middleware cache maybe return cached results maybe return cached results star invoke cache clj metabase mongo query processor middleware permissions check query permissions fn invoke permissions clj metabase mongo metabase enterprise advanced permissions query processor middleware permissions check download permissions fn invoke permissions clj metabase mongo metabase enterprise sandbox query processor middleware column level perms check maybe apply column level perms check fn invoke column level perms check clj metabase mongo query processor middleware mbql to native mbql gt native fn invoke mbql to native clj metabase mongo query processor fn combined post process combined post process star invoke query processor clj metabase mongo query processor fn combined pre process combined pre process star invoke query processor clj metabase mongo query processor middleware resolve database and driver resolve database and driver fn fn invoke resolve database and driver clj metabase mongo driver do with driver invokestatic driver clj metabase mongo driver do with driver invoke driver clj metabase mongo query processor middleware resolve database and driver resolve database and driver fn invoke resolve database and driver clj metabase mongo query processor middleware fetch source query resolve card id source tables fn invoke fetch source query clj metabase mongo query processor middleware store initialize store fn fn invoke store clj metabase mongo query processor store do with store invokestatic store clj metabase mongo query processor store do with store invoke store clj metabase mongo query processor middleware store initialize store fn invoke store clj metabase mongo query processor middleware normalize query normalize fn invoke normalize query clj metabase mongo metabase enterprise audit app query processor middleware handle audit queries handle internal queries fn invoke handle audit queries clj metabase mongo query processor middleware constraints add default userland constraints fn invoke constraints clj metabase mongo query processor middleware process userland query process userland query fn invoke process userland query clj metabase mongo query processor middleware catch exceptions catch exceptions fn invoke catch exceptions clj metabase mongo query processor reducible async qp qp star thunk invoke reducible clj metabase mongo query processor reducible async qp qp star fn invoke reducible clj metabase mongo card id metabase mongo context dashboard metabase mongo error metabase mongo command failed with error badvalue regex has to be a string on server mongo the full response is ok errmsg regex has to be a string code codename badvalue metabase mongo row count metabase mongo running time metabase mongo data rows cols metabase mongo to reproduce create a native question in mongodb like match source regex source options i project id id address address email email password password name name city city longitude longitude state state source source birth date birth date zip zip id id latitude latitude created at created at add the question to a dashboard add a text filter to that dashboard connect the text filter to the question and insert facebook hit the issue misteriously when you use the filter in the question it works expected behavior filters should work on every screen screenshots videos na information about your metabase installation metabase version severity additional context na
1
550,639
16,116,894,117
IssuesEvent
2021-04-28 08:32:27
avantifellows/plio-backend
https://api.github.com/repos/avantifellows/plio-backend
closed
Better way to handle dependencies
good first issue priority : medium
**Describe the task** We currently use requirements.txt and requirements-dev.txt. Though they help in segregating the environment-level dependencies, still managing them is not very easy. Some use cases that do not go well are: 1. In case I install a dependency using pip install and then decide its not "the one", there's no clear way to uninstall or remove it. 2. Once a dev-dependency is installed, we don't know what to do for next prod-level dependency as `pip freeze requirements.txt` will also add the dev-dependency. 3. There are many sub-dependencies that get added when doing `pip freeze`. If I need to remove the main dependency from requirements.txt, those sub-dependencies still remain. **Expected behavior** While doing some research in [Zappa](https://github.com/zappa/zappa), I saw a `requirements.in` that contained some dependencies. There was also a `requirements.txt` file that was auto-generated. Upon some more research, there seems to be a bit better way to manage dependencies in Python. Do some more research on how Zappa manages the dependencies and if we can adopt something similar. **References** Check out this blog on what `requirements.in` is and how it helps managing dependencies. Also see other alternatives mentioned. https://alysivji.github.io/python-managing-dependencies-with-pip-tools.html
1.0
Better way to handle dependencies - **Describe the task** We currently use requirements.txt and requirements-dev.txt. Though they help in segregating the environment-level dependencies, still managing them is not very easy. Some use cases that do not go well are: 1. In case I install a dependency using pip install and then decide its not "the one", there's no clear way to uninstall or remove it. 2. Once a dev-dependency is installed, we don't know what to do for next prod-level dependency as `pip freeze requirements.txt` will also add the dev-dependency. 3. There are many sub-dependencies that get added when doing `pip freeze`. If I need to remove the main dependency from requirements.txt, those sub-dependencies still remain. **Expected behavior** While doing some research in [Zappa](https://github.com/zappa/zappa), I saw a `requirements.in` that contained some dependencies. There was also a `requirements.txt` file that was auto-generated. Upon some more research, there seems to be a bit better way to manage dependencies in Python. Do some more research on how Zappa manages the dependencies and if we can adopt something similar. **References** Check out this blog on what `requirements.in` is and how it helps managing dependencies. Also see other alternatives mentioned. https://alysivji.github.io/python-managing-dependencies-with-pip-tools.html
non_process
better way to handle dependencies describe the task we currently use requirements txt and requirements dev txt though they help in segregating the environment level dependencies still managing them is not very easy some use cases that do not go well are in case i install a dependency using pip install and then decide its not the one there s no clear way to uninstall or remove it once a dev dependency is installed we don t know what to do for next prod level dependency as pip freeze requirements txt will also add the dev dependency there are many sub dependencies that get added when doing pip freeze if i need to remove the main dependency from requirements txt those sub dependencies still remain expected behavior while doing some research in i saw a requirements in that contained some dependencies there was also a requirements txt file that was auto generated upon some more research there seems to be a bit better way to manage dependencies in python do some more research on how zappa manages the dependencies and if we can adopt something similar references check out this blog on what requirements in is and how it helps managing dependencies also see other alternatives mentioned
0
40,736
5,258,565,060
IssuesEvent
2017-02-02 23:48:16
aspnet/Security
https://api.github.com/repos/aspnet/Security
closed
Make it harder (or impossible) to have more than one auth provider have "automatic" challenges
1 - Ready needs design
From discussion with @HaoK @Tratcher @blowdart and others, we suggest either: * Try to detect this scenario at runtime and alert the developer that this is an unsupported scenario (having more than one “automatic” challenge registered). * Or change how automatic challenges are registered so that only one auth provider per app pipeline can be automatic. For example, instead of each provider having its own bool setting, you would instead have one "global" setting (i.e. per-IAppBuilder) where you specify which auth protocol is the automatic one (and all others are not automatic).
1.0
Make it harder (or impossible) to have more than one auth provider have "automatic" challenges - From discussion with @HaoK @Tratcher @blowdart and others, we suggest either: * Try to detect this scenario at runtime and alert the developer that this is an unsupported scenario (having more than one “automatic” challenge registered). * Or change how automatic challenges are registered so that only one auth provider per app pipeline can be automatic. For example, instead of each provider having its own bool setting, you would instead have one "global" setting (i.e. per-IAppBuilder) where you specify which auth protocol is the automatic one (and all others are not automatic).
non_process
make it harder or impossible to have more than one auth provider have automatic challenges from discussion with haok tratcher blowdart and others we suggest either try to detect this scenario at runtime and alert the developer that this is an unsupported scenario having more than one “automatic” challenge registered or change how automatic challenges are registered so that only one auth provider per app pipeline can be automatic for example instead of each provider having its own bool setting you would instead have one global setting i e per iappbuilder where you specify which auth protocol is the automatic one and all others are not automatic
0
124,432
10,311,408,697
IssuesEvent
2019-08-29 17:17:44
radare/radare2
https://api.github.com/repos/radare/radare2
closed
Expose the analyze jump table as a command and API
RAnal test-required
This function takes the begining of the jump table, the address of the UJMP, the size of the delta pointers ,... the idea is that there must be an `a` subcommand that will get the required arguments to add the edge nodes to the current basic block, this can be useful to manually reanalize jump tables with different options, or to experiment with unsupported jump tables
1.0
Expose the analyze jump table as a command and API - This function takes the begining of the jump table, the address of the UJMP, the size of the delta pointers ,... the idea is that there must be an `a` subcommand that will get the required arguments to add the edge nodes to the current basic block, this can be useful to manually reanalize jump tables with different options, or to experiment with unsupported jump tables
non_process
expose the analyze jump table as a command and api this function takes the begining of the jump table the address of the ujmp the size of the delta pointers the idea is that there must be an a subcommand that will get the required arguments to add the edge nodes to the current basic block this can be useful to manually reanalize jump tables with different options or to experiment with unsupported jump tables
0
13,686
16,444,328,274
IssuesEvent
2021-05-20 17:40:52
GoogleCloudPlatform/java-docs-samples
https://api.github.com/repos/GoogleCloudPlatform/java-docs-samples
closed
Action Required: Fix Renovate Configuration
samples type: process
There is an error with this repository's Renovate configuration that needs to be fixed. As a precaution, Renovate will stop PRs until it is resolved. Location: `renovate.json` Error type: The renovate configuration file contains some invalid settings Message: `packageRules[1]: Each packageRule must contain at least one match* or exclude* selector. Rule: {"allowedVersions":"!/.+-sp\\.[0-9]+$/"}`
1.0
Action Required: Fix Renovate Configuration - There is an error with this repository's Renovate configuration that needs to be fixed. As a precaution, Renovate will stop PRs until it is resolved. Location: `renovate.json` Error type: The renovate configuration file contains some invalid settings Message: `packageRules[1]: Each packageRule must contain at least one match* or exclude* selector. Rule: {"allowedVersions":"!/.+-sp\\.[0-9]+$/"}`
process
action required fix renovate configuration there is an error with this repository s renovate configuration that needs to be fixed as a precaution renovate will stop prs until it is resolved location renovate json error type the renovate configuration file contains some invalid settings message packagerules each packagerule must contain at least one match or exclude selector rule allowedversions sp
1
11,440
14,260,941,868
IssuesEvent
2020-11-20 10:33:46
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
closed
GO:0002227 innate immune response in mucosa
Other term-related request multi-species process quick fix
GO:0002227 innate immune response in mucosa 21 EXP annotations Is this term required? mucosa seems to be epithelial cells and most of these have cell type in extensions.
1.0
GO:0002227 innate immune response in mucosa - GO:0002227 innate immune response in mucosa 21 EXP annotations Is this term required? mucosa seems to be epithelial cells and most of these have cell type in extensions.
process
go innate immune response in mucosa go innate immune response in mucosa exp annotations is this term required mucosa seems to be epithelial cells and most of these have cell type in extensions
1
12,027
14,738,565,826
IssuesEvent
2021-01-07 05:07:33
kdjstudios/SABillingGitlab
https://api.github.com/repos/kdjstudios/SABillingGitlab
closed
Keener - Staged fee being added to duplicate invoices.
anc-ops anc-process anp-important ant-bug ant-parent/primary ant-support
In GitLab by @kdjstudios on Jun 19, 2018, 09:25 **Submitted by:** Gaylan **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-06-04-57342/conversation **Server:** External **Client/Site:** Kenner **Account:** Multi **Issue:** For the two terminated account Channel Islands and Clarity Eye Group, those should both be corrected. It appears that when I did a debit to zero out the accounts, the debit was processed twice, therefore, leaving a balance on the account that should have been at zero. I am not sure how the debit was done twice since I did it as a staged fee and created the invoice in between billing cycles but it looks as if it did the staged fee again when the actual billing was done for 5/23 but now sure how it could have done that since the staged fee was used when I created the invoice.
1.0
Keener - Staged fee being added to duplicate invoices. - In GitLab by @kdjstudios on Jun 19, 2018, 09:25 **Submitted by:** Gaylan **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-06-04-57342/conversation **Server:** External **Client/Site:** Kenner **Account:** Multi **Issue:** For the two terminated account Channel Islands and Clarity Eye Group, those should both be corrected. It appears that when I did a debit to zero out the accounts, the debit was processed twice, therefore, leaving a balance on the account that should have been at zero. I am not sure how the debit was done twice since I did it as a staged fee and created the invoice in between billing cycles but it looks as if it did the staged fee again when the actual billing was done for 5/23 but now sure how it could have done that since the staged fee was used when I created the invoice.
process
keener staged fee being added to duplicate invoices in gitlab by kdjstudios on jun submitted by gaylan helpdesk server external client site kenner account multi issue for the two terminated account channel islands and clarity eye group those should both be corrected it appears that when i did a debit to zero out the accounts the debit was processed twice therefore leaving a balance on the account that should have been at zero i am not sure how the debit was done twice since i did it as a staged fee and created the invoice in between billing cycles but it looks as if it did the staged fee again when the actual billing was done for but now sure how it could have done that since the staged fee was used when i created the invoice
1
2,934
5,919,367,829
IssuesEvent
2017-05-22 17:31:56
dita-ot/dita-ot
https://api.github.com/repos/dita-ot/dita-ot
closed
Too many files generated in XHTML output folder when using conref situation
preprocess/conref
Let's say I have a DITA Map: ``` <!DOCTYPE map PUBLIC "-//OASIS//DTD DITA Map//EN" "map.dtd"> <map> <title>DITA Topic Map</title> <topicref href="a.dita"/> </map> ``` and 3 topics. a.dita: ``` <!DOCTYPE topic PUBLIC "-//OASIS//DTD DITA Topic//EN" "topic.dtd"> <topic id="topic_zlx_4gd_hp"> <title>ABC</title> <body> <p>ADITA </p> <p conref="b.dita#topic_gwk_qgd_hp/p_k5d_tgd_hp"/> </body> </topic> ``` b.dita: ``` <!DOCTYPE topic PUBLIC "-//OASIS//DTD DITA Topic//EN" "topic.dtd"> <topic id="topic_gwk_qgd_hp"> <title>B</title> <body> <p id="p_k5d_tgd_hp">BDITA</p> <p><xref href="c.dita"></xref></p> </body> </topic> ``` c.dita: ``` <!DOCTYPE topic PUBLIC "-//OASIS//DTD DITA Topic//EN" "topic.dtd"> <topic id="topic_snj_chd_hp"> <title>CCC</title> <body> <p>CDITA</p> </body> </topic> ``` When publishing to XHTML, only the "a.html" should be present in the output folder. Right now in the output folder we also have the "c.html" which is referenced from a part of "b.dita" which should have been ignored by the processing.
1.0
Too many files generated in XHTML output folder when using conref situation - Let's say I have a DITA Map: ``` <!DOCTYPE map PUBLIC "-//OASIS//DTD DITA Map//EN" "map.dtd"> <map> <title>DITA Topic Map</title> <topicref href="a.dita"/> </map> ``` and 3 topics. a.dita: ``` <!DOCTYPE topic PUBLIC "-//OASIS//DTD DITA Topic//EN" "topic.dtd"> <topic id="topic_zlx_4gd_hp"> <title>ABC</title> <body> <p>ADITA </p> <p conref="b.dita#topic_gwk_qgd_hp/p_k5d_tgd_hp"/> </body> </topic> ``` b.dita: ``` <!DOCTYPE topic PUBLIC "-//OASIS//DTD DITA Topic//EN" "topic.dtd"> <topic id="topic_gwk_qgd_hp"> <title>B</title> <body> <p id="p_k5d_tgd_hp">BDITA</p> <p><xref href="c.dita"></xref></p> </body> </topic> ``` c.dita: ``` <!DOCTYPE topic PUBLIC "-//OASIS//DTD DITA Topic//EN" "topic.dtd"> <topic id="topic_snj_chd_hp"> <title>CCC</title> <body> <p>CDITA</p> </body> </topic> ``` When publishing to XHTML, only the "a.html" should be present in the output folder. Right now in the output folder we also have the "c.html" which is referenced from a part of "b.dita" which should have been ignored by the processing.
process
too many files generated in xhtml output folder when using conref situation let s say i have a dita map dita topic map and topics a dita abc adita b dita b bdita c dita ccc cdita when publishing to xhtml only the a html should be present in the output folder right now in the output folder we also have the c html which is referenced from a part of b dita which should have been ignored by the processing
1
867
3,329,186,411
IssuesEvent
2015-11-11 00:21:59
beesmart-it/trend-hrm
https://api.github.com/repos/beesmart-it/trend-hrm
opened
Define use of startDate on selection process
enhancement selection process
Show processes only when current date > start date Use READY status, change to IN-PROCESS in a daily batch process
1.0
Define use of startDate on selection process - Show processes only when current date > start date Use READY status, change to IN-PROCESS in a daily batch process
process
define use of startdate on selection process show processes only when current date start date use ready status change to in process in a daily batch process
1
86,524
10,509,368,859
IssuesEvent
2019-09-27 10:48:49
GEOLYTIX/xyz
https://api.github.com/repos/GEOLYTIX/xyz
opened
_new locations
Documentation
By default the locations.select function will query the endpoint for the locations data from the connected database. A new flag can be used to prevent the select method to query the location from the database. This is for example used when aggregate locations should be added the locations list and map. ``` _xyz.locations.select({ _new: true, geometry: JSON.parse(e.target.response.geomj), infoj: e.target.response.infoj, layer: layer, }); ```
1.0
_new locations - By default the locations.select function will query the endpoint for the locations data from the connected database. A new flag can be used to prevent the select method to query the location from the database. This is for example used when aggregate locations should be added the locations list and map. ``` _xyz.locations.select({ _new: true, geometry: JSON.parse(e.target.response.geomj), infoj: e.target.response.infoj, layer: layer, }); ```
non_process
new locations by default the locations select function will query the endpoint for the locations data from the connected database a new flag can be used to prevent the select method to query the location from the database this is for example used when aggregate locations should be added the locations list and map xyz locations select new true geometry json parse e target response geomj infoj e target response infoj layer layer
0
9,335
12,340,785,072
IssuesEvent
2020-05-14 20:35:23
DiSSCo/user-stories
https://api.github.com/repos/DiSSCo/user-stories
opened
a CMS independent annotation system
2. University/Research institute 4. Data processing ICEDIG-SURVEY Research Specimen level
As a Scientist I want to visit a collection and annotate additional information of specimens through an Unified Curation and Annotation System UCAS so that I an capture information on geographical coordinates, locality, scientific name, accession number, collector name, and relevant measurements of specimens for this I need a CMS independent annotation system
1.0
a CMS independent annotation system - As a Scientist I want to visit a collection and annotate additional information of specimens through an Unified Curation and Annotation System UCAS so that I an capture information on geographical coordinates, locality, scientific name, accession number, collector name, and relevant measurements of specimens for this I need a CMS independent annotation system
process
a cms independent annotation system as a scientist i want to visit a collection and annotate additional information of specimens through an unified curation and annotation system ucas so that i an capture information on geographical coordinates locality scientific name accession number collector name and relevant measurements of specimens for this i need a cms independent annotation system
1
8,562
11,734,709,447
IssuesEvent
2020-03-11 09:50:17
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Can't open file
Pri3 automation/svc cxp process-automation/subsvc product-question triaged
I am getting the following error when I try sudo python /opt/microsoft/omsconfig/modules/nxOMSAutomationWorker/DSCResources/MSFT_nxOMSAutomationWorkerResource/automationworker/scripts/onboarding.py --register -w `<LogAnalyticsworkspaceId>` -k `<AutomationSharedKey>` -g `<hybridgroupname>` -e `<automationendpoint>` (with corresponding values) python: can't open file '/opt/microsoft/omsconfig/modules/nxOMSAutomationWorker/DSCResources/MSFT_nxOMSAutomationWorkerResource/automationworker/scripts/onboarding.py': [Errno 2] No such file or directory So I am stuck at this stage and no way how to proceed from here. I have run out of ideas. Can someone help me please? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: e38be5b8-d76d-a4f1-c014-7bf9248be2de * Version Independent ID: 976e5e90-b28c-d7ba-0495-69d92e62ea46 * Content: [Azure Automation Linux Hybrid Runbook Worker](https://docs.microsoft.com/en-us/azure/automation/automation-linux-hrw-install) * Content Source: [articles/automation/automation-linux-hrw-install.md](https://github.com/Microsoft/azure-docs/blob/master/articles/automation/automation-linux-hrw-install.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @MGoedtel * Microsoft Alias: **magoedte**
1.0
Can't open file - I am getting the following error when I try sudo python /opt/microsoft/omsconfig/modules/nxOMSAutomationWorker/DSCResources/MSFT_nxOMSAutomationWorkerResource/automationworker/scripts/onboarding.py --register -w `<LogAnalyticsworkspaceId>` -k `<AutomationSharedKey>` -g `<hybridgroupname>` -e `<automationendpoint>` (with corresponding values) python: can't open file '/opt/microsoft/omsconfig/modules/nxOMSAutomationWorker/DSCResources/MSFT_nxOMSAutomationWorkerResource/automationworker/scripts/onboarding.py': [Errno 2] No such file or directory So I am stuck at this stage and no way how to proceed from here. I have run out of ideas. Can someone help me please? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: e38be5b8-d76d-a4f1-c014-7bf9248be2de * Version Independent ID: 976e5e90-b28c-d7ba-0495-69d92e62ea46 * Content: [Azure Automation Linux Hybrid Runbook Worker](https://docs.microsoft.com/en-us/azure/automation/automation-linux-hrw-install) * Content Source: [articles/automation/automation-linux-hrw-install.md](https://github.com/Microsoft/azure-docs/blob/master/articles/automation/automation-linux-hrw-install.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @MGoedtel * Microsoft Alias: **magoedte**
process
can t open file i am getting the following error when i try sudo python opt microsoft omsconfig modules nxomsautomationworker dscresources msft nxomsautomationworkerresource automationworker scripts onboarding py register w k g e with corresponding values python can t open file opt microsoft omsconfig modules nxomsautomationworker dscresources msft nxomsautomationworkerresource automationworker scripts onboarding py no such file or directory so i am stuck at this stage and no way how to proceed from here i have run out of ideas can someone help me please document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login mgoedtel microsoft alias magoedte
1
7,194
10,331,817,056
IssuesEvent
2019-09-02 20:00:09
Ultimate-Hosts-Blacklist/whitelist
https://api.github.com/repos/Ultimate-Hosts-Blacklist/whitelist
closed
static.xx.fbcdn.net
whitelisting process
*@xxcriticxx commented on Dec 1, 2018, 6:11 AM UTC:* breaks fb Match found in [https://hosts.ubuntu101.co.za/hosts](https://hosts.ubuntu101.co.za/hosts): static.xx.fbcdn.net [![golden boy boxing home facebook](https://user-images.githubusercontent.com/15007183/49324896-ef1b8580-f505-11e8-9bde-eb689658a443.png)](https://user-images.githubusercontent.com/15007183/49324896-ef1b8580-f505-11e8-9bde-eb689658a443.png) *This issue was moved by [funilrys](https://github.com/funilrys) from [mitchellkrogza/Ultimate.Hosts.Blacklist#476](https://github.com/mitchellkrogza/Ultimate.Hosts.Blacklist/issues/476).*
1.0
static.xx.fbcdn.net - *@xxcriticxx commented on Dec 1, 2018, 6:11 AM UTC:* breaks fb Match found in [https://hosts.ubuntu101.co.za/hosts](https://hosts.ubuntu101.co.za/hosts): static.xx.fbcdn.net [![golden boy boxing home facebook](https://user-images.githubusercontent.com/15007183/49324896-ef1b8580-f505-11e8-9bde-eb689658a443.png)](https://user-images.githubusercontent.com/15007183/49324896-ef1b8580-f505-11e8-9bde-eb689658a443.png) *This issue was moved by [funilrys](https://github.com/funilrys) from [mitchellkrogza/Ultimate.Hosts.Blacklist#476](https://github.com/mitchellkrogza/Ultimate.Hosts.Blacklist/issues/476).*
process
static xx fbcdn net xxcriticxx commented on dec am utc breaks fb match found in static xx fbcdn net this issue was moved by from
1
14,757
18,040,654,991
IssuesEvent
2021-09-18 02:01:36
ooi-data/CE04OSPD-DP01B-03-FLCDRA103-recovered_inst-dpc_flcdrtd_instrument_recovered
https://api.github.com/repos/ooi-data/CE04OSPD-DP01B-03-FLCDRA103-recovered_inst-dpc_flcdrtd_instrument_recovered
opened
🛑 Processing failed: ResponseParserError
process
## Overview `ResponseParserError` found in `processing_task` task during run ended on 2021-09-18T02:01:36.099256. ## Details Flow name: `CE04OSPD-DP01B-03-FLCDRA103-recovered_inst-dpc_flcdrtd_instrument_recovered` Task name: `processing_task` Error type: `ResponseParserError` Error message: Unable to parse response (no element found: line 2, column 0), invalid XML received. Further retries may succeed: b'<?xml version="1.0" encoding="UTF-8"?>\n' <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 452, in _parse_xml_string_to_dom root = parser.close() xml.etree.ElementTree.ParseError: no element found: line 2, column 0 During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.8/site-packages/ooi_harvester/processor/pipeline.py", line 101, in processing final_path = finalize_zarr( File "/srv/conda/envs/notebook/lib/python3.8/site-packages/ooi_harvester/processor/__init__.py", line 359, in finalize_zarr source_store.fs.delete(source_store.root, recursive=True) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/fsspec/spec.py", line 1187, in delete return self.rm(path, recursive=recursive, maxdepth=maxdepth) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/fsspec/asyn.py", line 88, in wrapper return sync(self.loop, func, *args, **kwargs) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/fsspec/asyn.py", line 69, in sync raise result[0] File "/srv/conda/envs/notebook/lib/python3.8/site-packages/fsspec/asyn.py", line 25, in _runner result[0] = await coro File "/srv/conda/envs/notebook/lib/python3.8/site-packages/s3fs/core.py", line 1677, in _rm await asyncio.gather( File "/srv/conda/envs/notebook/lib/python3.8/site-packages/s3fs/core.py", line 1657, in _bulk_delete await self._call_s3("delete_objects", kwargs, Bucket=bucket, Delete=delete_keys) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/s3fs/core.py", line 268, in _call_s3 raise err File "/srv/conda/envs/notebook/lib/python3.8/site-packages/s3fs/core.py", line 248, in _call_s3 out = await method(**additional_kwargs) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/aiobotocore/client.py", line 141, in _make_api_call http, parsed_response = await self._make_request( File "/srv/conda/envs/notebook/lib/python3.8/site-packages/aiobotocore/client.py", line 161, in _make_request return await self._endpoint.make_request(operation_model, request_dict) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/aiobotocore/endpoint.py", line 93, in _send_request success_response, exception = await self._get_response( File "/srv/conda/envs/notebook/lib/python3.8/site-packages/aiobotocore/endpoint.py", line 112, in _get_response success_response, exception = await self._do_get_response( File "/srv/conda/envs/notebook/lib/python3.8/site-packages/aiobotocore/endpoint.py", line 177, in _do_get_response parsed_response = parser.parse( File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 245, in parse parsed = self._do_parse(response, shape) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 813, in _do_parse self._add_modeled_parse(response, shape, final_parsed) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 822, in _add_modeled_parse self._parse_payload(response, shape, member_shapes, final_parsed) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 862, in _parse_payload original_parsed = self._initial_body_parse(response['body']) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 948, in _initial_body_parse return self._parse_xml_string_to_dom(xml_string) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 454, in _parse_xml_string_to_dom raise ResponseParserError( botocore.parsers.ResponseParserError: Unable to parse response (no element found: line 2, column 0), invalid XML received. Further retries may succeed: b'<?xml version="1.0" encoding="UTF-8"?>\n' ``` </details>
1.0
🛑 Processing failed: ResponseParserError - ## Overview `ResponseParserError` found in `processing_task` task during run ended on 2021-09-18T02:01:36.099256. ## Details Flow name: `CE04OSPD-DP01B-03-FLCDRA103-recovered_inst-dpc_flcdrtd_instrument_recovered` Task name: `processing_task` Error type: `ResponseParserError` Error message: Unable to parse response (no element found: line 2, column 0), invalid XML received. Further retries may succeed: b'<?xml version="1.0" encoding="UTF-8"?>\n' <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 452, in _parse_xml_string_to_dom root = parser.close() xml.etree.ElementTree.ParseError: no element found: line 2, column 0 During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.8/site-packages/ooi_harvester/processor/pipeline.py", line 101, in processing final_path = finalize_zarr( File "/srv/conda/envs/notebook/lib/python3.8/site-packages/ooi_harvester/processor/__init__.py", line 359, in finalize_zarr source_store.fs.delete(source_store.root, recursive=True) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/fsspec/spec.py", line 1187, in delete return self.rm(path, recursive=recursive, maxdepth=maxdepth) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/fsspec/asyn.py", line 88, in wrapper return sync(self.loop, func, *args, **kwargs) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/fsspec/asyn.py", line 69, in sync raise result[0] File "/srv/conda/envs/notebook/lib/python3.8/site-packages/fsspec/asyn.py", line 25, in _runner result[0] = await coro File "/srv/conda/envs/notebook/lib/python3.8/site-packages/s3fs/core.py", line 1677, in _rm await asyncio.gather( File "/srv/conda/envs/notebook/lib/python3.8/site-packages/s3fs/core.py", line 1657, in _bulk_delete await self._call_s3("delete_objects", kwargs, Bucket=bucket, Delete=delete_keys) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/s3fs/core.py", line 268, in _call_s3 raise err File "/srv/conda/envs/notebook/lib/python3.8/site-packages/s3fs/core.py", line 248, in _call_s3 out = await method(**additional_kwargs) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/aiobotocore/client.py", line 141, in _make_api_call http, parsed_response = await self._make_request( File "/srv/conda/envs/notebook/lib/python3.8/site-packages/aiobotocore/client.py", line 161, in _make_request return await self._endpoint.make_request(operation_model, request_dict) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/aiobotocore/endpoint.py", line 93, in _send_request success_response, exception = await self._get_response( File "/srv/conda/envs/notebook/lib/python3.8/site-packages/aiobotocore/endpoint.py", line 112, in _get_response success_response, exception = await self._do_get_response( File "/srv/conda/envs/notebook/lib/python3.8/site-packages/aiobotocore/endpoint.py", line 177, in _do_get_response parsed_response = parser.parse( File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 245, in parse parsed = self._do_parse(response, shape) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 813, in _do_parse self._add_modeled_parse(response, shape, final_parsed) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 822, in _add_modeled_parse self._parse_payload(response, shape, member_shapes, final_parsed) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 862, in _parse_payload original_parsed = self._initial_body_parse(response['body']) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 948, in _initial_body_parse return self._parse_xml_string_to_dom(xml_string) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 454, in _parse_xml_string_to_dom raise ResponseParserError( botocore.parsers.ResponseParserError: Unable to parse response (no element found: line 2, column 0), invalid XML received. Further retries may succeed: b'<?xml version="1.0" encoding="UTF-8"?>\n' ``` </details>
process
🛑 processing failed responseparsererror overview responseparsererror found in processing task task during run ended on details flow name recovered inst dpc flcdrtd instrument recovered task name processing task error type responseparsererror error message unable to parse response no element found line column invalid xml received further retries may succeed b n traceback traceback most recent call last file srv conda envs notebook lib site packages botocore parsers py line in parse xml string to dom root parser close xml etree elementtree parseerror no element found line column during handling of the above exception another exception occurred traceback most recent call last file srv conda envs notebook lib site packages ooi harvester processor pipeline py line in processing final path finalize zarr file srv conda envs notebook lib site packages ooi harvester processor init py line in finalize zarr source store fs delete source store root recursive true file srv conda envs notebook lib site packages fsspec spec py line in delete return self rm path recursive recursive maxdepth maxdepth file srv conda envs notebook lib site packages fsspec asyn py line in wrapper return sync self loop func args kwargs file srv conda envs notebook lib site packages fsspec asyn py line in sync raise result file srv conda envs notebook lib site packages fsspec asyn py line in runner result await coro file srv conda envs notebook lib site packages core py line in rm await asyncio gather file srv conda envs notebook lib site packages core py line in bulk delete await self call delete objects kwargs bucket bucket delete delete keys file srv conda envs notebook lib site packages core py line in call raise err file srv conda envs notebook lib site packages core py line in call out await method additional kwargs file srv conda envs notebook lib site packages aiobotocore client py line in make api call http parsed response await self make request file srv conda envs notebook lib site packages aiobotocore client py line in make request return await self endpoint make request operation model request dict file srv conda envs notebook lib site packages aiobotocore endpoint py line in send request success response exception await self get response file srv conda envs notebook lib site packages aiobotocore endpoint py line in get response success response exception await self do get response file srv conda envs notebook lib site packages aiobotocore endpoint py line in do get response parsed response parser parse file srv conda envs notebook lib site packages botocore parsers py line in parse parsed self do parse response shape file srv conda envs notebook lib site packages botocore parsers py line in do parse self add modeled parse response shape final parsed file srv conda envs notebook lib site packages botocore parsers py line in add modeled parse self parse payload response shape member shapes final parsed file srv conda envs notebook lib site packages botocore parsers py line in parse payload original parsed self initial body parse response file srv conda envs notebook lib site packages botocore parsers py line in initial body parse return self parse xml string to dom xml string file srv conda envs notebook lib site packages botocore parsers py line in parse xml string to dom raise responseparsererror botocore parsers responseparsererror unable to parse response no element found line column invalid xml received further retries may succeed b n
1
11,668
14,530,486,439
IssuesEvent
2020-12-14 19:20:10
AdrianArnaiz/Brain-MRI-Autoencoder
https://api.github.com/repos/AdrianArnaiz/Brain-MRI-Autoencoder
closed
Keras Data Loader Development
data-preprocessing enhancement experiments
Create a Data Loader which loads images into model. It could make some preprocessing or data aumentation. Posibilities: * [**ImageDataGenerator I**](https://www.tensorflow.org/api_docs/python/tf/keras/preprocessing/image/ImageDataGenerator?version=nightly) * [How to use it (*Towards Data Science*)](https://towardsdatascience.com/keras-data-generators-and-how-to-use-them-b69129ed779c) * [**ImageDataGenerator II**](https://keras.io/api/preprocessing/image/) * [**Data Generator**](https://stanford.edu/~shervine/blog/keras-how-to-generate-data-on-the-fly)
1.0
Keras Data Loader Development - Create a Data Loader which loads images into model. It could make some preprocessing or data aumentation. Posibilities: * [**ImageDataGenerator I**](https://www.tensorflow.org/api_docs/python/tf/keras/preprocessing/image/ImageDataGenerator?version=nightly) * [How to use it (*Towards Data Science*)](https://towardsdatascience.com/keras-data-generators-and-how-to-use-them-b69129ed779c) * [**ImageDataGenerator II**](https://keras.io/api/preprocessing/image/) * [**Data Generator**](https://stanford.edu/~shervine/blog/keras-how-to-generate-data-on-the-fly)
process
keras data loader development create a data loader which loads images into model it could make some preprocessing or data aumentation posibilities
1
123,563
10,273,024,702
IssuesEvent
2019-08-23 18:06:09
dotnet/coreclr
https://api.github.com/repos/dotnet/coreclr
closed
Test failed: System.Net.WebSockets.Client.Tests.ConnectTest/ConnectAsync_NonStandardRequestHeaders_HeadersAddedWithoutValidation
arch-x64 os-linux test-corefx
**Job:** https://mc.dot.net/#/user/coreclr-corefx-jitstress/ci~2Fdotnet~2Fcoreclr~2Frefs~2Fheads~2Fmaster/test~2Ffunctional~2Fcorefx~2F/20190617.1/workItem/System.Net.WebSockets.Client.Tests/analysis/xunit/System.Net.WebSockets.Client.Tests.ConnectTest~2FConnectAsync_NonStandardRequestHeaders_HeadersAddedWithoutValidation **Failed tests:** System.Net.WebSockets.Client.Tests.ConnectTest/ConnectAsync_NonStandardRequestHeaders_HeadersAddedWithoutValidation **Log:** ``` System.Threading.Tasks.TaskCanceledException : A task was canceled. at System.Threading.Tasks.TaskTimeoutExtensions.WhenAllOrAnyFailed(Task[] tasks) in /_/src/Common/tests/System/Threading/Tasks/TaskTimeoutExtensions.cs:line 83 at System.Threading.Tasks.TaskTimeoutExtensions.WhenAllOrAnyFailed(Task[] tasks) in /_/src/Common/tests/System/Threading/Tasks/TaskTimeoutExtensions.cs:line 111 at System.Net.Test.Common.LoopbackServer.<>c__DisplayClass11_0.<<CreateClientAndServerAsync>b__0>d.MoveNext() in /_/src/Common/tests/System/Net/Http/LoopbackServer.cs:line 83 --- End of stack trace from previous location where exception was thrown --- at System.Net.Test.Common.LoopbackServer.CreateServerAsync(Func`2 funcAsync, Options options) in /_/src/Common/tests/System/Net/Http/LoopbackServer.cs:line 66 at System.Net.WebSockets.Client.Tests.ConnectTest.ConnectAsync_NonStandardRequestHeaders_HeadersAddedWithoutValidation() in /_/src/System.Net.WebSockets.Client/tests/ConnectTest.cs:line 237 --- End of stack trace from previous location where exception was thrown --- ```
1.0
Test failed: System.Net.WebSockets.Client.Tests.ConnectTest/ConnectAsync_NonStandardRequestHeaders_HeadersAddedWithoutValidation - **Job:** https://mc.dot.net/#/user/coreclr-corefx-jitstress/ci~2Fdotnet~2Fcoreclr~2Frefs~2Fheads~2Fmaster/test~2Ffunctional~2Fcorefx~2F/20190617.1/workItem/System.Net.WebSockets.Client.Tests/analysis/xunit/System.Net.WebSockets.Client.Tests.ConnectTest~2FConnectAsync_NonStandardRequestHeaders_HeadersAddedWithoutValidation **Failed tests:** System.Net.WebSockets.Client.Tests.ConnectTest/ConnectAsync_NonStandardRequestHeaders_HeadersAddedWithoutValidation **Log:** ``` System.Threading.Tasks.TaskCanceledException : A task was canceled. at System.Threading.Tasks.TaskTimeoutExtensions.WhenAllOrAnyFailed(Task[] tasks) in /_/src/Common/tests/System/Threading/Tasks/TaskTimeoutExtensions.cs:line 83 at System.Threading.Tasks.TaskTimeoutExtensions.WhenAllOrAnyFailed(Task[] tasks) in /_/src/Common/tests/System/Threading/Tasks/TaskTimeoutExtensions.cs:line 111 at System.Net.Test.Common.LoopbackServer.<>c__DisplayClass11_0.<<CreateClientAndServerAsync>b__0>d.MoveNext() in /_/src/Common/tests/System/Net/Http/LoopbackServer.cs:line 83 --- End of stack trace from previous location where exception was thrown --- at System.Net.Test.Common.LoopbackServer.CreateServerAsync(Func`2 funcAsync, Options options) in /_/src/Common/tests/System/Net/Http/LoopbackServer.cs:line 66 at System.Net.WebSockets.Client.Tests.ConnectTest.ConnectAsync_NonStandardRequestHeaders_HeadersAddedWithoutValidation() in /_/src/System.Net.WebSockets.Client/tests/ConnectTest.cs:line 237 --- End of stack trace from previous location where exception was thrown --- ```
non_process
test failed system net websockets client tests connecttest connectasync nonstandardrequestheaders headersaddedwithoutvalidation job failed tests system net websockets client tests connecttest connectasync nonstandardrequestheaders headersaddedwithoutvalidation log system threading tasks taskcanceledexception a task was canceled at system threading tasks tasktimeoutextensions whenalloranyfailed task tasks in src common tests system threading tasks tasktimeoutextensions cs line at system threading tasks tasktimeoutextensions whenalloranyfailed task tasks in src common tests system threading tasks tasktimeoutextensions cs line at system net test common loopbackserver c b d movenext in src common tests system net http loopbackserver cs line end of stack trace from previous location where exception was thrown at system net test common loopbackserver createserverasync func funcasync options options in src common tests system net http loopbackserver cs line at system net websockets client tests connecttest connectasync nonstandardrequestheaders headersaddedwithoutvalidation in src system net websockets client tests connecttest cs line end of stack trace from previous location where exception was thrown
0
1,933
4,762,103,959
IssuesEvent
2016-10-25 10:21:16
opentrials/opentrials
https://api.github.com/repos/opentrials/opentrials
closed
Refactor processors tests to use a real database
Processors
All tests for writers should use a real database similar to [the one used in `TestDocumentWriter`](https://github.com/opentrials/processors/blob/917a231e5b6e55c59418ad254102d78a63449cb2/tests/processors/base/writers/test_document.py#L140-L162).
1.0
Refactor processors tests to use a real database - All tests for writers should use a real database similar to [the one used in `TestDocumentWriter`](https://github.com/opentrials/processors/blob/917a231e5b6e55c59418ad254102d78a63449cb2/tests/processors/base/writers/test_document.py#L140-L162).
process
refactor processors tests to use a real database all tests for writers should use a real database similar to
1
5,525
8,381,048,704
IssuesEvent
2018-10-07 20:47:49
MichiganDataScienceTeam/googleanalytics
https://api.github.com/repos/MichiganDataScienceTeam/googleanalytics
opened
Preprocess: u'trafficSource.adwordsClickInfo.isVideoAd', u'trafficSource.adwordsClickInfo.page', u'trafficSource.adwordsClickInfo.slot',
easy preprocessing
Preprocess the following features: u'trafficSource.adwordsClickInfo.isVideoAd', u'trafficSource.adwordsClickInfo.page', u'trafficSource.adwordsClickInfo.slot', 1. Standardization: [http://scikit-learn.org/stable/modules/preprocessing.html#standardization-or-mean-removal-and-variance-scaling](http://scikit-learn.org/stable/modules/preprocessing.html#standardization-or-mean-removal-and-variance-scaling) 2. Impute missing values: [http://scikit-learn.org/stable/modules/impute.html](http://scikit-learn.org/stable/modules/impute.html) 3. Normalization: [http://scikit-learn.org/stable/modules/preprocessing.html#normalization](http://scikit-learn.org/stable/modules/preprocessing.html#normalization) 4. Encode categorical features (optional): [http://scikit-learn.org/stable/modules/preprocessing.html#encoding-categorical-features](http://scikit-learn.org/stable/modules/preprocessing.html#encoding-categorical-features) 5. Discretization (optional): [http://scikit-learn.org/stable/modules/preprocessing.html#discretization](http://scikit-learn.org/stable/modules/preprocessing.html#discretization) [http://scikit-learn.org/stable/modules/preprocessing.html](http://scikit-learn.org/stable/modules/preprocessing.html)
1.0
Preprocess: u'trafficSource.adwordsClickInfo.isVideoAd', u'trafficSource.adwordsClickInfo.page', u'trafficSource.adwordsClickInfo.slot', - Preprocess the following features: u'trafficSource.adwordsClickInfo.isVideoAd', u'trafficSource.adwordsClickInfo.page', u'trafficSource.adwordsClickInfo.slot', 1. Standardization: [http://scikit-learn.org/stable/modules/preprocessing.html#standardization-or-mean-removal-and-variance-scaling](http://scikit-learn.org/stable/modules/preprocessing.html#standardization-or-mean-removal-and-variance-scaling) 2. Impute missing values: [http://scikit-learn.org/stable/modules/impute.html](http://scikit-learn.org/stable/modules/impute.html) 3. Normalization: [http://scikit-learn.org/stable/modules/preprocessing.html#normalization](http://scikit-learn.org/stable/modules/preprocessing.html#normalization) 4. Encode categorical features (optional): [http://scikit-learn.org/stable/modules/preprocessing.html#encoding-categorical-features](http://scikit-learn.org/stable/modules/preprocessing.html#encoding-categorical-features) 5. Discretization (optional): [http://scikit-learn.org/stable/modules/preprocessing.html#discretization](http://scikit-learn.org/stable/modules/preprocessing.html#discretization) [http://scikit-learn.org/stable/modules/preprocessing.html](http://scikit-learn.org/stable/modules/preprocessing.html)
process
preprocess u trafficsource adwordsclickinfo isvideoad u trafficsource adwordsclickinfo page u trafficsource adwordsclickinfo slot preprocess the following features u trafficsource adwordsclickinfo isvideoad u trafficsource adwordsclickinfo page u trafficsource adwordsclickinfo slot standardization impute missing values normalization encode categorical features optional discretization optional
1
99,345
12,419,040,404
IssuesEvent
2020-05-23 03:32:54
masakudamatsu/line-height-picker
https://api.github.com/repos/masakudamatsu/line-height-picker
closed
Style the font name display, sample paragraphs and CSS code
design
These three elements are the most important feedback to the user. We apply the linear light at the top and bottom of these elements. The linear light is defined as: ``` background: hsla(0, 0%, 100%, 0.9); box-shadow: 0 0 10px 0 hsla(0, 0%, 100%, 0.9), 0 0 20px 0 hsla(0, 0%, 100%, 0.9), 0 0 40px 0 hsla(0, 0%, 100%, 0.9); height: 1px; width: 100%; ``` Previously, we tried the cove lighting effect: https://codepen.io/masakudamatsu/pen/wvKjLZY - But this involves the complicated logic of the use of z-index. - Light spills to the sides of a page; so the sliding page transition (#194 ) reveals it.
1.0
Style the font name display, sample paragraphs and CSS code - These three elements are the most important feedback to the user. We apply the linear light at the top and bottom of these elements. The linear light is defined as: ``` background: hsla(0, 0%, 100%, 0.9); box-shadow: 0 0 10px 0 hsla(0, 0%, 100%, 0.9), 0 0 20px 0 hsla(0, 0%, 100%, 0.9), 0 0 40px 0 hsla(0, 0%, 100%, 0.9); height: 1px; width: 100%; ``` Previously, we tried the cove lighting effect: https://codepen.io/masakudamatsu/pen/wvKjLZY - But this involves the complicated logic of the use of z-index. - Light spills to the sides of a page; so the sliding page transition (#194 ) reveals it.
non_process
style the font name display sample paragraphs and css code these three elements are the most important feedback to the user we apply the linear light at the top and bottom of these elements the linear light is defined as background hsla box shadow hsla hsla hsla height width previously we tried the cove lighting effect but this involves the complicated logic of the use of z index light spills to the sides of a page so the sliding page transition reveals it
0
63,280
12,299,928,340
IssuesEvent
2020-05-11 13:13:31
AlexGidge/SkiPenguin
https://api.github.com/repos/AlexGidge/SkiPenguin
closed
Move previous & next level logic into the JSON data
2 Code Improvement
Instead of calculating previous/next, first/last
1.0
Move previous & next level logic into the JSON data - Instead of calculating previous/next, first/last
non_process
move previous next level logic into the json data instead of calculating previous next first last
0
13,624
16,237,670,539
IssuesEvent
2021-05-07 04:18:19
ooi-data/CE04OSPD-DP01B-01-CTDPFL105-recovered_inst-dpc_ctd_instrument_recovered
https://api.github.com/repos/ooi-data/CE04OSPD-DP01B-01-CTDPFL105-recovered_inst-dpc_ctd_instrument_recovered
opened
🛑 Processing failed: ResponseParserError
process
## Overview `ResponseParserError` found in `processing_task` task during run ended on 2021-05-07T04:18:19.229808. ## Details Flow name: `CE04OSPD-DP01B-01-CTDPFL105-recovered_inst-dpc_ctd_instrument_recovered` Task name: `processing_task` Error type: `ResponseParserError` Error message: Unable to parse response (no element found: line 2, column 0), invalid XML received. Further retries may succeed: b'<?xml version="1.0" encoding="UTF-8"?>\n' <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 452, in _parse_xml_string_to_dom root = parser.close() File "<string>", line None xml.etree.ElementTree.ParseError: no element found: line 2, column 0 During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/share/miniconda/envs/harvester/lib/python3.8/site-packages/ooi_harvester/processor/pipeline.py", line 71, in processing_task File "/srv/conda/envs/notebook/lib/python3.8/site-packages/ooi_harvester/processor/__init__.py", line 311, in finalize_zarr source_store.fs.delete(source_store.root, recursive=True) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/fsspec/spec.py", line 1151, in delete return self.rm(path, recursive=recursive, maxdepth=maxdepth) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/fsspec/asyn.py", line 72, in wrapper return sync(self.loop, func, *args, **kwargs) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/fsspec/asyn.py", line 53, in sync raise result[0] File "/srv/conda/envs/notebook/lib/python3.8/site-packages/fsspec/asyn.py", line 20, in _runner result[0] = await coro File "/srv/conda/envs/notebook/lib/python3.8/site-packages/s3fs/core.py", line 1510, in _rm await asyncio.gather( File "/srv/conda/envs/notebook/lib/python3.8/site-packages/s3fs/core.py", line 1498, in _bulk_delete await self._call_s3("delete_objects", kwargs, Bucket=bucket, Delete=delete_keys) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/s3fs/core.py", line 252, in _call_s3 raise translate_boto_error(err) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/s3fs/core.py", line 233, in _call_s3 out = await method(**additional_kwargs) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/aiobotocore/client.py", line 140, in _make_api_call http, parsed_response = await self._make_request( File "/srv/conda/envs/notebook/lib/python3.8/site-packages/aiobotocore/client.py", line 160, in _make_request return await self._endpoint.make_request(operation_model, request_dict) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/aiobotocore/endpoint.py", line 101, in _send_request success_response, exception = await self._get_response( File "/srv/conda/envs/notebook/lib/python3.8/site-packages/aiobotocore/endpoint.py", line 120, in _get_response success_response, exception = await self._do_get_response( File "/srv/conda/envs/notebook/lib/python3.8/site-packages/aiobotocore/endpoint.py", line 180, in _do_get_response parsed_response = parser.parse( File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 245, in parse parsed = self._do_parse(response, shape) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 809, in _do_parse self._add_modeled_parse(response, shape, final_parsed) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 818, in _add_modeled_parse self._parse_payload(response, shape, member_shapes, final_parsed) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 858, in _parse_payload original_parsed = self._initial_body_parse(response['body']) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 944, in _initial_body_parse return self._parse_xml_string_to_dom(xml_string) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 454, in _parse_xml_string_to_dom raise ResponseParserError( botocore.parsers.ResponseParserError: Unable to parse response (no element found: line 2, column 0), invalid XML received. Further retries may succeed: b'<?xml version="1.0" encoding="UTF-8"?>\n' ``` </details>
1.0
🛑 Processing failed: ResponseParserError - ## Overview `ResponseParserError` found in `processing_task` task during run ended on 2021-05-07T04:18:19.229808. ## Details Flow name: `CE04OSPD-DP01B-01-CTDPFL105-recovered_inst-dpc_ctd_instrument_recovered` Task name: `processing_task` Error type: `ResponseParserError` Error message: Unable to parse response (no element found: line 2, column 0), invalid XML received. Further retries may succeed: b'<?xml version="1.0" encoding="UTF-8"?>\n' <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 452, in _parse_xml_string_to_dom root = parser.close() File "<string>", line None xml.etree.ElementTree.ParseError: no element found: line 2, column 0 During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/share/miniconda/envs/harvester/lib/python3.8/site-packages/ooi_harvester/processor/pipeline.py", line 71, in processing_task File "/srv/conda/envs/notebook/lib/python3.8/site-packages/ooi_harvester/processor/__init__.py", line 311, in finalize_zarr source_store.fs.delete(source_store.root, recursive=True) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/fsspec/spec.py", line 1151, in delete return self.rm(path, recursive=recursive, maxdepth=maxdepth) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/fsspec/asyn.py", line 72, in wrapper return sync(self.loop, func, *args, **kwargs) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/fsspec/asyn.py", line 53, in sync raise result[0] File "/srv/conda/envs/notebook/lib/python3.8/site-packages/fsspec/asyn.py", line 20, in _runner result[0] = await coro File "/srv/conda/envs/notebook/lib/python3.8/site-packages/s3fs/core.py", line 1510, in _rm await asyncio.gather( File "/srv/conda/envs/notebook/lib/python3.8/site-packages/s3fs/core.py", line 1498, in _bulk_delete await self._call_s3("delete_objects", kwargs, Bucket=bucket, Delete=delete_keys) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/s3fs/core.py", line 252, in _call_s3 raise translate_boto_error(err) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/s3fs/core.py", line 233, in _call_s3 out = await method(**additional_kwargs) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/aiobotocore/client.py", line 140, in _make_api_call http, parsed_response = await self._make_request( File "/srv/conda/envs/notebook/lib/python3.8/site-packages/aiobotocore/client.py", line 160, in _make_request return await self._endpoint.make_request(operation_model, request_dict) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/aiobotocore/endpoint.py", line 101, in _send_request success_response, exception = await self._get_response( File "/srv/conda/envs/notebook/lib/python3.8/site-packages/aiobotocore/endpoint.py", line 120, in _get_response success_response, exception = await self._do_get_response( File "/srv/conda/envs/notebook/lib/python3.8/site-packages/aiobotocore/endpoint.py", line 180, in _do_get_response parsed_response = parser.parse( File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 245, in parse parsed = self._do_parse(response, shape) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 809, in _do_parse self._add_modeled_parse(response, shape, final_parsed) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 818, in _add_modeled_parse self._parse_payload(response, shape, member_shapes, final_parsed) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 858, in _parse_payload original_parsed = self._initial_body_parse(response['body']) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 944, in _initial_body_parse return self._parse_xml_string_to_dom(xml_string) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 454, in _parse_xml_string_to_dom raise ResponseParserError( botocore.parsers.ResponseParserError: Unable to parse response (no element found: line 2, column 0), invalid XML received. Further retries may succeed: b'<?xml version="1.0" encoding="UTF-8"?>\n' ``` </details>
process
🛑 processing failed responseparsererror overview responseparsererror found in processing task task during run ended on details flow name recovered inst dpc ctd instrument recovered task name processing task error type responseparsererror error message unable to parse response no element found line column invalid xml received further retries may succeed b n traceback traceback most recent call last file srv conda envs notebook lib site packages botocore parsers py line in parse xml string to dom root parser close file line none xml etree elementtree parseerror no element found line column during handling of the above exception another exception occurred traceback most recent call last file usr share miniconda envs harvester lib site packages ooi harvester processor pipeline py line in processing task file srv conda envs notebook lib site packages ooi harvester processor init py line in finalize zarr source store fs delete source store root recursive true file srv conda envs notebook lib site packages fsspec spec py line in delete return self rm path recursive recursive maxdepth maxdepth file srv conda envs notebook lib site packages fsspec asyn py line in wrapper return sync self loop func args kwargs file srv conda envs notebook lib site packages fsspec asyn py line in sync raise result file srv conda envs notebook lib site packages fsspec asyn py line in runner result await coro file srv conda envs notebook lib site packages core py line in rm await asyncio gather file srv conda envs notebook lib site packages core py line in bulk delete await self call delete objects kwargs bucket bucket delete delete keys file srv conda envs notebook lib site packages core py line in call raise translate boto error err file srv conda envs notebook lib site packages core py line in call out await method additional kwargs file srv conda envs notebook lib site packages aiobotocore client py line in make api call http parsed response await self make request file srv conda envs notebook lib site packages aiobotocore client py line in make request return await self endpoint make request operation model request dict file srv conda envs notebook lib site packages aiobotocore endpoint py line in send request success response exception await self get response file srv conda envs notebook lib site packages aiobotocore endpoint py line in get response success response exception await self do get response file srv conda envs notebook lib site packages aiobotocore endpoint py line in do get response parsed response parser parse file srv conda envs notebook lib site packages botocore parsers py line in parse parsed self do parse response shape file srv conda envs notebook lib site packages botocore parsers py line in do parse self add modeled parse response shape final parsed file srv conda envs notebook lib site packages botocore parsers py line in add modeled parse self parse payload response shape member shapes final parsed file srv conda envs notebook lib site packages botocore parsers py line in parse payload original parsed self initial body parse response file srv conda envs notebook lib site packages botocore parsers py line in initial body parse return self parse xml string to dom xml string file srv conda envs notebook lib site packages botocore parsers py line in parse xml string to dom raise responseparsererror botocore parsers responseparsererror unable to parse response no element found line column invalid xml received further retries may succeed b n
1
2,867
5,825,219,825
IssuesEvent
2017-05-07 19:38:33
sysown/proxysql
https://api.github.com/repos/sysown/proxysql
closed
Support for PCRE
QUERY PROCESSOR
It would be interesting to have support for Perl regexp (pcre) instead of the current implementation. This will help for creating datamasking rules.
1.0
Support for PCRE - It would be interesting to have support for Perl regexp (pcre) instead of the current implementation. This will help for creating datamasking rules.
process
support for pcre it would be interesting to have support for perl regexp pcre instead of the current implementation this will help for creating datamasking rules
1
4,364
7,260,514,535
IssuesEvent
2018-02-18 10:53:36
qgis/QGIS-Documentation
https://api.github.com/repos/qgis/QGIS-Documentation
closed
[FEATURE][processing] Remove duplicate create points along lines script
Automatic new feature Processing
Original commit: https://github.com/qgis/QGIS/commit/6bb4934907ea26bcff802bb4ae9abbb47eae7d5f by nyalldawson This functionality is duplicated by the PointsAlongGeometry algorithm Keep the algorithm version, since it has unit tests and supports things like progress reports. (marked as feature for inclusion in release notes)
1.0
[FEATURE][processing] Remove duplicate create points along lines script - Original commit: https://github.com/qgis/QGIS/commit/6bb4934907ea26bcff802bb4ae9abbb47eae7d5f by nyalldawson This functionality is duplicated by the PointsAlongGeometry algorithm Keep the algorithm version, since it has unit tests and supports things like progress reports. (marked as feature for inclusion in release notes)
process
remove duplicate create points along lines script original commit by nyalldawson this functionality is duplicated by the pointsalonggeometry algorithm keep the algorithm version since it has unit tests and supports things like progress reports marked as feature for inclusion in release notes
1
7,156
10,307,392,937
IssuesEvent
2019-08-29 08:37:12
vaerilius/angular8-course
https://api.github.com/repos/vaerilius/angular8-course
closed
Section 22: Angular Modules & Optimizing Angular Apps
inProcess
- [x] 317. Module Introduction - [x] 318. What are Modules? - [x] 319. Analyzing the AppModule - [x] 320. Getting Started with Feature Modules - [x] 321. Splitting Modules Correctly - [x] 322. Adding Routes to Feature Modules - [x] 323. Component Declarations - [x] 324. The ShoppingList Feature Module - [x] 325. Understanding Shared Modules - [x] 326. Understanding the Core Module - [x] 327. Adding an Auth Feature Module - [x] 328. Understanding Lazy Loading - [x] 329. Implementing Lazy Loading - [x] 330. Alternative Lazy Loading Syntax - [x] 331. More Lazy Loading - [x] 332. Preloading Lazy-Loaded Code - [x] 333. Modules & Services - [x] 334. Loading Services Differently - [x] 335. Ahead-of-Time Compilation - [x] 336. Wrap Up - [x] 337. Useful Resources & Links
1.0
Section 22: Angular Modules & Optimizing Angular Apps - - [x] 317. Module Introduction - [x] 318. What are Modules? - [x] 319. Analyzing the AppModule - [x] 320. Getting Started with Feature Modules - [x] 321. Splitting Modules Correctly - [x] 322. Adding Routes to Feature Modules - [x] 323. Component Declarations - [x] 324. The ShoppingList Feature Module - [x] 325. Understanding Shared Modules - [x] 326. Understanding the Core Module - [x] 327. Adding an Auth Feature Module - [x] 328. Understanding Lazy Loading - [x] 329. Implementing Lazy Loading - [x] 330. Alternative Lazy Loading Syntax - [x] 331. More Lazy Loading - [x] 332. Preloading Lazy-Loaded Code - [x] 333. Modules & Services - [x] 334. Loading Services Differently - [x] 335. Ahead-of-Time Compilation - [x] 336. Wrap Up - [x] 337. Useful Resources & Links
process
section angular modules optimizing angular apps module introduction what are modules analyzing the appmodule getting started with feature modules splitting modules correctly adding routes to feature modules component declarations the shoppinglist feature module understanding shared modules understanding the core module adding an auth feature module understanding lazy loading implementing lazy loading alternative lazy loading syntax more lazy loading preloading lazy loaded code modules services loading services differently ahead of time compilation wrap up useful resources links
1
263,816
23,083,685,432
IssuesEvent
2022-07-26 09:28:17
Tencent/bk-ci
https://api.github.com/repos/Tencent/bk-ci
closed
feat: 【流水线插件SDK】权限控制机制接口增加缓存
for gray for test kind/feat/tech kind/enhancement area/ci/backend tested streams/for gray streams/grayed streams/done area/stream/backend
**What would you like to be added**: 关联需求 #3349 **Why is this needed**: 目前插件调用敏感接口,会验证权限,但是接口访问量很大,访问的数据基本都是不变的,希望可以增加本地缓存,减少接口的调用量
2.0
feat: 【流水线插件SDK】权限控制机制接口增加缓存 - **What would you like to be added**: 关联需求 #3349 **Why is this needed**: 目前插件调用敏感接口,会验证权限,但是接口访问量很大,访问的数据基本都是不变的,希望可以增加本地缓存,减少接口的调用量
non_process
feat 【流水线插件sdk】权限控制机制接口增加缓存 what would you like to be added 关联需求 why is this needed 目前插件调用敏感接口,会验证权限,但是接口访问量很大,访问的数据基本都是不变的,希望可以增加本地缓存,减少接口的调用量
0
22,511
31,563,172,073
IssuesEvent
2023-09-03 13:52:46
h4sh5/npm-auto-scanner
https://api.github.com/repos/h4sh5/npm-auto-scanner
opened
create-creta 0.5.2 has 1 guarddog issues
npm-silent-process-execution
```{"npm-silent-process-execution":[{"code":"\t\tconst child = spawn(\n\t\t\tpath.join(EXE_DIR, 'updater.exe'),\n\t\t\t['-p', `${process.pid}`, '-e', EXE_PATH],\n\t\t\t{\n\t\t\t\tdetached: true,\n\t\t\t\tcwd: EXE_DIR,\n\t\t\t\tstdio: 'ignore',\n\t\t\t}\n\t\t);","location":"package/templates/default/src/main/service/updateService/win32.ts:19","message":"This package is silently executing another executable"}]}```
1.0
create-creta 0.5.2 has 1 guarddog issues - ```{"npm-silent-process-execution":[{"code":"\t\tconst child = spawn(\n\t\t\tpath.join(EXE_DIR, 'updater.exe'),\n\t\t\t['-p', `${process.pid}`, '-e', EXE_PATH],\n\t\t\t{\n\t\t\t\tdetached: true,\n\t\t\t\tcwd: EXE_DIR,\n\t\t\t\tstdio: 'ignore',\n\t\t\t}\n\t\t);","location":"package/templates/default/src/main/service/updateService/win32.ts:19","message":"This package is silently executing another executable"}]}```
process
create creta has guarddog issues npm silent process execution n t t t n t t t tdetached true n t t t tcwd exe dir n t t t tstdio ignore n t t t n t t location package templates default src main service updateservice ts message this package is silently executing another executable
1
11,384
14,222,927,440
IssuesEvent
2020-11-17 17:30:32
unicode-org/icu4x
https://api.github.com/repos/unicode-org/icu4x
closed
Add CI copyright headers check
C-process T-task
This task is split off from #84 to represent just the requirement of doing checks on copyright headers on source files. We want every file to have headers indicating that indicates that the file is a part of ICU4X and references a link to the license. Anything else I'm missing? @hsivonen
1.0
Add CI copyright headers check - This task is split off from #84 to represent just the requirement of doing checks on copyright headers on source files. We want every file to have headers indicating that indicates that the file is a part of ICU4X and references a link to the license. Anything else I'm missing? @hsivonen
process
add ci copyright headers check this task is split off from to represent just the requirement of doing checks on copyright headers on source files we want every file to have headers indicating that indicates that the file is a part of and references a link to the license anything else i m missing hsivonen
1
185,011
21,785,051,196
IssuesEvent
2022-05-14 02:17:59
directoryxx/Laravel-Jenkins-Docker
https://api.github.com/repos/directoryxx/Laravel-Jenkins-Docker
closed
CVE-2018-19826 (Medium) detected in node-sass-v4.11.0 - autoclosed
security vulnerability
## CVE-2018-19826 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-sassv4.11.0</b></p></summary> <p> <p>:rainbow: Node.js bindings to libsass</p> <p>Library home page: <a href=https://github.com/sass/node-sass.git>https://github.com/sass/node-sass.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/directoryxx/Laravel-Jenkins-Docker/commit/4e22a0a83e7c15832599b09020dba4934c4389d3">4e22a0a83e7c15832599b09020dba4934c4389d3</a></p> </p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Library Source Files (4)</summary> <p></p> <p> * The source files were matched to this source library based on a best effort match. Source libraries are selected from a list of probable public libraries.</p> <p> - /Laravel-Jenkins-Docker/node_modules/node-sass/src/binding.cpp - /Laravel-Jenkins-Docker/node_modules/node-sass/src/libsass/src/inspect.cpp - /Laravel-Jenkins-Docker/node_modules/node-sass/src/libsass/src/operators.cpp - /Laravel-Jenkins-Docker/node_modules/node-sass/src/libsass/src/parser.cpp </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> ** DISPUTED ** In inspect.cpp in LibSass 3.5.5, a high memory footprint caused by an endless loop (containing a Sass::Inspect::operator()(Sass::String_Quoted*) stack frame) may cause a Denial of Service via crafted sass input files with stray '&' or '/' characters. NOTE: Upstream comments indicate this issue is closed as "won't fix" and "works as intended" by design. <p>Publish Date: 2018-12-03 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19826>CVE-2018-19826</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-19826 (Medium) detected in node-sass-v4.11.0 - autoclosed - ## CVE-2018-19826 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-sassv4.11.0</b></p></summary> <p> <p>:rainbow: Node.js bindings to libsass</p> <p>Library home page: <a href=https://github.com/sass/node-sass.git>https://github.com/sass/node-sass.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/directoryxx/Laravel-Jenkins-Docker/commit/4e22a0a83e7c15832599b09020dba4934c4389d3">4e22a0a83e7c15832599b09020dba4934c4389d3</a></p> </p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Library Source Files (4)</summary> <p></p> <p> * The source files were matched to this source library based on a best effort match. Source libraries are selected from a list of probable public libraries.</p> <p> - /Laravel-Jenkins-Docker/node_modules/node-sass/src/binding.cpp - /Laravel-Jenkins-Docker/node_modules/node-sass/src/libsass/src/inspect.cpp - /Laravel-Jenkins-Docker/node_modules/node-sass/src/libsass/src/operators.cpp - /Laravel-Jenkins-Docker/node_modules/node-sass/src/libsass/src/parser.cpp </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> ** DISPUTED ** In inspect.cpp in LibSass 3.5.5, a high memory footprint caused by an endless loop (containing a Sass::Inspect::operator()(Sass::String_Quoted*) stack frame) may cause a Denial of Service via crafted sass input files with stray '&' or '/' characters. NOTE: Upstream comments indicate this issue is closed as "won't fix" and "works as intended" by design. <p>Publish Date: 2018-12-03 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19826>CVE-2018-19826</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in node sass autoclosed cve medium severity vulnerability vulnerable library node rainbow node js bindings to libsass library home page a href found in head commit a href library source files the source files were matched to this source library based on a best effort match source libraries are selected from a list of probable public libraries laravel jenkins docker node modules node sass src binding cpp laravel jenkins docker node modules node sass src libsass src inspect cpp laravel jenkins docker node modules node sass src libsass src operators cpp laravel jenkins docker node modules node sass src libsass src parser cpp vulnerability details disputed in inspect cpp in libsass a high memory footprint caused by an endless loop containing a sass inspect operator sass string quoted stack frame may cause a denial of service via crafted sass input files with stray or characters note upstream comments indicate this issue is closed as won t fix and works as intended by design publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href step up your open source security game with whitesource
0
18,246
24,323,960,997
IssuesEvent
2022-09-30 13:18:21
km4ack/patmenu2
https://api.github.com/repos/km4ack/patmenu2
closed
Manage Pat Menu - PAT Menu 2.10.0 Issues
bug in process
Under Manage Pat: Pat Login-Logout menu: (1)If Cancel is selected, the application does not return to the preceding menu level, Pat Menu disappears and has to be reselected from the Desktop Menu. (2) It does not show current settings so you are not sure what you are changing. GPS Grid Update: If this is selected I get the message "Ruby needed to read GPS..." I believe that I already had installed Ruby as conky is getting GPS and grid updates. None the less I allowed ruby to install. I had intentionally manually entered an incorrect grid square and following the ruby update it corrected the Winlink grid square. The same issue repeats if I again select the GPS Grid Update. Manual Grid Update: Does not display current setting so you are not sure what will be changed.
1.0
Manage Pat Menu - PAT Menu 2.10.0 Issues - Under Manage Pat: Pat Login-Logout menu: (1)If Cancel is selected, the application does not return to the preceding menu level, Pat Menu disappears and has to be reselected from the Desktop Menu. (2) It does not show current settings so you are not sure what you are changing. GPS Grid Update: If this is selected I get the message "Ruby needed to read GPS..." I believe that I already had installed Ruby as conky is getting GPS and grid updates. None the less I allowed ruby to install. I had intentionally manually entered an incorrect grid square and following the ruby update it corrected the Winlink grid square. The same issue repeats if I again select the GPS Grid Update. Manual Grid Update: Does not display current setting so you are not sure what will be changed.
process
manage pat menu pat menu issues under manage pat pat login logout menu if cancel is selected the application does not return to the preceding menu level pat menu disappears and has to be reselected from the desktop menu it does not show current settings so you are not sure what you are changing gps grid update if this is selected i get the message ruby needed to read gps i believe that i already had installed ruby as conky is getting gps and grid updates none the less i allowed ruby to install i had intentionally manually entered an incorrect grid square and following the ruby update it corrected the winlink grid square the same issue repeats if i again select the gps grid update manual grid update does not display current setting so you are not sure what will be changed
1
158,810
12,425,709,989
IssuesEvent
2020-05-24 17:36:00
ethereum/go-ethereum
https://api.github.com/repos/ethereum/go-ethereum
closed
export transaction evm trace via test rpc
retesteth
geth retesteth during the test creation/debug it is required to have evm trace to see what is happening with memory/storage and which instructions are called. we need a client to export that log via test RPC Martin mentioned that there is a software console GUI to view that log, so following that format would be great. since the log might be too huge, the method should ask for a range of evm steps. Something like this: `test_traceTransaction( blockNumber, transactionHash, evmPosStart, evmPosLength)` this is essential to new state test development as the debug is a key feature for test development the evm debug flag might be transfered in set_chainParams method
1.0
export transaction evm trace via test rpc - geth retesteth during the test creation/debug it is required to have evm trace to see what is happening with memory/storage and which instructions are called. we need a client to export that log via test RPC Martin mentioned that there is a software console GUI to view that log, so following that format would be great. since the log might be too huge, the method should ask for a range of evm steps. Something like this: `test_traceTransaction( blockNumber, transactionHash, evmPosStart, evmPosLength)` this is essential to new state test development as the debug is a key feature for test development the evm debug flag might be transfered in set_chainParams method
non_process
export transaction evm trace via test rpc geth retesteth during the test creation debug it is required to have evm trace to see what is happening with memory storage and which instructions are called we need a client to export that log via test rpc martin mentioned that there is a software console gui to view that log so following that format would be great since the log might be too huge the method should ask for a range of evm steps something like this test tracetransaction blocknumber transactionhash evmposstart evmposlength this is essential to new state test development as the debug is a key feature for test development the evm debug flag might be transfered in set chainparams method
0
7,040
10,197,458,704
IssuesEvent
2019-08-13 00:28:11
Ultimate-Hosts-Blacklist/whitelist
https://api.github.com/repos/Ultimate-Hosts-Blacklist/whitelist
closed
gp.se and www.gp.se
whitelisting process
@denizdogan [said](https://github.com/mitchellkrogza/Ultimate.Hosts.Blacklist/issues/518): > This is the biggest newspaper in Gothenburg, Sweden and should not be blocked.
1.0
gp.se and www.gp.se - @denizdogan [said](https://github.com/mitchellkrogza/Ultimate.Hosts.Blacklist/issues/518): > This is the biggest newspaper in Gothenburg, Sweden and should not be blocked.
process
gp se and denizdogan this is the biggest newspaper in gothenburg sweden and should not be blocked
1
7,920
11,098,040,264
IssuesEvent
2019-12-16 14:30:13
kubeflow/kfctl
https://api.github.com/repos/kubeflow/kfctl
opened
Migrate C2D code from kubeflow/kubeflow into kubeflow/kfctl
area/c2d area/kfctl kind/process priority/p2
Should we migrate the click to deploy code (C2D) from kubeflow/kubeflow to kubeflow/kfctl? The C2D code is here: https://github.com/kubeflow/kubeflow/tree/master/bootstrap https://github.com/kubeflow/kubeflow/tree/master/components/gcp-click-to-deploy I think it makes sense to move into kubeflow/kfctl so all the Kubeflow control plane code is in one place. @kunmingg What do you think? Related to: kubeflow/kfctl#7
1.0
Migrate C2D code from kubeflow/kubeflow into kubeflow/kfctl - Should we migrate the click to deploy code (C2D) from kubeflow/kubeflow to kubeflow/kfctl? The C2D code is here: https://github.com/kubeflow/kubeflow/tree/master/bootstrap https://github.com/kubeflow/kubeflow/tree/master/components/gcp-click-to-deploy I think it makes sense to move into kubeflow/kfctl so all the Kubeflow control plane code is in one place. @kunmingg What do you think? Related to: kubeflow/kfctl#7
process
migrate code from kubeflow kubeflow into kubeflow kfctl should we migrate the click to deploy code from kubeflow kubeflow to kubeflow kfctl the code is here i think it makes sense to move into kubeflow kfctl so all the kubeflow control plane code is in one place kunmingg what do you think related to kubeflow kfctl
1
161,396
12,542,977,098
IssuesEvent
2020-06-05 14:50:33
flutter/flutter
https://api.github.com/repos/flutter/flutter
opened
[e2e] Make it easier for an e2e test to obtain the timeline
P3 a: tests p: e2e
For example, flutter_driver has ways to obtain the timeline for a specific block of code. Package:e2e should offer similar functionality. More generally, it should offer functionality to measure performance metrics over the course of an integration test.
1.0
[e2e] Make it easier for an e2e test to obtain the timeline - For example, flutter_driver has ways to obtain the timeline for a specific block of code. Package:e2e should offer similar functionality. More generally, it should offer functionality to measure performance metrics over the course of an integration test.
non_process
make it easier for an test to obtain the timeline for example flutter driver has ways to obtain the timeline for a specific block of code package should offer similar functionality more generally it should offer functionality to measure performance metrics over the course of an integration test
0
16,781
21,967,016,196
IssuesEvent
2022-05-24 21:29:54
GoogleCloudPlatform/emblem
https://api.github.com/repos/GoogleCloudPlatform/emblem
opened
Organize Maintainer Environments into 3-project mode
type: process priority: p1 component: delivery
## Objective The current model of our deployed environments evolved organically from early project deployment. We want to standardize this to follow the three-project mode. There are two paths forward: * Manually clean up and align to the 3-project approach * Get terraform working as expected and use it as part of a migration to more formal management I will edit more details into this issue later this week, filing this as a planning placeholder.
1.0
Organize Maintainer Environments into 3-project mode - ## Objective The current model of our deployed environments evolved organically from early project deployment. We want to standardize this to follow the three-project mode. There are two paths forward: * Manually clean up and align to the 3-project approach * Get terraform working as expected and use it as part of a migration to more formal management I will edit more details into this issue later this week, filing this as a planning placeholder.
process
organize maintainer environments into project mode objective the current model of our deployed environments evolved organically from early project deployment we want to standardize this to follow the three project mode there are two paths forward manually clean up and align to the project approach get terraform working as expected and use it as part of a migration to more formal management i will edit more details into this issue later this week filing this as a planning placeholder
1
3,479
6,553,090,366
IssuesEvent
2017-09-05 20:59:08
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
closed
MP: GO:0090615 mitochondrial mRNA processing
low priority PomBase RNA processes
GO:0090615 mitochondrial mRNA processing is defined Steps involved in processing precursor RNAs arising from transcription of operons in the mitochondrial genome into mature mRNAs. should it be a descendent of GO:0031426 polycistronic mRNA processing The conversion of a primary mRNA transcript containing more than one complete protein-coding region into individual mature mRNA molecules. ?
1.0
MP: GO:0090615 mitochondrial mRNA processing - GO:0090615 mitochondrial mRNA processing is defined Steps involved in processing precursor RNAs arising from transcription of operons in the mitochondrial genome into mature mRNAs. should it be a descendent of GO:0031426 polycistronic mRNA processing The conversion of a primary mRNA transcript containing more than one complete protein-coding region into individual mature mRNA molecules. ?
process
mp go mitochondrial mrna processing go mitochondrial mrna processing is defined steps involved in processing precursor rnas arising from transcription of operons in the mitochondrial genome into mature mrnas should it be a descendent of go polycistronic mrna processing the conversion of a primary mrna transcript containing more than one complete protein coding region into individual mature mrna molecules
1
189,473
15,189,381,507
IssuesEvent
2021-02-15 16:20:11
ixpantia/ixpantia.introR
https://api.github.com/repos/ixpantia/ixpantia.introR
closed
Generar vignette introductorio sobre cómo preparar la computadora para la instalación del paquete
documentation
Necesitamos documentación para usuarios completamente nuevos a R que necesiten seguir un conjunto de instrucciones con imágenes sobre cómo realizar la instalación del paquete mismo. Para esto necesitan familiarizarse con la interfaz de RStudio, saber dónde escribir el comando de instalación y demás. UN vignette que contenga este tipo de ayuda será una gran contribución para aquellos usuarios que por primera vez conocen R y antes del curso deben de instalar el paquete.
1.0
Generar vignette introductorio sobre cómo preparar la computadora para la instalación del paquete - Necesitamos documentación para usuarios completamente nuevos a R que necesiten seguir un conjunto de instrucciones con imágenes sobre cómo realizar la instalación del paquete mismo. Para esto necesitan familiarizarse con la interfaz de RStudio, saber dónde escribir el comando de instalación y demás. UN vignette que contenga este tipo de ayuda será una gran contribución para aquellos usuarios que por primera vez conocen R y antes del curso deben de instalar el paquete.
non_process
generar vignette introductorio sobre cómo preparar la computadora para la instalación del paquete necesitamos documentación para usuarios completamente nuevos a r que necesiten seguir un conjunto de instrucciones con imágenes sobre cómo realizar la instalación del paquete mismo para esto necesitan familiarizarse con la interfaz de rstudio saber dónde escribir el comando de instalación y demás un vignette que contenga este tipo de ayuda será una gran contribución para aquellos usuarios que por primera vez conocen r y antes del curso deben de instalar el paquete
0
13,546
16,089,120,200
IssuesEvent
2021-04-26 14:42:59
MicrosoftDocs/azure-devops-docs
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
closed
The example doesn't seem to update the Build.BuildNumber
Pri2 cba devops-cicd-process/tech devops/prod support-request
I'm trying to create a nuget package numbering in a yaml file but whatever i try the build number is not chaning. Even the given example ``` name: $(TeamProject)_$(Build.DefinitionName)_$(SourceBranchName)_$(Date:yyyyMMdd)$(Rev:.r) steps: - script: echo '$(Build.BuildNumber)' # outputs customized build number like project_def_master_20200828.1 ``` Just returns the defauld YYYY.nn back = > '20210425.44' I've no cuw what I'm missing could you give me soe feedback? I just what to do: ``` variables: name: '1.0.0.$(Build.BuildId)' MyRunNumber: '1.0.0.$(Build.BuildId)' steps: - script: echo '$(Build.BuildNumber)' #is wrong - script: echo '$(MyRunNumber)' #is ok - script: echo '$(name)' #is ok .... some other stuff and not sure i need this - task: Assembly-Info-NetCore@2 inputs: Path: 'scr/xxx/xxxx.csproj' FileNames: '**/*.csproj' InsertAttributes: true FileEncoding: 'auto' WriteBOM: false VersionNumber: $(name) FileVersionNumber: $(name) PackageVersion: $(name) ... than the error command - task: NuGetCommand@2 inputs: command: 'pack' packagesToPack: 'scr/xxx/xxxx.csproj' versioningScheme: byBuildNumber # versioningScheme: byEnvVar # versionEnvVar: $(name) # not regonized ``` --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: a57f8545-bb15-3a71-1876-3a9ec1a59b93 * Version Independent ID: 28c87c8d-c28d-7493-0c7c-8c38b04fbcd7 * Content: [Run (build) number - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/run-number?view=azure-devops&tabs=yaml) * Content Source: [docs/pipelines/process/run-number.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/run-number.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
1.0
The example doesn't seem to update the Build.BuildNumber - I'm trying to create a nuget package numbering in a yaml file but whatever i try the build number is not chaning. Even the given example ``` name: $(TeamProject)_$(Build.DefinitionName)_$(SourceBranchName)_$(Date:yyyyMMdd)$(Rev:.r) steps: - script: echo '$(Build.BuildNumber)' # outputs customized build number like project_def_master_20200828.1 ``` Just returns the defauld YYYY.nn back = > '20210425.44' I've no cuw what I'm missing could you give me soe feedback? I just what to do: ``` variables: name: '1.0.0.$(Build.BuildId)' MyRunNumber: '1.0.0.$(Build.BuildId)' steps: - script: echo '$(Build.BuildNumber)' #is wrong - script: echo '$(MyRunNumber)' #is ok - script: echo '$(name)' #is ok .... some other stuff and not sure i need this - task: Assembly-Info-NetCore@2 inputs: Path: 'scr/xxx/xxxx.csproj' FileNames: '**/*.csproj' InsertAttributes: true FileEncoding: 'auto' WriteBOM: false VersionNumber: $(name) FileVersionNumber: $(name) PackageVersion: $(name) ... than the error command - task: NuGetCommand@2 inputs: command: 'pack' packagesToPack: 'scr/xxx/xxxx.csproj' versioningScheme: byBuildNumber # versioningScheme: byEnvVar # versionEnvVar: $(name) # not regonized ``` --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: a57f8545-bb15-3a71-1876-3a9ec1a59b93 * Version Independent ID: 28c87c8d-c28d-7493-0c7c-8c38b04fbcd7 * Content: [Run (build) number - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/run-number?view=azure-devops&tabs=yaml) * Content Source: [docs/pipelines/process/run-number.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/run-number.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
process
the example doesn t seem to update the build buildnumber i m trying to create a nuget package numbering in a yaml file but whatever i try the build number is not chaning even the given example name teamproject build definitionname sourcebranchname date yyyymmdd rev r steps script echo build buildnumber outputs customized build number like project def master just returns the defauld yyyy nn back i ve no cuw what i m missing could you give me soe feedback i just what to do variables name build buildid myrunnumber build buildid steps script echo build buildnumber is wrong script echo myrunnumber is ok script echo name is ok some other stuff and not sure i need this task assembly info netcore inputs path scr xxx xxxx csproj filenames csproj insertattributes true fileencoding auto writebom false versionnumber name fileversionnumber name packageversion name than the error command task nugetcommand inputs command pack packagestopack scr xxx xxxx csproj versioningscheme bybuildnumber versioningscheme byenvvar versionenvvar name not regonized document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
1
11
2,491,275,015
IssuesEvent
2015-01-03 07:03:22
golang/go
https://api.github.com/repos/golang/go
opened
builders: track bug for OS X virtualization
builder
I'd like to get OS X builders running in VMs too. (Related to #9492 and https://golang.org/s/builderplan) Looks like virtualizing OS X 10.7+ is legal (but not 10.6) as long as it's only 1 copy, and on official Apple hardware. Considering that we already run Go builders on official Mac hardware in the office, we can continue to do so, but with a VM solution. And looks like VMWare Fusion has the "vmrun" command, documented at http://www.vmware.com/pdf/vix162_vmrun_command.pdf , so we can write a little API server that runs on the OS X host and calls vmrun. /cc @adg
1.0
builders: track bug for OS X virtualization - I'd like to get OS X builders running in VMs too. (Related to #9492 and https://golang.org/s/builderplan) Looks like virtualizing OS X 10.7+ is legal (but not 10.6) as long as it's only 1 copy, and on official Apple hardware. Considering that we already run Go builders on official Mac hardware in the office, we can continue to do so, but with a VM solution. And looks like VMWare Fusion has the "vmrun" command, documented at http://www.vmware.com/pdf/vix162_vmrun_command.pdf , so we can write a little API server that runs on the OS X host and calls vmrun. /cc @adg
non_process
builders track bug for os x virtualization i d like to get os x builders running in vms too related to and looks like virtualizing os x is legal but not as long as it s only copy and on official apple hardware considering that we already run go builders on official mac hardware in the office we can continue to do so but with a vm solution and looks like vmware fusion has the vmrun command documented at so we can write a little api server that runs on the os x host and calls vmrun cc adg
0