Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
17,715
23,616,802,795
IssuesEvent
2022-08-24 16:33:50
celo-org/celo-monorepo
https://api.github.com/repos/celo-org/celo-monorepo
closed
Draft Core-Contracts 8 release notes.
release-process Component: Identity ASv2
Draft release notes for [Core Contracts Release 8](https://github.com/celo-org/celo-monorepo/releases/tag/core-contracts.v8.pre-audit) (pre-release). In particular the following sections: - [x] "Key updates in this release" (blurb with description of changes) - [x] "Specific Version Updates" (table with contract versions) You can use [Core Contracts Release 7](https://github.com/celo-org/celo-monorepo/releases/tag/core-contracts.v7.post-audit) as a template. I (@0xarthurxyz) will add the Hacken.io audit report to the "Audits" section when it's published in the next days.
1.0
Draft Core-Contracts 8 release notes. - Draft release notes for [Core Contracts Release 8](https://github.com/celo-org/celo-monorepo/releases/tag/core-contracts.v8.pre-audit) (pre-release). In particular the following sections: - [x] "Key updates in this release" (blurb with description of changes) - [x] "Specific Version Updates" (table with contract versions) You can use [Core Contracts Release 7](https://github.com/celo-org/celo-monorepo/releases/tag/core-contracts.v7.post-audit) as a template. I (@0xarthurxyz) will add the Hacken.io audit report to the "Audits" section when it's published in the next days.
process
draft core contracts release notes draft release notes for pre release in particular the following sections key updates in this release blurb with description of changes specific version updates table with contract versions you can use as a template i will add the hacken io audit report to the audits section when it s published in the next days
1
110,875
16,995,012,752
IssuesEvent
2021-07-01 04:41:18
avallete/yt-playlists-delete-enhancer
https://api.github.com/repos/avallete/yt-playlists-delete-enhancer
closed
CVE-2018-20821 (Medium) detected in opennmsopennms-source-26.0.0-1, node-sass-4.14.1.tgz
no-issue-activity security vulnerability
## CVE-2018-20821 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>opennmsopennms-source-26.0.0-1</b>, <b>node-sass-4.14.1.tgz</b></p></summary> <p> <details><summary><b>node-sass-4.14.1.tgz</b></p></summary> <p>Wrapper around libsass</p> <p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz</a></p> <p>Path to dependency file: yt-playlists-delete-enhancer/package.json</p> <p>Path to vulnerable library: yt-playlists-delete-enhancer/node_modules/node-sass/package.json</p> <p> Dependency Hierarchy: - userscripter-1.4.0.tgz (Root Library) - :x: **node-sass-4.14.1.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/avallete/yt-playlists-delete-enhancer/commit/9c336b0fa3155406498ca56519999606da4494b5">9c336b0fa3155406498ca56519999606da4494b5</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The parsing component in LibSass through 3.5.5 allows attackers to cause a denial-of-service (uncontrolled recursion in Sass::Parser::parse_css_variable_value in parser.cpp). <p>Publish Date: 2019-04-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20821>CVE-2018-20821</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20821">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20821</a></p> <p>Release Date: 2019-04-23</p> <p>Fix Resolution: LibSass - 3.6.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-20821 (Medium) detected in opennmsopennms-source-26.0.0-1, node-sass-4.14.1.tgz - ## CVE-2018-20821 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>opennmsopennms-source-26.0.0-1</b>, <b>node-sass-4.14.1.tgz</b></p></summary> <p> <details><summary><b>node-sass-4.14.1.tgz</b></p></summary> <p>Wrapper around libsass</p> <p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz</a></p> <p>Path to dependency file: yt-playlists-delete-enhancer/package.json</p> <p>Path to vulnerable library: yt-playlists-delete-enhancer/node_modules/node-sass/package.json</p> <p> Dependency Hierarchy: - userscripter-1.4.0.tgz (Root Library) - :x: **node-sass-4.14.1.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/avallete/yt-playlists-delete-enhancer/commit/9c336b0fa3155406498ca56519999606da4494b5">9c336b0fa3155406498ca56519999606da4494b5</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The parsing component in LibSass through 3.5.5 allows attackers to cause a denial-of-service (uncontrolled recursion in Sass::Parser::parse_css_variable_value in parser.cpp). <p>Publish Date: 2019-04-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20821>CVE-2018-20821</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20821">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20821</a></p> <p>Release Date: 2019-04-23</p> <p>Fix Resolution: LibSass - 3.6.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in opennmsopennms source node sass tgz cve medium severity vulnerability vulnerable libraries opennmsopennms source node sass tgz node sass tgz wrapper around libsass library home page a href path to dependency file yt playlists delete enhancer package json path to vulnerable library yt playlists delete enhancer node modules node sass package json dependency hierarchy userscripter tgz root library x node sass tgz vulnerable library found in head commit a href found in base branch master vulnerability details the parsing component in libsass through allows attackers to cause a denial of service uncontrolled recursion in sass parser parse css variable value in parser cpp publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution libsass step up your open source security game with whitesource
0
282,826
21,315,965,065
IssuesEvent
2022-04-16 09:23:44
johnmcmonigle/pe
https://api.github.com/repos/johnmcmonigle/pe
opened
'Add Ward' sequence diagram needs calling arrows touching the tops of activation bars, and returning arrows touching the bottoms
severity.Low type.DocumentationBug
![image.png](https://raw.githubusercontent.com/johnmcmonigle/pe/main/files/73226150-85c1-4067-8101-e9d88d8f866e.png) Many of the calling and returning arrows point to the middle of activation bars <!--session: 1650094551158-c98fb3d6-71d0-4d9b-987f-11636894934a--> <!--Version: Web v3.4.2-->
1.0
'Add Ward' sequence diagram needs calling arrows touching the tops of activation bars, and returning arrows touching the bottoms - ![image.png](https://raw.githubusercontent.com/johnmcmonigle/pe/main/files/73226150-85c1-4067-8101-e9d88d8f866e.png) Many of the calling and returning arrows point to the middle of activation bars <!--session: 1650094551158-c98fb3d6-71d0-4d9b-987f-11636894934a--> <!--Version: Web v3.4.2-->
non_process
add ward sequence diagram needs calling arrows touching the tops of activation bars and returning arrows touching the bottoms many of the calling and returning arrows point to the middle of activation bars
0
93,905
15,946,437,611
IssuesEvent
2021-04-15 01:04:05
jgeraigery/core
https://api.github.com/repos/jgeraigery/core
opened
CVE-2018-10237 (Medium) detected in guava-18.0.jar
security vulnerability
## CVE-2018-10237 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>guava-18.0.jar</b></p></summary> <p>Guava is a suite of core and expanded libraries that include utility classes, google's collections, io classes, and much much more. Guava has only one code dependency - javax.annotation, per the JSR-305 spec.</p> <p>Library home page: <a href="http://code.google.com/p/guava-libraries">http://code.google.com/p/guava-libraries</a></p> <p>Path to dependency file: core/nimbus-core/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/google/guava/guava/18.0/guava-18.0.jar,/home/wss-scanner/.m2/repository/com/google/guava/guava/18.0/guava-18.0.jar,/home/wss-scanner/.m2/repository/com/google/guava/guava/18.0/guava-18.0.jar</p> <p> Dependency Hierarchy: - springfox-swagger2-2.4.0.jar (Root Library) - :x: **guava-18.0.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Unbounded memory allocation in Google Guava 11.0 through 24.x before 24.1.1 allows remote attackers to conduct denial of service attacks against servers that depend on this library and deserialize attacker-provided data, because the AtomicDoubleArray class (when serialized with Java serialization) and the CompoundOrdering class (when serialized with GWT serialization) perform eager allocation without appropriate checks on what a client has sent and whether the data size is reasonable. <p>Publish Date: 2018-04-26 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-10237>CVE-2018-10237</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-10237">https://nvd.nist.gov/vuln/detail/CVE-2018-10237</a></p> <p>Release Date: 2018-04-26</p> <p>Fix Resolution: 24.1.1-jre, 24.1.1-android</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.google.guava","packageName":"guava","packageVersion":"18.0","packageFilePaths":["/nimbus-core/pom.xml","/nimbus-entity-dsl/pom.xml","/nimbus-test/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"io.springfox:springfox-swagger2:2.4.0;com.google.guava:guava:18.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"24.1.1-jre, 24.1.1-android"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2018-10237","vulnerabilityDetails":"Unbounded memory allocation in Google Guava 11.0 through 24.x before 24.1.1 allows remote attackers to conduct denial of service attacks against servers that depend on this library and deserialize attacker-provided data, because the AtomicDoubleArray class (when serialized with Java serialization) and the CompoundOrdering class (when serialized with GWT serialization) perform eager allocation without appropriate checks on what a client has sent and whether the data size is reasonable.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-10237","cvss3Severity":"medium","cvss3Score":"5.9","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2018-10237 (Medium) detected in guava-18.0.jar - ## CVE-2018-10237 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>guava-18.0.jar</b></p></summary> <p>Guava is a suite of core and expanded libraries that include utility classes, google's collections, io classes, and much much more. Guava has only one code dependency - javax.annotation, per the JSR-305 spec.</p> <p>Library home page: <a href="http://code.google.com/p/guava-libraries">http://code.google.com/p/guava-libraries</a></p> <p>Path to dependency file: core/nimbus-core/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/google/guava/guava/18.0/guava-18.0.jar,/home/wss-scanner/.m2/repository/com/google/guava/guava/18.0/guava-18.0.jar,/home/wss-scanner/.m2/repository/com/google/guava/guava/18.0/guava-18.0.jar</p> <p> Dependency Hierarchy: - springfox-swagger2-2.4.0.jar (Root Library) - :x: **guava-18.0.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Unbounded memory allocation in Google Guava 11.0 through 24.x before 24.1.1 allows remote attackers to conduct denial of service attacks against servers that depend on this library and deserialize attacker-provided data, because the AtomicDoubleArray class (when serialized with Java serialization) and the CompoundOrdering class (when serialized with GWT serialization) perform eager allocation without appropriate checks on what a client has sent and whether the data size is reasonable. <p>Publish Date: 2018-04-26 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-10237>CVE-2018-10237</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-10237">https://nvd.nist.gov/vuln/detail/CVE-2018-10237</a></p> <p>Release Date: 2018-04-26</p> <p>Fix Resolution: 24.1.1-jre, 24.1.1-android</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.google.guava","packageName":"guava","packageVersion":"18.0","packageFilePaths":["/nimbus-core/pom.xml","/nimbus-entity-dsl/pom.xml","/nimbus-test/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"io.springfox:springfox-swagger2:2.4.0;com.google.guava:guava:18.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"24.1.1-jre, 24.1.1-android"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2018-10237","vulnerabilityDetails":"Unbounded memory allocation in Google Guava 11.0 through 24.x before 24.1.1 allows remote attackers to conduct denial of service attacks against servers that depend on this library and deserialize attacker-provided data, because the AtomicDoubleArray class (when serialized with Java serialization) and the CompoundOrdering class (when serialized with GWT serialization) perform eager allocation without appropriate checks on what a client has sent and whether the data size is reasonable.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-10237","cvss3Severity":"medium","cvss3Score":"5.9","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_process
cve medium detected in guava jar cve medium severity vulnerability vulnerable library guava jar guava is a suite of core and expanded libraries that include utility classes google s collections io classes and much much more guava has only one code dependency javax annotation per the jsr spec library home page a href path to dependency file core nimbus core pom xml path to vulnerable library home wss scanner repository com google guava guava guava jar home wss scanner repository com google guava guava guava jar home wss scanner repository com google guava guava guava jar dependency hierarchy springfox jar root library x guava jar vulnerable library found in base branch master vulnerability details unbounded memory allocation in google guava through x before allows remote attackers to conduct denial of service attacks against servers that depend on this library and deserialize attacker provided data because the atomicdoublearray class when serialized with java serialization and the compoundordering class when serialized with gwt serialization perform eager allocation without appropriate checks on what a client has sent and whether the data size is reasonable publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jre android isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree io springfox springfox com google guava guava isminimumfixversionavailable true minimumfixversion jre android basebranches vulnerabilityidentifier cve vulnerabilitydetails unbounded memory allocation in google guava through x before allows remote attackers to conduct denial of service attacks against servers that depend on this library and deserialize attacker provided data because the atomicdoublearray class when serialized with java serialization and the compoundordering class when serialized with gwt serialization perform eager allocation without appropriate checks on what a client has sent and whether the data size is reasonable vulnerabilityurl
0
11,693
3,218,596,110
IssuesEvent
2015-10-08 02:51:16
eris-ltd/eris-cli
https://api.github.com/repos/eris-ltd/eris-cli
opened
tests should randomize ports
area/test suite
thus no need to stop all running containers when testing ... then start them back up again
1.0
tests should randomize ports - thus no need to stop all running containers when testing ... then start them back up again
non_process
tests should randomize ports thus no need to stop all running containers when testing then start them back up again
0
52,868
27,808,822,399
IssuesEvent
2023-03-17 23:37:50
enso-org/enso
https://api.github.com/repos/enso-org/enso
closed
Consider parallelization of ZIO's setup during startup
p-medium x-chore --low-performance -language-server
The current implementation of Language Server makes [heavy usage](https://github.com/enso-org/enso/blob/develop/engine/language-server/src/main/scala/org/enso/languageserver/boot/MainModule.scala#L107) of ZIO. As per profiling data ![image](https://user-images.githubusercontent.com/292128/224128736-adf24a4a-70c3-45a7-b96a-8fe69b17d698.png) it also takes a considerable amount of time to initialize. Unfortunately the current implementation of Language Server integrates this 3rd party library rather heavily and it is not easy to remove it without significant refactorings. But ZIO's runtime is also not needed until serving some requests which makes it a potential candidate for non-blocking initialization, potentially reducing the time spent in startup.
True
Consider parallelization of ZIO's setup during startup - The current implementation of Language Server makes [heavy usage](https://github.com/enso-org/enso/blob/develop/engine/language-server/src/main/scala/org/enso/languageserver/boot/MainModule.scala#L107) of ZIO. As per profiling data ![image](https://user-images.githubusercontent.com/292128/224128736-adf24a4a-70c3-45a7-b96a-8fe69b17d698.png) it also takes a considerable amount of time to initialize. Unfortunately the current implementation of Language Server integrates this 3rd party library rather heavily and it is not easy to remove it without significant refactorings. But ZIO's runtime is also not needed until serving some requests which makes it a potential candidate for non-blocking initialization, potentially reducing the time spent in startup.
non_process
consider parallelization of zio s setup during startup the current implementation of language server makes of zio as per profiling data it also takes a considerable amount of time to initialize unfortunately the current implementation of language server integrates this party library rather heavily and it is not easy to remove it without significant refactorings but zio s runtime is also not needed until serving some requests which makes it a potential candidate for non blocking initialization potentially reducing the time spent in startup
0
10,254
13,108,751,954
IssuesEvent
2020-08-04 17:24:15
googleapis/code-suggester
https://api.github.com/repos/googleapis/code-suggester
closed
Implement CLI for creating a forked pull request (PR)
enhancement type: process
- [x] Implement CLI for creating a forked pull request (PR) ### Description The user should be able to create a PR in a fork from either a non-git or git directory. For non-git directory specified: all of the files in the directory get added on top of the upstream repository For a git directory specified: all of the un-tracked git changes are made into a PR, except for the `.gitignore` files. ### Parameters The framework-core should be invoked through the command-line. Specifically, support the parameters for: 1. `--upstream-repo`: The upstream repository to base a forked repository off of. Alias: `repo`, `r` 2. `--upstream-owner`: "The owner of the upstream repository. Alias: `owner`, `o` 3. `--pr-description`: The description of the PR to be made. Alias: `description`, `d` 4. `--pr-title`: The description of the PR title to be made. Alias: `title`, `t` 5. `--branch-name`: The name of the working branch to apply changes to. Alias: `branch`, `b` 6. `--change.work-dir`: The location of non-git changes to apply on top of an upstream repository. 7. `--change.git-dir`: The location of any un-tracked changes that should be made into a PR. Files in the `.gitignore` are ignored ### Environment variables - `ACCESS_TOKEN`: an account that has permissions to create a fork. ### Additional context - automating a PR from un-tracked git changes is useful because it reduces the number of commands - creating a PR from a non-git directory is useful because it allows a repository to be initialized with a template. For instance if you have a `LICENSE` directory hosted locally, you can specify the directory of that LICENSE file and upload that to the upstream repository
1.0
Implement CLI for creating a forked pull request (PR) - - [x] Implement CLI for creating a forked pull request (PR) ### Description The user should be able to create a PR in a fork from either a non-git or git directory. For non-git directory specified: all of the files in the directory get added on top of the upstream repository For a git directory specified: all of the un-tracked git changes are made into a PR, except for the `.gitignore` files. ### Parameters The framework-core should be invoked through the command-line. Specifically, support the parameters for: 1. `--upstream-repo`: The upstream repository to base a forked repository off of. Alias: `repo`, `r` 2. `--upstream-owner`: "The owner of the upstream repository. Alias: `owner`, `o` 3. `--pr-description`: The description of the PR to be made. Alias: `description`, `d` 4. `--pr-title`: The description of the PR title to be made. Alias: `title`, `t` 5. `--branch-name`: The name of the working branch to apply changes to. Alias: `branch`, `b` 6. `--change.work-dir`: The location of non-git changes to apply on top of an upstream repository. 7. `--change.git-dir`: The location of any un-tracked changes that should be made into a PR. Files in the `.gitignore` are ignored ### Environment variables - `ACCESS_TOKEN`: an account that has permissions to create a fork. ### Additional context - automating a PR from un-tracked git changes is useful because it reduces the number of commands - creating a PR from a non-git directory is useful because it allows a repository to be initialized with a template. For instance if you have a `LICENSE` directory hosted locally, you can specify the directory of that LICENSE file and upload that to the upstream repository
process
implement cli for creating a forked pull request pr implement cli for creating a forked pull request pr description the user should be able to create a pr in a fork from either a non git or git directory for non git directory specified all of the files in the directory get added on top of the upstream repository for a git directory specified all of the un tracked git changes are made into a pr except for the gitignore files parameters the framework core should be invoked through the command line specifically support the parameters for upstream repo the upstream repository to base a forked repository off of alias repo r upstream owner the owner of the upstream repository alias owner o pr description the description of the pr to be made alias description d pr title the description of the pr title to be made alias title t branch name the name of the working branch to apply changes to alias branch b change work dir the location of non git changes to apply on top of an upstream repository change git dir the location of any un tracked changes that should be made into a pr files in the gitignore are ignored environment variables access token an account that has permissions to create a fork additional context automating a pr from un tracked git changes is useful because it reduces the number of commands creating a pr from a non git directory is useful because it allows a repository to be initialized with a template for instance if you have a license directory hosted locally you can specify the directory of that license file and upload that to the upstream repository
1
6,233
9,180,956,179
IssuesEvent
2019-03-05 09:04:08
kmycode/sangokukmy
https://api.github.com/repos/kmycode/sangokukmy
closed
農民反乱
enhancement func-oldkmy process-pending
三国志NET KMY Versionでは、農民反乱を実装しようと考えています。ほぼ以前あったものと同じです。 異民族( #10 )が若干強力すぎるので、農民反乱は地味なままでもいいかなと思ってます。 ## 趣旨・目的 密偵の謀略により、農民反乱が起きるようにする。戦争前、他国への侵攻作戦を考えるのと同時に、自国の防衛作戦を考慮する必要が生じ、戦争は開戦する前からすでに始まっているという、簡単に油断できない雰囲気を作る ## 発生条件 * 密偵が扇動を実行している * 民忠がゼロである * その都市に武将が誰もいない ## できること * 密偵( #12 )を放たれ、毎ターン扇動を実行された都市では、非常に低い確率で農民反乱が発生する * 農民反乱が発生した都市は、農業・商業・技術がすべてゼロになったうえで、城壁1000、守兵500の無所属都市となる * 2~3分に1の確率で、密偵を放った武将の所属国に寝返る * 無所属都市へは、戦争準備中の国だけでなく、無関係の国も侵攻できる。もちろん建国もできる ## 以前との違い 特になし(城壁、守兵くらい?)
1.0
農民反乱 - 三国志NET KMY Versionでは、農民反乱を実装しようと考えています。ほぼ以前あったものと同じです。 異民族( #10 )が若干強力すぎるので、農民反乱は地味なままでもいいかなと思ってます。 ## 趣旨・目的 密偵の謀略により、農民反乱が起きるようにする。戦争前、他国への侵攻作戦を考えるのと同時に、自国の防衛作戦を考慮する必要が生じ、戦争は開戦する前からすでに始まっているという、簡単に油断できない雰囲気を作る ## 発生条件 * 密偵が扇動を実行している * 民忠がゼロである * その都市に武将が誰もいない ## できること * 密偵( #12 )を放たれ、毎ターン扇動を実行された都市では、非常に低い確率で農民反乱が発生する * 農民反乱が発生した都市は、農業・商業・技術がすべてゼロになったうえで、城壁1000、守兵500の無所属都市となる * 2~3分に1の確率で、密偵を放った武将の所属国に寝返る * 無所属都市へは、戦争準備中の国だけでなく、無関係の国も侵攻できる。もちろん建国もできる ## 以前との違い 特になし(城壁、守兵くらい?)
process
農民反乱 三国志net kmy versionでは、農民反乱を実装しようと考えています。ほぼ以前あったものと同じです。 異民族 が若干強力すぎるので、農民反乱は地味なままでもいいかなと思ってます。 趣旨・目的 密偵の謀略により、農民反乱が起きるようにする。戦争前、他国への侵攻作戦を考えるのと同時に、自国の防衛作戦を考慮する必要が生じ、戦争は開戦する前からすでに始まっているという、簡単に油断できない雰囲気を作る 発生条件 密偵が扇動を実行している 民忠がゼロである その都市に武将が誰もいない できること 密偵 を放たれ、毎ターン扇動を実行された都市では、非常に低い確率で農民反乱が発生する 農民反乱が発生した都市は、農業・商業・技術がすべてゼロになったうえで、 、 ~ 、密偵を放った武将の所属国に寝返る 無所属都市へは、戦争準備中の国だけでなく、無関係の国も侵攻できる。もちろん建国もできる 以前との違い 特になし(城壁、守兵くらい?)
1
16,065
20,205,911,749
IssuesEvent
2022-02-11 20:20:05
createwithrani/superlist
https://api.github.com/repos/createwithrani/superlist
opened
Have a Main and Develop branch
Process
With #29 adding automated deployment from GitHub to SVN, should we have a `main` branch that is `stable` and a `development` branch that is where we merge new features and prepare for new releases? The advantage of such a setup is that we can then use [WordPress Plugin Readme/Assets Update action](https://github.com/marketplace/actions/wordpress-plugin-readme-assets-update) to update just readme files, assets, and the 'tested up to' frontmatter without affecting the rest of the plugin.
1.0
Have a Main and Develop branch - With #29 adding automated deployment from GitHub to SVN, should we have a `main` branch that is `stable` and a `development` branch that is where we merge new features and prepare for new releases? The advantage of such a setup is that we can then use [WordPress Plugin Readme/Assets Update action](https://github.com/marketplace/actions/wordpress-plugin-readme-assets-update) to update just readme files, assets, and the 'tested up to' frontmatter without affecting the rest of the plugin.
process
have a main and develop branch with adding automated deployment from github to svn should we have a main branch that is stable and a development branch that is where we merge new features and prepare for new releases the advantage of such a setup is that we can then use to update just readme files assets and the tested up to frontmatter without affecting the rest of the plugin
1
19,946
26,419,644,632
IssuesEvent
2023-01-13 19:05:36
ORNL-AMO/AMO-Tools-Desktop
https://api.github.com/repos/ORNL-AMO/AMO-Tools-Desktop
opened
Higher Heating Value Calculators
Process Heating Intern To Do
Turn New Gas Fuel and New Solid/Liquid fuel modals into stand alone calcs
1.0
Higher Heating Value Calculators - Turn New Gas Fuel and New Solid/Liquid fuel modals into stand alone calcs
process
higher heating value calculators turn new gas fuel and new solid liquid fuel modals into stand alone calcs
1
14,010
16,815,860,171
IssuesEvent
2021-06-17 07:17:14
didi/mpx
https://api.github.com/repos/didi/mpx
closed
[Bug report] 无法集成 thread-loader
processing
``` { test: /\.mpx$/, use: [ { loader: 'thread-loader' }, MpxWebpackPlugin.loader(currentMpxLoaderConf) ] } ``` ``` Module build failed (from ./node_modules/thread-loader/dist/cjs.js): Thread Loader (Worker 0) Cannot read property 'compiler' of undefined at PoolWorker.fromErrorObj (/test/node_modules/thread-loader/dist/WorkerPool.js:346:12) at /test/node_modules/thread-loader/dist/WorkerPool.js:219:29 at module.exports (/test/node_modules/@mpxjs/webpack-plugin/lib/utils/get-main-compilation.js:2:22) at Object.module.exports (/test/node_modules/@mpxjs/webpack-plugin/lib/loader.js:23:27) Build failed with errors. ```
1.0
[Bug report] 无法集成 thread-loader - ``` { test: /\.mpx$/, use: [ { loader: 'thread-loader' }, MpxWebpackPlugin.loader(currentMpxLoaderConf) ] } ``` ``` Module build failed (from ./node_modules/thread-loader/dist/cjs.js): Thread Loader (Worker 0) Cannot read property 'compiler' of undefined at PoolWorker.fromErrorObj (/test/node_modules/thread-loader/dist/WorkerPool.js:346:12) at /test/node_modules/thread-loader/dist/WorkerPool.js:219:29 at module.exports (/test/node_modules/@mpxjs/webpack-plugin/lib/utils/get-main-compilation.js:2:22) at Object.module.exports (/test/node_modules/@mpxjs/webpack-plugin/lib/loader.js:23:27) Build failed with errors. ```
process
无法集成 thread loader test mpx use loader thread loader mpxwebpackplugin loader currentmpxloaderconf module build failed from node modules thread loader dist cjs js thread loader worker cannot read property compiler of undefined at poolworker fromerrorobj test node modules thread loader dist workerpool js at test node modules thread loader dist workerpool js at module exports test node modules mpxjs webpack plugin lib utils get main compilation js at object module exports test node modules mpxjs webpack plugin lib loader js build failed with errors
1
20,681
27,352,924,987
IssuesEvent
2023-02-27 10:50:19
camunda/issues
https://api.github.com/repos/camunda/issues
opened
Process Instance Modification
component:operate component:zeebe component:zeebe-process-automation public feature-parity version:8.1 riskClass:medium
### Value Proposition Statement Repair process instances that ended up in the wrong state by repeating or skipping steps. Move running flow nodes, add new or cancel existing ones in a process instance easily via our UI. ### User Problem - During execution process instances can end up in the wrong state. Currently to solve this, the user has to cancel the instance and start a new instance from beginning, requiring to rerun all previous steps. This requires a lot of considerations during process design and implementation. - Some use-cases include: - Repairing process instances in which some steps have to be repeated or skipped - Testing: Activities can be skipped or repeated for isolated testing of individual process segments ### User Stories - As a Developer, I can modify one process instances (single instance) via API - As a Developer, I can interact with API via Java Client - As a Operations Engineer / Admin I can modify one process instance in Operate UI - As a Product Manager I can track the usage of the feature (both UI and API) in Mixpanel ### Implementation Notes - Operate Swagger is updated - file and UI - New endpoints documented in Postman repository - Mixpanel backend and frontend events trackable and discoverable ### Breakdown <!-- Please link to sub-issues / -tasks contributing to respective epic phase or phase results where appropriate. --> #### Discovery phase ## <!-- Example: "Conduct customer interview with xyz" --> #### Define phase ## <!-- Example: link to "Define User-Journey Flow" or "Define target architecture" --> Design Planning * Reviewed by design: May 9, 2022 * Designer assigned: Yes * Assignee: @nazlikaya * [Design Brief](https://docs.google.com/document/d/12xVBJSjpCbG9f5gkHsxSA5mCdIFsFl889OQYQ6gJQcU/edit?usp=sharing) * Research Brief - none Design Deliverables - [Prototype](https://github.com/camunda/product-design/issues/11) (Expected delivery date June 10) - Delivered: June 7 - [Specifications](https://github.com/camunda/product-design/issues/13) (Expected delivery date June 10) - Delivered: June 7 - [High fidelity design](https://github.com/camunda/product-design/issues/14) (Expected delivery date June 10)- Delivered: June 9 - [Handover Recording](https://camunda.zoom.us/rec/share/rTRBwu6JVfMIzPB6ZnxZObbrsV5T3CoXiZLCtoztGaxXNdIrbjRz7HqBKzzHn5-r.PhmtEN5Z5UsQvgz3?startTime=1654844508000) Passcode: B5vuqi*$ [Risk Assessment](https://github.com/camunda/product-hub/issues/322) - Risk Class: Medium - Action: Mitigate #### Implement phase ## - [x] https://github.com/camunda/zeebe/issues/9570 - [x] https://github.com/camunda/operate/issues/2955 #### Validate phase ## - https://eu.mixpanel.com/s/qwRNk
1.0
Process Instance Modification - ### Value Proposition Statement Repair process instances that ended up in the wrong state by repeating or skipping steps. Move running flow nodes, add new or cancel existing ones in a process instance easily via our UI. ### User Problem - During execution process instances can end up in the wrong state. Currently to solve this, the user has to cancel the instance and start a new instance from beginning, requiring to rerun all previous steps. This requires a lot of considerations during process design and implementation. - Some use-cases include: - Repairing process instances in which some steps have to be repeated or skipped - Testing: Activities can be skipped or repeated for isolated testing of individual process segments ### User Stories - As a Developer, I can modify one process instances (single instance) via API - As a Developer, I can interact with API via Java Client - As a Operations Engineer / Admin I can modify one process instance in Operate UI - As a Product Manager I can track the usage of the feature (both UI and API) in Mixpanel ### Implementation Notes - Operate Swagger is updated - file and UI - New endpoints documented in Postman repository - Mixpanel backend and frontend events trackable and discoverable ### Breakdown <!-- Please link to sub-issues / -tasks contributing to respective epic phase or phase results where appropriate. --> #### Discovery phase ## <!-- Example: "Conduct customer interview with xyz" --> #### Define phase ## <!-- Example: link to "Define User-Journey Flow" or "Define target architecture" --> Design Planning * Reviewed by design: May 9, 2022 * Designer assigned: Yes * Assignee: @nazlikaya * [Design Brief](https://docs.google.com/document/d/12xVBJSjpCbG9f5gkHsxSA5mCdIFsFl889OQYQ6gJQcU/edit?usp=sharing) * Research Brief - none Design Deliverables - [Prototype](https://github.com/camunda/product-design/issues/11) (Expected delivery date June 10) - Delivered: June 7 - [Specifications](https://github.com/camunda/product-design/issues/13) (Expected delivery date June 10) - Delivered: June 7 - [High fidelity design](https://github.com/camunda/product-design/issues/14) (Expected delivery date June 10)- Delivered: June 9 - [Handover Recording](https://camunda.zoom.us/rec/share/rTRBwu6JVfMIzPB6ZnxZObbrsV5T3CoXiZLCtoztGaxXNdIrbjRz7HqBKzzHn5-r.PhmtEN5Z5UsQvgz3?startTime=1654844508000) Passcode: B5vuqi*$ [Risk Assessment](https://github.com/camunda/product-hub/issues/322) - Risk Class: Medium - Action: Mitigate #### Implement phase ## - [x] https://github.com/camunda/zeebe/issues/9570 - [x] https://github.com/camunda/operate/issues/2955 #### Validate phase ## - https://eu.mixpanel.com/s/qwRNk
process
process instance modification value proposition statement repair process instances that ended up in the wrong state by repeating or skipping steps move running flow nodes add new or cancel existing ones in a process instance easily via our ui user problem during execution process instances can end up in the wrong state currently to solve this the user has to cancel the instance and start a new instance from beginning requiring to rerun all previous steps this requires a lot of considerations during process design and implementation some use cases include repairing process instances in which some steps have to be repeated or skipped testing activities can be skipped or repeated for isolated testing of individual process segments user stories as a developer i can modify one process instances single instance via api as a developer i can interact with api via java client as a operations engineer admin i can modify one process instance in operate ui as a product manager i can track the usage of the feature both ui and api in mixpanel implementation notes operate swagger is updated file and ui new endpoints documented in postman repository mixpanel backend and frontend events trackable and discoverable breakdown discovery phase define phase design planning reviewed by design may designer assigned yes assignee nazlikaya research brief none design deliverables expected delivery date june delivered june expected delivery date june delivered june expected delivery date june delivered june passcode risk class medium action mitigate implement phase validate phase
1
29,085
23,707,318,039
IssuesEvent
2022-08-30 03:21:53
UBCSailbot/.github
https://api.github.com/repos/UBCSailbot/.github
opened
Create PR template
infrastructure
### Purpose Guidelines for what to include in a PR ### Changes - Write PR template file - Look at the second resource to see if there is anything that we could do to improve our issues templates ### Resources - https://embeddedartistry.com/blog/2017/08/04/a-github-pull-request-template-for-your-projects/ - https://github.com/stevemao/github-issue-templates
1.0
Create PR template - ### Purpose Guidelines for what to include in a PR ### Changes - Write PR template file - Look at the second resource to see if there is anything that we could do to improve our issues templates ### Resources - https://embeddedartistry.com/blog/2017/08/04/a-github-pull-request-template-for-your-projects/ - https://github.com/stevemao/github-issue-templates
non_process
create pr template purpose guidelines for what to include in a pr changes write pr template file look at the second resource to see if there is anything that we could do to improve our issues templates resources
0
14,349
9,084,639,972
IssuesEvent
2019-02-18 04:47:03
OctopusDeploy/Issues
https://api.github.com/repos/OctopusDeploy/Issues
opened
Cant tell what type of deployment target a given machine is
area/usability kind/enhancement
# Prerequisites - [x ] I have searched [open](https://github.com/OctopusDeploy/Issues/issues) and [closed](https://github.com/OctopusDeploy/Issues/issues?utf8=%E2%9C%93&q=is%3Aissue+is%3Aclosed) issues to make sure it isn't already requested - [x] I have written a descriptive issue title - [x] I have linked the original source of this feature request - [x] I have tagged the issue appropriately (area/*, kind/enhancement) # The enhancement It's currently not possible to see what kind of deployment target you're looking at when viewing a deployment target, unless you've memorized the icons. We should show the deployment target type (`Polling Tentacle`, `Kubernetes Cluster`, etc) on the screen. ## Mockup ![image](https://user-images.githubusercontent.com/373389/52928494-20348500-3394-11e9-9b24-6e08a5a6b6ed.png) ## Workarounds Use the Infrastructure -> deployment targets screen to filter by target name and communication style until you find it ## Links source: internally reported.
True
Cant tell what type of deployment target a given machine is - # Prerequisites - [x ] I have searched [open](https://github.com/OctopusDeploy/Issues/issues) and [closed](https://github.com/OctopusDeploy/Issues/issues?utf8=%E2%9C%93&q=is%3Aissue+is%3Aclosed) issues to make sure it isn't already requested - [x] I have written a descriptive issue title - [x] I have linked the original source of this feature request - [x] I have tagged the issue appropriately (area/*, kind/enhancement) # The enhancement It's currently not possible to see what kind of deployment target you're looking at when viewing a deployment target, unless you've memorized the icons. We should show the deployment target type (`Polling Tentacle`, `Kubernetes Cluster`, etc) on the screen. ## Mockup ![image](https://user-images.githubusercontent.com/373389/52928494-20348500-3394-11e9-9b24-6e08a5a6b6ed.png) ## Workarounds Use the Infrastructure -> deployment targets screen to filter by target name and communication style until you find it ## Links source: internally reported.
non_process
cant tell what type of deployment target a given machine is prerequisites i have searched and issues to make sure it isn t already requested i have written a descriptive issue title i have linked the original source of this feature request i have tagged the issue appropriately area kind enhancement the enhancement it s currently not possible to see what kind of deployment target you re looking at when viewing a deployment target unless you ve memorized the icons we should show the deployment target type polling tentacle kubernetes cluster etc on the screen mockup workarounds use the infrastructure deployment targets screen to filter by target name and communication style until you find it links source internally reported
0
22,084
30,606,733,796
IssuesEvent
2023-07-23 04:57:52
hashgraph/hedera-json-rpc-relay
https://api.github.com/repos/hashgraph/hedera-json-rpc-relay
opened
Separate http and ws metrics
enhancement P2 process
### Problem Currently the http and ws server both utilize the relay and other busines logic. However, since the registries are created in the classes it's not clear the metrics are intended for the ws server when it's running ### Solution Separate metrics flags or make sure there's a different layer for calls originating for one over the other ### Alternatives _No response_
1.0
Separate http and ws metrics - ### Problem Currently the http and ws server both utilize the relay and other busines logic. However, since the registries are created in the classes it's not clear the metrics are intended for the ws server when it's running ### Solution Separate metrics flags or make sure there's a different layer for calls originating for one over the other ### Alternatives _No response_
process
separate http and ws metrics problem currently the http and ws server both utilize the relay and other busines logic however since the registries are created in the classes it s not clear the metrics are intended for the ws server when it s running solution separate metrics flags or make sure there s a different layer for calls originating for one over the other alternatives no response
1
338,024
30,277,400,647
IssuesEvent
2023-07-07 21:09:19
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
pkg/sql/schemachanger/schemachanger_test: TestValidateMixedVersionElements_drop_column_with_partial_index failed
C-test-failure O-robot branch-master T-sql-foundations
pkg/sql/schemachanger/schemachanger_test.TestValidateMixedVersionElements_drop_column_with_partial_index [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Ci_TestsAwsLinuxArm64_UnitTests/10812861?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Ci_TestsAwsLinuxArm64_UnitTests/10812861?buildTab=artifacts#/) on master @ [1137aee0089fc2fe161bab0f1c899d2ddc997c90](https://github.com/cockroachdb/cockroach/commits/1137aee0089fc2fe161bab0f1c899d2ddc997c90): ``` === RUN TestValidateMixedVersionElements_drop_column_with_partial_index test_log_scope.go:167: test logs captured to: /artifacts/tmp/_tmp/6e404cfaa494ad6aa388bb776400b270/logTestValidateMixedVersionElements_drop_column_with_partial_index1152271934 test_log_scope.go:81: use -show-logs to present logs inline datadriven.go:143: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/7298/execroot/com_github_cockroachdb_cockroach/bazel-out/aarch64-fastbuild/bin/pkg/sql/schemachanger/schemachanger_test_/schemachanger_test.runfiles/com_github_cockroachdb_cockroach/pkg/sql/schemachanger/testdata/end_to_end/drop_column_with_partial_index/drop_column_with_partial_index.definition:4: setup [0 args] CREATE TABLE t (i INT PRIMARY KEY, j INT, INDEX (j) WHERE i >=0) ---- datadriven.go:143: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/7298/execroot/com_github_cockroachdb_cockroach/bazel-out/aarch64-fastbuild/bin/pkg/sql/schemachanger/schemachanger_test_/schemachanger_test.runfiles/com_github_cockroachdb_cockroach/pkg/sql/schemachanger/testdata/end_to_end/drop_column_with_partial_index/drop_column_with_partial_index.definition:9: stage-exec [2 args] INSERT INTO t (i) VALUES($stageKey); INSERT INTO t (i) VALUES($stageKey + 1); ---- datadriven.go:143: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/7298/execroot/com_github_cockroachdb_cockroach/bazel-out/aarch64-fastbuild/bin/pkg/sql/schemachanger/schemachanger_test_/schemachanger_test.runfiles/com_github_cockroachdb_cockroach/pkg/sql/schemachanger/testdata/end_to_end/drop_column_with_partial_index/drop_column_with_partial_index.definition:16: stage-query [2 args] SELECT count(*)=$successfulStageCount*2 FROM t; ---- true datadriven.go:143: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/7298/execroot/com_github_cockroachdb_cockroach/bazel-out/aarch64-fastbuild/bin/pkg/sql/schemachanger/schemachanger_test_/schemachanger_test.runfiles/com_github_cockroachdb_cockroach/pkg/sql/schemachanger/testdata/end_to_end/drop_column_with_partial_index/drop_column_with_partial_index.definition:22: stage-exec [2 args] INSERT INTO t (i) VALUES($stageKey); INSERT INTO t (i) VALUES($stageKey + 1); ---- datadriven.go:143: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/7298/execroot/com_github_cockroachdb_cockroach/bazel-out/aarch64-fastbuild/bin/pkg/sql/schemachanger/schemachanger_test_/schemachanger_test.runfiles/com_github_cockroachdb_cockroach/pkg/sql/schemachanger/testdata/end_to_end/drop_column_with_partial_index/drop_column_with_partial_index.definition:29: stage-query [2 args] SELECT count(*)=$successfulStageCount*2 FROM t; ---- true === CONT TestValidateMixedVersionElements_drop_column_with_partial_index panic.go:522: -- test log scope end -- test logs left over in: /artifacts/tmp/_tmp/6e404cfaa494ad6aa388bb776400b270/logTestValidateMixedVersionElements_drop_column_with_partial_index1152271934 --- FAIL: TestValidateMixedVersionElements_drop_column_with_partial_index (23.39s) === RUN TestValidateMixedVersionElements_drop_column_with_partial_index/pause_upgrade_and_resume_at_each_stage --- FAIL: TestValidateMixedVersionElements_drop_column_with_partial_index/pause_upgrade_and_resume_at_each_stage (22.60s) === RUN TestValidateMixedVersionElements_drop_column_with_partial_index/pause_upgrade_and_resume_at_each_stage/PostCommitNonRevertiblePhase_3_of_3 cumulative.go:1724: Error Trace: github.com/cockroachdb/cockroach/pkg/sql/schemachanger/sctest/cumulative.go:1724 GOROOT/src/runtime/asm_arm64.s:1172 Error: Received unexpected error: dial tcp 127.0.0.1:40613: connect: connection refused Test: TestValidateMixedVersionElements_drop_column_with_partial_index/pause_upgrade_and_resume_at_each_stage/PostCommitNonRevertiblePhase_3_of_3 --- FAIL: TestValidateMixedVersionElements_drop_column_with_partial_index/pause_upgrade_and_resume_at_each_stage/PostCommitNonRevertiblePhase_3_of_3 (4.76s) ``` <details><summary>Help</summary> <p> See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) </p> </details> /cc @cockroachdb/sql-foundations <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestValidateMixedVersionElements_drop_column_with_partial_index.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-29530
1.0
pkg/sql/schemachanger/schemachanger_test: TestValidateMixedVersionElements_drop_column_with_partial_index failed - pkg/sql/schemachanger/schemachanger_test.TestValidateMixedVersionElements_drop_column_with_partial_index [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Ci_TestsAwsLinuxArm64_UnitTests/10812861?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Ci_TestsAwsLinuxArm64_UnitTests/10812861?buildTab=artifacts#/) on master @ [1137aee0089fc2fe161bab0f1c899d2ddc997c90](https://github.com/cockroachdb/cockroach/commits/1137aee0089fc2fe161bab0f1c899d2ddc997c90): ``` === RUN TestValidateMixedVersionElements_drop_column_with_partial_index test_log_scope.go:167: test logs captured to: /artifacts/tmp/_tmp/6e404cfaa494ad6aa388bb776400b270/logTestValidateMixedVersionElements_drop_column_with_partial_index1152271934 test_log_scope.go:81: use -show-logs to present logs inline datadriven.go:143: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/7298/execroot/com_github_cockroachdb_cockroach/bazel-out/aarch64-fastbuild/bin/pkg/sql/schemachanger/schemachanger_test_/schemachanger_test.runfiles/com_github_cockroachdb_cockroach/pkg/sql/schemachanger/testdata/end_to_end/drop_column_with_partial_index/drop_column_with_partial_index.definition:4: setup [0 args] CREATE TABLE t (i INT PRIMARY KEY, j INT, INDEX (j) WHERE i >=0) ---- datadriven.go:143: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/7298/execroot/com_github_cockroachdb_cockroach/bazel-out/aarch64-fastbuild/bin/pkg/sql/schemachanger/schemachanger_test_/schemachanger_test.runfiles/com_github_cockroachdb_cockroach/pkg/sql/schemachanger/testdata/end_to_end/drop_column_with_partial_index/drop_column_with_partial_index.definition:9: stage-exec [2 args] INSERT INTO t (i) VALUES($stageKey); INSERT INTO t (i) VALUES($stageKey + 1); ---- datadriven.go:143: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/7298/execroot/com_github_cockroachdb_cockroach/bazel-out/aarch64-fastbuild/bin/pkg/sql/schemachanger/schemachanger_test_/schemachanger_test.runfiles/com_github_cockroachdb_cockroach/pkg/sql/schemachanger/testdata/end_to_end/drop_column_with_partial_index/drop_column_with_partial_index.definition:16: stage-query [2 args] SELECT count(*)=$successfulStageCount*2 FROM t; ---- true datadriven.go:143: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/7298/execroot/com_github_cockroachdb_cockroach/bazel-out/aarch64-fastbuild/bin/pkg/sql/schemachanger/schemachanger_test_/schemachanger_test.runfiles/com_github_cockroachdb_cockroach/pkg/sql/schemachanger/testdata/end_to_end/drop_column_with_partial_index/drop_column_with_partial_index.definition:22: stage-exec [2 args] INSERT INTO t (i) VALUES($stageKey); INSERT INTO t (i) VALUES($stageKey + 1); ---- datadriven.go:143: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/7298/execroot/com_github_cockroachdb_cockroach/bazel-out/aarch64-fastbuild/bin/pkg/sql/schemachanger/schemachanger_test_/schemachanger_test.runfiles/com_github_cockroachdb_cockroach/pkg/sql/schemachanger/testdata/end_to_end/drop_column_with_partial_index/drop_column_with_partial_index.definition:29: stage-query [2 args] SELECT count(*)=$successfulStageCount*2 FROM t; ---- true === CONT TestValidateMixedVersionElements_drop_column_with_partial_index panic.go:522: -- test log scope end -- test logs left over in: /artifacts/tmp/_tmp/6e404cfaa494ad6aa388bb776400b270/logTestValidateMixedVersionElements_drop_column_with_partial_index1152271934 --- FAIL: TestValidateMixedVersionElements_drop_column_with_partial_index (23.39s) === RUN TestValidateMixedVersionElements_drop_column_with_partial_index/pause_upgrade_and_resume_at_each_stage --- FAIL: TestValidateMixedVersionElements_drop_column_with_partial_index/pause_upgrade_and_resume_at_each_stage (22.60s) === RUN TestValidateMixedVersionElements_drop_column_with_partial_index/pause_upgrade_and_resume_at_each_stage/PostCommitNonRevertiblePhase_3_of_3 cumulative.go:1724: Error Trace: github.com/cockroachdb/cockroach/pkg/sql/schemachanger/sctest/cumulative.go:1724 GOROOT/src/runtime/asm_arm64.s:1172 Error: Received unexpected error: dial tcp 127.0.0.1:40613: connect: connection refused Test: TestValidateMixedVersionElements_drop_column_with_partial_index/pause_upgrade_and_resume_at_each_stage/PostCommitNonRevertiblePhase_3_of_3 --- FAIL: TestValidateMixedVersionElements_drop_column_with_partial_index/pause_upgrade_and_resume_at_each_stage/PostCommitNonRevertiblePhase_3_of_3 (4.76s) ``` <details><summary>Help</summary> <p> See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) </p> </details> /cc @cockroachdb/sql-foundations <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestValidateMixedVersionElements_drop_column_with_partial_index.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-29530
non_process
pkg sql schemachanger schemachanger test testvalidatemixedversionelements drop column with partial index failed pkg sql schemachanger schemachanger test testvalidatemixedversionelements drop column with partial index with on master run testvalidatemixedversionelements drop column with partial index test log scope go test logs captured to artifacts tmp tmp logtestvalidatemixedversionelements drop column with partial test log scope go use show logs to present logs inline datadriven go home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql schemachanger schemachanger test schemachanger test runfiles com github cockroachdb cockroach pkg sql schemachanger testdata end to end drop column with partial index drop column with partial index definition setup create table t i int primary key j int index j where i datadriven go home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql schemachanger schemachanger test schemachanger test runfiles com github cockroachdb cockroach pkg sql schemachanger testdata end to end drop column with partial index drop column with partial index definition stage exec insert into t i values stagekey insert into t i values stagekey datadriven go home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql schemachanger schemachanger test schemachanger test runfiles com github cockroachdb cockroach pkg sql schemachanger testdata end to end drop column with partial index drop column with partial index definition stage query select count successfulstagecount from t true datadriven go home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql schemachanger schemachanger test schemachanger test runfiles com github cockroachdb cockroach pkg sql schemachanger testdata end to end drop column with partial index drop column with partial index definition stage exec insert into t i values stagekey insert into t i values stagekey datadriven go home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql schemachanger schemachanger test schemachanger test runfiles com github cockroachdb cockroach pkg sql schemachanger testdata end to end drop column with partial index drop column with partial index definition stage query select count successfulstagecount from t true cont testvalidatemixedversionelements drop column with partial index panic go test log scope end test logs left over in artifacts tmp tmp logtestvalidatemixedversionelements drop column with partial fail testvalidatemixedversionelements drop column with partial index run testvalidatemixedversionelements drop column with partial index pause upgrade and resume at each stage fail testvalidatemixedversionelements drop column with partial index pause upgrade and resume at each stage run testvalidatemixedversionelements drop column with partial index pause upgrade and resume at each stage postcommitnonrevertiblephase of cumulative go error trace github com cockroachdb cockroach pkg sql schemachanger sctest cumulative go goroot src runtime asm s error received unexpected error dial tcp connect connection refused test testvalidatemixedversionelements drop column with partial index pause upgrade and resume at each stage postcommitnonrevertiblephase of fail testvalidatemixedversionelements drop column with partial index pause upgrade and resume at each stage postcommitnonrevertiblephase of help see also cc cockroachdb sql foundations jira issue crdb
0
787,020
27,701,928,353
IssuesEvent
2023-03-14 08:40:57
hoangnguyen92dn/survey-flutter-ic
https://api.github.com/repos/hoangnguyen92dn/survey-flutter-ic
closed
[UI] As a user, I can sign in with email and password
type: feature priority: high @0.1.0 epic: authentication
## Why Users must be authenticated before taking or viewing any surveys. Our surveys have access restrictions. Some surveys are targeted to a specific group of users only for a more accurate survey result. The mobile app should allow users to authenticate with their email and password. ## Acceptance Criteria - Display the Nimble logo. - Display the blurred background. - Display the Email text input. - Set the placeholder text to Email. - Display the Password text input. - Set the placeholder text to Password. - Mask the password input with the native-OS feature. - Display the Login button. - **DO NOT** display the Forgot Password button yet. We will work on this on another separate epic. ## Design ![image](https://user-images.githubusercontent.com/6950766/222625614-8c4ab3c5-480a-4932-9402-c228d0c19789.png) ## Resource https://www.figma.com/file/GjRPOjDyZ6f4EDL3wKarRK/Challenge---Mobile-App?node-id=31%3A195&t=2O4uXmma4d3QrhkU-4
1.0
[UI] As a user, I can sign in with email and password - ## Why Users must be authenticated before taking or viewing any surveys. Our surveys have access restrictions. Some surveys are targeted to a specific group of users only for a more accurate survey result. The mobile app should allow users to authenticate with their email and password. ## Acceptance Criteria - Display the Nimble logo. - Display the blurred background. - Display the Email text input. - Set the placeholder text to Email. - Display the Password text input. - Set the placeholder text to Password. - Mask the password input with the native-OS feature. - Display the Login button. - **DO NOT** display the Forgot Password button yet. We will work on this on another separate epic. ## Design ![image](https://user-images.githubusercontent.com/6950766/222625614-8c4ab3c5-480a-4932-9402-c228d0c19789.png) ## Resource https://www.figma.com/file/GjRPOjDyZ6f4EDL3wKarRK/Challenge---Mobile-App?node-id=31%3A195&t=2O4uXmma4d3QrhkU-4
non_process
as a user i can sign in with email and password why users must be authenticated before taking or viewing any surveys our surveys have access restrictions some surveys are targeted to a specific group of users only for a more accurate survey result the mobile app should allow users to authenticate with their email and password acceptance criteria display the nimble logo display the blurred background display the email text input set the placeholder text to email display the password text input set the placeholder text to password mask the password input with the native os feature display the login button do not display the forgot password button yet we will work on this on another separate epic design resource
0
118,245
15,262,598,455
IssuesEvent
2021-02-22 00:04:39
PyTorchLightning/pytorch-lightning
https://api.github.com/repos/PyTorchLightning/pytorch-lightning
closed
on_{validation,test}_epoch_end functions should have an outputs parameter
API / design duplicate enhancement help wanted
## 🚀 Feature https://github.com/PyTorchLightning/pytorch-lightning/blob/3b0e4e0b2bc5b62bba09df5976e1460774ae7337/pytorch_lightning/core/hooks.py#L255 https://github.com/PyTorchLightning/pytorch-lightning/blob/3b0e4e0b2bc5b62bba09df5976e1460774ae7337/pytorch_lightning/core/hooks.py#L267 Should have an `outputs` parameter as https://github.com/PyTorchLightning/pytorch-lightning/blob/3b0e4e0b2bc5b62bba09df5976e1460774ae7337/pytorch_lightning/core/hooks.py#L243
1.0
on_{validation,test}_epoch_end functions should have an outputs parameter - ## 🚀 Feature https://github.com/PyTorchLightning/pytorch-lightning/blob/3b0e4e0b2bc5b62bba09df5976e1460774ae7337/pytorch_lightning/core/hooks.py#L255 https://github.com/PyTorchLightning/pytorch-lightning/blob/3b0e4e0b2bc5b62bba09df5976e1460774ae7337/pytorch_lightning/core/hooks.py#L267 Should have an `outputs` parameter as https://github.com/PyTorchLightning/pytorch-lightning/blob/3b0e4e0b2bc5b62bba09df5976e1460774ae7337/pytorch_lightning/core/hooks.py#L243
non_process
on validation test epoch end functions should have an outputs parameter 🚀 feature should have an outputs parameter as
0
21,709
30,209,012,197
IssuesEvent
2023-07-05 11:29:18
camunda/issues
https://api.github.com/repos/camunda/issues
closed
Catch errors without errorCode
component:desktopModeler component:operate component:optimize component:webModeler component:zeebe-process-automation public kind:epic feature-parity version:8.2
### Value Proposition Statement Error catch events without error codes enable users to model a specific response for a known escalation code, and a general response for unknown escalation codes. ### User Problem Currently, I can only model Error catch events for known error codes. However, if a BPMN Error with an unknown error code is thrown, an incident is raised. ### User Stories As a user, I can model a general response to any BPMN Error thrown in a process. <sup>:robot: This issue is automatically synced from: [source](https://github.com/camunda/product-hub/issues/771)</sup> <!-- copiedFromSourceIssue: https://github.com/camunda/product-hub/issues/771 -->
1.0
Catch errors without errorCode - ### Value Proposition Statement Error catch events without error codes enable users to model a specific response for a known escalation code, and a general response for unknown escalation codes. ### User Problem Currently, I can only model Error catch events for known error codes. However, if a BPMN Error with an unknown error code is thrown, an incident is raised. ### User Stories As a user, I can model a general response to any BPMN Error thrown in a process. <sup>:robot: This issue is automatically synced from: [source](https://github.com/camunda/product-hub/issues/771)</sup> <!-- copiedFromSourceIssue: https://github.com/camunda/product-hub/issues/771 -->
process
catch errors without errorcode value proposition statement error catch events without error codes enable users to model a specific response for a known escalation code and a general response for unknown escalation codes user problem currently i can only model error catch events for known error codes however if a bpmn error with an unknown error code is thrown an incident is raised user stories as a user i can model a general response to any bpmn error thrown in a process robot this issue is automatically synced from
1
206,677
15,767,814,368
IssuesEvent
2021-03-31 16:30:27
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
ccl/importccl: TestProtectedTimestampsDuringImportInto failed
C-test-failure O-robot branch-master
[(ccl/importccl).TestProtectedTimestampsDuringImportInto failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2833638&tab=buildLog) on [master@c6125c3c5f4e416382c19adfaebe6c2190977190](https://github.com/cockroachdb/cockroach/commits/c6125c3c5f4e416382c19adfaebe6c2190977190): ``` === RUN TestProtectedTimestampsDuringImportInto test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logTestProtectedTimestampsDuringImportInto005524695 test_log_scope.go:74: use -show-logs to present logs inline n1: found liveness E210331 12:12:05.331931 550735 storage/cloudimpl/http_storage.go:195 [n1,job=645930425636913153,import-distsql-ingest] 1 HTTP:Req error: err=error response from server: 500 Internal Server Error "" (attempt 0) E210331 12:12:05.365749 550586 jobs/adopt.go:260 [n1] 2 job 645930425636913153: adoption completed with error error response from server: 500 Internal Server Error "" sql_runner.go:224: error executing 'SELECT * FROM foo': pq: relation "foo" is offline: importing E210331 12:12:05.377353 549479 jobs/registry.go:696 [-] 3 error claiming jobs: could not query jobs table: claim-jobs: context canceled E210331 12:12:05.379465 549479 jobs/registry.go:706 [-] 4 error processing claimed jobs: could not query for claimed jobs: select-running/get-claimed-jobs: context canceled panic.go:617: -- test log scope end -- test logs left over in: /go/src/github.com/cockroachdb/cockroach/artifacts/logTestProtectedTimestampsDuringImportInto005524695 --- FAIL: TestProtectedTimestampsDuringImportInto (3.52s) ``` <details><summary>More</summary><p> Parameters: - GOFLAGS=-json ``` make stressrace TESTS=TestProtectedTimestampsDuringImportInto PKG=./pkg/ccl/importccl TESTTIMEOUT=5m STRESSFLAGS='-timeout 5m' 2>&1 ``` [See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2ATestProtectedTimestampsDuringImportInto.%2A&sort=title&restgroup=false&display=lastcommented+project) <sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
1.0
ccl/importccl: TestProtectedTimestampsDuringImportInto failed - [(ccl/importccl).TestProtectedTimestampsDuringImportInto failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2833638&tab=buildLog) on [master@c6125c3c5f4e416382c19adfaebe6c2190977190](https://github.com/cockroachdb/cockroach/commits/c6125c3c5f4e416382c19adfaebe6c2190977190): ``` === RUN TestProtectedTimestampsDuringImportInto test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logTestProtectedTimestampsDuringImportInto005524695 test_log_scope.go:74: use -show-logs to present logs inline n1: found liveness E210331 12:12:05.331931 550735 storage/cloudimpl/http_storage.go:195 [n1,job=645930425636913153,import-distsql-ingest] 1 HTTP:Req error: err=error response from server: 500 Internal Server Error "" (attempt 0) E210331 12:12:05.365749 550586 jobs/adopt.go:260 [n1] 2 job 645930425636913153: adoption completed with error error response from server: 500 Internal Server Error "" sql_runner.go:224: error executing 'SELECT * FROM foo': pq: relation "foo" is offline: importing E210331 12:12:05.377353 549479 jobs/registry.go:696 [-] 3 error claiming jobs: could not query jobs table: claim-jobs: context canceled E210331 12:12:05.379465 549479 jobs/registry.go:706 [-] 4 error processing claimed jobs: could not query for claimed jobs: select-running/get-claimed-jobs: context canceled panic.go:617: -- test log scope end -- test logs left over in: /go/src/github.com/cockroachdb/cockroach/artifacts/logTestProtectedTimestampsDuringImportInto005524695 --- FAIL: TestProtectedTimestampsDuringImportInto (3.52s) ``` <details><summary>More</summary><p> Parameters: - GOFLAGS=-json ``` make stressrace TESTS=TestProtectedTimestampsDuringImportInto PKG=./pkg/ccl/importccl TESTTIMEOUT=5m STRESSFLAGS='-timeout 5m' 2>&1 ``` [See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2ATestProtectedTimestampsDuringImportInto.%2A&sort=title&restgroup=false&display=lastcommented+project) <sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
non_process
ccl importccl testprotectedtimestampsduringimportinto failed on run testprotectedtimestampsduringimportinto test log scope go test logs captured to go src github com cockroachdb cockroach artifacts test log scope go use show logs to present logs inline found liveness storage cloudimpl http storage go http req error err error response from server internal server error attempt jobs adopt go job adoption completed with error error response from server internal server error sql runner go error executing select from foo pq relation foo is offline importing jobs registry go error claiming jobs could not query jobs table claim jobs context canceled jobs registry go error processing claimed jobs could not query for claimed jobs select running get claimed jobs context canceled panic go test log scope end test logs left over in go src github com cockroachdb cockroach artifacts fail testprotectedtimestampsduringimportinto more parameters goflags json make stressrace tests testprotectedtimestampsduringimportinto pkg pkg ccl importccl testtimeout stressflags timeout powered by
0
98,631
16,387,781,418
IssuesEvent
2021-05-17 12:47:06
fitzinbox/Exomiser
https://api.github.com/repos/fitzinbox/Exomiser
opened
CVE-2020-36184 (High) detected in jackson-databind-2.9.8.jar
security vulnerability
## CVE-2020-36184 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.8.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: Exomiser/exomiser-data-genome/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar,Exomiser/exomiser-cli/target/lib/jackson-databind-2.9.8.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar,canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.9.8.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/fitzinbox/Exomiser/commit/3a0ae5a0b72ae7a7e59a638af862c28aa80dcdf6">3a0ae5a0b72ae7a7e59a638af862c28aa80dcdf6</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.tomcat.dbcp.dbcp2.datasources.PerUserPoolDataSource. <p>Publish Date: 2021-01-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36184>CVE-2020-36184</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2998">https://github.com/FasterXML/jackson-databind/issues/2998</a></p> <p>Release Date: 2021-01-06</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-36184 (High) detected in jackson-databind-2.9.8.jar - ## CVE-2020-36184 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.8.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: Exomiser/exomiser-data-genome/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar,Exomiser/exomiser-cli/target/lib/jackson-databind-2.9.8.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar,canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.9.8.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/fitzinbox/Exomiser/commit/3a0ae5a0b72ae7a7e59a638af862c28aa80dcdf6">3a0ae5a0b72ae7a7e59a638af862c28aa80dcdf6</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.tomcat.dbcp.dbcp2.datasources.PerUserPoolDataSource. <p>Publish Date: 2021-01-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36184>CVE-2020-36184</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2998">https://github.com/FasterXML/jackson-databind/issues/2998</a></p> <p>Release Date: 2021-01-06</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file exomiser exomiser data genome pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar exomiser exomiser cli target lib jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar canner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache tomcat dbcp datasources peruserpooldatasource publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind step up your open source security game with whitesource
0
2,814
5,738,746,685
IssuesEvent
2017-04-23 07:58:13
dita-ot/dita-ot
https://api.github.com/repos/dita-ot/dita-ot
closed
Keydef containing uplevels
bug P2 preprocess/keyref
In my DITA Map I define a key to an image: `<keydef keys="arch_diagram" href="../../introduction/images/arch_diagram" format="png"/>` With the keyref: `<image keyref="arch_diagram"/>` It's not expanded in 2.3.3 or 2.4.6 but is in 1.8. Other keyrefs to keys without uplevels and all other images are impacted. Also, I have the same issue as #2420, except the conref has also uplevels.
1.0
Keydef containing uplevels - In my DITA Map I define a key to an image: `<keydef keys="arch_diagram" href="../../introduction/images/arch_diagram" format="png"/>` With the keyref: `<image keyref="arch_diagram"/>` It's not expanded in 2.3.3 or 2.4.6 but is in 1.8. Other keyrefs to keys without uplevels and all other images are impacted. Also, I have the same issue as #2420, except the conref has also uplevels.
process
keydef containing uplevels in my dita map i define a key to an image with the keyref it s not expanded in or but is in other keyrefs to keys without uplevels and all other images are impacted also i have the same issue as except the conref has also uplevels
1
35,208
30,841,392,211
IssuesEvent
2023-08-02 10:54:00
woowacourse-teams/2023-zipgo
https://api.github.com/repos/woowacourse-teams/2023-zipgo
closed
HTTPS + 도메인 연결
🌍 Infrastructure 🕋 Backend 🧚🏻‍♀️ Support
### 🧚🏻‍♀️ Describe https와 도메인을 연결합니다 ### ✅ Tasks - [x] nginx + certbot을 이용한 도메인과 https 연결 - [x] nginx - [x] 도메인 연결 - [x] https 연결 ### 🕖 예상 작업 소요 시간 - 3시간 - 늦어도 오늘 안엔 가능 ### 🙋🏻 More [적용 완료!](https://zipgo.pet)
1.0
HTTPS + 도메인 연결 - ### 🧚🏻‍♀️ Describe https와 도메인을 연결합니다 ### ✅ Tasks - [x] nginx + certbot을 이용한 도메인과 https 연결 - [x] nginx - [x] 도메인 연결 - [x] https 연결 ### 🕖 예상 작업 소요 시간 - 3시간 - 늦어도 오늘 안엔 가능 ### 🙋🏻 More [적용 완료!](https://zipgo.pet)
non_process
https 도메인 연결 🧚🏻‍♀️ describe https와 도메인을 연결합니다 ✅ tasks nginx certbot을 이용한 도메인과 https 연결 nginx 도메인 연결 https 연결 🕖 예상 작업 소요 시간 늦어도 오늘 안엔 가능 🙋🏻 more
0
2,396
5,191,905,518
IssuesEvent
2017-01-22 01:37:01
mitchellh/packer
https://api.github.com/repos/mitchellh/packer
closed
Packer crash with vagrant post-processor
crash post-processor/vagrant
https://gist.github.com/cbednarski/26e40b91a1dba233cc78 No crash when I remove the use of the vagrant post processor.
1.0
Packer crash with vagrant post-processor - https://gist.github.com/cbednarski/26e40b91a1dba233cc78 No crash when I remove the use of the vagrant post processor.
process
packer crash with vagrant post processor no crash when i remove the use of the vagrant post processor
1
753,760
26,360,642,490
IssuesEvent
2023-01-11 13:08:55
OpenNebula/one
https://api.github.com/repos/OpenNebula/one
closed
Update VM templates after renaming the components referenced
Type: Backlog Category: Core & System Type: Feature Sponsored Category: API Priority: Low
**Description** After issuing a rename API Call, like `one.image.rename`, the object new name doesn't appear updated on the VM Templates that reference it. **Use case** The calls could issue the name update for consistency purposes. Even though the operation remains fully functional due to the object ID being used for it. **Additional Context** This behavior can be implemented with API Hooks triggering on the rename calls. <!--////////////////////////////////////////////--> <!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM --> <!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS --> <!-- PROGRESS WILL BE REFLECTED HERE --> <!--////////////////////////////////////////////--> ## Progress Status - [ ] Code committed - [ ] Testing - QA - [ ] Documentation (Release notes - resolved issues, compatibility, known issues)
1.0
Update VM templates after renaming the components referenced - **Description** After issuing a rename API Call, like `one.image.rename`, the object new name doesn't appear updated on the VM Templates that reference it. **Use case** The calls could issue the name update for consistency purposes. Even though the operation remains fully functional due to the object ID being used for it. **Additional Context** This behavior can be implemented with API Hooks triggering on the rename calls. <!--////////////////////////////////////////////--> <!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM --> <!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS --> <!-- PROGRESS WILL BE REFLECTED HERE --> <!--////////////////////////////////////////////--> ## Progress Status - [ ] Code committed - [ ] Testing - QA - [ ] Documentation (Release notes - resolved issues, compatibility, known issues)
non_process
update vm templates after renaming the components referenced description after issuing a rename api call like one image rename the object new name doesn t appear updated on the vm templates that reference it use case the calls could issue the name update for consistency purposes even though the operation remains fully functional due to the object id being used for it additional context this behavior can be implemented with api hooks triggering on the rename calls progress status code committed testing qa documentation release notes resolved issues compatibility known issues
0
39,728
20,170,913,819
IssuesEvent
2022-02-10 10:19:14
milessabin/shapeless
https://api.github.com/repos/milessabin/shapeless
closed
performance issue with combined `Length` and `ToSizedHList` implicit derivation
Performance
There seems to be a performance problem with the combined implicit derivation of `Length` and `ToSizedHList`. When `n` grows, the compilation time becomes quickly unpractical. I reproduced the problem with shapeless 2.3.7 and scala 2.12.14/2.13.6 . ```scala val n = Nat(100) val f = Fill[n.N, String] def test[N <: Nat](implicit l: Length.Aux[f.Out, N], v: ToSizedHList[List, String, N]): Unit = {} // perf issue test // no perf issue test[n.N] val l = Length[f.Out] ToSizedHList[List, String, l.Out] ```
True
performance issue with combined `Length` and `ToSizedHList` implicit derivation - There seems to be a performance problem with the combined implicit derivation of `Length` and `ToSizedHList`. When `n` grows, the compilation time becomes quickly unpractical. I reproduced the problem with shapeless 2.3.7 and scala 2.12.14/2.13.6 . ```scala val n = Nat(100) val f = Fill[n.N, String] def test[N <: Nat](implicit l: Length.Aux[f.Out, N], v: ToSizedHList[List, String, N]): Unit = {} // perf issue test // no perf issue test[n.N] val l = Length[f.Out] ToSizedHList[List, String, l.Out] ```
non_process
performance issue with combined length and tosizedhlist implicit derivation there seems to be a performance problem with the combined implicit derivation of length and tosizedhlist when n grows the compilation time becomes quickly unpractical i reproduced the problem with shapeless and scala scala val n nat val f fill def test implicit l length aux v tosizedhlist unit perf issue test no perf issue test val l length tosizedhlist
0
22,305
30,859,670,789
IssuesEvent
2023-08-03 01:08:02
emily-writes-poems/emily-writes-poems-processing
https://api.github.com/repos/emily-writes-poems/emily-writes-poems-processing
closed
editing poems in existing collections
script migration processing
being able to add/remove poems from collections. currently the collection poems list is stored as 2 arrays in Mongo (poem_ids and poem_titles) easiest way is probably to pass around an array/list of poem ids and then revise the poem_titles array with a query from main poems coll?
1.0
editing poems in existing collections - being able to add/remove poems from collections. currently the collection poems list is stored as 2 arrays in Mongo (poem_ids and poem_titles) easiest way is probably to pass around an array/list of poem ids and then revise the poem_titles array with a query from main poems coll?
process
editing poems in existing collections being able to add remove poems from collections currently the collection poems list is stored as arrays in mongo poem ids and poem titles easiest way is probably to pass around an array list of poem ids and then revise the poem titles array with a query from main poems coll
1
205,762
15,686,436,957
IssuesEvent
2021-03-25 12:31:38
Slimefun/Slimefun4
https://api.github.com/repos/Slimefun/Slimefun4
opened
Automated Ignition Chamber dupe items
🎯 Needs testing 🐞 Bug Report
## :round_pushpin: Description (REQUIRED) When the Automated Ignition Chamber is broken, items from it are doubled ## :bookmark_tabs: Steps to reproduce the Issue (REQUIRED) https://youtu.be/6hAKlnGN41A ## :bulb: Expected behavior (REQUIRED) Items do not multiply ## :compass: Environment (REQUIRED) ![image](https://user-images.githubusercontent.com/21084913/112472806-2336a680-8da0-11eb-805a-5aca1e8490a5.png)
1.0
Automated Ignition Chamber dupe items - ## :round_pushpin: Description (REQUIRED) When the Automated Ignition Chamber is broken, items from it are doubled ## :bookmark_tabs: Steps to reproduce the Issue (REQUIRED) https://youtu.be/6hAKlnGN41A ## :bulb: Expected behavior (REQUIRED) Items do not multiply ## :compass: Environment (REQUIRED) ![image](https://user-images.githubusercontent.com/21084913/112472806-2336a680-8da0-11eb-805a-5aca1e8490a5.png)
non_process
automated ignition chamber dupe items round pushpin description required when the automated ignition chamber is broken items from it are doubled bookmark tabs steps to reproduce the issue required bulb expected behavior required items do not multiply compass environment required
0
296,674
9,125,136,293
IssuesEvent
2019-02-24 11:00:46
python/mypy
https://api.github.com/repos/python/mypy
closed
Wrong type inferred for union containing restricted type variable
bug false-positive priority-1-normal topic-union-types
``` from typing import Generic, TypeVar, Union T = TypeVar('T') class G(Generic[T]): pass class A(object): pass class B(object): pass g_a = None # type: G[A] g_b = None # type: G[B] AB = TypeVar('AB', A, B) def f(x): # type: (Union[G[AB],AB]) -> G[AB] pass f(A()) f(B()) f(g_a) f(g_b) # E: Argument 1 to "f" has incompatible type G[B]; expected "Union[G[A], A]" ``` The error is incorrect--`f(g_b)` should be valid. Swapping `A` and `B` in the definition of `AB` results in an error on the `f(g_a)` call instead. I'm not sure this title clearly describes the issue; improvements welcome.
1.0
Wrong type inferred for union containing restricted type variable - ``` from typing import Generic, TypeVar, Union T = TypeVar('T') class G(Generic[T]): pass class A(object): pass class B(object): pass g_a = None # type: G[A] g_b = None # type: G[B] AB = TypeVar('AB', A, B) def f(x): # type: (Union[G[AB],AB]) -> G[AB] pass f(A()) f(B()) f(g_a) f(g_b) # E: Argument 1 to "f" has incompatible type G[B]; expected "Union[G[A], A]" ``` The error is incorrect--`f(g_b)` should be valid. Swapping `A` and `B` in the definition of `AB` results in an error on the `f(g_a)` call instead. I'm not sure this title clearly describes the issue; improvements welcome.
non_process
wrong type inferred for union containing restricted type variable from typing import generic typevar union t typevar t class g generic pass class a object pass class b object pass g a none type g g b none type g ab typevar ab a b def f x type union ab g pass f a f b f g a f g b e argument to f has incompatible type g expected union a the error is incorrect f g b should be valid swapping a and b in the definition of ab results in an error on the f g a call instead i m not sure this title clearly describes the issue improvements welcome
0
3,169
6,224,106,515
IssuesEvent
2017-07-10 13:36:07
dzhw/zofar
https://api.github.com/repos/dzhw/zofar
opened
Monitoring-Bridge
category: technical.processes prio: 9999 status: discussion type: backlog.item
Weiterreichung von Metrikdaten. Durch Trennung von HIS-IT: Überlegung über den Aufbau einer Virtualisierung eigener Metriken! engl. tba
1.0
Monitoring-Bridge - Weiterreichung von Metrikdaten. Durch Trennung von HIS-IT: Überlegung über den Aufbau einer Virtualisierung eigener Metriken! engl. tba
process
monitoring bridge weiterreichung von metrikdaten durch trennung von his it überlegung über den aufbau einer virtualisierung eigener metriken engl tba
1
7,953
11,137,562,938
IssuesEvent
2019-12-20 19:42:22
openopps/openopps-platform
https://api.github.com/repos/openopps/openopps-platform
closed
Display education in sorted order on application review page
Apply Process Requirements Ready State Dept.
Who: Student applicants What: Display education data by sort preference on the application review Why: to allow applicants to review the application correctly Acceptance Criteria: - The education data will now have a sort order (either default from USAJOBS or resorted in open opps). Display the education data in the correct sort order on the application review page
1.0
Display education in sorted order on application review page - Who: Student applicants What: Display education data by sort preference on the application review Why: to allow applicants to review the application correctly Acceptance Criteria: - The education data will now have a sort order (either default from USAJOBS or resorted in open opps). Display the education data in the correct sort order on the application review page
process
display education in sorted order on application review page who student applicants what display education data by sort preference on the application review why to allow applicants to review the application correctly acceptance criteria the education data will now have a sort order either default from usajobs or resorted in open opps display the education data in the correct sort order on the application review page
1
1,046
3,513,113,129
IssuesEvent
2016-01-11 08:33:04
nodejs/node
https://api.github.com/repos/nodejs/node
closed
Error in process._tickCallback
process
After upgrading from node v0.12.7 to v4.2.3 our application crashes every few hours or after 1 or 2 days with the following uncaught exception: ``` TypeError: Cannot read property 'callback' of undefined at process._tickCallback (node.js:341:26) ``` Could this be a bug in node.js? The code in src/node.js doesn't look like `tock` is expected to be undefined: ```javascript // Run callbacks that have no domain. // Using domains will cause this to be overridden. function _tickCallback() { var callback, args, tock; do { while (tickInfo[kIndex] < tickInfo[kLength]) { tock = nextTickQueue[tickInfo[kIndex]++]; callback = tock.callback; // <------- line 341. tock is undefined args = tock.args; ```
1.0
Error in process._tickCallback - After upgrading from node v0.12.7 to v4.2.3 our application crashes every few hours or after 1 or 2 days with the following uncaught exception: ``` TypeError: Cannot read property 'callback' of undefined at process._tickCallback (node.js:341:26) ``` Could this be a bug in node.js? The code in src/node.js doesn't look like `tock` is expected to be undefined: ```javascript // Run callbacks that have no domain. // Using domains will cause this to be overridden. function _tickCallback() { var callback, args, tock; do { while (tickInfo[kIndex] < tickInfo[kLength]) { tock = nextTickQueue[tickInfo[kIndex]++]; callback = tock.callback; // <------- line 341. tock is undefined args = tock.args; ```
process
error in process tickcallback after upgrading from node to our application crashes every few hours or after or days with the following uncaught exception typeerror cannot read property callback of undefined at process tickcallback node js could this be a bug in node js the code in src node js doesn t look like tock is expected to be undefined javascript run callbacks that have no domain using domains will cause this to be overridden function tickcallback var callback args tock do while tickinfo tickinfo tock nexttickqueue callback tock callback line tock is undefined args tock args
1
21,604
30,005,553,287
IssuesEvent
2023-06-26 12:13:31
open-telemetry/opentelemetry-collector-contrib
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
closed
processor: add resource_attributes group in metadata.yaml files
enhancement processor/k8sattributes processor/resourcedetection cmd/mdatagen
### Component(s) k8sattributesprocessor, resourcedetectionprocessor ### Describe the issue you're reporting The group `resource_attributes` was introduced in the following [PR](https://github.com/open-telemetry/opentelemetry-collector-contrib/pull/21664). As of today, no processors use the `resource_attributes` group and its generated config structs. What I'm proposing is adding the `resource_attributes` to processors where this is relevant: - [x] [k8sattributesprocessor](https://github.com/open-telemetry/opentelemetry-collector-contrib/issues/23152) - [x] [resourcedetectionprocessor](https://github.com/open-telemetry/opentelemetry-collector-contrib/issues/21482)
2.0
processor: add resource_attributes group in metadata.yaml files - ### Component(s) k8sattributesprocessor, resourcedetectionprocessor ### Describe the issue you're reporting The group `resource_attributes` was introduced in the following [PR](https://github.com/open-telemetry/opentelemetry-collector-contrib/pull/21664). As of today, no processors use the `resource_attributes` group and its generated config structs. What I'm proposing is adding the `resource_attributes` to processors where this is relevant: - [x] [k8sattributesprocessor](https://github.com/open-telemetry/opentelemetry-collector-contrib/issues/23152) - [x] [resourcedetectionprocessor](https://github.com/open-telemetry/opentelemetry-collector-contrib/issues/21482)
process
processor add resource attributes group in metadata yaml files component s resourcedetectionprocessor describe the issue you re reporting the group resource attributes was introduced in the following as of today no processors use the resource attributes group and its generated config structs what i m proposing is adding the resource attributes to processors where this is relevant
1
370,922
10,958,539,906
IssuesEvent
2019-11-27 09:36:39
krzychu124/Cities-Skylines-Traffic-Manager-President-Edition
https://api.github.com/repos/krzychu124/Cities-Skylines-Traffic-Manager-President-Edition
opened
Speed limits window too constrained on QHD resolution
BUG SPEED LIMITS UI confirmed low priority
On this resolution: ![image](https://user-images.githubusercontent.com/1386719/69711681-6116b280-10f9-11ea-9cea-6172b19cf879.png) The speed limits panel is too constrained and can't be dragged further towards the bottom-right than shown: ![image](https://user-images.githubusercontent.com/1386719/69711588-3af11280-10f9-11ea-948a-9a6d71d50ea3.png)
1.0
Speed limits window too constrained on QHD resolution - On this resolution: ![image](https://user-images.githubusercontent.com/1386719/69711681-6116b280-10f9-11ea-9cea-6172b19cf879.png) The speed limits panel is too constrained and can't be dragged further towards the bottom-right than shown: ![image](https://user-images.githubusercontent.com/1386719/69711588-3af11280-10f9-11ea-948a-9a6d71d50ea3.png)
non_process
speed limits window too constrained on qhd resolution on this resolution the speed limits panel is too constrained and can t be dragged further towards the bottom right than shown
0
3,104
6,111,644,469
IssuesEvent
2017-06-21 17:30:59
NeuroTechX/eeg-101
https://api.github.com/repos/NeuroTechX/eeg-101
closed
Create ClassifierDataSource class to integrate with ClassifierModule and the GaussianNaiveBayesClassifier
Java Signal Processing
The job of this class will be to do three things: 1. Contain a MuseDataListener and runnable thread that fills up a buffer with streaming EEG data 2. Pre-process data for classifier - filter - epoch - FFT - extract band powers 3. Concatenate data into appropriate format to pass to classifier
1.0
Create ClassifierDataSource class to integrate with ClassifierModule and the GaussianNaiveBayesClassifier - The job of this class will be to do three things: 1. Contain a MuseDataListener and runnable thread that fills up a buffer with streaming EEG data 2. Pre-process data for classifier - filter - epoch - FFT - extract band powers 3. Concatenate data into appropriate format to pass to classifier
process
create classifierdatasource class to integrate with classifiermodule and the gaussiannaivebayesclassifier the job of this class will be to do three things contain a musedatalistener and runnable thread that fills up a buffer with streaming eeg data pre process data for classifier filter epoch fft extract band powers concatenate data into appropriate format to pass to classifier
1
816,289
30,596,510,264
IssuesEvent
2023-07-21 23:00:09
googleapis/python-storage
https://api.github.com/repos/googleapis/python-storage
closed
tests.system.test_bucket: test_bucket_list_blobs_paginated_w_offset failed
api: storage type: bug priority: p2 flakybot: issue flakybot: flaky
Note: #969 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky. ---- commit: f9462179f4a4b08eea7471a5ffb4aa5071fc5a5e buildURL: [Build Status](https://source.cloud.google.com/results/invocations/ee1c078c-b3c3-4ce6-9627-cae04bcaee3d), [Sponge](http://sponge2/ee1c078c-b3c3-4ce6-9627-cae04bcaee3d) status: failed <details><summary>Test output</summary><br><pre>storage_client = <google.cloud.storage.client.Client object at 0x7f22b672bca0> listable_bucket_name = 'gcp-systest-listable-1689908513911' file_data = {'big': {'hash': b'cEome4a+NYd7YIXzXQnR5Q==', 'path': '/tmpfs/src/github/python-storage/tests/data/five-point-one-mb-f...g'}, 'simple': {'hash': b'3Hkwjv2WvCnKjNR6Z3CboA==', 'path': '/tmpfs/src/github/python-storage/tests/data/simple.txt'}} @pytest.fixture(scope="session") def listable_bucket(storage_client, listable_bucket_name, file_data): bucket = storage_client.bucket(listable_bucket_name) > _helpers.retry_429_503(bucket.create)() tests/system/conftest.py:105: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ .nox/system-3-8/lib/python3.8/site-packages/test_utils/retry.py:100: in wrapped_function return to_wrap(*args, **kwargs) google/cloud/storage/bucket.py:972: in create client.create_bucket( google/cloud/storage/client.py:961: in create_bucket api_response = self._post_resource( google/cloud/storage/client.py:625: in _post_resource return self._connection.api_request( google/cloud/storage/_http.py:72: in api_request return call() .nox/system-3-8/lib/python3.8/site-packages/google/api_core/retry.py:349: in retry_wrapped_func return retry_target( .nox/system-3-8/lib/python3.8/site-packages/google/api_core/retry.py:191: in retry_target return target() _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <google.cloud.storage._http.Connection object at 0x7f22b683b520> method = 'POST', path = '/b', query_params = {'project': 'precise-truck-742'} data = '{"name": "gcp-systest-listable-1689908513911"}' content_type = 'application/json', headers = None, api_base_url = None api_version = None, expect_json = True _target_object = <Bucket: gcp-systest-listable-1689908513911>, timeout = 60 extra_api_info = 'gccl-invocation-id/f0b8f74a-b400-440a-a7d5-0869c725a434' def api_request( self, method, path, query_params=None, data=None, content_type=None, headers=None, api_base_url=None, api_version=None, expect_json=True, _target_object=None, timeout=_DEFAULT_TIMEOUT, extra_api_info=None, ): """Make a request over the HTTP transport to the API. You shouldn't need to use this method, but if you plan to interact with the API using these primitives, this is the correct one to use. :type method: str :param method: The HTTP method name (ie, ``GET``, ``POST``, etc). Required. :type path: str :param path: The path to the resource (ie, ``'/b/bucket-name'``). Required. :type query_params: dict or list :param query_params: A dictionary of keys and values (or list of key-value pairs) to insert into the query string of the URL. :type data: str :param data: The data to send as the body of the request. Default is the empty string. :type content_type: str :param content_type: The proper MIME type of the data provided. Default is None. :type headers: dict :param headers: extra HTTP headers to be sent with the request. :type api_base_url: str :param api_base_url: The base URL for the API endpoint. Typically you won't have to provide this. Default is the standard API base URL. :type api_version: str :param api_version: The version of the API to call. Typically you shouldn't provide this and instead use the default for the library. Default is the latest API version supported by google-cloud-python. :type expect_json: bool :param expect_json: If True, this method will try to parse the response as JSON and raise an exception if that cannot be done. Default is True. :type _target_object: :class:`object` :param _target_object: (Optional) Protected argument to be used by library callers. This can allow custom behavior, for example, to defer an HTTP request and complete initialization of the object at a later time. :type timeout: float or tuple :param timeout: (optional) The amount of time, in seconds, to wait for the server response. Can also be passed as a tuple (connect_timeout, read_timeout). See :meth:`requests.Session.request` documentation for details. :type extra_api_info: string :param extra_api_info: (optional) Extra api info to be appended to the X-Goog-API-Client header :raises ~google.cloud.exceptions.GoogleCloudError: if the response code is not 200 OK. :raises ValueError: if the response content type is not JSON. :rtype: dict or str :returns: The API response payload, either as a raw string or a dictionary if the response is valid JSON. """ url = self.build_api_url( path=path, query_params=query_params, api_base_url=api_base_url, api_version=api_version, ) # Making the executive decision that any dictionary # data will be sent properly as JSON. if data and isinstance(data, dict): data = json.dumps(data) content_type = "application/json" response = self._make_request( method=method, url=url, data=data, content_type=content_type, headers=headers, target_object=_target_object, timeout=timeout, extra_api_info=extra_api_info, ) if not 200 <= response.status_code < 300: > raise exceptions.from_http_response(response) E google.api_core.exceptions.Conflict: 409 POST https://storage-preprod-test-unified.googleusercontent.com/storage/v1_preprod/b?project=precise-truck-742&prettyPrint=false: Your previous request to create the named bucket succeeded and you already own it. .nox/system-3-8/lib/python3.8/site-packages/google/cloud/_http/__init__.py:494: Conflict</pre></details>
1.0
tests.system.test_bucket: test_bucket_list_blobs_paginated_w_offset failed - Note: #969 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky. ---- commit: f9462179f4a4b08eea7471a5ffb4aa5071fc5a5e buildURL: [Build Status](https://source.cloud.google.com/results/invocations/ee1c078c-b3c3-4ce6-9627-cae04bcaee3d), [Sponge](http://sponge2/ee1c078c-b3c3-4ce6-9627-cae04bcaee3d) status: failed <details><summary>Test output</summary><br><pre>storage_client = <google.cloud.storage.client.Client object at 0x7f22b672bca0> listable_bucket_name = 'gcp-systest-listable-1689908513911' file_data = {'big': {'hash': b'cEome4a+NYd7YIXzXQnR5Q==', 'path': '/tmpfs/src/github/python-storage/tests/data/five-point-one-mb-f...g'}, 'simple': {'hash': b'3Hkwjv2WvCnKjNR6Z3CboA==', 'path': '/tmpfs/src/github/python-storage/tests/data/simple.txt'}} @pytest.fixture(scope="session") def listable_bucket(storage_client, listable_bucket_name, file_data): bucket = storage_client.bucket(listable_bucket_name) > _helpers.retry_429_503(bucket.create)() tests/system/conftest.py:105: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ .nox/system-3-8/lib/python3.8/site-packages/test_utils/retry.py:100: in wrapped_function return to_wrap(*args, **kwargs) google/cloud/storage/bucket.py:972: in create client.create_bucket( google/cloud/storage/client.py:961: in create_bucket api_response = self._post_resource( google/cloud/storage/client.py:625: in _post_resource return self._connection.api_request( google/cloud/storage/_http.py:72: in api_request return call() .nox/system-3-8/lib/python3.8/site-packages/google/api_core/retry.py:349: in retry_wrapped_func return retry_target( .nox/system-3-8/lib/python3.8/site-packages/google/api_core/retry.py:191: in retry_target return target() _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <google.cloud.storage._http.Connection object at 0x7f22b683b520> method = 'POST', path = '/b', query_params = {'project': 'precise-truck-742'} data = '{"name": "gcp-systest-listable-1689908513911"}' content_type = 'application/json', headers = None, api_base_url = None api_version = None, expect_json = True _target_object = <Bucket: gcp-systest-listable-1689908513911>, timeout = 60 extra_api_info = 'gccl-invocation-id/f0b8f74a-b400-440a-a7d5-0869c725a434' def api_request( self, method, path, query_params=None, data=None, content_type=None, headers=None, api_base_url=None, api_version=None, expect_json=True, _target_object=None, timeout=_DEFAULT_TIMEOUT, extra_api_info=None, ): """Make a request over the HTTP transport to the API. You shouldn't need to use this method, but if you plan to interact with the API using these primitives, this is the correct one to use. :type method: str :param method: The HTTP method name (ie, ``GET``, ``POST``, etc). Required. :type path: str :param path: The path to the resource (ie, ``'/b/bucket-name'``). Required. :type query_params: dict or list :param query_params: A dictionary of keys and values (or list of key-value pairs) to insert into the query string of the URL. :type data: str :param data: The data to send as the body of the request. Default is the empty string. :type content_type: str :param content_type: The proper MIME type of the data provided. Default is None. :type headers: dict :param headers: extra HTTP headers to be sent with the request. :type api_base_url: str :param api_base_url: The base URL for the API endpoint. Typically you won't have to provide this. Default is the standard API base URL. :type api_version: str :param api_version: The version of the API to call. Typically you shouldn't provide this and instead use the default for the library. Default is the latest API version supported by google-cloud-python. :type expect_json: bool :param expect_json: If True, this method will try to parse the response as JSON and raise an exception if that cannot be done. Default is True. :type _target_object: :class:`object` :param _target_object: (Optional) Protected argument to be used by library callers. This can allow custom behavior, for example, to defer an HTTP request and complete initialization of the object at a later time. :type timeout: float or tuple :param timeout: (optional) The amount of time, in seconds, to wait for the server response. Can also be passed as a tuple (connect_timeout, read_timeout). See :meth:`requests.Session.request` documentation for details. :type extra_api_info: string :param extra_api_info: (optional) Extra api info to be appended to the X-Goog-API-Client header :raises ~google.cloud.exceptions.GoogleCloudError: if the response code is not 200 OK. :raises ValueError: if the response content type is not JSON. :rtype: dict or str :returns: The API response payload, either as a raw string or a dictionary if the response is valid JSON. """ url = self.build_api_url( path=path, query_params=query_params, api_base_url=api_base_url, api_version=api_version, ) # Making the executive decision that any dictionary # data will be sent properly as JSON. if data and isinstance(data, dict): data = json.dumps(data) content_type = "application/json" response = self._make_request( method=method, url=url, data=data, content_type=content_type, headers=headers, target_object=_target_object, timeout=timeout, extra_api_info=extra_api_info, ) if not 200 <= response.status_code < 300: > raise exceptions.from_http_response(response) E google.api_core.exceptions.Conflict: 409 POST https://storage-preprod-test-unified.googleusercontent.com/storage/v1_preprod/b?project=precise-truck-742&prettyPrint=false: Your previous request to create the named bucket succeeded and you already own it. .nox/system-3-8/lib/python3.8/site-packages/google/cloud/_http/__init__.py:494: Conflict</pre></details>
non_process
tests system test bucket test bucket list blobs paginated w offset failed note was also for this test but it was closed more than days ago so i didn t mark it flaky commit buildurl status failed test output storage client listable bucket name gcp systest listable file data big hash b path tmpfs src github python storage tests data five point one mb f g simple hash b path tmpfs src github python storage tests data simple txt pytest fixture scope session def listable bucket storage client listable bucket name file data bucket storage client bucket listable bucket name helpers retry bucket create tests system conftest py nox system lib site packages test utils retry py in wrapped function return to wrap args kwargs google cloud storage bucket py in create client create bucket google cloud storage client py in create bucket api response self post resource google cloud storage client py in post resource return self connection api request google cloud storage http py in api request return call nox system lib site packages google api core retry py in retry wrapped func return retry target nox system lib site packages google api core retry py in retry target return target self method post path b query params project precise truck data name gcp systest listable content type application json headers none api base url none api version none expect json true target object timeout extra api info gccl invocation id def api request self method path query params none data none content type none headers none api base url none api version none expect json true target object none timeout default timeout extra api info none make a request over the http transport to the api you shouldn t need to use this method but if you plan to interact with the api using these primitives this is the correct one to use type method str param method the http method name ie get post etc required type path str param path the path to the resource ie b bucket name required type query params dict or list param query params a dictionary of keys and values or list of key value pairs to insert into the query string of the url type data str param data the data to send as the body of the request default is the empty string type content type str param content type the proper mime type of the data provided default is none type headers dict param headers extra http headers to be sent with the request type api base url str param api base url the base url for the api endpoint typically you won t have to provide this default is the standard api base url type api version str param api version the version of the api to call typically you shouldn t provide this and instead use the default for the library default is the latest api version supported by google cloud python type expect json bool param expect json if true this method will try to parse the response as json and raise an exception if that cannot be done default is true type target object class object param target object optional protected argument to be used by library callers this can allow custom behavior for example to defer an http request and complete initialization of the object at a later time type timeout float or tuple param timeout optional the amount of time in seconds to wait for the server response can also be passed as a tuple connect timeout read timeout see meth requests session request documentation for details type extra api info string param extra api info optional extra api info to be appended to the x goog api client header raises google cloud exceptions googleclouderror if the response code is not ok raises valueerror if the response content type is not json rtype dict or str returns the api response payload either as a raw string or a dictionary if the response is valid json url self build api url path path query params query params api base url api base url api version api version making the executive decision that any dictionary data will be sent properly as json if data and isinstance data dict data json dumps data content type application json response self make request method method url url data data content type content type headers headers target object target object timeout timeout extra api info extra api info if not response status code raise exceptions from http response response e google api core exceptions conflict post your previous request to create the named bucket succeeded and you already own it nox system lib site packages google cloud http init py conflict
0
16,094
20,263,434,083
IssuesEvent
2022-02-15 09:48:47
quark-engine/quark-engine
https://api.github.com/repos/quark-engine/quark-engine
closed
Update README to indicate the compatible Rizin versions
work-in-progress issue-processing-state-04
**Is your feature request related to a problem? Please describe.** An API change in Rizin v0.3.0 has led the Rizin-based analysis to fail. To avoid users experiencing errors similar to #305, we need to indicate this restriction before fixing it. **Describe the solution you'd like.** Indicate the compatible Rizin versions in the README. For now, Quark only works with Rizin v0.2.0 and v0.2.1.
1.0
Update README to indicate the compatible Rizin versions - **Is your feature request related to a problem? Please describe.** An API change in Rizin v0.3.0 has led the Rizin-based analysis to fail. To avoid users experiencing errors similar to #305, we need to indicate this restriction before fixing it. **Describe the solution you'd like.** Indicate the compatible Rizin versions in the README. For now, Quark only works with Rizin v0.2.0 and v0.2.1.
process
update readme to indicate the compatible rizin versions is your feature request related to a problem please describe an api change in rizin has led the rizin based analysis to fail to avoid users experiencing errors similar to we need to indicate this restriction before fixing it describe the solution you d like indicate the compatible rizin versions in the readme for now quark only works with rizin and
1
17,762
23,691,256,724
IssuesEvent
2022-08-29 11:01:28
pyanodon/pybugreports
https://api.github.com/repos/pyanodon/pybugreports
closed
Inserter mode gets stomped excessively
bug mod:pycoalprocessing
### Mod source PyAE Beta ### Which mod are you having an issue with? - [ ] pyalienlife - [ ] pyalternativeenergy - [X] pycoalprocessing - [ ] pyfusionenergy - [ ] pyhightech - [ ] pyindustry - [ ] pypetroleumhandling - [ ] pypostprocessing - [ ] pyrawores ### Operating system >=Windows 10 ### What kind of issue is this? - [ ] Compatibility - [ ] Locale (names, descriptions, unknown keys) - [ ] Graphical - [ ] Crash - [ ] Progression - [ ] Balance - [ ] Pypostprocessing failure - [X] Other ### What is the problem? The burner-inserter and inserter forcing to blacklist from either on_built_entity or on_configuration_changed is a bit too heavy-handed at times. It would be helpful if it considered a circuit attachment to be the same as having a filter set as far as avoiding the force back to blacklist. ### Steps to reproduce _No response_ ### Additional context _No response_ ### Log file _No response_
1.0
Inserter mode gets stomped excessively - ### Mod source PyAE Beta ### Which mod are you having an issue with? - [ ] pyalienlife - [ ] pyalternativeenergy - [X] pycoalprocessing - [ ] pyfusionenergy - [ ] pyhightech - [ ] pyindustry - [ ] pypetroleumhandling - [ ] pypostprocessing - [ ] pyrawores ### Operating system >=Windows 10 ### What kind of issue is this? - [ ] Compatibility - [ ] Locale (names, descriptions, unknown keys) - [ ] Graphical - [ ] Crash - [ ] Progression - [ ] Balance - [ ] Pypostprocessing failure - [X] Other ### What is the problem? The burner-inserter and inserter forcing to blacklist from either on_built_entity or on_configuration_changed is a bit too heavy-handed at times. It would be helpful if it considered a circuit attachment to be the same as having a filter set as far as avoiding the force back to blacklist. ### Steps to reproduce _No response_ ### Additional context _No response_ ### Log file _No response_
process
inserter mode gets stomped excessively mod source pyae beta which mod are you having an issue with pyalienlife pyalternativeenergy pycoalprocessing pyfusionenergy pyhightech pyindustry pypetroleumhandling pypostprocessing pyrawores operating system windows what kind of issue is this compatibility locale names descriptions unknown keys graphical crash progression balance pypostprocessing failure other what is the problem the burner inserter and inserter forcing to blacklist from either on built entity or on configuration changed is a bit too heavy handed at times it would be helpful if it considered a circuit attachment to be the same as having a filter set as far as avoiding the force back to blacklist steps to reproduce no response additional context no response log file no response
1
5,736
8,580,069,955
IssuesEvent
2018-11-13 10:53:42
threefoldtech/rivine
https://api.github.com/repos/threefoldtech/rivine
closed
Hierarchical Deterministic wallets
process_duplicate type_feature type_story
https://github.com/bitcoin/bips/blob/master/bip-0032.mediawiki Ou deterministic wallet typically consist of a single "chain" of keypairs. The fact that there is only one chain means that sharing a wallet happens on an all-or-nothing basis. However, in some cases one only wants some (public) keys to be shared and recoverable. In the example of a webshop, the webserver does not need access to all public keys of the merchant's wallet; only to those addresses which are used to receive customer's payments, and not for example the change addresses that are generated when the merchant spends money. Hierarchical deterministic wallets allow such selective sharing by supporting multiple keypair chains, derived from a single root. This would also mean that a 3 bot can have it's own chain ,just as a mobile or jumpscale wallet can, all derived from a single root which can be locked away safely.
1.0
Hierarchical Deterministic wallets - https://github.com/bitcoin/bips/blob/master/bip-0032.mediawiki Ou deterministic wallet typically consist of a single "chain" of keypairs. The fact that there is only one chain means that sharing a wallet happens on an all-or-nothing basis. However, in some cases one only wants some (public) keys to be shared and recoverable. In the example of a webshop, the webserver does not need access to all public keys of the merchant's wallet; only to those addresses which are used to receive customer's payments, and not for example the change addresses that are generated when the merchant spends money. Hierarchical deterministic wallets allow such selective sharing by supporting multiple keypair chains, derived from a single root. This would also mean that a 3 bot can have it's own chain ,just as a mobile or jumpscale wallet can, all derived from a single root which can be locked away safely.
process
hierarchical deterministic wallets ou deterministic wallet typically consist of a single chain of keypairs the fact that there is only one chain means that sharing a wallet happens on an all or nothing basis however in some cases one only wants some public keys to be shared and recoverable in the example of a webshop the webserver does not need access to all public keys of the merchant s wallet only to those addresses which are used to receive customer s payments and not for example the change addresses that are generated when the merchant spends money hierarchical deterministic wallets allow such selective sharing by supporting multiple keypair chains derived from a single root this would also mean that a bot can have it s own chain just as a mobile or jumpscale wallet can all derived from a single root which can be locked away safely
1
17,323
23,142,780,670
IssuesEvent
2022-07-28 20:17:46
USGS-R/drb-do-ml
https://api.github.com/repos/USGS-R/drb-do-ml
closed
Check correlation between temperature and DO predictions
process-guidance
We are interested to see the correlation between model DO predictions and input air temperature in the Baseline model.
1.0
Check correlation between temperature and DO predictions - We are interested to see the correlation between model DO predictions and input air temperature in the Baseline model.
process
check correlation between temperature and do predictions we are interested to see the correlation between model do predictions and input air temperature in the baseline model
1
18,803
10,231,609,955
IssuesEvent
2019-08-18 11:05:58
TryGhost/Ghost
https://api.github.com/repos/TryGhost/Ghost
closed
Delete all content triggers a 504 gateway timeout
admin-api api help wanted performance server stale
### Issue Summary Using the Delete all content feature in Labs times out with a collection exceeding 3400 stories. This is not a limit I have explored, but the current situation I am in. ### To Reproduce Import 3400+ stories in to ghost, try to delete all content. I had to batch this import as the 5.5Mb file times out also, but that is probably due to my connection speed. **Any other info e.g. Why do you consider this to be a bug? What did you expect to happen instead?** I expected the request not to timeout. This is a development environment, and I would not expect this operation to be carried out under normal production circumstances. Of course I can nuke the instance and start again with no issues. Just though I would let you know of the issue. ### Technical details: * Ghost-CLI version: 1.9.9 * Node Version: 10.15.0 * Browser/OS: N/A * Database: mysql Ver 14.14 Distrib 5.7.25, for Linux (x86_64) using EditLine wrapper
True
Delete all content triggers a 504 gateway timeout - ### Issue Summary Using the Delete all content feature in Labs times out with a collection exceeding 3400 stories. This is not a limit I have explored, but the current situation I am in. ### To Reproduce Import 3400+ stories in to ghost, try to delete all content. I had to batch this import as the 5.5Mb file times out also, but that is probably due to my connection speed. **Any other info e.g. Why do you consider this to be a bug? What did you expect to happen instead?** I expected the request not to timeout. This is a development environment, and I would not expect this operation to be carried out under normal production circumstances. Of course I can nuke the instance and start again with no issues. Just though I would let you know of the issue. ### Technical details: * Ghost-CLI version: 1.9.9 * Node Version: 10.15.0 * Browser/OS: N/A * Database: mysql Ver 14.14 Distrib 5.7.25, for Linux (x86_64) using EditLine wrapper
non_process
delete all content triggers a gateway timeout issue summary using the delete all content feature in labs times out with a collection exceeding stories this is not a limit i have explored but the current situation i am in to reproduce import stories in to ghost try to delete all content i had to batch this import as the file times out also but that is probably due to my connection speed any other info e g why do you consider this to be a bug what did you expect to happen instead i expected the request not to timeout this is a development environment and i would not expect this operation to be carried out under normal production circumstances of course i can nuke the instance and start again with no issues just though i would let you know of the issue technical details ghost cli version node version browser os n a database mysql ver distrib for linux using editline wrapper
0
218,208
16,976,709,180
IssuesEvent
2021-06-30 00:45:10
microsoft/vscode
https://api.github.com/repos/microsoft/vscode
closed
VS Code Live Server - June Iteration Plans
feature-request live-server on-testplan
<!-- ⚠️⚠️ Do Not Delete This! feature_request_template ⚠️⚠️ --> <!-- Please read our Rules of Conduct: https://opensource.microsoft.com/codeofconduct/ --> <!-- Please search existing issues to avoid creating duplicates. --> <!-- Describe the feature you'd like. --> This tracks the upcoming work on the Live Server, which is described in #124607. Some of these may be deferred to the July iteration. See backlog #125343 Bug Fixes: - [x] External links on browser are broken. #125200 - [x] Relative file links in sub-folders are not working. #125201 - [x] Make hosting port flexible + have error handling for if ports are occupied. #125152 - [x] Two refreshes when file change happens #125174 - [x] Filenames end in slash in index menu #125178 - [x] Handle situation if `index.html` does not exist (can show page with all of the files to pick). - [x] Fix `Upgrade Required` when using localhost instead of 127.0.0.1 - [x] Support Emojis on html files Features: - [x] Basic Implementation of Default Page Index/Filesystem Navigation - [x] Basic Implementation of "Page Not Found" - [x] Allow window reload to show last visible file in preview (currently always shows `index.html`) - [x] Allow users to choose which file to display using active editor. #125775 - [x] Implement more visual way of showing server is on or off. #125854 - [x] Allow user settings to tweak experience #125855 - [x] Explore best UX experience for users. #125949 #126307 - [x] Server run using tasks #126309 - [x] Page title support on embedded preview. - [x] Non-workspace Support #126715 - [x] 💪SVG Preview Support #126093 - [x] Release Docs + Code Cleanup #126308 - [x] Link Preview #127019 - [x] Multi-root workspaces #126622 Admin: - [x] Transfer repo ownership to Microsoft - [x] Look into extension bundling for smaller `*.vsix` size (bundling) - [x] Upload to marketplace.
1.0
VS Code Live Server - June Iteration Plans - <!-- ⚠️⚠️ Do Not Delete This! feature_request_template ⚠️⚠️ --> <!-- Please read our Rules of Conduct: https://opensource.microsoft.com/codeofconduct/ --> <!-- Please search existing issues to avoid creating duplicates. --> <!-- Describe the feature you'd like. --> This tracks the upcoming work on the Live Server, which is described in #124607. Some of these may be deferred to the July iteration. See backlog #125343 Bug Fixes: - [x] External links on browser are broken. #125200 - [x] Relative file links in sub-folders are not working. #125201 - [x] Make hosting port flexible + have error handling for if ports are occupied. #125152 - [x] Two refreshes when file change happens #125174 - [x] Filenames end in slash in index menu #125178 - [x] Handle situation if `index.html` does not exist (can show page with all of the files to pick). - [x] Fix `Upgrade Required` when using localhost instead of 127.0.0.1 - [x] Support Emojis on html files Features: - [x] Basic Implementation of Default Page Index/Filesystem Navigation - [x] Basic Implementation of "Page Not Found" - [x] Allow window reload to show last visible file in preview (currently always shows `index.html`) - [x] Allow users to choose which file to display using active editor. #125775 - [x] Implement more visual way of showing server is on or off. #125854 - [x] Allow user settings to tweak experience #125855 - [x] Explore best UX experience for users. #125949 #126307 - [x] Server run using tasks #126309 - [x] Page title support on embedded preview. - [x] Non-workspace Support #126715 - [x] 💪SVG Preview Support #126093 - [x] Release Docs + Code Cleanup #126308 - [x] Link Preview #127019 - [x] Multi-root workspaces #126622 Admin: - [x] Transfer repo ownership to Microsoft - [x] Look into extension bundling for smaller `*.vsix` size (bundling) - [x] Upload to marketplace.
non_process
vs code live server june iteration plans this tracks the upcoming work on the live server which is described in some of these may be deferred to the july iteration see backlog bug fixes external links on browser are broken relative file links in sub folders are not working make hosting port flexible have error handling for if ports are occupied two refreshes when file change happens filenames end in slash in index menu handle situation if index html does not exist can show page with all of the files to pick fix upgrade required when using localhost instead of support emojis on html files features basic implementation of default page index filesystem navigation basic implementation of page not found allow window reload to show last visible file in preview currently always shows index html allow users to choose which file to display using active editor implement more visual way of showing server is on or off allow user settings to tweak experience explore best ux experience for users server run using tasks page title support on embedded preview non workspace support 💪svg preview support release docs code cleanup link preview multi root workspaces admin transfer repo ownership to microsoft look into extension bundling for smaller vsix size bundling upload to marketplace
0
7,404
10,523,658,825
IssuesEvent
2019-09-30 11:34:10
teleporthq/teleport-code-generators
https://api.github.com/repos/teleporthq/teleport-code-generators
closed
HTML formatting utility
enhancement post-processors
We're relying on `prettier` for formatting html chunks (which are represented as strings). However, perf tests (#64) showed that prettier is running very slow when the UIDL gets bigger (ex: 2000 nodes). We have two ways of approaching this: ### Simple post-processor function that formats html tags This should run after the chunks are linked and should just align the html tags and content accordingly. Important to note that some frameworks might have non-standard html elements to render. ### HAST formatter This should run at link time, when the html is a HAST structure represented internally. some utilities from the unified ecosystem could be checked out for this.
1.0
HTML formatting utility - We're relying on `prettier` for formatting html chunks (which are represented as strings). However, perf tests (#64) showed that prettier is running very slow when the UIDL gets bigger (ex: 2000 nodes). We have two ways of approaching this: ### Simple post-processor function that formats html tags This should run after the chunks are linked and should just align the html tags and content accordingly. Important to note that some frameworks might have non-standard html elements to render. ### HAST formatter This should run at link time, when the html is a HAST structure represented internally. some utilities from the unified ecosystem could be checked out for this.
process
html formatting utility we re relying on prettier for formatting html chunks which are represented as strings however perf tests showed that prettier is running very slow when the uidl gets bigger ex nodes we have two ways of approaching this simple post processor function that formats html tags this should run after the chunks are linked and should just align the html tags and content accordingly important to note that some frameworks might have non standard html elements to render hast formatter this should run at link time when the html is a hast structure represented internally some utilities from the unified ecosystem could be checked out for this
1
26,913
2,688,753,225
IssuesEvent
2015-03-31 03:36:52
ChrisMahlke/nominate-2
https://api.github.com/repos/ChrisMahlke/nominate-2
closed
My Profile tab is not turning blue on Refresh
MediumPriority
After fixing the “My Profile” section to receive full points, the red text and border does not change until I click on another tab. This should change to blue automatically after reaching the desired score. I really appreciate the “Refresh” button, but it would be nice if it happened automatically.
1.0
My Profile tab is not turning blue on Refresh - After fixing the “My Profile” section to receive full points, the red text and border does not change until I click on another tab. This should change to blue automatically after reaching the desired score. I really appreciate the “Refresh” button, but it would be nice if it happened automatically.
non_process
my profile tab is not turning blue on refresh after fixing the “my profile” section to receive full points the red text and border does not change until i click on another tab this should change to blue automatically after reaching the desired score i really appreciate the “refresh” button but it would be nice if it happened automatically
0
10,158
13,044,162,640
IssuesEvent
2020-07-29 03:47:34
tikv/tikv
https://api.github.com/repos/tikv/tikv
closed
UCP: Migrate scalar function `AesEncryptIV` from TiDB
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
## Description Port the scalar function `AesEncryptIV` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @lonng ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
2.0
UCP: Migrate scalar function `AesEncryptIV` from TiDB - ## Description Port the scalar function `AesEncryptIV` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @lonng ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
process
ucp migrate scalar function aesencryptiv from tidb description port the scalar function aesencryptiv from tidb to coprocessor score mentor s lonng recommended skills rust programming learning materials already implemented expressions ported from tidb
1
4,683
7,522,205,207
IssuesEvent
2018-04-12 19:39:28
rubberduck-vba/Rubberduck
https://api.github.com/repos/rubberduck-vba/Rubberduck
closed
Members of implicit default object property are not resolved
bug navigation parse-tree-preprocessing
The default member of Workbooks is `[_Default]`, which is roughly equivalent to `Item`. Both `[_Default]` and `Item` both return a `Workbook` object, so RD should be able to resolve the `Workbook` members. But, when the default member call is implicit, RD fails to resolve the `Workbook` properties ```vb Debug.Print Workbooks.[_Default](ThisWorkbook.Name).ReadOnly 'ReadOnly is recognized Debug.Print Workbooks.Item(ThisWorkbook.Name).ReadOnly 'ReadOnly is recognized Debug.Print Workbooks(ThisWorkbook.Name).ReadOnly 'ReadOnly NOT recognized ```
1.0
Members of implicit default object property are not resolved - The default member of Workbooks is `[_Default]`, which is roughly equivalent to `Item`. Both `[_Default]` and `Item` both return a `Workbook` object, so RD should be able to resolve the `Workbook` members. But, when the default member call is implicit, RD fails to resolve the `Workbook` properties ```vb Debug.Print Workbooks.[_Default](ThisWorkbook.Name).ReadOnly 'ReadOnly is recognized Debug.Print Workbooks.Item(ThisWorkbook.Name).ReadOnly 'ReadOnly is recognized Debug.Print Workbooks(ThisWorkbook.Name).ReadOnly 'ReadOnly NOT recognized ```
process
members of implicit default object property are not resolved the default member of workbooks is which is roughly equivalent to item both and item both return a workbook object so rd should be able to resolve the workbook members but when the default member call is implicit rd fails to resolve the workbook properties vb debug print workbooks thisworkbook name readonly readonly is recognized debug print workbooks item thisworkbook name readonly readonly is recognized debug print workbooks thisworkbook name readonly readonly not recognized
1
516,271
14,978,503,001
IssuesEvent
2021-01-28 10:55:12
nextcloud/mail
https://api.github.com/repos/nextcloud/mail
closed
Can not access mail account in nextcloud after password change on imap server
4. to release bug priority:medium
### Expected behavior There should be an option to set the new password. ### Actual behavior Mail account is not shown anymore in nextcloud. The cronjob gives "OCA\Mail\Exception\ServiceException: IMAP errorMail server denied authentication." Can not access the mail account, so I even can not delete it. ### Mail app **Mail app version:** 1.4.1 **Mailserver or service:** imap.1und1.de ### Server configuration **Operating system**: Ubuntu 18.04 **Web server:** Apache 2.4.46 **Database:** MariaDB 10.4.15 **PHP version:** 7.3.23 **Nextcloud Version:** 19.0.4 #### Client configuration **Browser:** Firefox 82.0
1.0
Can not access mail account in nextcloud after password change on imap server - ### Expected behavior There should be an option to set the new password. ### Actual behavior Mail account is not shown anymore in nextcloud. The cronjob gives "OCA\Mail\Exception\ServiceException: IMAP errorMail server denied authentication." Can not access the mail account, so I even can not delete it. ### Mail app **Mail app version:** 1.4.1 **Mailserver or service:** imap.1und1.de ### Server configuration **Operating system**: Ubuntu 18.04 **Web server:** Apache 2.4.46 **Database:** MariaDB 10.4.15 **PHP version:** 7.3.23 **Nextcloud Version:** 19.0.4 #### Client configuration **Browser:** Firefox 82.0
non_process
can not access mail account in nextcloud after password change on imap server expected behavior there should be an option to set the new password actual behavior mail account is not shown anymore in nextcloud the cronjob gives oca mail exception serviceexception imap errormail server denied authentication can not access the mail account so i even can not delete it mail app mail app version mailserver or service imap de server configuration operating system ubuntu web server apache database mariadb php version nextcloud version client configuration browser firefox
0
8,643
7,349,176,814
IssuesEvent
2018-03-08 09:45:59
vector-im/riot-android
https://api.github.com/repos/vector-im/riot-android
closed
Embedded images must be restricted to mxc: URIs
P1 bug security
Right now img tags are allowed arbitrary linking, which is a security issue!
True
Embedded images must be restricted to mxc: URIs - Right now img tags are allowed arbitrary linking, which is a security issue!
non_process
embedded images must be restricted to mxc uris right now img tags are allowed arbitrary linking which is a security issue
0
9,722
12,717,208,300
IssuesEvent
2020-06-24 04:24:47
kubeflow/pipelines
https://api.github.com/repos/kubeflow/pipelines
closed
Jest Snapshot Test didn't well handle time string
area/frontend kind/process lifecycle/stale needs investigation priority/p2 status/triaged
In our codes we use Date.toLocalDateString which depends on local time format. It may generate different snapshot.
1.0
Jest Snapshot Test didn't well handle time string - In our codes we use Date.toLocalDateString which depends on local time format. It may generate different snapshot.
process
jest snapshot test didn t well handle time string in our codes we use date tolocaldatestring which depends on local time format it may generate different snapshot
1
18,401
4,266,248,177
IssuesEvent
2016-07-12 14:02:31
morepath/morepath
https://api.github.com/repos/morepath/morepath
closed
update the morepath example applications to follow cookiecutter-style setup
documentation entry level help wanted
Now that we have a cookiecutter setup we should update the example applications to follow a similar setup, with install instructions in README.txt. The sample applications right now use buildout and depend on development versions of Morepath, but that's not really useful anymore. We should also retire any example applications that have served their purpose. Spun off from #404
1.0
update the morepath example applications to follow cookiecutter-style setup - Now that we have a cookiecutter setup we should update the example applications to follow a similar setup, with install instructions in README.txt. The sample applications right now use buildout and depend on development versions of Morepath, but that's not really useful anymore. We should also retire any example applications that have served their purpose. Spun off from #404
non_process
update the morepath example applications to follow cookiecutter style setup now that we have a cookiecutter setup we should update the example applications to follow a similar setup with install instructions in readme txt the sample applications right now use buildout and depend on development versions of morepath but that s not really useful anymore we should also retire any example applications that have served their purpose spun off from
0
88
2,534,358,013
IssuesEvent
2015-01-24 21:44:41
rhattersley/docbook2asciidoc
https://api.github.com/repos/rhattersley/docbook2asciidoc
opened
Remove formatting from subtitle
pre-process
```xml <subtitle>Version 1.7.2 <emphasis role="bold">DRAFT</emphasis>, 28 March, 2014</subtitle> ``` converts to: ``` Version 1.7.2 **DRAFT**, 28 March, 2014 ``` which renders the "*" characters literally.
1.0
Remove formatting from subtitle - ```xml <subtitle>Version 1.7.2 <emphasis role="bold">DRAFT</emphasis>, 28 March, 2014</subtitle> ``` converts to: ``` Version 1.7.2 **DRAFT**, 28 March, 2014 ``` which renders the "*" characters literally.
process
remove formatting from subtitle xml version draft march converts to version draft march which renders the characters literally
1
11,889
14,682,445,293
IssuesEvent
2020-12-31 16:27:46
darktable-org/darktable
https://api.github.com/repos/darktable-org/darktable
closed
fix to #6734 breaks pipe with multi-instanciated modules
bug: pending difficulty: average priority: high scope: image processing
**Describe the bug** <!-- A clear and concise description of what the bug is. --> The issue #6734 has been fixed in darktable 3.4 but the fix introduced a new issue when there are multiple instances of the same module in the development pipe. The photos is more or less scrambled and you can't access anymore to the extra instances of the multi-instanced module. N.B. **If there is no multi-instanciated module in the pipe everything seems to be OK** **To Reproduce** Same thing as for #6734 except the result is different. 1. Edit a photo with for instance two instances of the exposure module 1. Go to 'Light Table' 2. Select the photo 3. Apply a style (I tested with a style which adds a watermark) 4. See that the thumbnail is correctly updated 5. Press Ctrl-Z to "undo" 6. See that the thumbnail is incorrectly updated (it is more or less scrambled, sometimes looking a little bit like a raw under/overexposure warning, sometimes looking like heavy noise). 7. Go to 'Dark Room' 8. See that there is a lot of "garbage" as in the thumbnail 9. See that the second instance of the exposure module is no more in the pipe, but your history stack is not affected 10. See that Darktable throws this kind of error message to stderr : ``` cannot get iop-order for exposure instance 1 [dt_ioppr_check_iop_order] history module not used but enabled!! exposure 1(2147483647) image 8747 (dt_dev_read_history_no_image end) ``` **Expected behavior** No garbage Second instance of the exposure module kept in the pipe **Screenshots** <!-- If applicable, add screenshots to help explain your problem. --> ![Screenshot_20201230_172401](https://user-images.githubusercontent.com/31566103/103366902-0271f400-4ac4-11eb-95c5-9e1514c6fbde.png) **Platform (please complete the following information):** * Darktable Version: [3.4.0] * OS: [ Ubuntu 20.04 with KDE 5 ] * OpenCL not activated * Graphics card : Cape Verde XT [Radeon HD 7770/8760 / R7 250X] **Additional context** I experimented it with two instances of exposure module and with two instances of color balance module. The xmp sidecar file seems to be OK but removing the photo from the database and re-importing it with its xmp file gets the same garbage. The only way to partially recover your work is to go back in the history stack before the first duplicate instance of whatever module is duplicated. Or if you have important settings after that to manually edit your xmp file and remove what is related to multiple instances of modules (keep only the first instance). Then re-import it
1.0
fix to #6734 breaks pipe with multi-instanciated modules - **Describe the bug** <!-- A clear and concise description of what the bug is. --> The issue #6734 has been fixed in darktable 3.4 but the fix introduced a new issue when there are multiple instances of the same module in the development pipe. The photos is more or less scrambled and you can't access anymore to the extra instances of the multi-instanced module. N.B. **If there is no multi-instanciated module in the pipe everything seems to be OK** **To Reproduce** Same thing as for #6734 except the result is different. 1. Edit a photo with for instance two instances of the exposure module 1. Go to 'Light Table' 2. Select the photo 3. Apply a style (I tested with a style which adds a watermark) 4. See that the thumbnail is correctly updated 5. Press Ctrl-Z to "undo" 6. See that the thumbnail is incorrectly updated (it is more or less scrambled, sometimes looking a little bit like a raw under/overexposure warning, sometimes looking like heavy noise). 7. Go to 'Dark Room' 8. See that there is a lot of "garbage" as in the thumbnail 9. See that the second instance of the exposure module is no more in the pipe, but your history stack is not affected 10. See that Darktable throws this kind of error message to stderr : ``` cannot get iop-order for exposure instance 1 [dt_ioppr_check_iop_order] history module not used but enabled!! exposure 1(2147483647) image 8747 (dt_dev_read_history_no_image end) ``` **Expected behavior** No garbage Second instance of the exposure module kept in the pipe **Screenshots** <!-- If applicable, add screenshots to help explain your problem. --> ![Screenshot_20201230_172401](https://user-images.githubusercontent.com/31566103/103366902-0271f400-4ac4-11eb-95c5-9e1514c6fbde.png) **Platform (please complete the following information):** * Darktable Version: [3.4.0] * OS: [ Ubuntu 20.04 with KDE 5 ] * OpenCL not activated * Graphics card : Cape Verde XT [Radeon HD 7770/8760 / R7 250X] **Additional context** I experimented it with two instances of exposure module and with two instances of color balance module. The xmp sidecar file seems to be OK but removing the photo from the database and re-importing it with its xmp file gets the same garbage. The only way to partially recover your work is to go back in the history stack before the first duplicate instance of whatever module is duplicated. Or if you have important settings after that to manually edit your xmp file and remove what is related to multiple instances of modules (keep only the first instance). Then re-import it
process
fix to breaks pipe with multi instanciated modules describe the bug the issue has been fixed in darktable but the fix introduced a new issue when there are multiple instances of the same module in the development pipe the photos is more or less scrambled and you can t access anymore to the extra instances of the multi instanced module n b if there is no multi instanciated module in the pipe everything seems to be ok to reproduce same thing as for except the result is different edit a photo with for instance two instances of the exposure module go to light table select the photo apply a style i tested with a style which adds a watermark see that the thumbnail is correctly updated press ctrl z to undo see that the thumbnail is incorrectly updated it is more or less scrambled sometimes looking a little bit like a raw under overexposure warning sometimes looking like heavy noise go to dark room see that there is a lot of garbage as in the thumbnail see that the second instance of the exposure module is no more in the pipe but your history stack is not affected see that darktable throws this kind of error message to stderr cannot get iop order for exposure instance history module not used but enabled exposure image dt dev read history no image end expected behavior no garbage second instance of the exposure module kept in the pipe screenshots platform please complete the following information darktable version os opencl not activated graphics card cape verde xt additional context i experimented it with two instances of exposure module and with two instances of color balance module the xmp sidecar file seems to be ok but removing the photo from the database and re importing it with its xmp file gets the same garbage the only way to partially recover your work is to go back in the history stack before the first duplicate instance of whatever module is duplicated or if you have important settings after that to manually edit your xmp file and remove what is related to multiple instances of modules keep only the first instance then re import it
1
53,648
13,262,047,358
IssuesEvent
2020-08-20 21:00:31
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
closed
[ppc] clang 3.8 errors (Trac #1809)
Migrated from Trac combo simulation defect
These errors were found compiling with clang 3.8: ```text In file included from /scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/ppc.cxx:696: /scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:22:7: error: no member named 'w' in 'cl_float4' p.n.w=type>0?-int(type):-128; ~~~ ^ /scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:75:7: error: no member named 'w' in 'cl_float4' p.n.w=0, p.f=0; ~~~ ^ /scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:171:7: error: no member named 'w' in 'cl_float4' p.n.w=dr; ~~~ ^ /scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:343:7: error: no member named 'w' in 'cl_float4' p.r.w=t; p.r.x=rx; p.r.y=ry; p.r.z=rz; ~~~ ^ /scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:343:16: error: no member named 'x' in 'cl_float4' p.r.w=t; p.r.x=rx; p.r.y=ry; p.r.z=rz; ~~~ ^ /scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:343:26: error: no member named 'y' in 'cl_float4' p.r.w=t; p.r.x=rx; p.r.y=ry; p.r.z=rz; ~~~ ^ /scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:343:36: error: no member named 'z' in 'cl_float4' p.r.w=t; p.r.x=rx; p.r.y=ry; p.r.z=rz; ~~~ ^ /scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:379:17: error: no member named 'x' in 'cl_float4' p.q=flne; p.n.x=nx; p.n.y=ny; p.n.z=nz; ~~~ ^ /scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:379:27: error: no member named 'y' in 'cl_float4' p.q=flne; p.n.x=nx; p.n.y=ny; p.n.z=nz; ~~~ ^ /scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:379:37: error: no member named 'z' in 'cl_float4' p.q=flne; p.n.x=nx; p.n.y=ny; p.n.z=nz; ~~~ ^ 10 errors generated. ``` This post may be related: http://stackoverflow.com/questions/10979487/opencl-cl-datatypes-arithmetic Note that clang defaults to c++14, which probably disables the macro `__GNUC__`. <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1809">https://code.icecube.wisc.edu/projects/icecube/ticket/1809</a>, reported by david.schultzand owned by dima</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:12:38", "_ts": "1550067158057333", "description": "These errors were found compiling with clang 3.8:\n\n{{{\nIn file included from /scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/ppc.cxx:696:\n/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:22:7: error: no member named 'w' in 'cl_float4'\np.n.w=type>0?-int(type):-128;\n ~~~ ^\n/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:75:7: error: no member named 'w' in 'cl_float4'\np.n.w=0, p.f=0;\n ~~~ ^\n/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:171:7: error: no member named 'w' in 'cl_float4'\np.n.w=dr;\n ~~~ ^\n/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:343:7: error: no member named 'w' in 'cl_float4'\np.r.w=t; p.r.x=rx; p.r.y=ry; p.r.z=rz;\n ~~~ ^\n/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:343:16: error: no member named 'x' in 'cl_float4'\np.r.w=t; p.r.x=rx; p.r.y=ry; p.r.z=rz;\n ~~~ ^\n/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:343:26: error: no member named 'y' in 'cl_float4'\np.r.w=t; p.r.x=rx; p.r.y=ry; p.r.z=rz;\n ~~~ ^\n/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:343:36: error: no member named 'z' in 'cl_float4'\np.r.w=t; p.r.x=rx; p.r.y=ry; p.r.z=rz;\n ~~~ ^\n/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:379:17: error: no member named 'x' in 'cl_float4'\np.q=flne; p.n.x=nx; p.n.y=ny; p.n.z=nz;\n ~~~ ^\n/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:379:27: error: no member named 'y' in 'cl_float4'\np.q=flne; p.n.x=nx; p.n.y=ny; p.n.z=nz;\n ~~~ ^\n/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:379:37: error: no member named 'z' in 'cl_float4'\np.q=flne; p.n.x=nx; p.n.y=ny; p.n.z=nz;\n ~~~ ^\n10 errors generated.\n}}}\n\nThis post may be related:\nhttp://stackoverflow.com/questions/10979487/opencl-cl-datatypes-arithmetic\n\nNote that clang defaults to c++14, which probably disables the macro `__GNUC__`.", "reporter": "david.schultz", "cc": "olivas", "resolution": "fixed", "time": "2016-07-29T20:44:11", "component": "combo simulation", "summary": "[ppc] clang 3.8 errors", "priority": "major", "keywords": "", "milestone": "", "owner": "dima", "type": "defect" } ``` </p> </details>
1.0
[ppc] clang 3.8 errors (Trac #1809) - These errors were found compiling with clang 3.8: ```text In file included from /scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/ppc.cxx:696: /scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:22:7: error: no member named 'w' in 'cl_float4' p.n.w=type>0?-int(type):-128; ~~~ ^ /scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:75:7: error: no member named 'w' in 'cl_float4' p.n.w=0, p.f=0; ~~~ ^ /scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:171:7: error: no member named 'w' in 'cl_float4' p.n.w=dr; ~~~ ^ /scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:343:7: error: no member named 'w' in 'cl_float4' p.r.w=t; p.r.x=rx; p.r.y=ry; p.r.z=rz; ~~~ ^ /scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:343:16: error: no member named 'x' in 'cl_float4' p.r.w=t; p.r.x=rx; p.r.y=ry; p.r.z=rz; ~~~ ^ /scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:343:26: error: no member named 'y' in 'cl_float4' p.r.w=t; p.r.x=rx; p.r.y=ry; p.r.z=rz; ~~~ ^ /scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:343:36: error: no member named 'z' in 'cl_float4' p.r.w=t; p.r.x=rx; p.r.y=ry; p.r.z=rz; ~~~ ^ /scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:379:17: error: no member named 'x' in 'cl_float4' p.q=flne; p.n.x=nx; p.n.y=ny; p.n.z=nz; ~~~ ^ /scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:379:27: error: no member named 'y' in 'cl_float4' p.q=flne; p.n.x=nx; p.n.y=ny; p.n.z=nz; ~~~ ^ /scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:379:37: error: no member named 'z' in 'cl_float4' p.q=flne; p.n.x=nx; p.n.y=ny; p.n.z=nz; ~~~ ^ 10 errors generated. ``` This post may be related: http://stackoverflow.com/questions/10979487/opencl-cl-datatypes-arithmetic Note that clang defaults to c++14, which probably disables the macro `__GNUC__`. <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1809">https://code.icecube.wisc.edu/projects/icecube/ticket/1809</a>, reported by david.schultzand owned by dima</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:12:38", "_ts": "1550067158057333", "description": "These errors were found compiling with clang 3.8:\n\n{{{\nIn file included from /scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/ppc.cxx:696:\n/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:22:7: error: no member named 'w' in 'cl_float4'\np.n.w=type>0?-int(type):-128;\n ~~~ ^\n/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:75:7: error: no member named 'w' in 'cl_float4'\np.n.w=0, p.f=0;\n ~~~ ^\n/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:171:7: error: no member named 'w' in 'cl_float4'\np.n.w=dr;\n ~~~ ^\n/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:343:7: error: no member named 'w' in 'cl_float4'\np.r.w=t; p.r.x=rx; p.r.y=ry; p.r.z=rz;\n ~~~ ^\n/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:343:16: error: no member named 'x' in 'cl_float4'\np.r.w=t; p.r.x=rx; p.r.y=ry; p.r.z=rz;\n ~~~ ^\n/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:343:26: error: no member named 'y' in 'cl_float4'\np.r.w=t; p.r.x=rx; p.r.y=ry; p.r.z=rz;\n ~~~ ^\n/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:343:36: error: no member named 'z' in 'cl_float4'\np.r.w=t; p.r.x=rx; p.r.y=ry; p.r.z=rz;\n ~~~ ^\n/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:379:17: error: no member named 'x' in 'cl_float4'\np.q=flne; p.n.x=nx; p.n.y=ny; p.n.z=nz;\n ~~~ ^\n/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:379:27: error: no member named 'y' in 'cl_float4'\np.q=flne; p.n.x=nx; p.n.y=ny; p.n.z=nz;\n ~~~ ^\n/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:379:37: error: no member named 'z' in 'cl_float4'\np.q=flne; p.n.x=nx; p.n.y=ny; p.n.z=nz;\n ~~~ ^\n10 errors generated.\n}}}\n\nThis post may be related:\nhttp://stackoverflow.com/questions/10979487/opencl-cl-datatypes-arithmetic\n\nNote that clang defaults to c++14, which probably disables the macro `__GNUC__`.", "reporter": "david.schultz", "cc": "olivas", "resolution": "fixed", "time": "2016-07-29T20:44:11", "component": "combo simulation", "summary": "[ppc] clang 3.8 errors", "priority": "major", "keywords": "", "milestone": "", "owner": "dima", "type": "defect" } ``` </p> </details>
non_process
clang errors trac these errors were found compiling with clang text in file included from scratch dschultz icetray profiling src ppc private ppc ocl ppc cxx scratch dschultz icetray profiling src ppc private ppc ocl cxx error no member named w in cl p n w type int type scratch dschultz icetray profiling src ppc private ppc ocl cxx error no member named w in cl p n w p f scratch dschultz icetray profiling src ppc private ppc ocl cxx error no member named w in cl p n w dr scratch dschultz icetray profiling src ppc private ppc ocl cxx error no member named w in cl p r w t p r x rx p r y ry p r z rz scratch dschultz icetray profiling src ppc private ppc ocl cxx error no member named x in cl p r w t p r x rx p r y ry p r z rz scratch dschultz icetray profiling src ppc private ppc ocl cxx error no member named y in cl p r w t p r x rx p r y ry p r z rz scratch dschultz icetray profiling src ppc private ppc ocl cxx error no member named z in cl p r w t p r x rx p r y ry p r z rz scratch dschultz icetray profiling src ppc private ppc ocl cxx error no member named x in cl p q flne p n x nx p n y ny p n z nz scratch dschultz icetray profiling src ppc private ppc ocl cxx error no member named y in cl p q flne p n x nx p n y ny p n z nz scratch dschultz icetray profiling src ppc private ppc ocl cxx error no member named z in cl p q flne p n x nx p n y ny p n z nz errors generated this post may be related note that clang defaults to c which probably disables the macro gnuc migrated from json status closed changetime ts description these errors were found compiling with clang n n nin file included from scratch dschultz icetray profiling src ppc private ppc ocl ppc cxx n scratch dschultz icetray profiling src ppc private ppc ocl cxx error no member named w in cl np n w type int type n n scratch dschultz icetray profiling src ppc private ppc ocl cxx error no member named w in cl np n w p f n n scratch dschultz icetray profiling src ppc private ppc ocl cxx error no member named w in cl np n w dr n n scratch dschultz icetray profiling src ppc private ppc ocl cxx error no member named w in cl np r w t p r x rx p r y ry p r z rz n n scratch dschultz icetray profiling src ppc private ppc ocl cxx error no member named x in cl np r w t p r x rx p r y ry p r z rz n n scratch dschultz icetray profiling src ppc private ppc ocl cxx error no member named y in cl np r w t p r x rx p r y ry p r z rz n n scratch dschultz icetray profiling src ppc private ppc ocl cxx error no member named z in cl np r w t p r x rx p r y ry p r z rz n n scratch dschultz icetray profiling src ppc private ppc ocl cxx error no member named x in cl np q flne p n x nx p n y ny p n z nz n n scratch dschultz icetray profiling src ppc private ppc ocl cxx error no member named y in cl np q flne p n x nx p n y ny p n z nz n n scratch dschultz icetray profiling src ppc private ppc ocl cxx error no member named z in cl np q flne p n x nx p n y ny p n z nz n errors generated n n nthis post may be related n that clang defaults to c which probably disables the macro gnuc reporter david schultz cc olivas resolution fixed time component combo simulation summary clang errors priority major keywords milestone owner dima type defect
0
91,939
18,755,813,031
IssuesEvent
2021-11-05 10:35:12
julia-vscode/julia-vscode
https://api.github.com/repos/julia-vscode/julia-vscode
closed
entering new line clears the in-line evaluation popup cell
bug area-code-execution
MWE. type ``` rand() ``` press ALT+ENTER. A popup appears. Then press ENTER to start a new line and write whatever other command. The popup to the right of `rand` has dissapeared. This isn't nice for interactively writing scripts.
1.0
entering new line clears the in-line evaluation popup cell - MWE. type ``` rand() ``` press ALT+ENTER. A popup appears. Then press ENTER to start a new line and write whatever other command. The popup to the right of `rand` has dissapeared. This isn't nice for interactively writing scripts.
non_process
entering new line clears the in line evaluation popup cell mwe type rand press alt enter a popup appears then press enter to start a new line and write whatever other command the popup to the right of rand has dissapeared this isn t nice for interactively writing scripts
0
228,362
18,172,119,282
IssuesEvent
2021-09-27 21:19:03
microsoft/vscode
https://api.github.com/repos/microsoft/vscode
opened
Test: Outline button in Jupyter notebooks
testplan-item
Refs: https://github.com/microsoft/vscode-jupyter/issues/7305 - [ ] anyOS - [ ] anyOS Complexity: 1 Author: @IanMatthewHuff --- File bugs on the Jupyter repo here: https://github.com/microsoft/vscode-jupyter/issues Jupyter notebook users have been continually asking for Table of Contents control in their notebooks, and this functionality is currently provided by the VS Code Outline window. However, Jupyter users are not finding this control, and we are still getting issues and upvotes from users about a missing Table of Contents control. To alleviate this, a button has been added to the Jupyter notebook toolbar to focus the outline window to point users to this. Testing: 1. Jupyter extension installed 2. Open up a .ipynb file 3. For viewing a useful outline it works best with some markdown cells with # header blocks in them (as they go into the outline view) 4. On the toolbar there should be an Outline button next to the variables button ![image](https://user-images.githubusercontent.com/812783/134986622-f5aced2e-e1e1-4144-9272-b870bac606a6.png) - [ ] If the VS Code Outline view is collapsed or offscreen the button should be enabled. Pressing the button should focus the outline view. - [ ] If the VS Code Outline view is onscreen and expanded the button should be disabled. - [ ] "Jupyter: Show Table Of Contents (Outline View)" command should be available for the same function also when the view is offscreen or unavailable. The naming here is specifically for notebook customers searching for table of contents in the command list. ![image](https://user-images.githubusercontent.com/812783/134986714-20360a9f-2d6a-4654-b090-7ed25db31dac.png)
1.0
Test: Outline button in Jupyter notebooks - Refs: https://github.com/microsoft/vscode-jupyter/issues/7305 - [ ] anyOS - [ ] anyOS Complexity: 1 Author: @IanMatthewHuff --- File bugs on the Jupyter repo here: https://github.com/microsoft/vscode-jupyter/issues Jupyter notebook users have been continually asking for Table of Contents control in their notebooks, and this functionality is currently provided by the VS Code Outline window. However, Jupyter users are not finding this control, and we are still getting issues and upvotes from users about a missing Table of Contents control. To alleviate this, a button has been added to the Jupyter notebook toolbar to focus the outline window to point users to this. Testing: 1. Jupyter extension installed 2. Open up a .ipynb file 3. For viewing a useful outline it works best with some markdown cells with # header blocks in them (as they go into the outline view) 4. On the toolbar there should be an Outline button next to the variables button ![image](https://user-images.githubusercontent.com/812783/134986622-f5aced2e-e1e1-4144-9272-b870bac606a6.png) - [ ] If the VS Code Outline view is collapsed or offscreen the button should be enabled. Pressing the button should focus the outline view. - [ ] If the VS Code Outline view is onscreen and expanded the button should be disabled. - [ ] "Jupyter: Show Table Of Contents (Outline View)" command should be available for the same function also when the view is offscreen or unavailable. The naming here is specifically for notebook customers searching for table of contents in the command list. ![image](https://user-images.githubusercontent.com/812783/134986714-20360a9f-2d6a-4654-b090-7ed25db31dac.png)
non_process
test outline button in jupyter notebooks refs anyos anyos complexity author ianmatthewhuff file bugs on the jupyter repo here jupyter notebook users have been continually asking for table of contents control in their notebooks and this functionality is currently provided by the vs code outline window however jupyter users are not finding this control and we are still getting issues and upvotes from users about a missing table of contents control to alleviate this a button has been added to the jupyter notebook toolbar to focus the outline window to point users to this testing jupyter extension installed open up a ipynb file for viewing a useful outline it works best with some markdown cells with header blocks in them as they go into the outline view on the toolbar there should be an outline button next to the variables button if the vs code outline view is collapsed or offscreen the button should be enabled pressing the button should focus the outline view if the vs code outline view is onscreen and expanded the button should be disabled jupyter show table of contents outline view command should be available for the same function also when the view is offscreen or unavailable the naming here is specifically for notebook customers searching for table of contents in the command list
0
12,807
15,184,213,153
IssuesEvent
2021-02-15 09:15:53
topcoder-platform/community-app
https://api.github.com/repos/topcoder-platform/community-app
opened
Member skills are not getting updated when a member wins a challenge
BE P1 ShapeupProcess challenge- recommender-tool member-skill-extractor
The member skill extractor is not updating the members skills when a member is placed in a challenge. Example: member: tester1234 challenge: https://www.topcoder-dev.com/challenges/cfc3f821-64e4-4585-8fdc-744928d2bc9f the member_skills_history table is not updated with the skills for the this member cc @lakshmiathreya
1.0
Member skills are not getting updated when a member wins a challenge - The member skill extractor is not updating the members skills when a member is placed in a challenge. Example: member: tester1234 challenge: https://www.topcoder-dev.com/challenges/cfc3f821-64e4-4585-8fdc-744928d2bc9f the member_skills_history table is not updated with the skills for the this member cc @lakshmiathreya
process
member skills are not getting updated when a member wins a challenge the member skill extractor is not updating the members skills when a member is placed in a challenge example member challenge the member skills history table is not updated with the skills for the this member cc lakshmiathreya
1
14,127
17,020,220,057
IssuesEvent
2021-07-02 17:43:18
darktable-org/darktable
https://api.github.com/repos/darktable-org/darktable
closed
Perspective module: cannot set the automatic cropping value untile some perspective correction is set
bug: pending scope: image processing
If I try to set automatic cropping to some value which is not off before changing the rotation, lens shift, etc. values, I get an error message on the image "automatic cropping failed".
1.0
Perspective module: cannot set the automatic cropping value untile some perspective correction is set - If I try to set automatic cropping to some value which is not off before changing the rotation, lens shift, etc. values, I get an error message on the image "automatic cropping failed".
process
perspective module cannot set the automatic cropping value untile some perspective correction is set if i try to set automatic cropping to some value which is not off before changing the rotation lens shift etc values i get an error message on the image automatic cropping failed
1
1,754
4,460,688,074
IssuesEvent
2016-08-24 00:46:52
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
closed
StandardOutput and StandardError receive an extra line
System.Diagnostics.Process
On Linux, compared to other languages, say Python, .NET is appending (somehow) an extra line on redirected StandardOutput and StandardError. For example, if I execute the following script: ```bash #!/bin/sh exit 42 ``` In Python (e.g., subprocess.check_output) I receive, as I would expect, an empty string. In .NET, however, if I redirect through ProcessStartInfo, I receive an event in my OutputDataReceived handler with an empty string (and the my code cannot reject empty string out-of-hand since it may be a legal part of output).
1.0
StandardOutput and StandardError receive an extra line - On Linux, compared to other languages, say Python, .NET is appending (somehow) an extra line on redirected StandardOutput and StandardError. For example, if I execute the following script: ```bash #!/bin/sh exit 42 ``` In Python (e.g., subprocess.check_output) I receive, as I would expect, an empty string. In .NET, however, if I redirect through ProcessStartInfo, I receive an event in my OutputDataReceived handler with an empty string (and the my code cannot reject empty string out-of-hand since it may be a legal part of output).
process
standardoutput and standarderror receive an extra line on linux compared to other languages say python net is appending somehow an extra line on redirected standardoutput and standarderror for example if i execute the following script bash bin sh exit in python e g subprocess check output i receive as i would expect an empty string in net however if i redirect through processstartinfo i receive an event in my outputdatareceived handler with an empty string and the my code cannot reject empty string out of hand since it may be a legal part of output
1
75,349
9,850,986,273
IssuesEvent
2019-06-19 09:24:59
mlr-org/mlr
https://api.github.com/repos/mlr-org/mlr
opened
`mRMRe::mrmr` differs from `praznik::mrmr`
type-documentation
I don't think we can do anything about this but it is important to know. Note that the absolute values are not of interest here but only the ranking of the features. ``` r suppressPackageStartupMessages(library(mlr)) library(magrittr) bh.task = dropFeatures(bh.task, "chas") fv_mrmr = generateFilterValuesData(bh.task, "mrmr") fv_mrmr_praznik = generateFilterValuesData(bh.task, "praznik_MRMR") purrr::map(list(fv_mrmr$data, fv_mrmr_praznik$data), ~ dplyr::arrange(.x, method, desc(value))) %>% print() #> [[1]] #> # A tibble: 12 x 4 #> name type method value #> <chr> <chr> <chr> <dbl> #> 1 lstat numeric mrmr 0.393 #> 2 rm numeric mrmr 0.0940 #> 3 ptratio numeric mrmr 0.0776 #> 4 b numeric mrmr 0.0269 #> 5 indus numeric mrmr 0.0189 #> 6 crim numeric mrmr 0.0106 #> 7 zn numeric mrmr -0.00239 #> 8 tax numeric mrmr -0.0295 #> 9 age numeric mrmr -0.0495 #> 10 nox numeric mrmr -0.0911 #> 11 rad numeric mrmr -0.135 #> 12 dis numeric mrmr -0.160 #> #> [[2]] #> # A tibble: 12 x 4 #> name type method value #> <chr> <chr> <chr> <dbl> #> 1 lstat numeric praznik_MRMR 1 #> 2 ptratio numeric praznik_MRMR 0.917 #> 3 rm numeric praznik_MRMR 0.833 #> 4 crim numeric praznik_MRMR 0.75 #> 5 age numeric praznik_MRMR 0.667 #> 6 b numeric praznik_MRMR 0.583 #> 7 nox numeric praznik_MRMR 0.5 #> 8 zn numeric praznik_MRMR 0.417 #> 9 tax numeric praznik_MRMR 0.333 #> 10 rad numeric praznik_MRMR 0.25 #> 11 dis numeric praznik_MRMR 0.167 #> 12 indus numeric praznik_MRMR 0.0833 ``` <sup>Created on 2019-06-19 by the [reprex package](https://reprex.tidyverse.org) (v0.3.0)</sup> In addition, here is a runtime comparison for a dataset with ~ 7k features: ``` build_times(fv_nri_praznik_mrmr, fv_nri_mrmr) # A tibble: 2 x 4 target elapsed user system <chr> <S4: Duration> <S4: Duration> <S4: Duration> 1 fv_nri_mrmr 6487s (~1.8 hours) 6495s (~1.8 hours) 0.111s 2 fv_nri_praznik_mrmr 144s (~2.4 minutes) 290s (~4.83 minutes) 0.144s ```
1.0
`mRMRe::mrmr` differs from `praznik::mrmr` - I don't think we can do anything about this but it is important to know. Note that the absolute values are not of interest here but only the ranking of the features. ``` r suppressPackageStartupMessages(library(mlr)) library(magrittr) bh.task = dropFeatures(bh.task, "chas") fv_mrmr = generateFilterValuesData(bh.task, "mrmr") fv_mrmr_praznik = generateFilterValuesData(bh.task, "praznik_MRMR") purrr::map(list(fv_mrmr$data, fv_mrmr_praznik$data), ~ dplyr::arrange(.x, method, desc(value))) %>% print() #> [[1]] #> # A tibble: 12 x 4 #> name type method value #> <chr> <chr> <chr> <dbl> #> 1 lstat numeric mrmr 0.393 #> 2 rm numeric mrmr 0.0940 #> 3 ptratio numeric mrmr 0.0776 #> 4 b numeric mrmr 0.0269 #> 5 indus numeric mrmr 0.0189 #> 6 crim numeric mrmr 0.0106 #> 7 zn numeric mrmr -0.00239 #> 8 tax numeric mrmr -0.0295 #> 9 age numeric mrmr -0.0495 #> 10 nox numeric mrmr -0.0911 #> 11 rad numeric mrmr -0.135 #> 12 dis numeric mrmr -0.160 #> #> [[2]] #> # A tibble: 12 x 4 #> name type method value #> <chr> <chr> <chr> <dbl> #> 1 lstat numeric praznik_MRMR 1 #> 2 ptratio numeric praznik_MRMR 0.917 #> 3 rm numeric praznik_MRMR 0.833 #> 4 crim numeric praznik_MRMR 0.75 #> 5 age numeric praznik_MRMR 0.667 #> 6 b numeric praznik_MRMR 0.583 #> 7 nox numeric praznik_MRMR 0.5 #> 8 zn numeric praznik_MRMR 0.417 #> 9 tax numeric praznik_MRMR 0.333 #> 10 rad numeric praznik_MRMR 0.25 #> 11 dis numeric praznik_MRMR 0.167 #> 12 indus numeric praznik_MRMR 0.0833 ``` <sup>Created on 2019-06-19 by the [reprex package](https://reprex.tidyverse.org) (v0.3.0)</sup> In addition, here is a runtime comparison for a dataset with ~ 7k features: ``` build_times(fv_nri_praznik_mrmr, fv_nri_mrmr) # A tibble: 2 x 4 target elapsed user system <chr> <S4: Duration> <S4: Duration> <S4: Duration> 1 fv_nri_mrmr 6487s (~1.8 hours) 6495s (~1.8 hours) 0.111s 2 fv_nri_praznik_mrmr 144s (~2.4 minutes) 290s (~4.83 minutes) 0.144s ```
non_process
mrmre mrmr differs from praznik mrmr i don t think we can do anything about this but it is important to know note that the absolute values are not of interest here but only the ranking of the features r suppresspackagestartupmessages library mlr library magrittr bh task dropfeatures bh task chas fv mrmr generatefiltervaluesdata bh task mrmr fv mrmr praznik generatefiltervaluesdata bh task praznik mrmr purrr map list fv mrmr data fv mrmr praznik data dplyr arrange x method desc value print a tibble x name type method value lstat numeric mrmr rm numeric mrmr ptratio numeric mrmr b numeric mrmr indus numeric mrmr crim numeric mrmr zn numeric mrmr tax numeric mrmr age numeric mrmr nox numeric mrmr rad numeric mrmr dis numeric mrmr a tibble x name type method value lstat numeric praznik mrmr ptratio numeric praznik mrmr rm numeric praznik mrmr crim numeric praznik mrmr age numeric praznik mrmr b numeric praznik mrmr nox numeric praznik mrmr zn numeric praznik mrmr tax numeric praznik mrmr rad numeric praznik mrmr dis numeric praznik mrmr indus numeric praznik mrmr created on by the in addition here is a runtime comparison for a dataset with features build times fv nri praznik mrmr fv nri mrmr a tibble x target elapsed user system fv nri mrmr hours hours fv nri praznik mrmr minutes minutes
0
273,983
20,821,837,286
IssuesEvent
2022-03-18 16:07:35
sjefferson99/Boatman-pico-uart-hub
https://api.github.com/repos/sjefferson99/Boatman-pico-uart-hub
closed
Improve documentation in bmserial module
documentation
The bmserial module should document that this is for the serial comms including a definition of the data structure. Functions should be boatman module agnostic and parse data to be passed on to python modules for the appropriate boatman module.
1.0
Improve documentation in bmserial module - The bmserial module should document that this is for the serial comms including a definition of the data structure. Functions should be boatman module agnostic and parse data to be passed on to python modules for the appropriate boatman module.
non_process
improve documentation in bmserial module the bmserial module should document that this is for the serial comms including a definition of the data structure functions should be boatman module agnostic and parse data to be passed on to python modules for the appropriate boatman module
0
34,681
7,853,763,386
IssuesEvent
2018-06-20 18:29:10
kobotoolbox/kpi
https://api.github.com/repos/kobotoolbox/kpi
closed
Opening tags moves project details up
bug coded low priority ui
Minor one, but irritating - after opening tags editor, the project details (name, date, etc.) move up by few pixels; plus the amount of whitespace between name and tags editor is too big: <img width="1005" alt="screen shot 2018-06-09 at 15 40 47" src="https://user-images.githubusercontent.com/2521888/41192318-24bf0e02-6bfc-11e8-9224-c68df359ae9e.png"> <img width="998" alt="screen shot 2018-06-09 at 15 41 01" src="https://user-images.githubusercontent.com/2521888/41192319-24e1c55a-6bfc-11e8-8895-129c1cfa9a88.png"> PS. While doing this, verify if this function could be deleted - from `tagInput.es6`: ```es6 handleChange(tags) { var transformed = tags.map(function(tag) { // Behavior should match KpiTaggableManager.add() return tag.trim().replace(/ /g, '-'); }); this.setState({tags: transformed}); var uid = this.props.uid; actions.resources.updateAsset(uid, { tag_string: transformed.join(',') }); } ```
1.0
Opening tags moves project details up - Minor one, but irritating - after opening tags editor, the project details (name, date, etc.) move up by few pixels; plus the amount of whitespace between name and tags editor is too big: <img width="1005" alt="screen shot 2018-06-09 at 15 40 47" src="https://user-images.githubusercontent.com/2521888/41192318-24bf0e02-6bfc-11e8-9224-c68df359ae9e.png"> <img width="998" alt="screen shot 2018-06-09 at 15 41 01" src="https://user-images.githubusercontent.com/2521888/41192319-24e1c55a-6bfc-11e8-8895-129c1cfa9a88.png"> PS. While doing this, verify if this function could be deleted - from `tagInput.es6`: ```es6 handleChange(tags) { var transformed = tags.map(function(tag) { // Behavior should match KpiTaggableManager.add() return tag.trim().replace(/ /g, '-'); }); this.setState({tags: transformed}); var uid = this.props.uid; actions.resources.updateAsset(uid, { tag_string: transformed.join(',') }); } ```
non_process
opening tags moves project details up minor one but irritating after opening tags editor the project details name date etc move up by few pixels plus the amount of whitespace between name and tags editor is too big img width alt screen shot at src img width alt screen shot at src ps while doing this verify if this function could be deleted from taginput handlechange tags var transformed tags map function tag behavior should match kpitaggablemanager add return tag trim replace g this setstate tags transformed var uid this props uid actions resources updateasset uid tag string transformed join
0
5,835
8,666,148,624
IssuesEvent
2018-11-29 02:36:36
w3c/w3process
https://api.github.com/repos/w3c/w3process
opened
TAG appointment should be via IETF style nomcom
Process2020Candidate
The TAG is currently elected. We could move to an IETF-style NomCom appointment. This ensures the right people for the role are selected and a balanced TAG can be achieved. Comments are welcome. - The NomCom consists of a random set of volunteers who meet a set criteria - e.g. have chaired a group or published a draft in the last X years. The AB could be members of the NomCom. The NomCom should have advising members. They cannot vote, only advise. The TAG Chair, a Developer Invited Expert would be good choices. - Nominations for groups go direct to the NomCom. Nominations can come from any source. The NomCom will see via the application what the source of the nomination is - the NomCom can choose to weigh an AC Nomination in a positive manner if they wish. This means that AC nominations count for something, but other groups (e.g. independent developers) are not blocked from applying. - NomCom sees nominations as they come in. If they feel a diverse pool isn’t being reached (diversity for all items including gender, technical knowledge of particular topics, organisation representation size, global location, etc.) the NomCom will begin pushing for more candidates. - NomCom starts the process of selecting candidates. This is as follows: - The NomCom meets before evaluating the candidates. It agrees on key qualities it is looking for in the candidates for the open roles. It is likely these qualities will be the same in each election, although sometimes they will be different (e.g. “Chair of the TAG" will require different qualities to “TAG Member”). Consensus on these qualities is reached. - All applications are read. Obvious “no” candidates are discarded. - The NomCom meets to discuss the “possible” candidates remaining. Further candidates are discarded for reasons the NomCom deem appropriate. These could be because there has been an influx of candidates which represent one industry, or many candidates are experts in only one field. - The NomCom is left with the final set of “possible candidates”. These candidates are interviewed for the role at TPAC, at the AC Meeting or via teleconference. Video Conference interviews are acceptable. - The NomCom announces the selected candidates. An objection process opens to allow formal objections from the AC. Formal objections will be handled as normal. After 28 days, the formal objection window closes and the newly selected individuals begin their term.
1.0
TAG appointment should be via IETF style nomcom - The TAG is currently elected. We could move to an IETF-style NomCom appointment. This ensures the right people for the role are selected and a balanced TAG can be achieved. Comments are welcome. - The NomCom consists of a random set of volunteers who meet a set criteria - e.g. have chaired a group or published a draft in the last X years. The AB could be members of the NomCom. The NomCom should have advising members. They cannot vote, only advise. The TAG Chair, a Developer Invited Expert would be good choices. - Nominations for groups go direct to the NomCom. Nominations can come from any source. The NomCom will see via the application what the source of the nomination is - the NomCom can choose to weigh an AC Nomination in a positive manner if they wish. This means that AC nominations count for something, but other groups (e.g. independent developers) are not blocked from applying. - NomCom sees nominations as they come in. If they feel a diverse pool isn’t being reached (diversity for all items including gender, technical knowledge of particular topics, organisation representation size, global location, etc.) the NomCom will begin pushing for more candidates. - NomCom starts the process of selecting candidates. This is as follows: - The NomCom meets before evaluating the candidates. It agrees on key qualities it is looking for in the candidates for the open roles. It is likely these qualities will be the same in each election, although sometimes they will be different (e.g. “Chair of the TAG" will require different qualities to “TAG Member”). Consensus on these qualities is reached. - All applications are read. Obvious “no” candidates are discarded. - The NomCom meets to discuss the “possible” candidates remaining. Further candidates are discarded for reasons the NomCom deem appropriate. These could be because there has been an influx of candidates which represent one industry, or many candidates are experts in only one field. - The NomCom is left with the final set of “possible candidates”. These candidates are interviewed for the role at TPAC, at the AC Meeting or via teleconference. Video Conference interviews are acceptable. - The NomCom announces the selected candidates. An objection process opens to allow formal objections from the AC. Formal objections will be handled as normal. After 28 days, the formal objection window closes and the newly selected individuals begin their term.
process
tag appointment should be via ietf style nomcom the tag is currently elected we could move to an ietf style nomcom appointment this ensures the right people for the role are selected and a balanced tag can be achieved comments are welcome the nomcom consists of a random set of volunteers who meet a set criteria e g have chaired a group or published a draft in the last x years the ab could be members of the nomcom the nomcom should have advising members they cannot vote only advise the tag chair a developer invited expert would be good choices nominations for groups go direct to the nomcom nominations can come from any source the nomcom will see via the application what the source of the nomination is the nomcom can choose to weigh an ac nomination in a positive manner if they wish this means that ac nominations count for something but other groups e g independent developers are not blocked from applying nomcom sees nominations as they come in if they feel a diverse pool isn’t being reached diversity for all items including gender technical knowledge of particular topics organisation representation size global location etc the nomcom will begin pushing for more candidates nomcom starts the process of selecting candidates this is as follows the nomcom meets before evaluating the candidates it agrees on key qualities it is looking for in the candidates for the open roles it is likely these qualities will be the same in each election although sometimes they will be different e g “chair of the tag will require different qualities to “tag member” consensus on these qualities is reached all applications are read obvious “no” candidates are discarded the nomcom meets to discuss the “possible” candidates remaining further candidates are discarded for reasons the nomcom deem appropriate these could be because there has been an influx of candidates which represent one industry or many candidates are experts in only one field the nomcom is left with the final set of “possible candidates” these candidates are interviewed for the role at tpac at the ac meeting or via teleconference video conference interviews are acceptable the nomcom announces the selected candidates an objection process opens to allow formal objections from the ac formal objections will be handled as normal after days the formal objection window closes and the newly selected individuals begin their term
1
201,384
15,802,252,754
IssuesEvent
2021-04-03 08:49:38
LJ-37/ped
https://api.github.com/repos/LJ-37/ped
opened
Typo in UG
severity.Medium type.DocumentationBug
Tag for view food category should be f/ but is written as a/ in the UG under view diet data. ![image.png](https://raw.githubusercontent.com/LJ-37/ped/main/files/bcedb91b-8cb7-4b27-8741-b3eea04f6248.png) <!--session: 1617437380999-4a5bfdbc-3d33-43fd-9397-daa368b5b349-->
1.0
Typo in UG - Tag for view food category should be f/ but is written as a/ in the UG under view diet data. ![image.png](https://raw.githubusercontent.com/LJ-37/ped/main/files/bcedb91b-8cb7-4b27-8741-b3eea04f6248.png) <!--session: 1617437380999-4a5bfdbc-3d33-43fd-9397-daa368b5b349-->
non_process
typo in ug tag for view food category should be f but is written as a in the ug under view diet data
0
15,296
2,850,599,479
IssuesEvent
2015-05-31 18:21:31
damonkohler/sl4a
https://api.github.com/repos/damonkohler/sl4a
opened
SL4A Force Close on droid.startActivityIntent(chooserIntent)
auto-migrated Priority-Medium Type-Defect
_From @GoogleCodeExporter on May 31, 2015 11:30_ ``` What device(s) are you experiencing the problem on? Samsung Vibrant (SGH-T959) What firmware version are you running on the device? 2.2 What steps will reproduce the problem? 1. Run the attached Python script on an Android device with SL4Ar4 -or- 1. Make an intent with action="android.intent.action.SEND", uriType = "text/plain", extras = {"EXTRA_TEXT":"my message", "EXTRA_SUBJECT":"and my subject"}, and everything else = None. 2. Make another intent with action = "android.intent.action.CHOOSER", extras = {"EXTRA_INTENT":theFirstIntent}, and everything else = None. 3. droid.startActivityIntent(yourSecondIntent) 4. SL4A force closes What is the expected output? What do you see instead? I'm trying to get a chooser for sharing text via other apps (Gmail, Messaging, etc.). It should be just like the one from the "Share" option in the menu of SL4A's script editor activity. Instead, SL4A force closes, and there's no indication whatsoever to the Python interpreter. What version of the product are you using? On what operating system? I'm using SL4Ar4 on Samung's OEM build FROYO.UVKB5 My development machine is Windows 7 64-bit, but running the script directly on the phone still crashes, so that's probably not the issue. Please provide any additional information below. I tried to adapt the code from the "Share" option of the SL4A script editor by manually making the same intent as Intent.createChooser() as described in the Android documentation. My script is adapted from lines 198-202 here: http://code.google.com/p/android-scripting/source/diff?spec=svndeba523e8a9b44ea6 ef6e2497e21914084b109a8&r=deba523e8a9b44ea6ef6e2497e21914084b109a8&format=side&p ath=/android/ScriptingLayerForAndroid/src/com/googlecode/android_scripting/activ ity/ScriptEditor.java And this describes how to make the chooser intent: http://developer.android.com/reference/android/content/Intent.html#ACTION_CHOOSE R ``` Original issue reported on code.google.com by `armadaje...@gmail.com` on 5 Dec 2011 at 12:55 Attachments: * [logcat.txt](https://storage.googleapis.com/google-code-attachments/android-scripting/issue-588/comment-0/logcat.txt) * [chooser.py](https://storage.googleapis.com/google-code-attachments/android-scripting/issue-588/comment-0/chooser.py) _Copied from original issue: damonkohler/android-scripting#588_
1.0
SL4A Force Close on droid.startActivityIntent(chooserIntent) - _From @GoogleCodeExporter on May 31, 2015 11:30_ ``` What device(s) are you experiencing the problem on? Samsung Vibrant (SGH-T959) What firmware version are you running on the device? 2.2 What steps will reproduce the problem? 1. Run the attached Python script on an Android device with SL4Ar4 -or- 1. Make an intent with action="android.intent.action.SEND", uriType = "text/plain", extras = {"EXTRA_TEXT":"my message", "EXTRA_SUBJECT":"and my subject"}, and everything else = None. 2. Make another intent with action = "android.intent.action.CHOOSER", extras = {"EXTRA_INTENT":theFirstIntent}, and everything else = None. 3. droid.startActivityIntent(yourSecondIntent) 4. SL4A force closes What is the expected output? What do you see instead? I'm trying to get a chooser for sharing text via other apps (Gmail, Messaging, etc.). It should be just like the one from the "Share" option in the menu of SL4A's script editor activity. Instead, SL4A force closes, and there's no indication whatsoever to the Python interpreter. What version of the product are you using? On what operating system? I'm using SL4Ar4 on Samung's OEM build FROYO.UVKB5 My development machine is Windows 7 64-bit, but running the script directly on the phone still crashes, so that's probably not the issue. Please provide any additional information below. I tried to adapt the code from the "Share" option of the SL4A script editor by manually making the same intent as Intent.createChooser() as described in the Android documentation. My script is adapted from lines 198-202 here: http://code.google.com/p/android-scripting/source/diff?spec=svndeba523e8a9b44ea6 ef6e2497e21914084b109a8&r=deba523e8a9b44ea6ef6e2497e21914084b109a8&format=side&p ath=/android/ScriptingLayerForAndroid/src/com/googlecode/android_scripting/activ ity/ScriptEditor.java And this describes how to make the chooser intent: http://developer.android.com/reference/android/content/Intent.html#ACTION_CHOOSE R ``` Original issue reported on code.google.com by `armadaje...@gmail.com` on 5 Dec 2011 at 12:55 Attachments: * [logcat.txt](https://storage.googleapis.com/google-code-attachments/android-scripting/issue-588/comment-0/logcat.txt) * [chooser.py](https://storage.googleapis.com/google-code-attachments/android-scripting/issue-588/comment-0/chooser.py) _Copied from original issue: damonkohler/android-scripting#588_
non_process
force close on droid startactivityintent chooserintent from googlecodeexporter on may what device s are you experiencing the problem on samsung vibrant sgh what firmware version are you running on the device what steps will reproduce the problem run the attached python script on an android device with or make an intent with action android intent action send uritype text plain extras extra text my message extra subject and my subject and everything else none make another intent with action android intent action chooser extras extra intent thefirstintent and everything else none droid startactivityintent yoursecondintent force closes what is the expected output what do you see instead i m trying to get a chooser for sharing text via other apps gmail messaging etc it should be just like the one from the share option in the menu of s script editor activity instead force closes and there s no indication whatsoever to the python interpreter what version of the product are you using on what operating system i m using on samung s oem build froyo my development machine is windows bit but running the script directly on the phone still crashes so that s probably not the issue please provide any additional information below i tried to adapt the code from the share option of the script editor by manually making the same intent as intent createchooser as described in the android documentation my script is adapted from lines here r format side p ath android scriptinglayerforandroid src com googlecode android scripting activ ity scripteditor java and this describes how to make the chooser intent r original issue reported on code google com by armadaje gmail com on dec at attachments copied from original issue damonkohler android scripting
0
10,455
13,234,960,625
IssuesEvent
2020-08-18 17:11:26
googleapis/repo-automation-bots
https://api.github.com/repos/googleapis/repo-automation-bots
closed
bug: refactor for octokit/webhooks
type: process
We either should refactor release-please to not use octokit/webhooks or add it to ignore modules See #816
1.0
bug: refactor for octokit/webhooks - We either should refactor release-please to not use octokit/webhooks or add it to ignore modules See #816
process
bug refactor for octokit webhooks we either should refactor release please to not use octokit webhooks or add it to ignore modules see
1
340,736
24,668,635,804
IssuesEvent
2022-10-18 12:16:52
mantidproject/mantid
https://api.github.com/repos/mantidproject/mantid
opened
dev-docs pages not removed on site update
Documentation ISIS Team: Core
Found here: https://github.com/mantidproject/mantid/issues/34357#issuecomment-1248311915 **Describe the bug** A dev docs file was renamed. The new page exists on the dev docs website, but so does the old. Old https://developer.mantidproject.org/Testing/SANSGUI/SANSGUITests.html New https://developer.mantidproject.org/Testing/SANSGUI/ISISSANSGUITests.html **Expected behavior** If a file is removed (or renamed) from the dev docs (or the user docs), the old page is removed from the website. A change is needed here: https://github.com/mantidproject/mantid/blob/main/buildconfig/Jenkins/Conda/build-and-publish-devsite.sh Possibly do not use the old clone of the website. **Screenshots** ![Screenshot 2022-10-18 at 13 15 27](https://user-images.githubusercontent.com/55980573/196426684-3adfbd01-d66d-4922-a878-d3b062ffb984.png) ![Screenshot 2022-10-18 at 13 16 10](https://user-images.githubusercontent.com/55980573/196426835-39b489ea-9bac-4b20-aea5-17659663841f.png)
1.0
dev-docs pages not removed on site update - Found here: https://github.com/mantidproject/mantid/issues/34357#issuecomment-1248311915 **Describe the bug** A dev docs file was renamed. The new page exists on the dev docs website, but so does the old. Old https://developer.mantidproject.org/Testing/SANSGUI/SANSGUITests.html New https://developer.mantidproject.org/Testing/SANSGUI/ISISSANSGUITests.html **Expected behavior** If a file is removed (or renamed) from the dev docs (or the user docs), the old page is removed from the website. A change is needed here: https://github.com/mantidproject/mantid/blob/main/buildconfig/Jenkins/Conda/build-and-publish-devsite.sh Possibly do not use the old clone of the website. **Screenshots** ![Screenshot 2022-10-18 at 13 15 27](https://user-images.githubusercontent.com/55980573/196426684-3adfbd01-d66d-4922-a878-d3b062ffb984.png) ![Screenshot 2022-10-18 at 13 16 10](https://user-images.githubusercontent.com/55980573/196426835-39b489ea-9bac-4b20-aea5-17659663841f.png)
non_process
dev docs pages not removed on site update found here describe the bug a dev docs file was renamed the new page exists on the dev docs website but so does the old old new expected behavior if a file is removed or renamed from the dev docs or the user docs the old page is removed from the website a change is needed here possibly do not use the old clone of the website screenshots
0
1,681
2,658,826,471
IssuesEvent
2015-03-18 17:34:09
phetsims/pendulum-lab
https://api.github.com/repos/phetsims/pendulum-lab
opened
Missing assets/*-screenshot.png
code review
Noticed during code review #56. Not sure who should be assigned for this.
1.0
Missing assets/*-screenshot.png - Noticed during code review #56. Not sure who should be assigned for this.
non_process
missing assets screenshot png noticed during code review not sure who should be assigned for this
0
202,660
15,837,159,305
IssuesEvent
2021-04-06 20:22:37
HARDTECHIO/dhoa-front
https://api.github.com/repos/HARDTECHIO/dhoa-front
closed
Aplicar Componentes do Template/MENU
documentation enhancement
## Tarefas - [x] Criar Template Menu; - [x] Criar NavBar com os componentes com Bootstrap;
1.0
Aplicar Componentes do Template/MENU - ## Tarefas - [x] Criar Template Menu; - [x] Criar NavBar com os componentes com Bootstrap;
non_process
aplicar componentes do template menu tarefas criar template menu criar navbar com os componentes com bootstrap
0
136,853
12,736,517,785
IssuesEvent
2020-06-25 17:03:23
ualberta-smr/LibCompPlugin
https://api.github.com/repos/ualberta-smr/LibCompPlugin
closed
Finalize Paper
documentation
As I work on enhancements to LibComp, stay on track and continue to update the paper after each task. Working on paper will take longer for the user study as I will have to analyze the data and re-write the entire section so I set up some time specifically for this task. Goal: A good publishable paper for a tool tract.
1.0
Finalize Paper - As I work on enhancements to LibComp, stay on track and continue to update the paper after each task. Working on paper will take longer for the user study as I will have to analyze the data and re-write the entire section so I set up some time specifically for this task. Goal: A good publishable paper for a tool tract.
non_process
finalize paper as i work on enhancements to libcomp stay on track and continue to update the paper after each task working on paper will take longer for the user study as i will have to analyze the data and re write the entire section so i set up some time specifically for this task goal a good publishable paper for a tool tract
0
131,172
18,244,879,937
IssuesEvent
2021-10-01 17:01:11
protocol/nft-website
https://api.github.com/repos/protocol/nft-website
closed
[CONTENT] Write section summary content
help wanted P2 kind/enhancement dif/medium effort/hours topic/design-content
Write brief, TLDR-style summary pages for each content section: - [ ] `/concepts/` - [ ] `/tutorial/` - [ ] `/how-to/` - [ ] `/reference/` Each of these already appears as a `README.md` in their respective directories, just without any page body content. When done, do two things to make them appear: - [ ] Remove their redirects to front-of-site from `docs/.vuepress/redirects` - [ ] In site nav in `docs/.vuepress/config.js`, add a line for `link: 'foo'`after each section's `title:`, so nav section headers become clickable links to summary pages - [ ] May be necessary to retitle or modify `/contribute/' in nav for consistency
1.0
[CONTENT] Write section summary content - Write brief, TLDR-style summary pages for each content section: - [ ] `/concepts/` - [ ] `/tutorial/` - [ ] `/how-to/` - [ ] `/reference/` Each of these already appears as a `README.md` in their respective directories, just without any page body content. When done, do two things to make them appear: - [ ] Remove their redirects to front-of-site from `docs/.vuepress/redirects` - [ ] In site nav in `docs/.vuepress/config.js`, add a line for `link: 'foo'`after each section's `title:`, so nav section headers become clickable links to summary pages - [ ] May be necessary to retitle or modify `/contribute/' in nav for consistency
non_process
write section summary content write brief tldr style summary pages for each content section concepts tutorial how to reference each of these already appears as a readme md in their respective directories just without any page body content when done do two things to make them appear remove their redirects to front of site from docs vuepress redirects in site nav in docs vuepress config js add a line for link foo after each section s title so nav section headers become clickable links to summary pages may be necessary to retitle or modify contribute in nav for consistency
0
29,696
8,392,143,480
IssuesEvent
2018-10-09 16:46:22
trilinos/Trilinos
https://api.github.com/repos/trilinos/Trilinos
opened
PyTrilinos: Standardize Configuration Macros
PyTrilinos build
@trilinos/pytrilinos ## Expectations Configuration macros for PyTrilinos should be consistently named and distinct from other packages. ## Current Behavior Some configuration macros are not distinct (`HAVE_EPETRA`), some _are_ distinct (`HAVE_PYTRILINOS_AZTECOO`), and some are redundant (`HAVE_PYTRILINOS_EPETRA`). ## Motivation and Context @mhoemmen made some suggestions while reviewing #3575, which led to the observation that this needs to be cleaned up. ## Definition of Done - [ ] Eliminate `HAVE_<PACKAGE>` macros in favor of `HAVE_PYTRILINOS_<PACKAGE>` macros in `PyTrilinos_config.h.in` - [ ] Make same changes in PyTrilinos headers and source files - [ ] Test for Teuchos-only configuration - [ ] Test for linear solver only configuration - [ ] Test for linear and nonlinear solver only configuration - [ ] Test for full configuration
1.0
PyTrilinos: Standardize Configuration Macros - @trilinos/pytrilinos ## Expectations Configuration macros for PyTrilinos should be consistently named and distinct from other packages. ## Current Behavior Some configuration macros are not distinct (`HAVE_EPETRA`), some _are_ distinct (`HAVE_PYTRILINOS_AZTECOO`), and some are redundant (`HAVE_PYTRILINOS_EPETRA`). ## Motivation and Context @mhoemmen made some suggestions while reviewing #3575, which led to the observation that this needs to be cleaned up. ## Definition of Done - [ ] Eliminate `HAVE_<PACKAGE>` macros in favor of `HAVE_PYTRILINOS_<PACKAGE>` macros in `PyTrilinos_config.h.in` - [ ] Make same changes in PyTrilinos headers and source files - [ ] Test for Teuchos-only configuration - [ ] Test for linear solver only configuration - [ ] Test for linear and nonlinear solver only configuration - [ ] Test for full configuration
non_process
pytrilinos standardize configuration macros trilinos pytrilinos expectations configuration macros for pytrilinos should be consistently named and distinct from other packages current behavior some configuration macros are not distinct have epetra some are distinct have pytrilinos aztecoo and some are redundant have pytrilinos epetra motivation and context mhoemmen made some suggestions while reviewing which led to the observation that this needs to be cleaned up definition of done eliminate have macros in favor of have pytrilinos macros in pytrilinos config h in make same changes in pytrilinos headers and source files test for teuchos only configuration test for linear solver only configuration test for linear and nonlinear solver only configuration test for full configuration
0
15,051
18,762,895,012
IssuesEvent
2021-11-05 18:46:18
GoogleCloudPlatform/ai-platform-samples
https://api.github.com/repos/GoogleCloudPlatform/ai-platform-samples
closed
Python 3.5 CI builds are failing
type: process
Python 3.5 CI builds are failing with this error: `FileNotFoundError: [Errno 2] No such file or directory: '/tmpfs/src/envs/python3.5/venv'` [Example failed PR build](https://github.com/GoogleCloudPlatform/ai-platform-samples/pull/514)
1.0
Python 3.5 CI builds are failing - Python 3.5 CI builds are failing with this error: `FileNotFoundError: [Errno 2] No such file or directory: '/tmpfs/src/envs/python3.5/venv'` [Example failed PR build](https://github.com/GoogleCloudPlatform/ai-platform-samples/pull/514)
process
python ci builds are failing python ci builds are failing with this error filenotfounderror no such file or directory tmpfs src envs venv
1
7,248
9,527,289,699
IssuesEvent
2019-04-29 02:59:04
Lothrazar/Cyclic
https://api.github.com/repos/Lothrazar/Cyclic
closed
Garden Scythe Harvesting Bugs
bug: gameplay mod compatibility
Minecraft version & Mod Version: - Forge v14.23.5.2796-1.12.2 - Cyclic v1.17.11 Single player or Server: - Probably Both Describe problem (what you were doing / what happened): - There are multiple bugs that I noticed when harvesting modded crops using the garden scythe. (See below for id's) 1. The Red Orchid from [Extra Utilities 2](https://minecraft.curseforge.com/projects/extra-utilities) is harvested, but not replanted. 2. The Enderlilly from [Extra Utilities 2](https://minecraft.curseforge.com/projects/extra-utilities) is not harvested or replanted. 3. The Corn from [Simple Corn](https://minecraft.curseforge.com/projects/simple-corn) plants a 'ghost' plant. It simply disappears after updating the block. 4. The Grape Stem from [Rustic](https://minecraft.curseforge.com/projects/rustic) is still being harvested instead of the fruit. (See image below) 5. Tomatoes and Peppers from [Rustic](https://minecraft.curseforge.com/projects/rustic) can both grow 2-3 blocks tall on the stake, but harvesting sets it back to stage 1 instead of keeping it at stage 3. (See image below) 6. The Blueberry Bush from [Nifty](https://minecraft.curseforge.com/projects/niftyblocks?gameCategorySlug=mc-mods&projectID=228189) gives me the bush instead of the fruit when the scythe is used. 7. The Bushes from [Minestrappolation 5](https://minecraft.curseforge.com/projects/minestrappolation-5) also gives me bushes instead of the corresponding fruit. 8. The Wildberry Bush from [Rustic](https://minecraft.curseforge.com/projects/rustic) can't be harvested using the scythe. 9. The Bushes from [Plants 2](https://minecraft.curseforge.com/projects/plants) also aren't able to be harvested for fruit using the scythe. 10. The Bushes and the fruit hanging on trees from [Plant Mega Pack](http://10paksmods.net/pmp.html) are harvested from, but it doesn't give the fruit to you. 11. The Climbing Vine from [Plant Mega Pack gives](http://10paksmods.net/pmp.html) the vine instead of the fruit. ``` extrautils2:enderlilly extrautils2:redorchid minestrapp:blueberry_bush minestrapp:blackberry_bush minestrapp:raspberry_bush minestrapp:strawberry_bush minestrapp:mana_bush minestrapp:voidberry_bush nifty:blueberry_bush plants2:harvest_0:* plants2:harvest_1:* plants2:nether_harvest:* plants2:double_harvest_0 plants2:bush:* pmp:bbush_beaut pmp:bbush_black pmp:bbush_blue pmp:bbush_elder pmp:bbush_goose pmp:bbush_huckl pmp:bbush_orang pmp:bbush_snow pmp:bbush_straw pmp:climb_porce rustic:wildberry_bush rustic:grape_stem rustic:grape_leaves rustic:grapes rustic:chili_crop rustic:tomato_crop simplecorn:corn ``` ![2018-12-19_11 49 57](https://user-images.githubusercontent.com/12000593/50392347-4a125800-0713-11e9-9ef2-d1d5e7d82366.png) ![2018-12-19_11 49 26](https://user-images.githubusercontent.com/12000593/50393459-fbb58700-071b-11e9-86c5-857c88274aa5.png)
True
Garden Scythe Harvesting Bugs - Minecraft version & Mod Version: - Forge v14.23.5.2796-1.12.2 - Cyclic v1.17.11 Single player or Server: - Probably Both Describe problem (what you were doing / what happened): - There are multiple bugs that I noticed when harvesting modded crops using the garden scythe. (See below for id's) 1. The Red Orchid from [Extra Utilities 2](https://minecraft.curseforge.com/projects/extra-utilities) is harvested, but not replanted. 2. The Enderlilly from [Extra Utilities 2](https://minecraft.curseforge.com/projects/extra-utilities) is not harvested or replanted. 3. The Corn from [Simple Corn](https://minecraft.curseforge.com/projects/simple-corn) plants a 'ghost' plant. It simply disappears after updating the block. 4. The Grape Stem from [Rustic](https://minecraft.curseforge.com/projects/rustic) is still being harvested instead of the fruit. (See image below) 5. Tomatoes and Peppers from [Rustic](https://minecraft.curseforge.com/projects/rustic) can both grow 2-3 blocks tall on the stake, but harvesting sets it back to stage 1 instead of keeping it at stage 3. (See image below) 6. The Blueberry Bush from [Nifty](https://minecraft.curseforge.com/projects/niftyblocks?gameCategorySlug=mc-mods&projectID=228189) gives me the bush instead of the fruit when the scythe is used. 7. The Bushes from [Minestrappolation 5](https://minecraft.curseforge.com/projects/minestrappolation-5) also gives me bushes instead of the corresponding fruit. 8. The Wildberry Bush from [Rustic](https://minecraft.curseforge.com/projects/rustic) can't be harvested using the scythe. 9. The Bushes from [Plants 2](https://minecraft.curseforge.com/projects/plants) also aren't able to be harvested for fruit using the scythe. 10. The Bushes and the fruit hanging on trees from [Plant Mega Pack](http://10paksmods.net/pmp.html) are harvested from, but it doesn't give the fruit to you. 11. The Climbing Vine from [Plant Mega Pack gives](http://10paksmods.net/pmp.html) the vine instead of the fruit. ``` extrautils2:enderlilly extrautils2:redorchid minestrapp:blueberry_bush minestrapp:blackberry_bush minestrapp:raspberry_bush minestrapp:strawberry_bush minestrapp:mana_bush minestrapp:voidberry_bush nifty:blueberry_bush plants2:harvest_0:* plants2:harvest_1:* plants2:nether_harvest:* plants2:double_harvest_0 plants2:bush:* pmp:bbush_beaut pmp:bbush_black pmp:bbush_blue pmp:bbush_elder pmp:bbush_goose pmp:bbush_huckl pmp:bbush_orang pmp:bbush_snow pmp:bbush_straw pmp:climb_porce rustic:wildberry_bush rustic:grape_stem rustic:grape_leaves rustic:grapes rustic:chili_crop rustic:tomato_crop simplecorn:corn ``` ![2018-12-19_11 49 57](https://user-images.githubusercontent.com/12000593/50392347-4a125800-0713-11e9-9ef2-d1d5e7d82366.png) ![2018-12-19_11 49 26](https://user-images.githubusercontent.com/12000593/50393459-fbb58700-071b-11e9-86c5-857c88274aa5.png)
non_process
garden scythe harvesting bugs minecraft version mod version forge cyclic single player or server probably both describe problem what you were doing what happened there are multiple bugs that i noticed when harvesting modded crops using the garden scythe see below for id s the red orchid from is harvested but not replanted the enderlilly from is not harvested or replanted the corn from plants a ghost plant it simply disappears after updating the block the grape stem from is still being harvested instead of the fruit see image below tomatoes and peppers from can both grow blocks tall on the stake but harvesting sets it back to stage instead of keeping it at stage see image below the blueberry bush from gives me the bush instead of the fruit when the scythe is used the bushes from also gives me bushes instead of the corresponding fruit the wildberry bush from can t be harvested using the scythe the bushes from also aren t able to be harvested for fruit using the scythe the bushes and the fruit hanging on trees from are harvested from but it doesn t give the fruit to you the climbing vine from the vine instead of the fruit enderlilly redorchid minestrapp blueberry bush minestrapp blackberry bush minestrapp raspberry bush minestrapp strawberry bush minestrapp mana bush minestrapp voidberry bush nifty blueberry bush harvest harvest nether harvest double harvest bush pmp bbush beaut pmp bbush black pmp bbush blue pmp bbush elder pmp bbush goose pmp bbush huckl pmp bbush orang pmp bbush snow pmp bbush straw pmp climb porce rustic wildberry bush rustic grape stem rustic grape leaves rustic grapes rustic chili crop rustic tomato crop simplecorn corn
0
233,521
25,765,525,473
IssuesEvent
2022-12-09 01:17:21
jasonjberry/CDM
https://api.github.com/repos/jasonjberry/CDM
opened
CVE-2022-23491 (Medium) detected in certifi-2019.6.16-py2.py3-none-any.whl
security vulnerability
## CVE-2022-23491 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>certifi-2019.6.16-py2.py3-none-any.whl</b></p></summary> <p>Python package for providing Mozilla's CA Bundle.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/69/1b/b853c7a9d4f6a6d00749e94eb6f3a041e342a885b87340b79c1ef73e3a78/certifi-2019.6.16-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/69/1b/b853c7a9d4f6a6d00749e94eb6f3a041e342a885b87340b79c1ef73e3a78/certifi-2019.6.16-py2.py3-none-any.whl</a></p> <p>Path to dependency file: /objectModel/Python</p> <p>Path to vulnerable library: /objectModel/Python,/objectModel/Python/requirements.txt</p> <p> Dependency Hierarchy: - :x: **certifi-2019.6.16-py2.py3-none-any.whl** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jasonjberry/CDM/commit/2f532047f0b9e231c1602df6a076c534fba6d73a">2f532047f0b9e231c1602df6a076c534fba6d73a</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Certifi is a curated collection of Root Certificates for validating the trustworthiness of SSL certificates while verifying the identity of TLS hosts. Certifi 2022.12.07 removes root certificates from "TrustCor" from the root store. These are in the process of being removed from Mozilla's trust store. TrustCor's root certificates are being removed pursuant to an investigation prompted by media reporting that TrustCor's ownership also operated a business that produced spyware. Conclusions of Mozilla's investigation can be found in the linked google group discussion. <p>Publish Date: 2022-12-07 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-23491>CVE-2022-23491</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Changed - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2022-23491">https://www.cve.org/CVERecord?id=CVE-2022-23491</a></p> <p>Release Date: 2022-12-07</p> <p>Fix Resolution: certifi - 2022.12.07</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-23491 (Medium) detected in certifi-2019.6.16-py2.py3-none-any.whl - ## CVE-2022-23491 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>certifi-2019.6.16-py2.py3-none-any.whl</b></p></summary> <p>Python package for providing Mozilla's CA Bundle.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/69/1b/b853c7a9d4f6a6d00749e94eb6f3a041e342a885b87340b79c1ef73e3a78/certifi-2019.6.16-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/69/1b/b853c7a9d4f6a6d00749e94eb6f3a041e342a885b87340b79c1ef73e3a78/certifi-2019.6.16-py2.py3-none-any.whl</a></p> <p>Path to dependency file: /objectModel/Python</p> <p>Path to vulnerable library: /objectModel/Python,/objectModel/Python/requirements.txt</p> <p> Dependency Hierarchy: - :x: **certifi-2019.6.16-py2.py3-none-any.whl** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jasonjberry/CDM/commit/2f532047f0b9e231c1602df6a076c534fba6d73a">2f532047f0b9e231c1602df6a076c534fba6d73a</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Certifi is a curated collection of Root Certificates for validating the trustworthiness of SSL certificates while verifying the identity of TLS hosts. Certifi 2022.12.07 removes root certificates from "TrustCor" from the root store. These are in the process of being removed from Mozilla's trust store. TrustCor's root certificates are being removed pursuant to an investigation prompted by media reporting that TrustCor's ownership also operated a business that produced spyware. Conclusions of Mozilla's investigation can be found in the linked google group discussion. <p>Publish Date: 2022-12-07 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-23491>CVE-2022-23491</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Changed - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2022-23491">https://www.cve.org/CVERecord?id=CVE-2022-23491</a></p> <p>Release Date: 2022-12-07</p> <p>Fix Resolution: certifi - 2022.12.07</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in certifi none any whl cve medium severity vulnerability vulnerable library certifi none any whl python package for providing mozilla s ca bundle library home page a href path to dependency file objectmodel python path to vulnerable library objectmodel python objectmodel python requirements txt dependency hierarchy x certifi none any whl vulnerable library found in head commit a href found in base branch master vulnerability details certifi is a curated collection of root certificates for validating the trustworthiness of ssl certificates while verifying the identity of tls hosts certifi removes root certificates from trustcor from the root store these are in the process of being removed from mozilla s trust store trustcor s root certificates are being removed pursuant to an investigation prompted by media reporting that trustcor s ownership also operated a business that produced spyware conclusions of mozilla s investigation can be found in the linked google group discussion publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required high user interaction none scope changed impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution certifi step up your open source security game with mend
0
7,483
10,574,172,509
IssuesEvent
2019-10-07 13:32:04
prisma/lift
https://api.github.com/repos/prisma/lift
closed
Incorrect database name error
bug/2-confirmed kind/bug process/candidate
A database created via lift fails at `lift save`, the database name in this case is `p2-mysql-views` To reproduce: 1. Create a blank project starter 2. Choose MySQL and create a new database with name `p2-mysql-view` 3. Run `prisma2 lift save --name init` Panics with ``` divyendusingh [p2-mysql-view]$ prisma2 lift save --name init ERROR Oops, an unexpected error occured! Error in migration engine: thread 'main' panicked at 'CommandError(Generic { code: 1000, error: "QueryError(QueryError(MyS qlError { ERROR 1102 (42000): Incorrect database name \'\' }\n\nstack backtrace:\n 0: backtrace::backtrace::trace\n 1: backtrace::capture::Backtrace::new_unresolved\n 2: failure::backtrace::internal::InternalBacktrace::new\n 3: <failure ::backtrace::Backtrace as core::default::Default>::default\n 4: prisma_query::connector::mysql::error::<impl core::conve rt::From<mysql::error::Error> for prisma_query::error::Error>::from\n 5: prisma_query::connector::metrics::query\n 6: <prisma_query::connector::mysql::Mysql as prisma_query::connector::queryable::Queryable>::query_raw\n 7: <sql_migration_ connector::migration_database::Mysql as sql_migration_connector::migration_database::MigrationDatabase>::query_raw\n 8: <sql_migration_connector::SqlMigrationConnector as migration_connector::MigrationConnector>::initialize\n 9: migration_e ngine::migration_engine::MigrationEngine<C,D>::new\n 10: migration_engine::api::rpc::RpcApi::new\n 11: migration_engine: :api::rpc::RpcApi::new_async\n 12: migration_engine::main\n 13: std::rt::lang_start::{{closure}}\n 14: std::panicking:: ERROR Oops, an unexpected error occured! Error in migration engine: thread 'main' panicked at 'CommandError(Generic { code: 1000, error: "QueryError(QueryError(MyS qlError { ERROR 1102 (42000): Incorrect database name \'\' }\n\nstack backtrace:\n 0: backtrace::backtrace::trace\n 1: backtrace::capture::Backtrace::new_unresolved\n 2: failure::backtrace::internal::InternalBacktrace::new\n 3: <failure ::backtrace::Backtrace as core::default::Default>::default\n 4: prisma_query::connector::mysql::error::<impl core::conve rt::From<mysql::error::Error> for prisma_query::error::Error>::from\n 5: prisma_query::connector::metrics::query\n 6: <prisma_query::connector::mysql::Mysql as prisma_query::connector::queryable::Queryable>::query_raw\n 7: <sql_migration_ connector::migration_database::Mysql as sql_migration_connector::migration_database::MigrationDatabase>::query_raw\n 8: <sql_migration_connector::SqlMigrationConnector as migration_connector::MigrationConnector>::initialize\n 9: migration_e ngine::migration_engine::MigrationEngine<C,D>::new\n 10: migration_engine::api::rpc::RpcApi::new\n 11: migration_engine: :api::rpc::RpcApi::new_async\n 12: migration_engine::main\n 13: std::rt::lang_start::{{closure}}\n 14: std::panicking:: try::do_call\n 15: __rust_maybe_catch_panic\n 16: std::rt::lang_start_internal\n 17: main\n)\n\nstack backtrace:\n 0: backtrace::backtrace::trace\n 1: backtrace::capture::Backtrace::new_unresolved\n 2: failure::backtrace::internal::Int ernalBacktrace::new\n 3: <failure::backtrace::Backtrace as core::default::Default>::default\n 4: <migration_connector ```
1.0
Incorrect database name error - A database created via lift fails at `lift save`, the database name in this case is `p2-mysql-views` To reproduce: 1. Create a blank project starter 2. Choose MySQL and create a new database with name `p2-mysql-view` 3. Run `prisma2 lift save --name init` Panics with ``` divyendusingh [p2-mysql-view]$ prisma2 lift save --name init ERROR Oops, an unexpected error occured! Error in migration engine: thread 'main' panicked at 'CommandError(Generic { code: 1000, error: "QueryError(QueryError(MyS qlError { ERROR 1102 (42000): Incorrect database name \'\' }\n\nstack backtrace:\n 0: backtrace::backtrace::trace\n 1: backtrace::capture::Backtrace::new_unresolved\n 2: failure::backtrace::internal::InternalBacktrace::new\n 3: <failure ::backtrace::Backtrace as core::default::Default>::default\n 4: prisma_query::connector::mysql::error::<impl core::conve rt::From<mysql::error::Error> for prisma_query::error::Error>::from\n 5: prisma_query::connector::metrics::query\n 6: <prisma_query::connector::mysql::Mysql as prisma_query::connector::queryable::Queryable>::query_raw\n 7: <sql_migration_ connector::migration_database::Mysql as sql_migration_connector::migration_database::MigrationDatabase>::query_raw\n 8: <sql_migration_connector::SqlMigrationConnector as migration_connector::MigrationConnector>::initialize\n 9: migration_e ngine::migration_engine::MigrationEngine<C,D>::new\n 10: migration_engine::api::rpc::RpcApi::new\n 11: migration_engine: :api::rpc::RpcApi::new_async\n 12: migration_engine::main\n 13: std::rt::lang_start::{{closure}}\n 14: std::panicking:: ERROR Oops, an unexpected error occured! Error in migration engine: thread 'main' panicked at 'CommandError(Generic { code: 1000, error: "QueryError(QueryError(MyS qlError { ERROR 1102 (42000): Incorrect database name \'\' }\n\nstack backtrace:\n 0: backtrace::backtrace::trace\n 1: backtrace::capture::Backtrace::new_unresolved\n 2: failure::backtrace::internal::InternalBacktrace::new\n 3: <failure ::backtrace::Backtrace as core::default::Default>::default\n 4: prisma_query::connector::mysql::error::<impl core::conve rt::From<mysql::error::Error> for prisma_query::error::Error>::from\n 5: prisma_query::connector::metrics::query\n 6: <prisma_query::connector::mysql::Mysql as prisma_query::connector::queryable::Queryable>::query_raw\n 7: <sql_migration_ connector::migration_database::Mysql as sql_migration_connector::migration_database::MigrationDatabase>::query_raw\n 8: <sql_migration_connector::SqlMigrationConnector as migration_connector::MigrationConnector>::initialize\n 9: migration_e ngine::migration_engine::MigrationEngine<C,D>::new\n 10: migration_engine::api::rpc::RpcApi::new\n 11: migration_engine: :api::rpc::RpcApi::new_async\n 12: migration_engine::main\n 13: std::rt::lang_start::{{closure}}\n 14: std::panicking:: try::do_call\n 15: __rust_maybe_catch_panic\n 16: std::rt::lang_start_internal\n 17: main\n)\n\nstack backtrace:\n 0: backtrace::backtrace::trace\n 1: backtrace::capture::Backtrace::new_unresolved\n 2: failure::backtrace::internal::Int ernalBacktrace::new\n 3: <failure::backtrace::Backtrace as core::default::Default>::default\n 4: <migration_connector ```
process
incorrect database name error a database created via lift fails at lift save the database name in this case is mysql views to reproduce create a blank project starter choose mysql and create a new database with name mysql view run lift save name init panics with divyendusingh lift save name init error oops an unexpected error occured error in migration engine thread main panicked at commanderror generic code error queryerror queryerror mys qlerror error incorrect database name n nstack backtrace n backtrace backtrace trace n backtrace capture backtrace new unresolved n failure backtrace internal internalbacktrace new n failure backtrace backtrace as core default default default n prisma query connector mysql error impl core conve rt from for prisma query error error from n prisma query connector metrics query n query raw n sql migration connector migration database mysql as sql migration connector migration database migrationdatabase query raw n initialize n migration e ngine migration engine migrationengine new n migration engine api rpc rpcapi new n migration engine api rpc rpcapi new async n migration engine main n std rt lang start closure n std panicking error oops an unexpected error occured error in migration engine thread main panicked at commanderror generic code error queryerror queryerror mys qlerror error incorrect database name n nstack backtrace n backtrace backtrace trace n backtrace capture backtrace new unresolved n failure backtrace internal internalbacktrace new n failure backtrace backtrace as core default default default n prisma query connector mysql error impl core conve rt from for prisma query error error from n prisma query connector metrics query n query raw n sql migration connector migration database mysql as sql migration connector migration database migrationdatabase query raw n initialize n migration e ngine migration engine migrationengine new n migration engine api rpc rpcapi new n migration engine api rpc rpcapi new async n migration engine main n std rt lang start closure n std panicking try do call n rust maybe catch panic n std rt lang start internal n main n n nstack backtrace n backtrace backtrace trace n backtrace capture backtrace new unresolved n failure backtrace internal int ernalbacktrace new n default n migration connector
1
275,243
30,219,150,649
IssuesEvent
2023-07-05 17:53:48
temporalio/samples-go
https://api.github.com/repos/temporalio/samples-go
closed
go.temporal.io/sdk/contrib/tools/workflowcheck-v0.0.0-20230612164027-11c2cb9e7d2d: 3 vulnerabilities (highest severity is: 6.1) - autoclosed
Mend: dependency security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>go.temporal.io/sdk/contrib/tools/workflowcheck-v0.0.0-20230612164027-11c2cb9e7d2d</b></p></summary> <p></p> <p> <p>Found in HEAD commit: <a href="https://github.com/temporalio/samples-go/commit/d54b5f7bd51ab2c52ba1eb744b6f87ab9bc5596f">d54b5f7bd51ab2c52ba1eb744b6f87ab9bc5596f</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (go.temporal.io/sdk/contrib/tools/workflowcheck-v0.0.0 version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2020-11022](https://www.mend.io/vulnerability-database/CVE-2020-11022) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 6.1 | golang.org/x/tools-v0.9.3 | Transitive | N/A* | &#10060; | | [CVE-2015-9251](https://www.mend.io/vulnerability-database/CVE-2015-9251) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 6.1 | golang.org/x/tools-v0.9.3 | Transitive | N/A* | &#10060; | | [CVE-2012-6708](https://www.mend.io/vulnerability-database/CVE-2012-6708) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 6.1 | golang.org/x/tools-v0.9.3 | Transitive | N/A* | &#10060; | <p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the "Details" section below to see if there is a version of transitive dependency where vulnerability is fixed.</p> ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2020-11022</summary> ### Vulnerable Library - <b>golang.org/x/tools-v0.9.3</b></p> <p></p> <p>Library home page: <a href="https://proxy.golang.org/golang.org/x/tools/@v/v0.9.3.zip">https://proxy.golang.org/golang.org/x/tools/@v/v0.9.3.zip</a></p> <p> Dependency Hierarchy: - go.temporal.io/sdk/contrib/tools/workflowcheck-v0.0.0-20230612164027-11c2cb9e7d2d (Root Library) - :x: **golang.org/x/tools-v0.9.3** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/temporalio/samples-go/commit/d54b5f7bd51ab2c52ba1eb744b6f87ab9bc5596f">d54b5f7bd51ab2c52ba1eb744b6f87ab9bc5596f</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-11022>CVE-2020-11022</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11022">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11022</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jQuery - 3.5.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2015-9251</summary> ### Vulnerable Library - <b>golang.org/x/tools-v0.9.3</b></p> <p></p> <p>Library home page: <a href="https://proxy.golang.org/golang.org/x/tools/@v/v0.9.3.zip">https://proxy.golang.org/golang.org/x/tools/@v/v0.9.3.zip</a></p> <p> Dependency Hierarchy: - go.temporal.io/sdk/contrib/tools/workflowcheck-v0.0.0-20230612164027-11c2cb9e7d2d (Root Library) - :x: **golang.org/x/tools-v0.9.3** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/temporalio/samples-go/commit/d54b5f7bd51ab2c52ba1eb744b6f87ab9bc5596f">d54b5f7bd51ab2c52ba1eb744b6f87ab9bc5596f</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed. <p>Publish Date: 2018-01-18 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2015-9251>CVE-2015-9251</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p> <p>Release Date: 2018-01-18</p> <p>Fix Resolution: jQuery - 3.0.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2012-6708</summary> ### Vulnerable Library - <b>golang.org/x/tools-v0.9.3</b></p> <p></p> <p>Library home page: <a href="https://proxy.golang.org/golang.org/x/tools/@v/v0.9.3.zip">https://proxy.golang.org/golang.org/x/tools/@v/v0.9.3.zip</a></p> <p> Dependency Hierarchy: - go.temporal.io/sdk/contrib/tools/workflowcheck-v0.0.0-20230612164027-11c2cb9e7d2d (Root Library) - :x: **golang.org/x/tools-v0.9.3** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/temporalio/samples-go/commit/d54b5f7bd51ab2c52ba1eb744b6f87ab9bc5596f">d54b5f7bd51ab2c52ba1eb744b6f87ab9bc5596f</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> jQuery before 1.9.0 is vulnerable to Cross-site Scripting (XSS) attacks. The jQuery(strInput) function does not differentiate selectors from HTML in a reliable fashion. In vulnerable versions, jQuery determined whether the input was HTML by looking for the '<' character anywhere in the string, giving attackers more flexibility when attempting to construct a malicious payload. In fixed versions, jQuery only deems the input to be HTML if it explicitly starts with the '<' character, limiting exploitability only to attackers who can control the beginning of a string, which is far less common. <p>Publish Date: 2018-01-18 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2012-6708>CVE-2012-6708</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2012-6708">https://nvd.nist.gov/vuln/detail/CVE-2012-6708</a></p> <p>Release Date: 2018-01-18</p> <p>Fix Resolution: jQuery - v1.9.0</p> </p> <p></p> </details>
True
go.temporal.io/sdk/contrib/tools/workflowcheck-v0.0.0-20230612164027-11c2cb9e7d2d: 3 vulnerabilities (highest severity is: 6.1) - autoclosed - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>go.temporal.io/sdk/contrib/tools/workflowcheck-v0.0.0-20230612164027-11c2cb9e7d2d</b></p></summary> <p></p> <p> <p>Found in HEAD commit: <a href="https://github.com/temporalio/samples-go/commit/d54b5f7bd51ab2c52ba1eb744b6f87ab9bc5596f">d54b5f7bd51ab2c52ba1eb744b6f87ab9bc5596f</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (go.temporal.io/sdk/contrib/tools/workflowcheck-v0.0.0 version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2020-11022](https://www.mend.io/vulnerability-database/CVE-2020-11022) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 6.1 | golang.org/x/tools-v0.9.3 | Transitive | N/A* | &#10060; | | [CVE-2015-9251](https://www.mend.io/vulnerability-database/CVE-2015-9251) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 6.1 | golang.org/x/tools-v0.9.3 | Transitive | N/A* | &#10060; | | [CVE-2012-6708](https://www.mend.io/vulnerability-database/CVE-2012-6708) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 6.1 | golang.org/x/tools-v0.9.3 | Transitive | N/A* | &#10060; | <p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the "Details" section below to see if there is a version of transitive dependency where vulnerability is fixed.</p> ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2020-11022</summary> ### Vulnerable Library - <b>golang.org/x/tools-v0.9.3</b></p> <p></p> <p>Library home page: <a href="https://proxy.golang.org/golang.org/x/tools/@v/v0.9.3.zip">https://proxy.golang.org/golang.org/x/tools/@v/v0.9.3.zip</a></p> <p> Dependency Hierarchy: - go.temporal.io/sdk/contrib/tools/workflowcheck-v0.0.0-20230612164027-11c2cb9e7d2d (Root Library) - :x: **golang.org/x/tools-v0.9.3** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/temporalio/samples-go/commit/d54b5f7bd51ab2c52ba1eb744b6f87ab9bc5596f">d54b5f7bd51ab2c52ba1eb744b6f87ab9bc5596f</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-11022>CVE-2020-11022</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11022">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11022</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jQuery - 3.5.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2015-9251</summary> ### Vulnerable Library - <b>golang.org/x/tools-v0.9.3</b></p> <p></p> <p>Library home page: <a href="https://proxy.golang.org/golang.org/x/tools/@v/v0.9.3.zip">https://proxy.golang.org/golang.org/x/tools/@v/v0.9.3.zip</a></p> <p> Dependency Hierarchy: - go.temporal.io/sdk/contrib/tools/workflowcheck-v0.0.0-20230612164027-11c2cb9e7d2d (Root Library) - :x: **golang.org/x/tools-v0.9.3** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/temporalio/samples-go/commit/d54b5f7bd51ab2c52ba1eb744b6f87ab9bc5596f">d54b5f7bd51ab2c52ba1eb744b6f87ab9bc5596f</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed. <p>Publish Date: 2018-01-18 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2015-9251>CVE-2015-9251</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p> <p>Release Date: 2018-01-18</p> <p>Fix Resolution: jQuery - 3.0.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2012-6708</summary> ### Vulnerable Library - <b>golang.org/x/tools-v0.9.3</b></p> <p></p> <p>Library home page: <a href="https://proxy.golang.org/golang.org/x/tools/@v/v0.9.3.zip">https://proxy.golang.org/golang.org/x/tools/@v/v0.9.3.zip</a></p> <p> Dependency Hierarchy: - go.temporal.io/sdk/contrib/tools/workflowcheck-v0.0.0-20230612164027-11c2cb9e7d2d (Root Library) - :x: **golang.org/x/tools-v0.9.3** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/temporalio/samples-go/commit/d54b5f7bd51ab2c52ba1eb744b6f87ab9bc5596f">d54b5f7bd51ab2c52ba1eb744b6f87ab9bc5596f</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> jQuery before 1.9.0 is vulnerable to Cross-site Scripting (XSS) attacks. The jQuery(strInput) function does not differentiate selectors from HTML in a reliable fashion. In vulnerable versions, jQuery determined whether the input was HTML by looking for the '<' character anywhere in the string, giving attackers more flexibility when attempting to construct a malicious payload. In fixed versions, jQuery only deems the input to be HTML if it explicitly starts with the '<' character, limiting exploitability only to attackers who can control the beginning of a string, which is far less common. <p>Publish Date: 2018-01-18 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2012-6708>CVE-2012-6708</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2012-6708">https://nvd.nist.gov/vuln/detail/CVE-2012-6708</a></p> <p>Release Date: 2018-01-18</p> <p>Fix Resolution: jQuery - v1.9.0</p> </p> <p></p> </details>
non_process
go temporal io sdk contrib tools workflowcheck vulnerabilities highest severity is autoclosed vulnerable library go temporal io sdk contrib tools workflowcheck found in head commit a href vulnerabilities cve severity cvss dependency type fixed in go temporal io sdk contrib tools workflowcheck version remediation available medium golang org x tools transitive n a medium golang org x tools transitive n a medium golang org x tools transitive n a for some transitive vulnerabilities there is no version of direct dependency with a fix check the details section below to see if there is a version of transitive dependency where vulnerability is fixed details cve vulnerable library golang org x tools library home page a href dependency hierarchy go temporal io sdk contrib tools workflowcheck root library x golang org x tools vulnerable library found in head commit a href found in base branch main vulnerability details in jquery versions greater than or equal to and before passing html from untrusted sources even after sanitizing it to one of jquery s dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery cve vulnerable library golang org x tools library home page a href dependency hierarchy go temporal io sdk contrib tools workflowcheck root library x golang org x tools vulnerable library found in head commit a href found in base branch main vulnerability details jquery before is vulnerable to cross site scripting xss attacks when a cross domain ajax request is performed without the datatype option causing text javascript responses to be executed publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery cve vulnerable library golang org x tools library home page a href dependency hierarchy go temporal io sdk contrib tools workflowcheck root library x golang org x tools vulnerable library found in head commit a href found in base branch main vulnerability details jquery before is vulnerable to cross site scripting xss attacks the jquery strinput function does not differentiate selectors from html in a reliable fashion in vulnerable versions jquery determined whether the input was html by looking for the character anywhere in the string giving attackers more flexibility when attempting to construct a malicious payload in fixed versions jquery only deems the input to be html if it explicitly starts with the character limiting exploitability only to attackers who can control the beginning of a string which is far less common publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery
0
238,603
18,245,597,930
IssuesEvent
2021-10-01 17:57:14
uriahf/rtichoke
https://api.github.com/repos/uriahf/rtichoke
closed
Change naming convention, from "performance_table" to "performance_data"
documentation rtichoke function
"performance_data" is a better name than "performance_table" and it is usefull to distinguish between the data as an object and a rendered table. The function `create_performance_table()` should be renamed to `prepare_performance_data()`. The output should be tibble instead of data.frame.
1.0
Change naming convention, from "performance_table" to "performance_data" - "performance_data" is a better name than "performance_table" and it is usefull to distinguish between the data as an object and a rendered table. The function `create_performance_table()` should be renamed to `prepare_performance_data()`. The output should be tibble instead of data.frame.
non_process
change naming convention from performance table to performance data performance data is a better name than performance table and it is usefull to distinguish between the data as an object and a rendered table the function create performance table should be renamed to prepare performance data the output should be tibble instead of data frame
0
12,844
15,225,037,093
IssuesEvent
2021-02-18 06:33:53
googleapis/java-functions
https://api.github.com/repos/googleapis/java-functions
reopened
Dependency Dashboard
api: cloudfunctions type: process
This issue contains a list of Renovate updates and their statuses. ## Open These updates have all been created already. Click a checkbox below to force a retry/rebase of any. - [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-functions-1.x -->[chore(deps): update dependency com.google.cloud:google-cloud-functions to v1.0.6](../pull/134) --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
1.0
Dependency Dashboard - This issue contains a list of Renovate updates and their statuses. ## Open These updates have all been created already. Click a checkbox below to force a retry/rebase of any. - [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-functions-1.x -->[chore(deps): update dependency com.google.cloud:google-cloud-functions to v1.0.6](../pull/134) --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
process
dependency dashboard this issue contains a list of renovate updates and their statuses open these updates have all been created already click a checkbox below to force a retry rebase of any pull check this box to trigger a request for renovate to run again on this repository
1
21,905
30,353,041,507
IssuesEvent
2023-07-11 20:36:30
metabase/metabase
https://api.github.com/repos/metabase/metabase
closed
config.yml fails if password contains special characters
Type:Bug Priority:P2 Operation/Serialization .Team/QueryProcessor :hammer_and_wrench:
### Describe the bug Trying to initialize Metabase with the following config file: ``` version: 1 config: users: - first_name: Luiz last_name: Arakaki password: MetaPa$$123{{> email: admin@metabase.com ``` It fails with `clojure.lang.ExceptionInfo: Invalid query: found [[ or {{ with no matching ]] or }} {:type :invalid-query}` ### To Reproduce Use the config file above ### Expected behavior It should start Metabase and the admin must use that password ### Logs _No response_ ### Information about your Metabase installation ```JSON 46.3 ``` ### Severity P2 ### Additional context _No response_
1.0
config.yml fails if password contains special characters - ### Describe the bug Trying to initialize Metabase with the following config file: ``` version: 1 config: users: - first_name: Luiz last_name: Arakaki password: MetaPa$$123{{> email: admin@metabase.com ``` It fails with `clojure.lang.ExceptionInfo: Invalid query: found [[ or {{ with no matching ]] or }} {:type :invalid-query}` ### To Reproduce Use the config file above ### Expected behavior It should start Metabase and the admin must use that password ### Logs _No response_ ### Information about your Metabase installation ```JSON 46.3 ``` ### Severity P2 ### Additional context _No response_
process
config yml fails if password contains special characters describe the bug trying to initialize metabase with the following config file version config users first name luiz last name arakaki password metapa email admin metabase com it fails with clojure lang exceptioninfo invalid query found or type invalid query to reproduce use the config file above expected behavior it should start metabase and the admin must use that password logs no response information about your metabase installation json severity additional context no response
1
15,708
19,848,660,502
IssuesEvent
2022-01-21 09:47:37
ooi-data/CE07SHSM-SBD12-04-PCO2AA000-recovered_host-pco2a_a_dcl_instrument_water_recovered
https://api.github.com/repos/ooi-data/CE07SHSM-SBD12-04-PCO2AA000-recovered_host-pco2a_a_dcl_instrument_water_recovered
opened
🛑 Processing failed: ValueError
process
## Overview `ValueError` found in `processing_task` task during run ended on 2022-01-21T09:47:36.956819. ## Details Flow name: `CE07SHSM-SBD12-04-PCO2AA000-recovered_host-pco2a_a_dcl_instrument_water_recovered` Task name: `processing_task` Error type: `ValueError` Error message: not enough values to unpack (expected 3, got 0) <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing final_path = finalize_data_stream( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream append_to_zarr(mod_ds, final_store, enc, logger=logger) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr _append_zarr(store, mod_ds) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr existing_arr.append(var_data.values) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values return _as_array_or_item(self._data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item data = np.asarray(data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__ x = self.compute() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute (result,) = compute(self, traverse=False, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute results = schedule(dsk, keys, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get results = get_async( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async raise_exception(exc, tb) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise raise exc File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task result = _execute_task(task, data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task return func(*(_execute_task(a, cache) for a in args)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter c = np.asarray(c) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__ self._ensure_cached() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached self.array = NumpyIndexingAdapter(np.asarray(self.array)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__ return array[key.tuple] File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__ return self.get_basic_selection(selection, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection return self._get_basic_selection_nd(selection=selection, out=out, File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd return self._get_selection(indexer=indexer, out=out, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection lchunk_coords, lchunk_selection, lout_selection = zip(*indexer) ValueError: not enough values to unpack (expected 3, got 0) ``` </details>
1.0
🛑 Processing failed: ValueError - ## Overview `ValueError` found in `processing_task` task during run ended on 2022-01-21T09:47:36.956819. ## Details Flow name: `CE07SHSM-SBD12-04-PCO2AA000-recovered_host-pco2a_a_dcl_instrument_water_recovered` Task name: `processing_task` Error type: `ValueError` Error message: not enough values to unpack (expected 3, got 0) <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing final_path = finalize_data_stream( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream append_to_zarr(mod_ds, final_store, enc, logger=logger) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr _append_zarr(store, mod_ds) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr existing_arr.append(var_data.values) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values return _as_array_or_item(self._data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item data = np.asarray(data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__ x = self.compute() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute (result,) = compute(self, traverse=False, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute results = schedule(dsk, keys, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get results = get_async( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async raise_exception(exc, tb) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise raise exc File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task result = _execute_task(task, data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task return func(*(_execute_task(a, cache) for a in args)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter c = np.asarray(c) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__ self._ensure_cached() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached self.array = NumpyIndexingAdapter(np.asarray(self.array)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__ return array[key.tuple] File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__ return self.get_basic_selection(selection, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection return self._get_basic_selection_nd(selection=selection, out=out, File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd return self._get_selection(indexer=indexer, out=out, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection lchunk_coords, lchunk_selection, lout_selection = zip(*indexer) ValueError: not enough values to unpack (expected 3, got 0) ``` </details>
process
🛑 processing failed valueerror overview valueerror found in processing task task during run ended on details flow name recovered host a dcl instrument water recovered task name processing task error type valueerror error message not enough values to unpack expected got traceback traceback most recent call last file srv conda envs notebook lib site packages ooi harvester processor pipeline py line in processing final path finalize data stream file srv conda envs notebook lib site packages ooi harvester processor init py line in finalize data stream append to zarr mod ds final store enc logger logger file srv conda envs notebook lib site packages ooi harvester processor init py line in append to zarr append zarr store mod ds file srv conda envs notebook lib site packages ooi harvester processor utils py line in append zarr existing arr append var data values file srv conda envs notebook lib site packages xarray core variable py line in values return as array or item self data file srv conda envs notebook lib site packages xarray core variable py line in as array or item data np asarray data file srv conda envs notebook lib site packages dask array core py line in array x self compute file srv conda envs notebook lib site packages dask base py line in compute result compute self traverse false kwargs file srv conda envs notebook lib site packages dask base py line in compute results schedule dsk keys kwargs file srv conda envs notebook lib site packages dask threaded py line in get results get async file srv conda envs notebook lib site packages dask local py line in get async raise exception exc tb file srv conda envs notebook lib site packages dask local py line in reraise raise exc file srv conda envs notebook lib site packages dask local py line in execute task result execute task task data file srv conda envs notebook lib site packages dask core py line in execute task return func execute task a cache for a in args file srv conda envs notebook lib site packages dask array core py line in getter c np asarray c file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array self ensure cached file srv conda envs notebook lib site packages xarray core indexing py line in ensure cached self array numpyindexingadapter np asarray self array file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray backends zarr py line in getitem return array file srv conda envs notebook lib site packages zarr core py line in getitem return self get basic selection selection fields fields file srv conda envs notebook lib site packages zarr core py line in get basic selection return self get basic selection nd selection selection out out file srv conda envs notebook lib site packages zarr core py line in get basic selection nd return self get selection indexer indexer out out fields fields file srv conda envs notebook lib site packages zarr core py line in get selection lchunk coords lchunk selection lout selection zip indexer valueerror not enough values to unpack expected got
1
148,814
5,696,848,853
IssuesEvent
2017-04-16 16:00:29
mreishman/Log-Hog
https://api.github.com/repos/mreishman/Log-Hog
opened
Reset changes button next to save button
enhancement Priority - 3 - Medium
- [ ] Reset changes button next to save button (resets back to values from on page load)
1.0
Reset changes button next to save button - - [ ] Reset changes button next to save button (resets back to values from on page load)
non_process
reset changes button next to save button reset changes button next to save button resets back to values from on page load
0
2,382
5,187,201,391
IssuesEvent
2017-01-20 16:12:06
openvstorage/framework-alba-plugin
https://api.github.com/repos/openvstorage/framework-alba-plugin
closed
Failed removing disk /dev/sdb: 'NoneType' object has no attribute 'partitions'
priority_minor process_cantreproduce type_bug
Initialized an SSD with 2 ASD and a bit later removed it. Didn't claim the ASDs. Resulting error: Failed removing disk /dev/sdb: 'NoneType' object has no attribute 'partitions'
1.0
Failed removing disk /dev/sdb: 'NoneType' object has no attribute 'partitions' - Initialized an SSD with 2 ASD and a bit later removed it. Didn't claim the ASDs. Resulting error: Failed removing disk /dev/sdb: 'NoneType' object has no attribute 'partitions'
process
failed removing disk dev sdb nonetype object has no attribute partitions initialized an ssd with asd and a bit later removed it didn t claim the asds resulting error failed removing disk dev sdb nonetype object has no attribute partitions
1
13,839
3,363,004,739
IssuesEvent
2015-11-20 09:57:52
NativeScript/nativescript-cli
https://api.github.com/repos/NativeScript/nativescript-cli
closed
"tns test" fails with Unknown command 'dev-test|android'.
2 - Ready For Test bug critical unit testing
C:\Git>rm -rf TestApp C:\Git>tns create TestApp Project TestApp was successfully created C:\Git>tns test init --path TestApp ? Select testing framework: jasmine npm WARN package.json @ No description npm WARN package.json @ No repository field. npm WARN package.json @ No README data karma@0.13.15 TestApp\node_modules\karma ├── batch@0.5.3 ├── di@0.0.1 ├── rimraf@2.4.3 ├── graceful-fs@4.1.2 ├── mime@1.3.4 ├── colors@1.1.2 ├── http-proxy@1.12.0 (eventemitter3@1.1.1, requires-port@0.0.1) ├── minimatch@3.0.0 (brace-expansion@1.1.1) ├── glob@5.0.15 (path-is-absolute@1.0.0, inherits@2.0.1, inflight@1.0.4, once@1.3.2) ├── dom-serialize@2.2.1 (custom-event@1.0.0, void-elements@2.0.1, extend@3.0.0, ent@2.2.0) ├── optimist@0.6.1 (wordwrap@0.0.3, minimist@0.0.10) ├── connect@3.4.0 (utils-merge@1.0.0, parseurl@1.3.0, debug@2.2.0, finalhandler@0.4.0) ├── useragent@2.1.7 (lru-cache@2.2.4) ├── expand-braces@0.1.2 (array-unique@0.2.1, array-slice@0.2.3, braces@0.1.5) ├── bluebird@2.10.2 ├── body-parser@1.14.1 (bytes@2.1.0, content-type@1.0.1, depd@1.1.0, qs@5.1.0, http-errors@1.3.1, on-finished@2.3.0, raw-body@2.1.4, debug@2.2.0, iconv-lite@0.4.12, type-is@1.6.9) ├── source-map@0.5.3 ├── chokidar@1.2.0 (arrify@1.0.0, path-is-absolute@1.0.0, glob-parent@2.0.0, async-each@0.1.6, is-binary-path@1.0.1, is-glob@2.0.1, lodash.flatten@3.0.2, readdirp@2.0.0, anymatch@1.3.0) ├── log4js@0.6.27 (async@0.2.10, underscore@1.8.2, readable-stream@1.0.33, semver@4.3.6) ├── lodash@3.10.1 ├── core-js@1.2.6 └── socket.io@1.3.7 (debug@2.1.0, has-binary-data@0.1.3, engine.io@1.5.4, socket.io-adapter@0.3.1, socket.io-parser@2.2.4, socket.io-client@1.3.7) npm WARN peerDependencies The peer dependency jasmine-core@* included from karma-jasmine will no npm WARN peerDependencies longer be automatically installed to fulfill the peerDependency npm WARN peerDependencies in npm 3+. Your application will need to depend on it explicitly. jasmine-core@2.3.4 TestApp\node_modules\jasmine-core karma-jasmine@0.3.6 TestApp\node_modules\karma-jasmine karma-nativescript-launcher@0.2.5 TestApp\node_modules\karma-nativescript-launcher nativescript-unit-test-runner@0.2.8 TestApp\node_modules\nativescript-unit-test-runner Successfully installed plugin nativescript-unit-test-runner. Example test file created in app/tests/ Run your tests using the "$ tns test <platform>" command. C:\Git> C:\Git> C:\Git>tns test android --path TestApp 16 11 2015 17:20:33.594:INFO [karma]: Karma v0.13.15 server started at http://localhost:9876/ 16 11 2015 17:20:33.601:INFO [launcher]: Starting browser NativeScript Unit Test Runner Copying template files... spawn EBADF Unknown command 'dev-test|android'. Try '$ tns help' for a full list of supported commands. 16 11 2015 17:22:08.033:INFO [launcher]: NativeScript deployment completed with code 127
2.0
"tns test" fails with Unknown command 'dev-test|android'. - C:\Git>rm -rf TestApp C:\Git>tns create TestApp Project TestApp was successfully created C:\Git>tns test init --path TestApp ? Select testing framework: jasmine npm WARN package.json @ No description npm WARN package.json @ No repository field. npm WARN package.json @ No README data karma@0.13.15 TestApp\node_modules\karma ├── batch@0.5.3 ├── di@0.0.1 ├── rimraf@2.4.3 ├── graceful-fs@4.1.2 ├── mime@1.3.4 ├── colors@1.1.2 ├── http-proxy@1.12.0 (eventemitter3@1.1.1, requires-port@0.0.1) ├── minimatch@3.0.0 (brace-expansion@1.1.1) ├── glob@5.0.15 (path-is-absolute@1.0.0, inherits@2.0.1, inflight@1.0.4, once@1.3.2) ├── dom-serialize@2.2.1 (custom-event@1.0.0, void-elements@2.0.1, extend@3.0.0, ent@2.2.0) ├── optimist@0.6.1 (wordwrap@0.0.3, minimist@0.0.10) ├── connect@3.4.0 (utils-merge@1.0.0, parseurl@1.3.0, debug@2.2.0, finalhandler@0.4.0) ├── useragent@2.1.7 (lru-cache@2.2.4) ├── expand-braces@0.1.2 (array-unique@0.2.1, array-slice@0.2.3, braces@0.1.5) ├── bluebird@2.10.2 ├── body-parser@1.14.1 (bytes@2.1.0, content-type@1.0.1, depd@1.1.0, qs@5.1.0, http-errors@1.3.1, on-finished@2.3.0, raw-body@2.1.4, debug@2.2.0, iconv-lite@0.4.12, type-is@1.6.9) ├── source-map@0.5.3 ├── chokidar@1.2.0 (arrify@1.0.0, path-is-absolute@1.0.0, glob-parent@2.0.0, async-each@0.1.6, is-binary-path@1.0.1, is-glob@2.0.1, lodash.flatten@3.0.2, readdirp@2.0.0, anymatch@1.3.0) ├── log4js@0.6.27 (async@0.2.10, underscore@1.8.2, readable-stream@1.0.33, semver@4.3.6) ├── lodash@3.10.1 ├── core-js@1.2.6 └── socket.io@1.3.7 (debug@2.1.0, has-binary-data@0.1.3, engine.io@1.5.4, socket.io-adapter@0.3.1, socket.io-parser@2.2.4, socket.io-client@1.3.7) npm WARN peerDependencies The peer dependency jasmine-core@* included from karma-jasmine will no npm WARN peerDependencies longer be automatically installed to fulfill the peerDependency npm WARN peerDependencies in npm 3+. Your application will need to depend on it explicitly. jasmine-core@2.3.4 TestApp\node_modules\jasmine-core karma-jasmine@0.3.6 TestApp\node_modules\karma-jasmine karma-nativescript-launcher@0.2.5 TestApp\node_modules\karma-nativescript-launcher nativescript-unit-test-runner@0.2.8 TestApp\node_modules\nativescript-unit-test-runner Successfully installed plugin nativescript-unit-test-runner. Example test file created in app/tests/ Run your tests using the "$ tns test <platform>" command. C:\Git> C:\Git> C:\Git>tns test android --path TestApp 16 11 2015 17:20:33.594:INFO [karma]: Karma v0.13.15 server started at http://localhost:9876/ 16 11 2015 17:20:33.601:INFO [launcher]: Starting browser NativeScript Unit Test Runner Copying template files... spawn EBADF Unknown command 'dev-test|android'. Try '$ tns help' for a full list of supported commands. 16 11 2015 17:22:08.033:INFO [launcher]: NativeScript deployment completed with code 127
non_process
tns test fails with unknown command dev test android c git rm rf testapp c git tns create testapp project testapp was successfully created c git tns test init path testapp select testing framework jasmine npm warn package json no description npm warn package json no repository field npm warn package json no readme data karma testapp node modules karma ├── batch ├── di ├── rimraf ├── graceful fs ├── mime ├── colors ├── http proxy requires port ├── minimatch brace expansion ├── glob path is absolute inherits inflight once ├── dom serialize custom event void elements extend ent ├── optimist wordwrap minimist ├── connect utils merge parseurl debug finalhandler ├── useragent lru cache ├── expand braces array unique array slice braces ├── bluebird ├── body parser bytes content type depd qs http errors on finished raw body debug iconv lite type is ├── source map ├── chokidar arrify path is absolute glob parent async each is binary path is glob lodash flatten readdirp anymatch ├── async underscore readable stream semver ├── lodash ├── core js └── socket io debug has binary data engine io socket io adapter socket io parser socket io client npm warn peerdependencies the peer dependency jasmine core included from karma jasmine will no npm warn peerdependencies longer be automatically installed to fulfill the peerdependency npm warn peerdependencies in npm your application will need to depend on it explicitly jasmine core testapp node modules jasmine core karma jasmine testapp node modules karma jasmine karma nativescript launcher testapp node modules karma nativescript launcher nativescript unit test runner testapp node modules nativescript unit test runner successfully installed plugin nativescript unit test runner example test file created in app tests run your tests using the tns test command c git c git c git tns test android path testapp info karma server started at info starting browser nativescript unit test runner copying template files spawn ebadf unknown command dev test android try tns help for a full list of supported commands info nativescript deployment completed with code
0
4,873
11,776,362,344
IssuesEvent
2020-03-16 13:08:09
eclipse/omr
https://api.github.com/repos/eclipse/omr
opened
WIP: JitBuilder 2.0 proposal
architecture review pending comp:jitbuilder discussion enhancement
I would like to begin discussion for an enhancement to the JitBuilder API that is significant enough I have code named it "JitBuilder 2.0". For the last two years, I have been collecting feedback and experiences with the use of the JitBuilder API, and this issue is an attempt to start to crystallize all of that into a more concrete proposal. The process was kicked off at the OMR architecture meeting on March 13, which you can listen to here if you have an hour or so: https://youtu.be/ylJmi2xbCQc I wanted to get this issue created, so people could track the effort. Stay tuned for more information!
1.0
WIP: JitBuilder 2.0 proposal - I would like to begin discussion for an enhancement to the JitBuilder API that is significant enough I have code named it "JitBuilder 2.0". For the last two years, I have been collecting feedback and experiences with the use of the JitBuilder API, and this issue is an attempt to start to crystallize all of that into a more concrete proposal. The process was kicked off at the OMR architecture meeting on March 13, which you can listen to here if you have an hour or so: https://youtu.be/ylJmi2xbCQc I wanted to get this issue created, so people could track the effort. Stay tuned for more information!
non_process
wip jitbuilder proposal i would like to begin discussion for an enhancement to the jitbuilder api that is significant enough i have code named it jitbuilder for the last two years i have been collecting feedback and experiences with the use of the jitbuilder api and this issue is an attempt to start to crystallize all of that into a more concrete proposal the process was kicked off at the omr architecture meeting on march which you can listen to here if you have an hour or so i wanted to get this issue created so people could track the effort stay tuned for more information
0
112,038
24,231,168,113
IssuesEvent
2022-09-26 18:25:19
FerretDB/FerretDB
https://api.github.com/repos/FerretDB/FerretDB
closed
Fuzz `wire` package with recorded data
code/chore fuzz
When FerretDB is run with `task run` or similar commands, it stores recorded client sessions in the `records/` directory. We should use them in our `wire` package fuzzing. See `fuzzMessages` function in tests. We should read all files in this directory, read wire messages in each file and use `f.Add` to add them to the seed corpus one-by-one. No changes in CI configuration, etc.
1.0
Fuzz `wire` package with recorded data - When FerretDB is run with `task run` or similar commands, it stores recorded client sessions in the `records/` directory. We should use them in our `wire` package fuzzing. See `fuzzMessages` function in tests. We should read all files in this directory, read wire messages in each file and use `f.Add` to add them to the seed corpus one-by-one. No changes in CI configuration, etc.
non_process
fuzz wire package with recorded data when ferretdb is run with task run or similar commands it stores recorded client sessions in the records directory we should use them in our wire package fuzzing see fuzzmessages function in tests we should read all files in this directory read wire messages in each file and use f add to add them to the seed corpus one by one no changes in ci configuration etc
0
4,592
7,432,096,857
IssuesEvent
2018-03-25 21:08:19
brucemiller/LaTeXML
https://api.github.com/repos/brucemiller/LaTeXML
closed
Latexmlc (LaTeXML 0.8.2) epub generation fails on Windows 10
bug postprocessing
There are multiple issues with epub generation. Command line latexmlc --dest "%1.epub" "%1.tex" produces an epub with only mimetype file and META-INF folder containing container.xml. The same script when run through TeXWorks throws an error Can't locate loadable object for module XML::LibXML in @INC (@INC contains: C:\Strawberry\perl\site\lib D:/texlive/2016/tlpkg/tlperl/lib d:/texlive/2016/tlpkg/tlperl/site/lib d:/texlive/2016/tlpkg/tlperl/lib .) at C:\Strawberry\perl\site\lib/XML/LibXML.pm line 156. and does not generate any epub. On removing the path definition d:\texlive\2016\tlpkg\tlperl\bin from TeXWorks preferences, produces a new error Cwd.c: loadable library and perl binaries are mismatched (got handshake key 0AA80080, needed 0AF00080)
1.0
Latexmlc (LaTeXML 0.8.2) epub generation fails on Windows 10 - There are multiple issues with epub generation. Command line latexmlc --dest "%1.epub" "%1.tex" produces an epub with only mimetype file and META-INF folder containing container.xml. The same script when run through TeXWorks throws an error Can't locate loadable object for module XML::LibXML in @INC (@INC contains: C:\Strawberry\perl\site\lib D:/texlive/2016/tlpkg/tlperl/lib d:/texlive/2016/tlpkg/tlperl/site/lib d:/texlive/2016/tlpkg/tlperl/lib .) at C:\Strawberry\perl\site\lib/XML/LibXML.pm line 156. and does not generate any epub. On removing the path definition d:\texlive\2016\tlpkg\tlperl\bin from TeXWorks preferences, produces a new error Cwd.c: loadable library and perl binaries are mismatched (got handshake key 0AA80080, needed 0AF00080)
process
latexmlc latexml epub generation fails on windows there are multiple issues with epub generation command line latexmlc dest epub tex produces an epub with only mimetype file and meta inf folder containing container xml the same script when run through texworks throws an error can t locate loadable object for module xml libxml in inc inc contains c strawberry perl site lib d texlive tlpkg tlperl lib d texlive tlpkg tlperl site lib d texlive tlpkg tlperl lib at c strawberry perl site lib xml libxml pm line and does not generate any epub on removing the path definition d texlive tlpkg tlperl bin from texworks preferences produces a new error cwd c loadable library and perl binaries are mismatched got handshake key needed
1
11,145
13,957,692,756
IssuesEvent
2020-10-24 08:10:47
alexanderkotsev/geoportal
https://api.github.com/repos/alexanderkotsev/geoportal
opened
BE: harvesting
BE - Belgium Geoportal Harvesting process
Dear Angelo, I hope you are all right. I have a question about harvesting: I don&#39;t remember what we finally decided in Ispra but I noticed we haven&#39;t been harvested since January. Should I command harvesting by myself or could you do it by yourself ? Regards, Beno&icirc;t
1.0
BE: harvesting - Dear Angelo, I hope you are all right. I have a question about harvesting: I don&#39;t remember what we finally decided in Ispra but I noticed we haven&#39;t been harvested since January. Should I command harvesting by myself or could you do it by yourself ? Regards, Beno&icirc;t
process
be harvesting dear angelo i hope you are all right i have a question about harvesting i don t remember what we finally decided in ispra but i noticed we haven t been harvested since january should i command harvesting by myself or could you do it by yourself regards beno icirc t
1
11,567
14,440,502,290
IssuesEvent
2020-12-07 15:38:57
MicrosoftDocs/azure-devops-docs
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
reopened
Pipeline Trigger Issues
Pri1 devops-cicd-process/tech devops/prod doc-enhancement
In the section **YAML resource trigger issues support and traceability** it states there is a menu item in the pipeline definition page. Where on the page? Which pipeline definition, source or destination? Can we get a screenshot showing the menu item on the page? I've clicked on just about everything I can think of and not seeing it :( https://docs.microsoft.com/en-us/azure/devops/pipelines/process/resources?view=azure-devops&tabs=schema#yaml-resource-trigger-issues-support-and-traceability --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: ee4ec9d0-e0d5-4fb4-7c3e-b84abfa290c2 * Version Independent ID: 3e2b80d9-30e5-0c48-49f0-4fcdfedf5eee * Content: [Resources - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/resources?view=azure-devops&tabs=schema) * Content Source: [docs/pipelines/process/resources.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/resources.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
1.0
Pipeline Trigger Issues - In the section **YAML resource trigger issues support and traceability** it states there is a menu item in the pipeline definition page. Where on the page? Which pipeline definition, source or destination? Can we get a screenshot showing the menu item on the page? I've clicked on just about everything I can think of and not seeing it :( https://docs.microsoft.com/en-us/azure/devops/pipelines/process/resources?view=azure-devops&tabs=schema#yaml-resource-trigger-issues-support-and-traceability --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: ee4ec9d0-e0d5-4fb4-7c3e-b84abfa290c2 * Version Independent ID: 3e2b80d9-30e5-0c48-49f0-4fcdfedf5eee * Content: [Resources - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/resources?view=azure-devops&tabs=schema) * Content Source: [docs/pipelines/process/resources.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/resources.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
process
pipeline trigger issues in the section yaml resource trigger issues support and traceability it states there is a menu item in the pipeline definition page where on the page which pipeline definition source or destination can we get a screenshot showing the menu item on the page i ve clicked on just about everything i can think of and not seeing it document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
1
22,389
31,142,286,048
IssuesEvent
2023-08-16 01:44:16
cypress-io/cypress
https://api.github.com/repos/cypress-io/cypress
closed
Flaky test: iframe input retains focus when browser is out of focus [chrome]
OS: linux process: flaky test topic: flake ❄️ stage: flake stale
### Link to dashboard or CircleCI failure https://app.circleci.com/pipelines/github/cypress-io/cypress/41761/workflows/4f6901d8-6339-4309-b284-e03055666f1c/jobs/1732048/tests#failed-test-0 ### Link to failing test in GitHub https://github.com/cypress-io/cypress/blob/develop/packages/data-context/test/unit/sources/GitDataSource.spec.ts#L46 ### Analysis <img width="964" alt="Screen Shot 2022-08-12 at 11 23 50 AM" src="https://user-images.githubusercontent.com/26726429/184420695-b9cc4553-61ad-496a-a9b2-d8b4b4dc1d94.png"> ### Cypress Version 10.4.0 ### Other Search for this issue number in the codebase to find the test(s) skipped until this issue is fixed
1.0
Flaky test: iframe input retains focus when browser is out of focus [chrome] - ### Link to dashboard or CircleCI failure https://app.circleci.com/pipelines/github/cypress-io/cypress/41761/workflows/4f6901d8-6339-4309-b284-e03055666f1c/jobs/1732048/tests#failed-test-0 ### Link to failing test in GitHub https://github.com/cypress-io/cypress/blob/develop/packages/data-context/test/unit/sources/GitDataSource.spec.ts#L46 ### Analysis <img width="964" alt="Screen Shot 2022-08-12 at 11 23 50 AM" src="https://user-images.githubusercontent.com/26726429/184420695-b9cc4553-61ad-496a-a9b2-d8b4b4dc1d94.png"> ### Cypress Version 10.4.0 ### Other Search for this issue number in the codebase to find the test(s) skipped until this issue is fixed
process
flaky test iframe input retains focus when browser is out of focus link to dashboard or circleci failure link to failing test in github analysis img width alt screen shot at am src cypress version other search for this issue number in the codebase to find the test s skipped until this issue is fixed
1
12,527
14,968,753,301
IssuesEvent
2021-01-27 17:14:11
googleapis/python-bigquery
https://api.github.com/repos/googleapis/python-bigquery
opened
Run samples tests with Python 3.9, too
type: process
Now that `pyarrow` extra supports Python 3.9, BigQuery can support it, too. BigQuery samples should thus include Python 3.9 in their nox test sessions. Requires #462 (BigQuery with Python 3.9 support needs to be released first).
1.0
Run samples tests with Python 3.9, too - Now that `pyarrow` extra supports Python 3.9, BigQuery can support it, too. BigQuery samples should thus include Python 3.9 in their nox test sessions. Requires #462 (BigQuery with Python 3.9 support needs to be released first).
process
run samples tests with python too now that pyarrow extra supports python bigquery can support it too bigquery samples should thus include python in their nox test sessions requires bigquery with python support needs to be released first
1
7,972
11,156,967,382
IssuesEvent
2019-12-25 09:58:46
bisq-network/bisq
https://api.github.com/repos/bisq-network/bisq
closed
A Timeout occurred
an:investigation in:trade-process was:dropped
Attaching .log file as requested by the client. Error when taking offer to sell BTC for BSQ An error occurred when taking the offer. A timeout occurred. [TradeProtocolTimeoutreached-bisq.log](https://github.com/bisq-network/bisq/files/3542524/TradeProtocolTimeoutreached-bisq.log) Edit update: I got the timeout error attempting the same offer 3 times. I switched to a different offer and the trade confirmed successfully.
1.0
A Timeout occurred - Attaching .log file as requested by the client. Error when taking offer to sell BTC for BSQ An error occurred when taking the offer. A timeout occurred. [TradeProtocolTimeoutreached-bisq.log](https://github.com/bisq-network/bisq/files/3542524/TradeProtocolTimeoutreached-bisq.log) Edit update: I got the timeout error attempting the same offer 3 times. I switched to a different offer and the trade confirmed successfully.
process
a timeout occurred attaching log file as requested by the client error when taking offer to sell btc for bsq an error occurred when taking the offer a timeout occurred edit update i got the timeout error attempting the same offer times i switched to a different offer and the trade confirmed successfully
1
36,122
12,396,873,049
IssuesEvent
2020-05-20 21:23:49
rammatzkvosky/888
https://api.github.com/repos/rammatzkvosky/888
opened
CVE-2020-9547 (High) detected in jackson-databind-2.8.8.jar
security vulnerability
## CVE-2020-9547 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.8.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /tmp/ws-scm/888/pom.xml</p> <p>Path to vulnerable library: epository/com/fasterxml/jackson/core/jackson-databind/2.8.8/jackson-databind-2.8.8.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.8.8.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/rammatzkvosky/888/commit/26c91f41d9c3b8d120fec834c102a2d177b36807">26c91f41d9c3b8d120fec834c102a2d177b36807</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to com.ibatis.sqlmap.engine.transaction.jta.JtaTransactionConfig (aka ibatis-sqlmap). <p>Publish Date: 2020-03-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9547>CVE-2020-9547</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-9547">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-9547</a></p> <p>Release Date: 2020-03-02</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.10.3</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.8","isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.8.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.10.3"}],"vulnerabilityIdentifier":"CVE-2020-9547","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to com.ibatis.sqlmap.engine.transaction.jta.JtaTransactionConfig (aka ibatis-sqlmap).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9547","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-9547 (High) detected in jackson-databind-2.8.8.jar - ## CVE-2020-9547 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.8.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /tmp/ws-scm/888/pom.xml</p> <p>Path to vulnerable library: epository/com/fasterxml/jackson/core/jackson-databind/2.8.8/jackson-databind-2.8.8.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.8.8.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/rammatzkvosky/888/commit/26c91f41d9c3b8d120fec834c102a2d177b36807">26c91f41d9c3b8d120fec834c102a2d177b36807</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to com.ibatis.sqlmap.engine.transaction.jta.JtaTransactionConfig (aka ibatis-sqlmap). <p>Publish Date: 2020-03-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9547>CVE-2020-9547</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-9547">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-9547</a></p> <p>Release Date: 2020-03-02</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.10.3</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.8","isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.8.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.10.3"}],"vulnerabilityIdentifier":"CVE-2020-9547","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to com.ibatis.sqlmap.engine.transaction.jta.JtaTransactionConfig (aka ibatis-sqlmap).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9547","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_process
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file tmp ws scm pom xml path to vulnerable library epository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to com ibatis sqlmap engine transaction jta jtatransactionconfig aka ibatis sqlmap publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to com ibatis sqlmap engine transaction jta jtatransactionconfig aka ibatis sqlmap vulnerabilityurl
0
9,831
12,827,978,829
IssuesEvent
2020-07-06 19:35:16
googleapis/code-suggester
https://api.github.com/repos/googleapis/code-suggester
opened
Framework-core: handle when access token does not have permissions for a fork-PR
type: process
- [ ] Error message for when there is an access token that does not have forking permissions - [ ] Error message when the token does not have permissions to make a branch on that fork - [ ] Error message when the token does not have permissions to write to that branch on its fork - [ ] Error message when the token does not have permissions to create a PR from its fork to an upstream-repository ### Description An access token without permissions will be rejected ### Solution On the error messages received from the [GitHub API](https://developer.github.com/v3) gracefully fail.
1.0
Framework-core: handle when access token does not have permissions for a fork-PR - - [ ] Error message for when there is an access token that does not have forking permissions - [ ] Error message when the token does not have permissions to make a branch on that fork - [ ] Error message when the token does not have permissions to write to that branch on its fork - [ ] Error message when the token does not have permissions to create a PR from its fork to an upstream-repository ### Description An access token without permissions will be rejected ### Solution On the error messages received from the [GitHub API](https://developer.github.com/v3) gracefully fail.
process
framework core handle when access token does not have permissions for a fork pr error message for when there is an access token that does not have forking permissions error message when the token does not have permissions to make a branch on that fork error message when the token does not have permissions to write to that branch on its fork error message when the token does not have permissions to create a pr from its fork to an upstream repository description an access token without permissions will be rejected solution on the error messages received from the gracefully fail
1
18,146
24,186,818,559
IssuesEvent
2022-09-23 13:58:14
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
closed
NTR disruption of cellular component of another organism
New term request multi-species process
To group GO:0051673 membrane disruption in another organism GO:0044278 cell wall disruption in another organism' (these were under '[GO:0035821 modulation of process of another organism](http://amigo.geneontology.org/amigo/term/GO:0035821#display-lineage-tab)'
1.0
NTR disruption of cellular component of another organism - To group GO:0051673 membrane disruption in another organism GO:0044278 cell wall disruption in another organism' (these were under '[GO:0035821 modulation of process of another organism](http://amigo.geneontology.org/amigo/term/GO:0035821#display-lineage-tab)'
process
ntr disruption of cellular component of another organism to group go membrane disruption in another organism go cell wall disruption in another organism these were under
1
68,659
13,167,020,017
IssuesEvent
2020-08-11 09:32:41
mozilla-mobile/fenix
https://api.github.com/repos/mozilla-mobile/fenix
reopened
FNX3-15859 ⁃ [Bug] issue with QR code scanner if Fenix does not yet have camera permission
Feature: QRCode 🐞 bug
### Steps to reproduce 1. make sure in your Android app permissions that Fenix does not yet have permission to use the camera 2. use the QR code scanner of Fenix => A dialog appears to give Fenix the permission to use the camera 3. Give Fenix the permission ### Expected behavior After giving the permission the QR code scanner should be active. ### Actual behavior The "Scan" button has a visible "active" state but the QR code scanner is not active. Instead you have to press again the "Scan" button. ### Device information * Android device: HTC U11 / Android 8.0 * Fenix version: Fenix revision 080dadcb7e1cc0aa53c020595505f6f493e4a143
1.0
FNX3-15859 ⁃ [Bug] issue with QR code scanner if Fenix does not yet have camera permission - ### Steps to reproduce 1. make sure in your Android app permissions that Fenix does not yet have permission to use the camera 2. use the QR code scanner of Fenix => A dialog appears to give Fenix the permission to use the camera 3. Give Fenix the permission ### Expected behavior After giving the permission the QR code scanner should be active. ### Actual behavior The "Scan" button has a visible "active" state but the QR code scanner is not active. Instead you have to press again the "Scan" button. ### Device information * Android device: HTC U11 / Android 8.0 * Fenix version: Fenix revision 080dadcb7e1cc0aa53c020595505f6f493e4a143
non_process
⁃ issue with qr code scanner if fenix does not yet have camera permission steps to reproduce make sure in your android app permissions that fenix does not yet have permission to use the camera use the qr code scanner of fenix a dialog appears to give fenix the permission to use the camera give fenix the permission expected behavior after giving the permission the qr code scanner should be active actual behavior the scan button has a visible active state but the qr code scanner is not active instead you have to press again the scan button device information android device htc android fenix version fenix revision
0
12,528
14,969,402,883
IssuesEvent
2021-01-27 18:06:04
hashgraph/hedera-mirror-node
https://api.github.com/repos/hashgraph/hedera-mirror-node
closed
Integrate REST API build into Maven
P3 enhancement process rest
**Problem** Currently, if you want to run or build the entirety of the mirror node, you have to install multiple tools and run multiple commands. To make it easier for the community to run, we should reduce the number of external tools like npm that they have to install and make it more self contained. There should be a single command to run the build lifecycle. **Solution** The Maven wrapper should be the only command needed to build the mirror node: - Use frontend-maven-plugin to: - Install node and npm - npm install hedera-mirror-rest - npm test hedera-mirror-rest - npm package hedera-mirror-rest - npm install monitoring_api - npm test monitoring_api - Remove npm specific commands or jobs from CircleCI config - Update documentation to replace npm commands with maven, if applicable **Alternatives** Keep as is **Additional Context**
1.0
Integrate REST API build into Maven - **Problem** Currently, if you want to run or build the entirety of the mirror node, you have to install multiple tools and run multiple commands. To make it easier for the community to run, we should reduce the number of external tools like npm that they have to install and make it more self contained. There should be a single command to run the build lifecycle. **Solution** The Maven wrapper should be the only command needed to build the mirror node: - Use frontend-maven-plugin to: - Install node and npm - npm install hedera-mirror-rest - npm test hedera-mirror-rest - npm package hedera-mirror-rest - npm install monitoring_api - npm test monitoring_api - Remove npm specific commands or jobs from CircleCI config - Update documentation to replace npm commands with maven, if applicable **Alternatives** Keep as is **Additional Context**
process
integrate rest api build into maven problem currently if you want to run or build the entirety of the mirror node you have to install multiple tools and run multiple commands to make it easier for the community to run we should reduce the number of external tools like npm that they have to install and make it more self contained there should be a single command to run the build lifecycle solution the maven wrapper should be the only command needed to build the mirror node use frontend maven plugin to install node and npm npm install hedera mirror rest npm test hedera mirror rest npm package hedera mirror rest npm install monitoring api npm test monitoring api remove npm specific commands or jobs from circleci config update documentation to replace npm commands with maven if applicable alternatives keep as is additional context
1
19,173
25,281,889,538
IssuesEvent
2022-11-16 16:20:03
sillsdev/silnlp
https://api.github.com/repos/sillsdev/silnlp
opened
Error messages when files are missing.
enhancement good first issue pipeline 3: preprocess
I'm getting this error when trying to run an experiment on ClearML. It seems to occur when a required Scripture file is missing. It looks like Issue #85 ``` File "/root/.clearml/venvs-builds/3.8/task_repository/silnlp/.venv/lib/python3.8/site-packages/transformers/trainer.py", line 841, in get_train_dataloader raise ValueError("Trainer: training requires a train_dataset.") ValueError: Trainer: training requires a train_dataset. 2022-11-16 12:49:39 Process failed, exit code 1 ```
1.0
Error messages when files are missing. - I'm getting this error when trying to run an experiment on ClearML. It seems to occur when a required Scripture file is missing. It looks like Issue #85 ``` File "/root/.clearml/venvs-builds/3.8/task_repository/silnlp/.venv/lib/python3.8/site-packages/transformers/trainer.py", line 841, in get_train_dataloader raise ValueError("Trainer: training requires a train_dataset.") ValueError: Trainer: training requires a train_dataset. 2022-11-16 12:49:39 Process failed, exit code 1 ```
process
error messages when files are missing i m getting this error when trying to run an experiment on clearml it seems to occur when a required scripture file is missing it looks like issue file root clearml venvs builds task repository silnlp venv lib site packages transformers trainer py line in get train dataloader raise valueerror trainer training requires a train dataset valueerror trainer training requires a train dataset process failed exit code
1
11,924
14,703,943,695
IssuesEvent
2021-01-04 15:46:22
GetTerminus/terminus-oss
https://api.github.com/repos/GetTerminus/terminus-oss
closed
Add top-level changelog pointing to sub changelogs
Focus: community Goal: Process Improvement Type: chore
So that consumers can quickly see there are changelogs despite not having one for the repo as a whole. ie https://github.com/firebase/firebase-js-sdk/blob/master/CHANGELOG.md
1.0
Add top-level changelog pointing to sub changelogs - So that consumers can quickly see there are changelogs despite not having one for the repo as a whole. ie https://github.com/firebase/firebase-js-sdk/blob/master/CHANGELOG.md
process
add top level changelog pointing to sub changelogs so that consumers can quickly see there are changelogs despite not having one for the repo as a whole ie
1
147,351
11,782,836,962
IssuesEvent
2020-03-17 03:21:18
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
closed
network-based TorchHub tests contribute to CI flakiness
high priority topic: flaky-tests triage review
Perhaps this suite of tests can be run on nightly instead of on each `master` commit? See this [example failure on `master` branch](https://circleci.com/gh/pytorch/pytorch/4763382): ``` Mar 11 05:06:21 ====================================================================== Mar 11 05:06:21 ERROR [0.403s]: test_download_url_to_file (__main__.TestHub) Mar 11 05:06:21 ---------------------------------------------------------------------- Mar 11 05:06:21 Traceback (most recent call last): Mar 11 05:06:21 File "test_utils.py", line 547, in test_download_url_to_file Mar 11 05:06:21 hub.download_url_to_file(TORCHHUB_EXAMPLE_RELEASE_URL, temp_file, progress=False) Mar 11 05:06:21 File "/opt/conda/lib/python3.6/site-packages/torch/hub.py", line 394, in download_url_to_file Mar 11 05:06:21 u = urlopen(url) Mar 11 05:06:21 File "/opt/conda/lib/python3.6/urllib/request.py", line 223, in urlopen Mar 11 05:06:21 return opener.open(url, data, timeout) Mar 11 05:06:21 File "/opt/conda/lib/python3.6/urllib/request.py", line 532, in open Mar 11 05:06:21 response = meth(req, response) Mar 11 05:06:21 File "/opt/conda/lib/python3.6/urllib/request.py", line 642, in http_response Mar 11 05:06:21 'http', request, response, code, msg, hdrs) Mar 11 05:06:21 File "/opt/conda/lib/python3.6/urllib/request.py", line 570, in error Mar 11 05:06:21 return self._call_chain(*args) Mar 11 05:06:21 File "/opt/conda/lib/python3.6/urllib/request.py", line 504, in _call_chain Mar 11 05:06:21 result = func(*args) Mar 11 05:06:21 File "/opt/conda/lib/python3.6/urllib/request.py", line 650, in http_error_default Mar 11 05:06:21 raise HTTPError(req.full_url, code, msg, hdrs, fp) Mar 11 05:06:21 urllib.error.HTTPError: HTTP Error 401: UNAUTHORIZED ``` cc @ezyang @gchanan @zou3519 @ailzhang, @suo
1.0
network-based TorchHub tests contribute to CI flakiness - Perhaps this suite of tests can be run on nightly instead of on each `master` commit? See this [example failure on `master` branch](https://circleci.com/gh/pytorch/pytorch/4763382): ``` Mar 11 05:06:21 ====================================================================== Mar 11 05:06:21 ERROR [0.403s]: test_download_url_to_file (__main__.TestHub) Mar 11 05:06:21 ---------------------------------------------------------------------- Mar 11 05:06:21 Traceback (most recent call last): Mar 11 05:06:21 File "test_utils.py", line 547, in test_download_url_to_file Mar 11 05:06:21 hub.download_url_to_file(TORCHHUB_EXAMPLE_RELEASE_URL, temp_file, progress=False) Mar 11 05:06:21 File "/opt/conda/lib/python3.6/site-packages/torch/hub.py", line 394, in download_url_to_file Mar 11 05:06:21 u = urlopen(url) Mar 11 05:06:21 File "/opt/conda/lib/python3.6/urllib/request.py", line 223, in urlopen Mar 11 05:06:21 return opener.open(url, data, timeout) Mar 11 05:06:21 File "/opt/conda/lib/python3.6/urllib/request.py", line 532, in open Mar 11 05:06:21 response = meth(req, response) Mar 11 05:06:21 File "/opt/conda/lib/python3.6/urllib/request.py", line 642, in http_response Mar 11 05:06:21 'http', request, response, code, msg, hdrs) Mar 11 05:06:21 File "/opt/conda/lib/python3.6/urllib/request.py", line 570, in error Mar 11 05:06:21 return self._call_chain(*args) Mar 11 05:06:21 File "/opt/conda/lib/python3.6/urllib/request.py", line 504, in _call_chain Mar 11 05:06:21 result = func(*args) Mar 11 05:06:21 File "/opt/conda/lib/python3.6/urllib/request.py", line 650, in http_error_default Mar 11 05:06:21 raise HTTPError(req.full_url, code, msg, hdrs, fp) Mar 11 05:06:21 urllib.error.HTTPError: HTTP Error 401: UNAUTHORIZED ``` cc @ezyang @gchanan @zou3519 @ailzhang, @suo
non_process
network based torchhub tests contribute to ci flakiness perhaps this suite of tests can be run on nightly instead of on each master commit see this mar mar error test download url to file main testhub mar mar traceback most recent call last mar file test utils py line in test download url to file mar hub download url to file torchhub example release url temp file progress false mar file opt conda lib site packages torch hub py line in download url to file mar u urlopen url mar file opt conda lib urllib request py line in urlopen mar return opener open url data timeout mar file opt conda lib urllib request py line in open mar response meth req response mar file opt conda lib urllib request py line in http response mar http request response code msg hdrs mar file opt conda lib urllib request py line in error mar return self call chain args mar file opt conda lib urllib request py line in call chain mar result func args mar file opt conda lib urllib request py line in http error default mar raise httperror req full url code msg hdrs fp mar urllib error httperror http error unauthorized cc ezyang gchanan ailzhang suo
0
8,112
11,301,081,137
IssuesEvent
2020-01-17 14:55:07
Open-EO/openeo-api
https://api.github.com/repos/Open-EO/openeo-api
closed
UDF Usage
feedback required help wanted interoperability no solution yet processes udfs
I'm currently trying to catch up with the current UDF implementations for R and Python to change/improve the R implementation. While going through the issues on the UDFs and related issues in the openeo-processes I think we have major problems with users getting into UDFs. As I see it stated in the issues on Open-EO/openeo-processes#42 and Open-EO/openeo-udf#10 the information an user who wants to implement a successfully running UDF is lacking the following: 1. The dimension specification of the data / data cube at a given point during the execution of a process graph 2. How the data has to be in terms of the dimension specification, when the UDF result is injected back into the data stream of the process graph 3. How they can test their UDF <s>implementation</s> scripts (where: endpoint, with what: sample data, what is expected to be returned) 4. Wheter the data is chunked (they don't work with the whole dataset, e.g. a spatial tile as a raster time series) Therefore I would like to bring up the following ideas for discussion: - 1.1 Track the changes on the dimensions like running a metadata execution before running the process graph with actual data. This might be done by the client or the backend. This information must be accessible for the clients, because then we can create sample data. - 2.1. Implement some sort of constraint on the dimensions of the data cube import into process - 2.2. The UDF as such has to declare its returned data in terms of ist dimensions - 3.1. I'm not sure if this is solveable at the moment. We should have some testing endpoint for UDFs, but the problem is that the UDF runtime is mostly part of a specific back-end. I have no possible solution for 4 yet, since this will be mostly an optimization within the back-ends, when it comes to parallelization.
1.0
UDF Usage - I'm currently trying to catch up with the current UDF implementations for R and Python to change/improve the R implementation. While going through the issues on the UDFs and related issues in the openeo-processes I think we have major problems with users getting into UDFs. As I see it stated in the issues on Open-EO/openeo-processes#42 and Open-EO/openeo-udf#10 the information an user who wants to implement a successfully running UDF is lacking the following: 1. The dimension specification of the data / data cube at a given point during the execution of a process graph 2. How the data has to be in terms of the dimension specification, when the UDF result is injected back into the data stream of the process graph 3. How they can test their UDF <s>implementation</s> scripts (where: endpoint, with what: sample data, what is expected to be returned) 4. Wheter the data is chunked (they don't work with the whole dataset, e.g. a spatial tile as a raster time series) Therefore I would like to bring up the following ideas for discussion: - 1.1 Track the changes on the dimensions like running a metadata execution before running the process graph with actual data. This might be done by the client or the backend. This information must be accessible for the clients, because then we can create sample data. - 2.1. Implement some sort of constraint on the dimensions of the data cube import into process - 2.2. The UDF as such has to declare its returned data in terms of ist dimensions - 3.1. I'm not sure if this is solveable at the moment. We should have some testing endpoint for UDFs, but the problem is that the UDF runtime is mostly part of a specific back-end. I have no possible solution for 4 yet, since this will be mostly an optimization within the back-ends, when it comes to parallelization.
process
udf usage i m currently trying to catch up with the current udf implementations for r and python to change improve the r implementation while going through the issues on the udfs and related issues in the openeo processes i think we have major problems with users getting into udfs as i see it stated in the issues on open eo openeo processes and open eo openeo udf the information an user who wants to implement a successfully running udf is lacking the following the dimension specification of the data data cube at a given point during the execution of a process graph how the data has to be in terms of the dimension specification when the udf result is injected back into the data stream of the process graph how they can test their udf implementation scripts where endpoint with what sample data what is expected to be returned wheter the data is chunked they don t work with the whole dataset e g a spatial tile as a raster time series therefore i would like to bring up the following ideas for discussion track the changes on the dimensions like running a metadata execution before running the process graph with actual data this might be done by the client or the backend this information must be accessible for the clients because then we can create sample data implement some sort of constraint on the dimensions of the data cube import into process the udf as such has to declare its returned data in terms of ist dimensions i m not sure if this is solveable at the moment we should have some testing endpoint for udfs but the problem is that the udf runtime is mostly part of a specific back end i have no possible solution for yet since this will be mostly an optimization within the back ends when it comes to parallelization
1