Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 7 112 | repo_url stringlengths 36 141 | action stringclasses 3 values | title stringlengths 1 744 | labels stringlengths 4 574 | body stringlengths 9 211k | index stringclasses 10 values | text_combine stringlengths 96 211k | label stringclasses 2 values | text stringlengths 96 188k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
17,715 | 23,616,802,795 | IssuesEvent | 2022-08-24 16:33:50 | celo-org/celo-monorepo | https://api.github.com/repos/celo-org/celo-monorepo | closed | Draft Core-Contracts 8 release notes. | release-process Component: Identity ASv2 | Draft release notes for [Core Contracts Release 8](https://github.com/celo-org/celo-monorepo/releases/tag/core-contracts.v8.pre-audit) (pre-release). In particular the following sections:
- [x] "Key updates in this release" (blurb with description of changes)
- [x] "Specific Version Updates" (table with contract versions)
You can use [Core Contracts Release 7](https://github.com/celo-org/celo-monorepo/releases/tag/core-contracts.v7.post-audit) as a template.
I (@0xarthurxyz) will add the Hacken.io audit report to the "Audits" section when it's published in the next days. | 1.0 | Draft Core-Contracts 8 release notes. - Draft release notes for [Core Contracts Release 8](https://github.com/celo-org/celo-monorepo/releases/tag/core-contracts.v8.pre-audit) (pre-release). In particular the following sections:
- [x] "Key updates in this release" (blurb with description of changes)
- [x] "Specific Version Updates" (table with contract versions)
You can use [Core Contracts Release 7](https://github.com/celo-org/celo-monorepo/releases/tag/core-contracts.v7.post-audit) as a template.
I (@0xarthurxyz) will add the Hacken.io audit report to the "Audits" section when it's published in the next days. | process | draft core contracts release notes draft release notes for pre release in particular the following sections key updates in this release blurb with description of changes specific version updates table with contract versions you can use as a template i will add the hacken io audit report to the audits section when it s published in the next days | 1 |
110,875 | 16,995,012,752 | IssuesEvent | 2021-07-01 04:41:18 | avallete/yt-playlists-delete-enhancer | https://api.github.com/repos/avallete/yt-playlists-delete-enhancer | closed | CVE-2018-20821 (Medium) detected in opennmsopennms-source-26.0.0-1, node-sass-4.14.1.tgz | no-issue-activity security vulnerability | ## CVE-2018-20821 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>opennmsopennms-source-26.0.0-1</b>, <b>node-sass-4.14.1.tgz</b></p></summary>
<p>
<details><summary><b>node-sass-4.14.1.tgz</b></p></summary>
<p>Wrapper around libsass</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz</a></p>
<p>Path to dependency file: yt-playlists-delete-enhancer/package.json</p>
<p>Path to vulnerable library: yt-playlists-delete-enhancer/node_modules/node-sass/package.json</p>
<p>
Dependency Hierarchy:
- userscripter-1.4.0.tgz (Root Library)
- :x: **node-sass-4.14.1.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/avallete/yt-playlists-delete-enhancer/commit/9c336b0fa3155406498ca56519999606da4494b5">9c336b0fa3155406498ca56519999606da4494b5</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The parsing component in LibSass through 3.5.5 allows attackers to cause a denial-of-service (uncontrolled recursion in Sass::Parser::parse_css_variable_value in parser.cpp).
<p>Publish Date: 2019-04-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20821>CVE-2018-20821</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20821">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20821</a></p>
<p>Release Date: 2019-04-23</p>
<p>Fix Resolution: LibSass - 3.6.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2018-20821 (Medium) detected in opennmsopennms-source-26.0.0-1, node-sass-4.14.1.tgz - ## CVE-2018-20821 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>opennmsopennms-source-26.0.0-1</b>, <b>node-sass-4.14.1.tgz</b></p></summary>
<p>
<details><summary><b>node-sass-4.14.1.tgz</b></p></summary>
<p>Wrapper around libsass</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz</a></p>
<p>Path to dependency file: yt-playlists-delete-enhancer/package.json</p>
<p>Path to vulnerable library: yt-playlists-delete-enhancer/node_modules/node-sass/package.json</p>
<p>
Dependency Hierarchy:
- userscripter-1.4.0.tgz (Root Library)
- :x: **node-sass-4.14.1.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/avallete/yt-playlists-delete-enhancer/commit/9c336b0fa3155406498ca56519999606da4494b5">9c336b0fa3155406498ca56519999606da4494b5</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The parsing component in LibSass through 3.5.5 allows attackers to cause a denial-of-service (uncontrolled recursion in Sass::Parser::parse_css_variable_value in parser.cpp).
<p>Publish Date: 2019-04-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20821>CVE-2018-20821</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20821">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20821</a></p>
<p>Release Date: 2019-04-23</p>
<p>Fix Resolution: LibSass - 3.6.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_process | cve medium detected in opennmsopennms source node sass tgz cve medium severity vulnerability vulnerable libraries opennmsopennms source node sass tgz node sass tgz wrapper around libsass library home page a href path to dependency file yt playlists delete enhancer package json path to vulnerable library yt playlists delete enhancer node modules node sass package json dependency hierarchy userscripter tgz root library x node sass tgz vulnerable library found in head commit a href found in base branch master vulnerability details the parsing component in libsass through allows attackers to cause a denial of service uncontrolled recursion in sass parser parse css variable value in parser cpp publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution libsass step up your open source security game with whitesource | 0 |
282,826 | 21,315,965,065 | IssuesEvent | 2022-04-16 09:23:44 | johnmcmonigle/pe | https://api.github.com/repos/johnmcmonigle/pe | opened | 'Add Ward' sequence diagram needs calling arrows touching the tops of activation bars, and returning arrows touching the bottoms | severity.Low type.DocumentationBug | 
Many of the calling and returning arrows point to the middle of activation bars
<!--session: 1650094551158-c98fb3d6-71d0-4d9b-987f-11636894934a-->
<!--Version: Web v3.4.2--> | 1.0 | 'Add Ward' sequence diagram needs calling arrows touching the tops of activation bars, and returning arrows touching the bottoms - 
Many of the calling and returning arrows point to the middle of activation bars
<!--session: 1650094551158-c98fb3d6-71d0-4d9b-987f-11636894934a-->
<!--Version: Web v3.4.2--> | non_process | add ward sequence diagram needs calling arrows touching the tops of activation bars and returning arrows touching the bottoms many of the calling and returning arrows point to the middle of activation bars | 0 |
93,905 | 15,946,437,611 | IssuesEvent | 2021-04-15 01:04:05 | jgeraigery/core | https://api.github.com/repos/jgeraigery/core | opened | CVE-2018-10237 (Medium) detected in guava-18.0.jar | security vulnerability | ## CVE-2018-10237 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>guava-18.0.jar</b></p></summary>
<p>Guava is a suite of core and expanded libraries that include
utility classes, google's collections, io classes, and much
much more.
Guava has only one code dependency - javax.annotation,
per the JSR-305 spec.</p>
<p>Library home page: <a href="http://code.google.com/p/guava-libraries">http://code.google.com/p/guava-libraries</a></p>
<p>Path to dependency file: core/nimbus-core/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/google/guava/guava/18.0/guava-18.0.jar,/home/wss-scanner/.m2/repository/com/google/guava/guava/18.0/guava-18.0.jar,/home/wss-scanner/.m2/repository/com/google/guava/guava/18.0/guava-18.0.jar</p>
<p>
Dependency Hierarchy:
- springfox-swagger2-2.4.0.jar (Root Library)
- :x: **guava-18.0.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Unbounded memory allocation in Google Guava 11.0 through 24.x before 24.1.1 allows remote attackers to conduct denial of service attacks against servers that depend on this library and deserialize attacker-provided data, because the AtomicDoubleArray class (when serialized with Java serialization) and the CompoundOrdering class (when serialized with GWT serialization) perform eager allocation without appropriate checks on what a client has sent and whether the data size is reasonable.
<p>Publish Date: 2018-04-26
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-10237>CVE-2018-10237</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-10237">https://nvd.nist.gov/vuln/detail/CVE-2018-10237</a></p>
<p>Release Date: 2018-04-26</p>
<p>Fix Resolution: 24.1.1-jre, 24.1.1-android</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.google.guava","packageName":"guava","packageVersion":"18.0","packageFilePaths":["/nimbus-core/pom.xml","/nimbus-entity-dsl/pom.xml","/nimbus-test/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"io.springfox:springfox-swagger2:2.4.0;com.google.guava:guava:18.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"24.1.1-jre, 24.1.1-android"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2018-10237","vulnerabilityDetails":"Unbounded memory allocation in Google Guava 11.0 through 24.x before 24.1.1 allows remote attackers to conduct denial of service attacks against servers that depend on this library and deserialize attacker-provided data, because the AtomicDoubleArray class (when serialized with Java serialization) and the CompoundOrdering class (when serialized with GWT serialization) perform eager allocation without appropriate checks on what a client has sent and whether the data size is reasonable.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-10237","cvss3Severity":"medium","cvss3Score":"5.9","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | True | CVE-2018-10237 (Medium) detected in guava-18.0.jar - ## CVE-2018-10237 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>guava-18.0.jar</b></p></summary>
<p>Guava is a suite of core and expanded libraries that include
utility classes, google's collections, io classes, and much
much more.
Guava has only one code dependency - javax.annotation,
per the JSR-305 spec.</p>
<p>Library home page: <a href="http://code.google.com/p/guava-libraries">http://code.google.com/p/guava-libraries</a></p>
<p>Path to dependency file: core/nimbus-core/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/google/guava/guava/18.0/guava-18.0.jar,/home/wss-scanner/.m2/repository/com/google/guava/guava/18.0/guava-18.0.jar,/home/wss-scanner/.m2/repository/com/google/guava/guava/18.0/guava-18.0.jar</p>
<p>
Dependency Hierarchy:
- springfox-swagger2-2.4.0.jar (Root Library)
- :x: **guava-18.0.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Unbounded memory allocation in Google Guava 11.0 through 24.x before 24.1.1 allows remote attackers to conduct denial of service attacks against servers that depend on this library and deserialize attacker-provided data, because the AtomicDoubleArray class (when serialized with Java serialization) and the CompoundOrdering class (when serialized with GWT serialization) perform eager allocation without appropriate checks on what a client has sent and whether the data size is reasonable.
<p>Publish Date: 2018-04-26
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-10237>CVE-2018-10237</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-10237">https://nvd.nist.gov/vuln/detail/CVE-2018-10237</a></p>
<p>Release Date: 2018-04-26</p>
<p>Fix Resolution: 24.1.1-jre, 24.1.1-android</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.google.guava","packageName":"guava","packageVersion":"18.0","packageFilePaths":["/nimbus-core/pom.xml","/nimbus-entity-dsl/pom.xml","/nimbus-test/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"io.springfox:springfox-swagger2:2.4.0;com.google.guava:guava:18.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"24.1.1-jre, 24.1.1-android"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2018-10237","vulnerabilityDetails":"Unbounded memory allocation in Google Guava 11.0 through 24.x before 24.1.1 allows remote attackers to conduct denial of service attacks against servers that depend on this library and deserialize attacker-provided data, because the AtomicDoubleArray class (when serialized with Java serialization) and the CompoundOrdering class (when serialized with GWT serialization) perform eager allocation without appropriate checks on what a client has sent and whether the data size is reasonable.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-10237","cvss3Severity":"medium","cvss3Score":"5.9","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | non_process | cve medium detected in guava jar cve medium severity vulnerability vulnerable library guava jar guava is a suite of core and expanded libraries that include utility classes google s collections io classes and much much more guava has only one code dependency javax annotation per the jsr spec library home page a href path to dependency file core nimbus core pom xml path to vulnerable library home wss scanner repository com google guava guava guava jar home wss scanner repository com google guava guava guava jar home wss scanner repository com google guava guava guava jar dependency hierarchy springfox jar root library x guava jar vulnerable library found in base branch master vulnerability details unbounded memory allocation in google guava through x before allows remote attackers to conduct denial of service attacks against servers that depend on this library and deserialize attacker provided data because the atomicdoublearray class when serialized with java serialization and the compoundordering class when serialized with gwt serialization perform eager allocation without appropriate checks on what a client has sent and whether the data size is reasonable publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jre android isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree io springfox springfox com google guava guava isminimumfixversionavailable true minimumfixversion jre android basebranches vulnerabilityidentifier cve vulnerabilitydetails unbounded memory allocation in google guava through x before allows remote attackers to conduct denial of service attacks against servers that depend on this library and deserialize attacker provided data because the atomicdoublearray class when serialized with java serialization and the compoundordering class when serialized with gwt serialization perform eager allocation without appropriate checks on what a client has sent and whether the data size is reasonable vulnerabilityurl | 0 |
11,693 | 3,218,596,110 | IssuesEvent | 2015-10-08 02:51:16 | eris-ltd/eris-cli | https://api.github.com/repos/eris-ltd/eris-cli | opened | tests should randomize ports | area/test suite | thus no need to stop all running containers when testing ... then start them back up again | 1.0 | tests should randomize ports - thus no need to stop all running containers when testing ... then start them back up again | non_process | tests should randomize ports thus no need to stop all running containers when testing then start them back up again | 0 |
52,868 | 27,808,822,399 | IssuesEvent | 2023-03-17 23:37:50 | enso-org/enso | https://api.github.com/repos/enso-org/enso | closed | Consider parallelization of ZIO's setup during startup | p-medium x-chore --low-performance -language-server | The current implementation of Language Server makes [heavy usage](https://github.com/enso-org/enso/blob/develop/engine/language-server/src/main/scala/org/enso/languageserver/boot/MainModule.scala#L107) of ZIO.
As per profiling data

it also takes a considerable amount of time to initialize.
Unfortunately the current implementation of Language Server integrates this 3rd party library rather heavily and it is not easy to remove it without significant refactorings.
But ZIO's runtime is also not needed until serving some requests which makes it a potential candidate for non-blocking initialization, potentially reducing the time spent in startup. | True | Consider parallelization of ZIO's setup during startup - The current implementation of Language Server makes [heavy usage](https://github.com/enso-org/enso/blob/develop/engine/language-server/src/main/scala/org/enso/languageserver/boot/MainModule.scala#L107) of ZIO.
As per profiling data

it also takes a considerable amount of time to initialize.
Unfortunately the current implementation of Language Server integrates this 3rd party library rather heavily and it is not easy to remove it without significant refactorings.
But ZIO's runtime is also not needed until serving some requests which makes it a potential candidate for non-blocking initialization, potentially reducing the time spent in startup. | non_process | consider parallelization of zio s setup during startup the current implementation of language server makes of zio as per profiling data it also takes a considerable amount of time to initialize unfortunately the current implementation of language server integrates this party library rather heavily and it is not easy to remove it without significant refactorings but zio s runtime is also not needed until serving some requests which makes it a potential candidate for non blocking initialization potentially reducing the time spent in startup | 0 |
10,254 | 13,108,751,954 | IssuesEvent | 2020-08-04 17:24:15 | googleapis/code-suggester | https://api.github.com/repos/googleapis/code-suggester | closed | Implement CLI for creating a forked pull request (PR) | enhancement type: process | - [x] Implement CLI for creating a forked pull request (PR)
### Description
The user should be able to create a PR in a fork from either a non-git or git directory.
For non-git directory specified: all of the files in the directory get added on top of the upstream repository
For a git directory specified: all of the un-tracked git changes are made into a PR, except for the `.gitignore` files.
### Parameters
The framework-core should be invoked through the command-line. Specifically, support the parameters for:
1. `--upstream-repo`: The upstream repository to base a forked repository off of. Alias: `repo`, `r`
2. `--upstream-owner`: "The owner of the upstream repository. Alias: `owner`, `o`
3. `--pr-description`: The description of the PR to be made. Alias: `description`, `d`
4. `--pr-title`: The description of the PR title to be made. Alias: `title`, `t`
5. `--branch-name`: The name of the working branch to apply changes to. Alias: `branch`, `b`
6. `--change.work-dir`: The location of non-git changes to apply on top of an upstream repository.
7. `--change.git-dir`: The location of any un-tracked changes that should be made into a PR. Files in the `.gitignore` are ignored
### Environment variables
- `ACCESS_TOKEN`: an account that has permissions to create a fork.
### Additional context
- automating a PR from un-tracked git changes is useful because it reduces the number of commands
- creating a PR from a non-git directory is useful because it allows a repository to be initialized with a template. For instance if you have a `LICENSE` directory hosted locally, you can specify the directory of that LICENSE file and upload that to the upstream repository
| 1.0 | Implement CLI for creating a forked pull request (PR) - - [x] Implement CLI for creating a forked pull request (PR)
### Description
The user should be able to create a PR in a fork from either a non-git or git directory.
For non-git directory specified: all of the files in the directory get added on top of the upstream repository
For a git directory specified: all of the un-tracked git changes are made into a PR, except for the `.gitignore` files.
### Parameters
The framework-core should be invoked through the command-line. Specifically, support the parameters for:
1. `--upstream-repo`: The upstream repository to base a forked repository off of. Alias: `repo`, `r`
2. `--upstream-owner`: "The owner of the upstream repository. Alias: `owner`, `o`
3. `--pr-description`: The description of the PR to be made. Alias: `description`, `d`
4. `--pr-title`: The description of the PR title to be made. Alias: `title`, `t`
5. `--branch-name`: The name of the working branch to apply changes to. Alias: `branch`, `b`
6. `--change.work-dir`: The location of non-git changes to apply on top of an upstream repository.
7. `--change.git-dir`: The location of any un-tracked changes that should be made into a PR. Files in the `.gitignore` are ignored
### Environment variables
- `ACCESS_TOKEN`: an account that has permissions to create a fork.
### Additional context
- automating a PR from un-tracked git changes is useful because it reduces the number of commands
- creating a PR from a non-git directory is useful because it allows a repository to be initialized with a template. For instance if you have a `LICENSE` directory hosted locally, you can specify the directory of that LICENSE file and upload that to the upstream repository
| process | implement cli for creating a forked pull request pr implement cli for creating a forked pull request pr description the user should be able to create a pr in a fork from either a non git or git directory for non git directory specified all of the files in the directory get added on top of the upstream repository for a git directory specified all of the un tracked git changes are made into a pr except for the gitignore files parameters the framework core should be invoked through the command line specifically support the parameters for upstream repo the upstream repository to base a forked repository off of alias repo r upstream owner the owner of the upstream repository alias owner o pr description the description of the pr to be made alias description d pr title the description of the pr title to be made alias title t branch name the name of the working branch to apply changes to alias branch b change work dir the location of non git changes to apply on top of an upstream repository change git dir the location of any un tracked changes that should be made into a pr files in the gitignore are ignored environment variables access token an account that has permissions to create a fork additional context automating a pr from un tracked git changes is useful because it reduces the number of commands creating a pr from a non git directory is useful because it allows a repository to be initialized with a template for instance if you have a license directory hosted locally you can specify the directory of that license file and upload that to the upstream repository | 1 |
6,233 | 9,180,956,179 | IssuesEvent | 2019-03-05 09:04:08 | kmycode/sangokukmy | https://api.github.com/repos/kmycode/sangokukmy | closed | 農民反乱 | enhancement func-oldkmy process-pending | 三国志NET KMY Versionでは、農民反乱を実装しようと考えています。ほぼ以前あったものと同じです。
異民族( #10 )が若干強力すぎるので、農民反乱は地味なままでもいいかなと思ってます。
## 趣旨・目的
密偵の謀略により、農民反乱が起きるようにする。戦争前、他国への侵攻作戦を考えるのと同時に、自国の防衛作戦を考慮する必要が生じ、戦争は開戦する前からすでに始まっているという、簡単に油断できない雰囲気を作る
## 発生条件
* 密偵が扇動を実行している
* 民忠がゼロである
* その都市に武将が誰もいない
## できること
* 密偵( #12 )を放たれ、毎ターン扇動を実行された都市では、非常に低い確率で農民反乱が発生する
* 農民反乱が発生した都市は、農業・商業・技術がすべてゼロになったうえで、城壁1000、守兵500の無所属都市となる
* 2~3分に1の確率で、密偵を放った武将の所属国に寝返る
* 無所属都市へは、戦争準備中の国だけでなく、無関係の国も侵攻できる。もちろん建国もできる
## 以前との違い
特になし(城壁、守兵くらい?) | 1.0 | 農民反乱 - 三国志NET KMY Versionでは、農民反乱を実装しようと考えています。ほぼ以前あったものと同じです。
異民族( #10 )が若干強力すぎるので、農民反乱は地味なままでもいいかなと思ってます。
## 趣旨・目的
密偵の謀略により、農民反乱が起きるようにする。戦争前、他国への侵攻作戦を考えるのと同時に、自国の防衛作戦を考慮する必要が生じ、戦争は開戦する前からすでに始まっているという、簡単に油断できない雰囲気を作る
## 発生条件
* 密偵が扇動を実行している
* 民忠がゼロである
* その都市に武将が誰もいない
## できること
* 密偵( #12 )を放たれ、毎ターン扇動を実行された都市では、非常に低い確率で農民反乱が発生する
* 農民反乱が発生した都市は、農業・商業・技術がすべてゼロになったうえで、城壁1000、守兵500の無所属都市となる
* 2~3分に1の確率で、密偵を放った武将の所属国に寝返る
* 無所属都市へは、戦争準備中の国だけでなく、無関係の国も侵攻できる。もちろん建国もできる
## 以前との違い
特になし(城壁、守兵くらい?) | process | 農民反乱 三国志net kmy versionでは、農民反乱を実装しようと考えています。ほぼ以前あったものと同じです。 異民族 が若干強力すぎるので、農民反乱は地味なままでもいいかなと思ってます。 趣旨・目的 密偵の謀略により、農民反乱が起きるようにする。戦争前、他国への侵攻作戦を考えるのと同時に、自国の防衛作戦を考慮する必要が生じ、戦争は開戦する前からすでに始まっているという、簡単に油断できない雰囲気を作る 発生条件 密偵が扇動を実行している 民忠がゼロである その都市に武将が誰もいない できること 密偵 を放たれ、毎ターン扇動を実行された都市では、非常に低い確率で農民反乱が発生する 農民反乱が発生した都市は、農業・商業・技術がすべてゼロになったうえで、 、 ~ 、密偵を放った武将の所属国に寝返る 無所属都市へは、戦争準備中の国だけでなく、無関係の国も侵攻できる。もちろん建国もできる 以前との違い 特になし(城壁、守兵くらい?) | 1 |
16,065 | 20,205,911,749 | IssuesEvent | 2022-02-11 20:20:05 | createwithrani/superlist | https://api.github.com/repos/createwithrani/superlist | opened | Have a Main and Develop branch | Process | With #29 adding automated deployment from GitHub to SVN, should we have a `main` branch that is `stable` and a `development` branch that is where we merge new features and prepare for new releases?
The advantage of such a setup is that we can then use [WordPress Plugin Readme/Assets Update action](https://github.com/marketplace/actions/wordpress-plugin-readme-assets-update) to update just readme files, assets, and the 'tested up to' frontmatter without affecting the rest of the plugin. | 1.0 | Have a Main and Develop branch - With #29 adding automated deployment from GitHub to SVN, should we have a `main` branch that is `stable` and a `development` branch that is where we merge new features and prepare for new releases?
The advantage of such a setup is that we can then use [WordPress Plugin Readme/Assets Update action](https://github.com/marketplace/actions/wordpress-plugin-readme-assets-update) to update just readme files, assets, and the 'tested up to' frontmatter without affecting the rest of the plugin. | process | have a main and develop branch with adding automated deployment from github to svn should we have a main branch that is stable and a development branch that is where we merge new features and prepare for new releases the advantage of such a setup is that we can then use to update just readme files assets and the tested up to frontmatter without affecting the rest of the plugin | 1 |
19,946 | 26,419,644,632 | IssuesEvent | 2023-01-13 19:05:36 | ORNL-AMO/AMO-Tools-Desktop | https://api.github.com/repos/ORNL-AMO/AMO-Tools-Desktop | opened | Higher Heating Value Calculators | Process Heating Intern To Do | Turn New Gas Fuel and New Solid/Liquid fuel modals into stand alone calcs | 1.0 | Higher Heating Value Calculators - Turn New Gas Fuel and New Solid/Liquid fuel modals into stand alone calcs | process | higher heating value calculators turn new gas fuel and new solid liquid fuel modals into stand alone calcs | 1 |
14,010 | 16,815,860,171 | IssuesEvent | 2021-06-17 07:17:14 | didi/mpx | https://api.github.com/repos/didi/mpx | closed | [Bug report] 无法集成 thread-loader | processing | ```
{
test: /\.mpx$/,
use: [
{
loader: 'thread-loader'
},
MpxWebpackPlugin.loader(currentMpxLoaderConf)
]
}
```
```
Module build failed (from ./node_modules/thread-loader/dist/cjs.js):
Thread Loader (Worker 0)
Cannot read property 'compiler' of undefined
at PoolWorker.fromErrorObj (/test/node_modules/thread-loader/dist/WorkerPool.js:346:12)
at /test/node_modules/thread-loader/dist/WorkerPool.js:219:29
at module.exports (/test/node_modules/@mpxjs/webpack-plugin/lib/utils/get-main-compilation.js:2:22)
at Object.module.exports (/test/node_modules/@mpxjs/webpack-plugin/lib/loader.js:23:27)
Build failed with errors.
``` | 1.0 | [Bug report] 无法集成 thread-loader - ```
{
test: /\.mpx$/,
use: [
{
loader: 'thread-loader'
},
MpxWebpackPlugin.loader(currentMpxLoaderConf)
]
}
```
```
Module build failed (from ./node_modules/thread-loader/dist/cjs.js):
Thread Loader (Worker 0)
Cannot read property 'compiler' of undefined
at PoolWorker.fromErrorObj (/test/node_modules/thread-loader/dist/WorkerPool.js:346:12)
at /test/node_modules/thread-loader/dist/WorkerPool.js:219:29
at module.exports (/test/node_modules/@mpxjs/webpack-plugin/lib/utils/get-main-compilation.js:2:22)
at Object.module.exports (/test/node_modules/@mpxjs/webpack-plugin/lib/loader.js:23:27)
Build failed with errors.
``` | process | 无法集成 thread loader test mpx use loader thread loader mpxwebpackplugin loader currentmpxloaderconf module build failed from node modules thread loader dist cjs js thread loader worker cannot read property compiler of undefined at poolworker fromerrorobj test node modules thread loader dist workerpool js at test node modules thread loader dist workerpool js at module exports test node modules mpxjs webpack plugin lib utils get main compilation js at object module exports test node modules mpxjs webpack plugin lib loader js build failed with errors | 1 |
20,681 | 27,352,924,987 | IssuesEvent | 2023-02-27 10:50:19 | camunda/issues | https://api.github.com/repos/camunda/issues | opened | Process Instance Modification | component:operate component:zeebe component:zeebe-process-automation public feature-parity version:8.1 riskClass:medium | ### Value Proposition Statement
Repair process instances that ended up in the wrong state by repeating or skipping steps. Move running flow nodes, add new or cancel existing ones in a process instance easily via our UI.
### User Problem
- During execution process instances can end up in the wrong state. Currently to solve this, the user has to cancel the instance and start a new instance from beginning, requiring to rerun all previous steps. This requires a lot of considerations during process design and implementation.
- Some use-cases include:
- Repairing process instances in which some steps have to be repeated or skipped
- Testing: Activities can be skipped or repeated for isolated testing of individual process segments
### User Stories
- As a Developer, I can modify one process instances (single instance) via API
- As a Developer, I can interact with API via Java Client
- As a Operations Engineer / Admin I can modify one process instance in Operate UI
- As a Product Manager I can track the usage of the feature (both UI and API) in Mixpanel
### Implementation Notes
- Operate Swagger is updated - file and UI
- New endpoints documented in Postman repository
- Mixpanel backend and frontend events trackable and discoverable
### Breakdown
<!-- Please link to sub-issues / -tasks contributing to respective epic phase or phase results where appropriate. -->
#### Discovery phase ##
<!-- Example: "Conduct customer interview with xyz" -->
#### Define phase ##
<!-- Example: link to "Define User-Journey Flow" or "Define target architecture" -->
Design Planning
* Reviewed by design: May 9, 2022
* Designer assigned: Yes
* Assignee: @nazlikaya
* [Design Brief](https://docs.google.com/document/d/12xVBJSjpCbG9f5gkHsxSA5mCdIFsFl889OQYQ6gJQcU/edit?usp=sharing)
* Research Brief - none
Design Deliverables
- [Prototype](https://github.com/camunda/product-design/issues/11) (Expected delivery date June 10) - Delivered: June 7
- [Specifications](https://github.com/camunda/product-design/issues/13) (Expected delivery date June 10) - Delivered: June 7
- [High fidelity design](https://github.com/camunda/product-design/issues/14) (Expected delivery date June 10)- Delivered: June 9
- [Handover Recording](https://camunda.zoom.us/rec/share/rTRBwu6JVfMIzPB6ZnxZObbrsV5T3CoXiZLCtoztGaxXNdIrbjRz7HqBKzzHn5-r.PhmtEN5Z5UsQvgz3?startTime=1654844508000) Passcode: B5vuqi*$
[Risk Assessment](https://github.com/camunda/product-hub/issues/322)
- Risk Class: Medium
- Action: Mitigate
#### Implement phase ##
- [x] https://github.com/camunda/zeebe/issues/9570
- [x] https://github.com/camunda/operate/issues/2955
#### Validate phase ##
- https://eu.mixpanel.com/s/qwRNk | 1.0 | Process Instance Modification - ### Value Proposition Statement
Repair process instances that ended up in the wrong state by repeating or skipping steps. Move running flow nodes, add new or cancel existing ones in a process instance easily via our UI.
### User Problem
- During execution process instances can end up in the wrong state. Currently to solve this, the user has to cancel the instance and start a new instance from beginning, requiring to rerun all previous steps. This requires a lot of considerations during process design and implementation.
- Some use-cases include:
- Repairing process instances in which some steps have to be repeated or skipped
- Testing: Activities can be skipped or repeated for isolated testing of individual process segments
### User Stories
- As a Developer, I can modify one process instances (single instance) via API
- As a Developer, I can interact with API via Java Client
- As a Operations Engineer / Admin I can modify one process instance in Operate UI
- As a Product Manager I can track the usage of the feature (both UI and API) in Mixpanel
### Implementation Notes
- Operate Swagger is updated - file and UI
- New endpoints documented in Postman repository
- Mixpanel backend and frontend events trackable and discoverable
### Breakdown
<!-- Please link to sub-issues / -tasks contributing to respective epic phase or phase results where appropriate. -->
#### Discovery phase ##
<!-- Example: "Conduct customer interview with xyz" -->
#### Define phase ##
<!-- Example: link to "Define User-Journey Flow" or "Define target architecture" -->
Design Planning
* Reviewed by design: May 9, 2022
* Designer assigned: Yes
* Assignee: @nazlikaya
* [Design Brief](https://docs.google.com/document/d/12xVBJSjpCbG9f5gkHsxSA5mCdIFsFl889OQYQ6gJQcU/edit?usp=sharing)
* Research Brief - none
Design Deliverables
- [Prototype](https://github.com/camunda/product-design/issues/11) (Expected delivery date June 10) - Delivered: June 7
- [Specifications](https://github.com/camunda/product-design/issues/13) (Expected delivery date June 10) - Delivered: June 7
- [High fidelity design](https://github.com/camunda/product-design/issues/14) (Expected delivery date June 10)- Delivered: June 9
- [Handover Recording](https://camunda.zoom.us/rec/share/rTRBwu6JVfMIzPB6ZnxZObbrsV5T3CoXiZLCtoztGaxXNdIrbjRz7HqBKzzHn5-r.PhmtEN5Z5UsQvgz3?startTime=1654844508000) Passcode: B5vuqi*$
[Risk Assessment](https://github.com/camunda/product-hub/issues/322)
- Risk Class: Medium
- Action: Mitigate
#### Implement phase ##
- [x] https://github.com/camunda/zeebe/issues/9570
- [x] https://github.com/camunda/operate/issues/2955
#### Validate phase ##
- https://eu.mixpanel.com/s/qwRNk | process | process instance modification value proposition statement repair process instances that ended up in the wrong state by repeating or skipping steps move running flow nodes add new or cancel existing ones in a process instance easily via our ui user problem during execution process instances can end up in the wrong state currently to solve this the user has to cancel the instance and start a new instance from beginning requiring to rerun all previous steps this requires a lot of considerations during process design and implementation some use cases include repairing process instances in which some steps have to be repeated or skipped testing activities can be skipped or repeated for isolated testing of individual process segments user stories as a developer i can modify one process instances single instance via api as a developer i can interact with api via java client as a operations engineer admin i can modify one process instance in operate ui as a product manager i can track the usage of the feature both ui and api in mixpanel implementation notes operate swagger is updated file and ui new endpoints documented in postman repository mixpanel backend and frontend events trackable and discoverable breakdown discovery phase define phase design planning reviewed by design may designer assigned yes assignee nazlikaya research brief none design deliverables expected delivery date june delivered june expected delivery date june delivered june expected delivery date june delivered june passcode risk class medium action mitigate implement phase validate phase | 1 |
29,085 | 23,707,318,039 | IssuesEvent | 2022-08-30 03:21:53 | UBCSailbot/.github | https://api.github.com/repos/UBCSailbot/.github | opened | Create PR template | infrastructure | ### Purpose
Guidelines for what to include in a PR
### Changes
- Write PR template file
- Look at the second resource to see if there is anything that we could do to improve our issues templates
### Resources
- https://embeddedartistry.com/blog/2017/08/04/a-github-pull-request-template-for-your-projects/
- https://github.com/stevemao/github-issue-templates
| 1.0 | Create PR template - ### Purpose
Guidelines for what to include in a PR
### Changes
- Write PR template file
- Look at the second resource to see if there is anything that we could do to improve our issues templates
### Resources
- https://embeddedartistry.com/blog/2017/08/04/a-github-pull-request-template-for-your-projects/
- https://github.com/stevemao/github-issue-templates
| non_process | create pr template purpose guidelines for what to include in a pr changes write pr template file look at the second resource to see if there is anything that we could do to improve our issues templates resources | 0 |
14,349 | 9,084,639,972 | IssuesEvent | 2019-02-18 04:47:03 | OctopusDeploy/Issues | https://api.github.com/repos/OctopusDeploy/Issues | opened | Cant tell what type of deployment target a given machine is | area/usability kind/enhancement | # Prerequisites
- [x ] I have searched [open](https://github.com/OctopusDeploy/Issues/issues) and [closed](https://github.com/OctopusDeploy/Issues/issues?utf8=%E2%9C%93&q=is%3Aissue+is%3Aclosed) issues to make sure it isn't already requested
- [x] I have written a descriptive issue title
- [x] I have linked the original source of this feature request
- [x] I have tagged the issue appropriately (area/*, kind/enhancement)
# The enhancement
It's currently not possible to see what kind of deployment target you're looking at when viewing a deployment target, unless you've memorized the icons.
We should show the deployment target type (`Polling Tentacle`, `Kubernetes Cluster`, etc) on the screen.
## Mockup

## Workarounds
Use the Infrastructure -> deployment targets screen to filter by target name and communication style until you find it
## Links
source: internally reported. | True | Cant tell what type of deployment target a given machine is - # Prerequisites
- [x ] I have searched [open](https://github.com/OctopusDeploy/Issues/issues) and [closed](https://github.com/OctopusDeploy/Issues/issues?utf8=%E2%9C%93&q=is%3Aissue+is%3Aclosed) issues to make sure it isn't already requested
- [x] I have written a descriptive issue title
- [x] I have linked the original source of this feature request
- [x] I have tagged the issue appropriately (area/*, kind/enhancement)
# The enhancement
It's currently not possible to see what kind of deployment target you're looking at when viewing a deployment target, unless you've memorized the icons.
We should show the deployment target type (`Polling Tentacle`, `Kubernetes Cluster`, etc) on the screen.
## Mockup

## Workarounds
Use the Infrastructure -> deployment targets screen to filter by target name and communication style until you find it
## Links
source: internally reported. | non_process | cant tell what type of deployment target a given machine is prerequisites i have searched and issues to make sure it isn t already requested i have written a descriptive issue title i have linked the original source of this feature request i have tagged the issue appropriately area kind enhancement the enhancement it s currently not possible to see what kind of deployment target you re looking at when viewing a deployment target unless you ve memorized the icons we should show the deployment target type polling tentacle kubernetes cluster etc on the screen mockup workarounds use the infrastructure deployment targets screen to filter by target name and communication style until you find it links source internally reported | 0 |
22,084 | 30,606,733,796 | IssuesEvent | 2023-07-23 04:57:52 | hashgraph/hedera-json-rpc-relay | https://api.github.com/repos/hashgraph/hedera-json-rpc-relay | opened | Separate http and ws metrics | enhancement P2 process | ### Problem
Currently the http and ws server both utilize the relay and other busines logic.
However, since the registries are created in the classes it's not clear the metrics are intended for the ws server when it's running
### Solution
Separate metrics flags or make sure there's a different layer for calls originating for one over the other
### Alternatives
_No response_ | 1.0 | Separate http and ws metrics - ### Problem
Currently the http and ws server both utilize the relay and other busines logic.
However, since the registries are created in the classes it's not clear the metrics are intended for the ws server when it's running
### Solution
Separate metrics flags or make sure there's a different layer for calls originating for one over the other
### Alternatives
_No response_ | process | separate http and ws metrics problem currently the http and ws server both utilize the relay and other busines logic however since the registries are created in the classes it s not clear the metrics are intended for the ws server when it s running solution separate metrics flags or make sure there s a different layer for calls originating for one over the other alternatives no response | 1 |
338,024 | 30,277,400,647 | IssuesEvent | 2023-07-07 21:09:19 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | pkg/sql/schemachanger/schemachanger_test: TestValidateMixedVersionElements_drop_column_with_partial_index failed | C-test-failure O-robot branch-master T-sql-foundations | pkg/sql/schemachanger/schemachanger_test.TestValidateMixedVersionElements_drop_column_with_partial_index [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Ci_TestsAwsLinuxArm64_UnitTests/10812861?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Ci_TestsAwsLinuxArm64_UnitTests/10812861?buildTab=artifacts#/) on master @ [1137aee0089fc2fe161bab0f1c899d2ddc997c90](https://github.com/cockroachdb/cockroach/commits/1137aee0089fc2fe161bab0f1c899d2ddc997c90):
```
=== RUN TestValidateMixedVersionElements_drop_column_with_partial_index
test_log_scope.go:167: test logs captured to: /artifacts/tmp/_tmp/6e404cfaa494ad6aa388bb776400b270/logTestValidateMixedVersionElements_drop_column_with_partial_index1152271934
test_log_scope.go:81: use -show-logs to present logs inline
datadriven.go:143:
/home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/7298/execroot/com_github_cockroachdb_cockroach/bazel-out/aarch64-fastbuild/bin/pkg/sql/schemachanger/schemachanger_test_/schemachanger_test.runfiles/com_github_cockroachdb_cockroach/pkg/sql/schemachanger/testdata/end_to_end/drop_column_with_partial_index/drop_column_with_partial_index.definition:4:
setup [0 args]
CREATE TABLE t (i INT PRIMARY KEY, j INT, INDEX (j) WHERE i >=0)
----
datadriven.go:143:
/home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/7298/execroot/com_github_cockroachdb_cockroach/bazel-out/aarch64-fastbuild/bin/pkg/sql/schemachanger/schemachanger_test_/schemachanger_test.runfiles/com_github_cockroachdb_cockroach/pkg/sql/schemachanger/testdata/end_to_end/drop_column_with_partial_index/drop_column_with_partial_index.definition:9:
stage-exec [2 args]
INSERT INTO t (i) VALUES($stageKey);
INSERT INTO t (i) VALUES($stageKey + 1);
----
datadriven.go:143:
/home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/7298/execroot/com_github_cockroachdb_cockroach/bazel-out/aarch64-fastbuild/bin/pkg/sql/schemachanger/schemachanger_test_/schemachanger_test.runfiles/com_github_cockroachdb_cockroach/pkg/sql/schemachanger/testdata/end_to_end/drop_column_with_partial_index/drop_column_with_partial_index.definition:16:
stage-query [2 args]
SELECT count(*)=$successfulStageCount*2 FROM t;
----
true
datadriven.go:143:
/home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/7298/execroot/com_github_cockroachdb_cockroach/bazel-out/aarch64-fastbuild/bin/pkg/sql/schemachanger/schemachanger_test_/schemachanger_test.runfiles/com_github_cockroachdb_cockroach/pkg/sql/schemachanger/testdata/end_to_end/drop_column_with_partial_index/drop_column_with_partial_index.definition:22:
stage-exec [2 args]
INSERT INTO t (i) VALUES($stageKey);
INSERT INTO t (i) VALUES($stageKey + 1);
----
datadriven.go:143:
/home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/7298/execroot/com_github_cockroachdb_cockroach/bazel-out/aarch64-fastbuild/bin/pkg/sql/schemachanger/schemachanger_test_/schemachanger_test.runfiles/com_github_cockroachdb_cockroach/pkg/sql/schemachanger/testdata/end_to_end/drop_column_with_partial_index/drop_column_with_partial_index.definition:29:
stage-query [2 args]
SELECT count(*)=$successfulStageCount*2 FROM t;
----
true
=== CONT TestValidateMixedVersionElements_drop_column_with_partial_index
panic.go:522: -- test log scope end --
test logs left over in: /artifacts/tmp/_tmp/6e404cfaa494ad6aa388bb776400b270/logTestValidateMixedVersionElements_drop_column_with_partial_index1152271934
--- FAIL: TestValidateMixedVersionElements_drop_column_with_partial_index (23.39s)
=== RUN TestValidateMixedVersionElements_drop_column_with_partial_index/pause_upgrade_and_resume_at_each_stage
--- FAIL: TestValidateMixedVersionElements_drop_column_with_partial_index/pause_upgrade_and_resume_at_each_stage (22.60s)
=== RUN TestValidateMixedVersionElements_drop_column_with_partial_index/pause_upgrade_and_resume_at_each_stage/PostCommitNonRevertiblePhase_3_of_3
cumulative.go:1724:
Error Trace: github.com/cockroachdb/cockroach/pkg/sql/schemachanger/sctest/cumulative.go:1724
GOROOT/src/runtime/asm_arm64.s:1172
Error: Received unexpected error:
dial tcp 127.0.0.1:40613: connect: connection refused
Test: TestValidateMixedVersionElements_drop_column_with_partial_index/pause_upgrade_and_resume_at_each_stage/PostCommitNonRevertiblePhase_3_of_3
--- FAIL: TestValidateMixedVersionElements_drop_column_with_partial_index/pause_upgrade_and_resume_at_each_stage/PostCommitNonRevertiblePhase_3_of_3 (4.76s)
```
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)
</p>
</details>
/cc @cockroachdb/sql-foundations
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestValidateMixedVersionElements_drop_column_with_partial_index.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-29530 | 1.0 | pkg/sql/schemachanger/schemachanger_test: TestValidateMixedVersionElements_drop_column_with_partial_index failed - pkg/sql/schemachanger/schemachanger_test.TestValidateMixedVersionElements_drop_column_with_partial_index [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Ci_TestsAwsLinuxArm64_UnitTests/10812861?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Ci_TestsAwsLinuxArm64_UnitTests/10812861?buildTab=artifacts#/) on master @ [1137aee0089fc2fe161bab0f1c899d2ddc997c90](https://github.com/cockroachdb/cockroach/commits/1137aee0089fc2fe161bab0f1c899d2ddc997c90):
```
=== RUN TestValidateMixedVersionElements_drop_column_with_partial_index
test_log_scope.go:167: test logs captured to: /artifacts/tmp/_tmp/6e404cfaa494ad6aa388bb776400b270/logTestValidateMixedVersionElements_drop_column_with_partial_index1152271934
test_log_scope.go:81: use -show-logs to present logs inline
datadriven.go:143:
/home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/7298/execroot/com_github_cockroachdb_cockroach/bazel-out/aarch64-fastbuild/bin/pkg/sql/schemachanger/schemachanger_test_/schemachanger_test.runfiles/com_github_cockroachdb_cockroach/pkg/sql/schemachanger/testdata/end_to_end/drop_column_with_partial_index/drop_column_with_partial_index.definition:4:
setup [0 args]
CREATE TABLE t (i INT PRIMARY KEY, j INT, INDEX (j) WHERE i >=0)
----
datadriven.go:143:
/home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/7298/execroot/com_github_cockroachdb_cockroach/bazel-out/aarch64-fastbuild/bin/pkg/sql/schemachanger/schemachanger_test_/schemachanger_test.runfiles/com_github_cockroachdb_cockroach/pkg/sql/schemachanger/testdata/end_to_end/drop_column_with_partial_index/drop_column_with_partial_index.definition:9:
stage-exec [2 args]
INSERT INTO t (i) VALUES($stageKey);
INSERT INTO t (i) VALUES($stageKey + 1);
----
datadriven.go:143:
/home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/7298/execroot/com_github_cockroachdb_cockroach/bazel-out/aarch64-fastbuild/bin/pkg/sql/schemachanger/schemachanger_test_/schemachanger_test.runfiles/com_github_cockroachdb_cockroach/pkg/sql/schemachanger/testdata/end_to_end/drop_column_with_partial_index/drop_column_with_partial_index.definition:16:
stage-query [2 args]
SELECT count(*)=$successfulStageCount*2 FROM t;
----
true
datadriven.go:143:
/home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/7298/execroot/com_github_cockroachdb_cockroach/bazel-out/aarch64-fastbuild/bin/pkg/sql/schemachanger/schemachanger_test_/schemachanger_test.runfiles/com_github_cockroachdb_cockroach/pkg/sql/schemachanger/testdata/end_to_end/drop_column_with_partial_index/drop_column_with_partial_index.definition:22:
stage-exec [2 args]
INSERT INTO t (i) VALUES($stageKey);
INSERT INTO t (i) VALUES($stageKey + 1);
----
datadriven.go:143:
/home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/7298/execroot/com_github_cockroachdb_cockroach/bazel-out/aarch64-fastbuild/bin/pkg/sql/schemachanger/schemachanger_test_/schemachanger_test.runfiles/com_github_cockroachdb_cockroach/pkg/sql/schemachanger/testdata/end_to_end/drop_column_with_partial_index/drop_column_with_partial_index.definition:29:
stage-query [2 args]
SELECT count(*)=$successfulStageCount*2 FROM t;
----
true
=== CONT TestValidateMixedVersionElements_drop_column_with_partial_index
panic.go:522: -- test log scope end --
test logs left over in: /artifacts/tmp/_tmp/6e404cfaa494ad6aa388bb776400b270/logTestValidateMixedVersionElements_drop_column_with_partial_index1152271934
--- FAIL: TestValidateMixedVersionElements_drop_column_with_partial_index (23.39s)
=== RUN TestValidateMixedVersionElements_drop_column_with_partial_index/pause_upgrade_and_resume_at_each_stage
--- FAIL: TestValidateMixedVersionElements_drop_column_with_partial_index/pause_upgrade_and_resume_at_each_stage (22.60s)
=== RUN TestValidateMixedVersionElements_drop_column_with_partial_index/pause_upgrade_and_resume_at_each_stage/PostCommitNonRevertiblePhase_3_of_3
cumulative.go:1724:
Error Trace: github.com/cockroachdb/cockroach/pkg/sql/schemachanger/sctest/cumulative.go:1724
GOROOT/src/runtime/asm_arm64.s:1172
Error: Received unexpected error:
dial tcp 127.0.0.1:40613: connect: connection refused
Test: TestValidateMixedVersionElements_drop_column_with_partial_index/pause_upgrade_and_resume_at_each_stage/PostCommitNonRevertiblePhase_3_of_3
--- FAIL: TestValidateMixedVersionElements_drop_column_with_partial_index/pause_upgrade_and_resume_at_each_stage/PostCommitNonRevertiblePhase_3_of_3 (4.76s)
```
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)
</p>
</details>
/cc @cockroachdb/sql-foundations
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestValidateMixedVersionElements_drop_column_with_partial_index.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-29530 | non_process | pkg sql schemachanger schemachanger test testvalidatemixedversionelements drop column with partial index failed pkg sql schemachanger schemachanger test testvalidatemixedversionelements drop column with partial index with on master run testvalidatemixedversionelements drop column with partial index test log scope go test logs captured to artifacts tmp tmp logtestvalidatemixedversionelements drop column with partial test log scope go use show logs to present logs inline datadriven go home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql schemachanger schemachanger test schemachanger test runfiles com github cockroachdb cockroach pkg sql schemachanger testdata end to end drop column with partial index drop column with partial index definition setup create table t i int primary key j int index j where i datadriven go home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql schemachanger schemachanger test schemachanger test runfiles com github cockroachdb cockroach pkg sql schemachanger testdata end to end drop column with partial index drop column with partial index definition stage exec insert into t i values stagekey insert into t i values stagekey datadriven go home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql schemachanger schemachanger test schemachanger test runfiles com github cockroachdb cockroach pkg sql schemachanger testdata end to end drop column with partial index drop column with partial index definition stage query select count successfulstagecount from t true datadriven go home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql schemachanger schemachanger test schemachanger test runfiles com github cockroachdb cockroach pkg sql schemachanger testdata end to end drop column with partial index drop column with partial index definition stage exec insert into t i values stagekey insert into t i values stagekey datadriven go home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql schemachanger schemachanger test schemachanger test runfiles com github cockroachdb cockroach pkg sql schemachanger testdata end to end drop column with partial index drop column with partial index definition stage query select count successfulstagecount from t true cont testvalidatemixedversionelements drop column with partial index panic go test log scope end test logs left over in artifacts tmp tmp logtestvalidatemixedversionelements drop column with partial fail testvalidatemixedversionelements drop column with partial index run testvalidatemixedversionelements drop column with partial index pause upgrade and resume at each stage fail testvalidatemixedversionelements drop column with partial index pause upgrade and resume at each stage run testvalidatemixedversionelements drop column with partial index pause upgrade and resume at each stage postcommitnonrevertiblephase of cumulative go error trace github com cockroachdb cockroach pkg sql schemachanger sctest cumulative go goroot src runtime asm s error received unexpected error dial tcp connect connection refused test testvalidatemixedversionelements drop column with partial index pause upgrade and resume at each stage postcommitnonrevertiblephase of fail testvalidatemixedversionelements drop column with partial index pause upgrade and resume at each stage postcommitnonrevertiblephase of help see also cc cockroachdb sql foundations jira issue crdb | 0 |
787,020 | 27,701,928,353 | IssuesEvent | 2023-03-14 08:40:57 | hoangnguyen92dn/survey-flutter-ic | https://api.github.com/repos/hoangnguyen92dn/survey-flutter-ic | closed | [UI] As a user, I can sign in with email and password | type: feature priority: high @0.1.0 epic: authentication | ## Why
Users must be authenticated before taking or viewing any surveys. Our surveys have access restrictions. Some surveys are targeted to a specific group of users only for a more accurate survey result.
The mobile app should allow users to authenticate with their email and password.
## Acceptance Criteria
- Display the Nimble logo.
- Display the blurred background.
- Display the Email text input.
- Set the placeholder text to Email.
- Display the Password text input.
- Set the placeholder text to Password.
- Mask the password input with the native-OS feature.
- Display the Login button.
- **DO NOT** display the Forgot Password button yet. We will work on this on another separate epic.
## Design

## Resource
https://www.figma.com/file/GjRPOjDyZ6f4EDL3wKarRK/Challenge---Mobile-App?node-id=31%3A195&t=2O4uXmma4d3QrhkU-4 | 1.0 | [UI] As a user, I can sign in with email and password - ## Why
Users must be authenticated before taking or viewing any surveys. Our surveys have access restrictions. Some surveys are targeted to a specific group of users only for a more accurate survey result.
The mobile app should allow users to authenticate with their email and password.
## Acceptance Criteria
- Display the Nimble logo.
- Display the blurred background.
- Display the Email text input.
- Set the placeholder text to Email.
- Display the Password text input.
- Set the placeholder text to Password.
- Mask the password input with the native-OS feature.
- Display the Login button.
- **DO NOT** display the Forgot Password button yet. We will work on this on another separate epic.
## Design

## Resource
https://www.figma.com/file/GjRPOjDyZ6f4EDL3wKarRK/Challenge---Mobile-App?node-id=31%3A195&t=2O4uXmma4d3QrhkU-4 | non_process | as a user i can sign in with email and password why users must be authenticated before taking or viewing any surveys our surveys have access restrictions some surveys are targeted to a specific group of users only for a more accurate survey result the mobile app should allow users to authenticate with their email and password acceptance criteria display the nimble logo display the blurred background display the email text input set the placeholder text to email display the password text input set the placeholder text to password mask the password input with the native os feature display the login button do not display the forgot password button yet we will work on this on another separate epic design resource | 0 |
118,245 | 15,262,598,455 | IssuesEvent | 2021-02-22 00:04:39 | PyTorchLightning/pytorch-lightning | https://api.github.com/repos/PyTorchLightning/pytorch-lightning | closed | on_{validation,test}_epoch_end functions should have an outputs parameter | API / design duplicate enhancement help wanted | ## 🚀 Feature
https://github.com/PyTorchLightning/pytorch-lightning/blob/3b0e4e0b2bc5b62bba09df5976e1460774ae7337/pytorch_lightning/core/hooks.py#L255
https://github.com/PyTorchLightning/pytorch-lightning/blob/3b0e4e0b2bc5b62bba09df5976e1460774ae7337/pytorch_lightning/core/hooks.py#L267
Should have an `outputs` parameter as
https://github.com/PyTorchLightning/pytorch-lightning/blob/3b0e4e0b2bc5b62bba09df5976e1460774ae7337/pytorch_lightning/core/hooks.py#L243 | 1.0 | on_{validation,test}_epoch_end functions should have an outputs parameter - ## 🚀 Feature
https://github.com/PyTorchLightning/pytorch-lightning/blob/3b0e4e0b2bc5b62bba09df5976e1460774ae7337/pytorch_lightning/core/hooks.py#L255
https://github.com/PyTorchLightning/pytorch-lightning/blob/3b0e4e0b2bc5b62bba09df5976e1460774ae7337/pytorch_lightning/core/hooks.py#L267
Should have an `outputs` parameter as
https://github.com/PyTorchLightning/pytorch-lightning/blob/3b0e4e0b2bc5b62bba09df5976e1460774ae7337/pytorch_lightning/core/hooks.py#L243 | non_process | on validation test epoch end functions should have an outputs parameter 🚀 feature should have an outputs parameter as | 0 |
21,709 | 30,209,012,197 | IssuesEvent | 2023-07-05 11:29:18 | camunda/issues | https://api.github.com/repos/camunda/issues | closed | Catch errors without errorCode | component:desktopModeler component:operate component:optimize component:webModeler component:zeebe-process-automation public kind:epic feature-parity version:8.2 |
### Value Proposition Statement
Error catch events without error codes enable users to model a specific response for a known escalation code, and a general response for unknown escalation codes.
### User Problem
Currently, I can only model Error catch events for known error codes. However, if a BPMN Error with an unknown error code is thrown, an incident is raised.
### User Stories
As a user, I can model a general response to any BPMN Error thrown in a process.
<sup>:robot: This issue is automatically synced from: [source](https://github.com/camunda/product-hub/issues/771)</sup>
<!-- copiedFromSourceIssue: https://github.com/camunda/product-hub/issues/771 --> | 1.0 | Catch errors without errorCode -
### Value Proposition Statement
Error catch events without error codes enable users to model a specific response for a known escalation code, and a general response for unknown escalation codes.
### User Problem
Currently, I can only model Error catch events for known error codes. However, if a BPMN Error with an unknown error code is thrown, an incident is raised.
### User Stories
As a user, I can model a general response to any BPMN Error thrown in a process.
<sup>:robot: This issue is automatically synced from: [source](https://github.com/camunda/product-hub/issues/771)</sup>
<!-- copiedFromSourceIssue: https://github.com/camunda/product-hub/issues/771 --> | process | catch errors without errorcode value proposition statement error catch events without error codes enable users to model a specific response for a known escalation code and a general response for unknown escalation codes user problem currently i can only model error catch events for known error codes however if a bpmn error with an unknown error code is thrown an incident is raised user stories as a user i can model a general response to any bpmn error thrown in a process robot this issue is automatically synced from | 1 |
206,677 | 15,767,814,368 | IssuesEvent | 2021-03-31 16:30:27 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | ccl/importccl: TestProtectedTimestampsDuringImportInto failed | C-test-failure O-robot branch-master | [(ccl/importccl).TestProtectedTimestampsDuringImportInto failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2833638&tab=buildLog) on [master@c6125c3c5f4e416382c19adfaebe6c2190977190](https://github.com/cockroachdb/cockroach/commits/c6125c3c5f4e416382c19adfaebe6c2190977190):
```
=== RUN TestProtectedTimestampsDuringImportInto
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logTestProtectedTimestampsDuringImportInto005524695
test_log_scope.go:74: use -show-logs to present logs inline
n1: found liveness
E210331 12:12:05.331931 550735 storage/cloudimpl/http_storage.go:195 [n1,job=645930425636913153,import-distsql-ingest] 1 HTTP:Req error: err=error response from server: 500 Internal Server Error "" (attempt 0)
E210331 12:12:05.365749 550586 jobs/adopt.go:260 [n1] 2 job 645930425636913153: adoption completed with error error response from server: 500 Internal Server Error ""
sql_runner.go:224: error executing 'SELECT * FROM foo': pq: relation "foo" is offline: importing
E210331 12:12:05.377353 549479 jobs/registry.go:696 [-] 3 error claiming jobs: could not query jobs table: claim-jobs: context canceled
E210331 12:12:05.379465 549479 jobs/registry.go:706 [-] 4 error processing claimed jobs: could not query for claimed jobs: select-running/get-claimed-jobs: context canceled
panic.go:617: -- test log scope end --
test logs left over in: /go/src/github.com/cockroachdb/cockroach/artifacts/logTestProtectedTimestampsDuringImportInto005524695
--- FAIL: TestProtectedTimestampsDuringImportInto (3.52s)
```
<details><summary>More</summary><p>
Parameters:
- GOFLAGS=-json
```
make stressrace TESTS=TestProtectedTimestampsDuringImportInto PKG=./pkg/ccl/importccl TESTTIMEOUT=5m STRESSFLAGS='-timeout 5m' 2>&1
```
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2ATestProtectedTimestampsDuringImportInto.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
| 1.0 | ccl/importccl: TestProtectedTimestampsDuringImportInto failed - [(ccl/importccl).TestProtectedTimestampsDuringImportInto failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2833638&tab=buildLog) on [master@c6125c3c5f4e416382c19adfaebe6c2190977190](https://github.com/cockroachdb/cockroach/commits/c6125c3c5f4e416382c19adfaebe6c2190977190):
```
=== RUN TestProtectedTimestampsDuringImportInto
test_log_scope.go:73: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logTestProtectedTimestampsDuringImportInto005524695
test_log_scope.go:74: use -show-logs to present logs inline
n1: found liveness
E210331 12:12:05.331931 550735 storage/cloudimpl/http_storage.go:195 [n1,job=645930425636913153,import-distsql-ingest] 1 HTTP:Req error: err=error response from server: 500 Internal Server Error "" (attempt 0)
E210331 12:12:05.365749 550586 jobs/adopt.go:260 [n1] 2 job 645930425636913153: adoption completed with error error response from server: 500 Internal Server Error ""
sql_runner.go:224: error executing 'SELECT * FROM foo': pq: relation "foo" is offline: importing
E210331 12:12:05.377353 549479 jobs/registry.go:696 [-] 3 error claiming jobs: could not query jobs table: claim-jobs: context canceled
E210331 12:12:05.379465 549479 jobs/registry.go:706 [-] 4 error processing claimed jobs: could not query for claimed jobs: select-running/get-claimed-jobs: context canceled
panic.go:617: -- test log scope end --
test logs left over in: /go/src/github.com/cockroachdb/cockroach/artifacts/logTestProtectedTimestampsDuringImportInto005524695
--- FAIL: TestProtectedTimestampsDuringImportInto (3.52s)
```
<details><summary>More</summary><p>
Parameters:
- GOFLAGS=-json
```
make stressrace TESTS=TestProtectedTimestampsDuringImportInto PKG=./pkg/ccl/importccl TESTTIMEOUT=5m STRESSFLAGS='-timeout 5m' 2>&1
```
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2ATestProtectedTimestampsDuringImportInto.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
| non_process | ccl importccl testprotectedtimestampsduringimportinto failed on run testprotectedtimestampsduringimportinto test log scope go test logs captured to go src github com cockroachdb cockroach artifacts test log scope go use show logs to present logs inline found liveness storage cloudimpl http storage go http req error err error response from server internal server error attempt jobs adopt go job adoption completed with error error response from server internal server error sql runner go error executing select from foo pq relation foo is offline importing jobs registry go error claiming jobs could not query jobs table claim jobs context canceled jobs registry go error processing claimed jobs could not query for claimed jobs select running get claimed jobs context canceled panic go test log scope end test logs left over in go src github com cockroachdb cockroach artifacts fail testprotectedtimestampsduringimportinto more parameters goflags json make stressrace tests testprotectedtimestampsduringimportinto pkg pkg ccl importccl testtimeout stressflags timeout powered by | 0 |
98,631 | 16,387,781,418 | IssuesEvent | 2021-05-17 12:47:06 | fitzinbox/Exomiser | https://api.github.com/repos/fitzinbox/Exomiser | opened | CVE-2020-36184 (High) detected in jackson-databind-2.9.8.jar | security vulnerability | ## CVE-2020-36184 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.8.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: Exomiser/exomiser-data-genome/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar,Exomiser/exomiser-cli/target/lib/jackson-databind-2.9.8.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar,canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.8.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/fitzinbox/Exomiser/commit/3a0ae5a0b72ae7a7e59a638af862c28aa80dcdf6">3a0ae5a0b72ae7a7e59a638af862c28aa80dcdf6</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.tomcat.dbcp.dbcp2.datasources.PerUserPoolDataSource.
<p>Publish Date: 2021-01-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36184>CVE-2020-36184</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2998">https://github.com/FasterXML/jackson-databind/issues/2998</a></p>
<p>Release Date: 2021-01-06</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-36184 (High) detected in jackson-databind-2.9.8.jar - ## CVE-2020-36184 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.8.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: Exomiser/exomiser-data-genome/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar,Exomiser/exomiser-cli/target/lib/jackson-databind-2.9.8.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar,canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.8.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/fitzinbox/Exomiser/commit/3a0ae5a0b72ae7a7e59a638af862c28aa80dcdf6">3a0ae5a0b72ae7a7e59a638af862c28aa80dcdf6</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.tomcat.dbcp.dbcp2.datasources.PerUserPoolDataSource.
<p>Publish Date: 2021-01-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36184>CVE-2020-36184</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2998">https://github.com/FasterXML/jackson-databind/issues/2998</a></p>
<p>Release Date: 2021-01-06</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_process | cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file exomiser exomiser data genome pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar exomiser exomiser cli target lib jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar canner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache tomcat dbcp datasources peruserpooldatasource publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind step up your open source security game with whitesource | 0 |
2,814 | 5,738,746,685 | IssuesEvent | 2017-04-23 07:58:13 | dita-ot/dita-ot | https://api.github.com/repos/dita-ot/dita-ot | closed | Keydef containing uplevels | bug P2 preprocess/keyref | In my DITA Map I define a key to an image:
`<keydef keys="arch_diagram" href="../../introduction/images/arch_diagram" format="png"/>`
With the keyref:
`<image keyref="arch_diagram"/>`
It's not expanded in 2.3.3 or 2.4.6 but is in 1.8.
Other keyrefs to keys without uplevels and all other images are impacted.
Also, I have the same issue as #2420, except the conref has also uplevels. | 1.0 | Keydef containing uplevels - In my DITA Map I define a key to an image:
`<keydef keys="arch_diagram" href="../../introduction/images/arch_diagram" format="png"/>`
With the keyref:
`<image keyref="arch_diagram"/>`
It's not expanded in 2.3.3 or 2.4.6 but is in 1.8.
Other keyrefs to keys without uplevels and all other images are impacted.
Also, I have the same issue as #2420, except the conref has also uplevels. | process | keydef containing uplevels in my dita map i define a key to an image with the keyref it s not expanded in or but is in other keyrefs to keys without uplevels and all other images are impacted also i have the same issue as except the conref has also uplevels | 1 |
35,208 | 30,841,392,211 | IssuesEvent | 2023-08-02 10:54:00 | woowacourse-teams/2023-zipgo | https://api.github.com/repos/woowacourse-teams/2023-zipgo | closed | HTTPS + 도메인 연결 | 🌍 Infrastructure 🕋 Backend 🧚🏻♀️ Support | ### 🧚🏻♀️ Describe
https와 도메인을 연결합니다
### ✅ Tasks
- [x] nginx + certbot을 이용한 도메인과 https 연결
- [x] nginx
- [x] 도메인 연결
- [x] https 연결
### 🕖 예상 작업 소요 시간
- 3시간
- 늦어도 오늘 안엔 가능
### 🙋🏻 More
[적용 완료!](https://zipgo.pet) | 1.0 | HTTPS + 도메인 연결 - ### 🧚🏻♀️ Describe
https와 도메인을 연결합니다
### ✅ Tasks
- [x] nginx + certbot을 이용한 도메인과 https 연결
- [x] nginx
- [x] 도메인 연결
- [x] https 연결
### 🕖 예상 작업 소요 시간
- 3시간
- 늦어도 오늘 안엔 가능
### 🙋🏻 More
[적용 완료!](https://zipgo.pet) | non_process | https 도메인 연결 🧚🏻♀️ describe https와 도메인을 연결합니다 ✅ tasks nginx certbot을 이용한 도메인과 https 연결 nginx 도메인 연결 https 연결 🕖 예상 작업 소요 시간 늦어도 오늘 안엔 가능 🙋🏻 more | 0 |
2,396 | 5,191,905,518 | IssuesEvent | 2017-01-22 01:37:01 | mitchellh/packer | https://api.github.com/repos/mitchellh/packer | closed | Packer crash with vagrant post-processor | crash post-processor/vagrant | https://gist.github.com/cbednarski/26e40b91a1dba233cc78
No crash when I remove the use of the vagrant post processor.
| 1.0 | Packer crash with vagrant post-processor - https://gist.github.com/cbednarski/26e40b91a1dba233cc78
No crash when I remove the use of the vagrant post processor.
| process | packer crash with vagrant post processor no crash when i remove the use of the vagrant post processor | 1 |
753,760 | 26,360,642,490 | IssuesEvent | 2023-01-11 13:08:55 | OpenNebula/one | https://api.github.com/repos/OpenNebula/one | closed | Update VM templates after renaming the components referenced | Type: Backlog Category: Core & System Type: Feature Sponsored Category: API Priority: Low | **Description**
After issuing a rename API Call, like `one.image.rename`, the object new name doesn't appear updated on the VM Templates that reference it.
**Use case**
The calls could issue the name update for consistency purposes. Even though the operation remains fully functional due to the object ID being used for it.
**Additional Context**
This behavior can be implemented with API Hooks triggering on the rename calls.
<!--////////////////////////////////////////////-->
<!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM -->
<!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS -->
<!-- PROGRESS WILL BE REFLECTED HERE -->
<!--////////////////////////////////////////////-->
## Progress Status
- [ ] Code committed
- [ ] Testing - QA
- [ ] Documentation (Release notes - resolved issues, compatibility, known issues)
| 1.0 | Update VM templates after renaming the components referenced - **Description**
After issuing a rename API Call, like `one.image.rename`, the object new name doesn't appear updated on the VM Templates that reference it.
**Use case**
The calls could issue the name update for consistency purposes. Even though the operation remains fully functional due to the object ID being used for it.
**Additional Context**
This behavior can be implemented with API Hooks triggering on the rename calls.
<!--////////////////////////////////////////////-->
<!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM -->
<!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS -->
<!-- PROGRESS WILL BE REFLECTED HERE -->
<!--////////////////////////////////////////////-->
## Progress Status
- [ ] Code committed
- [ ] Testing - QA
- [ ] Documentation (Release notes - resolved issues, compatibility, known issues)
| non_process | update vm templates after renaming the components referenced description after issuing a rename api call like one image rename the object new name doesn t appear updated on the vm templates that reference it use case the calls could issue the name update for consistency purposes even though the operation remains fully functional due to the object id being used for it additional context this behavior can be implemented with api hooks triggering on the rename calls progress status code committed testing qa documentation release notes resolved issues compatibility known issues | 0 |
39,728 | 20,170,913,819 | IssuesEvent | 2022-02-10 10:19:14 | milessabin/shapeless | https://api.github.com/repos/milessabin/shapeless | closed | performance issue with combined `Length` and `ToSizedHList` implicit derivation | Performance | There seems to be a performance problem with the combined implicit derivation of `Length` and `ToSizedHList`. When `n` grows, the compilation time becomes quickly unpractical. I reproduced the problem with shapeless 2.3.7 and scala 2.12.14/2.13.6 .
```scala
val n = Nat(100)
val f = Fill[n.N, String]
def test[N <: Nat](implicit l: Length.Aux[f.Out, N], v: ToSizedHList[List, String, N]): Unit = {}
// perf issue
test
// no perf issue
test[n.N]
val l = Length[f.Out]
ToSizedHList[List, String, l.Out]
``` | True | performance issue with combined `Length` and `ToSizedHList` implicit derivation - There seems to be a performance problem with the combined implicit derivation of `Length` and `ToSizedHList`. When `n` grows, the compilation time becomes quickly unpractical. I reproduced the problem with shapeless 2.3.7 and scala 2.12.14/2.13.6 .
```scala
val n = Nat(100)
val f = Fill[n.N, String]
def test[N <: Nat](implicit l: Length.Aux[f.Out, N], v: ToSizedHList[List, String, N]): Unit = {}
// perf issue
test
// no perf issue
test[n.N]
val l = Length[f.Out]
ToSizedHList[List, String, l.Out]
``` | non_process | performance issue with combined length and tosizedhlist implicit derivation there seems to be a performance problem with the combined implicit derivation of length and tosizedhlist when n grows the compilation time becomes quickly unpractical i reproduced the problem with shapeless and scala scala val n nat val f fill def test implicit l length aux v tosizedhlist unit perf issue test no perf issue test val l length tosizedhlist | 0 |
22,305 | 30,859,670,789 | IssuesEvent | 2023-08-03 01:08:02 | emily-writes-poems/emily-writes-poems-processing | https://api.github.com/repos/emily-writes-poems/emily-writes-poems-processing | closed | editing poems in existing collections | script migration processing | being able to add/remove poems from collections. currently the collection poems list is stored as 2 arrays in Mongo (poem_ids and poem_titles)
easiest way is probably to pass around an array/list of poem ids and then revise the poem_titles array with a query from main poems coll? | 1.0 | editing poems in existing collections - being able to add/remove poems from collections. currently the collection poems list is stored as 2 arrays in Mongo (poem_ids and poem_titles)
easiest way is probably to pass around an array/list of poem ids and then revise the poem_titles array with a query from main poems coll? | process | editing poems in existing collections being able to add remove poems from collections currently the collection poems list is stored as arrays in mongo poem ids and poem titles easiest way is probably to pass around an array list of poem ids and then revise the poem titles array with a query from main poems coll | 1 |
205,762 | 15,686,436,957 | IssuesEvent | 2021-03-25 12:31:38 | Slimefun/Slimefun4 | https://api.github.com/repos/Slimefun/Slimefun4 | opened | Automated Ignition Chamber dupe items | 🎯 Needs testing 🐞 Bug Report |
## :round_pushpin: Description (REQUIRED)
When the Automated Ignition Chamber is broken, items from it are doubled
## :bookmark_tabs: Steps to reproduce the Issue (REQUIRED)
https://youtu.be/6hAKlnGN41A
## :bulb: Expected behavior (REQUIRED)
Items do not multiply
## :compass: Environment (REQUIRED)

| 1.0 | Automated Ignition Chamber dupe items -
## :round_pushpin: Description (REQUIRED)
When the Automated Ignition Chamber is broken, items from it are doubled
## :bookmark_tabs: Steps to reproduce the Issue (REQUIRED)
https://youtu.be/6hAKlnGN41A
## :bulb: Expected behavior (REQUIRED)
Items do not multiply
## :compass: Environment (REQUIRED)

| non_process | automated ignition chamber dupe items round pushpin description required when the automated ignition chamber is broken items from it are doubled bookmark tabs steps to reproduce the issue required bulb expected behavior required items do not multiply compass environment required | 0 |
296,674 | 9,125,136,293 | IssuesEvent | 2019-02-24 11:00:46 | python/mypy | https://api.github.com/repos/python/mypy | closed | Wrong type inferred for union containing restricted type variable | bug false-positive priority-1-normal topic-union-types | ```
from typing import Generic, TypeVar, Union
T = TypeVar('T')
class G(Generic[T]): pass
class A(object): pass
class B(object): pass
g_a = None # type: G[A]
g_b = None # type: G[B]
AB = TypeVar('AB', A, B)
def f(x):
# type: (Union[G[AB],AB]) -> G[AB]
pass
f(A())
f(B())
f(g_a)
f(g_b) # E: Argument 1 to "f" has incompatible type G[B]; expected "Union[G[A], A]"
```
The error is incorrect--`f(g_b)` should be valid.
Swapping `A` and `B` in the definition of `AB` results in an error on the `f(g_a)` call instead.
I'm not sure this title clearly describes the issue; improvements welcome. | 1.0 | Wrong type inferred for union containing restricted type variable - ```
from typing import Generic, TypeVar, Union
T = TypeVar('T')
class G(Generic[T]): pass
class A(object): pass
class B(object): pass
g_a = None # type: G[A]
g_b = None # type: G[B]
AB = TypeVar('AB', A, B)
def f(x):
# type: (Union[G[AB],AB]) -> G[AB]
pass
f(A())
f(B())
f(g_a)
f(g_b) # E: Argument 1 to "f" has incompatible type G[B]; expected "Union[G[A], A]"
```
The error is incorrect--`f(g_b)` should be valid.
Swapping `A` and `B` in the definition of `AB` results in an error on the `f(g_a)` call instead.
I'm not sure this title clearly describes the issue; improvements welcome. | non_process | wrong type inferred for union containing restricted type variable from typing import generic typevar union t typevar t class g generic pass class a object pass class b object pass g a none type g g b none type g ab typevar ab a b def f x type union ab g pass f a f b f g a f g b e argument to f has incompatible type g expected union a the error is incorrect f g b should be valid swapping a and b in the definition of ab results in an error on the f g a call instead i m not sure this title clearly describes the issue improvements welcome | 0 |
3,169 | 6,224,106,515 | IssuesEvent | 2017-07-10 13:36:07 | dzhw/zofar | https://api.github.com/repos/dzhw/zofar | opened | Monitoring-Bridge | category: technical.processes prio: 9999 status: discussion type: backlog.item | Weiterreichung von Metrikdaten. Durch Trennung von HIS-IT: Überlegung über den Aufbau einer
Virtualisierung eigener Metriken!
engl. tba | 1.0 | Monitoring-Bridge - Weiterreichung von Metrikdaten. Durch Trennung von HIS-IT: Überlegung über den Aufbau einer
Virtualisierung eigener Metriken!
engl. tba | process | monitoring bridge weiterreichung von metrikdaten durch trennung von his it überlegung über den aufbau einer virtualisierung eigener metriken engl tba | 1 |
7,953 | 11,137,562,938 | IssuesEvent | 2019-12-20 19:42:22 | openopps/openopps-platform | https://api.github.com/repos/openopps/openopps-platform | closed | Display education in sorted order on application review page | Apply Process Requirements Ready State Dept. | Who: Student applicants
What: Display education data by sort preference on the application review
Why: to allow applicants to review the application correctly
Acceptance Criteria:
- The education data will now have a sort order (either default from USAJOBS or resorted in open opps). Display the education data in the correct sort order on the application review page | 1.0 | Display education in sorted order on application review page - Who: Student applicants
What: Display education data by sort preference on the application review
Why: to allow applicants to review the application correctly
Acceptance Criteria:
- The education data will now have a sort order (either default from USAJOBS or resorted in open opps). Display the education data in the correct sort order on the application review page | process | display education in sorted order on application review page who student applicants what display education data by sort preference on the application review why to allow applicants to review the application correctly acceptance criteria the education data will now have a sort order either default from usajobs or resorted in open opps display the education data in the correct sort order on the application review page | 1 |
1,046 | 3,513,113,129 | IssuesEvent | 2016-01-11 08:33:04 | nodejs/node | https://api.github.com/repos/nodejs/node | closed | Error in process._tickCallback | process | After upgrading from node v0.12.7 to v4.2.3 our application crashes every few hours or after 1 or 2 days with the following uncaught exception:
```
TypeError: Cannot read property 'callback' of undefined
at process._tickCallback (node.js:341:26)
```
Could this be a bug in node.js?
The code in src/node.js doesn't look like `tock` is expected to be undefined:
```javascript
// Run callbacks that have no domain.
// Using domains will cause this to be overridden.
function _tickCallback() {
var callback, args, tock;
do {
while (tickInfo[kIndex] < tickInfo[kLength]) {
tock = nextTickQueue[tickInfo[kIndex]++];
callback = tock.callback; // <------- line 341. tock is undefined
args = tock.args;
``` | 1.0 | Error in process._tickCallback - After upgrading from node v0.12.7 to v4.2.3 our application crashes every few hours or after 1 or 2 days with the following uncaught exception:
```
TypeError: Cannot read property 'callback' of undefined
at process._tickCallback (node.js:341:26)
```
Could this be a bug in node.js?
The code in src/node.js doesn't look like `tock` is expected to be undefined:
```javascript
// Run callbacks that have no domain.
// Using domains will cause this to be overridden.
function _tickCallback() {
var callback, args, tock;
do {
while (tickInfo[kIndex] < tickInfo[kLength]) {
tock = nextTickQueue[tickInfo[kIndex]++];
callback = tock.callback; // <------- line 341. tock is undefined
args = tock.args;
``` | process | error in process tickcallback after upgrading from node to our application crashes every few hours or after or days with the following uncaught exception typeerror cannot read property callback of undefined at process tickcallback node js could this be a bug in node js the code in src node js doesn t look like tock is expected to be undefined javascript run callbacks that have no domain using domains will cause this to be overridden function tickcallback var callback args tock do while tickinfo tickinfo tock nexttickqueue callback tock callback line tock is undefined args tock args | 1 |
21,604 | 30,005,553,287 | IssuesEvent | 2023-06-26 12:13:31 | open-telemetry/opentelemetry-collector-contrib | https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib | closed | processor: add resource_attributes group in metadata.yaml files | enhancement processor/k8sattributes processor/resourcedetection cmd/mdatagen | ### Component(s)
k8sattributesprocessor, resourcedetectionprocessor
### Describe the issue you're reporting
The group `resource_attributes` was introduced in the following [PR](https://github.com/open-telemetry/opentelemetry-collector-contrib/pull/21664).
As of today, no processors use the `resource_attributes` group and its generated config structs. What I'm proposing is adding the `resource_attributes` to processors where this is relevant:
- [x] [k8sattributesprocessor](https://github.com/open-telemetry/opentelemetry-collector-contrib/issues/23152)
- [x] [resourcedetectionprocessor](https://github.com/open-telemetry/opentelemetry-collector-contrib/issues/21482)
| 2.0 | processor: add resource_attributes group in metadata.yaml files - ### Component(s)
k8sattributesprocessor, resourcedetectionprocessor
### Describe the issue you're reporting
The group `resource_attributes` was introduced in the following [PR](https://github.com/open-telemetry/opentelemetry-collector-contrib/pull/21664).
As of today, no processors use the `resource_attributes` group and its generated config structs. What I'm proposing is adding the `resource_attributes` to processors where this is relevant:
- [x] [k8sattributesprocessor](https://github.com/open-telemetry/opentelemetry-collector-contrib/issues/23152)
- [x] [resourcedetectionprocessor](https://github.com/open-telemetry/opentelemetry-collector-contrib/issues/21482)
| process | processor add resource attributes group in metadata yaml files component s resourcedetectionprocessor describe the issue you re reporting the group resource attributes was introduced in the following as of today no processors use the resource attributes group and its generated config structs what i m proposing is adding the resource attributes to processors where this is relevant | 1 |
370,922 | 10,958,539,906 | IssuesEvent | 2019-11-27 09:36:39 | krzychu124/Cities-Skylines-Traffic-Manager-President-Edition | https://api.github.com/repos/krzychu124/Cities-Skylines-Traffic-Manager-President-Edition | opened | Speed limits window too constrained on QHD resolution | BUG SPEED LIMITS UI confirmed low priority | On this resolution:

The speed limits panel is too constrained and can't be dragged further towards the bottom-right than shown:

| 1.0 | Speed limits window too constrained on QHD resolution - On this resolution:

The speed limits panel is too constrained and can't be dragged further towards the bottom-right than shown:

| non_process | speed limits window too constrained on qhd resolution on this resolution the speed limits panel is too constrained and can t be dragged further towards the bottom right than shown | 0 |
3,104 | 6,111,644,469 | IssuesEvent | 2017-06-21 17:30:59 | NeuroTechX/eeg-101 | https://api.github.com/repos/NeuroTechX/eeg-101 | closed | Create ClassifierDataSource class to integrate with ClassifierModule and the GaussianNaiveBayesClassifier | Java Signal Processing | The job of this class will be to do three things:
1. Contain a MuseDataListener and runnable thread that fills up a buffer with streaming EEG data
2. Pre-process data for classifier
- filter
- epoch
- FFT
- extract band powers
3. Concatenate data into appropriate format to pass to classifier
| 1.0 | Create ClassifierDataSource class to integrate with ClassifierModule and the GaussianNaiveBayesClassifier - The job of this class will be to do three things:
1. Contain a MuseDataListener and runnable thread that fills up a buffer with streaming EEG data
2. Pre-process data for classifier
- filter
- epoch
- FFT
- extract band powers
3. Concatenate data into appropriate format to pass to classifier
| process | create classifierdatasource class to integrate with classifiermodule and the gaussiannaivebayesclassifier the job of this class will be to do three things contain a musedatalistener and runnable thread that fills up a buffer with streaming eeg data pre process data for classifier filter epoch fft extract band powers concatenate data into appropriate format to pass to classifier | 1 |
816,289 | 30,596,510,264 | IssuesEvent | 2023-07-21 23:00:09 | googleapis/python-storage | https://api.github.com/repos/googleapis/python-storage | closed | tests.system.test_bucket: test_bucket_list_blobs_paginated_w_offset failed | api: storage type: bug priority: p2 flakybot: issue flakybot: flaky | Note: #969 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky.
----
commit: f9462179f4a4b08eea7471a5ffb4aa5071fc5a5e
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/ee1c078c-b3c3-4ce6-9627-cae04bcaee3d), [Sponge](http://sponge2/ee1c078c-b3c3-4ce6-9627-cae04bcaee3d)
status: failed
<details><summary>Test output</summary><br><pre>storage_client = <google.cloud.storage.client.Client object at 0x7f22b672bca0>
listable_bucket_name = 'gcp-systest-listable-1689908513911'
file_data = {'big': {'hash': b'cEome4a+NYd7YIXzXQnR5Q==', 'path': '/tmpfs/src/github/python-storage/tests/data/five-point-one-mb-f...g'}, 'simple': {'hash': b'3Hkwjv2WvCnKjNR6Z3CboA==', 'path': '/tmpfs/src/github/python-storage/tests/data/simple.txt'}}
@pytest.fixture(scope="session")
def listable_bucket(storage_client, listable_bucket_name, file_data):
bucket = storage_client.bucket(listable_bucket_name)
> _helpers.retry_429_503(bucket.create)()
tests/system/conftest.py:105:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
.nox/system-3-8/lib/python3.8/site-packages/test_utils/retry.py:100: in wrapped_function
return to_wrap(*args, **kwargs)
google/cloud/storage/bucket.py:972: in create
client.create_bucket(
google/cloud/storage/client.py:961: in create_bucket
api_response = self._post_resource(
google/cloud/storage/client.py:625: in _post_resource
return self._connection.api_request(
google/cloud/storage/_http.py:72: in api_request
return call()
.nox/system-3-8/lib/python3.8/site-packages/google/api_core/retry.py:349: in retry_wrapped_func
return retry_target(
.nox/system-3-8/lib/python3.8/site-packages/google/api_core/retry.py:191: in retry_target
return target()
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <google.cloud.storage._http.Connection object at 0x7f22b683b520>
method = 'POST', path = '/b', query_params = {'project': 'precise-truck-742'}
data = '{"name": "gcp-systest-listable-1689908513911"}'
content_type = 'application/json', headers = None, api_base_url = None
api_version = None, expect_json = True
_target_object = <Bucket: gcp-systest-listable-1689908513911>, timeout = 60
extra_api_info = 'gccl-invocation-id/f0b8f74a-b400-440a-a7d5-0869c725a434'
def api_request(
self,
method,
path,
query_params=None,
data=None,
content_type=None,
headers=None,
api_base_url=None,
api_version=None,
expect_json=True,
_target_object=None,
timeout=_DEFAULT_TIMEOUT,
extra_api_info=None,
):
"""Make a request over the HTTP transport to the API.
You shouldn't need to use this method, but if you plan to
interact with the API using these primitives, this is the
correct one to use.
:type method: str
:param method: The HTTP method name (ie, ``GET``, ``POST``, etc).
Required.
:type path: str
:param path: The path to the resource (ie, ``'/b/bucket-name'``).
Required.
:type query_params: dict or list
:param query_params: A dictionary of keys and values (or list of
key-value pairs) to insert into the query
string of the URL.
:type data: str
:param data: The data to send as the body of the request. Default is
the empty string.
:type content_type: str
:param content_type: The proper MIME type of the data provided. Default
is None.
:type headers: dict
:param headers: extra HTTP headers to be sent with the request.
:type api_base_url: str
:param api_base_url: The base URL for the API endpoint.
Typically you won't have to provide this.
Default is the standard API base URL.
:type api_version: str
:param api_version: The version of the API to call. Typically
you shouldn't provide this and instead use
the default for the library. Default is the
latest API version supported by
google-cloud-python.
:type expect_json: bool
:param expect_json: If True, this method will try to parse the
response as JSON and raise an exception if
that cannot be done. Default is True.
:type _target_object: :class:`object`
:param _target_object:
(Optional) Protected argument to be used by library callers. This
can allow custom behavior, for example, to defer an HTTP request
and complete initialization of the object at a later time.
:type timeout: float or tuple
:param timeout: (optional) The amount of time, in seconds, to wait
for the server response.
Can also be passed as a tuple (connect_timeout, read_timeout).
See :meth:`requests.Session.request` documentation for details.
:type extra_api_info: string
:param extra_api_info: (optional) Extra api info to be appended to
the X-Goog-API-Client header
:raises ~google.cloud.exceptions.GoogleCloudError: if the response code
is not 200 OK.
:raises ValueError: if the response content type is not JSON.
:rtype: dict or str
:returns: The API response payload, either as a raw string or
a dictionary if the response is valid JSON.
"""
url = self.build_api_url(
path=path,
query_params=query_params,
api_base_url=api_base_url,
api_version=api_version,
)
# Making the executive decision that any dictionary
# data will be sent properly as JSON.
if data and isinstance(data, dict):
data = json.dumps(data)
content_type = "application/json"
response = self._make_request(
method=method,
url=url,
data=data,
content_type=content_type,
headers=headers,
target_object=_target_object,
timeout=timeout,
extra_api_info=extra_api_info,
)
if not 200 <= response.status_code < 300:
> raise exceptions.from_http_response(response)
E google.api_core.exceptions.Conflict: 409 POST https://storage-preprod-test-unified.googleusercontent.com/storage/v1_preprod/b?project=precise-truck-742&prettyPrint=false: Your previous request to create the named bucket succeeded and you already own it.
.nox/system-3-8/lib/python3.8/site-packages/google/cloud/_http/__init__.py:494: Conflict</pre></details> | 1.0 | tests.system.test_bucket: test_bucket_list_blobs_paginated_w_offset failed - Note: #969 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky.
----
commit: f9462179f4a4b08eea7471a5ffb4aa5071fc5a5e
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/ee1c078c-b3c3-4ce6-9627-cae04bcaee3d), [Sponge](http://sponge2/ee1c078c-b3c3-4ce6-9627-cae04bcaee3d)
status: failed
<details><summary>Test output</summary><br><pre>storage_client = <google.cloud.storage.client.Client object at 0x7f22b672bca0>
listable_bucket_name = 'gcp-systest-listable-1689908513911'
file_data = {'big': {'hash': b'cEome4a+NYd7YIXzXQnR5Q==', 'path': '/tmpfs/src/github/python-storage/tests/data/five-point-one-mb-f...g'}, 'simple': {'hash': b'3Hkwjv2WvCnKjNR6Z3CboA==', 'path': '/tmpfs/src/github/python-storage/tests/data/simple.txt'}}
@pytest.fixture(scope="session")
def listable_bucket(storage_client, listable_bucket_name, file_data):
bucket = storage_client.bucket(listable_bucket_name)
> _helpers.retry_429_503(bucket.create)()
tests/system/conftest.py:105:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
.nox/system-3-8/lib/python3.8/site-packages/test_utils/retry.py:100: in wrapped_function
return to_wrap(*args, **kwargs)
google/cloud/storage/bucket.py:972: in create
client.create_bucket(
google/cloud/storage/client.py:961: in create_bucket
api_response = self._post_resource(
google/cloud/storage/client.py:625: in _post_resource
return self._connection.api_request(
google/cloud/storage/_http.py:72: in api_request
return call()
.nox/system-3-8/lib/python3.8/site-packages/google/api_core/retry.py:349: in retry_wrapped_func
return retry_target(
.nox/system-3-8/lib/python3.8/site-packages/google/api_core/retry.py:191: in retry_target
return target()
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <google.cloud.storage._http.Connection object at 0x7f22b683b520>
method = 'POST', path = '/b', query_params = {'project': 'precise-truck-742'}
data = '{"name": "gcp-systest-listable-1689908513911"}'
content_type = 'application/json', headers = None, api_base_url = None
api_version = None, expect_json = True
_target_object = <Bucket: gcp-systest-listable-1689908513911>, timeout = 60
extra_api_info = 'gccl-invocation-id/f0b8f74a-b400-440a-a7d5-0869c725a434'
def api_request(
self,
method,
path,
query_params=None,
data=None,
content_type=None,
headers=None,
api_base_url=None,
api_version=None,
expect_json=True,
_target_object=None,
timeout=_DEFAULT_TIMEOUT,
extra_api_info=None,
):
"""Make a request over the HTTP transport to the API.
You shouldn't need to use this method, but if you plan to
interact with the API using these primitives, this is the
correct one to use.
:type method: str
:param method: The HTTP method name (ie, ``GET``, ``POST``, etc).
Required.
:type path: str
:param path: The path to the resource (ie, ``'/b/bucket-name'``).
Required.
:type query_params: dict or list
:param query_params: A dictionary of keys and values (or list of
key-value pairs) to insert into the query
string of the URL.
:type data: str
:param data: The data to send as the body of the request. Default is
the empty string.
:type content_type: str
:param content_type: The proper MIME type of the data provided. Default
is None.
:type headers: dict
:param headers: extra HTTP headers to be sent with the request.
:type api_base_url: str
:param api_base_url: The base URL for the API endpoint.
Typically you won't have to provide this.
Default is the standard API base URL.
:type api_version: str
:param api_version: The version of the API to call. Typically
you shouldn't provide this and instead use
the default for the library. Default is the
latest API version supported by
google-cloud-python.
:type expect_json: bool
:param expect_json: If True, this method will try to parse the
response as JSON and raise an exception if
that cannot be done. Default is True.
:type _target_object: :class:`object`
:param _target_object:
(Optional) Protected argument to be used by library callers. This
can allow custom behavior, for example, to defer an HTTP request
and complete initialization of the object at a later time.
:type timeout: float or tuple
:param timeout: (optional) The amount of time, in seconds, to wait
for the server response.
Can also be passed as a tuple (connect_timeout, read_timeout).
See :meth:`requests.Session.request` documentation for details.
:type extra_api_info: string
:param extra_api_info: (optional) Extra api info to be appended to
the X-Goog-API-Client header
:raises ~google.cloud.exceptions.GoogleCloudError: if the response code
is not 200 OK.
:raises ValueError: if the response content type is not JSON.
:rtype: dict or str
:returns: The API response payload, either as a raw string or
a dictionary if the response is valid JSON.
"""
url = self.build_api_url(
path=path,
query_params=query_params,
api_base_url=api_base_url,
api_version=api_version,
)
# Making the executive decision that any dictionary
# data will be sent properly as JSON.
if data and isinstance(data, dict):
data = json.dumps(data)
content_type = "application/json"
response = self._make_request(
method=method,
url=url,
data=data,
content_type=content_type,
headers=headers,
target_object=_target_object,
timeout=timeout,
extra_api_info=extra_api_info,
)
if not 200 <= response.status_code < 300:
> raise exceptions.from_http_response(response)
E google.api_core.exceptions.Conflict: 409 POST https://storage-preprod-test-unified.googleusercontent.com/storage/v1_preprod/b?project=precise-truck-742&prettyPrint=false: Your previous request to create the named bucket succeeded and you already own it.
.nox/system-3-8/lib/python3.8/site-packages/google/cloud/_http/__init__.py:494: Conflict</pre></details> | non_process | tests system test bucket test bucket list blobs paginated w offset failed note was also for this test but it was closed more than days ago so i didn t mark it flaky commit buildurl status failed test output storage client listable bucket name gcp systest listable file data big hash b path tmpfs src github python storage tests data five point one mb f g simple hash b path tmpfs src github python storage tests data simple txt pytest fixture scope session def listable bucket storage client listable bucket name file data bucket storage client bucket listable bucket name helpers retry bucket create tests system conftest py nox system lib site packages test utils retry py in wrapped function return to wrap args kwargs google cloud storage bucket py in create client create bucket google cloud storage client py in create bucket api response self post resource google cloud storage client py in post resource return self connection api request google cloud storage http py in api request return call nox system lib site packages google api core retry py in retry wrapped func return retry target nox system lib site packages google api core retry py in retry target return target self method post path b query params project precise truck data name gcp systest listable content type application json headers none api base url none api version none expect json true target object timeout extra api info gccl invocation id def api request self method path query params none data none content type none headers none api base url none api version none expect json true target object none timeout default timeout extra api info none make a request over the http transport to the api you shouldn t need to use this method but if you plan to interact with the api using these primitives this is the correct one to use type method str param method the http method name ie get post etc required type path str param path the path to the resource ie b bucket name required type query params dict or list param query params a dictionary of keys and values or list of key value pairs to insert into the query string of the url type data str param data the data to send as the body of the request default is the empty string type content type str param content type the proper mime type of the data provided default is none type headers dict param headers extra http headers to be sent with the request type api base url str param api base url the base url for the api endpoint typically you won t have to provide this default is the standard api base url type api version str param api version the version of the api to call typically you shouldn t provide this and instead use the default for the library default is the latest api version supported by google cloud python type expect json bool param expect json if true this method will try to parse the response as json and raise an exception if that cannot be done default is true type target object class object param target object optional protected argument to be used by library callers this can allow custom behavior for example to defer an http request and complete initialization of the object at a later time type timeout float or tuple param timeout optional the amount of time in seconds to wait for the server response can also be passed as a tuple connect timeout read timeout see meth requests session request documentation for details type extra api info string param extra api info optional extra api info to be appended to the x goog api client header raises google cloud exceptions googleclouderror if the response code is not ok raises valueerror if the response content type is not json rtype dict or str returns the api response payload either as a raw string or a dictionary if the response is valid json url self build api url path path query params query params api base url api base url api version api version making the executive decision that any dictionary data will be sent properly as json if data and isinstance data dict data json dumps data content type application json response self make request method method url url data data content type content type headers headers target object target object timeout timeout extra api info extra api info if not response status code raise exceptions from http response response e google api core exceptions conflict post your previous request to create the named bucket succeeded and you already own it nox system lib site packages google cloud http init py conflict | 0 |
16,094 | 20,263,434,083 | IssuesEvent | 2022-02-15 09:48:47 | quark-engine/quark-engine | https://api.github.com/repos/quark-engine/quark-engine | closed | Update README to indicate the compatible Rizin versions | work-in-progress issue-processing-state-04 | **Is your feature request related to a problem? Please describe.**
An API change in Rizin v0.3.0 has led the Rizin-based analysis to fail. To avoid users experiencing errors similar to #305, we need to indicate this restriction before fixing it.
**Describe the solution you'd like.**
Indicate the compatible Rizin versions in the README. For now, Quark only works with Rizin v0.2.0 and v0.2.1. | 1.0 | Update README to indicate the compatible Rizin versions - **Is your feature request related to a problem? Please describe.**
An API change in Rizin v0.3.0 has led the Rizin-based analysis to fail. To avoid users experiencing errors similar to #305, we need to indicate this restriction before fixing it.
**Describe the solution you'd like.**
Indicate the compatible Rizin versions in the README. For now, Quark only works with Rizin v0.2.0 and v0.2.1. | process | update readme to indicate the compatible rizin versions is your feature request related to a problem please describe an api change in rizin has led the rizin based analysis to fail to avoid users experiencing errors similar to we need to indicate this restriction before fixing it describe the solution you d like indicate the compatible rizin versions in the readme for now quark only works with rizin and | 1 |
17,762 | 23,691,256,724 | IssuesEvent | 2022-08-29 11:01:28 | pyanodon/pybugreports | https://api.github.com/repos/pyanodon/pybugreports | closed | Inserter mode gets stomped excessively | bug mod:pycoalprocessing | ### Mod source
PyAE Beta
### Which mod are you having an issue with?
- [ ] pyalienlife
- [ ] pyalternativeenergy
- [X] pycoalprocessing
- [ ] pyfusionenergy
- [ ] pyhightech
- [ ] pyindustry
- [ ] pypetroleumhandling
- [ ] pypostprocessing
- [ ] pyrawores
### Operating system
>=Windows 10
### What kind of issue is this?
- [ ] Compatibility
- [ ] Locale (names, descriptions, unknown keys)
- [ ] Graphical
- [ ] Crash
- [ ] Progression
- [ ] Balance
- [ ] Pypostprocessing failure
- [X] Other
### What is the problem?
The burner-inserter and inserter forcing to blacklist from either on_built_entity or on_configuration_changed is a bit too heavy-handed at times.
It would be helpful if it considered a circuit attachment to be the same as having a filter set as far as avoiding the force back to blacklist.
### Steps to reproduce
_No response_
### Additional context
_No response_
### Log file
_No response_ | 1.0 | Inserter mode gets stomped excessively - ### Mod source
PyAE Beta
### Which mod are you having an issue with?
- [ ] pyalienlife
- [ ] pyalternativeenergy
- [X] pycoalprocessing
- [ ] pyfusionenergy
- [ ] pyhightech
- [ ] pyindustry
- [ ] pypetroleumhandling
- [ ] pypostprocessing
- [ ] pyrawores
### Operating system
>=Windows 10
### What kind of issue is this?
- [ ] Compatibility
- [ ] Locale (names, descriptions, unknown keys)
- [ ] Graphical
- [ ] Crash
- [ ] Progression
- [ ] Balance
- [ ] Pypostprocessing failure
- [X] Other
### What is the problem?
The burner-inserter and inserter forcing to blacklist from either on_built_entity or on_configuration_changed is a bit too heavy-handed at times.
It would be helpful if it considered a circuit attachment to be the same as having a filter set as far as avoiding the force back to blacklist.
### Steps to reproduce
_No response_
### Additional context
_No response_
### Log file
_No response_ | process | inserter mode gets stomped excessively mod source pyae beta which mod are you having an issue with pyalienlife pyalternativeenergy pycoalprocessing pyfusionenergy pyhightech pyindustry pypetroleumhandling pypostprocessing pyrawores operating system windows what kind of issue is this compatibility locale names descriptions unknown keys graphical crash progression balance pypostprocessing failure other what is the problem the burner inserter and inserter forcing to blacklist from either on built entity or on configuration changed is a bit too heavy handed at times it would be helpful if it considered a circuit attachment to be the same as having a filter set as far as avoiding the force back to blacklist steps to reproduce no response additional context no response log file no response | 1 |
5,736 | 8,580,069,955 | IssuesEvent | 2018-11-13 10:53:42 | threefoldtech/rivine | https://api.github.com/repos/threefoldtech/rivine | closed | Hierarchical Deterministic wallets | process_duplicate type_feature type_story | https://github.com/bitcoin/bips/blob/master/bip-0032.mediawiki
Ou deterministic wallet typically consist of a single "chain" of keypairs. The fact that there is only one chain means that sharing a wallet happens on an all-or-nothing basis. However, in some cases one only wants some (public) keys to be shared and recoverable. In the example of a webshop, the webserver does not need access to all public keys of the merchant's wallet; only to those addresses which are used to receive customer's payments, and not for example the change addresses that are generated when the merchant spends money. Hierarchical deterministic wallets allow such selective sharing by supporting multiple keypair chains, derived from a single root.
This would also mean that a 3 bot can have it's own chain ,just as a mobile or jumpscale wallet can, all derived from a single root which can be locked away safely. | 1.0 | Hierarchical Deterministic wallets - https://github.com/bitcoin/bips/blob/master/bip-0032.mediawiki
Ou deterministic wallet typically consist of a single "chain" of keypairs. The fact that there is only one chain means that sharing a wallet happens on an all-or-nothing basis. However, in some cases one only wants some (public) keys to be shared and recoverable. In the example of a webshop, the webserver does not need access to all public keys of the merchant's wallet; only to those addresses which are used to receive customer's payments, and not for example the change addresses that are generated when the merchant spends money. Hierarchical deterministic wallets allow such selective sharing by supporting multiple keypair chains, derived from a single root.
This would also mean that a 3 bot can have it's own chain ,just as a mobile or jumpscale wallet can, all derived from a single root which can be locked away safely. | process | hierarchical deterministic wallets ou deterministic wallet typically consist of a single chain of keypairs the fact that there is only one chain means that sharing a wallet happens on an all or nothing basis however in some cases one only wants some public keys to be shared and recoverable in the example of a webshop the webserver does not need access to all public keys of the merchant s wallet only to those addresses which are used to receive customer s payments and not for example the change addresses that are generated when the merchant spends money hierarchical deterministic wallets allow such selective sharing by supporting multiple keypair chains derived from a single root this would also mean that a bot can have it s own chain just as a mobile or jumpscale wallet can all derived from a single root which can be locked away safely | 1 |
17,323 | 23,142,780,670 | IssuesEvent | 2022-07-28 20:17:46 | USGS-R/drb-do-ml | https://api.github.com/repos/USGS-R/drb-do-ml | closed | Check correlation between temperature and DO predictions | process-guidance | We are interested to see the correlation between model DO predictions and input air temperature in the Baseline model. | 1.0 | Check correlation between temperature and DO predictions - We are interested to see the correlation between model DO predictions and input air temperature in the Baseline model. | process | check correlation between temperature and do predictions we are interested to see the correlation between model do predictions and input air temperature in the baseline model | 1 |
18,803 | 10,231,609,955 | IssuesEvent | 2019-08-18 11:05:58 | TryGhost/Ghost | https://api.github.com/repos/TryGhost/Ghost | closed | Delete all content triggers a 504 gateway timeout | admin-api api help wanted performance server stale | ### Issue Summary
Using the Delete all content feature in Labs times out with a collection exceeding 3400 stories. This is not a limit I have explored, but the current situation I am in.
### To Reproduce
Import 3400+ stories in to ghost, try to delete all content.
I had to batch this import as the 5.5Mb file times out also, but that is probably due to my connection speed.
**Any other info e.g. Why do you consider this to be a bug? What did you expect to happen instead?**
I expected the request not to timeout.
This is a development environment, and I would not expect this operation to be carried out under normal production circumstances. Of course I can nuke the instance and start again with no issues.
Just though I would let you know of the issue.
### Technical details:
* Ghost-CLI version: 1.9.9
* Node Version: 10.15.0
* Browser/OS: N/A
* Database: mysql Ver 14.14 Distrib 5.7.25, for Linux (x86_64) using EditLine wrapper
| True | Delete all content triggers a 504 gateway timeout - ### Issue Summary
Using the Delete all content feature in Labs times out with a collection exceeding 3400 stories. This is not a limit I have explored, but the current situation I am in.
### To Reproduce
Import 3400+ stories in to ghost, try to delete all content.
I had to batch this import as the 5.5Mb file times out also, but that is probably due to my connection speed.
**Any other info e.g. Why do you consider this to be a bug? What did you expect to happen instead?**
I expected the request not to timeout.
This is a development environment, and I would not expect this operation to be carried out under normal production circumstances. Of course I can nuke the instance and start again with no issues.
Just though I would let you know of the issue.
### Technical details:
* Ghost-CLI version: 1.9.9
* Node Version: 10.15.0
* Browser/OS: N/A
* Database: mysql Ver 14.14 Distrib 5.7.25, for Linux (x86_64) using EditLine wrapper
| non_process | delete all content triggers a gateway timeout issue summary using the delete all content feature in labs times out with a collection exceeding stories this is not a limit i have explored but the current situation i am in to reproduce import stories in to ghost try to delete all content i had to batch this import as the file times out also but that is probably due to my connection speed any other info e g why do you consider this to be a bug what did you expect to happen instead i expected the request not to timeout this is a development environment and i would not expect this operation to be carried out under normal production circumstances of course i can nuke the instance and start again with no issues just though i would let you know of the issue technical details ghost cli version node version browser os n a database mysql ver distrib for linux using editline wrapper | 0 |
218,208 | 16,976,709,180 | IssuesEvent | 2021-06-30 00:45:10 | microsoft/vscode | https://api.github.com/repos/microsoft/vscode | closed | VS Code Live Server - June Iteration Plans | feature-request live-server on-testplan | <!-- ⚠️⚠️ Do Not Delete This! feature_request_template ⚠️⚠️ -->
<!-- Please read our Rules of Conduct: https://opensource.microsoft.com/codeofconduct/ -->
<!-- Please search existing issues to avoid creating duplicates. -->
<!-- Describe the feature you'd like. -->
This tracks the upcoming work on the Live Server, which is described in #124607. Some of these may be deferred to the July iteration. See backlog #125343
Bug Fixes:
- [x] External links on browser are broken. #125200
- [x] Relative file links in sub-folders are not working. #125201
- [x] Make hosting port flexible + have error handling for if ports are occupied. #125152
- [x] Two refreshes when file change happens #125174
- [x] Filenames end in slash in index menu #125178
- [x] Handle situation if `index.html` does not exist (can show page with all of the files to pick).
- [x] Fix `Upgrade Required` when using localhost instead of 127.0.0.1
- [x] Support Emojis on html files
Features:
- [x] Basic Implementation of Default Page Index/Filesystem Navigation
- [x] Basic Implementation of "Page Not Found"
- [x] Allow window reload to show last visible file in preview (currently always shows `index.html`)
- [x] Allow users to choose which file to display using active editor. #125775
- [x] Implement more visual way of showing server is on or off. #125854
- [x] Allow user settings to tweak experience #125855
- [x] Explore best UX experience for users. #125949 #126307
- [x] Server run using tasks #126309
- [x] Page title support on embedded preview.
- [x] Non-workspace Support #126715
- [x] 💪SVG Preview Support #126093
- [x] Release Docs + Code Cleanup #126308
- [x] Link Preview #127019
- [x] Multi-root workspaces #126622
Admin:
- [x] Transfer repo ownership to Microsoft
- [x] Look into extension bundling for smaller `*.vsix` size (bundling)
- [x] Upload to marketplace. | 1.0 | VS Code Live Server - June Iteration Plans - <!-- ⚠️⚠️ Do Not Delete This! feature_request_template ⚠️⚠️ -->
<!-- Please read our Rules of Conduct: https://opensource.microsoft.com/codeofconduct/ -->
<!-- Please search existing issues to avoid creating duplicates. -->
<!-- Describe the feature you'd like. -->
This tracks the upcoming work on the Live Server, which is described in #124607. Some of these may be deferred to the July iteration. See backlog #125343
Bug Fixes:
- [x] External links on browser are broken. #125200
- [x] Relative file links in sub-folders are not working. #125201
- [x] Make hosting port flexible + have error handling for if ports are occupied. #125152
- [x] Two refreshes when file change happens #125174
- [x] Filenames end in slash in index menu #125178
- [x] Handle situation if `index.html` does not exist (can show page with all of the files to pick).
- [x] Fix `Upgrade Required` when using localhost instead of 127.0.0.1
- [x] Support Emojis on html files
Features:
- [x] Basic Implementation of Default Page Index/Filesystem Navigation
- [x] Basic Implementation of "Page Not Found"
- [x] Allow window reload to show last visible file in preview (currently always shows `index.html`)
- [x] Allow users to choose which file to display using active editor. #125775
- [x] Implement more visual way of showing server is on or off. #125854
- [x] Allow user settings to tweak experience #125855
- [x] Explore best UX experience for users. #125949 #126307
- [x] Server run using tasks #126309
- [x] Page title support on embedded preview.
- [x] Non-workspace Support #126715
- [x] 💪SVG Preview Support #126093
- [x] Release Docs + Code Cleanup #126308
- [x] Link Preview #127019
- [x] Multi-root workspaces #126622
Admin:
- [x] Transfer repo ownership to Microsoft
- [x] Look into extension bundling for smaller `*.vsix` size (bundling)
- [x] Upload to marketplace. | non_process | vs code live server june iteration plans this tracks the upcoming work on the live server which is described in some of these may be deferred to the july iteration see backlog bug fixes external links on browser are broken relative file links in sub folders are not working make hosting port flexible have error handling for if ports are occupied two refreshes when file change happens filenames end in slash in index menu handle situation if index html does not exist can show page with all of the files to pick fix upgrade required when using localhost instead of support emojis on html files features basic implementation of default page index filesystem navigation basic implementation of page not found allow window reload to show last visible file in preview currently always shows index html allow users to choose which file to display using active editor implement more visual way of showing server is on or off allow user settings to tweak experience explore best ux experience for users server run using tasks page title support on embedded preview non workspace support 💪svg preview support release docs code cleanup link preview multi root workspaces admin transfer repo ownership to microsoft look into extension bundling for smaller vsix size bundling upload to marketplace | 0 |
7,404 | 10,523,658,825 | IssuesEvent | 2019-09-30 11:34:10 | teleporthq/teleport-code-generators | https://api.github.com/repos/teleporthq/teleport-code-generators | closed | HTML formatting utility | enhancement post-processors | We're relying on `prettier` for formatting html chunks (which are represented as strings). However, perf tests (#64) showed that prettier is running very slow when the UIDL gets bigger (ex: 2000 nodes).
We have two ways of approaching this:
### Simple post-processor function that formats html tags
This should run after the chunks are linked and should just align the html tags and content accordingly. Important to note that some frameworks might have non-standard html elements to render.
### HAST formatter
This should run at link time, when the html is a HAST structure represented internally. some utilities from the unified ecosystem could be checked out for this. | 1.0 | HTML formatting utility - We're relying on `prettier` for formatting html chunks (which are represented as strings). However, perf tests (#64) showed that prettier is running very slow when the UIDL gets bigger (ex: 2000 nodes).
We have two ways of approaching this:
### Simple post-processor function that formats html tags
This should run after the chunks are linked and should just align the html tags and content accordingly. Important to note that some frameworks might have non-standard html elements to render.
### HAST formatter
This should run at link time, when the html is a HAST structure represented internally. some utilities from the unified ecosystem could be checked out for this. | process | html formatting utility we re relying on prettier for formatting html chunks which are represented as strings however perf tests showed that prettier is running very slow when the uidl gets bigger ex nodes we have two ways of approaching this simple post processor function that formats html tags this should run after the chunks are linked and should just align the html tags and content accordingly important to note that some frameworks might have non standard html elements to render hast formatter this should run at link time when the html is a hast structure represented internally some utilities from the unified ecosystem could be checked out for this | 1 |
26,913 | 2,688,753,225 | IssuesEvent | 2015-03-31 03:36:52 | ChrisMahlke/nominate-2 | https://api.github.com/repos/ChrisMahlke/nominate-2 | closed | My Profile tab is not turning blue on Refresh | MediumPriority | After fixing the “My Profile” section to receive full points, the red text and border does not change until I click on another tab. This should change to blue automatically after reaching the desired score. I really appreciate the “Refresh” button, but it would be nice if it happened automatically.
| 1.0 | My Profile tab is not turning blue on Refresh - After fixing the “My Profile” section to receive full points, the red text and border does not change until I click on another tab. This should change to blue automatically after reaching the desired score. I really appreciate the “Refresh” button, but it would be nice if it happened automatically.
| non_process | my profile tab is not turning blue on refresh after fixing the “my profile” section to receive full points the red text and border does not change until i click on another tab this should change to blue automatically after reaching the desired score i really appreciate the “refresh” button but it would be nice if it happened automatically | 0 |
10,158 | 13,044,162,640 | IssuesEvent | 2020-07-29 03:47:34 | tikv/tikv | https://api.github.com/repos/tikv/tikv | closed | UCP: Migrate scalar function `AesEncryptIV` from TiDB | challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor |
## Description
Port the scalar function `AesEncryptIV` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @lonng
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
| 2.0 | UCP: Migrate scalar function `AesEncryptIV` from TiDB -
## Description
Port the scalar function `AesEncryptIV` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @lonng
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
| process | ucp migrate scalar function aesencryptiv from tidb description port the scalar function aesencryptiv from tidb to coprocessor score mentor s lonng recommended skills rust programming learning materials already implemented expressions ported from tidb | 1 |
4,683 | 7,522,205,207 | IssuesEvent | 2018-04-12 19:39:28 | rubberduck-vba/Rubberduck | https://api.github.com/repos/rubberduck-vba/Rubberduck | closed | Members of implicit default object property are not resolved | bug navigation parse-tree-preprocessing | The default member of Workbooks is `[_Default]`, which is roughly equivalent to `Item`. Both `[_Default]` and `Item` both return a `Workbook` object, so RD should be able to resolve the `Workbook` members.
But, when the default member call is implicit, RD fails to resolve the `Workbook` properties
```vb
Debug.Print Workbooks.[_Default](ThisWorkbook.Name).ReadOnly 'ReadOnly is recognized
Debug.Print Workbooks.Item(ThisWorkbook.Name).ReadOnly 'ReadOnly is recognized
Debug.Print Workbooks(ThisWorkbook.Name).ReadOnly 'ReadOnly NOT recognized
```
| 1.0 | Members of implicit default object property are not resolved - The default member of Workbooks is `[_Default]`, which is roughly equivalent to `Item`. Both `[_Default]` and `Item` both return a `Workbook` object, so RD should be able to resolve the `Workbook` members.
But, when the default member call is implicit, RD fails to resolve the `Workbook` properties
```vb
Debug.Print Workbooks.[_Default](ThisWorkbook.Name).ReadOnly 'ReadOnly is recognized
Debug.Print Workbooks.Item(ThisWorkbook.Name).ReadOnly 'ReadOnly is recognized
Debug.Print Workbooks(ThisWorkbook.Name).ReadOnly 'ReadOnly NOT recognized
```
| process | members of implicit default object property are not resolved the default member of workbooks is which is roughly equivalent to item both and item both return a workbook object so rd should be able to resolve the workbook members but when the default member call is implicit rd fails to resolve the workbook properties vb debug print workbooks thisworkbook name readonly readonly is recognized debug print workbooks item thisworkbook name readonly readonly is recognized debug print workbooks thisworkbook name readonly readonly not recognized | 1 |
516,271 | 14,978,503,001 | IssuesEvent | 2021-01-28 10:55:12 | nextcloud/mail | https://api.github.com/repos/nextcloud/mail | closed | Can not access mail account in nextcloud after password change on imap server | 4. to release bug priority:medium | ### Expected behavior
There should be an option to set the new password.
### Actual behavior
Mail account is not shown anymore in nextcloud. The cronjob gives "OCA\Mail\Exception\ServiceException: IMAP errorMail server denied authentication." Can not access the mail account, so I even can not delete it.
### Mail app
**Mail app version:** 1.4.1
**Mailserver or service:** imap.1und1.de
### Server configuration
**Operating system**: Ubuntu 18.04
**Web server:** Apache 2.4.46
**Database:** MariaDB 10.4.15
**PHP version:** 7.3.23
**Nextcloud Version:** 19.0.4
#### Client configuration
**Browser:** Firefox 82.0
| 1.0 | Can not access mail account in nextcloud after password change on imap server - ### Expected behavior
There should be an option to set the new password.
### Actual behavior
Mail account is not shown anymore in nextcloud. The cronjob gives "OCA\Mail\Exception\ServiceException: IMAP errorMail server denied authentication." Can not access the mail account, so I even can not delete it.
### Mail app
**Mail app version:** 1.4.1
**Mailserver or service:** imap.1und1.de
### Server configuration
**Operating system**: Ubuntu 18.04
**Web server:** Apache 2.4.46
**Database:** MariaDB 10.4.15
**PHP version:** 7.3.23
**Nextcloud Version:** 19.0.4
#### Client configuration
**Browser:** Firefox 82.0
| non_process | can not access mail account in nextcloud after password change on imap server expected behavior there should be an option to set the new password actual behavior mail account is not shown anymore in nextcloud the cronjob gives oca mail exception serviceexception imap errormail server denied authentication can not access the mail account so i even can not delete it mail app mail app version mailserver or service imap de server configuration operating system ubuntu web server apache database mariadb php version nextcloud version client configuration browser firefox | 0 |
8,643 | 7,349,176,814 | IssuesEvent | 2018-03-08 09:45:59 | vector-im/riot-android | https://api.github.com/repos/vector-im/riot-android | closed | Embedded images must be restricted to mxc: URIs | P1 bug security | Right now img tags are allowed arbitrary linking, which is a security issue! | True | Embedded images must be restricted to mxc: URIs - Right now img tags are allowed arbitrary linking, which is a security issue! | non_process | embedded images must be restricted to mxc uris right now img tags are allowed arbitrary linking which is a security issue | 0 |
9,722 | 12,717,208,300 | IssuesEvent | 2020-06-24 04:24:47 | kubeflow/pipelines | https://api.github.com/repos/kubeflow/pipelines | closed | Jest Snapshot Test didn't well handle time string | area/frontend kind/process lifecycle/stale needs investigation priority/p2 status/triaged | In our codes we use Date.toLocalDateString which depends on local time format. It may generate different snapshot. | 1.0 | Jest Snapshot Test didn't well handle time string - In our codes we use Date.toLocalDateString which depends on local time format. It may generate different snapshot. | process | jest snapshot test didn t well handle time string in our codes we use date tolocaldatestring which depends on local time format it may generate different snapshot | 1 |
18,401 | 4,266,248,177 | IssuesEvent | 2016-07-12 14:02:31 | morepath/morepath | https://api.github.com/repos/morepath/morepath | closed | update the morepath example applications to follow cookiecutter-style setup | documentation entry level help wanted | Now that we have a cookiecutter setup we should update the example applications to follow a similar setup, with install instructions in README.txt. The sample applications right now use buildout and depend on development versions of Morepath, but that's not really useful anymore. We should also retire any example applications that have served their purpose.
Spun off from #404 | 1.0 | update the morepath example applications to follow cookiecutter-style setup - Now that we have a cookiecutter setup we should update the example applications to follow a similar setup, with install instructions in README.txt. The sample applications right now use buildout and depend on development versions of Morepath, but that's not really useful anymore. We should also retire any example applications that have served their purpose.
Spun off from #404 | non_process | update the morepath example applications to follow cookiecutter style setup now that we have a cookiecutter setup we should update the example applications to follow a similar setup with install instructions in readme txt the sample applications right now use buildout and depend on development versions of morepath but that s not really useful anymore we should also retire any example applications that have served their purpose spun off from | 0 |
88 | 2,534,358,013 | IssuesEvent | 2015-01-24 21:44:41 | rhattersley/docbook2asciidoc | https://api.github.com/repos/rhattersley/docbook2asciidoc | opened | Remove formatting from subtitle | pre-process | ```xml
<subtitle>Version 1.7.2 <emphasis role="bold">DRAFT</emphasis>, 28 March,
2014</subtitle>
```
converts to:
```
Version 1.7.2 **DRAFT**, 28 March, 2014
```
which renders the "*" characters literally. | 1.0 | Remove formatting from subtitle - ```xml
<subtitle>Version 1.7.2 <emphasis role="bold">DRAFT</emphasis>, 28 March,
2014</subtitle>
```
converts to:
```
Version 1.7.2 **DRAFT**, 28 March, 2014
```
which renders the "*" characters literally. | process | remove formatting from subtitle xml version draft march converts to version draft march which renders the characters literally | 1 |
11,889 | 14,682,445,293 | IssuesEvent | 2020-12-31 16:27:46 | darktable-org/darktable | https://api.github.com/repos/darktable-org/darktable | closed | fix to #6734 breaks pipe with multi-instanciated modules | bug: pending difficulty: average priority: high scope: image processing | **Describe the bug**
<!-- A clear and concise description of what the bug is. -->
The issue #6734 has been fixed in darktable 3.4 but the fix introduced a new issue when there are multiple instances of the same module in the development pipe. The photos is more or less scrambled and you can't access anymore to the extra instances of the multi-instanced module.
N.B. **If there is no multi-instanciated module in the pipe everything seems to be OK**
**To Reproduce**
Same thing as for #6734 except the result is different.
1. Edit a photo with for instance two instances of the exposure module
1. Go to 'Light Table'
2. Select the photo
3. Apply a style (I tested with a style which adds a watermark)
4. See that the thumbnail is correctly updated
5. Press Ctrl-Z to "undo"
6. See that the thumbnail is incorrectly updated (it is more or less scrambled, sometimes looking a little bit like a raw under/overexposure warning, sometimes looking like heavy noise).
7. Go to 'Dark Room'
8. See that there is a lot of "garbage" as in the thumbnail
9. See that the second instance of the exposure module is no more in the pipe, but your history stack is not affected
10. See that Darktable throws this kind of error message to stderr :
```
cannot get iop-order for exposure instance 1
[dt_ioppr_check_iop_order] history module not used but enabled!! exposure 1(2147483647) image 8747 (dt_dev_read_history_no_image end)
```
**Expected behavior**
No garbage
Second instance of the exposure module kept in the pipe
**Screenshots**
<!-- If applicable, add screenshots to help explain your problem. -->

**Platform (please complete the following information):**
* Darktable Version: [3.4.0]
* OS: [ Ubuntu 20.04 with KDE 5 ]
* OpenCL not activated
* Graphics card : Cape Verde XT [Radeon HD 7770/8760 / R7 250X]
**Additional context**
I experimented it with two instances of exposure module and with two instances of color balance module.
The xmp sidecar file seems to be OK but removing the photo from the database and re-importing it with its xmp file gets the same garbage.
The only way to partially recover your work is to go back in the history stack before the first duplicate instance of whatever module is duplicated. Or if you have important settings after that to manually edit your xmp file and remove what is related to multiple instances of modules (keep only the first instance). Then re-import it | 1.0 | fix to #6734 breaks pipe with multi-instanciated modules - **Describe the bug**
<!-- A clear and concise description of what the bug is. -->
The issue #6734 has been fixed in darktable 3.4 but the fix introduced a new issue when there are multiple instances of the same module in the development pipe. The photos is more or less scrambled and you can't access anymore to the extra instances of the multi-instanced module.
N.B. **If there is no multi-instanciated module in the pipe everything seems to be OK**
**To Reproduce**
Same thing as for #6734 except the result is different.
1. Edit a photo with for instance two instances of the exposure module
1. Go to 'Light Table'
2. Select the photo
3. Apply a style (I tested with a style which adds a watermark)
4. See that the thumbnail is correctly updated
5. Press Ctrl-Z to "undo"
6. See that the thumbnail is incorrectly updated (it is more or less scrambled, sometimes looking a little bit like a raw under/overexposure warning, sometimes looking like heavy noise).
7. Go to 'Dark Room'
8. See that there is a lot of "garbage" as in the thumbnail
9. See that the second instance of the exposure module is no more in the pipe, but your history stack is not affected
10. See that Darktable throws this kind of error message to stderr :
```
cannot get iop-order for exposure instance 1
[dt_ioppr_check_iop_order] history module not used but enabled!! exposure 1(2147483647) image 8747 (dt_dev_read_history_no_image end)
```
**Expected behavior**
No garbage
Second instance of the exposure module kept in the pipe
**Screenshots**
<!-- If applicable, add screenshots to help explain your problem. -->

**Platform (please complete the following information):**
* Darktable Version: [3.4.0]
* OS: [ Ubuntu 20.04 with KDE 5 ]
* OpenCL not activated
* Graphics card : Cape Verde XT [Radeon HD 7770/8760 / R7 250X]
**Additional context**
I experimented it with two instances of exposure module and with two instances of color balance module.
The xmp sidecar file seems to be OK but removing the photo from the database and re-importing it with its xmp file gets the same garbage.
The only way to partially recover your work is to go back in the history stack before the first duplicate instance of whatever module is duplicated. Or if you have important settings after that to manually edit your xmp file and remove what is related to multiple instances of modules (keep only the first instance). Then re-import it | process | fix to breaks pipe with multi instanciated modules describe the bug the issue has been fixed in darktable but the fix introduced a new issue when there are multiple instances of the same module in the development pipe the photos is more or less scrambled and you can t access anymore to the extra instances of the multi instanced module n b if there is no multi instanciated module in the pipe everything seems to be ok to reproduce same thing as for except the result is different edit a photo with for instance two instances of the exposure module go to light table select the photo apply a style i tested with a style which adds a watermark see that the thumbnail is correctly updated press ctrl z to undo see that the thumbnail is incorrectly updated it is more or less scrambled sometimes looking a little bit like a raw under overexposure warning sometimes looking like heavy noise go to dark room see that there is a lot of garbage as in the thumbnail see that the second instance of the exposure module is no more in the pipe but your history stack is not affected see that darktable throws this kind of error message to stderr cannot get iop order for exposure instance history module not used but enabled exposure image dt dev read history no image end expected behavior no garbage second instance of the exposure module kept in the pipe screenshots platform please complete the following information darktable version os opencl not activated graphics card cape verde xt additional context i experimented it with two instances of exposure module and with two instances of color balance module the xmp sidecar file seems to be ok but removing the photo from the database and re importing it with its xmp file gets the same garbage the only way to partially recover your work is to go back in the history stack before the first duplicate instance of whatever module is duplicated or if you have important settings after that to manually edit your xmp file and remove what is related to multiple instances of modules keep only the first instance then re import it | 1 |
53,648 | 13,262,047,358 | IssuesEvent | 2020-08-20 21:00:31 | icecube-trac/tix4 | https://api.github.com/repos/icecube-trac/tix4 | closed | [ppc] clang 3.8 errors (Trac #1809) | Migrated from Trac combo simulation defect | These errors were found compiling with clang 3.8:
```text
In file included from /scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/ppc.cxx:696:
/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:22:7: error: no member named 'w' in 'cl_float4'
p.n.w=type>0?-int(type):-128;
~~~ ^
/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:75:7: error: no member named 'w' in 'cl_float4'
p.n.w=0, p.f=0;
~~~ ^
/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:171:7: error: no member named 'w' in 'cl_float4'
p.n.w=dr;
~~~ ^
/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:343:7: error: no member named 'w' in 'cl_float4'
p.r.w=t; p.r.x=rx; p.r.y=ry; p.r.z=rz;
~~~ ^
/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:343:16: error: no member named 'x' in 'cl_float4'
p.r.w=t; p.r.x=rx; p.r.y=ry; p.r.z=rz;
~~~ ^
/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:343:26: error: no member named 'y' in 'cl_float4'
p.r.w=t; p.r.x=rx; p.r.y=ry; p.r.z=rz;
~~~ ^
/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:343:36: error: no member named 'z' in 'cl_float4'
p.r.w=t; p.r.x=rx; p.r.y=ry; p.r.z=rz;
~~~ ^
/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:379:17: error: no member named 'x' in 'cl_float4'
p.q=flne; p.n.x=nx; p.n.y=ny; p.n.z=nz;
~~~ ^
/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:379:27: error: no member named 'y' in 'cl_float4'
p.q=flne; p.n.x=nx; p.n.y=ny; p.n.z=nz;
~~~ ^
/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:379:37: error: no member named 'z' in 'cl_float4'
p.q=flne; p.n.x=nx; p.n.y=ny; p.n.z=nz;
~~~ ^
10 errors generated.
```
This post may be related:
http://stackoverflow.com/questions/10979487/opencl-cl-datatypes-arithmetic
Note that clang defaults to c++14, which probably disables the macro `__GNUC__`.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1809">https://code.icecube.wisc.edu/projects/icecube/ticket/1809</a>, reported by david.schultzand owned by dima</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:12:38",
"_ts": "1550067158057333",
"description": "These errors were found compiling with clang 3.8:\n\n{{{\nIn file included from /scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/ppc.cxx:696:\n/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:22:7: error: no member named 'w' in 'cl_float4'\np.n.w=type>0?-int(type):-128;\n ~~~ ^\n/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:75:7: error: no member named 'w' in 'cl_float4'\np.n.w=0, p.f=0;\n ~~~ ^\n/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:171:7: error: no member named 'w' in 'cl_float4'\np.n.w=dr;\n ~~~ ^\n/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:343:7: error: no member named 'w' in 'cl_float4'\np.r.w=t; p.r.x=rx; p.r.y=ry; p.r.z=rz;\n ~~~ ^\n/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:343:16: error: no member named 'x' in 'cl_float4'\np.r.w=t; p.r.x=rx; p.r.y=ry; p.r.z=rz;\n ~~~ ^\n/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:343:26: error: no member named 'y' in 'cl_float4'\np.r.w=t; p.r.x=rx; p.r.y=ry; p.r.z=rz;\n ~~~ ^\n/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:343:36: error: no member named 'z' in 'cl_float4'\np.r.w=t; p.r.x=rx; p.r.y=ry; p.r.z=rz;\n ~~~ ^\n/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:379:17: error: no member named 'x' in 'cl_float4'\np.q=flne; p.n.x=nx; p.n.y=ny; p.n.z=nz;\n ~~~ ^\n/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:379:27: error: no member named 'y' in 'cl_float4'\np.q=flne; p.n.x=nx; p.n.y=ny; p.n.z=nz;\n ~~~ ^\n/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:379:37: error: no member named 'z' in 'cl_float4'\np.q=flne; p.n.x=nx; p.n.y=ny; p.n.z=nz;\n ~~~ ^\n10 errors generated.\n}}}\n\nThis post may be related:\nhttp://stackoverflow.com/questions/10979487/opencl-cl-datatypes-arithmetic\n\nNote that clang defaults to c++14, which probably disables the macro `__GNUC__`.",
"reporter": "david.schultz",
"cc": "olivas",
"resolution": "fixed",
"time": "2016-07-29T20:44:11",
"component": "combo simulation",
"summary": "[ppc] clang 3.8 errors",
"priority": "major",
"keywords": "",
"milestone": "",
"owner": "dima",
"type": "defect"
}
```
</p>
</details>
| 1.0 | [ppc] clang 3.8 errors (Trac #1809) - These errors were found compiling with clang 3.8:
```text
In file included from /scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/ppc.cxx:696:
/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:22:7: error: no member named 'w' in 'cl_float4'
p.n.w=type>0?-int(type):-128;
~~~ ^
/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:75:7: error: no member named 'w' in 'cl_float4'
p.n.w=0, p.f=0;
~~~ ^
/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:171:7: error: no member named 'w' in 'cl_float4'
p.n.w=dr;
~~~ ^
/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:343:7: error: no member named 'w' in 'cl_float4'
p.r.w=t; p.r.x=rx; p.r.y=ry; p.r.z=rz;
~~~ ^
/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:343:16: error: no member named 'x' in 'cl_float4'
p.r.w=t; p.r.x=rx; p.r.y=ry; p.r.z=rz;
~~~ ^
/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:343:26: error: no member named 'y' in 'cl_float4'
p.r.w=t; p.r.x=rx; p.r.y=ry; p.r.z=rz;
~~~ ^
/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:343:36: error: no member named 'z' in 'cl_float4'
p.r.w=t; p.r.x=rx; p.r.y=ry; p.r.z=rz;
~~~ ^
/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:379:17: error: no member named 'x' in 'cl_float4'
p.q=flne; p.n.x=nx; p.n.y=ny; p.n.z=nz;
~~~ ^
/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:379:27: error: no member named 'y' in 'cl_float4'
p.q=flne; p.n.x=nx; p.n.y=ny; p.n.z=nz;
~~~ ^
/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:379:37: error: no member named 'z' in 'cl_float4'
p.q=flne; p.n.x=nx; p.n.y=ny; p.n.z=nz;
~~~ ^
10 errors generated.
```
This post may be related:
http://stackoverflow.com/questions/10979487/opencl-cl-datatypes-arithmetic
Note that clang defaults to c++14, which probably disables the macro `__GNUC__`.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1809">https://code.icecube.wisc.edu/projects/icecube/ticket/1809</a>, reported by david.schultzand owned by dima</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:12:38",
"_ts": "1550067158057333",
"description": "These errors were found compiling with clang 3.8:\n\n{{{\nIn file included from /scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/ppc.cxx:696:\n/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:22:7: error: no member named 'w' in 'cl_float4'\np.n.w=type>0?-int(type):-128;\n ~~~ ^\n/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:75:7: error: no member named 'w' in 'cl_float4'\np.n.w=0, p.f=0;\n ~~~ ^\n/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:171:7: error: no member named 'w' in 'cl_float4'\np.n.w=dr;\n ~~~ ^\n/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:343:7: error: no member named 'w' in 'cl_float4'\np.r.w=t; p.r.x=rx; p.r.y=ry; p.r.z=rz;\n ~~~ ^\n/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:343:16: error: no member named 'x' in 'cl_float4'\np.r.w=t; p.r.x=rx; p.r.y=ry; p.r.z=rz;\n ~~~ ^\n/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:343:26: error: no member named 'y' in 'cl_float4'\np.r.w=t; p.r.x=rx; p.r.y=ry; p.r.z=rz;\n ~~~ ^\n/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:343:36: error: no member named 'z' in 'cl_float4'\np.r.w=t; p.r.x=rx; p.r.y=ry; p.r.z=rz;\n ~~~ ^\n/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:379:17: error: no member named 'x' in 'cl_float4'\np.q=flne; p.n.x=nx; p.n.y=ny; p.n.z=nz;\n ~~~ ^\n/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:379:27: error: no member named 'y' in 'cl_float4'\np.q=flne; p.n.x=nx; p.n.y=ny; p.n.z=nz;\n ~~~ ^\n/scratch/dschultz/icetray_profiling/tmptua13z/src/ppc/private/ppc/ocl/f2k.cxx:379:37: error: no member named 'z' in 'cl_float4'\np.q=flne; p.n.x=nx; p.n.y=ny; p.n.z=nz;\n ~~~ ^\n10 errors generated.\n}}}\n\nThis post may be related:\nhttp://stackoverflow.com/questions/10979487/opencl-cl-datatypes-arithmetic\n\nNote that clang defaults to c++14, which probably disables the macro `__GNUC__`.",
"reporter": "david.schultz",
"cc": "olivas",
"resolution": "fixed",
"time": "2016-07-29T20:44:11",
"component": "combo simulation",
"summary": "[ppc] clang 3.8 errors",
"priority": "major",
"keywords": "",
"milestone": "",
"owner": "dima",
"type": "defect"
}
```
</p>
</details>
| non_process | clang errors trac these errors were found compiling with clang text in file included from scratch dschultz icetray profiling src ppc private ppc ocl ppc cxx scratch dschultz icetray profiling src ppc private ppc ocl cxx error no member named w in cl p n w type int type scratch dschultz icetray profiling src ppc private ppc ocl cxx error no member named w in cl p n w p f scratch dschultz icetray profiling src ppc private ppc ocl cxx error no member named w in cl p n w dr scratch dschultz icetray profiling src ppc private ppc ocl cxx error no member named w in cl p r w t p r x rx p r y ry p r z rz scratch dschultz icetray profiling src ppc private ppc ocl cxx error no member named x in cl p r w t p r x rx p r y ry p r z rz scratch dschultz icetray profiling src ppc private ppc ocl cxx error no member named y in cl p r w t p r x rx p r y ry p r z rz scratch dschultz icetray profiling src ppc private ppc ocl cxx error no member named z in cl p r w t p r x rx p r y ry p r z rz scratch dschultz icetray profiling src ppc private ppc ocl cxx error no member named x in cl p q flne p n x nx p n y ny p n z nz scratch dschultz icetray profiling src ppc private ppc ocl cxx error no member named y in cl p q flne p n x nx p n y ny p n z nz scratch dschultz icetray profiling src ppc private ppc ocl cxx error no member named z in cl p q flne p n x nx p n y ny p n z nz errors generated this post may be related note that clang defaults to c which probably disables the macro gnuc migrated from json status closed changetime ts description these errors were found compiling with clang n n nin file included from scratch dschultz icetray profiling src ppc private ppc ocl ppc cxx n scratch dschultz icetray profiling src ppc private ppc ocl cxx error no member named w in cl np n w type int type n n scratch dschultz icetray profiling src ppc private ppc ocl cxx error no member named w in cl np n w p f n n scratch dschultz icetray profiling src ppc private ppc ocl cxx error no member named w in cl np n w dr n n scratch dschultz icetray profiling src ppc private ppc ocl cxx error no member named w in cl np r w t p r x rx p r y ry p r z rz n n scratch dschultz icetray profiling src ppc private ppc ocl cxx error no member named x in cl np r w t p r x rx p r y ry p r z rz n n scratch dschultz icetray profiling src ppc private ppc ocl cxx error no member named y in cl np r w t p r x rx p r y ry p r z rz n n scratch dschultz icetray profiling src ppc private ppc ocl cxx error no member named z in cl np r w t p r x rx p r y ry p r z rz n n scratch dschultz icetray profiling src ppc private ppc ocl cxx error no member named x in cl np q flne p n x nx p n y ny p n z nz n n scratch dschultz icetray profiling src ppc private ppc ocl cxx error no member named y in cl np q flne p n x nx p n y ny p n z nz n n scratch dschultz icetray profiling src ppc private ppc ocl cxx error no member named z in cl np q flne p n x nx p n y ny p n z nz n errors generated n n nthis post may be related n that clang defaults to c which probably disables the macro gnuc reporter david schultz cc olivas resolution fixed time component combo simulation summary clang errors priority major keywords milestone owner dima type defect | 0 |
91,939 | 18,755,813,031 | IssuesEvent | 2021-11-05 10:35:12 | julia-vscode/julia-vscode | https://api.github.com/repos/julia-vscode/julia-vscode | closed | entering new line clears the in-line evaluation popup cell | bug area-code-execution | MWE.
type
```
rand()
```
press ALT+ENTER. A popup appears.
Then press ENTER to start a new line and write whatever other command. The popup to the right of `rand` has dissapeared.
This isn't nice for interactively writing scripts. | 1.0 | entering new line clears the in-line evaluation popup cell - MWE.
type
```
rand()
```
press ALT+ENTER. A popup appears.
Then press ENTER to start a new line and write whatever other command. The popup to the right of `rand` has dissapeared.
This isn't nice for interactively writing scripts. | non_process | entering new line clears the in line evaluation popup cell mwe type rand press alt enter a popup appears then press enter to start a new line and write whatever other command the popup to the right of rand has dissapeared this isn t nice for interactively writing scripts | 0 |
228,362 | 18,172,119,282 | IssuesEvent | 2021-09-27 21:19:03 | microsoft/vscode | https://api.github.com/repos/microsoft/vscode | opened | Test: Outline button in Jupyter notebooks | testplan-item | Refs: https://github.com/microsoft/vscode-jupyter/issues/7305
- [ ] anyOS
- [ ] anyOS
Complexity: 1
Author: @IanMatthewHuff
---
File bugs on the Jupyter repo here: https://github.com/microsoft/vscode-jupyter/issues
Jupyter notebook users have been continually asking for Table of Contents control in their notebooks, and this functionality is currently provided by the VS Code Outline window. However, Jupyter users are not finding this control, and we are still getting issues and upvotes from users about a missing Table of Contents control. To alleviate this, a button has been added to the Jupyter notebook toolbar to focus the outline window to point users to this.
Testing:
1. Jupyter extension installed
2. Open up a .ipynb file
3. For viewing a useful outline it works best with some markdown cells with # header blocks in them (as they go into the outline view)
4. On the toolbar there should be an Outline button next to the variables button

- [ ] If the VS Code Outline view is collapsed or offscreen the button should be enabled. Pressing the button should focus the outline view.
- [ ] If the VS Code Outline view is onscreen and expanded the button should be disabled.
- [ ] "Jupyter: Show Table Of Contents (Outline View)" command should be available for the same function also when the view is offscreen or unavailable. The naming here is specifically for notebook customers searching for table of contents in the command list.

| 1.0 | Test: Outline button in Jupyter notebooks - Refs: https://github.com/microsoft/vscode-jupyter/issues/7305
- [ ] anyOS
- [ ] anyOS
Complexity: 1
Author: @IanMatthewHuff
---
File bugs on the Jupyter repo here: https://github.com/microsoft/vscode-jupyter/issues
Jupyter notebook users have been continually asking for Table of Contents control in their notebooks, and this functionality is currently provided by the VS Code Outline window. However, Jupyter users are not finding this control, and we are still getting issues and upvotes from users about a missing Table of Contents control. To alleviate this, a button has been added to the Jupyter notebook toolbar to focus the outline window to point users to this.
Testing:
1. Jupyter extension installed
2. Open up a .ipynb file
3. For viewing a useful outline it works best with some markdown cells with # header blocks in them (as they go into the outline view)
4. On the toolbar there should be an Outline button next to the variables button

- [ ] If the VS Code Outline view is collapsed or offscreen the button should be enabled. Pressing the button should focus the outline view.
- [ ] If the VS Code Outline view is onscreen and expanded the button should be disabled.
- [ ] "Jupyter: Show Table Of Contents (Outline View)" command should be available for the same function also when the view is offscreen or unavailable. The naming here is specifically for notebook customers searching for table of contents in the command list.

| non_process | test outline button in jupyter notebooks refs anyos anyos complexity author ianmatthewhuff file bugs on the jupyter repo here jupyter notebook users have been continually asking for table of contents control in their notebooks and this functionality is currently provided by the vs code outline window however jupyter users are not finding this control and we are still getting issues and upvotes from users about a missing table of contents control to alleviate this a button has been added to the jupyter notebook toolbar to focus the outline window to point users to this testing jupyter extension installed open up a ipynb file for viewing a useful outline it works best with some markdown cells with header blocks in them as they go into the outline view on the toolbar there should be an outline button next to the variables button if the vs code outline view is collapsed or offscreen the button should be enabled pressing the button should focus the outline view if the vs code outline view is onscreen and expanded the button should be disabled jupyter show table of contents outline view command should be available for the same function also when the view is offscreen or unavailable the naming here is specifically for notebook customers searching for table of contents in the command list | 0 |
12,807 | 15,184,213,153 | IssuesEvent | 2021-02-15 09:15:53 | topcoder-platform/community-app | https://api.github.com/repos/topcoder-platform/community-app | opened | Member skills are not getting updated when a member wins a challenge | BE P1 ShapeupProcess challenge- recommender-tool member-skill-extractor | The member skill extractor is not updating the members skills when a member is placed in a challenge.
Example:
member: tester1234
challenge: https://www.topcoder-dev.com/challenges/cfc3f821-64e4-4585-8fdc-744928d2bc9f
the member_skills_history table is not updated with the skills for the this member
cc @lakshmiathreya | 1.0 | Member skills are not getting updated when a member wins a challenge - The member skill extractor is not updating the members skills when a member is placed in a challenge.
Example:
member: tester1234
challenge: https://www.topcoder-dev.com/challenges/cfc3f821-64e4-4585-8fdc-744928d2bc9f
the member_skills_history table is not updated with the skills for the this member
cc @lakshmiathreya | process | member skills are not getting updated when a member wins a challenge the member skill extractor is not updating the members skills when a member is placed in a challenge example member challenge the member skills history table is not updated with the skills for the this member cc lakshmiathreya | 1 |
14,127 | 17,020,220,057 | IssuesEvent | 2021-07-02 17:43:18 | darktable-org/darktable | https://api.github.com/repos/darktable-org/darktable | closed | Perspective module: cannot set the automatic cropping value untile some perspective correction is set | bug: pending scope: image processing | If I try to set automatic cropping to some value which is not off before changing the rotation, lens shift, etc. values, I get an error message on the image "automatic cropping failed".
| 1.0 | Perspective module: cannot set the automatic cropping value untile some perspective correction is set - If I try to set automatic cropping to some value which is not off before changing the rotation, lens shift, etc. values, I get an error message on the image "automatic cropping failed".
| process | perspective module cannot set the automatic cropping value untile some perspective correction is set if i try to set automatic cropping to some value which is not off before changing the rotation lens shift etc values i get an error message on the image automatic cropping failed | 1 |
1,754 | 4,460,688,074 | IssuesEvent | 2016-08-24 00:46:52 | dotnet/corefx | https://api.github.com/repos/dotnet/corefx | closed | StandardOutput and StandardError receive an extra line | System.Diagnostics.Process | On Linux, compared to other languages, say Python, .NET is appending (somehow) an extra line on redirected StandardOutput and StandardError.
For example, if I execute the following script:
```bash
#!/bin/sh
exit 42
```
In Python (e.g., subprocess.check_output) I receive, as I would expect, an empty string. In .NET, however, if I redirect through ProcessStartInfo, I receive an event in my OutputDataReceived handler with an empty string (and the my code cannot reject empty string out-of-hand since it may be a legal part of output). | 1.0 | StandardOutput and StandardError receive an extra line - On Linux, compared to other languages, say Python, .NET is appending (somehow) an extra line on redirected StandardOutput and StandardError.
For example, if I execute the following script:
```bash
#!/bin/sh
exit 42
```
In Python (e.g., subprocess.check_output) I receive, as I would expect, an empty string. In .NET, however, if I redirect through ProcessStartInfo, I receive an event in my OutputDataReceived handler with an empty string (and the my code cannot reject empty string out-of-hand since it may be a legal part of output). | process | standardoutput and standarderror receive an extra line on linux compared to other languages say python net is appending somehow an extra line on redirected standardoutput and standarderror for example if i execute the following script bash bin sh exit in python e g subprocess check output i receive as i would expect an empty string in net however if i redirect through processstartinfo i receive an event in my outputdatareceived handler with an empty string and the my code cannot reject empty string out of hand since it may be a legal part of output | 1 |
75,349 | 9,850,986,273 | IssuesEvent | 2019-06-19 09:24:59 | mlr-org/mlr | https://api.github.com/repos/mlr-org/mlr | opened | `mRMRe::mrmr` differs from `praznik::mrmr` | type-documentation | I don't think we can do anything about this but it is important to know.
Note that the absolute values are not of interest here but only the ranking of the features.
``` r
suppressPackageStartupMessages(library(mlr))
library(magrittr)
bh.task = dropFeatures(bh.task, "chas")
fv_mrmr = generateFilterValuesData(bh.task, "mrmr")
fv_mrmr_praznik = generateFilterValuesData(bh.task, "praznik_MRMR")
purrr::map(list(fv_mrmr$data, fv_mrmr_praznik$data), ~
dplyr::arrange(.x, method, desc(value))) %>%
print()
#> [[1]]
#> # A tibble: 12 x 4
#> name type method value
#> <chr> <chr> <chr> <dbl>
#> 1 lstat numeric mrmr 0.393
#> 2 rm numeric mrmr 0.0940
#> 3 ptratio numeric mrmr 0.0776
#> 4 b numeric mrmr 0.0269
#> 5 indus numeric mrmr 0.0189
#> 6 crim numeric mrmr 0.0106
#> 7 zn numeric mrmr -0.00239
#> 8 tax numeric mrmr -0.0295
#> 9 age numeric mrmr -0.0495
#> 10 nox numeric mrmr -0.0911
#> 11 rad numeric mrmr -0.135
#> 12 dis numeric mrmr -0.160
#>
#> [[2]]
#> # A tibble: 12 x 4
#> name type method value
#> <chr> <chr> <chr> <dbl>
#> 1 lstat numeric praznik_MRMR 1
#> 2 ptratio numeric praznik_MRMR 0.917
#> 3 rm numeric praznik_MRMR 0.833
#> 4 crim numeric praznik_MRMR 0.75
#> 5 age numeric praznik_MRMR 0.667
#> 6 b numeric praznik_MRMR 0.583
#> 7 nox numeric praznik_MRMR 0.5
#> 8 zn numeric praznik_MRMR 0.417
#> 9 tax numeric praznik_MRMR 0.333
#> 10 rad numeric praznik_MRMR 0.25
#> 11 dis numeric praznik_MRMR 0.167
#> 12 indus numeric praznik_MRMR 0.0833
```
<sup>Created on 2019-06-19 by the [reprex package](https://reprex.tidyverse.org) (v0.3.0)</sup>
In addition, here is a runtime comparison for a dataset with ~ 7k features:
```
build_times(fv_nri_praznik_mrmr, fv_nri_mrmr)
# A tibble: 2 x 4
target elapsed user system
<chr> <S4: Duration> <S4: Duration> <S4: Duration>
1 fv_nri_mrmr 6487s (~1.8 hours) 6495s (~1.8 hours) 0.111s
2 fv_nri_praznik_mrmr 144s (~2.4 minutes) 290s (~4.83 minutes) 0.144s
``` | 1.0 | `mRMRe::mrmr` differs from `praznik::mrmr` - I don't think we can do anything about this but it is important to know.
Note that the absolute values are not of interest here but only the ranking of the features.
``` r
suppressPackageStartupMessages(library(mlr))
library(magrittr)
bh.task = dropFeatures(bh.task, "chas")
fv_mrmr = generateFilterValuesData(bh.task, "mrmr")
fv_mrmr_praznik = generateFilterValuesData(bh.task, "praznik_MRMR")
purrr::map(list(fv_mrmr$data, fv_mrmr_praznik$data), ~
dplyr::arrange(.x, method, desc(value))) %>%
print()
#> [[1]]
#> # A tibble: 12 x 4
#> name type method value
#> <chr> <chr> <chr> <dbl>
#> 1 lstat numeric mrmr 0.393
#> 2 rm numeric mrmr 0.0940
#> 3 ptratio numeric mrmr 0.0776
#> 4 b numeric mrmr 0.0269
#> 5 indus numeric mrmr 0.0189
#> 6 crim numeric mrmr 0.0106
#> 7 zn numeric mrmr -0.00239
#> 8 tax numeric mrmr -0.0295
#> 9 age numeric mrmr -0.0495
#> 10 nox numeric mrmr -0.0911
#> 11 rad numeric mrmr -0.135
#> 12 dis numeric mrmr -0.160
#>
#> [[2]]
#> # A tibble: 12 x 4
#> name type method value
#> <chr> <chr> <chr> <dbl>
#> 1 lstat numeric praznik_MRMR 1
#> 2 ptratio numeric praznik_MRMR 0.917
#> 3 rm numeric praznik_MRMR 0.833
#> 4 crim numeric praznik_MRMR 0.75
#> 5 age numeric praznik_MRMR 0.667
#> 6 b numeric praznik_MRMR 0.583
#> 7 nox numeric praznik_MRMR 0.5
#> 8 zn numeric praznik_MRMR 0.417
#> 9 tax numeric praznik_MRMR 0.333
#> 10 rad numeric praznik_MRMR 0.25
#> 11 dis numeric praznik_MRMR 0.167
#> 12 indus numeric praznik_MRMR 0.0833
```
<sup>Created on 2019-06-19 by the [reprex package](https://reprex.tidyverse.org) (v0.3.0)</sup>
In addition, here is a runtime comparison for a dataset with ~ 7k features:
```
build_times(fv_nri_praznik_mrmr, fv_nri_mrmr)
# A tibble: 2 x 4
target elapsed user system
<chr> <S4: Duration> <S4: Duration> <S4: Duration>
1 fv_nri_mrmr 6487s (~1.8 hours) 6495s (~1.8 hours) 0.111s
2 fv_nri_praznik_mrmr 144s (~2.4 minutes) 290s (~4.83 minutes) 0.144s
``` | non_process | mrmre mrmr differs from praznik mrmr i don t think we can do anything about this but it is important to know note that the absolute values are not of interest here but only the ranking of the features r suppresspackagestartupmessages library mlr library magrittr bh task dropfeatures bh task chas fv mrmr generatefiltervaluesdata bh task mrmr fv mrmr praznik generatefiltervaluesdata bh task praznik mrmr purrr map list fv mrmr data fv mrmr praznik data dplyr arrange x method desc value print a tibble x name type method value lstat numeric mrmr rm numeric mrmr ptratio numeric mrmr b numeric mrmr indus numeric mrmr crim numeric mrmr zn numeric mrmr tax numeric mrmr age numeric mrmr nox numeric mrmr rad numeric mrmr dis numeric mrmr a tibble x name type method value lstat numeric praznik mrmr ptratio numeric praznik mrmr rm numeric praznik mrmr crim numeric praznik mrmr age numeric praznik mrmr b numeric praznik mrmr nox numeric praznik mrmr zn numeric praznik mrmr tax numeric praznik mrmr rad numeric praznik mrmr dis numeric praznik mrmr indus numeric praznik mrmr created on by the in addition here is a runtime comparison for a dataset with features build times fv nri praznik mrmr fv nri mrmr a tibble x target elapsed user system fv nri mrmr hours hours fv nri praznik mrmr minutes minutes | 0 |
273,983 | 20,821,837,286 | IssuesEvent | 2022-03-18 16:07:35 | sjefferson99/Boatman-pico-uart-hub | https://api.github.com/repos/sjefferson99/Boatman-pico-uart-hub | closed | Improve documentation in bmserial module | documentation | The bmserial module should document that this is for the serial comms including a definition of the data structure. Functions should be boatman module agnostic and parse data to be passed on to python modules for the appropriate boatman module. | 1.0 | Improve documentation in bmserial module - The bmserial module should document that this is for the serial comms including a definition of the data structure. Functions should be boatman module agnostic and parse data to be passed on to python modules for the appropriate boatman module. | non_process | improve documentation in bmserial module the bmserial module should document that this is for the serial comms including a definition of the data structure functions should be boatman module agnostic and parse data to be passed on to python modules for the appropriate boatman module | 0 |
34,681 | 7,853,763,386 | IssuesEvent | 2018-06-20 18:29:10 | kobotoolbox/kpi | https://api.github.com/repos/kobotoolbox/kpi | closed | Opening tags moves project details up | bug coded low priority ui | Minor one, but irritating - after opening tags editor, the project details (name, date, etc.) move up by few pixels; plus the amount of whitespace between name and tags editor is too big:
<img width="1005" alt="screen shot 2018-06-09 at 15 40 47" src="https://user-images.githubusercontent.com/2521888/41192318-24bf0e02-6bfc-11e8-9224-c68df359ae9e.png">
<img width="998" alt="screen shot 2018-06-09 at 15 41 01" src="https://user-images.githubusercontent.com/2521888/41192319-24e1c55a-6bfc-11e8-8895-129c1cfa9a88.png">
PS. While doing this, verify if this function could be deleted - from `tagInput.es6`:
```es6
handleChange(tags) {
var transformed = tags.map(function(tag) {
// Behavior should match KpiTaggableManager.add()
return tag.trim().replace(/ /g, '-');
});
this.setState({tags: transformed});
var uid = this.props.uid;
actions.resources.updateAsset(uid, {
tag_string: transformed.join(',')
});
}
``` | 1.0 | Opening tags moves project details up - Minor one, but irritating - after opening tags editor, the project details (name, date, etc.) move up by few pixels; plus the amount of whitespace between name and tags editor is too big:
<img width="1005" alt="screen shot 2018-06-09 at 15 40 47" src="https://user-images.githubusercontent.com/2521888/41192318-24bf0e02-6bfc-11e8-9224-c68df359ae9e.png">
<img width="998" alt="screen shot 2018-06-09 at 15 41 01" src="https://user-images.githubusercontent.com/2521888/41192319-24e1c55a-6bfc-11e8-8895-129c1cfa9a88.png">
PS. While doing this, verify if this function could be deleted - from `tagInput.es6`:
```es6
handleChange(tags) {
var transformed = tags.map(function(tag) {
// Behavior should match KpiTaggableManager.add()
return tag.trim().replace(/ /g, '-');
});
this.setState({tags: transformed});
var uid = this.props.uid;
actions.resources.updateAsset(uid, {
tag_string: transformed.join(',')
});
}
``` | non_process | opening tags moves project details up minor one but irritating after opening tags editor the project details name date etc move up by few pixels plus the amount of whitespace between name and tags editor is too big img width alt screen shot at src img width alt screen shot at src ps while doing this verify if this function could be deleted from taginput handlechange tags var transformed tags map function tag behavior should match kpitaggablemanager add return tag trim replace g this setstate tags transformed var uid this props uid actions resources updateasset uid tag string transformed join | 0 |
5,835 | 8,666,148,624 | IssuesEvent | 2018-11-29 02:36:36 | w3c/w3process | https://api.github.com/repos/w3c/w3process | opened | TAG appointment should be via IETF style nomcom | Process2020Candidate | The TAG is currently elected. We could move to an IETF-style NomCom appointment. This ensures the right people for the role are selected and a balanced TAG can be achieved. Comments are welcome.
- The NomCom consists of a random set of volunteers who meet a set criteria - e.g. have chaired a group or published a draft in the last X years. The AB could be members of the NomCom. The NomCom should have advising members. They cannot vote, only advise. The TAG Chair, a Developer Invited Expert would be good choices.
- Nominations for groups go direct to the NomCom. Nominations can come from any source. The NomCom will see via the application what the source of the nomination is - the NomCom can choose to weigh an AC Nomination in a positive manner if they wish. This means that AC nominations count for something, but other groups (e.g. independent developers) are not blocked from applying.
- NomCom sees nominations as they come in. If they feel a diverse pool isn’t being reached (diversity for all items including gender, technical knowledge of particular topics, organisation representation size, global location, etc.) the NomCom will begin pushing for more candidates.
- NomCom starts the process of selecting candidates. This is as follows:
- The NomCom meets before evaluating the candidates. It agrees on key qualities it is looking for in the candidates for the open roles. It is likely these qualities will be the same in each election, although sometimes they will be different (e.g. “Chair of the TAG" will require different qualities to “TAG Member”). Consensus on these qualities is reached.
- All applications are read. Obvious “no” candidates are discarded.
- The NomCom meets to discuss the “possible” candidates remaining. Further candidates are discarded for reasons the NomCom deem appropriate. These could be because there has been an influx of candidates which represent one industry, or many candidates are experts in only one field.
- The NomCom is left with the final set of “possible candidates”. These candidates are interviewed for the role at TPAC, at the AC Meeting or via teleconference. Video Conference interviews are acceptable.
- The NomCom announces the selected candidates. An objection process opens to allow formal objections from the AC. Formal objections will be handled as normal. After 28 days, the formal objection window closes and the newly selected individuals begin their term.
| 1.0 | TAG appointment should be via IETF style nomcom - The TAG is currently elected. We could move to an IETF-style NomCom appointment. This ensures the right people for the role are selected and a balanced TAG can be achieved. Comments are welcome.
- The NomCom consists of a random set of volunteers who meet a set criteria - e.g. have chaired a group or published a draft in the last X years. The AB could be members of the NomCom. The NomCom should have advising members. They cannot vote, only advise. The TAG Chair, a Developer Invited Expert would be good choices.
- Nominations for groups go direct to the NomCom. Nominations can come from any source. The NomCom will see via the application what the source of the nomination is - the NomCom can choose to weigh an AC Nomination in a positive manner if they wish. This means that AC nominations count for something, but other groups (e.g. independent developers) are not blocked from applying.
- NomCom sees nominations as they come in. If they feel a diverse pool isn’t being reached (diversity for all items including gender, technical knowledge of particular topics, organisation representation size, global location, etc.) the NomCom will begin pushing for more candidates.
- NomCom starts the process of selecting candidates. This is as follows:
- The NomCom meets before evaluating the candidates. It agrees on key qualities it is looking for in the candidates for the open roles. It is likely these qualities will be the same in each election, although sometimes they will be different (e.g. “Chair of the TAG" will require different qualities to “TAG Member”). Consensus on these qualities is reached.
- All applications are read. Obvious “no” candidates are discarded.
- The NomCom meets to discuss the “possible” candidates remaining. Further candidates are discarded for reasons the NomCom deem appropriate. These could be because there has been an influx of candidates which represent one industry, or many candidates are experts in only one field.
- The NomCom is left with the final set of “possible candidates”. These candidates are interviewed for the role at TPAC, at the AC Meeting or via teleconference. Video Conference interviews are acceptable.
- The NomCom announces the selected candidates. An objection process opens to allow formal objections from the AC. Formal objections will be handled as normal. After 28 days, the formal objection window closes and the newly selected individuals begin their term.
| process | tag appointment should be via ietf style nomcom the tag is currently elected we could move to an ietf style nomcom appointment this ensures the right people for the role are selected and a balanced tag can be achieved comments are welcome the nomcom consists of a random set of volunteers who meet a set criteria e g have chaired a group or published a draft in the last x years the ab could be members of the nomcom the nomcom should have advising members they cannot vote only advise the tag chair a developer invited expert would be good choices nominations for groups go direct to the nomcom nominations can come from any source the nomcom will see via the application what the source of the nomination is the nomcom can choose to weigh an ac nomination in a positive manner if they wish this means that ac nominations count for something but other groups e g independent developers are not blocked from applying nomcom sees nominations as they come in if they feel a diverse pool isn’t being reached diversity for all items including gender technical knowledge of particular topics organisation representation size global location etc the nomcom will begin pushing for more candidates nomcom starts the process of selecting candidates this is as follows the nomcom meets before evaluating the candidates it agrees on key qualities it is looking for in the candidates for the open roles it is likely these qualities will be the same in each election although sometimes they will be different e g “chair of the tag will require different qualities to “tag member” consensus on these qualities is reached all applications are read obvious “no” candidates are discarded the nomcom meets to discuss the “possible” candidates remaining further candidates are discarded for reasons the nomcom deem appropriate these could be because there has been an influx of candidates which represent one industry or many candidates are experts in only one field the nomcom is left with the final set of “possible candidates” these candidates are interviewed for the role at tpac at the ac meeting or via teleconference video conference interviews are acceptable the nomcom announces the selected candidates an objection process opens to allow formal objections from the ac formal objections will be handled as normal after days the formal objection window closes and the newly selected individuals begin their term | 1 |
201,384 | 15,802,252,754 | IssuesEvent | 2021-04-03 08:49:38 | LJ-37/ped | https://api.github.com/repos/LJ-37/ped | opened | Typo in UG | severity.Medium type.DocumentationBug | Tag for view food category should be f/ but is written as a/ in the UG under view diet data.

<!--session: 1617437380999-4a5bfdbc-3d33-43fd-9397-daa368b5b349--> | 1.0 | Typo in UG - Tag for view food category should be f/ but is written as a/ in the UG under view diet data.

<!--session: 1617437380999-4a5bfdbc-3d33-43fd-9397-daa368b5b349--> | non_process | typo in ug tag for view food category should be f but is written as a in the ug under view diet data | 0 |
15,296 | 2,850,599,479 | IssuesEvent | 2015-05-31 18:21:31 | damonkohler/sl4a | https://api.github.com/repos/damonkohler/sl4a | opened | SL4A Force Close on droid.startActivityIntent(chooserIntent) | auto-migrated Priority-Medium Type-Defect | _From @GoogleCodeExporter on May 31, 2015 11:30_
```
What device(s) are you experiencing the problem on?
Samsung Vibrant (SGH-T959)
What firmware version are you running on the device?
2.2
What steps will reproduce the problem?
1. Run the attached Python script on an Android device with SL4Ar4
-or-
1. Make an intent with action="android.intent.action.SEND", uriType =
"text/plain", extras = {"EXTRA_TEXT":"my message", "EXTRA_SUBJECT":"and my
subject"}, and everything else = None.
2. Make another intent with action = "android.intent.action.CHOOSER", extras =
{"EXTRA_INTENT":theFirstIntent}, and everything else = None.
3. droid.startActivityIntent(yourSecondIntent)
4. SL4A force closes
What is the expected output? What do you see instead?
I'm trying to get a chooser for sharing text via other apps (Gmail, Messaging,
etc.). It should be just like the one from the "Share" option in the menu of
SL4A's script editor activity. Instead, SL4A force closes, and there's no
indication whatsoever to the Python interpreter.
What version of the product are you using? On what operating system?
I'm using SL4Ar4 on Samung's OEM build FROYO.UVKB5
My development machine is Windows 7 64-bit, but running the script directly on
the phone still crashes, so that's probably not the issue.
Please provide any additional information below.
I tried to adapt the code from the "Share" option of the SL4A script editor by
manually making the same intent as Intent.createChooser() as described in the
Android documentation.
My script is adapted from lines 198-202 here:
http://code.google.com/p/android-scripting/source/diff?spec=svndeba523e8a9b44ea6
ef6e2497e21914084b109a8&r=deba523e8a9b44ea6ef6e2497e21914084b109a8&format=side&p
ath=/android/ScriptingLayerForAndroid/src/com/googlecode/android_scripting/activ
ity/ScriptEditor.java
And this describes how to make the chooser intent:
http://developer.android.com/reference/android/content/Intent.html#ACTION_CHOOSE
R
```
Original issue reported on code.google.com by `armadaje...@gmail.com` on 5 Dec 2011 at 12:55
Attachments:
* [logcat.txt](https://storage.googleapis.com/google-code-attachments/android-scripting/issue-588/comment-0/logcat.txt)
* [chooser.py](https://storage.googleapis.com/google-code-attachments/android-scripting/issue-588/comment-0/chooser.py)
_Copied from original issue: damonkohler/android-scripting#588_ | 1.0 | SL4A Force Close on droid.startActivityIntent(chooserIntent) - _From @GoogleCodeExporter on May 31, 2015 11:30_
```
What device(s) are you experiencing the problem on?
Samsung Vibrant (SGH-T959)
What firmware version are you running on the device?
2.2
What steps will reproduce the problem?
1. Run the attached Python script on an Android device with SL4Ar4
-or-
1. Make an intent with action="android.intent.action.SEND", uriType =
"text/plain", extras = {"EXTRA_TEXT":"my message", "EXTRA_SUBJECT":"and my
subject"}, and everything else = None.
2. Make another intent with action = "android.intent.action.CHOOSER", extras =
{"EXTRA_INTENT":theFirstIntent}, and everything else = None.
3. droid.startActivityIntent(yourSecondIntent)
4. SL4A force closes
What is the expected output? What do you see instead?
I'm trying to get a chooser for sharing text via other apps (Gmail, Messaging,
etc.). It should be just like the one from the "Share" option in the menu of
SL4A's script editor activity. Instead, SL4A force closes, and there's no
indication whatsoever to the Python interpreter.
What version of the product are you using? On what operating system?
I'm using SL4Ar4 on Samung's OEM build FROYO.UVKB5
My development machine is Windows 7 64-bit, but running the script directly on
the phone still crashes, so that's probably not the issue.
Please provide any additional information below.
I tried to adapt the code from the "Share" option of the SL4A script editor by
manually making the same intent as Intent.createChooser() as described in the
Android documentation.
My script is adapted from lines 198-202 here:
http://code.google.com/p/android-scripting/source/diff?spec=svndeba523e8a9b44ea6
ef6e2497e21914084b109a8&r=deba523e8a9b44ea6ef6e2497e21914084b109a8&format=side&p
ath=/android/ScriptingLayerForAndroid/src/com/googlecode/android_scripting/activ
ity/ScriptEditor.java
And this describes how to make the chooser intent:
http://developer.android.com/reference/android/content/Intent.html#ACTION_CHOOSE
R
```
Original issue reported on code.google.com by `armadaje...@gmail.com` on 5 Dec 2011 at 12:55
Attachments:
* [logcat.txt](https://storage.googleapis.com/google-code-attachments/android-scripting/issue-588/comment-0/logcat.txt)
* [chooser.py](https://storage.googleapis.com/google-code-attachments/android-scripting/issue-588/comment-0/chooser.py)
_Copied from original issue: damonkohler/android-scripting#588_ | non_process | force close on droid startactivityintent chooserintent from googlecodeexporter on may what device s are you experiencing the problem on samsung vibrant sgh what firmware version are you running on the device what steps will reproduce the problem run the attached python script on an android device with or make an intent with action android intent action send uritype text plain extras extra text my message extra subject and my subject and everything else none make another intent with action android intent action chooser extras extra intent thefirstintent and everything else none droid startactivityintent yoursecondintent force closes what is the expected output what do you see instead i m trying to get a chooser for sharing text via other apps gmail messaging etc it should be just like the one from the share option in the menu of s script editor activity instead force closes and there s no indication whatsoever to the python interpreter what version of the product are you using on what operating system i m using on samung s oem build froyo my development machine is windows bit but running the script directly on the phone still crashes so that s probably not the issue please provide any additional information below i tried to adapt the code from the share option of the script editor by manually making the same intent as intent createchooser as described in the android documentation my script is adapted from lines here r format side p ath android scriptinglayerforandroid src com googlecode android scripting activ ity scripteditor java and this describes how to make the chooser intent r original issue reported on code google com by armadaje gmail com on dec at attachments copied from original issue damonkohler android scripting | 0 |
10,455 | 13,234,960,625 | IssuesEvent | 2020-08-18 17:11:26 | googleapis/repo-automation-bots | https://api.github.com/repos/googleapis/repo-automation-bots | closed | bug: refactor for octokit/webhooks | type: process | We either should refactor release-please to not use octokit/webhooks or add it to ignore modules
See #816 | 1.0 | bug: refactor for octokit/webhooks - We either should refactor release-please to not use octokit/webhooks or add it to ignore modules
See #816 | process | bug refactor for octokit webhooks we either should refactor release please to not use octokit webhooks or add it to ignore modules see | 1 |
340,736 | 24,668,635,804 | IssuesEvent | 2022-10-18 12:16:52 | mantidproject/mantid | https://api.github.com/repos/mantidproject/mantid | opened | dev-docs pages not removed on site update | Documentation ISIS Team: Core | Found here: https://github.com/mantidproject/mantid/issues/34357#issuecomment-1248311915
**Describe the bug**
A dev docs file was renamed. The new page exists on the dev docs website, but so does the old.
Old https://developer.mantidproject.org/Testing/SANSGUI/SANSGUITests.html
New https://developer.mantidproject.org/Testing/SANSGUI/ISISSANSGUITests.html
**Expected behavior**
If a file is removed (or renamed) from the dev docs (or the user docs), the old page is removed from the website.
A change is needed here: https://github.com/mantidproject/mantid/blob/main/buildconfig/Jenkins/Conda/build-and-publish-devsite.sh
Possibly do not use the old clone of the website.
**Screenshots**


| 1.0 | dev-docs pages not removed on site update - Found here: https://github.com/mantidproject/mantid/issues/34357#issuecomment-1248311915
**Describe the bug**
A dev docs file was renamed. The new page exists on the dev docs website, but so does the old.
Old https://developer.mantidproject.org/Testing/SANSGUI/SANSGUITests.html
New https://developer.mantidproject.org/Testing/SANSGUI/ISISSANSGUITests.html
**Expected behavior**
If a file is removed (or renamed) from the dev docs (or the user docs), the old page is removed from the website.
A change is needed here: https://github.com/mantidproject/mantid/blob/main/buildconfig/Jenkins/Conda/build-and-publish-devsite.sh
Possibly do not use the old clone of the website.
**Screenshots**


| non_process | dev docs pages not removed on site update found here describe the bug a dev docs file was renamed the new page exists on the dev docs website but so does the old old new expected behavior if a file is removed or renamed from the dev docs or the user docs the old page is removed from the website a change is needed here possibly do not use the old clone of the website screenshots | 0 |
1,681 | 2,658,826,471 | IssuesEvent | 2015-03-18 17:34:09 | phetsims/pendulum-lab | https://api.github.com/repos/phetsims/pendulum-lab | opened | Missing assets/*-screenshot.png | code review | Noticed during code review #56. Not sure who should be assigned for this. | 1.0 | Missing assets/*-screenshot.png - Noticed during code review #56. Not sure who should be assigned for this. | non_process | missing assets screenshot png noticed during code review not sure who should be assigned for this | 0 |
202,660 | 15,837,159,305 | IssuesEvent | 2021-04-06 20:22:37 | HARDTECHIO/dhoa-front | https://api.github.com/repos/HARDTECHIO/dhoa-front | closed | Aplicar Componentes do Template/MENU | documentation enhancement | ## Tarefas
- [x] Criar Template Menu;
- [x] Criar NavBar com os componentes com Bootstrap; | 1.0 | Aplicar Componentes do Template/MENU - ## Tarefas
- [x] Criar Template Menu;
- [x] Criar NavBar com os componentes com Bootstrap; | non_process | aplicar componentes do template menu tarefas criar template menu criar navbar com os componentes com bootstrap | 0 |
136,853 | 12,736,517,785 | IssuesEvent | 2020-06-25 17:03:23 | ualberta-smr/LibCompPlugin | https://api.github.com/repos/ualberta-smr/LibCompPlugin | closed | Finalize Paper | documentation | As I work on enhancements to LibComp, stay on track and continue to update the paper after each task.
Working on paper will take longer for the user study as I will have to analyze the data and re-write the entire section so I set up some time specifically for this task.
Goal: A good publishable paper for a tool tract. | 1.0 | Finalize Paper - As I work on enhancements to LibComp, stay on track and continue to update the paper after each task.
Working on paper will take longer for the user study as I will have to analyze the data and re-write the entire section so I set up some time specifically for this task.
Goal: A good publishable paper for a tool tract. | non_process | finalize paper as i work on enhancements to libcomp stay on track and continue to update the paper after each task working on paper will take longer for the user study as i will have to analyze the data and re write the entire section so i set up some time specifically for this task goal a good publishable paper for a tool tract | 0 |
131,172 | 18,244,879,937 | IssuesEvent | 2021-10-01 17:01:11 | protocol/nft-website | https://api.github.com/repos/protocol/nft-website | closed | [CONTENT] Write section summary content | help wanted P2 kind/enhancement dif/medium effort/hours topic/design-content | Write brief, TLDR-style summary pages for each content section:
- [ ] `/concepts/`
- [ ] `/tutorial/`
- [ ] `/how-to/`
- [ ] `/reference/`
Each of these already appears as a `README.md` in their respective directories, just without any page body content.
When done, do two things to make them appear:
- [ ] Remove their redirects to front-of-site from `docs/.vuepress/redirects`
- [ ] In site nav in `docs/.vuepress/config.js`, add a line for `link: 'foo'`after each section's `title:`, so nav section headers become clickable links to summary pages
- [ ] May be necessary to retitle or modify `/contribute/' in nav for consistency | 1.0 | [CONTENT] Write section summary content - Write brief, TLDR-style summary pages for each content section:
- [ ] `/concepts/`
- [ ] `/tutorial/`
- [ ] `/how-to/`
- [ ] `/reference/`
Each of these already appears as a `README.md` in their respective directories, just without any page body content.
When done, do two things to make them appear:
- [ ] Remove their redirects to front-of-site from `docs/.vuepress/redirects`
- [ ] In site nav in `docs/.vuepress/config.js`, add a line for `link: 'foo'`after each section's `title:`, so nav section headers become clickable links to summary pages
- [ ] May be necessary to retitle or modify `/contribute/' in nav for consistency | non_process | write section summary content write brief tldr style summary pages for each content section concepts tutorial how to reference each of these already appears as a readme md in their respective directories just without any page body content when done do two things to make them appear remove their redirects to front of site from docs vuepress redirects in site nav in docs vuepress config js add a line for link foo after each section s title so nav section headers become clickable links to summary pages may be necessary to retitle or modify contribute in nav for consistency | 0 |
29,696 | 8,392,143,480 | IssuesEvent | 2018-10-09 16:46:22 | trilinos/Trilinos | https://api.github.com/repos/trilinos/Trilinos | opened | PyTrilinos: Standardize Configuration Macros | PyTrilinos build | @trilinos/pytrilinos
## Expectations
Configuration macros for PyTrilinos should be consistently named and distinct from other packages.
## Current Behavior
Some configuration macros are not distinct (`HAVE_EPETRA`), some _are_ distinct (`HAVE_PYTRILINOS_AZTECOO`), and some are redundant (`HAVE_PYTRILINOS_EPETRA`).
## Motivation and Context
@mhoemmen made some suggestions while reviewing #3575, which led to the observation that this needs to be cleaned up.
## Definition of Done
- [ ] Eliminate `HAVE_<PACKAGE>` macros in favor of `HAVE_PYTRILINOS_<PACKAGE>` macros in `PyTrilinos_config.h.in`
- [ ] Make same changes in PyTrilinos headers and source files
- [ ] Test for Teuchos-only configuration
- [ ] Test for linear solver only configuration
- [ ] Test for linear and nonlinear solver only configuration
- [ ] Test for full configuration
| 1.0 | PyTrilinos: Standardize Configuration Macros - @trilinos/pytrilinos
## Expectations
Configuration macros for PyTrilinos should be consistently named and distinct from other packages.
## Current Behavior
Some configuration macros are not distinct (`HAVE_EPETRA`), some _are_ distinct (`HAVE_PYTRILINOS_AZTECOO`), and some are redundant (`HAVE_PYTRILINOS_EPETRA`).
## Motivation and Context
@mhoemmen made some suggestions while reviewing #3575, which led to the observation that this needs to be cleaned up.
## Definition of Done
- [ ] Eliminate `HAVE_<PACKAGE>` macros in favor of `HAVE_PYTRILINOS_<PACKAGE>` macros in `PyTrilinos_config.h.in`
- [ ] Make same changes in PyTrilinos headers and source files
- [ ] Test for Teuchos-only configuration
- [ ] Test for linear solver only configuration
- [ ] Test for linear and nonlinear solver only configuration
- [ ] Test for full configuration
| non_process | pytrilinos standardize configuration macros trilinos pytrilinos expectations configuration macros for pytrilinos should be consistently named and distinct from other packages current behavior some configuration macros are not distinct have epetra some are distinct have pytrilinos aztecoo and some are redundant have pytrilinos epetra motivation and context mhoemmen made some suggestions while reviewing which led to the observation that this needs to be cleaned up definition of done eliminate have macros in favor of have pytrilinos macros in pytrilinos config h in make same changes in pytrilinos headers and source files test for teuchos only configuration test for linear solver only configuration test for linear and nonlinear solver only configuration test for full configuration | 0 |
15,051 | 18,762,895,012 | IssuesEvent | 2021-11-05 18:46:18 | GoogleCloudPlatform/ai-platform-samples | https://api.github.com/repos/GoogleCloudPlatform/ai-platform-samples | closed | Python 3.5 CI builds are failing | type: process | Python 3.5 CI builds are failing with this error:
`FileNotFoundError: [Errno 2] No such file or directory: '/tmpfs/src/envs/python3.5/venv'`
[Example failed PR build](https://github.com/GoogleCloudPlatform/ai-platform-samples/pull/514) | 1.0 | Python 3.5 CI builds are failing - Python 3.5 CI builds are failing with this error:
`FileNotFoundError: [Errno 2] No such file or directory: '/tmpfs/src/envs/python3.5/venv'`
[Example failed PR build](https://github.com/GoogleCloudPlatform/ai-platform-samples/pull/514) | process | python ci builds are failing python ci builds are failing with this error filenotfounderror no such file or directory tmpfs src envs venv | 1 |
7,248 | 9,527,289,699 | IssuesEvent | 2019-04-29 02:59:04 | Lothrazar/Cyclic | https://api.github.com/repos/Lothrazar/Cyclic | closed | Garden Scythe Harvesting Bugs | bug: gameplay mod compatibility | Minecraft version & Mod Version:
- Forge v14.23.5.2796-1.12.2
- Cyclic v1.17.11
Single player or Server:
- Probably Both
Describe problem (what you were doing / what happened):
- There are multiple bugs that I noticed when harvesting modded crops using the garden scythe. (See below for id's)
1. The Red Orchid from [Extra Utilities 2](https://minecraft.curseforge.com/projects/extra-utilities) is harvested, but not replanted.
2. The Enderlilly from [Extra Utilities 2](https://minecraft.curseforge.com/projects/extra-utilities) is not harvested or replanted.
3. The Corn from [Simple Corn](https://minecraft.curseforge.com/projects/simple-corn) plants a 'ghost' plant. It simply disappears after updating the block.
4. The Grape Stem from [Rustic](https://minecraft.curseforge.com/projects/rustic) is still being harvested instead of the fruit. (See image below)
5. Tomatoes and Peppers from [Rustic](https://minecraft.curseforge.com/projects/rustic) can both grow 2-3 blocks tall on the stake, but harvesting sets it back to stage 1 instead of keeping it at stage 3. (See image below)
6. The Blueberry Bush from [Nifty](https://minecraft.curseforge.com/projects/niftyblocks?gameCategorySlug=mc-mods&projectID=228189) gives me the bush instead of the fruit when the scythe is used.
7. The Bushes from [Minestrappolation 5](https://minecraft.curseforge.com/projects/minestrappolation-5) also gives me bushes instead of the corresponding fruit.
8. The Wildberry Bush from [Rustic](https://minecraft.curseforge.com/projects/rustic) can't be harvested using the scythe.
9. The Bushes from [Plants 2](https://minecraft.curseforge.com/projects/plants) also aren't able to be harvested for fruit using the scythe.
10. The Bushes and the fruit hanging on trees from [Plant Mega Pack](http://10paksmods.net/pmp.html) are harvested from, but it doesn't give the fruit to you.
11. The Climbing Vine from [Plant Mega Pack gives](http://10paksmods.net/pmp.html) the vine instead of the fruit.
```
extrautils2:enderlilly
extrautils2:redorchid
minestrapp:blueberry_bush
minestrapp:blackberry_bush
minestrapp:raspberry_bush
minestrapp:strawberry_bush
minestrapp:mana_bush
minestrapp:voidberry_bush
nifty:blueberry_bush
plants2:harvest_0:*
plants2:harvest_1:*
plants2:nether_harvest:*
plants2:double_harvest_0
plants2:bush:*
pmp:bbush_beaut
pmp:bbush_black
pmp:bbush_blue
pmp:bbush_elder
pmp:bbush_goose
pmp:bbush_huckl
pmp:bbush_orang
pmp:bbush_snow
pmp:bbush_straw
pmp:climb_porce
rustic:wildberry_bush
rustic:grape_stem
rustic:grape_leaves
rustic:grapes
rustic:chili_crop
rustic:tomato_crop
simplecorn:corn
```


| True | Garden Scythe Harvesting Bugs - Minecraft version & Mod Version:
- Forge v14.23.5.2796-1.12.2
- Cyclic v1.17.11
Single player or Server:
- Probably Both
Describe problem (what you were doing / what happened):
- There are multiple bugs that I noticed when harvesting modded crops using the garden scythe. (See below for id's)
1. The Red Orchid from [Extra Utilities 2](https://minecraft.curseforge.com/projects/extra-utilities) is harvested, but not replanted.
2. The Enderlilly from [Extra Utilities 2](https://minecraft.curseforge.com/projects/extra-utilities) is not harvested or replanted.
3. The Corn from [Simple Corn](https://minecraft.curseforge.com/projects/simple-corn) plants a 'ghost' plant. It simply disappears after updating the block.
4. The Grape Stem from [Rustic](https://minecraft.curseforge.com/projects/rustic) is still being harvested instead of the fruit. (See image below)
5. Tomatoes and Peppers from [Rustic](https://minecraft.curseforge.com/projects/rustic) can both grow 2-3 blocks tall on the stake, but harvesting sets it back to stage 1 instead of keeping it at stage 3. (See image below)
6. The Blueberry Bush from [Nifty](https://minecraft.curseforge.com/projects/niftyblocks?gameCategorySlug=mc-mods&projectID=228189) gives me the bush instead of the fruit when the scythe is used.
7. The Bushes from [Minestrappolation 5](https://minecraft.curseforge.com/projects/minestrappolation-5) also gives me bushes instead of the corresponding fruit.
8. The Wildberry Bush from [Rustic](https://minecraft.curseforge.com/projects/rustic) can't be harvested using the scythe.
9. The Bushes from [Plants 2](https://minecraft.curseforge.com/projects/plants) also aren't able to be harvested for fruit using the scythe.
10. The Bushes and the fruit hanging on trees from [Plant Mega Pack](http://10paksmods.net/pmp.html) are harvested from, but it doesn't give the fruit to you.
11. The Climbing Vine from [Plant Mega Pack gives](http://10paksmods.net/pmp.html) the vine instead of the fruit.
```
extrautils2:enderlilly
extrautils2:redorchid
minestrapp:blueberry_bush
minestrapp:blackberry_bush
minestrapp:raspberry_bush
minestrapp:strawberry_bush
minestrapp:mana_bush
minestrapp:voidberry_bush
nifty:blueberry_bush
plants2:harvest_0:*
plants2:harvest_1:*
plants2:nether_harvest:*
plants2:double_harvest_0
plants2:bush:*
pmp:bbush_beaut
pmp:bbush_black
pmp:bbush_blue
pmp:bbush_elder
pmp:bbush_goose
pmp:bbush_huckl
pmp:bbush_orang
pmp:bbush_snow
pmp:bbush_straw
pmp:climb_porce
rustic:wildberry_bush
rustic:grape_stem
rustic:grape_leaves
rustic:grapes
rustic:chili_crop
rustic:tomato_crop
simplecorn:corn
```


| non_process | garden scythe harvesting bugs minecraft version mod version forge cyclic single player or server probably both describe problem what you were doing what happened there are multiple bugs that i noticed when harvesting modded crops using the garden scythe see below for id s the red orchid from is harvested but not replanted the enderlilly from is not harvested or replanted the corn from plants a ghost plant it simply disappears after updating the block the grape stem from is still being harvested instead of the fruit see image below tomatoes and peppers from can both grow blocks tall on the stake but harvesting sets it back to stage instead of keeping it at stage see image below the blueberry bush from gives me the bush instead of the fruit when the scythe is used the bushes from also gives me bushes instead of the corresponding fruit the wildberry bush from can t be harvested using the scythe the bushes from also aren t able to be harvested for fruit using the scythe the bushes and the fruit hanging on trees from are harvested from but it doesn t give the fruit to you the climbing vine from the vine instead of the fruit enderlilly redorchid minestrapp blueberry bush minestrapp blackberry bush minestrapp raspberry bush minestrapp strawberry bush minestrapp mana bush minestrapp voidberry bush nifty blueberry bush harvest harvest nether harvest double harvest bush pmp bbush beaut pmp bbush black pmp bbush blue pmp bbush elder pmp bbush goose pmp bbush huckl pmp bbush orang pmp bbush snow pmp bbush straw pmp climb porce rustic wildberry bush rustic grape stem rustic grape leaves rustic grapes rustic chili crop rustic tomato crop simplecorn corn | 0 |
233,521 | 25,765,525,473 | IssuesEvent | 2022-12-09 01:17:21 | jasonjberry/CDM | https://api.github.com/repos/jasonjberry/CDM | opened | CVE-2022-23491 (Medium) detected in certifi-2019.6.16-py2.py3-none-any.whl | security vulnerability | ## CVE-2022-23491 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>certifi-2019.6.16-py2.py3-none-any.whl</b></p></summary>
<p>Python package for providing Mozilla's CA Bundle.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/69/1b/b853c7a9d4f6a6d00749e94eb6f3a041e342a885b87340b79c1ef73e3a78/certifi-2019.6.16-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/69/1b/b853c7a9d4f6a6d00749e94eb6f3a041e342a885b87340b79c1ef73e3a78/certifi-2019.6.16-py2.py3-none-any.whl</a></p>
<p>Path to dependency file: /objectModel/Python</p>
<p>Path to vulnerable library: /objectModel/Python,/objectModel/Python/requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **certifi-2019.6.16-py2.py3-none-any.whl** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/jasonjberry/CDM/commit/2f532047f0b9e231c1602df6a076c534fba6d73a">2f532047f0b9e231c1602df6a076c534fba6d73a</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Certifi is a curated collection of Root Certificates for validating the trustworthiness of SSL certificates while verifying the identity of TLS hosts. Certifi 2022.12.07 removes root certificates from "TrustCor" from the root store. These are in the process of being removed from Mozilla's trust store. TrustCor's root certificates are being removed pursuant to an investigation prompted by media reporting that TrustCor's ownership also operated a business that produced spyware. Conclusions of Mozilla's investigation can be found in the linked google group discussion.
<p>Publish Date: 2022-12-07
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-23491>CVE-2022-23491</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2022-23491">https://www.cve.org/CVERecord?id=CVE-2022-23491</a></p>
<p>Release Date: 2022-12-07</p>
<p>Fix Resolution: certifi - 2022.12.07</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2022-23491 (Medium) detected in certifi-2019.6.16-py2.py3-none-any.whl - ## CVE-2022-23491 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>certifi-2019.6.16-py2.py3-none-any.whl</b></p></summary>
<p>Python package for providing Mozilla's CA Bundle.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/69/1b/b853c7a9d4f6a6d00749e94eb6f3a041e342a885b87340b79c1ef73e3a78/certifi-2019.6.16-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/69/1b/b853c7a9d4f6a6d00749e94eb6f3a041e342a885b87340b79c1ef73e3a78/certifi-2019.6.16-py2.py3-none-any.whl</a></p>
<p>Path to dependency file: /objectModel/Python</p>
<p>Path to vulnerable library: /objectModel/Python,/objectModel/Python/requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **certifi-2019.6.16-py2.py3-none-any.whl** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/jasonjberry/CDM/commit/2f532047f0b9e231c1602df6a076c534fba6d73a">2f532047f0b9e231c1602df6a076c534fba6d73a</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Certifi is a curated collection of Root Certificates for validating the trustworthiness of SSL certificates while verifying the identity of TLS hosts. Certifi 2022.12.07 removes root certificates from "TrustCor" from the root store. These are in the process of being removed from Mozilla's trust store. TrustCor's root certificates are being removed pursuant to an investigation prompted by media reporting that TrustCor's ownership also operated a business that produced spyware. Conclusions of Mozilla's investigation can be found in the linked google group discussion.
<p>Publish Date: 2022-12-07
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-23491>CVE-2022-23491</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2022-23491">https://www.cve.org/CVERecord?id=CVE-2022-23491</a></p>
<p>Release Date: 2022-12-07</p>
<p>Fix Resolution: certifi - 2022.12.07</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_process | cve medium detected in certifi none any whl cve medium severity vulnerability vulnerable library certifi none any whl python package for providing mozilla s ca bundle library home page a href path to dependency file objectmodel python path to vulnerable library objectmodel python objectmodel python requirements txt dependency hierarchy x certifi none any whl vulnerable library found in head commit a href found in base branch master vulnerability details certifi is a curated collection of root certificates for validating the trustworthiness of ssl certificates while verifying the identity of tls hosts certifi removes root certificates from trustcor from the root store these are in the process of being removed from mozilla s trust store trustcor s root certificates are being removed pursuant to an investigation prompted by media reporting that trustcor s ownership also operated a business that produced spyware conclusions of mozilla s investigation can be found in the linked google group discussion publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required high user interaction none scope changed impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution certifi step up your open source security game with mend | 0 |
7,483 | 10,574,172,509 | IssuesEvent | 2019-10-07 13:32:04 | prisma/lift | https://api.github.com/repos/prisma/lift | closed | Incorrect database name error | bug/2-confirmed kind/bug process/candidate | A database created via lift fails at `lift save`, the database name in this case is `p2-mysql-views`
To reproduce:
1. Create a blank project starter
2. Choose MySQL and create a new database with name `p2-mysql-view`
3. Run `prisma2 lift save --name init`
Panics with
```
divyendusingh [p2-mysql-view]$ prisma2 lift save --name init
ERROR Oops, an unexpected error occured!
Error in migration engine: thread 'main' panicked at 'CommandError(Generic { code: 1000, error: "QueryError(QueryError(MyS
qlError { ERROR 1102 (42000): Incorrect database name \'\' }\n\nstack backtrace:\n 0: backtrace::backtrace::trace\n 1:
backtrace::capture::Backtrace::new_unresolved\n 2: failure::backtrace::internal::InternalBacktrace::new\n 3: <failure
::backtrace::Backtrace as core::default::Default>::default\n 4: prisma_query::connector::mysql::error::<impl core::conve
rt::From<mysql::error::Error> for prisma_query::error::Error>::from\n 5: prisma_query::connector::metrics::query\n 6:
<prisma_query::connector::mysql::Mysql as prisma_query::connector::queryable::Queryable>::query_raw\n 7: <sql_migration_
connector::migration_database::Mysql as sql_migration_connector::migration_database::MigrationDatabase>::query_raw\n 8:
<sql_migration_connector::SqlMigrationConnector as migration_connector::MigrationConnector>::initialize\n 9: migration_e
ngine::migration_engine::MigrationEngine<C,D>::new\n 10: migration_engine::api::rpc::RpcApi::new\n 11: migration_engine:
:api::rpc::RpcApi::new_async\n 12: migration_engine::main\n 13: std::rt::lang_start::{{closure}}\n 14: std::panicking::
ERROR Oops, an unexpected error occured!
Error in migration engine: thread 'main' panicked at 'CommandError(Generic { code: 1000, error: "QueryError(QueryError(MyS
qlError { ERROR 1102 (42000): Incorrect database name \'\' }\n\nstack backtrace:\n 0: backtrace::backtrace::trace\n 1:
backtrace::capture::Backtrace::new_unresolved\n 2: failure::backtrace::internal::InternalBacktrace::new\n 3: <failure
::backtrace::Backtrace as core::default::Default>::default\n 4: prisma_query::connector::mysql::error::<impl core::conve
rt::From<mysql::error::Error> for prisma_query::error::Error>::from\n 5: prisma_query::connector::metrics::query\n 6:
<prisma_query::connector::mysql::Mysql as prisma_query::connector::queryable::Queryable>::query_raw\n 7: <sql_migration_
connector::migration_database::Mysql as sql_migration_connector::migration_database::MigrationDatabase>::query_raw\n 8:
<sql_migration_connector::SqlMigrationConnector as migration_connector::MigrationConnector>::initialize\n 9: migration_e
ngine::migration_engine::MigrationEngine<C,D>::new\n 10: migration_engine::api::rpc::RpcApi::new\n 11: migration_engine:
:api::rpc::RpcApi::new_async\n 12: migration_engine::main\n 13: std::rt::lang_start::{{closure}}\n 14: std::panicking::
try::do_call\n 15: __rust_maybe_catch_panic\n 16: std::rt::lang_start_internal\n 17: main\n)\n\nstack backtrace:\n 0:
backtrace::backtrace::trace\n 1: backtrace::capture::Backtrace::new_unresolved\n 2: failure::backtrace::internal::Int
ernalBacktrace::new\n 3: <failure::backtrace::Backtrace as core::default::Default>::default\n 4: <migration_connector
``` | 1.0 | Incorrect database name error - A database created via lift fails at `lift save`, the database name in this case is `p2-mysql-views`
To reproduce:
1. Create a blank project starter
2. Choose MySQL and create a new database with name `p2-mysql-view`
3. Run `prisma2 lift save --name init`
Panics with
```
divyendusingh [p2-mysql-view]$ prisma2 lift save --name init
ERROR Oops, an unexpected error occured!
Error in migration engine: thread 'main' panicked at 'CommandError(Generic { code: 1000, error: "QueryError(QueryError(MyS
qlError { ERROR 1102 (42000): Incorrect database name \'\' }\n\nstack backtrace:\n 0: backtrace::backtrace::trace\n 1:
backtrace::capture::Backtrace::new_unresolved\n 2: failure::backtrace::internal::InternalBacktrace::new\n 3: <failure
::backtrace::Backtrace as core::default::Default>::default\n 4: prisma_query::connector::mysql::error::<impl core::conve
rt::From<mysql::error::Error> for prisma_query::error::Error>::from\n 5: prisma_query::connector::metrics::query\n 6:
<prisma_query::connector::mysql::Mysql as prisma_query::connector::queryable::Queryable>::query_raw\n 7: <sql_migration_
connector::migration_database::Mysql as sql_migration_connector::migration_database::MigrationDatabase>::query_raw\n 8:
<sql_migration_connector::SqlMigrationConnector as migration_connector::MigrationConnector>::initialize\n 9: migration_e
ngine::migration_engine::MigrationEngine<C,D>::new\n 10: migration_engine::api::rpc::RpcApi::new\n 11: migration_engine:
:api::rpc::RpcApi::new_async\n 12: migration_engine::main\n 13: std::rt::lang_start::{{closure}}\n 14: std::panicking::
ERROR Oops, an unexpected error occured!
Error in migration engine: thread 'main' panicked at 'CommandError(Generic { code: 1000, error: "QueryError(QueryError(MyS
qlError { ERROR 1102 (42000): Incorrect database name \'\' }\n\nstack backtrace:\n 0: backtrace::backtrace::trace\n 1:
backtrace::capture::Backtrace::new_unresolved\n 2: failure::backtrace::internal::InternalBacktrace::new\n 3: <failure
::backtrace::Backtrace as core::default::Default>::default\n 4: prisma_query::connector::mysql::error::<impl core::conve
rt::From<mysql::error::Error> for prisma_query::error::Error>::from\n 5: prisma_query::connector::metrics::query\n 6:
<prisma_query::connector::mysql::Mysql as prisma_query::connector::queryable::Queryable>::query_raw\n 7: <sql_migration_
connector::migration_database::Mysql as sql_migration_connector::migration_database::MigrationDatabase>::query_raw\n 8:
<sql_migration_connector::SqlMigrationConnector as migration_connector::MigrationConnector>::initialize\n 9: migration_e
ngine::migration_engine::MigrationEngine<C,D>::new\n 10: migration_engine::api::rpc::RpcApi::new\n 11: migration_engine:
:api::rpc::RpcApi::new_async\n 12: migration_engine::main\n 13: std::rt::lang_start::{{closure}}\n 14: std::panicking::
try::do_call\n 15: __rust_maybe_catch_panic\n 16: std::rt::lang_start_internal\n 17: main\n)\n\nstack backtrace:\n 0:
backtrace::backtrace::trace\n 1: backtrace::capture::Backtrace::new_unresolved\n 2: failure::backtrace::internal::Int
ernalBacktrace::new\n 3: <failure::backtrace::Backtrace as core::default::Default>::default\n 4: <migration_connector
``` | process | incorrect database name error a database created via lift fails at lift save the database name in this case is mysql views to reproduce create a blank project starter choose mysql and create a new database with name mysql view run lift save name init panics with divyendusingh lift save name init error oops an unexpected error occured error in migration engine thread main panicked at commanderror generic code error queryerror queryerror mys qlerror error incorrect database name n nstack backtrace n backtrace backtrace trace n backtrace capture backtrace new unresolved n failure backtrace internal internalbacktrace new n failure backtrace backtrace as core default default default n prisma query connector mysql error impl core conve rt from for prisma query error error from n prisma query connector metrics query n query raw n sql migration connector migration database mysql as sql migration connector migration database migrationdatabase query raw n initialize n migration e ngine migration engine migrationengine new n migration engine api rpc rpcapi new n migration engine api rpc rpcapi new async n migration engine main n std rt lang start closure n std panicking error oops an unexpected error occured error in migration engine thread main panicked at commanderror generic code error queryerror queryerror mys qlerror error incorrect database name n nstack backtrace n backtrace backtrace trace n backtrace capture backtrace new unresolved n failure backtrace internal internalbacktrace new n failure backtrace backtrace as core default default default n prisma query connector mysql error impl core conve rt from for prisma query error error from n prisma query connector metrics query n query raw n sql migration connector migration database mysql as sql migration connector migration database migrationdatabase query raw n initialize n migration e ngine migration engine migrationengine new n migration engine api rpc rpcapi new n migration engine api rpc rpcapi new async n migration engine main n std rt lang start closure n std panicking try do call n rust maybe catch panic n std rt lang start internal n main n n nstack backtrace n backtrace backtrace trace n backtrace capture backtrace new unresolved n failure backtrace internal int ernalbacktrace new n default n migration connector | 1 |
275,243 | 30,219,150,649 | IssuesEvent | 2023-07-05 17:53:48 | temporalio/samples-go | https://api.github.com/repos/temporalio/samples-go | closed | go.temporal.io/sdk/contrib/tools/workflowcheck-v0.0.0-20230612164027-11c2cb9e7d2d: 3 vulnerabilities (highest severity is: 6.1) - autoclosed | Mend: dependency security vulnerability | <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>go.temporal.io/sdk/contrib/tools/workflowcheck-v0.0.0-20230612164027-11c2cb9e7d2d</b></p></summary>
<p></p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/temporalio/samples-go/commit/d54b5f7bd51ab2c52ba1eb744b6f87ab9bc5596f">d54b5f7bd51ab2c52ba1eb744b6f87ab9bc5596f</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (go.temporal.io/sdk/contrib/tools/workflowcheck-v0.0.0 version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2020-11022](https://www.mend.io/vulnerability-database/CVE-2020-11022) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 6.1 | golang.org/x/tools-v0.9.3 | Transitive | N/A* | ❌ |
| [CVE-2015-9251](https://www.mend.io/vulnerability-database/CVE-2015-9251) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 6.1 | golang.org/x/tools-v0.9.3 | Transitive | N/A* | ❌ |
| [CVE-2012-6708](https://www.mend.io/vulnerability-database/CVE-2012-6708) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 6.1 | golang.org/x/tools-v0.9.3 | Transitive | N/A* | ❌ |
<p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the "Details" section below to see if there is a version of transitive dependency where vulnerability is fixed.</p>
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2020-11022</summary>
### Vulnerable Library - <b>golang.org/x/tools-v0.9.3</b></p>
<p></p>
<p>Library home page: <a href="https://proxy.golang.org/golang.org/x/tools/@v/v0.9.3.zip">https://proxy.golang.org/golang.org/x/tools/@v/v0.9.3.zip</a></p>
<p>
Dependency Hierarchy:
- go.temporal.io/sdk/contrib/tools/workflowcheck-v0.0.0-20230612164027-11c2cb9e7d2d (Root Library)
- :x: **golang.org/x/tools-v0.9.3** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/temporalio/samples-go/commit/d54b5f7bd51ab2c52ba1eb744b6f87ab9bc5596f">d54b5f7bd51ab2c52ba1eb744b6f87ab9bc5596f</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.
<p>Publish Date: 2020-04-29
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-11022>CVE-2020-11022</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11022">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11022</a></p>
<p>Release Date: 2020-04-29</p>
<p>Fix Resolution: jQuery - 3.5.0</p>
</p>
<p></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2015-9251</summary>
### Vulnerable Library - <b>golang.org/x/tools-v0.9.3</b></p>
<p></p>
<p>Library home page: <a href="https://proxy.golang.org/golang.org/x/tools/@v/v0.9.3.zip">https://proxy.golang.org/golang.org/x/tools/@v/v0.9.3.zip</a></p>
<p>
Dependency Hierarchy:
- go.temporal.io/sdk/contrib/tools/workflowcheck-v0.0.0-20230612164027-11c2cb9e7d2d (Root Library)
- :x: **golang.org/x/tools-v0.9.3** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/temporalio/samples-go/commit/d54b5f7bd51ab2c52ba1eb744b6f87ab9bc5596f">d54b5f7bd51ab2c52ba1eb744b6f87ab9bc5596f</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed.
<p>Publish Date: 2018-01-18
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2015-9251>CVE-2015-9251</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p>
<p>Release Date: 2018-01-18</p>
<p>Fix Resolution: jQuery - 3.0.0</p>
</p>
<p></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2012-6708</summary>
### Vulnerable Library - <b>golang.org/x/tools-v0.9.3</b></p>
<p></p>
<p>Library home page: <a href="https://proxy.golang.org/golang.org/x/tools/@v/v0.9.3.zip">https://proxy.golang.org/golang.org/x/tools/@v/v0.9.3.zip</a></p>
<p>
Dependency Hierarchy:
- go.temporal.io/sdk/contrib/tools/workflowcheck-v0.0.0-20230612164027-11c2cb9e7d2d (Root Library)
- :x: **golang.org/x/tools-v0.9.3** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/temporalio/samples-go/commit/d54b5f7bd51ab2c52ba1eb744b6f87ab9bc5596f">d54b5f7bd51ab2c52ba1eb744b6f87ab9bc5596f</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
jQuery before 1.9.0 is vulnerable to Cross-site Scripting (XSS) attacks. The jQuery(strInput) function does not differentiate selectors from HTML in a reliable fashion. In vulnerable versions, jQuery determined whether the input was HTML by looking for the '<' character anywhere in the string, giving attackers more flexibility when attempting to construct a malicious payload. In fixed versions, jQuery only deems the input to be HTML if it explicitly starts with the '<' character, limiting exploitability only to attackers who can control the beginning of a string, which is far less common.
<p>Publish Date: 2018-01-18
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2012-6708>CVE-2012-6708</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2012-6708">https://nvd.nist.gov/vuln/detail/CVE-2012-6708</a></p>
<p>Release Date: 2018-01-18</p>
<p>Fix Resolution: jQuery - v1.9.0</p>
</p>
<p></p>
</details> | True | go.temporal.io/sdk/contrib/tools/workflowcheck-v0.0.0-20230612164027-11c2cb9e7d2d: 3 vulnerabilities (highest severity is: 6.1) - autoclosed - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>go.temporal.io/sdk/contrib/tools/workflowcheck-v0.0.0-20230612164027-11c2cb9e7d2d</b></p></summary>
<p></p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/temporalio/samples-go/commit/d54b5f7bd51ab2c52ba1eb744b6f87ab9bc5596f">d54b5f7bd51ab2c52ba1eb744b6f87ab9bc5596f</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (go.temporal.io/sdk/contrib/tools/workflowcheck-v0.0.0 version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2020-11022](https://www.mend.io/vulnerability-database/CVE-2020-11022) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 6.1 | golang.org/x/tools-v0.9.3 | Transitive | N/A* | ❌ |
| [CVE-2015-9251](https://www.mend.io/vulnerability-database/CVE-2015-9251) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 6.1 | golang.org/x/tools-v0.9.3 | Transitive | N/A* | ❌ |
| [CVE-2012-6708](https://www.mend.io/vulnerability-database/CVE-2012-6708) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 6.1 | golang.org/x/tools-v0.9.3 | Transitive | N/A* | ❌ |
<p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the "Details" section below to see if there is a version of transitive dependency where vulnerability is fixed.</p>
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2020-11022</summary>
### Vulnerable Library - <b>golang.org/x/tools-v0.9.3</b></p>
<p></p>
<p>Library home page: <a href="https://proxy.golang.org/golang.org/x/tools/@v/v0.9.3.zip">https://proxy.golang.org/golang.org/x/tools/@v/v0.9.3.zip</a></p>
<p>
Dependency Hierarchy:
- go.temporal.io/sdk/contrib/tools/workflowcheck-v0.0.0-20230612164027-11c2cb9e7d2d (Root Library)
- :x: **golang.org/x/tools-v0.9.3** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/temporalio/samples-go/commit/d54b5f7bd51ab2c52ba1eb744b6f87ab9bc5596f">d54b5f7bd51ab2c52ba1eb744b6f87ab9bc5596f</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.
<p>Publish Date: 2020-04-29
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-11022>CVE-2020-11022</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11022">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11022</a></p>
<p>Release Date: 2020-04-29</p>
<p>Fix Resolution: jQuery - 3.5.0</p>
</p>
<p></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2015-9251</summary>
### Vulnerable Library - <b>golang.org/x/tools-v0.9.3</b></p>
<p></p>
<p>Library home page: <a href="https://proxy.golang.org/golang.org/x/tools/@v/v0.9.3.zip">https://proxy.golang.org/golang.org/x/tools/@v/v0.9.3.zip</a></p>
<p>
Dependency Hierarchy:
- go.temporal.io/sdk/contrib/tools/workflowcheck-v0.0.0-20230612164027-11c2cb9e7d2d (Root Library)
- :x: **golang.org/x/tools-v0.9.3** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/temporalio/samples-go/commit/d54b5f7bd51ab2c52ba1eb744b6f87ab9bc5596f">d54b5f7bd51ab2c52ba1eb744b6f87ab9bc5596f</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed.
<p>Publish Date: 2018-01-18
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2015-9251>CVE-2015-9251</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p>
<p>Release Date: 2018-01-18</p>
<p>Fix Resolution: jQuery - 3.0.0</p>
</p>
<p></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2012-6708</summary>
### Vulnerable Library - <b>golang.org/x/tools-v0.9.3</b></p>
<p></p>
<p>Library home page: <a href="https://proxy.golang.org/golang.org/x/tools/@v/v0.9.3.zip">https://proxy.golang.org/golang.org/x/tools/@v/v0.9.3.zip</a></p>
<p>
Dependency Hierarchy:
- go.temporal.io/sdk/contrib/tools/workflowcheck-v0.0.0-20230612164027-11c2cb9e7d2d (Root Library)
- :x: **golang.org/x/tools-v0.9.3** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/temporalio/samples-go/commit/d54b5f7bd51ab2c52ba1eb744b6f87ab9bc5596f">d54b5f7bd51ab2c52ba1eb744b6f87ab9bc5596f</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
jQuery before 1.9.0 is vulnerable to Cross-site Scripting (XSS) attacks. The jQuery(strInput) function does not differentiate selectors from HTML in a reliable fashion. In vulnerable versions, jQuery determined whether the input was HTML by looking for the '<' character anywhere in the string, giving attackers more flexibility when attempting to construct a malicious payload. In fixed versions, jQuery only deems the input to be HTML if it explicitly starts with the '<' character, limiting exploitability only to attackers who can control the beginning of a string, which is far less common.
<p>Publish Date: 2018-01-18
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2012-6708>CVE-2012-6708</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2012-6708">https://nvd.nist.gov/vuln/detail/CVE-2012-6708</a></p>
<p>Release Date: 2018-01-18</p>
<p>Fix Resolution: jQuery - v1.9.0</p>
</p>
<p></p>
</details> | non_process | go temporal io sdk contrib tools workflowcheck vulnerabilities highest severity is autoclosed vulnerable library go temporal io sdk contrib tools workflowcheck found in head commit a href vulnerabilities cve severity cvss dependency type fixed in go temporal io sdk contrib tools workflowcheck version remediation available medium golang org x tools transitive n a medium golang org x tools transitive n a medium golang org x tools transitive n a for some transitive vulnerabilities there is no version of direct dependency with a fix check the details section below to see if there is a version of transitive dependency where vulnerability is fixed details cve vulnerable library golang org x tools library home page a href dependency hierarchy go temporal io sdk contrib tools workflowcheck root library x golang org x tools vulnerable library found in head commit a href found in base branch main vulnerability details in jquery versions greater than or equal to and before passing html from untrusted sources even after sanitizing it to one of jquery s dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery cve vulnerable library golang org x tools library home page a href dependency hierarchy go temporal io sdk contrib tools workflowcheck root library x golang org x tools vulnerable library found in head commit a href found in base branch main vulnerability details jquery before is vulnerable to cross site scripting xss attacks when a cross domain ajax request is performed without the datatype option causing text javascript responses to be executed publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery cve vulnerable library golang org x tools library home page a href dependency hierarchy go temporal io sdk contrib tools workflowcheck root library x golang org x tools vulnerable library found in head commit a href found in base branch main vulnerability details jquery before is vulnerable to cross site scripting xss attacks the jquery strinput function does not differentiate selectors from html in a reliable fashion in vulnerable versions jquery determined whether the input was html by looking for the character anywhere in the string giving attackers more flexibility when attempting to construct a malicious payload in fixed versions jquery only deems the input to be html if it explicitly starts with the character limiting exploitability only to attackers who can control the beginning of a string which is far less common publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery | 0 |
238,603 | 18,245,597,930 | IssuesEvent | 2021-10-01 17:57:14 | uriahf/rtichoke | https://api.github.com/repos/uriahf/rtichoke | closed | Change naming convention, from "performance_table" to "performance_data" | documentation rtichoke function | "performance_data" is a better name than "performance_table" and it is usefull to distinguish between the data as an object and a rendered table.
The function `create_performance_table()` should be renamed to `prepare_performance_data()`.
The output should be tibble instead of data.frame. | 1.0 | Change naming convention, from "performance_table" to "performance_data" - "performance_data" is a better name than "performance_table" and it is usefull to distinguish between the data as an object and a rendered table.
The function `create_performance_table()` should be renamed to `prepare_performance_data()`.
The output should be tibble instead of data.frame. | non_process | change naming convention from performance table to performance data performance data is a better name than performance table and it is usefull to distinguish between the data as an object and a rendered table the function create performance table should be renamed to prepare performance data the output should be tibble instead of data frame | 0 |
12,844 | 15,225,037,093 | IssuesEvent | 2021-02-18 06:33:53 | googleapis/java-functions | https://api.github.com/repos/googleapis/java-functions | reopened | Dependency Dashboard | api: cloudfunctions type: process | This issue contains a list of Renovate updates and their statuses.
## Open
These updates have all been created already. Click a checkbox below to force a retry/rebase of any.
- [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-functions-1.x -->[chore(deps): update dependency com.google.cloud:google-cloud-functions to v1.0.6](../pull/134)
---
- [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
| 1.0 | Dependency Dashboard - This issue contains a list of Renovate updates and their statuses.
## Open
These updates have all been created already. Click a checkbox below to force a retry/rebase of any.
- [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-functions-1.x -->[chore(deps): update dependency com.google.cloud:google-cloud-functions to v1.0.6](../pull/134)
---
- [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
| process | dependency dashboard this issue contains a list of renovate updates and their statuses open these updates have all been created already click a checkbox below to force a retry rebase of any pull check this box to trigger a request for renovate to run again on this repository | 1 |
21,905 | 30,353,041,507 | IssuesEvent | 2023-07-11 20:36:30 | metabase/metabase | https://api.github.com/repos/metabase/metabase | closed | config.yml fails if password contains special characters | Type:Bug Priority:P2 Operation/Serialization .Team/QueryProcessor :hammer_and_wrench: | ### Describe the bug
Trying to initialize Metabase with the following config file:
```
version: 1
config:
users:
- first_name: Luiz
last_name: Arakaki
password: MetaPa$$123{{>
email: admin@metabase.com
```
It fails with `clojure.lang.ExceptionInfo: Invalid query: found [[ or {{ with no matching ]] or }} {:type :invalid-query}`
### To Reproduce
Use the config file above
### Expected behavior
It should start Metabase and the admin must use that password
### Logs
_No response_
### Information about your Metabase installation
```JSON
46.3
```
### Severity
P2
### Additional context
_No response_ | 1.0 | config.yml fails if password contains special characters - ### Describe the bug
Trying to initialize Metabase with the following config file:
```
version: 1
config:
users:
- first_name: Luiz
last_name: Arakaki
password: MetaPa$$123{{>
email: admin@metabase.com
```
It fails with `clojure.lang.ExceptionInfo: Invalid query: found [[ or {{ with no matching ]] or }} {:type :invalid-query}`
### To Reproduce
Use the config file above
### Expected behavior
It should start Metabase and the admin must use that password
### Logs
_No response_
### Information about your Metabase installation
```JSON
46.3
```
### Severity
P2
### Additional context
_No response_ | process | config yml fails if password contains special characters describe the bug trying to initialize metabase with the following config file version config users first name luiz last name arakaki password metapa email admin metabase com it fails with clojure lang exceptioninfo invalid query found or type invalid query to reproduce use the config file above expected behavior it should start metabase and the admin must use that password logs no response information about your metabase installation json severity additional context no response | 1 |
15,708 | 19,848,660,502 | IssuesEvent | 2022-01-21 09:47:37 | ooi-data/CE07SHSM-SBD12-04-PCO2AA000-recovered_host-pco2a_a_dcl_instrument_water_recovered | https://api.github.com/repos/ooi-data/CE07SHSM-SBD12-04-PCO2AA000-recovered_host-pco2a_a_dcl_instrument_water_recovered | opened | 🛑 Processing failed: ValueError | process | ## Overview
`ValueError` found in `processing_task` task during run ended on 2022-01-21T09:47:36.956819.
## Details
Flow name: `CE07SHSM-SBD12-04-PCO2AA000-recovered_host-pco2a_a_dcl_instrument_water_recovered`
Task name: `processing_task`
Error type: `ValueError`
Error message: not enough values to unpack (expected 3, got 0)
<details>
<summary>Traceback</summary>
```
Traceback (most recent call last):
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing
final_path = finalize_data_stream(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream
append_to_zarr(mod_ds, final_store, enc, logger=logger)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr
_append_zarr(store, mod_ds)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr
existing_arr.append(var_data.values)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values
return _as_array_or_item(self._data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item
data = np.asarray(data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__
x = self.compute()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute
(result,) = compute(self, traverse=False, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute
results = schedule(dsk, keys, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get
results = get_async(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async
raise_exception(exc, tb)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise
raise exc
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task
result = _execute_task(task, data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task
return func(*(_execute_task(a, cache) for a in args))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter
c = np.asarray(c)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__
self._ensure_cached()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached
self.array = NumpyIndexingAdapter(np.asarray(self.array))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__
return np.asarray(array[self.key], dtype=None)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__
return array[key.tuple]
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__
return self.get_basic_selection(selection, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection
return self._get_basic_selection_nd(selection=selection, out=out,
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd
return self._get_selection(indexer=indexer, out=out, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection
lchunk_coords, lchunk_selection, lout_selection = zip(*indexer)
ValueError: not enough values to unpack (expected 3, got 0)
```
</details>
| 1.0 | 🛑 Processing failed: ValueError - ## Overview
`ValueError` found in `processing_task` task during run ended on 2022-01-21T09:47:36.956819.
## Details
Flow name: `CE07SHSM-SBD12-04-PCO2AA000-recovered_host-pco2a_a_dcl_instrument_water_recovered`
Task name: `processing_task`
Error type: `ValueError`
Error message: not enough values to unpack (expected 3, got 0)
<details>
<summary>Traceback</summary>
```
Traceback (most recent call last):
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing
final_path = finalize_data_stream(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream
append_to_zarr(mod_ds, final_store, enc, logger=logger)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr
_append_zarr(store, mod_ds)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr
existing_arr.append(var_data.values)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values
return _as_array_or_item(self._data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item
data = np.asarray(data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__
x = self.compute()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute
(result,) = compute(self, traverse=False, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute
results = schedule(dsk, keys, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get
results = get_async(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async
raise_exception(exc, tb)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise
raise exc
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task
result = _execute_task(task, data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task
return func(*(_execute_task(a, cache) for a in args))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter
c = np.asarray(c)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__
self._ensure_cached()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached
self.array = NumpyIndexingAdapter(np.asarray(self.array))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__
return np.asarray(array[self.key], dtype=None)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__
return array[key.tuple]
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__
return self.get_basic_selection(selection, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection
return self._get_basic_selection_nd(selection=selection, out=out,
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd
return self._get_selection(indexer=indexer, out=out, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection
lchunk_coords, lchunk_selection, lout_selection = zip(*indexer)
ValueError: not enough values to unpack (expected 3, got 0)
```
</details>
| process | 🛑 processing failed valueerror overview valueerror found in processing task task during run ended on details flow name recovered host a dcl instrument water recovered task name processing task error type valueerror error message not enough values to unpack expected got traceback traceback most recent call last file srv conda envs notebook lib site packages ooi harvester processor pipeline py line in processing final path finalize data stream file srv conda envs notebook lib site packages ooi harvester processor init py line in finalize data stream append to zarr mod ds final store enc logger logger file srv conda envs notebook lib site packages ooi harvester processor init py line in append to zarr append zarr store mod ds file srv conda envs notebook lib site packages ooi harvester processor utils py line in append zarr existing arr append var data values file srv conda envs notebook lib site packages xarray core variable py line in values return as array or item self data file srv conda envs notebook lib site packages xarray core variable py line in as array or item data np asarray data file srv conda envs notebook lib site packages dask array core py line in array x self compute file srv conda envs notebook lib site packages dask base py line in compute result compute self traverse false kwargs file srv conda envs notebook lib site packages dask base py line in compute results schedule dsk keys kwargs file srv conda envs notebook lib site packages dask threaded py line in get results get async file srv conda envs notebook lib site packages dask local py line in get async raise exception exc tb file srv conda envs notebook lib site packages dask local py line in reraise raise exc file srv conda envs notebook lib site packages dask local py line in execute task result execute task task data file srv conda envs notebook lib site packages dask core py line in execute task return func execute task a cache for a in args file srv conda envs notebook lib site packages dask array core py line in getter c np asarray c file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array self ensure cached file srv conda envs notebook lib site packages xarray core indexing py line in ensure cached self array numpyindexingadapter np asarray self array file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray backends zarr py line in getitem return array file srv conda envs notebook lib site packages zarr core py line in getitem return self get basic selection selection fields fields file srv conda envs notebook lib site packages zarr core py line in get basic selection return self get basic selection nd selection selection out out file srv conda envs notebook lib site packages zarr core py line in get basic selection nd return self get selection indexer indexer out out fields fields file srv conda envs notebook lib site packages zarr core py line in get selection lchunk coords lchunk selection lout selection zip indexer valueerror not enough values to unpack expected got | 1 |
148,814 | 5,696,848,853 | IssuesEvent | 2017-04-16 16:00:29 | mreishman/Log-Hog | https://api.github.com/repos/mreishman/Log-Hog | opened | Reset changes button next to save button | enhancement Priority - 3 - Medium | - [ ] Reset changes button next to save button
(resets back to values from on page load) | 1.0 | Reset changes button next to save button - - [ ] Reset changes button next to save button
(resets back to values from on page load) | non_process | reset changes button next to save button reset changes button next to save button resets back to values from on page load | 0 |
2,382 | 5,187,201,391 | IssuesEvent | 2017-01-20 16:12:06 | openvstorage/framework-alba-plugin | https://api.github.com/repos/openvstorage/framework-alba-plugin | closed | Failed removing disk /dev/sdb: 'NoneType' object has no attribute 'partitions' | priority_minor process_cantreproduce type_bug | Initialized an SSD with 2 ASD and a bit later removed it. Didn't claim the ASDs.
Resulting error:
Failed removing disk /dev/sdb: 'NoneType' object has no attribute 'partitions' | 1.0 | Failed removing disk /dev/sdb: 'NoneType' object has no attribute 'partitions' - Initialized an SSD with 2 ASD and a bit later removed it. Didn't claim the ASDs.
Resulting error:
Failed removing disk /dev/sdb: 'NoneType' object has no attribute 'partitions' | process | failed removing disk dev sdb nonetype object has no attribute partitions initialized an ssd with asd and a bit later removed it didn t claim the asds resulting error failed removing disk dev sdb nonetype object has no attribute partitions | 1 |
13,839 | 3,363,004,739 | IssuesEvent | 2015-11-20 09:57:52 | NativeScript/nativescript-cli | https://api.github.com/repos/NativeScript/nativescript-cli | closed | "tns test" fails with Unknown command 'dev-test|android'. | 2 - Ready For Test bug critical unit testing | C:\Git>rm -rf TestApp
C:\Git>tns create TestApp
Project TestApp was successfully created
C:\Git>tns test init --path TestApp
? Select testing framework: jasmine
npm WARN package.json @ No description
npm WARN package.json @ No repository field.
npm WARN package.json @ No README data
karma@0.13.15 TestApp\node_modules\karma
├── batch@0.5.3
├── di@0.0.1
├── rimraf@2.4.3
├── graceful-fs@4.1.2
├── mime@1.3.4
├── colors@1.1.2
├── http-proxy@1.12.0 (eventemitter3@1.1.1, requires-port@0.0.1)
├── minimatch@3.0.0 (brace-expansion@1.1.1)
├── glob@5.0.15 (path-is-absolute@1.0.0, inherits@2.0.1, inflight@1.0.4, once@1.3.2)
├── dom-serialize@2.2.1 (custom-event@1.0.0, void-elements@2.0.1, extend@3.0.0, ent@2.2.0)
├── optimist@0.6.1 (wordwrap@0.0.3, minimist@0.0.10)
├── connect@3.4.0 (utils-merge@1.0.0, parseurl@1.3.0, debug@2.2.0, finalhandler@0.4.0)
├── useragent@2.1.7 (lru-cache@2.2.4)
├── expand-braces@0.1.2 (array-unique@0.2.1, array-slice@0.2.3, braces@0.1.5)
├── bluebird@2.10.2
├── body-parser@1.14.1 (bytes@2.1.0, content-type@1.0.1, depd@1.1.0, qs@5.1.0, http-errors@1.3.1, on-finished@2.3.0, raw-body@2.1.4, debug@2.2.0, iconv-lite@0.4.12, type-is@1.6.9)
├── source-map@0.5.3
├── chokidar@1.2.0 (arrify@1.0.0, path-is-absolute@1.0.0, glob-parent@2.0.0, async-each@0.1.6, is-binary-path@1.0.1, is-glob@2.0.1, lodash.flatten@3.0.2, readdirp@2.0.0, anymatch@1.3.0)
├── log4js@0.6.27 (async@0.2.10, underscore@1.8.2, readable-stream@1.0.33, semver@4.3.6)
├── lodash@3.10.1
├── core-js@1.2.6
└── socket.io@1.3.7 (debug@2.1.0, has-binary-data@0.1.3, engine.io@1.5.4, socket.io-adapter@0.3.1, socket.io-parser@2.2.4, socket.io-client@1.3.7)
npm WARN peerDependencies The peer dependency jasmine-core@* included from karma-jasmine will no
npm WARN peerDependencies longer be automatically installed to fulfill the peerDependency
npm WARN peerDependencies in npm 3+. Your application will need to depend on it explicitly.
jasmine-core@2.3.4 TestApp\node_modules\jasmine-core
karma-jasmine@0.3.6 TestApp\node_modules\karma-jasmine
karma-nativescript-launcher@0.2.5 TestApp\node_modules\karma-nativescript-launcher
nativescript-unit-test-runner@0.2.8 TestApp\node_modules\nativescript-unit-test-runner
Successfully installed plugin nativescript-unit-test-runner.
Example test file created in app/tests/
Run your tests using the "$ tns test <platform>" command.
C:\Git>
C:\Git>
C:\Git>tns test android --path TestApp
16 11 2015 17:20:33.594:INFO [karma]: Karma v0.13.15 server started at http://localhost:9876/
16 11 2015 17:20:33.601:INFO [launcher]: Starting browser NativeScript Unit Test Runner
Copying template files...
spawn EBADF
Unknown command 'dev-test|android'. Try '$ tns help' for a full list of supported commands.
16 11 2015 17:22:08.033:INFO [launcher]: NativeScript deployment completed with code 127 | 2.0 | "tns test" fails with Unknown command 'dev-test|android'. - C:\Git>rm -rf TestApp
C:\Git>tns create TestApp
Project TestApp was successfully created
C:\Git>tns test init --path TestApp
? Select testing framework: jasmine
npm WARN package.json @ No description
npm WARN package.json @ No repository field.
npm WARN package.json @ No README data
karma@0.13.15 TestApp\node_modules\karma
├── batch@0.5.3
├── di@0.0.1
├── rimraf@2.4.3
├── graceful-fs@4.1.2
├── mime@1.3.4
├── colors@1.1.2
├── http-proxy@1.12.0 (eventemitter3@1.1.1, requires-port@0.0.1)
├── minimatch@3.0.0 (brace-expansion@1.1.1)
├── glob@5.0.15 (path-is-absolute@1.0.0, inherits@2.0.1, inflight@1.0.4, once@1.3.2)
├── dom-serialize@2.2.1 (custom-event@1.0.0, void-elements@2.0.1, extend@3.0.0, ent@2.2.0)
├── optimist@0.6.1 (wordwrap@0.0.3, minimist@0.0.10)
├── connect@3.4.0 (utils-merge@1.0.0, parseurl@1.3.0, debug@2.2.0, finalhandler@0.4.0)
├── useragent@2.1.7 (lru-cache@2.2.4)
├── expand-braces@0.1.2 (array-unique@0.2.1, array-slice@0.2.3, braces@0.1.5)
├── bluebird@2.10.2
├── body-parser@1.14.1 (bytes@2.1.0, content-type@1.0.1, depd@1.1.0, qs@5.1.0, http-errors@1.3.1, on-finished@2.3.0, raw-body@2.1.4, debug@2.2.0, iconv-lite@0.4.12, type-is@1.6.9)
├── source-map@0.5.3
├── chokidar@1.2.0 (arrify@1.0.0, path-is-absolute@1.0.0, glob-parent@2.0.0, async-each@0.1.6, is-binary-path@1.0.1, is-glob@2.0.1, lodash.flatten@3.0.2, readdirp@2.0.0, anymatch@1.3.0)
├── log4js@0.6.27 (async@0.2.10, underscore@1.8.2, readable-stream@1.0.33, semver@4.3.6)
├── lodash@3.10.1
├── core-js@1.2.6
└── socket.io@1.3.7 (debug@2.1.0, has-binary-data@0.1.3, engine.io@1.5.4, socket.io-adapter@0.3.1, socket.io-parser@2.2.4, socket.io-client@1.3.7)
npm WARN peerDependencies The peer dependency jasmine-core@* included from karma-jasmine will no
npm WARN peerDependencies longer be automatically installed to fulfill the peerDependency
npm WARN peerDependencies in npm 3+. Your application will need to depend on it explicitly.
jasmine-core@2.3.4 TestApp\node_modules\jasmine-core
karma-jasmine@0.3.6 TestApp\node_modules\karma-jasmine
karma-nativescript-launcher@0.2.5 TestApp\node_modules\karma-nativescript-launcher
nativescript-unit-test-runner@0.2.8 TestApp\node_modules\nativescript-unit-test-runner
Successfully installed plugin nativescript-unit-test-runner.
Example test file created in app/tests/
Run your tests using the "$ tns test <platform>" command.
C:\Git>
C:\Git>
C:\Git>tns test android --path TestApp
16 11 2015 17:20:33.594:INFO [karma]: Karma v0.13.15 server started at http://localhost:9876/
16 11 2015 17:20:33.601:INFO [launcher]: Starting browser NativeScript Unit Test Runner
Copying template files...
spawn EBADF
Unknown command 'dev-test|android'. Try '$ tns help' for a full list of supported commands.
16 11 2015 17:22:08.033:INFO [launcher]: NativeScript deployment completed with code 127 | non_process | tns test fails with unknown command dev test android c git rm rf testapp c git tns create testapp project testapp was successfully created c git tns test init path testapp select testing framework jasmine npm warn package json no description npm warn package json no repository field npm warn package json no readme data karma testapp node modules karma ├── batch ├── di ├── rimraf ├── graceful fs ├── mime ├── colors ├── http proxy requires port ├── minimatch brace expansion ├── glob path is absolute inherits inflight once ├── dom serialize custom event void elements extend ent ├── optimist wordwrap minimist ├── connect utils merge parseurl debug finalhandler ├── useragent lru cache ├── expand braces array unique array slice braces ├── bluebird ├── body parser bytes content type depd qs http errors on finished raw body debug iconv lite type is ├── source map ├── chokidar arrify path is absolute glob parent async each is binary path is glob lodash flatten readdirp anymatch ├── async underscore readable stream semver ├── lodash ├── core js └── socket io debug has binary data engine io socket io adapter socket io parser socket io client npm warn peerdependencies the peer dependency jasmine core included from karma jasmine will no npm warn peerdependencies longer be automatically installed to fulfill the peerdependency npm warn peerdependencies in npm your application will need to depend on it explicitly jasmine core testapp node modules jasmine core karma jasmine testapp node modules karma jasmine karma nativescript launcher testapp node modules karma nativescript launcher nativescript unit test runner testapp node modules nativescript unit test runner successfully installed plugin nativescript unit test runner example test file created in app tests run your tests using the tns test command c git c git c git tns test android path testapp info karma server started at info starting browser nativescript unit test runner copying template files spawn ebadf unknown command dev test android try tns help for a full list of supported commands info nativescript deployment completed with code | 0 |
4,873 | 11,776,362,344 | IssuesEvent | 2020-03-16 13:08:09 | eclipse/omr | https://api.github.com/repos/eclipse/omr | opened | WIP: JitBuilder 2.0 proposal | architecture review pending comp:jitbuilder discussion enhancement | I would like to begin discussion for an enhancement to the JitBuilder API that is significant enough I have code named it "JitBuilder 2.0".
For the last two years, I have been collecting feedback and experiences with the use of the JitBuilder API, and this issue is an attempt to start to crystallize all of that into a more concrete proposal.
The process was kicked off at the OMR architecture meeting on March 13, which you can listen to here if you have an hour or so: https://youtu.be/ylJmi2xbCQc
I wanted to get this issue created, so people could track the effort. Stay tuned for more information! | 1.0 | WIP: JitBuilder 2.0 proposal - I would like to begin discussion for an enhancement to the JitBuilder API that is significant enough I have code named it "JitBuilder 2.0".
For the last two years, I have been collecting feedback and experiences with the use of the JitBuilder API, and this issue is an attempt to start to crystallize all of that into a more concrete proposal.
The process was kicked off at the OMR architecture meeting on March 13, which you can listen to here if you have an hour or so: https://youtu.be/ylJmi2xbCQc
I wanted to get this issue created, so people could track the effort. Stay tuned for more information! | non_process | wip jitbuilder proposal i would like to begin discussion for an enhancement to the jitbuilder api that is significant enough i have code named it jitbuilder for the last two years i have been collecting feedback and experiences with the use of the jitbuilder api and this issue is an attempt to start to crystallize all of that into a more concrete proposal the process was kicked off at the omr architecture meeting on march which you can listen to here if you have an hour or so i wanted to get this issue created so people could track the effort stay tuned for more information | 0 |
112,038 | 24,231,168,113 | IssuesEvent | 2022-09-26 18:25:19 | FerretDB/FerretDB | https://api.github.com/repos/FerretDB/FerretDB | closed | Fuzz `wire` package with recorded data | code/chore fuzz | When FerretDB is run with `task run` or similar commands, it stores recorded client sessions in the `records/` directory. We should use them in our `wire` package fuzzing. See `fuzzMessages` function in tests.
We should read all files in this directory, read wire messages in each file and use `f.Add` to add them to the seed corpus one-by-one.
No changes in CI configuration, etc. | 1.0 | Fuzz `wire` package with recorded data - When FerretDB is run with `task run` or similar commands, it stores recorded client sessions in the `records/` directory. We should use them in our `wire` package fuzzing. See `fuzzMessages` function in tests.
We should read all files in this directory, read wire messages in each file and use `f.Add` to add them to the seed corpus one-by-one.
No changes in CI configuration, etc. | non_process | fuzz wire package with recorded data when ferretdb is run with task run or similar commands it stores recorded client sessions in the records directory we should use them in our wire package fuzzing see fuzzmessages function in tests we should read all files in this directory read wire messages in each file and use f add to add them to the seed corpus one by one no changes in ci configuration etc | 0 |
4,592 | 7,432,096,857 | IssuesEvent | 2018-03-25 21:08:19 | brucemiller/LaTeXML | https://api.github.com/repos/brucemiller/LaTeXML | closed | Latexmlc (LaTeXML 0.8.2) epub generation fails on Windows 10 | bug postprocessing | There are multiple issues with epub generation.
Command line latexmlc --dest "%1.epub" "%1.tex" produces an epub with only mimetype file and META-INF folder containing container.xml.
The same script when run through TeXWorks throws an error
Can't locate loadable object for module XML::LibXML in @INC (@INC contains: C:\Strawberry\perl\site\lib D:/texlive/2016/tlpkg/tlperl/lib d:/texlive/2016/tlpkg/tlperl/site/lib d:/texlive/2016/tlpkg/tlperl/lib .) at C:\Strawberry\perl\site\lib/XML/LibXML.pm line 156.
and does not generate any epub.
On removing the path definition d:\texlive\2016\tlpkg\tlperl\bin from TeXWorks preferences, produces a new error
Cwd.c: loadable library and perl binaries are mismatched (got handshake key 0AA80080, needed 0AF00080) | 1.0 | Latexmlc (LaTeXML 0.8.2) epub generation fails on Windows 10 - There are multiple issues with epub generation.
Command line latexmlc --dest "%1.epub" "%1.tex" produces an epub with only mimetype file and META-INF folder containing container.xml.
The same script when run through TeXWorks throws an error
Can't locate loadable object for module XML::LibXML in @INC (@INC contains: C:\Strawberry\perl\site\lib D:/texlive/2016/tlpkg/tlperl/lib d:/texlive/2016/tlpkg/tlperl/site/lib d:/texlive/2016/tlpkg/tlperl/lib .) at C:\Strawberry\perl\site\lib/XML/LibXML.pm line 156.
and does not generate any epub.
On removing the path definition d:\texlive\2016\tlpkg\tlperl\bin from TeXWorks preferences, produces a new error
Cwd.c: loadable library and perl binaries are mismatched (got handshake key 0AA80080, needed 0AF00080) | process | latexmlc latexml epub generation fails on windows there are multiple issues with epub generation command line latexmlc dest epub tex produces an epub with only mimetype file and meta inf folder containing container xml the same script when run through texworks throws an error can t locate loadable object for module xml libxml in inc inc contains c strawberry perl site lib d texlive tlpkg tlperl lib d texlive tlpkg tlperl site lib d texlive tlpkg tlperl lib at c strawberry perl site lib xml libxml pm line and does not generate any epub on removing the path definition d texlive tlpkg tlperl bin from texworks preferences produces a new error cwd c loadable library and perl binaries are mismatched got handshake key needed | 1 |
11,145 | 13,957,692,756 | IssuesEvent | 2020-10-24 08:10:47 | alexanderkotsev/geoportal | https://api.github.com/repos/alexanderkotsev/geoportal | opened | BE: harvesting | BE - Belgium Geoportal Harvesting process | Dear Angelo,
I hope you are all right. I have a question about harvesting: I don't remember what we finally decided in Ispra but I noticed we haven't been harvested since January. Should I command harvesting by myself or could you do it by yourself ?
Regards,
Benoît | 1.0 | BE: harvesting - Dear Angelo,
I hope you are all right. I have a question about harvesting: I don't remember what we finally decided in Ispra but I noticed we haven't been harvested since January. Should I command harvesting by myself or could you do it by yourself ?
Regards,
Benoît | process | be harvesting dear angelo i hope you are all right i have a question about harvesting i don t remember what we finally decided in ispra but i noticed we haven t been harvested since january should i command harvesting by myself or could you do it by yourself regards beno icirc t | 1 |
11,567 | 14,440,502,290 | IssuesEvent | 2020-12-07 15:38:57 | MicrosoftDocs/azure-devops-docs | https://api.github.com/repos/MicrosoftDocs/azure-devops-docs | reopened | Pipeline Trigger Issues | Pri1 devops-cicd-process/tech devops/prod doc-enhancement | In the section **YAML resource trigger issues support and traceability** it states there is a menu item in the pipeline definition page. Where on the page? Which pipeline definition, source or destination? Can we get a screenshot showing the menu item on the page? I've clicked on just about everything I can think of and not seeing it :(
https://docs.microsoft.com/en-us/azure/devops/pipelines/process/resources?view=azure-devops&tabs=schema#yaml-resource-trigger-issues-support-and-traceability
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: ee4ec9d0-e0d5-4fb4-7c3e-b84abfa290c2
* Version Independent ID: 3e2b80d9-30e5-0c48-49f0-4fcdfedf5eee
* Content: [Resources - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/resources?view=azure-devops&tabs=schema)
* Content Source: [docs/pipelines/process/resources.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/resources.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam** | 1.0 | Pipeline Trigger Issues - In the section **YAML resource trigger issues support and traceability** it states there is a menu item in the pipeline definition page. Where on the page? Which pipeline definition, source or destination? Can we get a screenshot showing the menu item on the page? I've clicked on just about everything I can think of and not seeing it :(
https://docs.microsoft.com/en-us/azure/devops/pipelines/process/resources?view=azure-devops&tabs=schema#yaml-resource-trigger-issues-support-and-traceability
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: ee4ec9d0-e0d5-4fb4-7c3e-b84abfa290c2
* Version Independent ID: 3e2b80d9-30e5-0c48-49f0-4fcdfedf5eee
* Content: [Resources - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/resources?view=azure-devops&tabs=schema)
* Content Source: [docs/pipelines/process/resources.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/resources.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam** | process | pipeline trigger issues in the section yaml resource trigger issues support and traceability it states there is a menu item in the pipeline definition page where on the page which pipeline definition source or destination can we get a screenshot showing the menu item on the page i ve clicked on just about everything i can think of and not seeing it document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam | 1 |
22,389 | 31,142,286,048 | IssuesEvent | 2023-08-16 01:44:16 | cypress-io/cypress | https://api.github.com/repos/cypress-io/cypress | closed | Flaky test: iframe input retains focus when browser is out of focus [chrome] | OS: linux process: flaky test topic: flake ❄️ stage: flake stale | ### Link to dashboard or CircleCI failure
https://app.circleci.com/pipelines/github/cypress-io/cypress/41761/workflows/4f6901d8-6339-4309-b284-e03055666f1c/jobs/1732048/tests#failed-test-0
### Link to failing test in GitHub
https://github.com/cypress-io/cypress/blob/develop/packages/data-context/test/unit/sources/GitDataSource.spec.ts#L46
### Analysis
<img width="964" alt="Screen Shot 2022-08-12 at 11 23 50 AM" src="https://user-images.githubusercontent.com/26726429/184420695-b9cc4553-61ad-496a-a9b2-d8b4b4dc1d94.png">
### Cypress Version
10.4.0
### Other
Search for this issue number in the codebase to find the test(s) skipped until this issue is fixed | 1.0 | Flaky test: iframe input retains focus when browser is out of focus [chrome] - ### Link to dashboard or CircleCI failure
https://app.circleci.com/pipelines/github/cypress-io/cypress/41761/workflows/4f6901d8-6339-4309-b284-e03055666f1c/jobs/1732048/tests#failed-test-0
### Link to failing test in GitHub
https://github.com/cypress-io/cypress/blob/develop/packages/data-context/test/unit/sources/GitDataSource.spec.ts#L46
### Analysis
<img width="964" alt="Screen Shot 2022-08-12 at 11 23 50 AM" src="https://user-images.githubusercontent.com/26726429/184420695-b9cc4553-61ad-496a-a9b2-d8b4b4dc1d94.png">
### Cypress Version
10.4.0
### Other
Search for this issue number in the codebase to find the test(s) skipped until this issue is fixed | process | flaky test iframe input retains focus when browser is out of focus link to dashboard or circleci failure link to failing test in github analysis img width alt screen shot at am src cypress version other search for this issue number in the codebase to find the test s skipped until this issue is fixed | 1 |
12,527 | 14,968,753,301 | IssuesEvent | 2021-01-27 17:14:11 | googleapis/python-bigquery | https://api.github.com/repos/googleapis/python-bigquery | opened | Run samples tests with Python 3.9, too | type: process | Now that `pyarrow` extra supports Python 3.9, BigQuery can support it, too. BigQuery samples should thus include Python 3.9 in their nox test sessions.
Requires #462 (BigQuery with Python 3.9 support needs to be released first).
| 1.0 | Run samples tests with Python 3.9, too - Now that `pyarrow` extra supports Python 3.9, BigQuery can support it, too. BigQuery samples should thus include Python 3.9 in their nox test sessions.
Requires #462 (BigQuery with Python 3.9 support needs to be released first).
| process | run samples tests with python too now that pyarrow extra supports python bigquery can support it too bigquery samples should thus include python in their nox test sessions requires bigquery with python support needs to be released first | 1 |
7,972 | 11,156,967,382 | IssuesEvent | 2019-12-25 09:58:46 | bisq-network/bisq | https://api.github.com/repos/bisq-network/bisq | closed | A Timeout occurred | an:investigation in:trade-process was:dropped | Attaching .log file as requested by the client.
Error when taking offer to sell BTC for BSQ
An error occurred when taking the offer.
A timeout occurred.
[TradeProtocolTimeoutreached-bisq.log](https://github.com/bisq-network/bisq/files/3542524/TradeProtocolTimeoutreached-bisq.log)
Edit update: I got the timeout error attempting the same offer 3 times. I switched to a different offer and the trade confirmed successfully.
| 1.0 | A Timeout occurred - Attaching .log file as requested by the client.
Error when taking offer to sell BTC for BSQ
An error occurred when taking the offer.
A timeout occurred.
[TradeProtocolTimeoutreached-bisq.log](https://github.com/bisq-network/bisq/files/3542524/TradeProtocolTimeoutreached-bisq.log)
Edit update: I got the timeout error attempting the same offer 3 times. I switched to a different offer and the trade confirmed successfully.
| process | a timeout occurred attaching log file as requested by the client error when taking offer to sell btc for bsq an error occurred when taking the offer a timeout occurred edit update i got the timeout error attempting the same offer times i switched to a different offer and the trade confirmed successfully | 1 |
36,122 | 12,396,873,049 | IssuesEvent | 2020-05-20 21:23:49 | rammatzkvosky/888 | https://api.github.com/repos/rammatzkvosky/888 | opened | CVE-2020-9547 (High) detected in jackson-databind-2.8.8.jar | security vulnerability | ## CVE-2020-9547 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.8.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /tmp/ws-scm/888/pom.xml</p>
<p>Path to vulnerable library: epository/com/fasterxml/jackson/core/jackson-databind/2.8.8/jackson-databind-2.8.8.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.8.8.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/rammatzkvosky/888/commit/26c91f41d9c3b8d120fec834c102a2d177b36807">26c91f41d9c3b8d120fec834c102a2d177b36807</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to com.ibatis.sqlmap.engine.transaction.jta.JtaTransactionConfig (aka ibatis-sqlmap).
<p>Publish Date: 2020-03-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9547>CVE-2020-9547</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-9547">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-9547</a></p>
<p>Release Date: 2020-03-02</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.10.3</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.8","isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.8.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.10.3"}],"vulnerabilityIdentifier":"CVE-2020-9547","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to com.ibatis.sqlmap.engine.transaction.jta.JtaTransactionConfig (aka ibatis-sqlmap).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9547","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2020-9547 (High) detected in jackson-databind-2.8.8.jar - ## CVE-2020-9547 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.8.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /tmp/ws-scm/888/pom.xml</p>
<p>Path to vulnerable library: epository/com/fasterxml/jackson/core/jackson-databind/2.8.8/jackson-databind-2.8.8.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.8.8.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/rammatzkvosky/888/commit/26c91f41d9c3b8d120fec834c102a2d177b36807">26c91f41d9c3b8d120fec834c102a2d177b36807</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to com.ibatis.sqlmap.engine.transaction.jta.JtaTransactionConfig (aka ibatis-sqlmap).
<p>Publish Date: 2020-03-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9547>CVE-2020-9547</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-9547">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-9547</a></p>
<p>Release Date: 2020-03-02</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.10.3</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.8","isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.8.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.10.3"}],"vulnerabilityIdentifier":"CVE-2020-9547","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to com.ibatis.sqlmap.engine.transaction.jta.JtaTransactionConfig (aka ibatis-sqlmap).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9547","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_process | cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file tmp ws scm pom xml path to vulnerable library epository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to com ibatis sqlmap engine transaction jta jtatransactionconfig aka ibatis sqlmap publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to com ibatis sqlmap engine transaction jta jtatransactionconfig aka ibatis sqlmap vulnerabilityurl | 0 |
9,831 | 12,827,978,829 | IssuesEvent | 2020-07-06 19:35:16 | googleapis/code-suggester | https://api.github.com/repos/googleapis/code-suggester | opened | Framework-core: handle when access token does not have permissions for a fork-PR | type: process | - [ ] Error message for when there is an access token that does not have forking permissions
- [ ] Error message when the token does not have permissions to make a branch on that fork
- [ ] Error message when the token does not have permissions to write to that branch on its fork
- [ ] Error message when the token does not have permissions to create a PR from its fork to an upstream-repository
### Description
An access token without permissions will be rejected
### Solution
On the error messages received from the [GitHub API](https://developer.github.com/v3) gracefully fail. | 1.0 | Framework-core: handle when access token does not have permissions for a fork-PR - - [ ] Error message for when there is an access token that does not have forking permissions
- [ ] Error message when the token does not have permissions to make a branch on that fork
- [ ] Error message when the token does not have permissions to write to that branch on its fork
- [ ] Error message when the token does not have permissions to create a PR from its fork to an upstream-repository
### Description
An access token without permissions will be rejected
### Solution
On the error messages received from the [GitHub API](https://developer.github.com/v3) gracefully fail. | process | framework core handle when access token does not have permissions for a fork pr error message for when there is an access token that does not have forking permissions error message when the token does not have permissions to make a branch on that fork error message when the token does not have permissions to write to that branch on its fork error message when the token does not have permissions to create a pr from its fork to an upstream repository description an access token without permissions will be rejected solution on the error messages received from the gracefully fail | 1 |
18,146 | 24,186,818,559 | IssuesEvent | 2022-09-23 13:58:14 | geneontology/go-ontology | https://api.github.com/repos/geneontology/go-ontology | closed | NTR disruption of cellular component of another organism | New term request multi-species process | To group
GO:0051673 membrane disruption in another organism
GO:0044278 cell wall disruption in another organism'
(these were under '[GO:0035821 modulation of process of another organism](http://amigo.geneontology.org/amigo/term/GO:0035821#display-lineage-tab)' | 1.0 | NTR disruption of cellular component of another organism - To group
GO:0051673 membrane disruption in another organism
GO:0044278 cell wall disruption in another organism'
(these were under '[GO:0035821 modulation of process of another organism](http://amigo.geneontology.org/amigo/term/GO:0035821#display-lineage-tab)' | process | ntr disruption of cellular component of another organism to group go membrane disruption in another organism go cell wall disruption in another organism these were under | 1 |
68,659 | 13,167,020,017 | IssuesEvent | 2020-08-11 09:32:41 | mozilla-mobile/fenix | https://api.github.com/repos/mozilla-mobile/fenix | reopened | FNX3-15859 ⁃ [Bug] issue with QR code scanner if Fenix does not yet have camera permission | Feature: QRCode 🐞 bug | ### Steps to reproduce
1. make sure in your Android app permissions that Fenix does not yet have permission to use the camera
2. use the QR code scanner of Fenix
=> A dialog appears to give Fenix the permission to use the camera
3. Give Fenix the permission
### Expected behavior
After giving the permission the QR code scanner should be active.
### Actual behavior
The "Scan" button has a visible "active" state but the QR code scanner is not active. Instead you have to press again the "Scan" button.
### Device information
* Android device: HTC U11 / Android 8.0
* Fenix version: Fenix revision 080dadcb7e1cc0aa53c020595505f6f493e4a143 | 1.0 | FNX3-15859 ⁃ [Bug] issue with QR code scanner if Fenix does not yet have camera permission - ### Steps to reproduce
1. make sure in your Android app permissions that Fenix does not yet have permission to use the camera
2. use the QR code scanner of Fenix
=> A dialog appears to give Fenix the permission to use the camera
3. Give Fenix the permission
### Expected behavior
After giving the permission the QR code scanner should be active.
### Actual behavior
The "Scan" button has a visible "active" state but the QR code scanner is not active. Instead you have to press again the "Scan" button.
### Device information
* Android device: HTC U11 / Android 8.0
* Fenix version: Fenix revision 080dadcb7e1cc0aa53c020595505f6f493e4a143 | non_process | ⁃ issue with qr code scanner if fenix does not yet have camera permission steps to reproduce make sure in your android app permissions that fenix does not yet have permission to use the camera use the qr code scanner of fenix a dialog appears to give fenix the permission to use the camera give fenix the permission expected behavior after giving the permission the qr code scanner should be active actual behavior the scan button has a visible active state but the qr code scanner is not active instead you have to press again the scan button device information android device htc android fenix version fenix revision | 0 |
12,528 | 14,969,402,883 | IssuesEvent | 2021-01-27 18:06:04 | hashgraph/hedera-mirror-node | https://api.github.com/repos/hashgraph/hedera-mirror-node | closed | Integrate REST API build into Maven | P3 enhancement process rest | **Problem**
Currently, if you want to run or build the entirety of the mirror node, you have to install multiple tools and run multiple commands. To make it easier for the community to run, we should reduce the number of external tools like npm that they have to install and make it more self contained. There should be a single command to run the build lifecycle.
**Solution**
The Maven wrapper should be the only command needed to build the mirror node:
- Use frontend-maven-plugin to:
- Install node and npm
- npm install hedera-mirror-rest
- npm test hedera-mirror-rest
- npm package hedera-mirror-rest
- npm install monitoring_api
- npm test monitoring_api
- Remove npm specific commands or jobs from CircleCI config
- Update documentation to replace npm commands with maven, if applicable
**Alternatives**
Keep as is
**Additional Context**
| 1.0 | Integrate REST API build into Maven - **Problem**
Currently, if you want to run or build the entirety of the mirror node, you have to install multiple tools and run multiple commands. To make it easier for the community to run, we should reduce the number of external tools like npm that they have to install and make it more self contained. There should be a single command to run the build lifecycle.
**Solution**
The Maven wrapper should be the only command needed to build the mirror node:
- Use frontend-maven-plugin to:
- Install node and npm
- npm install hedera-mirror-rest
- npm test hedera-mirror-rest
- npm package hedera-mirror-rest
- npm install monitoring_api
- npm test monitoring_api
- Remove npm specific commands or jobs from CircleCI config
- Update documentation to replace npm commands with maven, if applicable
**Alternatives**
Keep as is
**Additional Context**
| process | integrate rest api build into maven problem currently if you want to run or build the entirety of the mirror node you have to install multiple tools and run multiple commands to make it easier for the community to run we should reduce the number of external tools like npm that they have to install and make it more self contained there should be a single command to run the build lifecycle solution the maven wrapper should be the only command needed to build the mirror node use frontend maven plugin to install node and npm npm install hedera mirror rest npm test hedera mirror rest npm package hedera mirror rest npm install monitoring api npm test monitoring api remove npm specific commands or jobs from circleci config update documentation to replace npm commands with maven if applicable alternatives keep as is additional context | 1 |
19,173 | 25,281,889,538 | IssuesEvent | 2022-11-16 16:20:03 | sillsdev/silnlp | https://api.github.com/repos/sillsdev/silnlp | opened | Error messages when files are missing. | enhancement good first issue pipeline 3: preprocess | I'm getting this error when trying to run an experiment on ClearML. It seems to occur when a required Scripture file is missing.
It looks like Issue #85
```
File "/root/.clearml/venvs-builds/3.8/task_repository/silnlp/.venv/lib/python3.8/site-packages/transformers/trainer.py", line 841, in get_train_dataloader
raise ValueError("Trainer: training requires a train_dataset.")
ValueError: Trainer: training requires a train_dataset.
2022-11-16 12:49:39
Process failed, exit code 1
``` | 1.0 | Error messages when files are missing. - I'm getting this error when trying to run an experiment on ClearML. It seems to occur when a required Scripture file is missing.
It looks like Issue #85
```
File "/root/.clearml/venvs-builds/3.8/task_repository/silnlp/.venv/lib/python3.8/site-packages/transformers/trainer.py", line 841, in get_train_dataloader
raise ValueError("Trainer: training requires a train_dataset.")
ValueError: Trainer: training requires a train_dataset.
2022-11-16 12:49:39
Process failed, exit code 1
``` | process | error messages when files are missing i m getting this error when trying to run an experiment on clearml it seems to occur when a required scripture file is missing it looks like issue file root clearml venvs builds task repository silnlp venv lib site packages transformers trainer py line in get train dataloader raise valueerror trainer training requires a train dataset valueerror trainer training requires a train dataset process failed exit code | 1 |
11,924 | 14,703,943,695 | IssuesEvent | 2021-01-04 15:46:22 | GetTerminus/terminus-oss | https://api.github.com/repos/GetTerminus/terminus-oss | closed | Add top-level changelog pointing to sub changelogs | Focus: community Goal: Process Improvement Type: chore | So that consumers can quickly see there are changelogs despite not having one for the repo as a whole.
ie https://github.com/firebase/firebase-js-sdk/blob/master/CHANGELOG.md | 1.0 | Add top-level changelog pointing to sub changelogs - So that consumers can quickly see there are changelogs despite not having one for the repo as a whole.
ie https://github.com/firebase/firebase-js-sdk/blob/master/CHANGELOG.md | process | add top level changelog pointing to sub changelogs so that consumers can quickly see there are changelogs despite not having one for the repo as a whole ie | 1 |
147,351 | 11,782,836,962 | IssuesEvent | 2020-03-17 03:21:18 | pytorch/pytorch | https://api.github.com/repos/pytorch/pytorch | closed | network-based TorchHub tests contribute to CI flakiness | high priority topic: flaky-tests triage review | Perhaps this suite of tests can be run on nightly instead of on each `master` commit?
See this [example failure on `master` branch](https://circleci.com/gh/pytorch/pytorch/4763382):
```
Mar 11 05:06:21 ======================================================================
Mar 11 05:06:21 ERROR [0.403s]: test_download_url_to_file (__main__.TestHub)
Mar 11 05:06:21 ----------------------------------------------------------------------
Mar 11 05:06:21 Traceback (most recent call last):
Mar 11 05:06:21 File "test_utils.py", line 547, in test_download_url_to_file
Mar 11 05:06:21 hub.download_url_to_file(TORCHHUB_EXAMPLE_RELEASE_URL, temp_file, progress=False)
Mar 11 05:06:21 File "/opt/conda/lib/python3.6/site-packages/torch/hub.py", line 394, in download_url_to_file
Mar 11 05:06:21 u = urlopen(url)
Mar 11 05:06:21 File "/opt/conda/lib/python3.6/urllib/request.py", line 223, in urlopen
Mar 11 05:06:21 return opener.open(url, data, timeout)
Mar 11 05:06:21 File "/opt/conda/lib/python3.6/urllib/request.py", line 532, in open
Mar 11 05:06:21 response = meth(req, response)
Mar 11 05:06:21 File "/opt/conda/lib/python3.6/urllib/request.py", line 642, in http_response
Mar 11 05:06:21 'http', request, response, code, msg, hdrs)
Mar 11 05:06:21 File "/opt/conda/lib/python3.6/urllib/request.py", line 570, in error
Mar 11 05:06:21 return self._call_chain(*args)
Mar 11 05:06:21 File "/opt/conda/lib/python3.6/urllib/request.py", line 504, in _call_chain
Mar 11 05:06:21 result = func(*args)
Mar 11 05:06:21 File "/opt/conda/lib/python3.6/urllib/request.py", line 650, in http_error_default
Mar 11 05:06:21 raise HTTPError(req.full_url, code, msg, hdrs, fp)
Mar 11 05:06:21 urllib.error.HTTPError: HTTP Error 401: UNAUTHORIZED
```
cc @ezyang @gchanan @zou3519 @ailzhang, @suo | 1.0 | network-based TorchHub tests contribute to CI flakiness - Perhaps this suite of tests can be run on nightly instead of on each `master` commit?
See this [example failure on `master` branch](https://circleci.com/gh/pytorch/pytorch/4763382):
```
Mar 11 05:06:21 ======================================================================
Mar 11 05:06:21 ERROR [0.403s]: test_download_url_to_file (__main__.TestHub)
Mar 11 05:06:21 ----------------------------------------------------------------------
Mar 11 05:06:21 Traceback (most recent call last):
Mar 11 05:06:21 File "test_utils.py", line 547, in test_download_url_to_file
Mar 11 05:06:21 hub.download_url_to_file(TORCHHUB_EXAMPLE_RELEASE_URL, temp_file, progress=False)
Mar 11 05:06:21 File "/opt/conda/lib/python3.6/site-packages/torch/hub.py", line 394, in download_url_to_file
Mar 11 05:06:21 u = urlopen(url)
Mar 11 05:06:21 File "/opt/conda/lib/python3.6/urllib/request.py", line 223, in urlopen
Mar 11 05:06:21 return opener.open(url, data, timeout)
Mar 11 05:06:21 File "/opt/conda/lib/python3.6/urllib/request.py", line 532, in open
Mar 11 05:06:21 response = meth(req, response)
Mar 11 05:06:21 File "/opt/conda/lib/python3.6/urllib/request.py", line 642, in http_response
Mar 11 05:06:21 'http', request, response, code, msg, hdrs)
Mar 11 05:06:21 File "/opt/conda/lib/python3.6/urllib/request.py", line 570, in error
Mar 11 05:06:21 return self._call_chain(*args)
Mar 11 05:06:21 File "/opt/conda/lib/python3.6/urllib/request.py", line 504, in _call_chain
Mar 11 05:06:21 result = func(*args)
Mar 11 05:06:21 File "/opt/conda/lib/python3.6/urllib/request.py", line 650, in http_error_default
Mar 11 05:06:21 raise HTTPError(req.full_url, code, msg, hdrs, fp)
Mar 11 05:06:21 urllib.error.HTTPError: HTTP Error 401: UNAUTHORIZED
```
cc @ezyang @gchanan @zou3519 @ailzhang, @suo | non_process | network based torchhub tests contribute to ci flakiness perhaps this suite of tests can be run on nightly instead of on each master commit see this mar mar error test download url to file main testhub mar mar traceback most recent call last mar file test utils py line in test download url to file mar hub download url to file torchhub example release url temp file progress false mar file opt conda lib site packages torch hub py line in download url to file mar u urlopen url mar file opt conda lib urllib request py line in urlopen mar return opener open url data timeout mar file opt conda lib urllib request py line in open mar response meth req response mar file opt conda lib urllib request py line in http response mar http request response code msg hdrs mar file opt conda lib urllib request py line in error mar return self call chain args mar file opt conda lib urllib request py line in call chain mar result func args mar file opt conda lib urllib request py line in http error default mar raise httperror req full url code msg hdrs fp mar urllib error httperror http error unauthorized cc ezyang gchanan ailzhang suo | 0 |
8,112 | 11,301,081,137 | IssuesEvent | 2020-01-17 14:55:07 | Open-EO/openeo-api | https://api.github.com/repos/Open-EO/openeo-api | closed | UDF Usage | feedback required help wanted interoperability no solution yet processes udfs | I'm currently trying to catch up with the current UDF implementations for R and Python to change/improve the R implementation.
While going through the issues on the UDFs and related issues in the openeo-processes I think we have major problems with users getting into UDFs. As I see it stated in the issues on Open-EO/openeo-processes#42 and Open-EO/openeo-udf#10 the information an user who wants to implement a successfully running UDF is lacking the following:
1. The dimension specification of the data / data cube at a given point during the execution of a process graph
2. How the data has to be in terms of the dimension specification, when the UDF result is injected back into the data stream of the process graph
3. How they can test their UDF <s>implementation</s> scripts (where: endpoint, with what: sample data, what is expected to be returned)
4. Wheter the data is chunked (they don't work with the whole dataset, e.g. a spatial tile as a raster time series)
Therefore I would like to bring up the following ideas for discussion:
- 1.1 Track the changes on the dimensions like running a metadata execution before running the process graph with actual data. This might be done by the client or the backend. This information must be accessible for the clients, because then we can create sample data.
- 2.1. Implement some sort of constraint on the dimensions of the data cube import into process
- 2.2. The UDF as such has to declare its returned data in terms of ist dimensions
- 3.1. I'm not sure if this is solveable at the moment. We should have some testing endpoint for UDFs, but the problem is that the UDF runtime is mostly part of a specific back-end.
I have no possible solution for 4 yet, since this will be mostly an optimization within the back-ends, when it comes to parallelization. | 1.0 | UDF Usage - I'm currently trying to catch up with the current UDF implementations for R and Python to change/improve the R implementation.
While going through the issues on the UDFs and related issues in the openeo-processes I think we have major problems with users getting into UDFs. As I see it stated in the issues on Open-EO/openeo-processes#42 and Open-EO/openeo-udf#10 the information an user who wants to implement a successfully running UDF is lacking the following:
1. The dimension specification of the data / data cube at a given point during the execution of a process graph
2. How the data has to be in terms of the dimension specification, when the UDF result is injected back into the data stream of the process graph
3. How they can test their UDF <s>implementation</s> scripts (where: endpoint, with what: sample data, what is expected to be returned)
4. Wheter the data is chunked (they don't work with the whole dataset, e.g. a spatial tile as a raster time series)
Therefore I would like to bring up the following ideas for discussion:
- 1.1 Track the changes on the dimensions like running a metadata execution before running the process graph with actual data. This might be done by the client or the backend. This information must be accessible for the clients, because then we can create sample data.
- 2.1. Implement some sort of constraint on the dimensions of the data cube import into process
- 2.2. The UDF as such has to declare its returned data in terms of ist dimensions
- 3.1. I'm not sure if this is solveable at the moment. We should have some testing endpoint for UDFs, but the problem is that the UDF runtime is mostly part of a specific back-end.
I have no possible solution for 4 yet, since this will be mostly an optimization within the back-ends, when it comes to parallelization. | process | udf usage i m currently trying to catch up with the current udf implementations for r and python to change improve the r implementation while going through the issues on the udfs and related issues in the openeo processes i think we have major problems with users getting into udfs as i see it stated in the issues on open eo openeo processes and open eo openeo udf the information an user who wants to implement a successfully running udf is lacking the following the dimension specification of the data data cube at a given point during the execution of a process graph how the data has to be in terms of the dimension specification when the udf result is injected back into the data stream of the process graph how they can test their udf implementation scripts where endpoint with what sample data what is expected to be returned wheter the data is chunked they don t work with the whole dataset e g a spatial tile as a raster time series therefore i would like to bring up the following ideas for discussion track the changes on the dimensions like running a metadata execution before running the process graph with actual data this might be done by the client or the backend this information must be accessible for the clients because then we can create sample data implement some sort of constraint on the dimensions of the data cube import into process the udf as such has to declare its returned data in terms of ist dimensions i m not sure if this is solveable at the moment we should have some testing endpoint for udfs but the problem is that the udf runtime is mostly part of a specific back end i have no possible solution for yet since this will be mostly an optimization within the back ends when it comes to parallelization | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.