Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
189,677
| 6,800,668,521
|
IssuesEvent
|
2017-11-02 14:39:54
|
orange-alliance/the-orange-alliance
|
https://api.github.com/repos/orange-alliance/the-orange-alliance
|
opened
|
Allow API Key in the Header
|
enhancement High Priority
|
Allow the API Read Keys in the Header
DO NOT ADD TO THE WRITE PATHS
|
1.0
|
Allow API Key in the Header - Allow the API Read Keys in the Header
DO NOT ADD TO THE WRITE PATHS
|
non_process
|
allow api key in the header allow the api read keys in the header do not add to the write paths
| 0
|
3,090
| 6,106,524,094
|
IssuesEvent
|
2017-06-21 04:39:28
|
kerubistan/kerub
|
https://api.github.com/repos/kerubistan/kerub
|
closed
|
add a UI feature to add bmc info to the host
|
component:data processing component:ui enhancement
|
* type (ipmi/redfish)
* address
* authentication
|
1.0
|
add a UI feature to add bmc info to the host - * type (ipmi/redfish)
* address
* authentication
|
process
|
add a ui feature to add bmc info to the host type ipmi redfish address authentication
| 1
|
104,263
| 16,613,566,306
|
IssuesEvent
|
2021-06-02 14:15:35
|
Thanraj/linux-4.1.15
|
https://api.github.com/repos/Thanraj/linux-4.1.15
|
opened
|
CVE-2018-7191 (Medium) detected in linux-stable-rtv4.1.33, linuxlinux-4.1.17
|
security vulnerability
|
## CVE-2018-7191 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linux-stable-rtv4.1.33</b>, <b>linuxlinux-4.1.17</b></p></summary>
<p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In the tun subsystem in the Linux kernel before 4.13.14, dev_get_valid_name is not called before register_netdevice. This allows local users to cause a denial of service (NULL pointer dereference and panic) via an ioctl(TUNSETIFF) call with a dev name containing a / character. This is similar to CVE-2013-4343.
<p>Publish Date: 2019-05-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-7191>CVE-2018-7191</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-7191">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-7191</a></p>
<p>Release Date: 2019-05-17</p>
<p>Fix Resolution: v4.14-rc6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2018-7191 (Medium) detected in linux-stable-rtv4.1.33, linuxlinux-4.1.17 - ## CVE-2018-7191 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linux-stable-rtv4.1.33</b>, <b>linuxlinux-4.1.17</b></p></summary>
<p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In the tun subsystem in the Linux kernel before 4.13.14, dev_get_valid_name is not called before register_netdevice. This allows local users to cause a denial of service (NULL pointer dereference and panic) via an ioctl(TUNSETIFF) call with a dev name containing a / character. This is similar to CVE-2013-4343.
<p>Publish Date: 2019-05-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-7191>CVE-2018-7191</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-7191">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-7191</a></p>
<p>Release Date: 2019-05-17</p>
<p>Fix Resolution: v4.14-rc6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in linux stable linuxlinux cve medium severity vulnerability vulnerable libraries linux stable linuxlinux vulnerability details in the tun subsystem in the linux kernel before dev get valid name is not called before register netdevice this allows local users to cause a denial of service null pointer dereference and panic via an ioctl tunsetiff call with a dev name containing a character this is similar to cve publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
19,867
| 26,278,141,214
|
IssuesEvent
|
2023-01-07 02:04:19
|
rusefi/rusefi_documentation
|
https://api.github.com/repos/rusefi/rusefi_documentation
|
closed
|
What's the plan for wiki2-human https://github.com/rusefi/rusefi/wiki
|
wiki location & process change
|
Once we move to https://wiki.rusefi.com/ do we want to remove or adjust https://github.com/rusefi/rusefi/wiki in order to reduce potential confusion between https://github.com/rusefi/rusefi/wiki and https://wiki.rusefi.com/?
to some expend that depends on if we can easily support both or not
|
1.0
|
What's the plan for wiki2-human https://github.com/rusefi/rusefi/wiki - Once we move to https://wiki.rusefi.com/ do we want to remove or adjust https://github.com/rusefi/rusefi/wiki in order to reduce potential confusion between https://github.com/rusefi/rusefi/wiki and https://wiki.rusefi.com/?
to some expend that depends on if we can easily support both or not
|
process
|
what s the plan for human once we move to do we want to remove or adjust in order to reduce potential confusion between and to some expend that depends on if we can easily support both or not
| 1
|
409,487
| 27,740,366,356
|
IssuesEvent
|
2023-03-15 13:58:56
|
eic/EICrecon
|
https://api.github.com/repos/eic/EICrecon
|
closed
|
jana-generate plugin quibbles
|
part:documentation
|
### Environment: (where does this bug occur, have you tried other environments)
- Which branch (often `main` for latest released): All
- Which version (or `HEAD` for the most recent on git):all
- Any specific OS or system where the issue occurs?no
- Any special versions of ROOT or Geant4?no
### Steps to reproduce: (give a step by step account of how to trigger the bug)
1. jana-generate.py Plugin
### Expected Result: (what do you expect when you execute the steps above)
I expect skeleton code, with only lines of code that are needed in (commented out examples are fine.
### Actual Result: (what do you get when you execute the steps above)
I see a defined string "genfit" which ma not be necessary and a TFile/TDirectory which may not e needed. The Example histogram is actually there. Init does many things including define the histogram. In process the filling of this histogram is commented out. I'd expect either it is commented out everywhere or in everywhere. You could have an unobservant user who compiles this skeleton and run it. It might be shocking to to see it produce an empty root file (but wait the histogram is defined)
|
1.0
|
jana-generate plugin quibbles - ### Environment: (where does this bug occur, have you tried other environments)
- Which branch (often `main` for latest released): All
- Which version (or `HEAD` for the most recent on git):all
- Any specific OS or system where the issue occurs?no
- Any special versions of ROOT or Geant4?no
### Steps to reproduce: (give a step by step account of how to trigger the bug)
1. jana-generate.py Plugin
### Expected Result: (what do you expect when you execute the steps above)
I expect skeleton code, with only lines of code that are needed in (commented out examples are fine.
### Actual Result: (what do you get when you execute the steps above)
I see a defined string "genfit" which ma not be necessary and a TFile/TDirectory which may not e needed. The Example histogram is actually there. Init does many things including define the histogram. In process the filling of this histogram is commented out. I'd expect either it is commented out everywhere or in everywhere. You could have an unobservant user who compiles this skeleton and run it. It might be shocking to to see it produce an empty root file (but wait the histogram is defined)
|
non_process
|
jana generate plugin quibbles environment where does this bug occur have you tried other environments which branch often main for latest released all which version or head for the most recent on git all any specific os or system where the issue occurs no any special versions of root or no steps to reproduce give a step by step account of how to trigger the bug jana generate py plugin expected result what do you expect when you execute the steps above i expect skeleton code with only lines of code that are needed in commented out examples are fine actual result what do you get when you execute the steps above i see a defined string genfit which ma not be necessary and a tfile tdirectory which may not e needed the example histogram is actually there init does many things including define the histogram in process the filling of this histogram is commented out i d expect either it is commented out everywhere or in everywhere you could have an unobservant user who compiles this skeleton and run it it might be shocking to to see it produce an empty root file but wait the histogram is defined
| 0
|
215,672
| 16,614,424,830
|
IssuesEvent
|
2021-06-02 15:03:59
|
RedisBloom/redisbloom-go
|
https://api.github.com/repos/RedisBloom/redisbloom-go
|
opened
|
Count-Min Sketch creation, ingestion, and querying command examples
|
documentation
|
Sample example that can be used as reference:
https://github.com/RedisBloom/redisbloom-go/blob/master/example_client_test.go#L11
Further reference for commands:
https://oss.redislabs.com/redisbloom/CountMinSketch_Commands/
to check locally the documentation effects
```
# start http service
godoc -http=:6060
# open browser tab with
http://localhost:6060/pkg/github.com/RedisBloom/redisbloom-go/
```
|
1.0
|
Count-Min Sketch creation, ingestion, and querying command examples - Sample example that can be used as reference:
https://github.com/RedisBloom/redisbloom-go/blob/master/example_client_test.go#L11
Further reference for commands:
https://oss.redislabs.com/redisbloom/CountMinSketch_Commands/
to check locally the documentation effects
```
# start http service
godoc -http=:6060
# open browser tab with
http://localhost:6060/pkg/github.com/RedisBloom/redisbloom-go/
```
|
non_process
|
count min sketch creation ingestion and querying command examples sample example that can be used as reference further reference for commands to check locally the documentation effects start http service godoc http open browser tab with
| 0
|
31,216
| 25,448,846,523
|
IssuesEvent
|
2022-11-24 08:53:28
|
dotnet/roslyn
|
https://api.github.com/repos/dotnet/roslyn
|
closed
|
[Automated] PRs inserted in VS build main-33123.395
|
Area-Infrastructure untriaged vs-insertion
|
[View Complete Diff of Changes](https://github.com/dotnet/roslyn/compare/3f65b818a94d74f2230bda40382ea755702fe674...413319eb370210c93296ea90aa73987e41ee8521?w=1)
- [Add asserts for debug only test failures and fix tests (65565)](https://github.com/dotnet/roslyn/pull/65565)
- [Make sure the container syntax node is not GlobalStatements (65570)](https://github.com/dotnet/roslyn/pull/65570)
- [Increase CodeCleanUp test value (65575)](https://github.com/dotnet/roslyn/pull/65575)
- [Move to multi-column primary keys for our sqlite database. (65553)](https://github.com/dotnet/roslyn/pull/65553)
- [Update to latest version of protocol with new pull diagnostic types (65514)](https://github.com/dotnet/roslyn/pull/65514)
- [Semantic Snippets - Fix snippet priority in completion list (65103)](https://github.com/dotnet/roslyn/pull/65103)
- [[EnC] Allow reordering of top level statements (65560)](https://github.com/dotnet/roslyn/pull/65560)
- [Fix handling blocks in top level code. (65557)](https://github.com/dotnet/roslyn/pull/65557)
- [Add additional cases whre 'use coalesce expression' can simplify code. (65371)](https://github.com/dotnet/roslyn/pull/65371)
- [Fix MEF composition used for the LSIF generator (65512)](https://github.com/dotnet/roslyn/pull/65512)
- [Results of running the arrow-placement analyzer on roslyn (65475)](https://github.com/dotnet/roslyn/pull/65475)
- [Fix null ref in add-import (65549)](https://github.com/dotnet/roslyn/pull/65549)
- [Simplify how options are checked in the tagger. (65543)](https://github.com/dotnet/roslyn/pull/65543)
- [Fix spelling mistake (65541)](https://github.com/dotnet/roslyn/pull/65541)
- [Add experimental formatting option to control placement of `=>` (65476)](https://github.com/dotnet/roslyn/pull/65476)
- [[Port] Gracefully handle additional locations in DiagnosticData.Create (65532)](https://github.com/dotnet/roslyn/pull/65532)
- [Delete duplicate "SystemThreadingChannelsVersion" (65539)](https://github.com/dotnet/roslyn/pull/65539)
- [Fix bug in partial solutions after a document was removed and added (65349)](https://github.com/dotnet/roslyn/pull/65349)
- [Fix a NullReferenceException in Signature help (65372)](https://github.com/dotnet/roslyn/pull/65372)
- [Avoid ordering dependency in active statement map (65525)](https://github.com/dotnet/roslyn/pull/65525)
- [Ensure lvalue struct receivers are not copied by value due to interpolated string handler rewrite (65505)](https://github.com/dotnet/roslyn/pull/65505)
- [Do not import the VSWorkspace during Roslyn package initialization (65492)](https://github.com/dotnet/roslyn/pull/65492)
- [Optimize codegen for tuple swap scenarios (65327)](https://github.com/dotnet/roslyn/pull/65327)
- [Improve `NormalizeWhitespace` for object initializers (65249)](https://github.com/dotnet/roslyn/pull/65249)
- [Move context-sensitive parsing to context-free parsing. (65480)](https://github.com/dotnet/roslyn/pull/65480)
- [Add A/B test to allow us to turn off background compilation/parsing (65489)](https://github.com/dotnet/roslyn/pull/65489)
- [Run the conditional-token-placement fixer on roslyn (65469)](https://github.com/dotnet/roslyn/pull/65469)
- [Check type parameter index bounds when parsing `DocumentationCommentId` (65457)](https://github.com/dotnet/roslyn/pull/65457)
- [Convert parser warning into a binding warning. (65440)](https://github.com/dotnet/roslyn/pull/65440)
- [View Call Hierarchy - fix thread block (65452)](https://github.com/dotnet/roslyn/pull/65452)
- [Enhance rule metadata and suppression info in SARIF V2 errorlog (64277)](https://github.com/dotnet/roslyn/pull/64277)
- [Move IDS_FeatureXXX checks out of the parser (65413)](https://github.com/dotnet/roslyn/pull/65413)
- [Warn for unused parameters in source (65466)](https://github.com/dotnet/roslyn/pull/65466)
- [Pass in HostServices when creating the LSP server (65384)](https://github.com/dotnet/roslyn/pull/65384)
|
1.0
|
[Automated] PRs inserted in VS build main-33123.395 - [View Complete Diff of Changes](https://github.com/dotnet/roslyn/compare/3f65b818a94d74f2230bda40382ea755702fe674...413319eb370210c93296ea90aa73987e41ee8521?w=1)
- [Add asserts for debug only test failures and fix tests (65565)](https://github.com/dotnet/roslyn/pull/65565)
- [Make sure the container syntax node is not GlobalStatements (65570)](https://github.com/dotnet/roslyn/pull/65570)
- [Increase CodeCleanUp test value (65575)](https://github.com/dotnet/roslyn/pull/65575)
- [Move to multi-column primary keys for our sqlite database. (65553)](https://github.com/dotnet/roslyn/pull/65553)
- [Update to latest version of protocol with new pull diagnostic types (65514)](https://github.com/dotnet/roslyn/pull/65514)
- [Semantic Snippets - Fix snippet priority in completion list (65103)](https://github.com/dotnet/roslyn/pull/65103)
- [[EnC] Allow reordering of top level statements (65560)](https://github.com/dotnet/roslyn/pull/65560)
- [Fix handling blocks in top level code. (65557)](https://github.com/dotnet/roslyn/pull/65557)
- [Add additional cases whre 'use coalesce expression' can simplify code. (65371)](https://github.com/dotnet/roslyn/pull/65371)
- [Fix MEF composition used for the LSIF generator (65512)](https://github.com/dotnet/roslyn/pull/65512)
- [Results of running the arrow-placement analyzer on roslyn (65475)](https://github.com/dotnet/roslyn/pull/65475)
- [Fix null ref in add-import (65549)](https://github.com/dotnet/roslyn/pull/65549)
- [Simplify how options are checked in the tagger. (65543)](https://github.com/dotnet/roslyn/pull/65543)
- [Fix spelling mistake (65541)](https://github.com/dotnet/roslyn/pull/65541)
- [Add experimental formatting option to control placement of `=>` (65476)](https://github.com/dotnet/roslyn/pull/65476)
- [[Port] Gracefully handle additional locations in DiagnosticData.Create (65532)](https://github.com/dotnet/roslyn/pull/65532)
- [Delete duplicate "SystemThreadingChannelsVersion" (65539)](https://github.com/dotnet/roslyn/pull/65539)
- [Fix bug in partial solutions after a document was removed and added (65349)](https://github.com/dotnet/roslyn/pull/65349)
- [Fix a NullReferenceException in Signature help (65372)](https://github.com/dotnet/roslyn/pull/65372)
- [Avoid ordering dependency in active statement map (65525)](https://github.com/dotnet/roslyn/pull/65525)
- [Ensure lvalue struct receivers are not copied by value due to interpolated string handler rewrite (65505)](https://github.com/dotnet/roslyn/pull/65505)
- [Do not import the VSWorkspace during Roslyn package initialization (65492)](https://github.com/dotnet/roslyn/pull/65492)
- [Optimize codegen for tuple swap scenarios (65327)](https://github.com/dotnet/roslyn/pull/65327)
- [Improve `NormalizeWhitespace` for object initializers (65249)](https://github.com/dotnet/roslyn/pull/65249)
- [Move context-sensitive parsing to context-free parsing. (65480)](https://github.com/dotnet/roslyn/pull/65480)
- [Add A/B test to allow us to turn off background compilation/parsing (65489)](https://github.com/dotnet/roslyn/pull/65489)
- [Run the conditional-token-placement fixer on roslyn (65469)](https://github.com/dotnet/roslyn/pull/65469)
- [Check type parameter index bounds when parsing `DocumentationCommentId` (65457)](https://github.com/dotnet/roslyn/pull/65457)
- [Convert parser warning into a binding warning. (65440)](https://github.com/dotnet/roslyn/pull/65440)
- [View Call Hierarchy - fix thread block (65452)](https://github.com/dotnet/roslyn/pull/65452)
- [Enhance rule metadata and suppression info in SARIF V2 errorlog (64277)](https://github.com/dotnet/roslyn/pull/64277)
- [Move IDS_FeatureXXX checks out of the parser (65413)](https://github.com/dotnet/roslyn/pull/65413)
- [Warn for unused parameters in source (65466)](https://github.com/dotnet/roslyn/pull/65466)
- [Pass in HostServices when creating the LSP server (65384)](https://github.com/dotnet/roslyn/pull/65384)
|
non_process
|
prs inserted in vs build main allow reordering of top level statements gracefully handle additional locations in diagnosticdata create
| 0
|
18,637
| 25,953,548,394
|
IssuesEvent
|
2022-12-17 23:12:57
|
ikemen-engine/Ikemen-GO
|
https://api.github.com/repos/ikemen-engine/Ikemen-GO
|
closed
|
Different handling of stage zooming
|
bug compatibility
|
I've just noticed that stages that use tensionhigh and tensionlow have a different zoom behavior compared to Mugen 1.1. In Mugen, the stage will only zoom out if the characters are far apart, whereas in Ikemen it zooms out regardless of that.
Here's Mugen 1.1:

The characters are high above but the camera doesn't zoom out, because they are close to each other.
Ikemen GO (current repo):

Same scenario but the camera zooms out.
Here's the stage used:
[stage.zip](https://github.com/ikemen-engine/Ikemen-GO/files/9946463/stage.zip)
|
True
|
Different handling of stage zooming - I've just noticed that stages that use tensionhigh and tensionlow have a different zoom behavior compared to Mugen 1.1. In Mugen, the stage will only zoom out if the characters are far apart, whereas in Ikemen it zooms out regardless of that.
Here's Mugen 1.1:

The characters are high above but the camera doesn't zoom out, because they are close to each other.
Ikemen GO (current repo):

Same scenario but the camera zooms out.
Here's the stage used:
[stage.zip](https://github.com/ikemen-engine/Ikemen-GO/files/9946463/stage.zip)
|
non_process
|
different handling of stage zooming i ve just noticed that stages that use tensionhigh and tensionlow have a different zoom behavior compared to mugen in mugen the stage will only zoom out if the characters are far apart whereas in ikemen it zooms out regardless of that here s mugen the characters are high above but the camera doesn t zoom out because they are close to each other ikemen go current repo same scenario but the camera zooms out here s the stage used
| 0
|
17,596
| 23,424,464,549
|
IssuesEvent
|
2022-08-14 07:08:36
|
Battle-s/battle-school-backend
|
https://api.github.com/repos/Battle-s/battle-school-backend
|
closed
|
[FEAT] 학교 entity 생성
|
feature :computer: processing :hourglass_flowing_sand:
|
## 설명
## 체크사항
- [ ] 학교 관련 핵심 로직 설계 및 작성
- [ ] 회원가입 로직에 적용
## 참고자료
## 관련 논의
|
1.0
|
[FEAT] 학교 entity 생성 - ## 설명
## 체크사항
- [ ] 학교 관련 핵심 로직 설계 및 작성
- [ ] 회원가입 로직에 적용
## 참고자료
## 관련 논의
|
process
|
학교 entity 생성 설명 체크사항 학교 관련 핵심 로직 설계 및 작성 회원가입 로직에 적용 참고자료 관련 논의
| 1
|
63,639
| 12,359,948,475
|
IssuesEvent
|
2020-05-17 13:16:37
|
home-assistant/brands
|
https://api.github.com/repos/home-assistant/brands
|
closed
|
Foursquare is missing brand images
|
domain-missing has-codeowner
|
## The problem
The Foursquare integration does not have brand images in
this repository.
We recently started this Brands repository, to create a centralized storage of all brand-related images. These images are used on our website and the Home Assistant frontend.
The following images are missing and would ideally be added:
- `src/foursquare/icon.png`
- `src/foursquare/logo.png`
- `src/foursquare/icon@2x.png`
- `src/foursquare/logo@2x.png`
For image specifications and requirements, please see [README.md](https://github.com/home-assistant/brands/blob/master/README.md).
## Updating the documentation repository
Our documentation repository already has a logo for this integration, however, it does not meet the image requirements of this new Brands repository.
If adding images to this repository, please open up a PR to the documentation repository as well, removing the `logo: foursquare.png` line from this file:
<https://github.com/home-assistant/home-assistant.io/blob/current/source/_integrations/foursquare.markdown>
**Note**: The documentation PR needs to be opened against the `current` branch.
**Note2**: Please leave the actual logo file in the documentation repository. It will be cleaned up differently.
## Additional information
For more information about this repository, read the [README.md](https://github.com/home-assistant/brands/blob/master/README.md) file of this repository. It contains information on how this repository works, and image specification and requirements.
## Codeowner mention
Hi there, @robbiet480! Mind taking a look at this issue as it is with an integration (foursquare) you are listed as a [codeowner](https://github.com/home-assistant/core/blob/dev/homeassistant/components/foursquare/manifest.json) for? Thanks!
Resolving this issue is not limited to codeowners! If you want to help us out, feel free to resolve this issue! Thanks already!
|
1.0
|
Foursquare is missing brand images -
## The problem
The Foursquare integration does not have brand images in
this repository.
We recently started this Brands repository, to create a centralized storage of all brand-related images. These images are used on our website and the Home Assistant frontend.
The following images are missing and would ideally be added:
- `src/foursquare/icon.png`
- `src/foursquare/logo.png`
- `src/foursquare/icon@2x.png`
- `src/foursquare/logo@2x.png`
For image specifications and requirements, please see [README.md](https://github.com/home-assistant/brands/blob/master/README.md).
## Updating the documentation repository
Our documentation repository already has a logo for this integration, however, it does not meet the image requirements of this new Brands repository.
If adding images to this repository, please open up a PR to the documentation repository as well, removing the `logo: foursquare.png` line from this file:
<https://github.com/home-assistant/home-assistant.io/blob/current/source/_integrations/foursquare.markdown>
**Note**: The documentation PR needs to be opened against the `current` branch.
**Note2**: Please leave the actual logo file in the documentation repository. It will be cleaned up differently.
## Additional information
For more information about this repository, read the [README.md](https://github.com/home-assistant/brands/blob/master/README.md) file of this repository. It contains information on how this repository works, and image specification and requirements.
## Codeowner mention
Hi there, @robbiet480! Mind taking a look at this issue as it is with an integration (foursquare) you are listed as a [codeowner](https://github.com/home-assistant/core/blob/dev/homeassistant/components/foursquare/manifest.json) for? Thanks!
Resolving this issue is not limited to codeowners! If you want to help us out, feel free to resolve this issue! Thanks already!
|
non_process
|
foursquare is missing brand images the problem the foursquare integration does not have brand images in this repository we recently started this brands repository to create a centralized storage of all brand related images these images are used on our website and the home assistant frontend the following images are missing and would ideally be added src foursquare icon png src foursquare logo png src foursquare icon png src foursquare logo png for image specifications and requirements please see updating the documentation repository our documentation repository already has a logo for this integration however it does not meet the image requirements of this new brands repository if adding images to this repository please open up a pr to the documentation repository as well removing the logo foursquare png line from this file note the documentation pr needs to be opened against the current branch please leave the actual logo file in the documentation repository it will be cleaned up differently additional information for more information about this repository read the file of this repository it contains information on how this repository works and image specification and requirements codeowner mention hi there mind taking a look at this issue as it is with an integration foursquare you are listed as a for thanks resolving this issue is not limited to codeowners if you want to help us out feel free to resolve this issue thanks already
| 0
|
275,974
| 8,582,861,629
|
IssuesEvent
|
2018-11-13 18:06:02
|
sunjun-group/Ziyuan
|
https://api.github.com/repos/sunjun-group/Ziyuan
|
opened
|
Program Stuck
|
high priority
|
Hi Lyly,
Please check this, the program stuck for the following method:
org.apache.commons.math.analysis.integration.SimpsonIntegrator.integrate.70
|
1.0
|
Program Stuck - Hi Lyly,
Please check this, the program stuck for the following method:
org.apache.commons.math.analysis.integration.SimpsonIntegrator.integrate.70
|
non_process
|
program stuck hi lyly please check this the program stuck for the following method org apache commons math analysis integration simpsonintegrator integrate
| 0
|
435,139
| 30,487,264,873
|
IssuesEvent
|
2023-07-18 04:03:18
|
wehs7661/ensemble_md
|
https://api.github.com/repos/wehs7661/ensemble_md
|
closed
|
Enable coordinate modification in the EEXE framework
|
documentation enhancement
|
To expand the usage of EEXE, we want to enable coordinate manipulation at exchanges between replicas, which is most likely to be useful for estimating the free energy of multiple serial mutations using expanded ensemble simulations, such as mutating methane into ethane and then propane.
For example, we can have an EEXE simulation composed of two replicas mutating methane into ethane and ethane into propane, respectively, and only exchange the coordinates between replicas when they are at the end states, i.e., replica 1 being at λ=1 and replica 2 being at λ=0. In this example, we will have the following end states:
- Replica 1: Mutating methane to ethane
- State a: At λ = 0, we have methane with a dummy methyl group.
- State b: At λ = 1, we have ethane with a dummy H atom.
- Replica 2: Mutating ethane to propane
- State c: At λ = 0, we have ethane with a dummy methyl group.
- State d: At λ = 1, we have propane with a dummy H atom.
At exchanges, we will have two output `gro` files respectively from replicas 1 and 2, namely `rep1.gro` (state b, ethane with a dummy H atom at the first carbon) and `rep2.gro` (state c, ethane with a dummy ethyl group at the second carbon).
Note that in EEXE, each replica is bound to the transformation for its assigned alchemical range. In our case, this means that replica 1 will only be responsible for the mutation of a methane to an ethane, and replica 2 will only be responsible for mutating an ethane to a propane. Normally, we would just swap the `gro` files as is, so in the next iteration, replica 1 will be initialized with `rep2.gro` and sample the intermediate states along the mutation path between methane and ethane. However, `rep2.gro` is an ethane with a dummy methyl group, not an ethane with a dummy H atom that we need for such sampling. The same thing would happen when trying to initialize the next iteration of replica 2 using `rep1.gro`.
To address this issue, we can modify `rep2.gro` as follows and use it to proceed to the next iteration of replica 1:
- Remove the dummy methyl group at the second carbon atom from `rep2.gro`.
- Attach a dummy H atom to the first carbon atom in `rep2.gro`. Specifically, the coordinate of the dummy H atom can just be the coordinates of the second carbon atom. There won't be clashes since the dummy H atoms have no interactions with the rest of the system.
Similarly, we can modify `rep1.gro` as follows for the next iteration of replica 2:
- Remove the dummy H atom at the first carbon from `rep1.gro`.
- Attach a dummy methyl group at the second carbon atom in `rep1.gro`. Specifically, we can take the internal coordinates of the methyl group in `rep2.gro`, treat the group as rigid, rotate, and attach the group to the second carbon atom in `rep1.gro`.
Importantly, we can make the two modified `gro` files have the same potential energy, so the proposed exchange will always be adopted.
Here, we are not going to implement functions for coordinate manipulation in EEXE but modify the CLI `run_EEXE` (and the function `run_grompp` in `ensemble_EXE.py`, if necessary) to allow the flexibility of calling a user-defined function for coordinate manipulation from an input python module (where the user-defined function is defined).
|
1.0
|
Enable coordinate modification in the EEXE framework - To expand the usage of EEXE, we want to enable coordinate manipulation at exchanges between replicas, which is most likely to be useful for estimating the free energy of multiple serial mutations using expanded ensemble simulations, such as mutating methane into ethane and then propane.
For example, we can have an EEXE simulation composed of two replicas mutating methane into ethane and ethane into propane, respectively, and only exchange the coordinates between replicas when they are at the end states, i.e., replica 1 being at λ=1 and replica 2 being at λ=0. In this example, we will have the following end states:
- Replica 1: Mutating methane to ethane
- State a: At λ = 0, we have methane with a dummy methyl group.
- State b: At λ = 1, we have ethane with a dummy H atom.
- Replica 2: Mutating ethane to propane
- State c: At λ = 0, we have ethane with a dummy methyl group.
- State d: At λ = 1, we have propane with a dummy H atom.
At exchanges, we will have two output `gro` files respectively from replicas 1 and 2, namely `rep1.gro` (state b, ethane with a dummy H atom at the first carbon) and `rep2.gro` (state c, ethane with a dummy ethyl group at the second carbon).
Note that in EEXE, each replica is bound to the transformation for its assigned alchemical range. In our case, this means that replica 1 will only be responsible for the mutation of a methane to an ethane, and replica 2 will only be responsible for mutating an ethane to a propane. Normally, we would just swap the `gro` files as is, so in the next iteration, replica 1 will be initialized with `rep2.gro` and sample the intermediate states along the mutation path between methane and ethane. However, `rep2.gro` is an ethane with a dummy methyl group, not an ethane with a dummy H atom that we need for such sampling. The same thing would happen when trying to initialize the next iteration of replica 2 using `rep1.gro`.
To address this issue, we can modify `rep2.gro` as follows and use it to proceed to the next iteration of replica 1:
- Remove the dummy methyl group at the second carbon atom from `rep2.gro`.
- Attach a dummy H atom to the first carbon atom in `rep2.gro`. Specifically, the coordinate of the dummy H atom can just be the coordinates of the second carbon atom. There won't be clashes since the dummy H atoms have no interactions with the rest of the system.
Similarly, we can modify `rep1.gro` as follows for the next iteration of replica 2:
- Remove the dummy H atom at the first carbon from `rep1.gro`.
- Attach a dummy methyl group at the second carbon atom in `rep1.gro`. Specifically, we can take the internal coordinates of the methyl group in `rep2.gro`, treat the group as rigid, rotate, and attach the group to the second carbon atom in `rep1.gro`.
Importantly, we can make the two modified `gro` files have the same potential energy, so the proposed exchange will always be adopted.
Here, we are not going to implement functions for coordinate manipulation in EEXE but modify the CLI `run_EEXE` (and the function `run_grompp` in `ensemble_EXE.py`, if necessary) to allow the flexibility of calling a user-defined function for coordinate manipulation from an input python module (where the user-defined function is defined).
|
non_process
|
enable coordinate modification in the eexe framework to expand the usage of eexe we want to enable coordinate manipulation at exchanges between replicas which is most likely to be useful for estimating the free energy of multiple serial mutations using expanded ensemble simulations such as mutating methane into ethane and then propane for example we can have an eexe simulation composed of two replicas mutating methane into ethane and ethane into propane respectively and only exchange the coordinates between replicas when they are at the end states i e replica being at λ and replica being at λ in this example we will have the following end states replica mutating methane to ethane state a at λ we have methane with a dummy methyl group state b at λ we have ethane with a dummy h atom replica mutating ethane to propane state c at λ we have ethane with a dummy methyl group state d at λ we have propane with a dummy h atom at exchanges we will have two output gro files respectively from replicas and namely gro state b ethane with a dummy h atom at the first carbon and gro state c ethane with a dummy ethyl group at the second carbon note that in eexe each replica is bound to the transformation for its assigned alchemical range in our case this means that replica will only be responsible for the mutation of a methane to an ethane and replica will only be responsible for mutating an ethane to a propane normally we would just swap the gro files as is so in the next iteration replica will be initialized with gro and sample the intermediate states along the mutation path between methane and ethane however gro is an ethane with a dummy methyl group not an ethane with a dummy h atom that we need for such sampling the same thing would happen when trying to initialize the next iteration of replica using gro to address this issue we can modify gro as follows and use it to proceed to the next iteration of replica remove the dummy methyl group at the second carbon atom from gro attach a dummy h atom to the first carbon atom in gro specifically the coordinate of the dummy h atom can just be the coordinates of the second carbon atom there won t be clashes since the dummy h atoms have no interactions with the rest of the system similarly we can modify gro as follows for the next iteration of replica remove the dummy h atom at the first carbon from gro attach a dummy methyl group at the second carbon atom in gro specifically we can take the internal coordinates of the methyl group in gro treat the group as rigid rotate and attach the group to the second carbon atom in gro importantly we can make the two modified gro files have the same potential energy so the proposed exchange will always be adopted here we are not going to implement functions for coordinate manipulation in eexe but modify the cli run eexe and the function run grompp in ensemble exe py if necessary to allow the flexibility of calling a user defined function for coordinate manipulation from an input python module where the user defined function is defined
| 0
|
4,858
| 7,746,517,683
|
IssuesEvent
|
2018-05-29 22:02:05
|
AppFolioOnboarding/image-sharer-ChaoHuangAtAppfolio
|
https://api.github.com/repos/AppFolioOnboarding/image-sharer-ChaoHuangAtAppfolio
|
closed
|
Image Index
|
in process
|
#### As a user I want the homepage to display all saved images.
__Story__:
Now that we are saving images for our users, we want them to be able to view
the list of images in the system. You might be thinking, "what if my users only
want to view images of cats", or "what if my users only want to see the images
they uploaded". Those are great questions. Maybe you will have the opportunity
to add those features in a future story. For now, we will keep it simple and
quickly deliver a little more value to our users.
__Acceptance criteria__:
- [ ] New images that are added show up on the homepage.
- [ ] These images are persisted if the browser is closed or even if the
server is restarted.
- [ ] Images are not displayed wider than 400px.
- [ ] Newest images appear first.
__Dependencies__:
- Save Image Link
|
1.0
|
Image Index - #### As a user I want the homepage to display all saved images.
__Story__:
Now that we are saving images for our users, we want them to be able to view
the list of images in the system. You might be thinking, "what if my users only
want to view images of cats", or "what if my users only want to see the images
they uploaded". Those are great questions. Maybe you will have the opportunity
to add those features in a future story. For now, we will keep it simple and
quickly deliver a little more value to our users.
__Acceptance criteria__:
- [ ] New images that are added show up on the homepage.
- [ ] These images are persisted if the browser is closed or even if the
server is restarted.
- [ ] Images are not displayed wider than 400px.
- [ ] Newest images appear first.
__Dependencies__:
- Save Image Link
|
process
|
image index as a user i want the homepage to display all saved images story now that we are saving images for our users we want them to be able to view the list of images in the system you might be thinking what if my users only want to view images of cats or what if my users only want to see the images they uploaded those are great questions maybe you will have the opportunity to add those features in a future story for now we will keep it simple and quickly deliver a little more value to our users acceptance criteria new images that are added show up on the homepage these images are persisted if the browser is closed or even if the server is restarted images are not displayed wider than newest images appear first dependencies save image link
| 1
|
268,113
| 8,403,618,993
|
IssuesEvent
|
2018-10-11 10:17:23
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
gtc.lm.com - desktop site instead of mobile site
|
browser-firefox priority-normal
|
<!-- @browser: Firefox 63.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 6.3; Win64; x64; rv:63.0) Gecko/20100101 Firefox/63.0 -->
<!-- @reported_with: desktop-reporter -->
**URL**: https://gtc.lm.com/LM/Security/login?Jwt=/rL3cPEzRImDYcQqSxP5+j+sCq7T21DltWy8WKZq7E397TBPTruGYSiGZ86MQ2y8T40BPVJZOw4UrLX5ZgHQQwlIvw19HzJt5zhn2LgtEg7qdvPurNvUztiUlXoCZWyJNfToeEUKLn0Jxw0WqollxFKwjSKLR3N/nYJFT0uu1CGlSG+zo/U1INeJcCn5S/rbtpMy7+kZ4evWiW29pa3fewZgyHNkPS4hGcEEGYG4uUcNGRBUz7xzXCP2Sfelz5dRFM3KqR+MVBjgkd+Es2R2TwBW2cWaTwSi5TzktPwAC9fQ14cycs0N0IoWDxC3QyzPsqhN5lcM0Hy8bV1+qEFbdOlMghyeeqdB0I/V/iVEFV8=
**Browser / Version**: Firefox 63.0
**Operating System**: Windows 8.1
**Tested Another Browser**: Yes
**Problem type**: Desktop site instead of mobile site
**Description**: its not redirecting right
**Steps to Reproduce**:
i try several severs
<details>
<summary>Browser Configuration</summary>
<ul>
<li>mixed active content blocked: false</li><li>buildID: 20181004174654</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.all: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>channel: beta</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
gtc.lm.com - desktop site instead of mobile site - <!-- @browser: Firefox 63.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 6.3; Win64; x64; rv:63.0) Gecko/20100101 Firefox/63.0 -->
<!-- @reported_with: desktop-reporter -->
**URL**: https://gtc.lm.com/LM/Security/login?Jwt=/rL3cPEzRImDYcQqSxP5+j+sCq7T21DltWy8WKZq7E397TBPTruGYSiGZ86MQ2y8T40BPVJZOw4UrLX5ZgHQQwlIvw19HzJt5zhn2LgtEg7qdvPurNvUztiUlXoCZWyJNfToeEUKLn0Jxw0WqollxFKwjSKLR3N/nYJFT0uu1CGlSG+zo/U1INeJcCn5S/rbtpMy7+kZ4evWiW29pa3fewZgyHNkPS4hGcEEGYG4uUcNGRBUz7xzXCP2Sfelz5dRFM3KqR+MVBjgkd+Es2R2TwBW2cWaTwSi5TzktPwAC9fQ14cycs0N0IoWDxC3QyzPsqhN5lcM0Hy8bV1+qEFbdOlMghyeeqdB0I/V/iVEFV8=
**Browser / Version**: Firefox 63.0
**Operating System**: Windows 8.1
**Tested Another Browser**: Yes
**Problem type**: Desktop site instead of mobile site
**Description**: its not redirecting right
**Steps to Reproduce**:
i try several severs
<details>
<summary>Browser Configuration</summary>
<ul>
<li>mixed active content blocked: false</li><li>buildID: 20181004174654</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.all: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>channel: beta</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_process
|
gtc lm com desktop site instead of mobile site url browser version firefox operating system windows tested another browser yes problem type desktop site instead of mobile site description its not redirecting right steps to reproduce i try several severs browser configuration mixed active content blocked false buildid tracking content blocked false gfx webrender blob images true gfx webrender all false mixed passive content blocked false gfx webrender enabled false image mem shared true channel beta from with ❤️
| 0
|
20,956
| 27,817,074,295
|
IssuesEvent
|
2023-03-18 20:11:03
|
cse442-at-ub/project_s23-cinco
|
https://api.github.com/repos/cse442-at-ub/project_s23-cinco
|
closed
|
Connect database to react app.
|
Processing Task Sprint 2
|
Test
Navigate to https://www-student.cse.buffalo.edu/tools/db/phpmyadmin
Login to your account using your UBIT and UB person number
Go to CSE442 database "cse442_2023_spring_team_b_db", select Users table
Verify you can see username and password entries in database.
Prerequisite:
If you are not on campus and under UB wifi, you have to connect through VPN to access UB servers. Download AnyConnect here
https://www.buffalo.edu/ubit/service-guides/software/downloading/macintosh-software/managing-mac-software/anyconnect.html
and connect to vpn.buffalo.edu/UBVPN.
Test
Run Apache webserver using XAMPP.
Go to the root directory of the demo app and run "(sudo) npm start".
You are connected! Sadly, I don't think I can print out that you actually connected because echoing anything in the php file will assume it is the post response.
|
1.0
|
Connect database to react app. - Test
Navigate to https://www-student.cse.buffalo.edu/tools/db/phpmyadmin
Login to your account using your UBIT and UB person number
Go to CSE442 database "cse442_2023_spring_team_b_db", select Users table
Verify you can see username and password entries in database.
Prerequisite:
If you are not on campus and under UB wifi, you have to connect through VPN to access UB servers. Download AnyConnect here
https://www.buffalo.edu/ubit/service-guides/software/downloading/macintosh-software/managing-mac-software/anyconnect.html
and connect to vpn.buffalo.edu/UBVPN.
Test
Run Apache webserver using XAMPP.
Go to the root directory of the demo app and run "(sudo) npm start".
You are connected! Sadly, I don't think I can print out that you actually connected because echoing anything in the php file will assume it is the post response.
|
process
|
connect database to react app test navigate to login to your account using your ubit and ub person number go to database spring team b db select users table verify you can see username and password entries in database prerequisite if you are not on campus and under ub wifi you have to connect through vpn to access ub servers download anyconnect here and connect to vpn buffalo edu ubvpn test run apache webserver using xampp go to the root directory of the demo app and run sudo npm start you are connected sadly i don t think i can print out that you actually connected because echoing anything in the php file will assume it is the post response
| 1
|
22,010
| 30,513,683,550
|
IssuesEvent
|
2023-07-18 23:48:49
|
h4sh5/pypi-auto-scanner
|
https://api.github.com/repos/h4sh5/pypi-auto-scanner
|
opened
|
roblox-pyc 1.16.47 has 2 GuardDog issues
|
guarddog silent-process-execution
|
https://pypi.org/project/roblox-pyc
https://inspector.pypi.io/project/roblox-pyc
```{
"dependency": "roblox-pyc",
"version": "1.16.47",
"result": {
"issues": 2,
"errors": {},
"results": {
"silent-process-execution": [
{
"location": "roblox-pyc-1.16.47/src/robloxpy.py:115",
"code": " subprocess.call([\"luarocks\", \"--version\"], stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL, stdin=subprocess.DEVNULL)",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
},
{
"location": "roblox-pyc-1.16.47/src/robloxpy.py:122",
"code": " subprocess.call([\"moonc\", \"--version\"], stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL, stdin=subprocess.DEVNULL)",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
}
]
},
"path": "/tmp/tmp_ppbq5ad/roblox-pyc"
}
}```
|
1.0
|
roblox-pyc 1.16.47 has 2 GuardDog issues - https://pypi.org/project/roblox-pyc
https://inspector.pypi.io/project/roblox-pyc
```{
"dependency": "roblox-pyc",
"version": "1.16.47",
"result": {
"issues": 2,
"errors": {},
"results": {
"silent-process-execution": [
{
"location": "roblox-pyc-1.16.47/src/robloxpy.py:115",
"code": " subprocess.call([\"luarocks\", \"--version\"], stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL, stdin=subprocess.DEVNULL)",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
},
{
"location": "roblox-pyc-1.16.47/src/robloxpy.py:122",
"code": " subprocess.call([\"moonc\", \"--version\"], stdout=subprocess.DEVNULL, stderr=subprocess.DEVNULL, stdin=subprocess.DEVNULL)",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
}
]
},
"path": "/tmp/tmp_ppbq5ad/roblox-pyc"
}
}```
|
process
|
roblox pyc has guarddog issues dependency roblox pyc version result issues errors results silent process execution location roblox pyc src robloxpy py code subprocess call stdout subprocess devnull stderr subprocess devnull stdin subprocess devnull message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null location roblox pyc src robloxpy py code subprocess call stdout subprocess devnull stderr subprocess devnull stdin subprocess devnull message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null path tmp tmp roblox pyc
| 1
|
20,461
| 10,757,341,964
|
IssuesEvent
|
2019-10-31 13:06:49
|
humhub/humhub
|
https://api.github.com/repos/humhub/humhub
|
opened
|
Cache loaded ContentContainer
|
Kind:Enhancement Topic:API Topic:Performance
|
Sometimes there are many `$model->content->container` calls for different models. We should cache already loaded containers by `contentcontainer_id` in order to prevent reloading already fetched containers.
|
True
|
Cache loaded ContentContainer - Sometimes there are many `$model->content->container` calls for different models. We should cache already loaded containers by `contentcontainer_id` in order to prevent reloading already fetched containers.
|
non_process
|
cache loaded contentcontainer sometimes there are many model content container calls for different models we should cache already loaded containers by contentcontainer id in order to prevent reloading already fetched containers
| 0
|
4,664
| 7,497,255,108
|
IssuesEvent
|
2018-04-08 17:56:27
|
UnbFeelings/unb-feelings-GQA
|
https://api.github.com/repos/UnbFeelings/unb-feelings-GQA
|
opened
|
Analisar Processo e Artefatos
|
document process wiki
|
Analisar o processo definido pela [equipe de processo][e-processo] para identificar quais partes deste processo e quais artefatos serão auditados pela equipe GQA. Eu recomendo que a definição destes artefatos e processos tenham algum embasamento, seja ele por conta dos objetivos organizacionais, qualidade do produto, ou porque é necessário para a disciplina.
|
1.0
|
Analisar Processo e Artefatos - Analisar o processo definido pela [equipe de processo][e-processo] para identificar quais partes deste processo e quais artefatos serão auditados pela equipe GQA. Eu recomendo que a definição destes artefatos e processos tenham algum embasamento, seja ele por conta dos objetivos organizacionais, qualidade do produto, ou porque é necessário para a disciplina.
|
process
|
analisar processo e artefatos analisar o processo definido pela para identificar quais partes deste processo e quais artefatos serão auditados pela equipe gqa eu recomendo que a definição destes artefatos e processos tenham algum embasamento seja ele por conta dos objetivos organizacionais qualidade do produto ou porque é necessário para a disciplina
| 1
|
6,696
| 9,813,846,162
|
IssuesEvent
|
2019-06-13 08:55:53
|
cropmapteam/Scotland-crop-map
|
https://api.github.com/repos/cropmapteam/Scotland-crop-map
|
closed
|
Check / Edit digitised RFI polygon shapefiles
|
GIS process
|
The script #31 which does the masking out of manually digitised RFI locations provided as shapefiles does some validation before doing the masking in rasterio and detected some problems with some of the shapefiles / some of the records in some of the shapefiles.
There were 51 shapefiles provided. 43 images have been processed to set to nodata, all pixels that fell within the polygons held in each shapefile
why the difference of 8?
This seems to be due to the following 3 issues with the digitised RFI polygon shapefiles which need to be looked-at; confirmed and fixed.
**Issue 1:** there are multiple (2) digitised RFI polygons shapefiles associated with a single image.
3 tiff files seemed to each have 2 shapefiles associated with them. In some cases duplicate RFI features have been captured, one in each shapefile as shown here:

The script assumes 1 shapefile per image, so processing was skipped.
So for the image S1B_20180120_132_asc_174936_175001_DV_Gamma-0_GB_OSGB_RCTK_SpkRL.tif, there are these 2 shapefiles:
rfi/RFI_Interference_Shapefiles_Chrissy/S1B_20180120_132_asc_174936_175001_DV_Gamma-0_GB_OSGB_RCTK_SpkRL.shp
rfi/RFI_Interference_Shapefiles_Richard/S1B_20180120_132_asc_174936_175001_DV_Gamma-0_GB_OSGB_RCTK_SpkRL.shp
So for the image S1A_20180308_30_asc_175835_175900_DV_Gamma-0_GB_OSGB_RCTK_SpkRL.tif, there are these 2 shapefiles:
rfi/RFI_Interference_Shapefiles_Zara/ S1A_20180308_30_asc_175835_175900_DV_Gamma-0_GB_OSGB_RCTK_SpkRL.shp
rfi/RFI_Interference_Shapefiles_Chrissy/S1A_20180308_30_asc_175835_175900_DV_Gamma-0_GB_OSGB_RCTK_SpkRL.shp
So for the image S1A_20180107_30_asc_175836_175901_DV_Gamma-0_GB_OSGB_RCTK_SpkRL.tif, there are these 2 shapefiles:
rfi/RFI_Interference_Shapefiles_Chrissy/S1A_20180107_30_asc_175836_175901_DV_Gamma-0_GB_OSGB_RCTK_SpkRL.shp
rfi/RFI_Interference_Shapefiles_Richard/S1A_20180107_30_asc_175836_175901_DV_Gamma-0_GB_OSGB_RCTK_SpkRL.shp
Solution - in each case the shapefiles need to be checked and where duplicate RFI regions captured, a decision be made on which is best and the features from each of the 2 shapefiles then merged into a single shapefile.
**Issue 2:** Some of the digitised RFI polygons shapefiles contain no records at all or all the records contain invalid geometries.
For these images, either the associated shapefile contained no records at all or the records it did contain all had invalid geometries. Processing was skipped.
So for image S1B_20180309_132_asc_174935_175000_DV_Gamma-0_GB_OSGB_RCTK_SpkRL.tif the shapefile rfi/RFI_Interference_Shapefiles_Zara/S1B_20180309_132_asc_174935_175000_DV_Gamma-0_GB_OSGB_RCTK_SpkRL.shp seems to have no records
So for image S1B_20180618_30_asc_175812_175837_DV_Gamma-0_GB_OSGB_RCTK_SpkRL.tif the shapefile rfi/RFI_Interference_Shapefiles_Chrissy/S1B_20180618_30_asc_175812_175837_DV_Gamma-0_GB_OSGB_RCTK_SpkRL.shp seems to have no records with valid geometries
Solution - can the shapefiles / digitised polygons within them be checked?
**Issue3:** Some of the digitised RFI polygons shapefiles contain features with invalid geometries.
These seem to mostly be self-intersections like this:

In such cases, the pixels associated with polygons with valid geometries were masked out as normal and a new image created. The pixels associated with polygons with invalid geometries were skipped and the pixels retained in the new image.
This can be seen in this screenshot:

The pixels inside the valid polygons have been masked out. The pixels inside the invalid polygon have been retained. The following shapefiles were detected as containing features with invalid geometries:
Found invalid polygon with id 8 in rfi/RFI_Interference_Shapefiles_Chrissy/S1B_20180113_30_asc_175813_175838_DV_Gamma-0_GB_OSGB_RCTK_SpkRL.shp
Found invalid polygon with id 1 in rfi/RFI_Interference_Shapefiles_Zara/S1A_20180303_132_asc_175017_175042_DV_Gamma-0_GB_OSGB_RCTK_SpkRL.shp
Found invalid polygon with id 10 in rfi/RFI_Interference_Shapefiles_Chrissy/S1A_20180119_30_asc_175836_175901_DV_Gamma-0_GB_OSGB_RCTK_SpkRL.shp
Solution - can the digitised polygons be checked in these shapefiles. Once done, the image can be re-processed to mask out all the RFI regions.
|
1.0
|
Check / Edit digitised RFI polygon shapefiles - The script #31 which does the masking out of manually digitised RFI locations provided as shapefiles does some validation before doing the masking in rasterio and detected some problems with some of the shapefiles / some of the records in some of the shapefiles.
There were 51 shapefiles provided. 43 images have been processed to set to nodata, all pixels that fell within the polygons held in each shapefile
why the difference of 8?
This seems to be due to the following 3 issues with the digitised RFI polygon shapefiles which need to be looked-at; confirmed and fixed.
**Issue 1:** there are multiple (2) digitised RFI polygons shapefiles associated with a single image.
3 tiff files seemed to each have 2 shapefiles associated with them. In some cases duplicate RFI features have been captured, one in each shapefile as shown here:

The script assumes 1 shapefile per image, so processing was skipped.
So for the image S1B_20180120_132_asc_174936_175001_DV_Gamma-0_GB_OSGB_RCTK_SpkRL.tif, there are these 2 shapefiles:
rfi/RFI_Interference_Shapefiles_Chrissy/S1B_20180120_132_asc_174936_175001_DV_Gamma-0_GB_OSGB_RCTK_SpkRL.shp
rfi/RFI_Interference_Shapefiles_Richard/S1B_20180120_132_asc_174936_175001_DV_Gamma-0_GB_OSGB_RCTK_SpkRL.shp
So for the image S1A_20180308_30_asc_175835_175900_DV_Gamma-0_GB_OSGB_RCTK_SpkRL.tif, there are these 2 shapefiles:
rfi/RFI_Interference_Shapefiles_Zara/ S1A_20180308_30_asc_175835_175900_DV_Gamma-0_GB_OSGB_RCTK_SpkRL.shp
rfi/RFI_Interference_Shapefiles_Chrissy/S1A_20180308_30_asc_175835_175900_DV_Gamma-0_GB_OSGB_RCTK_SpkRL.shp
So for the image S1A_20180107_30_asc_175836_175901_DV_Gamma-0_GB_OSGB_RCTK_SpkRL.tif, there are these 2 shapefiles:
rfi/RFI_Interference_Shapefiles_Chrissy/S1A_20180107_30_asc_175836_175901_DV_Gamma-0_GB_OSGB_RCTK_SpkRL.shp
rfi/RFI_Interference_Shapefiles_Richard/S1A_20180107_30_asc_175836_175901_DV_Gamma-0_GB_OSGB_RCTK_SpkRL.shp
Solution - in each case the shapefiles need to be checked and where duplicate RFI regions captured, a decision be made on which is best and the features from each of the 2 shapefiles then merged into a single shapefile.
**Issue 2:** Some of the digitised RFI polygons shapefiles contain no records at all or all the records contain invalid geometries.
For these images, either the associated shapefile contained no records at all or the records it did contain all had invalid geometries. Processing was skipped.
So for image S1B_20180309_132_asc_174935_175000_DV_Gamma-0_GB_OSGB_RCTK_SpkRL.tif the shapefile rfi/RFI_Interference_Shapefiles_Zara/S1B_20180309_132_asc_174935_175000_DV_Gamma-0_GB_OSGB_RCTK_SpkRL.shp seems to have no records
So for image S1B_20180618_30_asc_175812_175837_DV_Gamma-0_GB_OSGB_RCTK_SpkRL.tif the shapefile rfi/RFI_Interference_Shapefiles_Chrissy/S1B_20180618_30_asc_175812_175837_DV_Gamma-0_GB_OSGB_RCTK_SpkRL.shp seems to have no records with valid geometries
Solution - can the shapefiles / digitised polygons within them be checked?
**Issue3:** Some of the digitised RFI polygons shapefiles contain features with invalid geometries.
These seem to mostly be self-intersections like this:

In such cases, the pixels associated with polygons with valid geometries were masked out as normal and a new image created. The pixels associated with polygons with invalid geometries were skipped and the pixels retained in the new image.
This can be seen in this screenshot:

The pixels inside the valid polygons have been masked out. The pixels inside the invalid polygon have been retained. The following shapefiles were detected as containing features with invalid geometries:
Found invalid polygon with id 8 in rfi/RFI_Interference_Shapefiles_Chrissy/S1B_20180113_30_asc_175813_175838_DV_Gamma-0_GB_OSGB_RCTK_SpkRL.shp
Found invalid polygon with id 1 in rfi/RFI_Interference_Shapefiles_Zara/S1A_20180303_132_asc_175017_175042_DV_Gamma-0_GB_OSGB_RCTK_SpkRL.shp
Found invalid polygon with id 10 in rfi/RFI_Interference_Shapefiles_Chrissy/S1A_20180119_30_asc_175836_175901_DV_Gamma-0_GB_OSGB_RCTK_SpkRL.shp
Solution - can the digitised polygons be checked in these shapefiles. Once done, the image can be re-processed to mask out all the RFI regions.
|
process
|
check edit digitised rfi polygon shapefiles the script which does the masking out of manually digitised rfi locations provided as shapefiles does some validation before doing the masking in rasterio and detected some problems with some of the shapefiles some of the records in some of the shapefiles there were shapefiles provided images have been processed to set to nodata all pixels that fell within the polygons held in each shapefile why the difference of this seems to be due to the following issues with the digitised rfi polygon shapefiles which need to be looked at confirmed and fixed issue there are multiple digitised rfi polygons shapefiles associated with a single image tiff files seemed to each have shapefiles associated with them in some cases duplicate rfi features have been captured one in each shapefile as shown here the script assumes shapefile per image so processing was skipped so for the image asc dv gamma gb osgb rctk spkrl tif there are these shapefiles rfi rfi interference shapefiles chrissy asc dv gamma gb osgb rctk spkrl shp rfi rfi interference shapefiles richard asc dv gamma gb osgb rctk spkrl shp so for the image asc dv gamma gb osgb rctk spkrl tif there are these shapefiles rfi rfi interference shapefiles zara asc dv gamma gb osgb rctk spkrl shp rfi rfi interference shapefiles chrissy asc dv gamma gb osgb rctk spkrl shp so for the image asc dv gamma gb osgb rctk spkrl tif there are these shapefiles rfi rfi interference shapefiles chrissy asc dv gamma gb osgb rctk spkrl shp rfi rfi interference shapefiles richard asc dv gamma gb osgb rctk spkrl shp solution in each case the shapefiles need to be checked and where duplicate rfi regions captured a decision be made on which is best and the features from each of the shapefiles then merged into a single shapefile issue some of the digitised rfi polygons shapefiles contain no records at all or all the records contain invalid geometries for these images either the associated shapefile contained no records at all or the records it did contain all had invalid geometries processing was skipped so for image asc dv gamma gb osgb rctk spkrl tif the shapefile rfi rfi interference shapefiles zara asc dv gamma gb osgb rctk spkrl shp seems to have no records so for image asc dv gamma gb osgb rctk spkrl tif the shapefile rfi rfi interference shapefiles chrissy asc dv gamma gb osgb rctk spkrl shp seems to have no records with valid geometries solution can the shapefiles digitised polygons within them be checked some of the digitised rfi polygons shapefiles contain features with invalid geometries these seem to mostly be self intersections like this in such cases the pixels associated with polygons with valid geometries were masked out as normal and a new image created the pixels associated with polygons with invalid geometries were skipped and the pixels retained in the new image this can be seen in this screenshot the pixels inside the valid polygons have been masked out the pixels inside the invalid polygon have been retained the following shapefiles were detected as containing features with invalid geometries found invalid polygon with id in rfi rfi interference shapefiles chrissy asc dv gamma gb osgb rctk spkrl shp found invalid polygon with id in rfi rfi interference shapefiles zara asc dv gamma gb osgb rctk spkrl shp found invalid polygon with id in rfi rfi interference shapefiles chrissy asc dv gamma gb osgb rctk spkrl shp solution can the digitised polygons be checked in these shapefiles once done the image can be re processed to mask out all the rfi regions
| 1
|
456,190
| 13,146,666,951
|
IssuesEvent
|
2020-08-08 11:16:27
|
kubernetes/website
|
https://api.github.com/repos/kubernetes/website
|
closed
|
kubectl generated docs show path of user running generation tools
|
lifecycle/rotten priority/backlog sig/release
|
**This is a Bug Report**
<!--Required Information-->
**Problem:**
Spotted as part of review for #18010
@daminisatya generated docs which ended up containing:
```html
<tr>
<td colspan="2">--cache-dir string Default: "/Users/dsatya/.kube/http-cache"</td>
</tr>
```
whereas the docs should really show something like:
```html
<tr>
<td colspan="2">--cache-dir string Default: ~/.kube/http-cache</td>
</tr>
```
**Proposed Solution:** This needs fixing either upstream in the generation code, or maybe in the [wrapper tooling](https://kubernetes.io/docs/contribute/generate-ref-docs/kubernetes-components/).
Alternatively, clearly document a manual cleanup step to happen during the release process.
**Page to Update:**
https://kubernetes.io/docs/reference/kubectl/kubectl/
**Kubernetes Version**: v1.17 (barely!)
**Additional Information**:
Closed issue #10081 may also be relevant
|
1.0
|
kubectl generated docs show path of user running generation tools - **This is a Bug Report**
<!--Required Information-->
**Problem:**
Spotted as part of review for #18010
@daminisatya generated docs which ended up containing:
```html
<tr>
<td colspan="2">--cache-dir string Default: "/Users/dsatya/.kube/http-cache"</td>
</tr>
```
whereas the docs should really show something like:
```html
<tr>
<td colspan="2">--cache-dir string Default: ~/.kube/http-cache</td>
</tr>
```
**Proposed Solution:** This needs fixing either upstream in the generation code, or maybe in the [wrapper tooling](https://kubernetes.io/docs/contribute/generate-ref-docs/kubernetes-components/).
Alternatively, clearly document a manual cleanup step to happen during the release process.
**Page to Update:**
https://kubernetes.io/docs/reference/kubectl/kubectl/
**Kubernetes Version**: v1.17 (barely!)
**Additional Information**:
Closed issue #10081 may also be relevant
|
non_process
|
kubectl generated docs show path of user running generation tools this is a bug report problem spotted as part of review for daminisatya generated docs which ended up containing html cache dir string nbsp nbsp nbsp nbsp nbsp default users dsatya kube http cache whereas the docs should really show something like html cache dir string nbsp nbsp nbsp nbsp nbsp default kube http cache proposed solution this needs fixing either upstream in the generation code or maybe in the alternatively clearly document a manual cleanup step to happen during the release process page to update kubernetes version barely additional information closed issue may also be relevant
| 0
|
326,783
| 9,961,045,081
|
IssuesEvent
|
2019-07-06 23:06:30
|
kubeflow/kubeflow
|
https://api.github.com/repos/kubeflow/kubeflow
|
closed
|
[kfctl] better platform isolation
|
area/kfctl lifecycle/stale priority/p2
|
Currently in bootstrap/pkg/utils we have multiple platforms included. need to make platform isolation better - move GCP only stuffs to pkg/kfapp/gcp/...
|
1.0
|
[kfctl] better platform isolation - Currently in bootstrap/pkg/utils we have multiple platforms included. need to make platform isolation better - move GCP only stuffs to pkg/kfapp/gcp/...
|
non_process
|
better platform isolation currently in bootstrap pkg utils we have multiple platforms included need to make platform isolation better move gcp only stuffs to pkg kfapp gcp
| 0
|
482,732
| 13,912,396,545
|
IssuesEvent
|
2020-10-20 18:48:37
|
cds-snc/report-a-cybercrime
|
https://api.github.com/repos/cds-snc/report-a-cybercrime
|
closed
|
Extra space after Personal information (in mobile version)
|
bug medium priority
|
## Summary
The extra space is needed after personal information field in "What do you think could be affected?" page when the report is viewed on phone

## Steps to reproduce
> How exactly can the bug be reproduced? Be very specific.
## Unresolved questions
> Are there any related issues you consider out of scope for this issue that could be addressed in the future?
|
1.0
|
Extra space after Personal information (in mobile version) - ## Summary
The extra space is needed after personal information field in "What do you think could be affected?" page when the report is viewed on phone

## Steps to reproduce
> How exactly can the bug be reproduced? Be very specific.
## Unresolved questions
> Are there any related issues you consider out of scope for this issue that could be addressed in the future?
|
non_process
|
extra space after personal information in mobile version summary the extra space is needed after personal information field in what do you think could be affected page when the report is viewed on phone steps to reproduce how exactly can the bug be reproduced be very specific unresolved questions are there any related issues you consider out of scope for this issue that could be addressed in the future
| 0
|
825,512
| 31,393,174,645
|
IssuesEvent
|
2023-08-26 15:49:13
|
tcet-opensource/erp-backend
|
https://api.github.com/repos/tcet-opensource/erp-backend
|
closed
|
[Feat]: [Create a model for exam]
|
enhancement good first issue Issue Size: 1 Priority: Medium Models
|
## Description
Create an exam model
This model will be named exam.js
create a schema with key and value pairs with appropriate conditions.
Models should be added to the model folder.
## Proposed Solution
```
date: date
startTime: time
duration: int
supervisor: Faculty
infrastructure: Infrastructure
course: Course
```
|
1.0
|
[Feat]: [Create a model for exam] - ## Description
Create an exam model
This model will be named exam.js
create a schema with key and value pairs with appropriate conditions.
Models should be added to the model folder.
## Proposed Solution
```
date: date
startTime: time
duration: int
supervisor: Faculty
infrastructure: Infrastructure
course: Course
```
|
non_process
|
description create an exam model this model will be named exam js create a schema with key and value pairs with appropriate conditions models should be added to the model folder proposed solution date date starttime time duration int supervisor faculty infrastructure infrastructure course course
| 0
|
5,486
| 8,359,255,328
|
IssuesEvent
|
2018-10-03 07:34:43
|
bitshares/bitshares-community-ui
|
https://api.github.com/repos/bitshares/bitshares-community-ui
|
closed
|
Authentication question
|
Login Signup process question
|
Should we make an app blocked from unauthorized user? So basically the user is redirected to login page if he is not authorized. Or the user always has access to main dashboard, and authorization provides additional features to him
|
1.0
|
Authentication question - Should we make an app blocked from unauthorized user? So basically the user is redirected to login page if he is not authorized. Or the user always has access to main dashboard, and authorization provides additional features to him
|
process
|
authentication question should we make an app blocked from unauthorized user so basically the user is redirected to login page if he is not authorized or the user always has access to main dashboard and authorization provides additional features to him
| 1
|
18,499
| 10,130,806,272
|
IssuesEvent
|
2019-08-01 17:53:49
|
embark-framework/embark
|
https://api.github.com/repos/embark-framework/embark
|
closed
|
Provide a way to optionally reduce filesize of contract objects by removing code attributes.
|
enhancement type: performance
|
### Feature Request
The contract JSON files have three attributes that represents a big % of the total file size: `code`, `runtimeBytecode` and `realRuntimeBytecode`, I see each of them have their use according to this link: https://www.reddit.com/r/ethereum/comments/3pq08g/some_quick_things_about_verifying_contracts/cw8qn0d/ but come with the drawback of drastically increasing the filesize.
Example: In a contract heavy project, like teller network, specifically for the Escrow contract, the json file containing these attributes has a file size of 157.3 kb, and after manually removing the attributes, the size is reduced to 31.2 kb. (80% reduction before minification)
A reduced file size could be beneficial in case a dApp is accessed via IPFS or using a mobile connection, because it means faster loading times and reduced cost, which leads me to believe that we could improve this situation by adding an optional post-processing to the contract files, in case the developer determines that they're not going to deploy new instances of a contract during the normal dApp use.
Maybe this could be done via a plugin, or by adding a new attribute to the contract configuration?
|
True
|
Provide a way to optionally reduce filesize of contract objects by removing code attributes. - ### Feature Request
The contract JSON files have three attributes that represents a big % of the total file size: `code`, `runtimeBytecode` and `realRuntimeBytecode`, I see each of them have their use according to this link: https://www.reddit.com/r/ethereum/comments/3pq08g/some_quick_things_about_verifying_contracts/cw8qn0d/ but come with the drawback of drastically increasing the filesize.
Example: In a contract heavy project, like teller network, specifically for the Escrow contract, the json file containing these attributes has a file size of 157.3 kb, and after manually removing the attributes, the size is reduced to 31.2 kb. (80% reduction before minification)
A reduced file size could be beneficial in case a dApp is accessed via IPFS or using a mobile connection, because it means faster loading times and reduced cost, which leads me to believe that we could improve this situation by adding an optional post-processing to the contract files, in case the developer determines that they're not going to deploy new instances of a contract during the normal dApp use.
Maybe this could be done via a plugin, or by adding a new attribute to the contract configuration?
|
non_process
|
provide a way to optionally reduce filesize of contract objects by removing code attributes feature request the contract json files have three attributes that represents a big of the total file size code runtimebytecode and realruntimebytecode i see each of them have their use according to this link but come with the drawback of drastically increasing the filesize example in a contract heavy project like teller network specifically for the escrow contract the json file containing these attributes has a file size of kb and after manually removing the attributes the size is reduced to kb reduction before minification a reduced file size could be beneficial in case a dapp is accessed via ipfs or using a mobile connection because it means faster loading times and reduced cost which leads me to believe that we could improve this situation by adding an optional post processing to the contract files in case the developer determines that they re not going to deploy new instances of a contract during the normal dapp use maybe this could be done via a plugin or by adding a new attribute to the contract configuration
| 0
|
316,845
| 9,658,070,500
|
IssuesEvent
|
2019-05-20 10:03:52
|
bbaumgartl/test
|
https://api.github.com/repos/bbaumgartl/test
|
closed
|
Support for 'module_sys_dmail_category'
|
Priority: Must have Status: Closed Tracker: Bug
|
---
Author Name: **Dan .** (Dan .)
Original Redmine Issue: 10145, https://forge.typo3.org/issues/10145
Original Date: 2010-10-07
Original Assignee: Bernhard Baumgartl, datamints GmbH
---
Hi
The select box of the column 'module_sys_dmail_category' produces an empty select box. My guess is that it is because the function showSelect in line 1683 looks for the number of items which are not given (since they are not fixed).
In addition, this should be rendered as check boxes ('renderMode' => 'checkbox',).
Thanks.
|
1.0
|
Support for 'module_sys_dmail_category' - ---
Author Name: **Dan .** (Dan .)
Original Redmine Issue: 10145, https://forge.typo3.org/issues/10145
Original Date: 2010-10-07
Original Assignee: Bernhard Baumgartl, datamints GmbH
---
Hi
The select box of the column 'module_sys_dmail_category' produces an empty select box. My guess is that it is because the function showSelect in line 1683 looks for the number of items which are not given (since they are not fixed).
In addition, this should be rendered as check boxes ('renderMode' => 'checkbox',).
Thanks.
|
non_process
|
support for module sys dmail category author name dan dan original redmine issue original date original assignee bernhard baumgartl datamints gmbh hi the select box of the column module sys dmail category produces an empty select box my guess is that it is because the function showselect in line looks for the number of items which are not given since they are not fixed in addition this should be rendered as check boxes rendermode checkbox thanks
| 0
|
3,812
| 6,796,374,582
|
IssuesEvent
|
2017-11-01 18:47:33
|
econtoolkit/continuous_time_methods
|
https://api.github.com/repos/econtoolkit/continuous_time_methods
|
opened
|
Basic code for discretizing non-uniform grid
|
matlab Stochastic Processes
|
@stevenzhangdx I am starting on this, but have added you as an observer. May hand over to you for testing/etc.
|
1.0
|
Basic code for discretizing non-uniform grid - @stevenzhangdx I am starting on this, but have added you as an observer. May hand over to you for testing/etc.
|
process
|
basic code for discretizing non uniform grid stevenzhangdx i am starting on this but have added you as an observer may hand over to you for testing etc
| 1
|
277,674
| 24,094,138,673
|
IssuesEvent
|
2022-09-19 17:09:43
|
chamilo/chamilo-lms
|
https://api.github.com/repos/chamilo/chamilo-lms
|
closed
|
Rename extra_field field_type and extra_field_type to item_type and data_type
|
Requires testing/validation
|
The extra_field-related tables are unnecessarily confusing because they have a "field_type" that identifies the type of value that should be stored in this extra field, and an "extra_field_type" that identifies the kind of item this extra_field relates to.
This should be clarified, renaming "field_type" to "data_type" (or "value_type") and renaming "extra_field_type" to "item_type".
Also, the term "value" in exta_field_values.value should be renamed to field_value because it is a reserved keyword in several DBMS: see https://www.petefreitag.com/tools/sql_reserved_words_checker/?word=value
|
1.0
|
Rename extra_field field_type and extra_field_type to item_type and data_type - The extra_field-related tables are unnecessarily confusing because they have a "field_type" that identifies the type of value that should be stored in this extra field, and an "extra_field_type" that identifies the kind of item this extra_field relates to.
This should be clarified, renaming "field_type" to "data_type" (or "value_type") and renaming "extra_field_type" to "item_type".
Also, the term "value" in exta_field_values.value should be renamed to field_value because it is a reserved keyword in several DBMS: see https://www.petefreitag.com/tools/sql_reserved_words_checker/?word=value
|
non_process
|
rename extra field field type and extra field type to item type and data type the extra field related tables are unnecessarily confusing because they have a field type that identifies the type of value that should be stored in this extra field and an extra field type that identifies the kind of item this extra field relates to this should be clarified renaming field type to data type or value type and renaming extra field type to item type also the term value in exta field values value should be renamed to field value because it is a reserved keyword in several dbms see
| 0
|
18,633
| 24,580,405,878
|
IssuesEvent
|
2022-10-13 15:12:48
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[FHIR ] [iOS] questionnaire response > JSON > Text choice response type with 'Other' option > Answer options are not getting saved in the FHIR store in the following scenario
|
Bug P0 iOS Process: Fixed Process: Tested QA Process: Tested dev
|
Steps:
1. SB > Create/edit study > Add questionnaire with Text choice response type along with 'Other' option
2. Launch or publish the study
3. Go to the mobile app
4. Enroll to the study
5. Click on the activity
6. Submit the response by selecting the 'Other' answer option
7. Go to the FHIR store and go to the Questionnaire response JSON file
8. Click on the response and observe
AR: Values are not getting mapped into the JSON file of the FHIR store
ER: Values should be mapped

|
3.0
|
[FHIR ] [iOS] questionnaire response > JSON > Text choice response type with 'Other' option > Answer options are not getting saved in the FHIR store in the following scenario - Steps:
1. SB > Create/edit study > Add questionnaire with Text choice response type along with 'Other' option
2. Launch or publish the study
3. Go to the mobile app
4. Enroll to the study
5. Click on the activity
6. Submit the response by selecting the 'Other' answer option
7. Go to the FHIR store and go to the Questionnaire response JSON file
8. Click on the response and observe
AR: Values are not getting mapped into the JSON file of the FHIR store
ER: Values should be mapped

|
process
|
questionnaire response json text choice response type with other option answer options are not getting saved in the fhir store in the following scenario steps sb create edit study add questionnaire with text choice response type along with other option launch or publish the study go to the mobile app enroll to the study click on the activity submit the response by selecting the other answer option go to the fhir store and go to the questionnaire response json file click on the response and observe ar values are not getting mapped into the json file of the fhir store er values should be mapped
| 1
|
18,354
| 24,481,016,210
|
IssuesEvent
|
2022-10-08 20:50:14
|
OpenDataScotland/the_od_bods
|
https://api.github.com/repos/OpenDataScotland/the_od_bods
|
closed
|
Fix Scottish Forestry Organisation
|
bug good first issue data processing back end
|
**Describe the bug**
A 2-part issue:
- Scottish Forestry is missing an entry in [jkan/_organizations ](https://github.com/OpenDataScotland/jkan/tree/gh-pages/_organizations)(twitter @scotforestry, website https://forestry.gov.scot/)
- Publisher name for Scottish Forestry is also not coming though correctly.
**To Reproduce**
Steps to reproduce the behavior:
1. Search for "Forestry" in page Organizations
2. Search for "Forestry" in page Datasets
**Expected behavior**
When searching for "Forestry" in Organizations, "Scottish Forestry" should appear.
When searching for "Forestry" in Datasets, the publisher name should be `Scottish Forestry` instead of `Scottish.Forestry`.
**Screenshots**

**Hardware and software used**
- Browser: Firefox 105.0.1 (64-bit)
**Additional context**
Not investigated where the `Scottish.Forestry` labelling is happening.
|
1.0
|
Fix Scottish Forestry Organisation - **Describe the bug**
A 2-part issue:
- Scottish Forestry is missing an entry in [jkan/_organizations ](https://github.com/OpenDataScotland/jkan/tree/gh-pages/_organizations)(twitter @scotforestry, website https://forestry.gov.scot/)
- Publisher name for Scottish Forestry is also not coming though correctly.
**To Reproduce**
Steps to reproduce the behavior:
1. Search for "Forestry" in page Organizations
2. Search for "Forestry" in page Datasets
**Expected behavior**
When searching for "Forestry" in Organizations, "Scottish Forestry" should appear.
When searching for "Forestry" in Datasets, the publisher name should be `Scottish Forestry` instead of `Scottish.Forestry`.
**Screenshots**

**Hardware and software used**
- Browser: Firefox 105.0.1 (64-bit)
**Additional context**
Not investigated where the `Scottish.Forestry` labelling is happening.
|
process
|
fix scottish forestry organisation describe the bug a part issue scottish forestry is missing an entry in scotforestry website publisher name for scottish forestry is also not coming though correctly to reproduce steps to reproduce the behavior search for forestry in page organizations search for forestry in page datasets expected behavior when searching for forestry in organizations scottish forestry should appear when searching for forestry in datasets the publisher name should be scottish forestry instead of scottish forestry screenshots hardware and software used browser firefox bit additional context not investigated where the scottish forestry labelling is happening
| 1
|
301
| 2,735,124,761
|
IssuesEvent
|
2015-04-18 03:24:33
|
sysown/proxysql-0.2
|
https://api.github.com/repos/sysown/proxysql-0.2
|
opened
|
Default schema per user
|
ADMIN AUTHENTICATION CONNECTION POOL development MYSQL PROTOCOL QUERY PROCESSOR ROUTING
|
Add two more columns to mysql_users table to specify:
a) default schema for user;
b) if the user has the option to issue any INIT_DB command
|
1.0
|
Default schema per user - Add two more columns to mysql_users table to specify:
a) default schema for user;
b) if the user has the option to issue any INIT_DB command
|
process
|
default schema per user add two more columns to mysql users table to specify a default schema for user b if the user has the option to issue any init db command
| 1
|
166,042
| 20,711,384,277
|
IssuesEvent
|
2022-03-12 01:15:14
|
snowflakedb/snowflake-jdbc
|
https://api.github.com/repos/snowflakedb/snowflake-jdbc
|
closed
|
SNOW-558888: CVE-2020-14061 (High) detected in jackson-databind-2.9.8.jar - autoclosed
|
security vulnerability
|
## CVE-2020-14061 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.8.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /tmp/ws-ua_20220312003410_KQQKCS/archiveExtraction_BELGGC/FUIDAN/20220312003410/snowflake-jdbc_depth_0/dependencies/arrow-vector-0.15.1/META-INF/maven/org.apache.arrow/arrow-vector/pom.xml</p>
<p>Path to vulnerable library: /sitory/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.8.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/snowflakedb/snowflake-jdbc/commit/8560bcca9d395d1ee02123536c2e958f7d386fe0">8560bcca9d395d1ee02123536c2e958f7d386fe0</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.5 mishandles the interaction between serialization gadgets and typing, related to oracle.jms.AQjmsQueueConnectionFactory, oracle.jms.AQjmsXATopicConnectionFactory, oracle.jms.AQjmsTopicConnectionFactory, oracle.jms.AQjmsXAQueueConnectionFactory, and oracle.jms.AQjmsXAConnectionFactory (aka weblogic/oracle-aqjms).
<p>Publish Date: 2020-06-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-14061>CVE-2020-14061</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-14061">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-14061</a></p>
<p>Release Date: 2020-06-14</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.10.0</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.8","packageFilePaths":["/tmp/ws-ua_20220312003410_KQQKCS/archiveExtraction_BELGGC/FUIDAN/20220312003410/snowflake-jdbc_depth_0/dependencies/arrow-vector-0.15.1/META-INF/maven/org.apache.arrow/arrow-vector/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.10.0","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-14061","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.5 mishandles the interaction between serialization gadgets and typing, related to oracle.jms.AQjmsQueueConnectionFactory, oracle.jms.AQjmsXATopicConnectionFactory, oracle.jms.AQjmsTopicConnectionFactory, oracle.jms.AQjmsXAQueueConnectionFactory, and oracle.jms.AQjmsXAConnectionFactory (aka weblogic/oracle-aqjms).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-14061","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
True
|
SNOW-558888: CVE-2020-14061 (High) detected in jackson-databind-2.9.8.jar - autoclosed - ## CVE-2020-14061 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.8.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /tmp/ws-ua_20220312003410_KQQKCS/archiveExtraction_BELGGC/FUIDAN/20220312003410/snowflake-jdbc_depth_0/dependencies/arrow-vector-0.15.1/META-INF/maven/org.apache.arrow/arrow-vector/pom.xml</p>
<p>Path to vulnerable library: /sitory/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.8.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/snowflakedb/snowflake-jdbc/commit/8560bcca9d395d1ee02123536c2e958f7d386fe0">8560bcca9d395d1ee02123536c2e958f7d386fe0</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.5 mishandles the interaction between serialization gadgets and typing, related to oracle.jms.AQjmsQueueConnectionFactory, oracle.jms.AQjmsXATopicConnectionFactory, oracle.jms.AQjmsTopicConnectionFactory, oracle.jms.AQjmsXAQueueConnectionFactory, and oracle.jms.AQjmsXAConnectionFactory (aka weblogic/oracle-aqjms).
<p>Publish Date: 2020-06-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-14061>CVE-2020-14061</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-14061">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-14061</a></p>
<p>Release Date: 2020-06-14</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.10.0</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.8","packageFilePaths":["/tmp/ws-ua_20220312003410_KQQKCS/archiveExtraction_BELGGC/FUIDAN/20220312003410/snowflake-jdbc_depth_0/dependencies/arrow-vector-0.15.1/META-INF/maven/org.apache.arrow/arrow-vector/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.10.0","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-14061","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.5 mishandles the interaction between serialization gadgets and typing, related to oracle.jms.AQjmsQueueConnectionFactory, oracle.jms.AQjmsXATopicConnectionFactory, oracle.jms.AQjmsTopicConnectionFactory, oracle.jms.AQjmsXAQueueConnectionFactory, and oracle.jms.AQjmsXAConnectionFactory (aka weblogic/oracle-aqjms).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-14061","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
non_process
|
snow cve high detected in jackson databind jar autoclosed cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file tmp ws ua kqqkcs archiveextraction belggc fuidan snowflake jdbc depth dependencies arrow vector meta inf maven org apache arrow arrow vector pom xml path to vulnerable library sitory com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to oracle jms aqjmsqueueconnectionfactory oracle jms aqjmsxatopicconnectionfactory oracle jms aqjmstopicconnectionfactory oracle jms aqjmsxaqueueconnectionfactory and oracle jms aqjmsxaconnectionfactory aka weblogic oracle aqjms publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to oracle jms aqjmsqueueconnectionfactory oracle jms aqjmsxatopicconnectionfactory oracle jms aqjmstopicconnectionfactory oracle jms aqjmsxaqueueconnectionfactory and oracle jms aqjmsxaconnectionfactory aka weblogic oracle aqjms vulnerabilityurl
| 0
|
8,786
| 11,905,980,000
|
IssuesEvent
|
2020-03-30 19:32:38
|
googleapis/google-cloud-cpp
|
https://api.github.com/repos/googleapis/google-cloud-cpp
|
closed
|
Announce that -bigquery will be moving in N days
|
api: bigquery type: process
|
For BigQuery, I think N = 1? Maybe as high as 7, but I don't think anyone uses it yet.
We want to give customers a heads up that the -bigquery repo will be moving into the monorepo at -cpp. We should cut a release and and include an announcement that this will be happening. We may also want to drop a note in our slack channel and anywhere else we can, like on the README.md.
|
1.0
|
Announce that -bigquery will be moving in N days - For BigQuery, I think N = 1? Maybe as high as 7, but I don't think anyone uses it yet.
We want to give customers a heads up that the -bigquery repo will be moving into the monorepo at -cpp. We should cut a release and and include an announcement that this will be happening. We may also want to drop a note in our slack channel and anywhere else we can, like on the README.md.
|
process
|
announce that bigquery will be moving in n days for bigquery i think n maybe as high as but i don t think anyone uses it yet we want to give customers a heads up that the bigquery repo will be moving into the monorepo at cpp we should cut a release and and include an announcement that this will be happening we may also want to drop a note in our slack channel and anywhere else we can like on the readme md
| 1
|
568,126
| 16,959,962,856
|
IssuesEvent
|
2021-06-29 01:22:18
|
knative/eventing
|
https://api.github.com/repos/knative/eventing
|
closed
|
Knative Tenant Targeted Logging
|
area/observability kind/feature-request kind/good-first-issue lifecycle/stale priority/important-soon
|
**Problem**
In a mult-tenant environment provide logging from kNative services that are specific to a particular tenant.
**[Persona:](https://github.com/knative/eventing/blob/master/docs/personas.md)**
The feature is for end users (tenant) needing logging that is necessary to diagnose issues with their Knative application. In such an mulit-tenant it is not feasible to give the tenant full Knative logs that includes information from other tenants, nor would it helpful to most users not knowing Knative internal workings.
[Knative Tenant Targeted Logging](https://docs.google.com/document/d/1FWekMACBK9d0Sc-dZJuSxRpFxHcH4tQI8pjIPVIX_Zg/edit#heading=h.9fc0yrfwqbi8) provides more details for this feature.
|
1.0
|
Knative Tenant Targeted Logging - **Problem**
In a mult-tenant environment provide logging from kNative services that are specific to a particular tenant.
**[Persona:](https://github.com/knative/eventing/blob/master/docs/personas.md)**
The feature is for end users (tenant) needing logging that is necessary to diagnose issues with their Knative application. In such an mulit-tenant it is not feasible to give the tenant full Knative logs that includes information from other tenants, nor would it helpful to most users not knowing Knative internal workings.
[Knative Tenant Targeted Logging](https://docs.google.com/document/d/1FWekMACBK9d0Sc-dZJuSxRpFxHcH4tQI8pjIPVIX_Zg/edit#heading=h.9fc0yrfwqbi8) provides more details for this feature.
|
non_process
|
knative tenant targeted logging problem in a mult tenant environment provide logging from knative services that are specific to a particular tenant the feature is for end users tenant needing logging that is necessary to diagnose issues with their knative application in such an mulit tenant it is not feasible to give the tenant full knative logs that includes information from other tenants nor would it helpful to most users not knowing knative internal workings provides more details for this feature
| 0
|
88,585
| 3,779,422,128
|
IssuesEvent
|
2016-03-18 08:12:33
|
sci-visus/visus-issues
|
https://api.github.com/repos/sci-visus/visus-issues
|
closed
|
importing visuspy without using it causes segfault on exit
|
Bug Priority Low
|
This isn't a critical issue. I'm reporting it because I spent a bunch of time chasing it and it's cause wasn't obvious.
```python
from visuspy import *
if __name__ == '__main__':
segfault=True
if not segfault:
app=Application()
app.setCommandLine("")
app.useModule(IdxModule.getSingleton())
else:
print "doing nothing"
```
I tried invoking pdb.set_trace at the end, which produces this suspicious exception for both setfault=True and segfault=False cases:
```
Exception AttributeError: "'NoneType' object has no attribute 'path'" in <function _remove at 0x10d8f6668> ignored
```
|
1.0
|
importing visuspy without using it causes segfault on exit - This isn't a critical issue. I'm reporting it because I spent a bunch of time chasing it and it's cause wasn't obvious.
```python
from visuspy import *
if __name__ == '__main__':
segfault=True
if not segfault:
app=Application()
app.setCommandLine("")
app.useModule(IdxModule.getSingleton())
else:
print "doing nothing"
```
I tried invoking pdb.set_trace at the end, which produces this suspicious exception for both setfault=True and segfault=False cases:
```
Exception AttributeError: "'NoneType' object has no attribute 'path'" in <function _remove at 0x10d8f6668> ignored
```
|
non_process
|
importing visuspy without using it causes segfault on exit this isn t a critical issue i m reporting it because i spent a bunch of time chasing it and it s cause wasn t obvious python from visuspy import if name main segfault true if not segfault app application app setcommandline app usemodule idxmodule getsingleton else print doing nothing i tried invoking pdb set trace at the end which produces this suspicious exception for both setfault true and segfault false cases exception attributeerror nonetype object has no attribute path in ignored
| 0
|
11,265
| 14,058,449,116
|
IssuesEvent
|
2020-11-03 00:14:52
|
panther-labs/panther
|
https://api.github.com/repos/panther-labs/panther
|
closed
|
Rule Testing: Show Dedup and Title Output
|
enhancement epic p0 team:data processing
|
### Describe the ideal solution
Today, rule tests only give an indication of if the rule returns true/false. The CLI workflow has slightly more functionality beyond that to validate the title() and dedup() workflows:
```
AWS.S3.ServerAccess.Unauthenticated
[PASS] Authenticated Access
[Title] Unauthenticated access to S3 bucket example-bucket
[Dedup] example-bucket
[PASS] Unauthenticated Access
[Title] Unauthenticated access to S3 bucket example-bucket
[Dedup] example-bucket
```
The UI workflow should also test these same functions.
### Describe your use cases
Validation of if the Title and Dedup functions work in addition to the Rule function.
### How frequently would you use such a feature
Daily
### Describe alternatives you have considered
CLI
## -
### Additional context

|
1.0
|
Rule Testing: Show Dedup and Title Output - ### Describe the ideal solution
Today, rule tests only give an indication of if the rule returns true/false. The CLI workflow has slightly more functionality beyond that to validate the title() and dedup() workflows:
```
AWS.S3.ServerAccess.Unauthenticated
[PASS] Authenticated Access
[Title] Unauthenticated access to S3 bucket example-bucket
[Dedup] example-bucket
[PASS] Unauthenticated Access
[Title] Unauthenticated access to S3 bucket example-bucket
[Dedup] example-bucket
```
The UI workflow should also test these same functions.
### Describe your use cases
Validation of if the Title and Dedup functions work in addition to the Rule function.
### How frequently would you use such a feature
Daily
### Describe alternatives you have considered
CLI
## -
### Additional context

|
process
|
rule testing show dedup and title output describe the ideal solution today rule tests only give an indication of if the rule returns true false the cli workflow has slightly more functionality beyond that to validate the title and dedup workflows aws serveraccess unauthenticated authenticated access unauthenticated access to bucket example bucket example bucket unauthenticated access unauthenticated access to bucket example bucket example bucket the ui workflow should also test these same functions describe your use cases validation of if the title and dedup functions work in addition to the rule function how frequently would you use such a feature daily describe alternatives you have considered cli additional context
| 1
|
5,533
| 8,391,255,490
|
IssuesEvent
|
2018-10-09 14:36:28
|
kiwicom/orbit-components
|
https://api.github.com/repos/kiwicom/orbit-components
|
closed
|
InputFile error and help text breaks out
|
bug processing
|
<!--- Provide a general summary of the issue in the Title above -->
Input and file text breaks out of the component. It is also broken on all the storybooks
## Expected Behavior
https://monosnap.com/file/yYHPZmLB7T0ybosmEvGadTzo1AKhDf
## Current Behavior
https://monosnap.com/file/rNQtCxDwDDQWk6gNFaWGIUevTKjljd
## Possible Solution
Remove `position: absolute` from the `FormFeedback` component.
I've currently hacked it by placing it in a div with `position: relative`, but would be nice if this wasn't needed
## Context (Environment)
Orbit 0.13.0
|
1.0
|
InputFile error and help text breaks out - <!--- Provide a general summary of the issue in the Title above -->
Input and file text breaks out of the component. It is also broken on all the storybooks
## Expected Behavior
https://monosnap.com/file/yYHPZmLB7T0ybosmEvGadTzo1AKhDf
## Current Behavior
https://monosnap.com/file/rNQtCxDwDDQWk6gNFaWGIUevTKjljd
## Possible Solution
Remove `position: absolute` from the `FormFeedback` component.
I've currently hacked it by placing it in a div with `position: relative`, but would be nice if this wasn't needed
## Context (Environment)
Orbit 0.13.0
|
process
|
inputfile error and help text breaks out input and file text breaks out of the component it is also broken on all the storybooks expected behavior current behavior possible solution remove position absolute from the formfeedback component i ve currently hacked it by placing it in a div with position relative but would be nice if this wasn t needed context environment orbit
| 1
|
10,103
| 13,044,162,127
|
IssuesEvent
|
2020-07-29 03:47:29
|
tikv/tikv
|
https://api.github.com/repos/tikv/tikv
|
closed
|
UCP: Migrate scalar function `UTCTimestampWithoutArg` from TiDB
|
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
|
## Description
Port the scalar function `UTCTimestampWithoutArg` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @iosmanthus
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
2.0
|
UCP: Migrate scalar function `UTCTimestampWithoutArg` from TiDB -
## Description
Port the scalar function `UTCTimestampWithoutArg` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @iosmanthus
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
process
|
ucp migrate scalar function utctimestampwithoutarg from tidb description port the scalar function utctimestampwithoutarg from tidb to coprocessor score mentor s iosmanthus recommended skills rust programming learning materials already implemented expressions ported from tidb
| 1
|
18,188
| 24,236,229,773
|
IssuesEvent
|
2022-09-26 23:42:45
|
keras-team/keras-cv
|
https://api.github.com/repos/keras-team/keras-cv
|
closed
|
Mosaic Augmentation for Object Detection
|
contribution-welcome preprocessing
|
Mosaic augmentation for object detection is used in Yolo-V4 literature, [FR](https://github.com/keras-team/keras-cv/issues/21). I'm not sure if it's used for classification tasks there. I was wondering if it's possible to do so for the image classification task. Here are two issues:
- **References. Not sure if it's ever used for classification only in any literature.**
- **Creation of class labels.**
For creating class labels, here is one possible solution, described [HERE](https://stackoverflow.com/a/65182855/9215780); using Dirichlet distribution.
```python
# for 2 images. Equivalent to λ and (1-λ)
>>> np.random.dirichlet((1, 1), 1)
array([[0.92870347, 0.07129653]])
>>> np.random.dirichlet((1, 1, 1), 1) # for 3 images.
array([[0.38712673, 0.46132787, 0.1515454 ]])
>>> np.random.dirichlet((1, 1, 1, 1), 1) # for 4 images.
array([[0.59482542, 0.0185333 , 0.33322484, 0.05341645]])
```
As mosaic takes 4 images.

---
update:
Mosaic augmentation for classification did use in literature. Please check: https://github.com/keras-team/keras-cv/issues/250#issuecomment-1130256698
|
1.0
|
Mosaic Augmentation for Object Detection - Mosaic augmentation for object detection is used in Yolo-V4 literature, [FR](https://github.com/keras-team/keras-cv/issues/21). I'm not sure if it's used for classification tasks there. I was wondering if it's possible to do so for the image classification task. Here are two issues:
- **References. Not sure if it's ever used for classification only in any literature.**
- **Creation of class labels.**
For creating class labels, here is one possible solution, described [HERE](https://stackoverflow.com/a/65182855/9215780); using Dirichlet distribution.
```python
# for 2 images. Equivalent to λ and (1-λ)
>>> np.random.dirichlet((1, 1), 1)
array([[0.92870347, 0.07129653]])
>>> np.random.dirichlet((1, 1, 1), 1) # for 3 images.
array([[0.38712673, 0.46132787, 0.1515454 ]])
>>> np.random.dirichlet((1, 1, 1, 1), 1) # for 4 images.
array([[0.59482542, 0.0185333 , 0.33322484, 0.05341645]])
```
As mosaic takes 4 images.

---
update:
Mosaic augmentation for classification did use in literature. Please check: https://github.com/keras-team/keras-cv/issues/250#issuecomment-1130256698
|
process
|
mosaic augmentation for object detection mosaic augmentation for object detection is used in yolo literature i m not sure if it s used for classification tasks there i was wondering if it s possible to do so for the image classification task here are two issues references not sure if it s ever used for classification only in any literature creation of class labels for creating class labels here is one possible solution described using dirichlet distribution python for images equivalent to λ and λ np random dirichlet array np random dirichlet for images array np random dirichlet for images array as mosaic takes images update mosaic augmentation for classification did use in literature please check
| 1
|
9,265
| 12,295,813,092
|
IssuesEvent
|
2020-05-11 05:18:51
|
labnote-ant/labnote
|
https://api.github.com/repos/labnote-ant/labnote
|
closed
|
How to show Chemicals in Process view
|
process-view
|
How do we show chemicals in Process view?
Chemicals contain time-related information (like their amounts) or can be some sort of compounds (like mixed, injected, or added ones).
Right now, we are just showing all the previously created chemicals. Like the below image example:

Above three red circled chemicals are technically the same chemicals, but later ones have gone through different processes (like heating and water bathing).
How can we create logic to show these chemicals?
|
1.0
|
How to show Chemicals in Process view - How do we show chemicals in Process view?
Chemicals contain time-related information (like their amounts) or can be some sort of compounds (like mixed, injected, or added ones).
Right now, we are just showing all the previously created chemicals. Like the below image example:

Above three red circled chemicals are technically the same chemicals, but later ones have gone through different processes (like heating and water bathing).
How can we create logic to show these chemicals?
|
process
|
how to show chemicals in process view how do we show chemicals in process view chemicals contain time related information like their amounts or can be some sort of compounds like mixed injected or added ones right now we are just showing all the previously created chemicals like the below image example above three red circled chemicals are technically the same chemicals but later ones have gone through different processes like heating and water bathing how can we create logic to show these chemicals
| 1
|
14,409
| 17,461,418,992
|
IssuesEvent
|
2021-08-06 11:02:08
|
prisma/prisma
|
https://api.github.com/repos/prisma/prisma
|
opened
|
Test CLI database creation for all supported databases
|
process/candidate topic: tests kind/tech team/migrations
|
Migrate commands like `db push` (and other Migrate commands) create a database if it does not exist. (See related bug: https://github.com/prisma/prisma/issues/8058) We should have explicit tests for all our supported databases, that make sure that this database creation really works.
|
1.0
|
Test CLI database creation for all supported databases - Migrate commands like `db push` (and other Migrate commands) create a database if it does not exist. (See related bug: https://github.com/prisma/prisma/issues/8058) We should have explicit tests for all our supported databases, that make sure that this database creation really works.
|
process
|
test cli database creation for all supported databases migrate commands like db push and other migrate commands create a database if it does not exist see related bug we should have explicit tests for all our supported databases that make sure that this database creation really works
| 1
|
18,735
| 10,212,418,892
|
IssuesEvent
|
2019-08-14 19:23:57
|
cmu-db/terrier
|
https://api.github.com/repos/cmu-db/terrier
|
opened
|
Reduce time spent on visibility check in indexes
|
performance question/discussion
|
### Background
@mbutrovich has identified that a large amount of time during the TPC-C benchmark is spent on visibility checks on tupleslots during index operations. The current method for the visibility check involves reading both two different sections of the block (allocation bitmap and the version pointer) and then unconditionally traversing the version chain as you would for materialization.
### Proposed Solution
Steal the low order bits of the version pointer to store visibility hints. Specifically, our undo records are 64-bit aligned which gives us three low order bits that do not currently carry information. We could use these bits to track allocation, "insert in version chain", and deleted. This would allow a visibility check to read atomically read the version pointer and current visibility state together, and in the common case (allocated, insert not in the version chain, and not deleted) return true.
### Details:
#### Flag definitions:
- **Unallocated**: Indicates whether the spot is unallocated (1) or allocated (0). This semantic makes the visibility check cleaner.
- **Insert in chain**: Indicates whether the existing version chain ends in an insert (1) or the insert is no longer visible (0) [definition means this flag is cleared when it is safe to unlink but before that has happened].
- **Deleted**: Indicates whether this tuple is logically deleted (1) or still exists (0). A deletion will set this when it first stages the delete.
#### SqlTable behavior
- On initial allocation of a block, this solution would require scanning down the version pointer column and initializing it to set the unallocated (incurs some cost on inserts, existing allocated bitmap would be removed).
- On an insert, the atomic swap of the undo record would include clearing the "unallocated" bit and setting the "insert in chain" bit.
- On unlink of an insert record, the GC would first clear the "insert in chain" bit through repeated CAS until successful and then proceed with the normal unlink process.
- On a delete, the atomic swap adding the delete record would include setting the "deleted" bit.
- On unlink of a deletion, the GC would ensure the deleted bit remains set and issue a deferred action to set the "unallocated" bit.
- Other operations would remain generally unmodified other than requiring the lowest three bits to be masked out before dereferencing the pointer.
#### Index Visibility Check
1. Read the version information: if low order three bits equals 0, return true. (Should generally be the common case)
2. Apply mask for the version pointer: if the pointer is nullptr, return false. (either unallocated or deleted)
3. Chase the version chain. This can potentially be optimized for early exiting, if needed.
|
True
|
Reduce time spent on visibility check in indexes - ### Background
@mbutrovich has identified that a large amount of time during the TPC-C benchmark is spent on visibility checks on tupleslots during index operations. The current method for the visibility check involves reading both two different sections of the block (allocation bitmap and the version pointer) and then unconditionally traversing the version chain as you would for materialization.
### Proposed Solution
Steal the low order bits of the version pointer to store visibility hints. Specifically, our undo records are 64-bit aligned which gives us three low order bits that do not currently carry information. We could use these bits to track allocation, "insert in version chain", and deleted. This would allow a visibility check to read atomically read the version pointer and current visibility state together, and in the common case (allocated, insert not in the version chain, and not deleted) return true.
### Details:
#### Flag definitions:
- **Unallocated**: Indicates whether the spot is unallocated (1) or allocated (0). This semantic makes the visibility check cleaner.
- **Insert in chain**: Indicates whether the existing version chain ends in an insert (1) or the insert is no longer visible (0) [definition means this flag is cleared when it is safe to unlink but before that has happened].
- **Deleted**: Indicates whether this tuple is logically deleted (1) or still exists (0). A deletion will set this when it first stages the delete.
#### SqlTable behavior
- On initial allocation of a block, this solution would require scanning down the version pointer column and initializing it to set the unallocated (incurs some cost on inserts, existing allocated bitmap would be removed).
- On an insert, the atomic swap of the undo record would include clearing the "unallocated" bit and setting the "insert in chain" bit.
- On unlink of an insert record, the GC would first clear the "insert in chain" bit through repeated CAS until successful and then proceed with the normal unlink process.
- On a delete, the atomic swap adding the delete record would include setting the "deleted" bit.
- On unlink of a deletion, the GC would ensure the deleted bit remains set and issue a deferred action to set the "unallocated" bit.
- Other operations would remain generally unmodified other than requiring the lowest three bits to be masked out before dereferencing the pointer.
#### Index Visibility Check
1. Read the version information: if low order three bits equals 0, return true. (Should generally be the common case)
2. Apply mask for the version pointer: if the pointer is nullptr, return false. (either unallocated or deleted)
3. Chase the version chain. This can potentially be optimized for early exiting, if needed.
|
non_process
|
reduce time spent on visibility check in indexes background mbutrovich has identified that a large amount of time during the tpc c benchmark is spent on visibility checks on tupleslots during index operations the current method for the visibility check involves reading both two different sections of the block allocation bitmap and the version pointer and then unconditionally traversing the version chain as you would for materialization proposed solution steal the low order bits of the version pointer to store visibility hints specifically our undo records are bit aligned which gives us three low order bits that do not currently carry information we could use these bits to track allocation insert in version chain and deleted this would allow a visibility check to read atomically read the version pointer and current visibility state together and in the common case allocated insert not in the version chain and not deleted return true details flag definitions unallocated indicates whether the spot is unallocated or allocated this semantic makes the visibility check cleaner insert in chain indicates whether the existing version chain ends in an insert or the insert is no longer visible deleted indicates whether this tuple is logically deleted or still exists a deletion will set this when it first stages the delete sqltable behavior on initial allocation of a block this solution would require scanning down the version pointer column and initializing it to set the unallocated incurs some cost on inserts existing allocated bitmap would be removed on an insert the atomic swap of the undo record would include clearing the unallocated bit and setting the insert in chain bit on unlink of an insert record the gc would first clear the insert in chain bit through repeated cas until successful and then proceed with the normal unlink process on a delete the atomic swap adding the delete record would include setting the deleted bit on unlink of a deletion the gc would ensure the deleted bit remains set and issue a deferred action to set the unallocated bit other operations would remain generally unmodified other than requiring the lowest three bits to be masked out before dereferencing the pointer index visibility check read the version information if low order three bits equals return true should generally be the common case apply mask for the version pointer if the pointer is nullptr return false either unallocated or deleted chase the version chain this can potentially be optimized for early exiting if needed
| 0
|
2,421
| 5,201,769,837
|
IssuesEvent
|
2017-01-24 06:43:52
|
jlm2017/jlm-video-subtitles
|
https://api.github.com/repos/jlm2017/jlm-video-subtitles
|
closed
|
[subtitles] [eng] #RDLS12 - JACQUELINE SAUVAGE, CHRISTINE LAGARDE, CAISSIÈRE AUCHAN, INSCRIPTION LISTES ÉLECTORALES
|
Language: English Process: [6] Approved
|
# Video title
#RDLS12 - JACQUELINE SAUVAGE, CHRISTINE LAGARDE, CAISSIÈRE AUCHAN, INSCRIPTION LISTES ÉLECTORALES
# URL
https://www.youtube.com/watch?v=OVp-swl3NuE
# Youtube subtitles language
Anglais
# Duration
23:10
# Subtitles URL
https://www.youtube.com/timedtext_editor?ui=hd&tab=captions&v=OVp-swl3NuE&lang=en&ref=player&bl=vmp&action_mde_edit_form=1
|
1.0
|
[subtitles] [eng] #RDLS12 - JACQUELINE SAUVAGE, CHRISTINE LAGARDE, CAISSIÈRE AUCHAN, INSCRIPTION LISTES ÉLECTORALES - # Video title
#RDLS12 - JACQUELINE SAUVAGE, CHRISTINE LAGARDE, CAISSIÈRE AUCHAN, INSCRIPTION LISTES ÉLECTORALES
# URL
https://www.youtube.com/watch?v=OVp-swl3NuE
# Youtube subtitles language
Anglais
# Duration
23:10
# Subtitles URL
https://www.youtube.com/timedtext_editor?ui=hd&tab=captions&v=OVp-swl3NuE&lang=en&ref=player&bl=vmp&action_mde_edit_form=1
|
process
|
jacqueline sauvage christine lagarde caissière auchan inscription listes électorales video title jacqueline sauvage christine lagarde caissière auchan inscription listes électorales url youtube subtitles language anglais duration subtitles url
| 1
|
26,400
| 6,767,171,249
|
IssuesEvent
|
2017-10-26 01:34:32
|
ahmedahamid/temp-third
|
https://api.github.com/repos/ahmedahamid/temp-third
|
closed
|
Create Example: CppAutomatePowerPoint
|
CodePlexMigrationInitiated Impact: Medium Office Task
|
demo automating PowerPoint from C++ app
#### This work item was migrated from CodePlex
CodePlex work item ID: '3216'
Vote count: '1'
|
1.0
|
Create Example: CppAutomatePowerPoint - demo automating PowerPoint from C++ app
#### This work item was migrated from CodePlex
CodePlex work item ID: '3216'
Vote count: '1'
|
non_process
|
create example cppautomatepowerpoint demo automating powerpoint from c app this work item was migrated from codeplex codeplex work item id vote count
| 0
|
136,448
| 18,734,121,911
|
IssuesEvent
|
2021-11-04 03:42:46
|
Dima2022/kaltura-sample-code
|
https://api.github.com/repos/Dima2022/kaltura-sample-code
|
opened
|
CVE-2018-19361 (High) detected in jackson-databind-2.8.9.jar
|
security vulnerability
|
## CVE-2018-19361 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.9.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: kaltura-sample-code/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.9/jackson-databind-2.8.9.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-1.5.6.RELEASE.jar (Root Library)
- :x: **jackson-databind-2.8.9.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Dima2022/kaltura-sample-code/commit/e022312630d07a000d07ace8bd86a7f0590a0f65">e022312630d07a000d07ace8bd86a7f0590a0f65</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.8 might allow attackers to have unspecified impact by leveraging failure to block the openjpa class from polymorphic deserialization.
<p>Publish Date: 2019-01-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-19361>CVE-2018-19361</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19361">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19361</a></p>
<p>Release Date: 2019-01-02</p>
<p>Fix Resolution: 2.9.8</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.9","packageFilePaths":["/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:1.5.6.RELEASE;com.fasterxml.jackson.core:jackson-databind:2.8.9","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.8"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2018-19361","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.8 might allow attackers to have unspecified impact by leveraging failure to block the openjpa class from polymorphic deserialization.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-19361","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2018-19361 (High) detected in jackson-databind-2.8.9.jar - ## CVE-2018-19361 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.9.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: kaltura-sample-code/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.9/jackson-databind-2.8.9.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-1.5.6.RELEASE.jar (Root Library)
- :x: **jackson-databind-2.8.9.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Dima2022/kaltura-sample-code/commit/e022312630d07a000d07ace8bd86a7f0590a0f65">e022312630d07a000d07ace8bd86a7f0590a0f65</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.8 might allow attackers to have unspecified impact by leveraging failure to block the openjpa class from polymorphic deserialization.
<p>Publish Date: 2019-01-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-19361>CVE-2018-19361</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19361">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19361</a></p>
<p>Release Date: 2019-01-02</p>
<p>Fix Resolution: 2.9.8</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.9","packageFilePaths":["/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:1.5.6.RELEASE;com.fasterxml.jackson.core:jackson-databind:2.8.9","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.8"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2018-19361","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.8 might allow attackers to have unspecified impact by leveraging failure to block the openjpa class from polymorphic deserialization.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-19361","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
non_process
|
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file kaltura sample code pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring boot starter web release jar root library x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details fasterxml jackson databind x before might allow attackers to have unspecified impact by leveraging failure to block the openjpa class from polymorphic deserialization publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org springframework boot spring boot starter web release com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before might allow attackers to have unspecified impact by leveraging failure to block the openjpa class from polymorphic deserialization vulnerabilityurl
| 0
|
178,270
| 13,770,994,106
|
IssuesEvent
|
2020-10-07 21:10:13
|
Azure/azure-sdk-for-cpp
|
https://api.github.com/repos/Azure/azure-sdk-for-cpp
|
closed
|
Transient core ctest failure
|
Azure.Core Client test bug
|
https://dev.azure.com/azure-sdk/public/_build/results?buildId=531116&view=logs&j=becb4c78-6d46-5b40-696d-491037a3748a&t=7da4ec77-80ef-5cae-b212-17e17d7c0a17&s=ff05ad62-bb9a-53b6-ce9f-72f329a63e7c
```
1: [==========] 50 tests from 8 test suites ran. (186524 ms total)
1: [ PASSED ] 48 tests.
1: [ FAILED ] 2 tests, listed below:
1: [ FAILED ] TransportAdapter.getMultiThread
1: [ FAILED ] TransportAdapter.ConnectionPoolCleaner
1:
1: 2 FAILED TESTS
1/1 Test #1: azure-core-test ..................***Failed 186.80 sec
```
This test failed once and then succeeded on a second run.
|
1.0
|
Transient core ctest failure - https://dev.azure.com/azure-sdk/public/_build/results?buildId=531116&view=logs&j=becb4c78-6d46-5b40-696d-491037a3748a&t=7da4ec77-80ef-5cae-b212-17e17d7c0a17&s=ff05ad62-bb9a-53b6-ce9f-72f329a63e7c
```
1: [==========] 50 tests from 8 test suites ran. (186524 ms total)
1: [ PASSED ] 48 tests.
1: [ FAILED ] 2 tests, listed below:
1: [ FAILED ] TransportAdapter.getMultiThread
1: [ FAILED ] TransportAdapter.ConnectionPoolCleaner
1:
1: 2 FAILED TESTS
1/1 Test #1: azure-core-test ..................***Failed 186.80 sec
```
This test failed once and then succeeded on a second run.
|
non_process
|
transient core ctest failure tests from test suites ran ms total tests tests listed below transportadapter getmultithread transportadapter connectionpoolcleaner failed tests test azure core test failed sec this test failed once and then succeeded on a second run
| 0
|
5,240
| 8,036,959,086
|
IssuesEvent
|
2018-07-30 10:54:24
|
Open-EO/openeo-api
|
https://api.github.com/repos/Open-EO/openeo-api
|
closed
|
Change the definition of a process_graph
|
process graphs vote work in progress
|
Issues such as #89 or #106 require a change in the process graph definition as objects are required to/should be used.
Values of a process argument are defined as follows:
`<Value> := <string|number|array|boolean|null|Process>`
Source: https://open-eo.github.io/openeo-api/v/0.3.0/processgraphs/index.html
For the mentioned issues (and probably others) we probably need to change the definition of values to:
`<Value> := <string|number|array|boolean|null|object|Process>`
@flahn Does this make things much more complicated for you? I remember that you voted to not allow objects in process graphs as argument values.
|
1.0
|
Change the definition of a process_graph - Issues such as #89 or #106 require a change in the process graph definition as objects are required to/should be used.
Values of a process argument are defined as follows:
`<Value> := <string|number|array|boolean|null|Process>`
Source: https://open-eo.github.io/openeo-api/v/0.3.0/processgraphs/index.html
For the mentioned issues (and probably others) we probably need to change the definition of values to:
`<Value> := <string|number|array|boolean|null|object|Process>`
@flahn Does this make things much more complicated for you? I remember that you voted to not allow objects in process graphs as argument values.
|
process
|
change the definition of a process graph issues such as or require a change in the process graph definition as objects are required to should be used values of a process argument are defined as follows source for the mentioned issues and probably others we probably need to change the definition of values to flahn does this make things much more complicated for you i remember that you voted to not allow objects in process graphs as argument values
| 1
|
7,652
| 10,739,205,794
|
IssuesEvent
|
2019-10-29 15:59:02
|
googleapis/java-bigtable
|
https://api.github.com/repos/googleapis/java-bigtable
|
closed
|
Add a BOM artifact
|
type: process
|
Please publish a google-cloud-bigtable-bom pom artifact that contains the synced versions of artifacts from this repository. We will consume them upstream in the google-cloud-bom as a single version rather than trying to keep 7+ artifact versions synced.
|
1.0
|
Add a BOM artifact - Please publish a google-cloud-bigtable-bom pom artifact that contains the synced versions of artifacts from this repository. We will consume them upstream in the google-cloud-bom as a single version rather than trying to keep 7+ artifact versions synced.
|
process
|
add a bom artifact please publish a google cloud bigtable bom pom artifact that contains the synced versions of artifacts from this repository we will consume them upstream in the google cloud bom as a single version rather than trying to keep artifact versions synced
| 1
|
256,988
| 22,140,198,354
|
IssuesEvent
|
2022-06-03 05:47:35
|
opensrp/opensrp-client-anc
|
https://api.github.com/repos/opensrp/opensrp-client-anc
|
closed
|
[Ona Support Request]: The min date and max date is not working well in Date picker widget in BS(Bikram Sambat)
|
medium priority qa+ Tech Partner (Nepal Team) client testing
|
### Affected App or Server Version
v1.6.10
### What kind of support do you need?
In registration form , when we enter date of birth .
The min date, and max date is not working as mentioned in code of <anc_register.json> .
### What is the acceptance criteria for your support request?
The min date, and max date need to work as mentioned in code of <anc_register.json> .
If such condition are on other form , we will check from our side after the issue in registration form is resolved.
### Relevant Information
_No response_
|
1.0
|
[Ona Support Request]: The min date and max date is not working well in Date picker widget in BS(Bikram Sambat) - ### Affected App or Server Version
v1.6.10
### What kind of support do you need?
In registration form , when we enter date of birth .
The min date, and max date is not working as mentioned in code of <anc_register.json> .
### What is the acceptance criteria for your support request?
The min date, and max date need to work as mentioned in code of <anc_register.json> .
If such condition are on other form , we will check from our side after the issue in registration form is resolved.
### Relevant Information
_No response_
|
non_process
|
the min date and max date is not working well in date picker widget in bs bikram sambat affected app or server version what kind of support do you need in registration form when we enter date of birth the min date and max date is not working as mentioned in code of what is the acceptance criteria for your support request the min date and max date need to work as mentioned in code of if such condition are on other form we will check from our side after the issue in registration form is resolved relevant information no response
| 0
|
9,820
| 12,827,013,055
|
IssuesEvent
|
2020-07-06 17:38:25
|
MicrosoftDocs/azure-devops-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
|
closed
|
Container Jobs and self hosted agents
|
Pri1 devops-cicd-process/tech devops/prod product-question
|
>You must specify "Hosted Ubuntu 1604" as the pool name in order to run Linux containers. Other pools won't work.
Does that effectively mean I cannot use this feature with a self hosted agent?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 3339a2e0-be29-1363-f588-b231d4472c02
* Version Independent ID: 72dd11a3-704d-d0fd-6dfa-cf49f3352de3
* Content: [Container Jobs in Azure Pipelines and TFS - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/container-phases?view=azure-devops)
* Content Source: [docs/pipelines/process/container-phases.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/container-phases.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
1.0
|
Container Jobs and self hosted agents - >You must specify "Hosted Ubuntu 1604" as the pool name in order to run Linux containers. Other pools won't work.
Does that effectively mean I cannot use this feature with a self hosted agent?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 3339a2e0-be29-1363-f588-b231d4472c02
* Version Independent ID: 72dd11a3-704d-d0fd-6dfa-cf49f3352de3
* Content: [Container Jobs in Azure Pipelines and TFS - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/container-phases?view=azure-devops)
* Content Source: [docs/pipelines/process/container-phases.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/container-phases.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
process
|
container jobs and self hosted agents you must specify hosted ubuntu as the pool name in order to run linux containers other pools won t work does that effectively mean i cannot use this feature with a self hosted agent document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
| 1
|
799,773
| 28,313,743,511
|
IssuesEvent
|
2023-04-10 17:42:26
|
asastats/channel
|
https://api.github.com/repos/asastats/channel
|
opened
|
Allow NFT amounts to be float numbers
|
bug medium priority mobile
|
Right now if a float is sent in API response then an error is raised, like in this example
https://www.asastats.com/6IW52JMHBXAWPFIAZTP6EVHZT43LA2HX7DKGOV56BXJCWCGGD2BZIV625I
for the 468359924:
```"nft_values": [{
"value": 13344.0,
"collection": "Alchemon",
"amount": 238.0,
"nfts": [{
...
...
{
"value": 0,
"id": 468359924,
"amount": 0.01,
"nft": {
"id": 468359924,
"name": "ART: MagicaL Bolt",
"unit": "ART6",
...
```
As a temporary solution, we've prevented raising error by converting those amounts to integers in API responses.
|
1.0
|
Allow NFT amounts to be float numbers - Right now if a float is sent in API response then an error is raised, like in this example
https://www.asastats.com/6IW52JMHBXAWPFIAZTP6EVHZT43LA2HX7DKGOV56BXJCWCGGD2BZIV625I
for the 468359924:
```"nft_values": [{
"value": 13344.0,
"collection": "Alchemon",
"amount": 238.0,
"nfts": [{
...
...
{
"value": 0,
"id": 468359924,
"amount": 0.01,
"nft": {
"id": 468359924,
"name": "ART: MagicaL Bolt",
"unit": "ART6",
...
```
As a temporary solution, we've prevented raising error by converting those amounts to integers in API responses.
|
non_process
|
allow nft amounts to be float numbers right now if a float is sent in api response then an error is raised like in this example for the nft values value collection alchemon amount nfts value id amount nft id name art magical bolt unit as a temporary solution we ve prevented raising error by converting those amounts to integers in api responses
| 0
|
18,845
| 24,756,079,989
|
IssuesEvent
|
2022-10-21 17:52:25
|
geneontology/go-ontology
|
https://api.github.com/repos/geneontology/go-ontology
|
closed
|
Hemeostasis: regulation of liquid surface tension: merge into surfactant homeostasis
|
organism-level process
|
Hello,
For the BP refactoring, we will split cellular and organism-level processes, as described in the [2022 bp refactoring plan - top level](https://docs.google.com/document/d/1k8yuWTMSkYVTnt2hRbrPVH8Ud3gDwC5IrGu4PqHkKds/edit).
GO:0050828 regulation of liquid surface tension is a direct child of [GO:0042592 homeostatic process](http://amigo.geneontology.org/amigo/term/GO:0042592#display-lineage-tab), which will be obsoleted, see #24218
This term will be merged and end up a child of 'GO:0048871 multicellular organismal homeostasis'
- [x] merge into GO:0050828 regulation of liquid surface tension into GO:0043129 surfactant homeostasis
- [x] obsolete GO:0101029 negative regulation of liquid surface tension & GO:0101028 positive regulation of liquid surface tension (no annotations)
- [x] change definition to "Any process involved in the maintenance of a steady-state level of a surface-active agent that maintains the surface tension of a liquid."
- [x] add reference: PMID:23708874
- [x] GO:0043129 surfactant homeostasis -> move under 'GO:0048871 multicellular organismal homeostasis'
Thanks, Pascale
|
1.0
|
Hemeostasis: regulation of liquid surface tension: merge into surfactant homeostasis - Hello,
For the BP refactoring, we will split cellular and organism-level processes, as described in the [2022 bp refactoring plan - top level](https://docs.google.com/document/d/1k8yuWTMSkYVTnt2hRbrPVH8Ud3gDwC5IrGu4PqHkKds/edit).
GO:0050828 regulation of liquid surface tension is a direct child of [GO:0042592 homeostatic process](http://amigo.geneontology.org/amigo/term/GO:0042592#display-lineage-tab), which will be obsoleted, see #24218
This term will be merged and end up a child of 'GO:0048871 multicellular organismal homeostasis'
- [x] merge into GO:0050828 regulation of liquid surface tension into GO:0043129 surfactant homeostasis
- [x] obsolete GO:0101029 negative regulation of liquid surface tension & GO:0101028 positive regulation of liquid surface tension (no annotations)
- [x] change definition to "Any process involved in the maintenance of a steady-state level of a surface-active agent that maintains the surface tension of a liquid."
- [x] add reference: PMID:23708874
- [x] GO:0043129 surfactant homeostasis -> move under 'GO:0048871 multicellular organismal homeostasis'
Thanks, Pascale
|
process
|
hemeostasis regulation of liquid surface tension merge into surfactant homeostasis hello for the bp refactoring we will split cellular and organism level processes as described in the go regulation of liquid surface tension is a direct child of which will be obsoleted see this term will be merged and end up a child of go multicellular organismal homeostasis merge into go regulation of liquid surface tension into go surfactant homeostasis obsolete go negative regulation of liquid surface tension go positive regulation of liquid surface tension no annotations change definition to any process involved in the maintenance of a steady state level of a surface active agent that maintains the surface tension of a liquid add reference pmid go surfactant homeostasis move under go multicellular organismal homeostasis thanks pascale
| 1
|
43,047
| 11,141,171,025
|
IssuesEvent
|
2019-12-21 20:19:12
|
neuronsimulator/nrn
|
https://api.github.com/repos/neuronsimulator/nrn
|
closed
|
Build issues with GCC v9
|
bug building
|
While building NEURON with CMake using GCC 9 (or 9.2), I am seeing below errors:
```
$ cmake .. -DNRN_ENABLE_CORENEURON=OFF -DNRN_ENABLE_INTERVIEWS=OFF
....
-- The C compiler identification is GNU 9.2.0
-- The CXX compiler identification is GNU 9.2.0
-- Check for working C compiler: /gpfs/bbp.cscs.ch/apps/hpc/jenkins/i_would_like_to_be_deleted---merge-2019-10/deploy/compilers/2019-10-02/linux-rhel7-x86_64/gcc-4.8.5/gcc-9.2.0-tuxi4uvc6t/bin/gcc
-- Check for working C compiler: /gpfs/bbp.cscs.ch/apps/hpc/jenkins/i_would_like_to_be_deleted---merge-2019-10/deploy/compilers/2019-10-02/linux-rhel7-x86_64/gcc-4.8.5/gcc-9.2.0-tuxi4uvc6t/bin/gcc -- works
-- Detecting C compiler ABI info
-- Detecting C compiler ABI info - done
-- Detecting C compile features
...
--
-- Configured NEURON 7.8
--
-- Some things you can do now:
-- --------------+--------------------------------------------------------------
-- Command | Description
-- --------------+--------------------------------------------------------------
-- make install | Will install NEURON to: /usr/local
-- | Change the install location of NEURON using:
-- | cmake <src_path> -DCMAKE_INSTALL_PREFIX=<install_path>
-- make docs | Build the API documentation
-- make uninstall| Removes files installed by make install (todo)
-- --------------+--------------------------------------------------------------
-- Build option | Status
-- --------------+--------------------------------------------------------------
-- Shared | ON
-- Legacy FR | ON
-- MPI | ON
-- INC | /opt/hpe/hpc/mpt/mpt-2.16/include
-- LIB | /lib64/libpthread.so
-- Python | ON
-- EXE | /gpfs/bbp.cscs.ch/apps/hpc/jenkins/deploy/external-libraries/2018-12-19/linux-rhel7-x86_64/gcc-6.4.0/python-3.6.5-ukuow6gd2f/bin/python
-- INC | /gpfs/bbp.cscs.ch/apps/hpc/jenkins/deploy/external-libraries/2018-12-19/linux-rhel7-x86_64/gcc-6.4.0/python-3.6.5-ukuow6gd2f/include/python3.6m
-- LIB | /gpfs/bbp.cscs.ch/apps/hpc/jenkins/deploy/external-libraries/2018-12-19/linux-rhel7-x86_64/gcc-6.4.0/python-3.6.5-ukuow6gd2f/lib/libpython3.6m.so
-- DYNAMIC | OFF
-- RX3D | ON
-- OptLevel | 0
-- Interviews | OFF
-- CoreNEURON | OFF
-- --------------+--------------------------------------------------------------
-- See documentation : https://www.neuron.yale.edu/neuron/
-- --------------+--------------------------------------------------------------
--
-- Configuring done
-- Generating done
-- Build files have been written to: /gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build
...
$ make VERBOSE=1
/gpfs/bbp.cscs.ch/apps/hpc/jenkins/deploy/tools/2018-12-19/linux-rhel7-x86_64/gcc-6.4.0/cmake-3.13.0-ivumccos42/bin/cmake -S/gpfs/bbp.cscs.ch/home/kumbhar/tmp/
....
[ 6%] Building CXX object src/nrniv/CMakeFiles/nrniv_lib.dir/bbsdirect.cpp.o
cd /gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build/src/nrniv && /gpfs/bbp.cscs.ch/apps/hpc/jenkins/i_would_like_to_be_deleted---merge-2019-10/deploy/compilers/2019-10-02/linux-rhel7-x86_64/gcc-4.8.5/gcc-9.2.0-tuxi4uvc6t/bin/g++ -DHAVE_CONFIG_H -Dnrniv_lib_EXPORTS -DCABLE=1 -DOOP=1 -DUSEBBS=1 -DUSECVODE=1 -DUSEMATRIX=1 -I/opt/hpe/hpc/mpt/mpt-2.16/include -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/ivoc -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrnoc -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/oc -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build/src/nrncvode -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build/src/nrnjava -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build/src/nrnoc -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build/src/nrnpython -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build/src/oc -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build/src/parallel -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build/src/sundials -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build/src/sundials/shared -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/gnu -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/memacs -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/mesch -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrncvode -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrnmpi -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrnpython -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/parallel -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/sparse -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/sparse13 -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/sundials -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/sundials/cvodes -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/sundials/ida -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/sundials/shared -I/gpfs/bbp.cscs.ch/apps/hpc/jenkins/deploy/external-libraries/2018-12-19/linux-rhel7-x86_64/gcc-6.4.0/python-3.6.5-ukuow6gd2f/include/python3.6m -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/nrniv_lib -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/ivos -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build/src/ivos -fPIC -std=c++98 -o CMakeFiles/nrniv_lib.dir/bbsdirect.cpp.o -c /gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/bbsdirect.cpp
In file included from /gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/bbsdirect.cpp:3:
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:35:28: error: expected template-name before ‘<’ token
35 | class KeepArgs : public map<int, bbsmpibuf*, ltint>{};
| ^
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:35:28: error: expected ‘{’ before ‘<’ token
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:35:28: error: expected unqualified-id before ‘<’ token
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp: In constructor ‘BBSDirect::BBSDirect()’:
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:48:27: error: invalid use of incomplete type ‘class KeepArgs’
48 | keepargs_ = new KeepArgs();
| ^
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:35:7: note: forward declaration of ‘class KeepArgs’
35 | class KeepArgs : public map<int, bbsmpibuf*, ltint>{};
| ^~~~~~~~
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp: In destructor ‘virtual BBSDirect::~BBSDirect()’:
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:56:9: warning: possible problem detected in invocation of delete operator: [-Wdelete-incomplete]
56 | delete keepargs_;
| ^~~~~~~~~
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:56:9: warning: invalid use of incomplete type ‘class KeepArgs’
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:35:7: note: forward declaration of ‘class KeepArgs’
35 | class KeepArgs : public map<int, bbsmpibuf*, ltint>{};
| ^~~~~~~~
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:56:9: note: neither the destructor nor the class-specific operator delete will be called, even if they are declared when the class is defined
56 | delete keepargs_;
| ^~~~~~~~~
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp: In member function ‘virtual void BBSDirect::save_args(int)’:
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:269:11: error: invalid use of incomplete type ‘class KeepArgs’
269 | keepargs_->insert(
| ^~
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:35:7: note: forward declaration of ‘class KeepArgs’
35 | class KeepArgs : public map<int, bbsmpibuf*, ltint>{};
| ^~~~~~~~
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:270:3: error: ‘pair’ was not declared in this scope; did you mean ‘std::pair’?
270 | pair<const int, bbsmpibuf*>(userid, sendbuf_)
| ^~~~
| std::pair
In file included from /gpfs/bbp.cscs.ch/apps/hpc/jenkins/i_would_like_to_be_deleted---merge-2019-10/deploy/compilers/2019-10-02/linux-rhel7-x86_64/gcc-4.8.5/gcc-9.2.0-tuxi4uvc6t/include/c++/9.2.0/bits/stl_algobase.h:64,
from /gpfs/bbp.cscs.ch/apps/hpc/jenkins/i_would_like_to_be_deleted---merge-2019-10/deploy/compilers/2019-10-02/linux-rhel7-x86_64/gcc-4.8.5/gcc-9.2.0-tuxi4uvc6t/include/c++/9.2.0/bits/char_traits.h:39,
from /gpfs/bbp.cscs.ch/apps/hpc/jenkins/i_would_like_to_be_deleted---merge-2019-10/deploy/compilers/2019-10-02/linux-rhel7-x86_64/gcc-4.8.5/gcc-9.2.0-tuxi4uvc6t/include/c++/9.2.0/ios:40,
from /gpfs/bbp.cscs.ch/apps/hpc/jenkins/i_would_like_to_be_deleted---merge-2019-10/deploy/compilers/2019-10-02/linux-rhel7-x86_64/gcc-4.8.5/gcc-9.2.0-tuxi4uvc6t/include/c++/9.2.0/ostream:38,
from /gpfs/bbp.cscs.ch/apps/hpc/jenkins/i_would_like_to_be_deleted---merge-2019-10/deploy/compilers/2019-10-02/linux-rhel7-x86_64/gcc-4.8.5/gcc-9.2.0-tuxi4uvc6t/include/c++/9.2.0/iostream:39,
from /gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/ivos/ivstream.h:4,
from /gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build/src/nrncvode/../../nrnconf.h:370,
from /gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:1,
from /gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/bbsdirect.cpp:3:
/gpfs/bbp.cscs.ch/apps/hpc/jenkins/i_would_like_to_be_deleted---merge-2019-10/deploy/compilers/2019-10-02/linux-rhel7-x86_64/gcc-4.8.5/gcc-9.2.0-tuxi4uvc6t/include/c++/9.2.0/bits/stl_pair.h:208:12: note: ‘std::pair’ declared here
208 | struct pair
| ^~~~
In file included from /gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/bbsdirect.cpp:3:
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:270:8: error: expected primary-expression before ‘const’
270 | pair<const int, bbsmpibuf*>(userid, sendbuf_)
| ^~~~~
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:270:28: error: expected primary-expression before ‘*’ token
270 | pair<const int, bbsmpibuf*>(userid, sendbuf_)
| ^
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:270:29: error: expected primary-expression before ‘>’ token
270 | pair<const int, bbsmpibuf*>(userid, sendbuf_)
| ^
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp: In member function ‘virtual void BBSDirect::return_args(int)’:
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:279:12: error: incomplete type ‘KeepArgs’ used in nested name specifier
279 | KeepArgs::iterator i = keepargs_->find(userid);
| ^~~~~~~~
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:282:6: error: ‘i’ was not declared in this scope
282 | if (i != keepargs_->end()) {
| ^
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:282:20: error: invalid use of incomplete type ‘class KeepArgs’
282 | if (i != keepargs_->end()) {
| ^~
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:35:7: note: forward declaration of ‘class KeepArgs’
35 | class KeepArgs : public map<int, bbsmpibuf*, ltint>{};
| ^~~~~~~~
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:284:12: error: invalid use of incomplete type ‘class KeepArgs’
284 | keepargs_->erase(i);
| ^~
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:35:7: note: forward declaration of ‘class KeepArgs’
35 | class KeepArgs : public map<int, bbsmpibuf*, ltint>{};
| ^~~~~~~~
make[2]: *** [src/nrniv/CMakeFiles/nrniv_lib.dir/bbsdirect.cpp.o] Error 1
make[2]: Leaving directory `/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build'
make[1]: *** [src/nrniv/CMakeFiles/nrniv_lib.dir/all] Error 2
make[1]: Leaving directory `/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build'
make: *** [all] Error 2
```
This seems to be related to #358 but I am surprised that this is only appearing with GCC v9.
I am shorty creating PR with the fix.
cc: @anilbey
|
1.0
|
Build issues with GCC v9 - While building NEURON with CMake using GCC 9 (or 9.2), I am seeing below errors:
```
$ cmake .. -DNRN_ENABLE_CORENEURON=OFF -DNRN_ENABLE_INTERVIEWS=OFF
....
-- The C compiler identification is GNU 9.2.0
-- The CXX compiler identification is GNU 9.2.0
-- Check for working C compiler: /gpfs/bbp.cscs.ch/apps/hpc/jenkins/i_would_like_to_be_deleted---merge-2019-10/deploy/compilers/2019-10-02/linux-rhel7-x86_64/gcc-4.8.5/gcc-9.2.0-tuxi4uvc6t/bin/gcc
-- Check for working C compiler: /gpfs/bbp.cscs.ch/apps/hpc/jenkins/i_would_like_to_be_deleted---merge-2019-10/deploy/compilers/2019-10-02/linux-rhel7-x86_64/gcc-4.8.5/gcc-9.2.0-tuxi4uvc6t/bin/gcc -- works
-- Detecting C compiler ABI info
-- Detecting C compiler ABI info - done
-- Detecting C compile features
...
--
-- Configured NEURON 7.8
--
-- Some things you can do now:
-- --------------+--------------------------------------------------------------
-- Command | Description
-- --------------+--------------------------------------------------------------
-- make install | Will install NEURON to: /usr/local
-- | Change the install location of NEURON using:
-- | cmake <src_path> -DCMAKE_INSTALL_PREFIX=<install_path>
-- make docs | Build the API documentation
-- make uninstall| Removes files installed by make install (todo)
-- --------------+--------------------------------------------------------------
-- Build option | Status
-- --------------+--------------------------------------------------------------
-- Shared | ON
-- Legacy FR | ON
-- MPI | ON
-- INC | /opt/hpe/hpc/mpt/mpt-2.16/include
-- LIB | /lib64/libpthread.so
-- Python | ON
-- EXE | /gpfs/bbp.cscs.ch/apps/hpc/jenkins/deploy/external-libraries/2018-12-19/linux-rhel7-x86_64/gcc-6.4.0/python-3.6.5-ukuow6gd2f/bin/python
-- INC | /gpfs/bbp.cscs.ch/apps/hpc/jenkins/deploy/external-libraries/2018-12-19/linux-rhel7-x86_64/gcc-6.4.0/python-3.6.5-ukuow6gd2f/include/python3.6m
-- LIB | /gpfs/bbp.cscs.ch/apps/hpc/jenkins/deploy/external-libraries/2018-12-19/linux-rhel7-x86_64/gcc-6.4.0/python-3.6.5-ukuow6gd2f/lib/libpython3.6m.so
-- DYNAMIC | OFF
-- RX3D | ON
-- OptLevel | 0
-- Interviews | OFF
-- CoreNEURON | OFF
-- --------------+--------------------------------------------------------------
-- See documentation : https://www.neuron.yale.edu/neuron/
-- --------------+--------------------------------------------------------------
--
-- Configuring done
-- Generating done
-- Build files have been written to: /gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build
...
$ make VERBOSE=1
/gpfs/bbp.cscs.ch/apps/hpc/jenkins/deploy/tools/2018-12-19/linux-rhel7-x86_64/gcc-6.4.0/cmake-3.13.0-ivumccos42/bin/cmake -S/gpfs/bbp.cscs.ch/home/kumbhar/tmp/
....
[ 6%] Building CXX object src/nrniv/CMakeFiles/nrniv_lib.dir/bbsdirect.cpp.o
cd /gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build/src/nrniv && /gpfs/bbp.cscs.ch/apps/hpc/jenkins/i_would_like_to_be_deleted---merge-2019-10/deploy/compilers/2019-10-02/linux-rhel7-x86_64/gcc-4.8.5/gcc-9.2.0-tuxi4uvc6t/bin/g++ -DHAVE_CONFIG_H -Dnrniv_lib_EXPORTS -DCABLE=1 -DOOP=1 -DUSEBBS=1 -DUSECVODE=1 -DUSEMATRIX=1 -I/opt/hpe/hpc/mpt/mpt-2.16/include -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/ivoc -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrnoc -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/oc -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build/src/nrncvode -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build/src/nrnjava -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build/src/nrnoc -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build/src/nrnpython -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build/src/oc -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build/src/parallel -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build/src/sundials -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build/src/sundials/shared -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/gnu -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/memacs -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/mesch -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrncvode -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrnmpi -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrnpython -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/parallel -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/sparse -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/sparse13 -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/sundials -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/sundials/cvodes -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/sundials/ida -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/sundials/shared -I/gpfs/bbp.cscs.ch/apps/hpc/jenkins/deploy/external-libraries/2018-12-19/linux-rhel7-x86_64/gcc-6.4.0/python-3.6.5-ukuow6gd2f/include/python3.6m -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/nrniv_lib -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/ivos -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build/src/ivos -fPIC -std=c++98 -o CMakeFiles/nrniv_lib.dir/bbsdirect.cpp.o -c /gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/bbsdirect.cpp
In file included from /gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/bbsdirect.cpp:3:
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:35:28: error: expected template-name before ‘<’ token
35 | class KeepArgs : public map<int, bbsmpibuf*, ltint>{};
| ^
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:35:28: error: expected ‘{’ before ‘<’ token
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:35:28: error: expected unqualified-id before ‘<’ token
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp: In constructor ‘BBSDirect::BBSDirect()’:
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:48:27: error: invalid use of incomplete type ‘class KeepArgs’
48 | keepargs_ = new KeepArgs();
| ^
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:35:7: note: forward declaration of ‘class KeepArgs’
35 | class KeepArgs : public map<int, bbsmpibuf*, ltint>{};
| ^~~~~~~~
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp: In destructor ‘virtual BBSDirect::~BBSDirect()’:
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:56:9: warning: possible problem detected in invocation of delete operator: [-Wdelete-incomplete]
56 | delete keepargs_;
| ^~~~~~~~~
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:56:9: warning: invalid use of incomplete type ‘class KeepArgs’
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:35:7: note: forward declaration of ‘class KeepArgs’
35 | class KeepArgs : public map<int, bbsmpibuf*, ltint>{};
| ^~~~~~~~
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:56:9: note: neither the destructor nor the class-specific operator delete will be called, even if they are declared when the class is defined
56 | delete keepargs_;
| ^~~~~~~~~
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp: In member function ‘virtual void BBSDirect::save_args(int)’:
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:269:11: error: invalid use of incomplete type ‘class KeepArgs’
269 | keepargs_->insert(
| ^~
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:35:7: note: forward declaration of ‘class KeepArgs’
35 | class KeepArgs : public map<int, bbsmpibuf*, ltint>{};
| ^~~~~~~~
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:270:3: error: ‘pair’ was not declared in this scope; did you mean ‘std::pair’?
270 | pair<const int, bbsmpibuf*>(userid, sendbuf_)
| ^~~~
| std::pair
In file included from /gpfs/bbp.cscs.ch/apps/hpc/jenkins/i_would_like_to_be_deleted---merge-2019-10/deploy/compilers/2019-10-02/linux-rhel7-x86_64/gcc-4.8.5/gcc-9.2.0-tuxi4uvc6t/include/c++/9.2.0/bits/stl_algobase.h:64,
from /gpfs/bbp.cscs.ch/apps/hpc/jenkins/i_would_like_to_be_deleted---merge-2019-10/deploy/compilers/2019-10-02/linux-rhel7-x86_64/gcc-4.8.5/gcc-9.2.0-tuxi4uvc6t/include/c++/9.2.0/bits/char_traits.h:39,
from /gpfs/bbp.cscs.ch/apps/hpc/jenkins/i_would_like_to_be_deleted---merge-2019-10/deploy/compilers/2019-10-02/linux-rhel7-x86_64/gcc-4.8.5/gcc-9.2.0-tuxi4uvc6t/include/c++/9.2.0/ios:40,
from /gpfs/bbp.cscs.ch/apps/hpc/jenkins/i_would_like_to_be_deleted---merge-2019-10/deploy/compilers/2019-10-02/linux-rhel7-x86_64/gcc-4.8.5/gcc-9.2.0-tuxi4uvc6t/include/c++/9.2.0/ostream:38,
from /gpfs/bbp.cscs.ch/apps/hpc/jenkins/i_would_like_to_be_deleted---merge-2019-10/deploy/compilers/2019-10-02/linux-rhel7-x86_64/gcc-4.8.5/gcc-9.2.0-tuxi4uvc6t/include/c++/9.2.0/iostream:39,
from /gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/ivos/ivstream.h:4,
from /gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build/src/nrncvode/../../nrnconf.h:370,
from /gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:1,
from /gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/bbsdirect.cpp:3:
/gpfs/bbp.cscs.ch/apps/hpc/jenkins/i_would_like_to_be_deleted---merge-2019-10/deploy/compilers/2019-10-02/linux-rhel7-x86_64/gcc-4.8.5/gcc-9.2.0-tuxi4uvc6t/include/c++/9.2.0/bits/stl_pair.h:208:12: note: ‘std::pair’ declared here
208 | struct pair
| ^~~~
In file included from /gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/bbsdirect.cpp:3:
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:270:8: error: expected primary-expression before ‘const’
270 | pair<const int, bbsmpibuf*>(userid, sendbuf_)
| ^~~~~
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:270:28: error: expected primary-expression before ‘*’ token
270 | pair<const int, bbsmpibuf*>(userid, sendbuf_)
| ^
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:270:29: error: expected primary-expression before ‘>’ token
270 | pair<const int, bbsmpibuf*>(userid, sendbuf_)
| ^
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp: In member function ‘virtual void BBSDirect::return_args(int)’:
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:279:12: error: incomplete type ‘KeepArgs’ used in nested name specifier
279 | KeepArgs::iterator i = keepargs_->find(userid);
| ^~~~~~~~
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:282:6: error: ‘i’ was not declared in this scope
282 | if (i != keepargs_->end()) {
| ^
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:282:20: error: invalid use of incomplete type ‘class KeepArgs’
282 | if (i != keepargs_->end()) {
| ^~
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:35:7: note: forward declaration of ‘class KeepArgs’
35 | class KeepArgs : public map<int, bbsmpibuf*, ltint>{};
| ^~~~~~~~
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:284:12: error: invalid use of incomplete type ‘class KeepArgs’
284 | keepargs_->erase(i);
| ^~
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:35:7: note: forward declaration of ‘class KeepArgs’
35 | class KeepArgs : public map<int, bbsmpibuf*, ltint>{};
| ^~~~~~~~
make[2]: *** [src/nrniv/CMakeFiles/nrniv_lib.dir/bbsdirect.cpp.o] Error 1
make[2]: Leaving directory `/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build'
make[1]: *** [src/nrniv/CMakeFiles/nrniv_lib.dir/all] Error 2
make[1]: Leaving directory `/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build'
make: *** [all] Error 2
```
This seems to be related to #358 but I am surprised that this is only appearing with GCC v9.
I am shorty creating PR with the fix.
cc: @anilbey
|
non_process
|
build issues with gcc while building neuron with cmake using gcc or i am seeing below errors cmake dnrn enable coreneuron off dnrn enable interviews off the c compiler identification is gnu the cxx compiler identification is gnu check for working c compiler gpfs bbp cscs ch apps hpc jenkins i would like to be deleted merge deploy compilers linux gcc gcc bin gcc check for working c compiler gpfs bbp cscs ch apps hpc jenkins i would like to be deleted merge deploy compilers linux gcc gcc bin gcc works detecting c compiler abi info detecting c compiler abi info done detecting c compile features configured neuron some things you can do now command description make install will install neuron to usr local change the install location of neuron using cmake dcmake install prefix make docs build the api documentation make uninstall removes files installed by make install todo build option status shared on legacy fr on mpi on inc opt hpe hpc mpt mpt include lib libpthread so python on exe gpfs bbp cscs ch apps hpc jenkins deploy external libraries linux gcc python bin python inc gpfs bbp cscs ch apps hpc jenkins deploy external libraries linux gcc python include lib gpfs bbp cscs ch apps hpc jenkins deploy external libraries linux gcc python lib so dynamic off on optlevel interviews off coreneuron off see documentation configuring done generating done build files have been written to gpfs bbp cscs ch home kumbhar tmp nrn build make verbose gpfs bbp cscs ch apps hpc jenkins deploy tools linux gcc cmake bin cmake s gpfs bbp cscs ch home kumbhar tmp building cxx object src nrniv cmakefiles nrniv lib dir bbsdirect cpp o cd gpfs bbp cscs ch home kumbhar tmp nrn build src nrniv gpfs bbp cscs ch apps hpc jenkins i would like to be deleted merge deploy compilers linux gcc gcc bin g dhave config h dnrniv lib exports dcable doop dusebbs dusecvode dusematrix i opt hpe hpc mpt mpt include i gpfs bbp cscs ch home kumbhar tmp nrn src ivoc i gpfs bbp cscs ch home kumbhar tmp nrn src nrniv i gpfs bbp cscs ch home kumbhar tmp nrn src nrnoc i gpfs bbp cscs ch home kumbhar tmp nrn src oc i gpfs bbp cscs ch home kumbhar tmp nrn build i gpfs bbp cscs ch home kumbhar tmp nrn build src nrncvode i gpfs bbp cscs ch home kumbhar tmp nrn build src nrnjava i gpfs bbp cscs ch home kumbhar tmp nrn build src nrnoc i gpfs bbp cscs ch home kumbhar tmp nrn build src nrnpython i gpfs bbp cscs ch home kumbhar tmp nrn build src oc i gpfs bbp cscs ch home kumbhar tmp nrn build src parallel i gpfs bbp cscs ch home kumbhar tmp nrn build src sundials i gpfs bbp cscs ch home kumbhar tmp nrn build src sundials shared i gpfs bbp cscs ch home kumbhar tmp nrn src i gpfs bbp cscs ch home kumbhar tmp nrn src gnu i gpfs bbp cscs ch home kumbhar tmp nrn src memacs i gpfs bbp cscs ch home kumbhar tmp nrn src mesch i gpfs bbp cscs ch home kumbhar tmp nrn src nrncvode i gpfs bbp cscs ch home kumbhar tmp nrn src nrnmpi i gpfs bbp cscs ch home kumbhar tmp nrn src nrnpython i gpfs bbp cscs ch home kumbhar tmp nrn src parallel i gpfs bbp cscs ch home kumbhar tmp nrn src sparse i gpfs bbp cscs ch home kumbhar tmp nrn src i gpfs bbp cscs ch home kumbhar tmp nrn src sundials i gpfs bbp cscs ch home kumbhar tmp nrn src sundials cvodes i gpfs bbp cscs ch home kumbhar tmp nrn src sundials ida i gpfs bbp cscs ch home kumbhar tmp nrn src sundials shared i gpfs bbp cscs ch apps hpc jenkins deploy external libraries linux gcc python include i gpfs bbp cscs ch home kumbhar tmp nrn src nrniv nrniv lib i gpfs bbp cscs ch home kumbhar tmp nrn src ivos i gpfs bbp cscs ch home kumbhar tmp nrn build src ivos fpic std c o cmakefiles nrniv lib dir bbsdirect cpp o c gpfs bbp cscs ch home kumbhar tmp nrn src nrniv bbsdirect cpp in file included from gpfs bbp cscs ch home kumbhar tmp nrn src nrniv bbsdirect cpp gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp error expected template name before ‘ ’ token class keepargs public map gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp error expected ‘ ’ before ‘ ’ token gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp error expected unqualified id before ‘ ’ token gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp in constructor ‘bbsdirect bbsdirect ’ gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp error invalid use of incomplete type ‘class keepargs’ keepargs new keepargs gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp note forward declaration of ‘class keepargs’ class keepargs public map gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp in destructor ‘virtual bbsdirect bbsdirect ’ gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp warning possible problem detected in invocation of delete operator delete keepargs gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp warning invalid use of incomplete type ‘class keepargs’ gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp note forward declaration of ‘class keepargs’ class keepargs public map gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp note neither the destructor nor the class specific operator delete will be called even if they are declared when the class is defined delete keepargs gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp in member function ‘virtual void bbsdirect save args int ’ gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp error invalid use of incomplete type ‘class keepargs’ keepargs insert gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp note forward declaration of ‘class keepargs’ class keepargs public map gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp error ‘pair’ was not declared in this scope did you mean ‘std pair’ pair userid sendbuf std pair in file included from gpfs bbp cscs ch apps hpc jenkins i would like to be deleted merge deploy compilers linux gcc gcc include c bits stl algobase h from gpfs bbp cscs ch apps hpc jenkins i would like to be deleted merge deploy compilers linux gcc gcc include c bits char traits h from gpfs bbp cscs ch apps hpc jenkins i would like to be deleted merge deploy compilers linux gcc gcc include c ios from gpfs bbp cscs ch apps hpc jenkins i would like to be deleted merge deploy compilers linux gcc gcc include c ostream from gpfs bbp cscs ch apps hpc jenkins i would like to be deleted merge deploy compilers linux gcc gcc include c iostream from gpfs bbp cscs ch home kumbhar tmp nrn src ivos ivstream h from gpfs bbp cscs ch home kumbhar tmp nrn build src nrncvode nrnconf h from gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp from gpfs bbp cscs ch home kumbhar tmp nrn src nrniv bbsdirect cpp gpfs bbp cscs ch apps hpc jenkins i would like to be deleted merge deploy compilers linux gcc gcc include c bits stl pair h note ‘std pair’ declared here struct pair in file included from gpfs bbp cscs ch home kumbhar tmp nrn src nrniv bbsdirect cpp gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp error expected primary expression before ‘const’ pair userid sendbuf gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp error expected primary expression before ‘ ’ token pair userid sendbuf gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp error expected primary expression before ‘ ’ token pair userid sendbuf gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp in member function ‘virtual void bbsdirect return args int ’ gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp error incomplete type ‘keepargs’ used in nested name specifier keepargs iterator i keepargs find userid gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp error ‘i’ was not declared in this scope if i keepargs end gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp error invalid use of incomplete type ‘class keepargs’ if i keepargs end gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp note forward declaration of ‘class keepargs’ class keepargs public map gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp error invalid use of incomplete type ‘class keepargs’ keepargs erase i gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp note forward declaration of ‘class keepargs’ class keepargs public map make error make leaving directory gpfs bbp cscs ch home kumbhar tmp nrn build make error make leaving directory gpfs bbp cscs ch home kumbhar tmp nrn build make error this seems to be related to but i am surprised that this is only appearing with gcc i am shorty creating pr with the fix cc anilbey
| 0
|
7,892
| 11,072,292,425
|
IssuesEvent
|
2019-12-12 10:00:42
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
Using node with process.stdin inside spawned shell
|
child_process
|
* **Version**: `v8.4.0`
* **Platform**: `Darwin duplo 15.6.0 Darwin Kernel Version 15.6.0: Tue Apr 11 16:00:51 PDT 2017; root:xnu-3248.60.11.5.3~1/RELEASE_X86_64 x86_64`
* **Subsystem**: net, stream or child_process (?)
Tested with:
- `GNU bash, version 3.2.57(1)-release (x86_64-apple-darwin15)`
- `GNU bash, version 4.2.25(1)-release (x86_64-pc-linux-gnu)`
<!-- Enter your issue details below this comment. -->
Not sure is this the same or another bug but in my environment it throws:
```
Error: write EPIPE
at _errnoException (util.js:1041:11)
at WriteWrap.afterWrite [as oncomplete] (net.js:858:14)
```
But when we played with it a little:
```
Error: This socket has been ended by the other party
at Socket.writeAfterFIN [as write] (net.js:352:12)
at __dirname (/Users/qfox/repos/trendbox-ci/omg.js:11:11)
at <anonymous>
```
How to repro:
```js
const { spawn } = require('child_process');
const pause = t => new Promise(resolve => setTimeout(resolve, t));
const { stdin } = spawn('bash', [/*'-i'*/], { stdio: ['pipe', 'inherit'] });
(async () => {
stdin.write('echo All is fine here\n');
await pause(200);
stdin.write('node -e process.stdin && echo After process.stdin\n');
await pause(2000); // It's important: the bigger pause, the more frequently occurs
stdin.write('echo After all\n');
})().catch(e => console.error(e.stack));
```
If we remove usage of `process.stdin` or force shell to be interactive (pass in `['-i']` flag) or just call `node -e process.stdin` with `</dev/stdin` — then the problem disappears.
Would be nice to have a good error message if it's unfixable ;'-(
upd:
See also: #947, #2339, https://github.com/nodejs/node/issues/13278
|
1.0
|
Using node with process.stdin inside spawned shell - * **Version**: `v8.4.0`
* **Platform**: `Darwin duplo 15.6.0 Darwin Kernel Version 15.6.0: Tue Apr 11 16:00:51 PDT 2017; root:xnu-3248.60.11.5.3~1/RELEASE_X86_64 x86_64`
* **Subsystem**: net, stream or child_process (?)
Tested with:
- `GNU bash, version 3.2.57(1)-release (x86_64-apple-darwin15)`
- `GNU bash, version 4.2.25(1)-release (x86_64-pc-linux-gnu)`
<!-- Enter your issue details below this comment. -->
Not sure is this the same or another bug but in my environment it throws:
```
Error: write EPIPE
at _errnoException (util.js:1041:11)
at WriteWrap.afterWrite [as oncomplete] (net.js:858:14)
```
But when we played with it a little:
```
Error: This socket has been ended by the other party
at Socket.writeAfterFIN [as write] (net.js:352:12)
at __dirname (/Users/qfox/repos/trendbox-ci/omg.js:11:11)
at <anonymous>
```
How to repro:
```js
const { spawn } = require('child_process');
const pause = t => new Promise(resolve => setTimeout(resolve, t));
const { stdin } = spawn('bash', [/*'-i'*/], { stdio: ['pipe', 'inherit'] });
(async () => {
stdin.write('echo All is fine here\n');
await pause(200);
stdin.write('node -e process.stdin && echo After process.stdin\n');
await pause(2000); // It's important: the bigger pause, the more frequently occurs
stdin.write('echo After all\n');
})().catch(e => console.error(e.stack));
```
If we remove usage of `process.stdin` or force shell to be interactive (pass in `['-i']` flag) or just call `node -e process.stdin` with `</dev/stdin` — then the problem disappears.
Would be nice to have a good error message if it's unfixable ;'-(
upd:
See also: #947, #2339, https://github.com/nodejs/node/issues/13278
|
process
|
using node with process stdin inside spawned shell version platform darwin duplo darwin kernel version tue apr pdt root xnu release subsystem net stream or child process tested with gnu bash version release apple gnu bash version release pc linux gnu not sure is this the same or another bug but in my environment it throws error write epipe at errnoexception util js at writewrap afterwrite net js but when we played with it a little error this socket has been ended by the other party at socket writeafterfin net js at dirname users qfox repos trendbox ci omg js at how to repro js const spawn require child process const pause t new promise resolve settimeout resolve t const stdin spawn bash stdio async stdin write echo all is fine here n await pause stdin write node e process stdin echo after process stdin n await pause it s important the bigger pause the more frequently occurs stdin write echo after all n catch e console error e stack if we remove usage of process stdin or force shell to be interactive pass in flag or just call node e process stdin with dev stdin — then the problem disappears would be nice to have a good error message if it s unfixable upd see also
| 1
|
14,567
| 10,194,500,728
|
IssuesEvent
|
2019-08-12 15:49:38
|
aws/aws-cdk
|
https://api.github.com/repos/aws/aws-cdk
|
closed
|
We don't have an AMI for kubernetes version 1.13
|
bug needs-repro service/eks
|
**Note: for support questions, please first reference our [documentation](https://docs.aws.amazon.com/cdk/api/latest), then use [Stackoverflow](https://stackoverflow.com/questions/ask?tags=aws-cdk)**. This repository's issues are intended for feature requests and bug reports.
* **I'm submitting a ...**
- [x] :beetle: bug report
- [ ] :rocket: feature request
- [ ] :books: construct library gap
- [ ] :phone: security issue or vulnerability => Please see [policy](https://github.com/awslabs/aws-cdk/security/policy)
- [ ] :question: support request => Please see note at the top of this template.
* **What is the current behavior?**
*If the current behavior is a :beetle:bug:beetle:: Please provide the steps to reproduce*
``` python
class DevelopmentStack(core.Stack):
def __init__(self, app: core.App, id: str, **kwargs) -> None:
super().__init__(app, id, **kwargs)
vpc = aws_ec2.Vpc(scope=self, id='Vpc', cidr='192.168.0.0./16')
eks_cluster = aws_eks.Cluster(scope=self, id='EksCluster', vpc=vpc, version='1.13')
eks_cluster.add_capacity(id='EksClusterOnDemandCapacity', 't2.large', desired_capacity=2)
app = core.App()
DevelopmentStack(
app=app,
id='Dev',
env={
'region': os.environ['CDK_DEFAULT_REGION'],
'account': os.environ['CDK_DEFAULT_ACCOUNT']
})
app.synth()
```
fails with: `jsii.errors.JSIIError: We don't have an AMI for kubernetes version 1.13`
* **What is the expected behavior (or behavior of feature suggested)?**
Template compiles.
* **What is the motivation / use case for changing the behavior or adding this feature?**
Cannot create EKS cluster with version 1.13
* **Please tell us about your environment:**
- CDK CLI Version: 0.36.0 (build 6d38487)
- Module Version: 0.36.0
- OS: [Ubuntu 19.4]
- Language: [Python]
* **Other information** (e.g. detailed explanation, stacktraces, related issues, suggestions how to fix, links for us to have context, eg. associated pull-request, stackoverflow, gitter, etc)
https://github.com/awslabs/aws-cdk/issues/3120
The template compiles only using autoscaling groups, that is via
`eks_cluster.add_auto_scaling_group(...)`
|
1.0
|
We don't have an AMI for kubernetes version 1.13 - **Note: for support questions, please first reference our [documentation](https://docs.aws.amazon.com/cdk/api/latest), then use [Stackoverflow](https://stackoverflow.com/questions/ask?tags=aws-cdk)**. This repository's issues are intended for feature requests and bug reports.
* **I'm submitting a ...**
- [x] :beetle: bug report
- [ ] :rocket: feature request
- [ ] :books: construct library gap
- [ ] :phone: security issue or vulnerability => Please see [policy](https://github.com/awslabs/aws-cdk/security/policy)
- [ ] :question: support request => Please see note at the top of this template.
* **What is the current behavior?**
*If the current behavior is a :beetle:bug:beetle:: Please provide the steps to reproduce*
``` python
class DevelopmentStack(core.Stack):
def __init__(self, app: core.App, id: str, **kwargs) -> None:
super().__init__(app, id, **kwargs)
vpc = aws_ec2.Vpc(scope=self, id='Vpc', cidr='192.168.0.0./16')
eks_cluster = aws_eks.Cluster(scope=self, id='EksCluster', vpc=vpc, version='1.13')
eks_cluster.add_capacity(id='EksClusterOnDemandCapacity', 't2.large', desired_capacity=2)
app = core.App()
DevelopmentStack(
app=app,
id='Dev',
env={
'region': os.environ['CDK_DEFAULT_REGION'],
'account': os.environ['CDK_DEFAULT_ACCOUNT']
})
app.synth()
```
fails with: `jsii.errors.JSIIError: We don't have an AMI for kubernetes version 1.13`
* **What is the expected behavior (or behavior of feature suggested)?**
Template compiles.
* **What is the motivation / use case for changing the behavior or adding this feature?**
Cannot create EKS cluster with version 1.13
* **Please tell us about your environment:**
- CDK CLI Version: 0.36.0 (build 6d38487)
- Module Version: 0.36.0
- OS: [Ubuntu 19.4]
- Language: [Python]
* **Other information** (e.g. detailed explanation, stacktraces, related issues, suggestions how to fix, links for us to have context, eg. associated pull-request, stackoverflow, gitter, etc)
https://github.com/awslabs/aws-cdk/issues/3120
The template compiles only using autoscaling groups, that is via
`eks_cluster.add_auto_scaling_group(...)`
|
non_process
|
we don t have an ami for kubernetes version note for support questions please first reference our then use this repository s issues are intended for feature requests and bug reports i m submitting a beetle bug report rocket feature request books construct library gap phone security issue or vulnerability please see question support request please see note at the top of this template what is the current behavior if the current behavior is a beetle bug beetle please provide the steps to reproduce python class developmentstack core stack def init self app core app id str kwargs none super init app id kwargs vpc aws vpc scope self id vpc cidr eks cluster aws eks cluster scope self id ekscluster vpc vpc version eks cluster add capacity id eksclusterondemandcapacity large desired capacity app core app developmentstack app app id dev env region os environ account os environ app synth fails with jsii errors jsiierror we don t have an ami for kubernetes version what is the expected behavior or behavior of feature suggested template compiles what is the motivation use case for changing the behavior or adding this feature cannot create eks cluster with version please tell us about your environment cdk cli version build module version os language other information e g detailed explanation stacktraces related issues suggestions how to fix links for us to have context eg associated pull request stackoverflow gitter etc the template compiles only using autoscaling groups that is via eks cluster add auto scaling group
| 0
|
10,098
| 13,044,162,089
|
IssuesEvent
|
2020-07-29 03:47:29
|
tikv/tikv
|
https://api.github.com/repos/tikv/tikv
|
closed
|
UCP: Migrate scalar function `AddStringAndString` from TiDB
|
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
|
## Description
Port the scalar function `AddStringAndString` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @andylokandy
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
2.0
|
UCP: Migrate scalar function `AddStringAndString` from TiDB -
## Description
Port the scalar function `AddStringAndString` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @andylokandy
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
process
|
ucp migrate scalar function addstringandstring from tidb description port the scalar function addstringandstring from tidb to coprocessor score mentor s andylokandy recommended skills rust programming learning materials already implemented expressions ported from tidb
| 1
|
65,181
| 19,214,358,363
|
IssuesEvent
|
2021-12-07 07:45:37
|
primefaces/primeng
|
https://api.github.com/repos/primefaces/primeng
|
closed
|
Calendar | Year Picker and Month Picker empty when value empty and using Range or Multiple Selection
|
defect
|
Similar to https://github.com/primefaces/primeng/issues/10906.
|
1.0
|
Calendar | Year Picker and Month Picker empty when value empty and using Range or Multiple Selection - Similar to https://github.com/primefaces/primeng/issues/10906.
|
non_process
|
calendar year picker and month picker empty when value empty and using range or multiple selection similar to
| 0
|
11,453
| 14,273,804,088
|
IssuesEvent
|
2020-11-21 23:43:18
|
carbon-design-system/ibm-cloud-cognitive
|
https://api.github.com/repos/carbon-design-system/ibm-cloud-cognitive
|
closed
|
Define CODEOWNERS and teams
|
type: process wontfix
|
## Summary
Identify and define CODEOWNERS and teams for design and development to automatically assign for PRs.
|
1.0
|
Define CODEOWNERS and teams - ## Summary
Identify and define CODEOWNERS and teams for design and development to automatically assign for PRs.
|
process
|
define codeowners and teams summary identify and define codeowners and teams for design and development to automatically assign for prs
| 1
|
11,132
| 13,957,691,537
|
IssuesEvent
|
2020-10-24 08:10:26
|
alexanderkotsev/geoportal
|
https://api.github.com/repos/alexanderkotsev/geoportal
|
opened
|
IE: INSPIRE National Discovery Service of Ireland not responding
|
Geoportal Harvesting process IE - Ireland
|
Dear Ben,
the registered National INSPIRE Discovery Service for Ireland has stopped responding:
https://www.geoportal.ie/geoportal/csw?SERVICE=CSW&VERSION=2.0.2&REQUEST=GetCapabilities
Best regards,
Angelo
|
1.0
|
IE: INSPIRE National Discovery Service of Ireland not responding - Dear Ben,
the registered National INSPIRE Discovery Service for Ireland has stopped responding:
https://www.geoportal.ie/geoportal/csw?SERVICE=CSW&VERSION=2.0.2&REQUEST=GetCapabilities
Best regards,
Angelo
|
process
|
ie inspire national discovery service of ireland not responding dear ben the registered national inspire discovery service for ireland has stopped responding best regards angelo
| 1
|
22,680
| 31,927,988,128
|
IssuesEvent
|
2023-09-19 04:30:05
|
ppy/osu-web
|
https://api.github.com/repos/ppy/osu-web
|
closed
|
Storyboarded maps can have icon not displayed on page due to incorrect file capitalisations.
|
area:beatmap-processing area:beatmap-info external component
|
This is a bit of a weird situation.
But as seen on this map: https://osu.ppy.sh/beatmapsets/1851103#osu/3802734
the storyboard icon and status is not placed even though a .osb file is found in the set.
The main issue is that the capitalisation of both the .osu and .osb files (although read correctly by stable) are not the same for the side of the website.

(For above example, the .osu file should be `twenty one pilots - Heathens (Magnetude Bootleg) (funny) [Marathon].osu` as is what's part of the correct metadata of said beatmap)
I have given a [problem report to the beatmap](https://osu.ppy.sh/beatmapsets/1851103/discussion/-/generalAll#/3437948) as it is currently in the Qualified stage, however, I do like to notify the issue on the website in case this little edge-case should be fixed.
|
1.0
|
Storyboarded maps can have icon not displayed on page due to incorrect file capitalisations. - This is a bit of a weird situation.
But as seen on this map: https://osu.ppy.sh/beatmapsets/1851103#osu/3802734
the storyboard icon and status is not placed even though a .osb file is found in the set.
The main issue is that the capitalisation of both the .osu and .osb files (although read correctly by stable) are not the same for the side of the website.

(For above example, the .osu file should be `twenty one pilots - Heathens (Magnetude Bootleg) (funny) [Marathon].osu` as is what's part of the correct metadata of said beatmap)
I have given a [problem report to the beatmap](https://osu.ppy.sh/beatmapsets/1851103/discussion/-/generalAll#/3437948) as it is currently in the Qualified stage, however, I do like to notify the issue on the website in case this little edge-case should be fixed.
|
process
|
storyboarded maps can have icon not displayed on page due to incorrect file capitalisations this is a bit of a weird situation but as seen on this map the storyboard icon and status is not placed even though a osb file is found in the set the main issue is that the capitalisation of both the osu and osb files although read correctly by stable are not the same for the side of the website for above example the osu file should be twenty one pilots heathens magnetude bootleg funny osu as is what s part of the correct metadata of said beatmap i have given a as it is currently in the qualified stage however i do like to notify the issue on the website in case this little edge case should be fixed
| 1
|
7,151
| 10,294,862,768
|
IssuesEvent
|
2019-08-27 20:47:13
|
Jeffail/benthos
|
https://api.github.com/repos/Jeffail/benthos
|
closed
|
XML Processor
|
enhancement processors
|
I think it's time we added an XML processor. This could be similar to the JSON processor we currently have where there are operators for different functions. However, I recommend starting simple and just having a `to_json` operator to start with.
|
1.0
|
XML Processor - I think it's time we added an XML processor. This could be similar to the JSON processor we currently have where there are operators for different functions. However, I recommend starting simple and just having a `to_json` operator to start with.
|
process
|
xml processor i think it s time we added an xml processor this could be similar to the json processor we currently have where there are operators for different functions however i recommend starting simple and just having a to json operator to start with
| 1
|
74,523
| 7,431,863,091
|
IssuesEvent
|
2018-03-25 18:51:16
|
smap-consulting/smapserver
|
https://api.github.com/repos/smap-consulting/smapserver
|
closed
|
BUG - 17.12 - Warning message "Question name not found" after uploading data
|
In Test Released bug
|
### SmapServer
### Version: 17.11
### Actual Behaviour: When you download data and upload again in Analysis there is a warning message about "question name missing":

### Steps to reproduce the problem: Upload attache file to form: Testusername
[TestUsername.zip](https://github.com/smap-consulting/smapserver/files/1575146/TestUsername.zip)
### Server: Kontrolid
### Organisation: Zarkman
### Project: TestAlex
### Form Name: Testusername
|
1.0
|
BUG - 17.12 - Warning message "Question name not found" after uploading data - ### SmapServer
### Version: 17.11
### Actual Behaviour: When you download data and upload again in Analysis there is a warning message about "question name missing":

### Steps to reproduce the problem: Upload attache file to form: Testusername
[TestUsername.zip](https://github.com/smap-consulting/smapserver/files/1575146/TestUsername.zip)
### Server: Kontrolid
### Organisation: Zarkman
### Project: TestAlex
### Form Name: Testusername
|
non_process
|
bug warning message question name not found after uploading data smapserver version actual behaviour when you download data and upload again in analysis there is a warning message about question name missing steps to reproduce the problem upload attache file to form testusername server kontrolid organisation zarkman project testalex form name testusername
| 0
|
17,033
| 22,409,044,416
|
IssuesEvent
|
2022-06-18 12:37:54
|
cgi-js/cgi-js
|
https://api.github.com/repos/cgi-js/cgi-js
|
closed
|
[REQUEST] Create templates for bugs/issues, features, and general requests
|
process
|
**Please describe your request.**
Create templates for bugs/issues, features, and general requests
**Describe the solution you'd like**
Templates for bugs/issues, features, and general requests
**Describe alternatives you've considered**
NA
**Additional context**
NA
|
1.0
|
[REQUEST] Create templates for bugs/issues, features, and general requests - **Please describe your request.**
Create templates for bugs/issues, features, and general requests
**Describe the solution you'd like**
Templates for bugs/issues, features, and general requests
**Describe alternatives you've considered**
NA
**Additional context**
NA
|
process
|
create templates for bugs issues features and general requests please describe your request create templates for bugs issues features and general requests describe the solution you d like templates for bugs issues features and general requests describe alternatives you ve considered na additional context na
| 1
|
1,826
| 4,613,581,221
|
IssuesEvent
|
2016-09-25 03:20:29
|
EBrown8534/StackExchangeStatisticsExplorer
|
https://api.github.com/repos/EBrown8534/StackExchangeStatisticsExplorer
|
opened
|
Export `/Sites/Compare/` results to a CSV
|
enhancement in process
|
Another self-explanatory issue, add a button to export `/Sites/Compare/` charts to CSV files and send to the user.
|
1.0
|
Export `/Sites/Compare/` results to a CSV - Another self-explanatory issue, add a button to export `/Sites/Compare/` charts to CSV files and send to the user.
|
process
|
export sites compare results to a csv another self explanatory issue add a button to export sites compare charts to csv files and send to the user
| 1
|
3,246
| 6,312,660,461
|
IssuesEvent
|
2017-07-24 04:52:29
|
gaocegege/Processing.R
|
https://api.github.com/repos/gaocegege/Processing.R
|
opened
|
Add more examples
|
community/processing difficulty/medium priority/p1 size/no-idea status/to-be-claimed type/enhancement
|
* R package examples I think are important for the same reason that me working through importLibrary Java examples was important
* We need some fancy, short examples to show the capability of P.R.
|
1.0
|
Add more examples - * R package examples I think are important for the same reason that me working through importLibrary Java examples was important
* We need some fancy, short examples to show the capability of P.R.
|
process
|
add more examples r package examples i think are important for the same reason that me working through importlibrary java examples was important we need some fancy short examples to show the capability of p r
| 1
|
42,517
| 2,870,950,466
|
IssuesEvent
|
2015-06-07 17:51:07
|
Naoghuman/Dream-Better-Worlds
|
https://api.github.com/repos/Naoghuman/Dream-Better-Worlds
|
closed
|
Remove in the menu Help the entry Help.
|
development refactoring issue fixed priority low
|
Remove in the menu Help the entry Help. Comment out only.
|
1.0
|
Remove in the menu Help the entry Help. - Remove in the menu Help the entry Help. Comment out only.
|
non_process
|
remove in the menu help the entry help remove in the menu help the entry help comment out only
| 0
|
7,035
| 10,196,634,055
|
IssuesEvent
|
2019-08-12 21:17:28
|
dotnet/coreclr
|
https://api.github.com/repos/dotnet/coreclr
|
closed
|
[Linux][Arm] assert currentProcessCpuCount == g_processAffinitySet.Count() in GC/API/GC/KeepAliveNull/KeepAliveNull.sh
|
area-PAL area-System.Diagnostics.Process ci-health
|
```
GC/API/GC/KeepAliveNull/KeepAliveNull.sh [FAIL]
Assert failure(PID 24052 [0x00005df4], Thread: 24052 [0x5df4]): currentProcessCpuCount == g_processAffinitySet.Count()
File: /__w/2/s/src/vm/gcenv.os.cpp Line: 114
Image: /ssd/dotnetbuild/work/66290456-48d2-448c-ae61-9d5ab09b9d72/Payload/corerun
/ssd/dotnetbuild/work/66290456-48d2-448c-ae61-9d5ab09b9d72/Work/ea6de819-2a26-464b-aba9-057e7137059f/Exec/GC/API/GC/KeepAliveNull/KeepAliveNull.sh: line 252: 24052 Aborted $LAUNCHER $ExePath "${CLRTestExecutionArguments[@]}"
Return code: 1
Raw output file: /ssd/dotnetbuild/work/66290456-48d2-448c-ae61-9d5ab09b9d72/Work/ea6de819-2a26-464b-aba9-057e7137059f/Exec/GC/API/Reports/GC.API/GC/KeepAliveNull/KeepAliveNull.output.txt
Raw output:
BEGIN EXECUTION
/ssd/dotnetbuild/work/66290456-48d2-448c-ae61-9d5ab09b9d72/Payload/corerun KeepAliveNull.exe ''
Expected: 100
Actual: 134
END EXECUTION - FAILED
Test Harness Exitcode is : 1
To run the test:
> set CORE_ROOT=/ssd/dotnetbuild/work/66290456-48d2-448c-ae61-9d5ab09b9d72/Payload
> /ssd/dotnetbuild/work/66290456-48d2-448c-ae61-9d5ab09b9d72/Work/ea6de819-2a26-464b-aba9-057e7137059f/Exec/GC/API/GC/KeepAliveNull/KeepAliveNull.sh
Expected: True
Actual: False
Stack Trace:
/__w/2/s/bin/tests/Linux.arm.Checked/TestWrappers/GC.API/GC.API.XUnitWrapper.cs(2856,0): at GC_API._GC_KeepAliveNull_KeepAliveNull_._GC_KeepAliveNull_KeepAliveNull_sh()
Output:
Assert failure(PID 24052 [0x00005df4], Thread: 24052 [0x5df4]): currentProcessCpuCount == g_processAffinitySet.Count()
File: /__w/2/s/src/vm/gcenv.os.cpp Line: 114
Image: /ssd/dotnetbuild/work/66290456-48d2-448c-ae61-9d5ab09b9d72/Payload/corerun
/ssd/dotnetbuild/work/66290456-48d2-448c-ae61-9d5ab09b9d72/Work/ea6de819-2a26-464b-aba9-057e7137059f/Exec/GC/API/GC/KeepAliveNull/KeepAliveNull.sh: line 252: 24052 Aborted $LAUNCHER $ExePath "${CLRTestExecutionArguments[@]}"
Return code: 1
Raw output file: /ssd/dotnetbuild/work/66290456-48d2-448c-ae61-9d5ab09b9d72/Work/ea6de819-2a26-464b-aba9-057e7137059f/Exec/GC/API/Reports/GC.API/GC/KeepAliveNull/KeepAliveNull.output.txt
Raw output:
BEGIN EXECUTION
/ssd/dotnetbuild/work/66290456-48d2-448c-ae61-9d5ab09b9d72/Payload/corerun KeepAliveNull.exe ''
Expected: 100
Actual: 134
END EXECUTION - FAILED
Test Harness Exitcode is : 1
To run the test:
> set CORE_ROOT=/ssd/dotnetbuild/work/66290456-48d2-448c-ae61-9d5ab09b9d72/Payload
> /ssd/dotnetbuild/work/66290456-48d2-448c-ae61-9d5ab09b9d72/Work/ea6de819-2a26-464b-aba9-057e7137059f/Exec/GC/API/GC/KeepAliveNull/KeepAliveNull.sh
```
|
1.0
|
[Linux][Arm] assert currentProcessCpuCount == g_processAffinitySet.Count() in GC/API/GC/KeepAliveNull/KeepAliveNull.sh - ```
GC/API/GC/KeepAliveNull/KeepAliveNull.sh [FAIL]
Assert failure(PID 24052 [0x00005df4], Thread: 24052 [0x5df4]): currentProcessCpuCount == g_processAffinitySet.Count()
File: /__w/2/s/src/vm/gcenv.os.cpp Line: 114
Image: /ssd/dotnetbuild/work/66290456-48d2-448c-ae61-9d5ab09b9d72/Payload/corerun
/ssd/dotnetbuild/work/66290456-48d2-448c-ae61-9d5ab09b9d72/Work/ea6de819-2a26-464b-aba9-057e7137059f/Exec/GC/API/GC/KeepAliveNull/KeepAliveNull.sh: line 252: 24052 Aborted $LAUNCHER $ExePath "${CLRTestExecutionArguments[@]}"
Return code: 1
Raw output file: /ssd/dotnetbuild/work/66290456-48d2-448c-ae61-9d5ab09b9d72/Work/ea6de819-2a26-464b-aba9-057e7137059f/Exec/GC/API/Reports/GC.API/GC/KeepAliveNull/KeepAliveNull.output.txt
Raw output:
BEGIN EXECUTION
/ssd/dotnetbuild/work/66290456-48d2-448c-ae61-9d5ab09b9d72/Payload/corerun KeepAliveNull.exe ''
Expected: 100
Actual: 134
END EXECUTION - FAILED
Test Harness Exitcode is : 1
To run the test:
> set CORE_ROOT=/ssd/dotnetbuild/work/66290456-48d2-448c-ae61-9d5ab09b9d72/Payload
> /ssd/dotnetbuild/work/66290456-48d2-448c-ae61-9d5ab09b9d72/Work/ea6de819-2a26-464b-aba9-057e7137059f/Exec/GC/API/GC/KeepAliveNull/KeepAliveNull.sh
Expected: True
Actual: False
Stack Trace:
/__w/2/s/bin/tests/Linux.arm.Checked/TestWrappers/GC.API/GC.API.XUnitWrapper.cs(2856,0): at GC_API._GC_KeepAliveNull_KeepAliveNull_._GC_KeepAliveNull_KeepAliveNull_sh()
Output:
Assert failure(PID 24052 [0x00005df4], Thread: 24052 [0x5df4]): currentProcessCpuCount == g_processAffinitySet.Count()
File: /__w/2/s/src/vm/gcenv.os.cpp Line: 114
Image: /ssd/dotnetbuild/work/66290456-48d2-448c-ae61-9d5ab09b9d72/Payload/corerun
/ssd/dotnetbuild/work/66290456-48d2-448c-ae61-9d5ab09b9d72/Work/ea6de819-2a26-464b-aba9-057e7137059f/Exec/GC/API/GC/KeepAliveNull/KeepAliveNull.sh: line 252: 24052 Aborted $LAUNCHER $ExePath "${CLRTestExecutionArguments[@]}"
Return code: 1
Raw output file: /ssd/dotnetbuild/work/66290456-48d2-448c-ae61-9d5ab09b9d72/Work/ea6de819-2a26-464b-aba9-057e7137059f/Exec/GC/API/Reports/GC.API/GC/KeepAliveNull/KeepAliveNull.output.txt
Raw output:
BEGIN EXECUTION
/ssd/dotnetbuild/work/66290456-48d2-448c-ae61-9d5ab09b9d72/Payload/corerun KeepAliveNull.exe ''
Expected: 100
Actual: 134
END EXECUTION - FAILED
Test Harness Exitcode is : 1
To run the test:
> set CORE_ROOT=/ssd/dotnetbuild/work/66290456-48d2-448c-ae61-9d5ab09b9d72/Payload
> /ssd/dotnetbuild/work/66290456-48d2-448c-ae61-9d5ab09b9d72/Work/ea6de819-2a26-464b-aba9-057e7137059f/Exec/GC/API/GC/KeepAliveNull/KeepAliveNull.sh
```
|
process
|
assert currentprocesscpucount g processaffinityset count in gc api gc keepalivenull keepalivenull sh gc api gc keepalivenull keepalivenull sh assert failure pid thread currentprocesscpucount g processaffinityset count file w s src vm gcenv os cpp line image ssd dotnetbuild work payload corerun ssd dotnetbuild work work exec gc api gc keepalivenull keepalivenull sh line aborted launcher exepath clrtestexecutionarguments return code raw output file ssd dotnetbuild work work exec gc api reports gc api gc keepalivenull keepalivenull output txt raw output begin execution ssd dotnetbuild work payload corerun keepalivenull exe expected actual end execution failed test harness exitcode is to run the test set core root ssd dotnetbuild work payload ssd dotnetbuild work work exec gc api gc keepalivenull keepalivenull sh expected true actual false stack trace w s bin tests linux arm checked testwrappers gc api gc api xunitwrapper cs at gc api gc keepalivenull keepalivenull gc keepalivenull keepalivenull sh output assert failure pid thread currentprocesscpucount g processaffinityset count file w s src vm gcenv os cpp line image ssd dotnetbuild work payload corerun ssd dotnetbuild work work exec gc api gc keepalivenull keepalivenull sh line aborted launcher exepath clrtestexecutionarguments return code raw output file ssd dotnetbuild work work exec gc api reports gc api gc keepalivenull keepalivenull output txt raw output begin execution ssd dotnetbuild work payload corerun keepalivenull exe expected actual end execution failed test harness exitcode is to run the test set core root ssd dotnetbuild work payload ssd dotnetbuild work work exec gc api gc keepalivenull keepalivenull sh
| 1
|
309,590
| 26,669,524,147
|
IssuesEvent
|
2023-01-26 09:01:26
|
mautic/mautic
|
https://api.github.com/repos/mautic/mautic
|
closed
|
Segment filters UI breaks on multiple edits
|
bug ready-to-test segments stale
|
### Mautic Version
4.4.x series
### PHP version
7.4
### What browsers are you seeing the problem on?
Firefox, Chrome
### What happened?
Changing filters of segments will be lost in case there is an error in one of the filters.
### How can we reproduce this issue?
This video shows the story behind the issue
https://www.loom.com/share/5207f6146c0f4721b9a2b1ae9b2eaaf8
Steps:
1. Create filters
2. Edit them on the same session
3. Have an error in one of the filters
4. Your changes are lost
### Relevant log output
_No response_
### Code of Conduct
- [X] I confirm that I have read and agree to follow this project's Code of Conduct
|
1.0
|
Segment filters UI breaks on multiple edits - ### Mautic Version
4.4.x series
### PHP version
7.4
### What browsers are you seeing the problem on?
Firefox, Chrome
### What happened?
Changing filters of segments will be lost in case there is an error in one of the filters.
### How can we reproduce this issue?
This video shows the story behind the issue
https://www.loom.com/share/5207f6146c0f4721b9a2b1ae9b2eaaf8
Steps:
1. Create filters
2. Edit them on the same session
3. Have an error in one of the filters
4. Your changes are lost
### Relevant log output
_No response_
### Code of Conduct
- [X] I confirm that I have read and agree to follow this project's Code of Conduct
|
non_process
|
segment filters ui breaks on multiple edits mautic version x series php version what browsers are you seeing the problem on firefox chrome what happened changing filters of segments will be lost in case there is an error in one of the filters how can we reproduce this issue this video shows the story behind the issue steps create filters edit them on the same session have an error in one of the filters your changes are lost relevant log output no response code of conduct i confirm that i have read and agree to follow this project s code of conduct
| 0
|
4,353
| 7,260,011,875
|
IssuesEvent
|
2018-02-18 03:15:54
|
muflihun/residue
|
https://api.github.com/repos/muflihun/residue
|
closed
|
Crash noticed during load testing
|
area: log-processing type: bug
|
```
22:30:51,352 [LogDispatcher] [FATAL] CRASH HANDLED; Application has crashed due to [SIGSEGV] signal
======= Backtrace: =========
[1] 2 residue 0x000000010c5c690b _Z23generalTerminateHandlerib + 267
[2] 3 libsystem_platform.dylib 0x00007fff6865ef5a _sigtramp + 26
[3] 4 ??? 0x0000000000000002 0x0 + 2
[4] 5 residue 0x000000010c66d08d _ZNSt3__16__treeINS_12__value_typeINS_12basic_stringIcNS_11char_traitsIcEENS_9allocatorIcEEEEN8nlohmann10basic_jsonINS_3mapENS_6vectorES7_bxydS5_NS8_14adl_serializerEEEEENS_19__map_value_compareIS7_SE_NS_4lessIS7_EELb1EEENS5_ISE_EEE7destroyEPNS_11__tree_nodeISE_PvEE + 29
[5] 6 residue 0x000000010c66d08d _ZNSt3__16__treeINS_12__value_typeINS_12basic_stringIcNS_11char_traitsIcEENS_9allocatorIcEEEEN8nlohmann10basic_jsonINS_3mapENS_6vectorES7_bxydS5_NS8_14adl_serializerEEEEENS_19__map_value_compareIS7_SE_NS_4lessIS7_EELb1EEENS5_ISE_EEE7destroyEPNS_11__tree_nodeISE_PvEE + 29
[6] 7 residue 0x000000010c66d08d _ZNSt3__16__treeINS_12__value_typeINS_12basic_stringIcNS_11char_traitsIcEENS_9allocatorIcEEEEN8nlohmann10basic_jsonINS_3mapENS_6vectorES7_bxydS5_NS8_14adl_serializerEEEEENS_19__map_value_compareIS7_SE_NS_4lessIS7_EELb1EEENS5_ISE_EEE7destroyEPNS_11__tree_nodeISE_PvEE + 29
[7] 8 residue 0x000000010c66d099 _ZNSt3__16__treeINS_12__value_typeINS_12basic_stringIcNS_11char_traitsIcEENS_9allocatorIcEEEEN8nlohmann10basic_jsonINS_3mapENS_6vectorES7_bxydS5_NS8_14adl_serializerEEEEENS_19__map_value_compareIS7_SE_NS_4lessIS7_EELb1EEENS5_ISE_EEE7destroyEPNS_11__tree_nodeISE_PvEE + 41
[8] 9 residue 0x000000010c66d099 _ZNSt3__16__treeINS_12__value_typeINS_12basic_stringIcNS_11char_traitsIcEENS_9allocatorIcEEEEN8nlohmann10basic_jsonINS_3mapENS_6vectorES7_bxydS5_NS8_14adl_serializerEEEEENS_19__map_value_compareIS7_SE_NS_4lessIS7_EELb1EEENS5_ISE_EEE7destroyEPNS_11__tree_nodeISE_PvEE + 41
[9] 10 residue 0x000000010c66d08d _ZNSt3__16__treeINS_12__value_typeINS_12basic_stringIcNS_11char_traitsIcEENS_9allocatorIcEEEEN8nlohmann10basic_jsonINS_3mapENS_6vectorES7_bxydS5_NS8_14adl_serializerEEEEENS_19__map_value_compareIS7_SE_NS_4lessIS7_EELb1EEENS5_ISE_EEE7destroyEPNS_11__tree_nodeISE_PvEE + 29
[10] 11 residue 0x000000010c66cf78 _ZN8nlohmann10basic_jsonINSt3__13mapENS1_6vectorENS1_12basic_stringIcNS1_11char_traitsIcEENS1_9allocatorIcEEEEbxydS7_NS_14adl_serializerEED2Ev + 56
[11] 12 residue 0x000000010c66cfad _ZN8nlohmann10basic_jsonINSt3__13mapENS1_6vectorENS1_12basic_stringIcNS1_11char_traitsIcEENS1_9allocatorIcEEEEbxydS7_NS_14adl_serializerEED2Ev + 109
[12] 13 residue 0x000000010c61d833 _ZN7residue7RequestD2Ev + 83
[13] 14 residue 0x000000010c619581 _ZN7residue17LogRequestHandler19processRequestQueueEv + 4449
[14] 15 residue 0x000000010c61d8e8 _ZNSt3__114__thread_proxyINS_5tupleIJNS_10unique_ptrINS_15__thread_structENS_14default_deleteIS3_EEEEZN7residue17LogRequestHandler5startEvE3$_0EEEEEPvSB_ + 152
[15] 16 libsystem_pthread.dylib 0x00007fff686686c1 _pthread_body + 340
[16] 17 libsystem_pthread.dylib 0x00007fff6866856d _pthread_body + 0
[17] 18 libsystem_pthread.dylib 0x00007fff68667c5d thread_start + 13
```
|
1.0
|
Crash noticed during load testing - ```
22:30:51,352 [LogDispatcher] [FATAL] CRASH HANDLED; Application has crashed due to [SIGSEGV] signal
======= Backtrace: =========
[1] 2 residue 0x000000010c5c690b _Z23generalTerminateHandlerib + 267
[2] 3 libsystem_platform.dylib 0x00007fff6865ef5a _sigtramp + 26
[3] 4 ??? 0x0000000000000002 0x0 + 2
[4] 5 residue 0x000000010c66d08d _ZNSt3__16__treeINS_12__value_typeINS_12basic_stringIcNS_11char_traitsIcEENS_9allocatorIcEEEEN8nlohmann10basic_jsonINS_3mapENS_6vectorES7_bxydS5_NS8_14adl_serializerEEEEENS_19__map_value_compareIS7_SE_NS_4lessIS7_EELb1EEENS5_ISE_EEE7destroyEPNS_11__tree_nodeISE_PvEE + 29
[5] 6 residue 0x000000010c66d08d _ZNSt3__16__treeINS_12__value_typeINS_12basic_stringIcNS_11char_traitsIcEENS_9allocatorIcEEEEN8nlohmann10basic_jsonINS_3mapENS_6vectorES7_bxydS5_NS8_14adl_serializerEEEEENS_19__map_value_compareIS7_SE_NS_4lessIS7_EELb1EEENS5_ISE_EEE7destroyEPNS_11__tree_nodeISE_PvEE + 29
[6] 7 residue 0x000000010c66d08d _ZNSt3__16__treeINS_12__value_typeINS_12basic_stringIcNS_11char_traitsIcEENS_9allocatorIcEEEEN8nlohmann10basic_jsonINS_3mapENS_6vectorES7_bxydS5_NS8_14adl_serializerEEEEENS_19__map_value_compareIS7_SE_NS_4lessIS7_EELb1EEENS5_ISE_EEE7destroyEPNS_11__tree_nodeISE_PvEE + 29
[7] 8 residue 0x000000010c66d099 _ZNSt3__16__treeINS_12__value_typeINS_12basic_stringIcNS_11char_traitsIcEENS_9allocatorIcEEEEN8nlohmann10basic_jsonINS_3mapENS_6vectorES7_bxydS5_NS8_14adl_serializerEEEEENS_19__map_value_compareIS7_SE_NS_4lessIS7_EELb1EEENS5_ISE_EEE7destroyEPNS_11__tree_nodeISE_PvEE + 41
[8] 9 residue 0x000000010c66d099 _ZNSt3__16__treeINS_12__value_typeINS_12basic_stringIcNS_11char_traitsIcEENS_9allocatorIcEEEEN8nlohmann10basic_jsonINS_3mapENS_6vectorES7_bxydS5_NS8_14adl_serializerEEEEENS_19__map_value_compareIS7_SE_NS_4lessIS7_EELb1EEENS5_ISE_EEE7destroyEPNS_11__tree_nodeISE_PvEE + 41
[9] 10 residue 0x000000010c66d08d _ZNSt3__16__treeINS_12__value_typeINS_12basic_stringIcNS_11char_traitsIcEENS_9allocatorIcEEEEN8nlohmann10basic_jsonINS_3mapENS_6vectorES7_bxydS5_NS8_14adl_serializerEEEEENS_19__map_value_compareIS7_SE_NS_4lessIS7_EELb1EEENS5_ISE_EEE7destroyEPNS_11__tree_nodeISE_PvEE + 29
[10] 11 residue 0x000000010c66cf78 _ZN8nlohmann10basic_jsonINSt3__13mapENS1_6vectorENS1_12basic_stringIcNS1_11char_traitsIcEENS1_9allocatorIcEEEEbxydS7_NS_14adl_serializerEED2Ev + 56
[11] 12 residue 0x000000010c66cfad _ZN8nlohmann10basic_jsonINSt3__13mapENS1_6vectorENS1_12basic_stringIcNS1_11char_traitsIcEENS1_9allocatorIcEEEEbxydS7_NS_14adl_serializerEED2Ev + 109
[12] 13 residue 0x000000010c61d833 _ZN7residue7RequestD2Ev + 83
[13] 14 residue 0x000000010c619581 _ZN7residue17LogRequestHandler19processRequestQueueEv + 4449
[14] 15 residue 0x000000010c61d8e8 _ZNSt3__114__thread_proxyINS_5tupleIJNS_10unique_ptrINS_15__thread_structENS_14default_deleteIS3_EEEEZN7residue17LogRequestHandler5startEvE3$_0EEEEEPvSB_ + 152
[15] 16 libsystem_pthread.dylib 0x00007fff686686c1 _pthread_body + 340
[16] 17 libsystem_pthread.dylib 0x00007fff6866856d _pthread_body + 0
[17] 18 libsystem_pthread.dylib 0x00007fff68667c5d thread_start + 13
```
|
process
|
crash noticed during load testing crash handled application has crashed due to signal backtrace residue libsystem platform dylib sigtramp residue treeins value typeins stringicns traitsiceens jsonins serializereeeeens map value se ns ise tree nodeise pvee residue treeins value typeins stringicns traitsiceens jsonins serializereeeeens map value se ns ise tree nodeise pvee residue treeins value typeins stringicns traitsiceens jsonins serializereeeeens map value se ns ise tree nodeise pvee residue treeins value typeins stringicns traitsiceens jsonins serializereeeeens map value se ns ise tree nodeise pvee residue treeins value typeins stringicns traitsiceens jsonins serializereeeeens map value se ns ise tree nodeise pvee residue treeins value typeins stringicns traitsiceens jsonins serializereeeeens map value se ns ise tree nodeise pvee residue ns residue ns residue residue residue thread proxyins ptrins thread structens libsystem pthread dylib pthread body libsystem pthread dylib pthread body libsystem pthread dylib thread start
| 1
|
15,574
| 19,703,507,047
|
IssuesEvent
|
2022-01-12 19:08:13
|
googleapis/nodejs-dms
|
https://api.github.com/repos/googleapis/nodejs-dms
|
opened
|
Your .repo-metadata.json file has a problem 🤒
|
type: process repo-metadata: lint
|
You have a problem with your .repo-metadata.json file:
Result of scan 📈:
* api_shortname 'dms' invalid in .repo-metadata.json
☝️ Once you correct these problems, you can close this issue.
Reach out to **go/github-automation** if you have any questions.
|
1.0
|
Your .repo-metadata.json file has a problem 🤒 - You have a problem with your .repo-metadata.json file:
Result of scan 📈:
* api_shortname 'dms' invalid in .repo-metadata.json
☝️ Once you correct these problems, you can close this issue.
Reach out to **go/github-automation** if you have any questions.
|
process
|
your repo metadata json file has a problem 🤒 you have a problem with your repo metadata json file result of scan 📈 api shortname dms invalid in repo metadata json ☝️ once you correct these problems you can close this issue reach out to go github automation if you have any questions
| 1
|
262,314
| 19,782,726,209
|
IssuesEvent
|
2022-01-18 00:06:30
|
Qwaxi-Lixard/yes-porn-allowed
|
https://api.github.com/repos/Qwaxi-Lixard/yes-porn-allowed
|
closed
|
Add [SITE] to Yes Porn Allow list
|
documentation
|
**What is the site you want to add?**
https://autonomic.zone/
**What kind of service are they?**
Web hosting, development, cloud storage and group chat hosting.
**How does the site tend to treat NSFW content?**
They are totally down with it. They gladly hosted my website and at least a few others.
**Do you know of any risk to the site?**
Nope.
**Any comment you want to add about the site?**
Autonomic are a UK-based co-operative. That means that they have no bosses, and all surplus money goes to the workers who get a fair wage. They are personal friends of mine, and create web resources for the industrial workers of the world, the union I organise with.
They're very supportive, and helpful for folks who are new to the whole webhosting thing. They also love a good chat.
|
1.0
|
Add [SITE] to Yes Porn Allow list - **What is the site you want to add?**
https://autonomic.zone/
**What kind of service are they?**
Web hosting, development, cloud storage and group chat hosting.
**How does the site tend to treat NSFW content?**
They are totally down with it. They gladly hosted my website and at least a few others.
**Do you know of any risk to the site?**
Nope.
**Any comment you want to add about the site?**
Autonomic are a UK-based co-operative. That means that they have no bosses, and all surplus money goes to the workers who get a fair wage. They are personal friends of mine, and create web resources for the industrial workers of the world, the union I organise with.
They're very supportive, and helpful for folks who are new to the whole webhosting thing. They also love a good chat.
|
non_process
|
add to yes porn allow list what is the site you want to add what kind of service are they web hosting development cloud storage and group chat hosting how does the site tend to treat nsfw content they are totally down with it they gladly hosted my website and at least a few others do you know of any risk to the site nope any comment you want to add about the site autonomic are a uk based co operative that means that they have no bosses and all surplus money goes to the workers who get a fair wage they are personal friends of mine and create web resources for the industrial workers of the world the union i organise with they re very supportive and helpful for folks who are new to the whole webhosting thing they also love a good chat
| 0
|
294
| 5,486,983,403
|
IssuesEvent
|
2017-03-14 02:17:08
|
alexandream/nuvm
|
https://api.github.com/repos/alexandream/nuvm
|
opened
|
Create a viable alternative for stdint.h
|
portability
|
For now we're using the host machine's `stdint.h` header for some stuff. In fact, we're even using an optional field of `stdint.h` as if it were always there (`intptr_t`).
We must have viable alternative implementations (in `common/compatibility/stdint.h`) for the pieces we need to make sure we can use those implementations when hosts do not provide them.
|
True
|
Create a viable alternative for stdint.h - For now we're using the host machine's `stdint.h` header for some stuff. In fact, we're even using an optional field of `stdint.h` as if it were always there (`intptr_t`).
We must have viable alternative implementations (in `common/compatibility/stdint.h`) for the pieces we need to make sure we can use those implementations when hosts do not provide them.
|
non_process
|
create a viable alternative for stdint h for now we re using the host machine s stdint h header for some stuff in fact we re even using an optional field of stdint h as if it were always there intptr t we must have viable alternative implementations in common compatibility stdint h for the pieces we need to make sure we can use those implementations when hosts do not provide them
| 0
|
11,192
| 13,957,699,960
|
IssuesEvent
|
2020-10-24 08:12:51
|
alexanderkotsev/geoportal
|
https://api.github.com/repos/alexanderkotsev/geoportal
|
opened
|
MT: Harvest
|
Geoportal Harvesting process MT - Malta
|
Dear Angelo,
Kindly can you perform a harvest on the Maltese CSW at your convenience as we need to check the outcome of some changes which were recently implemented.
Regards,
Rene
|
1.0
|
MT: Harvest - Dear Angelo,
Kindly can you perform a harvest on the Maltese CSW at your convenience as we need to check the outcome of some changes which were recently implemented.
Regards,
Rene
|
process
|
mt harvest dear angelo kindly can you perform a harvest on the maltese csw at your convenience as we need to check the outcome of some changes which were recently implemented regards rene
| 1
|
9,472
| 12,466,958,293
|
IssuesEvent
|
2020-05-28 16:16:42
|
varys-main/ps-tools
|
https://api.github.com/repos/varys-main/ps-tools
|
closed
|
Docker - Add-NavDockerContainer - includeCSide
|
processing
|
# User Story
- In der Funktion Add-NavDockerContainer soll es möglich sein festzulegen, ob CSide für den neuen Container aktiviert werden soll.
# Tasks
- [x] Abfrage, ob CSide aktiviert werden soll
- [x] SWITCH-Parameter in interne NEW-Funktion
- [x] Übergabe an navcontainerhelper
# Implementations
- Es erfolgt ein Abfrage, ob CSide aktiviert werden soll
# Known Problems
|
1.0
|
Docker - Add-NavDockerContainer - includeCSide - # User Story
- In der Funktion Add-NavDockerContainer soll es möglich sein festzulegen, ob CSide für den neuen Container aktiviert werden soll.
# Tasks
- [x] Abfrage, ob CSide aktiviert werden soll
- [x] SWITCH-Parameter in interne NEW-Funktion
- [x] Übergabe an navcontainerhelper
# Implementations
- Es erfolgt ein Abfrage, ob CSide aktiviert werden soll
# Known Problems
|
process
|
docker add navdockercontainer includecside user story in der funktion add navdockercontainer soll es möglich sein festzulegen ob cside für den neuen container aktiviert werden soll tasks abfrage ob cside aktiviert werden soll switch parameter in interne new funktion übergabe an navcontainerhelper implementations es erfolgt ein abfrage ob cside aktiviert werden soll known problems
| 1
|
8,939
| 12,055,317,088
|
IssuesEvent
|
2020-04-15 12:47:57
|
ArneSch/NF_Prj_MIMII_Dataset
|
https://api.github.com/repos/ArneSch/NF_Prj_MIMII_Dataset
|
closed
|
Sub-notebook request: Audio file loading
|
preprocessing utility
|
- Load all Audio files similar to the baseline model and store them within pickle files.
- Change pickle directory to ".\dataset\pickle"
- Store the pickle files within the dataset-folder to access it easier from different sub_notebooks
|
1.0
|
Sub-notebook request: Audio file loading - - Load all Audio files similar to the baseline model and store them within pickle files.
- Change pickle directory to ".\dataset\pickle"
- Store the pickle files within the dataset-folder to access it easier from different sub_notebooks
|
process
|
sub notebook request audio file loading load all audio files similar to the baseline model and store them within pickle files change pickle directory to dataset pickle store the pickle files within the dataset folder to access it easier from different sub notebooks
| 1
|
90,034
| 25,953,824,396
|
IssuesEvent
|
2022-12-18 00:14:37
|
QubesOS/updates-status
|
https://api.github.com/repos/QubesOS/updates-status
|
closed
|
app-linux-pdf-converter v2.1.14 (r4.2)
|
r4.2-vm-bullseye-building r4.2-host-cur-test r4.2-vm-bookworm-cur-test r4.2-vm-fc37-cur-test r4.2-vm-fc36-cur-test r4.2-vm-centos-stream8-failed
|
Update of app-linux-pdf-converter to v2.1.14 for Qubes r4.2, see comments below for details and build status.
From commit: https://github.com/QubesOS/qubes-app-linux-pdf-converter/commit/1a770cf330cecf783f76194b117fda721564c0b4
[Changes since previous version](https://github.com/QubesOS/qubes-app-linux-pdf-converter/compare/v2.1.13...v2.1.14):
QubesOS/qubes-app-linux-pdf-converter@1a770cf version 2.1.14
QubesOS/qubes-app-linux-pdf-converter@ebb4460 qvm_convert_pdf_nautilus.py: adjust to work with Nautilus API 4.0
Referenced issues:
QubesOS/qubes-issues#7916
If you're release manager, you can issue GPG-inline signed command:
* `Upload-component r4.2 app-linux-pdf-converter 1a770cf330cecf783f76194b117fda721564c0b4 current all` (available 5 days from now)
* `Upload-component r4.2 app-linux-pdf-converter 1a770cf330cecf783f76194b117fda721564c0b4 security-testing`
You can choose subset of distributions like:
* `Upload-component r4.2 app-linux-pdf-converter 1a770cf330cecf783f76194b117fda721564c0b4 current vm-bookworm,vm-fc37` (available 5 days from now)
Above commands will work only if packages in current-testing repository were built from given commit (i.e. no new version superseded it).
For more information on how to test this update, please take a look at https://www.qubes-os.org/doc/testing/#updates.
|
1.0
|
app-linux-pdf-converter v2.1.14 (r4.2) - Update of app-linux-pdf-converter to v2.1.14 for Qubes r4.2, see comments below for details and build status.
From commit: https://github.com/QubesOS/qubes-app-linux-pdf-converter/commit/1a770cf330cecf783f76194b117fda721564c0b4
[Changes since previous version](https://github.com/QubesOS/qubes-app-linux-pdf-converter/compare/v2.1.13...v2.1.14):
QubesOS/qubes-app-linux-pdf-converter@1a770cf version 2.1.14
QubesOS/qubes-app-linux-pdf-converter@ebb4460 qvm_convert_pdf_nautilus.py: adjust to work with Nautilus API 4.0
Referenced issues:
QubesOS/qubes-issues#7916
If you're release manager, you can issue GPG-inline signed command:
* `Upload-component r4.2 app-linux-pdf-converter 1a770cf330cecf783f76194b117fda721564c0b4 current all` (available 5 days from now)
* `Upload-component r4.2 app-linux-pdf-converter 1a770cf330cecf783f76194b117fda721564c0b4 security-testing`
You can choose subset of distributions like:
* `Upload-component r4.2 app-linux-pdf-converter 1a770cf330cecf783f76194b117fda721564c0b4 current vm-bookworm,vm-fc37` (available 5 days from now)
Above commands will work only if packages in current-testing repository were built from given commit (i.e. no new version superseded it).
For more information on how to test this update, please take a look at https://www.qubes-os.org/doc/testing/#updates.
|
non_process
|
app linux pdf converter update of app linux pdf converter to for qubes see comments below for details and build status from commit qubesos qubes app linux pdf converter version qubesos qubes app linux pdf converter qvm convert pdf nautilus py adjust to work with nautilus api referenced issues qubesos qubes issues if you re release manager you can issue gpg inline signed command upload component app linux pdf converter current all available days from now upload component app linux pdf converter security testing you can choose subset of distributions like upload component app linux pdf converter current vm bookworm vm available days from now above commands will work only if packages in current testing repository were built from given commit i e no new version superseded it for more information on how to test this update please take a look at
| 0
|
5,286
| 8,072,291,665
|
IssuesEvent
|
2018-08-06 15:32:07
|
cptechinc/soft-dplus-web
|
https://api.github.com/repos/cptechinc/soft-dplus-web
|
opened
|
Processwire Configs
|
Processwire
|
The next part to do is to abstract some of the content into Processwire's Backend.
So the first thing you'll do is you'll create a template called config
Then you'll make a field called site_displayname and apply it to the template,
Then add config to your pages, and change the site_displayname field to DistributionPlus.
After, any place where distribution plus is hard-coded then you'll just call to display that field value.
For reference, look in soft-dpluso we do it in the content/nav/yt-navigation.php file.
For your address, you'll want to change it to be a textarea field with htmlmarkup. You'll also want to add contact_name so you can hold CPTech as a value in there instead of having to have it hard-coded.
|
1.0
|
Processwire Configs - The next part to do is to abstract some of the content into Processwire's Backend.
So the first thing you'll do is you'll create a template called config
Then you'll make a field called site_displayname and apply it to the template,
Then add config to your pages, and change the site_displayname field to DistributionPlus.
After, any place where distribution plus is hard-coded then you'll just call to display that field value.
For reference, look in soft-dpluso we do it in the content/nav/yt-navigation.php file.
For your address, you'll want to change it to be a textarea field with htmlmarkup. You'll also want to add contact_name so you can hold CPTech as a value in there instead of having to have it hard-coded.
|
process
|
processwire configs the next part to do is to abstract some of the content into processwire s backend so the first thing you ll do is you ll create a template called config then you ll make a field called site displayname and apply it to the template then add config to your pages and change the site displayname field to distributionplus after any place where distribution plus is hard coded then you ll just call to display that field value for reference look in soft dpluso we do it in the content nav yt navigation php file for your address you ll want to change it to be a textarea field with htmlmarkup you ll also want to add contact name so you can hold cptech as a value in there instead of having to have it hard coded
| 1
|
9,036
| 12,130,107,919
|
IssuesEvent
|
2020-04-23 00:30:40
|
GoogleCloudPlatform/python-docs-samples
|
https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples
|
closed
|
remove gcp-devrel-py-tools from appengine/standard/endpoints-frameworks-v2/echo/requirements-test.txt
|
priority: p2 remove-gcp-devrel-py-tools type: process
|
remove gcp-devrel-py-tools from appengine/standard/endpoints-frameworks-v2/echo/requirements-test.txt
|
1.0
|
remove gcp-devrel-py-tools from appengine/standard/endpoints-frameworks-v2/echo/requirements-test.txt - remove gcp-devrel-py-tools from appengine/standard/endpoints-frameworks-v2/echo/requirements-test.txt
|
process
|
remove gcp devrel py tools from appengine standard endpoints frameworks echo requirements test txt remove gcp devrel py tools from appengine standard endpoints frameworks echo requirements test txt
| 1
|
4,697
| 7,540,181,518
|
IssuesEvent
|
2018-04-17 05:00:40
|
allinurl/goaccess
|
https://api.github.com/repos/allinurl/goaccess
|
closed
|
Unable to open the specified log file. No such file or directory
|
log-processing question
|

How can I solve this? Thanks!
|
1.0
|
Unable to open the specified log file. No such file or directory - 
How can I solve this? Thanks!
|
process
|
unable to open the specified log file no such file or directory how can i solve this thanks
| 1
|
4,589
| 7,431,225,223
|
IssuesEvent
|
2018-03-25 12:41:55
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
Flaky test-child-process-disconnect
|
CI / flaky test child_process test
|
* **Version**: master
* **Platform**: ubuntun 1204
* **Subsystem**: process
Seen in unrelated CI run
https://ci.nodejs.org/job/node-test-commit-linux/16231/nodes=ubuntu1204-64/consoleFull
```
not ok 166 parallel/test-child-process-disconnect
---
duration_ms: 0.513
severity: fail
stack: |-
...
```
|
1.0
|
Flaky test-child-process-disconnect - * **Version**: master
* **Platform**: ubuntun 1204
* **Subsystem**: process
Seen in unrelated CI run
https://ci.nodejs.org/job/node-test-commit-linux/16231/nodes=ubuntu1204-64/consoleFull
```
not ok 166 parallel/test-child-process-disconnect
---
duration_ms: 0.513
severity: fail
stack: |-
...
```
|
process
|
flaky test child process disconnect version master platform ubuntun subsystem process seen in unrelated ci run not ok parallel test child process disconnect duration ms severity fail stack
| 1
|
6,317
| 9,333,054,273
|
IssuesEvent
|
2019-03-28 13:39:58
|
plazi/arcadia-project
|
https://api.github.com/repos/plazi/arcadia-project
|
opened
|
QC introduction skype meeting, April 3
|
Article processing meetings
|
@gsautter will present the new QC tool for article post processing to @teodorgeorgiev @myrmoteras
2pm via skype, April 3
@teodorgeorgiev please confirm time
|
1.0
|
QC introduction skype meeting, April 3 - @gsautter will present the new QC tool for article post processing to @teodorgeorgiev @myrmoteras
2pm via skype, April 3
@teodorgeorgiev please confirm time
|
process
|
qc introduction skype meeting april gsautter will present the new qc tool for article post processing to teodorgeorgiev myrmoteras via skype april teodorgeorgiev please confirm time
| 1
|
7,008
| 2,596,356,595
|
IssuesEvent
|
2015-02-20 20:07:05
|
eliasjhojala/DMX-lighting-controller
|
https://api.github.com/repos/eliasjhojala/DMX-lighting-controller
|
opened
|
Fixture Configuration SubWindow
|
enhancement Priority 2
|
It's about time we fully get rid of the nasty old ControlP5 window. The plan is to offer all the functionality that the ControlP5 window offered for editing fixtures, possibly even more, in a more intuitive way, similar to how presets are created.
The trusses will also be getting a similar treatment.
|
1.0
|
Fixture Configuration SubWindow - It's about time we fully get rid of the nasty old ControlP5 window. The plan is to offer all the functionality that the ControlP5 window offered for editing fixtures, possibly even more, in a more intuitive way, similar to how presets are created.
The trusses will also be getting a similar treatment.
|
non_process
|
fixture configuration subwindow it s about time we fully get rid of the nasty old window the plan is to offer all the functionality that the window offered for editing fixtures possibly even more in a more intuitive way similar to how presets are created the trusses will also be getting a similar treatment
| 0
|
11,801
| 14,626,895,749
|
IssuesEvent
|
2020-12-23 11:12:15
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
No stdout event fired on 'data' for child_process
|
child_process events
|
I implemented a static method that takes two input parameters: (1) script/command (2) args to be passed to the given script/command. The issue is that I'm not able to get the output. Note I'm running on a mac machine if that makes sense.
Here is my code:
```typescript
import { spawn } from 'child_process';
import { StringBuilder } from 'typescript-string-operations';
export class Exec {
/**
* To execute a given command with its args in a separate process
* without creating a new shell
* @static
* @param {string} cmd
* @param {string[]} args
* @returns {Promise<any>}
* @memberof Exec
*/
static exec(cmd:string, args:string[]):Promise<string> {
return new Promise((resolve, reject) => {
try{
let out:StringBuilder = new StringBuilder();
let child = spawn(cmd, args);
child.stdout.on('data', (data) => {
out.Append(data.toString());
});
child.stderr.on('data', (data) => {
out.Append(data.toString())
});
child.on('disconnect', (code) => {
let msg:string = `${code} : Disconnected due to the parent process
manually calls the child.disconnect function` ;
reject(msg);
})
child.on('error', (code) => {
let msg:string = `${code} : The process could not be spawned or killed` ;
reject(msg) ;
})
child.on('close', (code) => {
let msg:string = `${code} : The stdio streams of a child process get closed`
Log.logger().info(msg);
if(code !== 0){
reject(msg);
}else{
resolve(out.ToString());
}
})
child.on('exit', (code, signal) => {
let msg:string = `Exited with code ${code} signal ${signal}` ;
if(code !== 0){
reject(msg);
}
});
} catch(err){
console.log(`Failed due to error: ${err}`);
reject(err.message);
}
});
}
}
// try to run the following line inside a function with 'async' word
let p = await Exec.exec('ps', ['ax']);
```
|
1.0
|
No stdout event fired on 'data' for child_process - I implemented a static method that takes two input parameters: (1) script/command (2) args to be passed to the given script/command. The issue is that I'm not able to get the output. Note I'm running on a mac machine if that makes sense.
Here is my code:
```typescript
import { spawn } from 'child_process';
import { StringBuilder } from 'typescript-string-operations';
export class Exec {
/**
* To execute a given command with its args in a separate process
* without creating a new shell
* @static
* @param {string} cmd
* @param {string[]} args
* @returns {Promise<any>}
* @memberof Exec
*/
static exec(cmd:string, args:string[]):Promise<string> {
return new Promise((resolve, reject) => {
try{
let out:StringBuilder = new StringBuilder();
let child = spawn(cmd, args);
child.stdout.on('data', (data) => {
out.Append(data.toString());
});
child.stderr.on('data', (data) => {
out.Append(data.toString())
});
child.on('disconnect', (code) => {
let msg:string = `${code} : Disconnected due to the parent process
manually calls the child.disconnect function` ;
reject(msg);
})
child.on('error', (code) => {
let msg:string = `${code} : The process could not be spawned or killed` ;
reject(msg) ;
})
child.on('close', (code) => {
let msg:string = `${code} : The stdio streams of a child process get closed`
Log.logger().info(msg);
if(code !== 0){
reject(msg);
}else{
resolve(out.ToString());
}
})
child.on('exit', (code, signal) => {
let msg:string = `Exited with code ${code} signal ${signal}` ;
if(code !== 0){
reject(msg);
}
});
} catch(err){
console.log(`Failed due to error: ${err}`);
reject(err.message);
}
});
}
}
// try to run the following line inside a function with 'async' word
let p = await Exec.exec('ps', ['ax']);
```
|
process
|
no stdout event fired on data for child process i implemented a static method that takes two input parameters script command args to be passed to the given script command the issue is that i m not able to get the output note i m running on a mac machine if that makes sense here is my code typescript import spawn from child process import stringbuilder from typescript string operations export class exec to execute a given command with its args in a separate process without creating a new shell static param string cmd param string args returns promise memberof exec static exec cmd string args string promise return new promise resolve reject try let out stringbuilder new stringbuilder let child spawn cmd args child stdout on data data out append data tostring child stderr on data data out append data tostring child on disconnect code let msg string code disconnected due to the parent process manually calls the child disconnect function reject msg child on error code let msg string code the process could not be spawned or killed reject msg child on close code let msg string code the stdio streams of a child process get closed log logger info msg if code reject msg else resolve out tostring child on exit code signal let msg string exited with code code signal signal if code reject msg catch err console log failed due to error err reject err message try to run the following line inside a function with async word let p await exec exec ps
| 1
|
3,183
| 6,258,913,548
|
IssuesEvent
|
2017-07-14 16:44:05
|
PeaceGeeksSociety/salesforce
|
https://api.github.com/repos/PeaceGeeksSociety/salesforce
|
opened
|
Filter volunteers (leads and contacts) to better place them and identify/report who is in our volunteer pool
|
Recruitment Processes
|
We would like to filter volunteers based on the following fields:
- By skills and experience
- Expertise
- Availability
- Affiliation
- Registration date
- Desired assignments/special interest areas
- Assignments to date with ability to read comments or feedback on success of participation
This is because it would allow us to better identify volunteers for specific tasks as volunteer skills and experiences aren't currently available in SF.
Done when: Volunteers can be filtered based on the above categories.
|
1.0
|
Filter volunteers (leads and contacts) to better place them and identify/report who is in our volunteer pool - We would like to filter volunteers based on the following fields:
- By skills and experience
- Expertise
- Availability
- Affiliation
- Registration date
- Desired assignments/special interest areas
- Assignments to date with ability to read comments or feedback on success of participation
This is because it would allow us to better identify volunteers for specific tasks as volunteer skills and experiences aren't currently available in SF.
Done when: Volunteers can be filtered based on the above categories.
|
process
|
filter volunteers leads and contacts to better place them and identify report who is in our volunteer pool we would like to filter volunteers based on the following fields by skills and experience expertise availability affiliation registration date desired assignments special interest areas assignments to date with ability to read comments or feedback on success of participation this is because it would allow us to better identify volunteers for specific tasks as volunteer skills and experiences aren t currently available in sf done when volunteers can be filtered based on the above categories
| 1
|
11,253
| 14,019,938,805
|
IssuesEvent
|
2020-10-29 18:54:47
|
kubeflow/kubeflow
|
https://api.github.com/repos/kubeflow/kubeflow
|
closed
|
Release - Kubeflow v1.2
|
area/1.2.0 area/docs area/engprod area/kfctl area/lifecycle community/maintenance effort/2-weeks+ kind/feature kind/process priority/p1
|
/kind process
**Why you need this:**
[Is your feature request related to a problem? Please describe in details]
This is an uber tracking issue for releasing related issues/bugs.
**Describe the solution you'd like:**
[A clear and concise description of what you want to happen.]
- [ ] Update [releasing docs](https://github.com/kubeflow/kubeflow/blob/master/docs_dev/releasing.md)
- [ ] Prow/GitHub Actions to automate some of the manual releasing process
**Anything else you would like to add:**
[Miscellaneous information that will assist in solving the issue.]
/cc @jlewi @animeshsingh @aronchick @Jeffwan @Bobgy @adrian555
|
1.0
|
Release - Kubeflow v1.2 - /kind process
**Why you need this:**
[Is your feature request related to a problem? Please describe in details]
This is an uber tracking issue for releasing related issues/bugs.
**Describe the solution you'd like:**
[A clear and concise description of what you want to happen.]
- [ ] Update [releasing docs](https://github.com/kubeflow/kubeflow/blob/master/docs_dev/releasing.md)
- [ ] Prow/GitHub Actions to automate some of the manual releasing process
**Anything else you would like to add:**
[Miscellaneous information that will assist in solving the issue.]
/cc @jlewi @animeshsingh @aronchick @Jeffwan @Bobgy @adrian555
|
process
|
release kubeflow kind process why you need this this is an uber tracking issue for releasing related issues bugs describe the solution you d like update prow github actions to automate some of the manual releasing process anything else you would like to add cc jlewi animeshsingh aronchick jeffwan bobgy
| 1
|
826,028
| 31,495,769,629
|
IssuesEvent
|
2023-08-31 02:06:51
|
kubernetes/ingress-nginx
|
https://api.github.com/repos/kubernetes/ingress-nginx
|
closed
|
Run make dev-env syntax error
|
kind/bug needs-triage needs-priority
|
<!--
Welcome to ingress-nginx! For a smooth issue process, try to answer the following questions.
Don't worry if they're not all applicable; just try to include what you can :-).
More info helps better understanding of the issue (needless to say).
If you need to include code snippets or logs, please put them in fenced code
blocks. If they're super-long, please use the details tag like
<details><summary>super-long log</summary> lots of stuff </details>
-->
<!--
IMPORTANT!!!
Please complete the next sections or the issue will be closed.
This questions are the first thing we need to know to understand the context.
-->
**What happened**:
```
$ make dev-env
3.12.2
build/dev-env.sh: line 50: [[: 3.12.2: syntax error: invalid arithmetic operator (error token is ".12.2")
[dev-env] building image
```
<!-- (please include exact error messages if you can) -->
**What you expected to happen**:
Run `make dev-env` without any error
**Version**:
Latest
<!-- If this is actually about documentation, uncomment the following block -->
<!--
/kind documentation
/remove-kind bug
-->
```[tasklist]
### Tasks
```
|
1.0
|
Run make dev-env syntax error - <!--
Welcome to ingress-nginx! For a smooth issue process, try to answer the following questions.
Don't worry if they're not all applicable; just try to include what you can :-).
More info helps better understanding of the issue (needless to say).
If you need to include code snippets or logs, please put them in fenced code
blocks. If they're super-long, please use the details tag like
<details><summary>super-long log</summary> lots of stuff </details>
-->
<!--
IMPORTANT!!!
Please complete the next sections or the issue will be closed.
This questions are the first thing we need to know to understand the context.
-->
**What happened**:
```
$ make dev-env
3.12.2
build/dev-env.sh: line 50: [[: 3.12.2: syntax error: invalid arithmetic operator (error token is ".12.2")
[dev-env] building image
```
<!-- (please include exact error messages if you can) -->
**What you expected to happen**:
Run `make dev-env` without any error
**Version**:
Latest
<!-- If this is actually about documentation, uncomment the following block -->
<!--
/kind documentation
/remove-kind bug
-->
```[tasklist]
### Tasks
```
|
non_process
|
run make dev env syntax error welcome to ingress nginx for a smooth issue process try to answer the following questions don t worry if they re not all applicable just try to include what you can more info helps better understanding of the issue needless to say if you need to include code snippets or logs please put them in fenced code blocks if they re super long please use the details tag like super long log lots of stuff important please complete the next sections or the issue will be closed this questions are the first thing we need to know to understand the context what happened make dev env build dev env sh line syntax error invalid arithmetic operator error token is building image what you expected to happen run make dev env without any error version latest kind documentation remove kind bug tasks
| 0
|
15,079
| 18,782,585,352
|
IssuesEvent
|
2021-11-08 08:46:19
|
tndd/alpaca_v2
|
https://api.github.com/repos/tndd/alpaca_v2
|
closed
|
テーブル検討 bars_close_price_movements
|
data_processing
|
# 概要
https://github.com/tndd/alpaca_v2/issues/1#issue-1043105037
前回の解析結果があまりに芳しく無いので、異なる方向性を検討する。
そこで以下のパラメータを与え、結果を検証する。
# テーブル定義
前回は説明変数に量的変数を用いていたが、それが予測モデルの複雑化を招き精度低下を引き起こしたのかもしれない。
そこで今回は目的変数をカテゴリ化し質的変数にしてみる。
## 説明変数
カラム名 | 型 | 説明
-- | -- | --
today | int | 今日と1日前の終値を比較した値動き (0: down, 1: up, 2: eq)
ago_1 | int | 1日前と2日前の終値を比較した値動き (0: down, 1: up, 2: eq)
ago_2 | int | 上と同パターン
ago_3 | int | 上と同パターン
## 目的変数
カラム名 | 型 | 説明
-- | -- | --
tomorrow | int | 明日と今日の終値を比較した値動き(0: down, 1: up, 2: eq)
|
1.0
|
テーブル検討 bars_close_price_movements - # 概要
https://github.com/tndd/alpaca_v2/issues/1#issue-1043105037
前回の解析結果があまりに芳しく無いので、異なる方向性を検討する。
そこで以下のパラメータを与え、結果を検証する。
# テーブル定義
前回は説明変数に量的変数を用いていたが、それが予測モデルの複雑化を招き精度低下を引き起こしたのかもしれない。
そこで今回は目的変数をカテゴリ化し質的変数にしてみる。
## 説明変数
カラム名 | 型 | 説明
-- | -- | --
today | int | 今日と1日前の終値を比較した値動き (0: down, 1: up, 2: eq)
ago_1 | int | 1日前と2日前の終値を比較した値動き (0: down, 1: up, 2: eq)
ago_2 | int | 上と同パターン
ago_3 | int | 上と同パターン
## 目的変数
カラム名 | 型 | 説明
-- | -- | --
tomorrow | int | 明日と今日の終値を比較した値動き(0: down, 1: up, 2: eq)
|
process
|
テーブル検討 bars close price movements 概要 前回の解析結果があまりに芳しく無いので、異なる方向性を検討する。 そこで以下のパラメータを与え、結果を検証する。 テーブル定義 前回は説明変数に量的変数を用いていたが、それが予測モデルの複雑化を招き精度低下を引き起こしたのかもしれない。 そこで今回は目的変数をカテゴリ化し質的変数にしてみる。 説明変数 カラム名 型 説明 today int down up eq ago int down up eq ago int 上と同パターン ago int 上と同パターン 目的変数 カラム名 型 説明 tomorrow int 明日と今日の終値を比較した値動き down up eq
| 1
|
15,798
| 19,986,923,154
|
IssuesEvent
|
2022-01-30 20:02:48
|
plazi/community
|
https://api.github.com/repos/plazi/community
|
closed
|
Lestid BiCIKL project
|
process request
|
@flsimoes
can you look into these treatmentCitation in Lestids
* markup of all the treatmentCitation
* get a list of all the bibRefs cited in treatment Citation
* check whether we have all the publications cited in the TC processed
Do you have access to the list of existing, processed articles?
|
1.0
|
Lestid BiCIKL project - @flsimoes
can you look into these treatmentCitation in Lestids
* markup of all the treatmentCitation
* get a list of all the bibRefs cited in treatment Citation
* check whether we have all the publications cited in the TC processed
Do you have access to the list of existing, processed articles?
|
process
|
lestid bicikl project flsimoes can you look into these treatmentcitation in lestids markup of all the treatmentcitation get a list of all the bibrefs cited in treatment citation check whether we have all the publications cited in the tc processed do you have access to the list of existing processed articles
| 1
|
757,242
| 26,502,087,756
|
IssuesEvent
|
2023-01-18 11:03:09
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
mdn.github.io - site is not usable
|
browser-firefox priority-important os-mac engine-gecko
|
<!-- @browser: Firefox 108.0 -->
<!-- @ua_header: Mozilla/5.0 (Macintosh; Intel Mac OS X 10.15; rv:108.0) Gecko/20100101 Firefox/108.0 -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/116788 -->
**URL**: https://mdn.github.io/dom-examples/web-share/
**Browser / Version**: Firefox 108.0
**Operating System**: Mac OS X 10.15
**Tested Another Browser**: Yes Safari
**Problem type**: Site is not usable
**Description**: Buttons or links not working
**Steps to Reproduce**:
Klick the 'share mdn' button:
Expected result:
share dialog should open
Actual result:
nothing happens
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
mdn.github.io - site is not usable - <!-- @browser: Firefox 108.0 -->
<!-- @ua_header: Mozilla/5.0 (Macintosh; Intel Mac OS X 10.15; rv:108.0) Gecko/20100101 Firefox/108.0 -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/116788 -->
**URL**: https://mdn.github.io/dom-examples/web-share/
**Browser / Version**: Firefox 108.0
**Operating System**: Mac OS X 10.15
**Tested Another Browser**: Yes Safari
**Problem type**: Site is not usable
**Description**: Buttons or links not working
**Steps to Reproduce**:
Klick the 'share mdn' button:
Expected result:
share dialog should open
Actual result:
nothing happens
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_process
|
mdn github io site is not usable url browser version firefox operating system mac os x tested another browser yes safari problem type site is not usable description buttons or links not working steps to reproduce klick the share mdn button expected result share dialog should open actual result nothing happens browser configuration none from with ❤️
| 0
|
20,309
| 26,950,591,912
|
IssuesEvent
|
2023-02-08 11:22:19
|
firebase/firebase-cpp-sdk
|
https://api.github.com/repos/firebase/firebase-cpp-sdk
|
reopened
|
[C++] Nightly Integration Testing Report
|
type: process nightly-testing
|
Note: This report excludes firestore. Please also check **[the report for firestore](https://github.com/firebase/firebase-cpp-sdk/issues/1178)**
***
<hidden value="integration-test-status-comment"></hidden>
### ✅ [build against repo] Integration test succeeded!
Requested by @DellaBitta on commit 5c0eebe6cdffa6007bd82cfac606c515ef9abb94
Last updated: Tue Feb 7 02:57 PST 2023
**[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/4112164378)**
<hidden value="integration-test-status-comment"></hidden>
***
### ✅ [build against SDK] Integration test succeeded!
Requested by @firebase-workflow-trigger[bot] on commit 5c0eebe6cdffa6007bd82cfac606c515ef9abb94
Last updated: Sun Feb 5 05:19 PST 2023
**[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/4096493964)**
<hidden value="integration-test-status-comment"></hidden>
|
1.0
|
[C++] Nightly Integration Testing Report - Note: This report excludes firestore. Please also check **[the report for firestore](https://github.com/firebase/firebase-cpp-sdk/issues/1178)**
***
<hidden value="integration-test-status-comment"></hidden>
### ✅ [build against repo] Integration test succeeded!
Requested by @DellaBitta on commit 5c0eebe6cdffa6007bd82cfac606c515ef9abb94
Last updated: Tue Feb 7 02:57 PST 2023
**[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/4112164378)**
<hidden value="integration-test-status-comment"></hidden>
***
### ✅ [build against SDK] Integration test succeeded!
Requested by @firebase-workflow-trigger[bot] on commit 5c0eebe6cdffa6007bd82cfac606c515ef9abb94
Last updated: Sun Feb 5 05:19 PST 2023
**[View integration test log & download artifacts](https://github.com/firebase/firebase-cpp-sdk/actions/runs/4096493964)**
<hidden value="integration-test-status-comment"></hidden>
|
process
|
nightly integration testing report note this report excludes firestore please also check ✅ nbsp integration test succeeded requested by dellabitta on commit last updated tue feb pst ✅ nbsp integration test succeeded requested by firebase workflow trigger on commit last updated sun feb pst
| 1
|
21,128
| 28,099,081,119
|
IssuesEvent
|
2023-03-30 17:59:53
|
metabase/metabase
|
https://api.github.com/repos/metabase/metabase
|
closed
|
[MLv2] QP support for pMBQL queries
|
.Team/QueryProcessor :hammer_and_wrench:
|
Dependent on #28696
The QP should recognize `:type :pipeline` queries and be able to convert them to legacy MBQL and run them. This should work for already normalized queries as well as ones coming in from the REST API
|
1.0
|
[MLv2] QP support for pMBQL queries - Dependent on #28696
The QP should recognize `:type :pipeline` queries and be able to convert them to legacy MBQL and run them. This should work for already normalized queries as well as ones coming in from the REST API
|
process
|
qp support for pmbql queries dependent on the qp should recognize type pipeline queries and be able to convert them to legacy mbql and run them this should work for already normalized queries as well as ones coming in from the rest api
| 1
|
17,091
| 22,600,564,119
|
IssuesEvent
|
2022-06-29 08:45:52
|
camunda/zeebe
|
https://api.github.com/repos/camunda/zeebe
|
opened
|
Reject create process instance command using logical transaction
|
kind/toil team/process-automation
|
**Description**
#9637 used a workaround to reject the Create Process Instance command when the engine fails to subscribe to the relevant catch events. The workaround is to throw an exception and let the ProcessingStateMachine handle the exception as an error. The workaround was necessary because of a lack of support for:
- #9420
As soon as logical transactions are supported, we should use the logical transaction to rollback the changes and reject the command.
|
1.0
|
Reject create process instance command using logical transaction - **Description**
#9637 used a workaround to reject the Create Process Instance command when the engine fails to subscribe to the relevant catch events. The workaround is to throw an exception and let the ProcessingStateMachine handle the exception as an error. The workaround was necessary because of a lack of support for:
- #9420
As soon as logical transactions are supported, we should use the logical transaction to rollback the changes and reject the command.
|
process
|
reject create process instance command using logical transaction description used a workaround to reject the create process instance command when the engine fails to subscribe to the relevant catch events the workaround is to throw an exception and let the processingstatemachine handle the exception as an error the workaround was necessary because of a lack of support for as soon as logical transactions are supported we should use the logical transaction to rollback the changes and reject the command
| 1
|
118,082
| 17,576,350,430
|
IssuesEvent
|
2021-08-15 17:32:21
|
turkdevops/brackets
|
https://api.github.com/repos/turkdevops/brackets
|
reopened
|
CVE-2013-7370 (Medium) detected in connect-1.8.7.tgz
|
security vulnerability
|
## CVE-2013-7370 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>connect-1.8.7.tgz</b></p></summary>
<p>High performance middleware framework</p>
<p>Library home page: <a href="https://registry.npmjs.org/connect/-/connect-1.8.7.tgz">https://registry.npmjs.org/connect/-/connect-1.8.7.tgz</a></p>
<p>Path to dependency file: brackets/src/extensions/default/JavaScriptQuickEdit/unittest-files/jquery-ui/package.json</p>
<p>Path to vulnerable library: brackets/src/extensions/default/JavaScriptQuickEdit/unittest-files/jquery-ui/node_modules/connect/package.json</p>
<p>
Dependency Hierarchy:
- grunt-0.3.9.tgz (Root Library)
- :x: **connect-1.8.7.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/brackets/commit/3fa45ae9a1a5190c6d0729a3d4b58907c1ad749e">3fa45ae9a1a5190c6d0729a3d4b58907c1ad749e</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
node-connect before 2.8.1 has XSS in the Sencha Labs Connect middleware
<p>Publish Date: 2019-12-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2013-7370>CVE-2013-7370</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2013-7370">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2013-7370</a></p>
<p>Release Date: 2013-07-01</p>
<p>Fix Resolution: 2.8.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2013-7370 (Medium) detected in connect-1.8.7.tgz - ## CVE-2013-7370 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>connect-1.8.7.tgz</b></p></summary>
<p>High performance middleware framework</p>
<p>Library home page: <a href="https://registry.npmjs.org/connect/-/connect-1.8.7.tgz">https://registry.npmjs.org/connect/-/connect-1.8.7.tgz</a></p>
<p>Path to dependency file: brackets/src/extensions/default/JavaScriptQuickEdit/unittest-files/jquery-ui/package.json</p>
<p>Path to vulnerable library: brackets/src/extensions/default/JavaScriptQuickEdit/unittest-files/jquery-ui/node_modules/connect/package.json</p>
<p>
Dependency Hierarchy:
- grunt-0.3.9.tgz (Root Library)
- :x: **connect-1.8.7.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/brackets/commit/3fa45ae9a1a5190c6d0729a3d4b58907c1ad749e">3fa45ae9a1a5190c6d0729a3d4b58907c1ad749e</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
node-connect before 2.8.1 has XSS in the Sencha Labs Connect middleware
<p>Publish Date: 2019-12-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2013-7370>CVE-2013-7370</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2013-7370">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2013-7370</a></p>
<p>Release Date: 2013-07-01</p>
<p>Fix Resolution: 2.8.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in connect tgz cve medium severity vulnerability vulnerable library connect tgz high performance middleware framework library home page a href path to dependency file brackets src extensions default javascriptquickedit unittest files jquery ui package json path to vulnerable library brackets src extensions default javascriptquickedit unittest files jquery ui node modules connect package json dependency hierarchy grunt tgz root library x connect tgz vulnerable library found in head commit a href found in base branch master vulnerability details node connect before has xss in the sencha labs connect middleware publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
97,596
| 8,663,221,032
|
IssuesEvent
|
2018-11-28 16:48:14
|
SME-Issues/issues
|
https://api.github.com/repos/SME-Issues/issues
|
closed
|
Test Summary - 28/11/2018 - 5004
|
NLP Api pulse_tests
|
### Canonical
- Payment Query Primary Expression Tests (20): **95%** pass (19), 1 failed understood (#1241)
|
1.0
|
Test Summary - 28/11/2018 - 5004 - ### Canonical
- Payment Query Primary Expression Tests (20): **95%** pass (19), 1 failed understood (#1241)
|
non_process
|
test summary canonical payment query primary expression tests pass failed understood
| 0
|
105,088
| 9,022,908,981
|
IssuesEvent
|
2019-02-07 04:17:03
|
InfiniteAdventures/ia-trilogy
|
https://api.github.com/repos/InfiniteAdventures/ia-trilogy
|
reopened
|
S. 282: "Die hundert Prozent Wahlstimmen..."
|
Grammatik Testdruck-Fehler ⚠ Logik
|
Erstens überzeugt mich der Satz inhaltlich nicht so wirklich (warum sollten nur 2% Island wählen, der gerade die Welt gerettet hat? Wie soll er weltweit so viele Stimmen kaufen?). Zweitens und wichtiger: Soll der Satz schon die Wahl selbst beschreiben? Oben geht es ja nur um eine Umfrage. Aber eine Seite weiter wird dann einfach davon ausgegangen, dass Island gewählt wurde. Das Plusquamperfekt im Satz irritiert halt auch irgendwie.
Wenn es um die Wahl selbst geht, würde ich das vielleicht an der Stelle deutlicher machen.
|
1.0
|
S. 282: "Die hundert Prozent Wahlstimmen..." - Erstens überzeugt mich der Satz inhaltlich nicht so wirklich (warum sollten nur 2% Island wählen, der gerade die Welt gerettet hat? Wie soll er weltweit so viele Stimmen kaufen?). Zweitens und wichtiger: Soll der Satz schon die Wahl selbst beschreiben? Oben geht es ja nur um eine Umfrage. Aber eine Seite weiter wird dann einfach davon ausgegangen, dass Island gewählt wurde. Das Plusquamperfekt im Satz irritiert halt auch irgendwie.
Wenn es um die Wahl selbst geht, würde ich das vielleicht an der Stelle deutlicher machen.
|
non_process
|
s die hundert prozent wahlstimmen erstens überzeugt mich der satz inhaltlich nicht so wirklich warum sollten nur island wählen der gerade die welt gerettet hat wie soll er weltweit so viele stimmen kaufen zweitens und wichtiger soll der satz schon die wahl selbst beschreiben oben geht es ja nur um eine umfrage aber eine seite weiter wird dann einfach davon ausgegangen dass island gewählt wurde das plusquamperfekt im satz irritiert halt auch irgendwie wenn es um die wahl selbst geht würde ich das vielleicht an der stelle deutlicher machen
| 0
|
4,511
| 7,354,920,148
|
IssuesEvent
|
2018-03-09 09:10:29
|
UKHomeOffice/dq-aws-transition
|
https://api.github.com/repos/UKHomeOffice/dq-aws-transition
|
closed
|
Test End-to-End Job_55_SMM_ACL Wherescape Job in Prod
|
DQ Data Pipeline Production SSM processing
|
Task Estimate: 3 hours
All tasks complete and expected files and data
- [x] End-to-End Job_55_SMM_ACL tested
- [x] Job running in Prod from Wherescape
|
1.0
|
Test End-to-End Job_55_SMM_ACL Wherescape Job in Prod - Task Estimate: 3 hours
All tasks complete and expected files and data
- [x] End-to-End Job_55_SMM_ACL tested
- [x] Job running in Prod from Wherescape
|
process
|
test end to end job smm acl wherescape job in prod task estimate hours all tasks complete and expected files and data end to end job smm acl tested job running in prod from wherescape
| 1
|
375,949
| 26,182,779,979
|
IssuesEvent
|
2023-01-02 18:04:34
|
bounswe/bounswe2022group9
|
https://api.github.com/repos/bounswe/bounswe2022group9
|
opened
|
[Documentation] Documenting Individual Contribution Report of İbrahim Buğra Elmas
|
Documentation Priority: High
|
I should add my contributions report to our project repo.
Deadline: 02.02.2023 @23.59
|
1.0
|
[Documentation] Documenting Individual Contribution Report of İbrahim Buğra Elmas - I should add my contributions report to our project repo.
Deadline: 02.02.2023 @23.59
|
non_process
|
documenting individual contribution report of i̇brahim buğra elmas i should add my contributions report to our project repo deadline
| 0
|
11,377
| 14,220,153,040
|
IssuesEvent
|
2020-11-17 14:12:34
|
mpi-forum/mpi-issues
|
https://api.github.com/repos/mpi-forum/mpi-issues
|
opened
|
Process Management Chapter RC Review
|
Chapter Committee Change chap-process mpi-4.0
|
# Problem
@bgoglin found this during the RC review:
### Section 11.1
"in the presense of" -> presence ?
|
1.0
|
Process Management Chapter RC Review - # Problem
@bgoglin found this during the RC review:
### Section 11.1
"in the presense of" -> presence ?
|
process
|
process management chapter rc review problem bgoglin found this during the rc review section in the presense of presence
| 1
|
728,046
| 25,063,753,204
|
IssuesEvent
|
2022-11-07 05:59:24
|
voxel51/fiftyone
|
https://api.github.com/repos/voxel51/fiftyone
|
closed
|
[BUG] View bars swallows first new stage I try to create
|
bug app high priority
|
On `develop`, the view bar will reliably fail to load a `Limit(10)` stage when I am viewing the `quickstart-geo`, `quickstart-groups`, and `bdd100k` datasets. This does not happen on the `quickstart` dataset. Likely only a problem for larger datasets?
This only happens the first time I try to use the view bar after loading any of these datasets. The next time I try to use the view bar, it works (until I change datasets again).

|
1.0
|
[BUG] View bars swallows first new stage I try to create - On `develop`, the view bar will reliably fail to load a `Limit(10)` stage when I am viewing the `quickstart-geo`, `quickstart-groups`, and `bdd100k` datasets. This does not happen on the `quickstart` dataset. Likely only a problem for larger datasets?
This only happens the first time I try to use the view bar after loading any of these datasets. The next time I try to use the view bar, it works (until I change datasets again).

|
non_process
|
view bars swallows first new stage i try to create on develop the view bar will reliably fail to load a limit stage when i am viewing the quickstart geo quickstart groups and datasets this does not happen on the quickstart dataset likely only a problem for larger datasets this only happens the first time i try to use the view bar after loading any of these datasets the next time i try to use the view bar it works until i change datasets again
| 0
|
7,154
| 10,300,398,854
|
IssuesEvent
|
2019-08-28 13:49:11
|
codacy/codacy-meta
|
https://api.github.com/repos/codacy/codacy-meta
|
closed
|
Check security leaks on OSS CI builds
|
Processes Tech
|
We might be exposing sensitive information on CI logs, preventing us from allowing builds to run for fork PRs. We should find a solution in order not to compromise any secrets and still allow builds on contributed PRs
|
1.0
|
Check security leaks on OSS CI builds - We might be exposing sensitive information on CI logs, preventing us from allowing builds to run for fork PRs. We should find a solution in order not to compromise any secrets and still allow builds on contributed PRs
|
process
|
check security leaks on oss ci builds we might be exposing sensitive information on ci logs preventing us from allowing builds to run for fork prs we should find a solution in order not to compromise any secrets and still allow builds on contributed prs
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.