Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
22,346
| 31,022,451,519
|
IssuesEvent
|
2023-08-10 06:45:03
|
bazelbuild/bazel
|
https://api.github.com/repos/bazelbuild/bazel
|
closed
|
[Mirror] https://go.dev/dl/go1.20.7.linux-amd64.tar.gz
|
P2 type: process team-OSS mirror request
|
### Please list the URLs of the archives you'd like to mirror:
https://go.dev/dl/go1.20.7.linux-amd64.tar.gz
This is to upgrade go toolchain in https://github.com/bazelbuild/remote-apis-sdks/pull/471.
|
1.0
|
[Mirror] https://go.dev/dl/go1.20.7.linux-amd64.tar.gz - ### Please list the URLs of the archives you'd like to mirror:
https://go.dev/dl/go1.20.7.linux-amd64.tar.gz
This is to upgrade go toolchain in https://github.com/bazelbuild/remote-apis-sdks/pull/471.
|
process
|
please list the urls of the archives you d like to mirror this is to upgrade go toolchain in
| 1
|
54,813
| 13,453,801,575
|
IssuesEvent
|
2020-09-09 02:00:09
|
golang/go
|
https://api.github.com/repos/golang/go
|
closed
|
x/build: 2 darwin builders are missing
|
Builders NeedsInvestigation
|
From https://farmer.golang.org/#health:
```
# "macs" status: MacStadium Mac VMs
# Notes: https://github.com/golang/build/tree/master/env/darwin/macstadium
Warn: macstadium_host08a missing, not seen for 55h42m8s
Warn: macstadium_host08b missing, not seen for 55h42m6s
Warn: 2 machines missing, 10% of capacity
Warn: makemac daemon: vm.destroy("mac_10_12_host08a") = govc vm.destroy ...: exit status 1, govc: Unable to communicate with the remote host, since it is disconnected.
Warn: makemac daemon: vm.destroy("mac_10_12_host08b") = govc vm.destroy ...: exit status 1, govc: Unable to communicate with the remote host, since it is disconnected.
```
The machine hosting `macstadium_host08a` and `macstadium_host08b` is being unresponsive.
I've taken some initial steps to bring it back up, but this might need more work. I'll resume tomorrow.
/cc @toothrot @andybons
|
1.0
|
x/build: 2 darwin builders are missing - From https://farmer.golang.org/#health:
```
# "macs" status: MacStadium Mac VMs
# Notes: https://github.com/golang/build/tree/master/env/darwin/macstadium
Warn: macstadium_host08a missing, not seen for 55h42m8s
Warn: macstadium_host08b missing, not seen for 55h42m6s
Warn: 2 machines missing, 10% of capacity
Warn: makemac daemon: vm.destroy("mac_10_12_host08a") = govc vm.destroy ...: exit status 1, govc: Unable to communicate with the remote host, since it is disconnected.
Warn: makemac daemon: vm.destroy("mac_10_12_host08b") = govc vm.destroy ...: exit status 1, govc: Unable to communicate with the remote host, since it is disconnected.
```
The machine hosting `macstadium_host08a` and `macstadium_host08b` is being unresponsive.
I've taken some initial steps to bring it back up, but this might need more work. I'll resume tomorrow.
/cc @toothrot @andybons
|
non_process
|
x build darwin builders are missing from macs status macstadium mac vms notes warn macstadium missing not seen for warn macstadium missing not seen for warn machines missing of capacity warn makemac daemon vm destroy mac govc vm destroy exit status govc unable to communicate with the remote host since it is disconnected warn makemac daemon vm destroy mac govc vm destroy exit status govc unable to communicate with the remote host since it is disconnected the machine hosting macstadium and macstadium is being unresponsive i ve taken some initial steps to bring it back up but this might need more work i ll resume tomorrow cc toothrot andybons
| 0
|
103,083
| 8,877,190,488
|
IssuesEvent
|
2019-01-12 21:58:48
|
swe-ms-boun/2018fall-swe574-g2
|
https://api.github.com/repos/swe-ms-boun/2018fall-swe574-g2
|
closed
|
Selenium Web-Driver UI Test Automation Framework
|
test
|
Acceptance test cases should be automated by using selenium web driver when the UI is deployed.
|
1.0
|
Selenium Web-Driver UI Test Automation Framework - Acceptance test cases should be automated by using selenium web driver when the UI is deployed.
|
non_process
|
selenium web driver ui test automation framework acceptance test cases should be automated by using selenium web driver when the ui is deployed
| 0
|
5,317
| 8,130,529,603
|
IssuesEvent
|
2018-08-17 18:49:36
|
WalkthroughVR/Handbook
|
https://api.github.com/repos/WalkthroughVR/Handbook
|
opened
|
v0.8 - Automating the 360 connection point mapping. Having the 360 model automatically map where the last 360 was taken.
|
Photography Processing Automation
|
## Problem:
- N/A
## Hypothesis/Goal:
- N/A
## Solution/Requirements:
- [ ] N/A
## What we should learn:
- N/A
## Additional notes
- N/A
|
1.0
|
v0.8 - Automating the 360 connection point mapping. Having the 360 model automatically map where the last 360 was taken. - ## Problem:
- N/A
## Hypothesis/Goal:
- N/A
## Solution/Requirements:
- [ ] N/A
## What we should learn:
- N/A
## Additional notes
- N/A
|
process
|
automating the connection point mapping having the model automatically map where the last was taken problem n a hypothesis goal n a solution requirements n a what we should learn n a additional notes n a
| 1
|
20,435
| 27,098,849,907
|
IssuesEvent
|
2023-02-15 06:43:34
|
bazelbuild/bazel
|
https://api.github.com/repos/bazelbuild/bazel
|
closed
|
Move Python rules documentation from Bazel site to rules_python
|
P3 type: process team-Rules-Python stale
|
Blocked on bazelbuild/rules_python#202 and bazelbuild/rules_python#144, both of which will be fixed imminently. (But I don't think this will be prioritized anywhere near as soon as those.)
|
1.0
|
Move Python rules documentation from Bazel site to rules_python - Blocked on bazelbuild/rules_python#202 and bazelbuild/rules_python#144, both of which will be fixed imminently. (But I don't think this will be prioritized anywhere near as soon as those.)
|
process
|
move python rules documentation from bazel site to rules python blocked on bazelbuild rules python and bazelbuild rules python both of which will be fixed imminently but i don t think this will be prioritized anywhere near as soon as those
| 1
|
217,946
| 24,351,688,644
|
IssuesEvent
|
2022-10-03 01:09:57
|
benlazarine/cas-overlay
|
https://api.github.com/repos/benlazarine/cas-overlay
|
opened
|
CVE-2022-38751 (Medium) detected in snakeyaml-1.17.jar
|
security vulnerability
|
## CVE-2022-38751 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>snakeyaml-1.17.jar</b></p></summary>
<p>YAML 1.1 parser and emitter for Java</p>
<p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /root/.m2/repository/org/yaml/snakeyaml/1.17/snakeyaml-1.17.jar</p>
<p>
Dependency Hierarchy:
- cas-server-support-oauth-webflow-5.3.7.jar (Root Library)
- spring-boot-starter-websocket-1.5.18.RELEASE.jar
- spring-boot-starter-1.5.18.RELEASE.jar
- :x: **snakeyaml-1.17.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Using snakeYAML to parse untrusted YAML files may be vulnerable to Denial of Service attacks (DOS). If the parser is running on user supplied input, an attacker may supply content that causes the parser to crash by stackoverflow.
<p>Publish Date: 2022-09-05
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-38751>CVE-2022-38751</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://bugs.chromium.org/p/oss-fuzz/issues/detail?id=47039">https://bugs.chromium.org/p/oss-fuzz/issues/detail?id=47039</a></p>
<p>Release Date: 2022-09-05</p>
<p>Fix Resolution: org.yaml:snakeyaml:1.31</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2022-38751 (Medium) detected in snakeyaml-1.17.jar - ## CVE-2022-38751 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>snakeyaml-1.17.jar</b></p></summary>
<p>YAML 1.1 parser and emitter for Java</p>
<p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /root/.m2/repository/org/yaml/snakeyaml/1.17/snakeyaml-1.17.jar</p>
<p>
Dependency Hierarchy:
- cas-server-support-oauth-webflow-5.3.7.jar (Root Library)
- spring-boot-starter-websocket-1.5.18.RELEASE.jar
- spring-boot-starter-1.5.18.RELEASE.jar
- :x: **snakeyaml-1.17.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Using snakeYAML to parse untrusted YAML files may be vulnerable to Denial of Service attacks (DOS). If the parser is running on user supplied input, an attacker may supply content that causes the parser to crash by stackoverflow.
<p>Publish Date: 2022-09-05
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-38751>CVE-2022-38751</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://bugs.chromium.org/p/oss-fuzz/issues/detail?id=47039">https://bugs.chromium.org/p/oss-fuzz/issues/detail?id=47039</a></p>
<p>Release Date: 2022-09-05</p>
<p>Fix Resolution: org.yaml:snakeyaml:1.31</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in snakeyaml jar cve medium severity vulnerability vulnerable library snakeyaml jar yaml parser and emitter for java library home page a href path to dependency file pom xml path to vulnerable library root repository org yaml snakeyaml snakeyaml jar dependency hierarchy cas server support oauth webflow jar root library spring boot starter websocket release jar spring boot starter release jar x snakeyaml jar vulnerable library vulnerability details using snakeyaml to parse untrusted yaml files may be vulnerable to denial of service attacks dos if the parser is running on user supplied input an attacker may supply content that causes the parser to crash by stackoverflow publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org yaml snakeyaml step up your open source security game with mend
| 0
|
11,524
| 14,401,981,829
|
IssuesEvent
|
2020-12-03 14:23:28
|
pystatgen/sgkit
|
https://api.github.com/repos/pystatgen/sgkit
|
closed
|
png files from documentation should be git ignored?
|
bug process + tools
|
`make clean html` will create `png` files that are not git ignored (making git dirty). Files:
```
mydask.png
order.png
```
|
1.0
|
png files from documentation should be git ignored? - `make clean html` will create `png` files that are not git ignored (making git dirty). Files:
```
mydask.png
order.png
```
|
process
|
png files from documentation should be git ignored make clean html will create png files that are not git ignored making git dirty files mydask png order png
| 1
|
1,641
| 4,259,980,857
|
IssuesEvent
|
2016-07-11 13:02:18
|
e-government-ua/iBP
|
https://api.github.com/repos/e-government-ua/iBP
|
closed
|
Нетишин: раскрітие и уточнения в "Розірвання договору оренди земельної ділянки"
|
In process of testing in work test
|
Розірвання договору оренди земельної ділянки або припинення права постійного користування земельною ділянкою
Зауваження:
1. Потрібна можливість завантажити сканкопії наступних документів:
- копія довідки з податкової, що людина не має заборгованості по сплаті;
- копія договору оренди.
2. Додати графу: підстава для розірвання.
|
1.0
|
Нетишин: раскрітие и уточнения в "Розірвання договору оренди земельної ділянки" - Розірвання договору оренди земельної ділянки або припинення права постійного користування земельною ділянкою
Зауваження:
1. Потрібна можливість завантажити сканкопії наступних документів:
- копія довідки з податкової, що людина не має заборгованості по сплаті;
- копія договору оренди.
2. Додати графу: підстава для розірвання.
|
process
|
нетишин раскрітие и уточнения в розірвання договору оренди земельної ділянки розірвання договору оренди земельної ділянки або припинення права постійного користування земельною ділянкою зауваження потрібна можливість завантажити сканкопії наступних документів копія довідки з податкової що людина не має заборгованості по сплаті копія договору оренди додати графу підстава для розірвання
| 1
|
20,145
| 26,694,734,309
|
IssuesEvent
|
2023-01-27 09:21:47
|
UnitTestBot/UTBotJava
|
https://api.github.com/repos/UnitTestBot/UTBotJava
|
closed
|
`UtExecutionInstrumentation` does not respect timeout in concrete execution
|
ctg-bug comp-contest-estimator comp-instrumented-process
|
**Description**
Currently `UtExecutionInstrumentation` measures only concrete execution, but not model construction for parameters and execution result.
Sometimes model construction might take too much time and will hang Instrumentation process indefenitely. This occurs too often in Contest Estimator, but problem relates to plugin too.
**To Reproduce**
Steps to reproduce the behavior:
1. In `ContestEstimator.kt` set:
- set `javaHome` to JDK8
- `timeLimit=120`
- `projectFilter = listOf("guava-26.0")`
- set `utbot-junit-contest/src/main/resources/classes/guava-26.0/list` to
```
com.google.common.primitives.Booleans
com.google.common.primitives.Shorts
```
2. Start `ContestEstimator`
3. Tests for class `Shorts` once in a 2-3 generations will not be generated at all
**Expected behavior**
Everything works
**Actual behavior**
Tests for `Short` does not generate, Instrumentation process hangs on some requrest.
|
1.0
|
`UtExecutionInstrumentation` does not respect timeout in concrete execution - **Description**
Currently `UtExecutionInstrumentation` measures only concrete execution, but not model construction for parameters and execution result.
Sometimes model construction might take too much time and will hang Instrumentation process indefenitely. This occurs too often in Contest Estimator, but problem relates to plugin too.
**To Reproduce**
Steps to reproduce the behavior:
1. In `ContestEstimator.kt` set:
- set `javaHome` to JDK8
- `timeLimit=120`
- `projectFilter = listOf("guava-26.0")`
- set `utbot-junit-contest/src/main/resources/classes/guava-26.0/list` to
```
com.google.common.primitives.Booleans
com.google.common.primitives.Shorts
```
2. Start `ContestEstimator`
3. Tests for class `Shorts` once in a 2-3 generations will not be generated at all
**Expected behavior**
Everything works
**Actual behavior**
Tests for `Short` does not generate, Instrumentation process hangs on some requrest.
|
process
|
utexecutioninstrumentation does not respect timeout in concrete execution description currently utexecutioninstrumentation measures only concrete execution but not model construction for parameters and execution result sometimes model construction might take too much time and will hang instrumentation process indefenitely this occurs too often in contest estimator but problem relates to plugin too to reproduce steps to reproduce the behavior in contestestimator kt set set javahome to timelimit projectfilter listof guava set utbot junit contest src main resources classes guava list to com google common primitives booleans com google common primitives shorts start contestestimator tests for class shorts once in a generations will not be generated at all expected behavior everything works actual behavior tests for short does not generate instrumentation process hangs on some requrest
| 1
|
4,893
| 7,763,824,683
|
IssuesEvent
|
2018-06-01 17:58:28
|
StrikeNP/trac_test
|
https://api.github.com/repos/StrikeNP/trac_test
|
closed
|
GABLS2 rtm, rtp2, thlm, and thlp2 are set to zero when plotgen is run manually (but not for the nightly tests) (Trac #24)
|
Migrated from Trac enhancement post_processing senkbeil@uwm.edu
|
Some time ago, in order to test CLUBB's scalars, Brandon changed plotgen so that it outputs scalars in place of rtm and thlm. The nightly plots work great.
However, if CLUBB is run manually without outputting scalars, and then plotgen is executed manually, then rtm, thlm, rtp2, and thlp2 are set to zero. For manual runs, typically we don't want to check scalars; we just want to plot standard versions of rtm, thlm, rtp2, and thlp2. I probably forgot to mention this earlier.
Is it feasible to insert some nightly flags or re-arrange some code so that the nightly plots test the scalars, but the manual plots simply plot rtm, thlm, rtp2, and thlp2? I believe that this is what is done for other specialized nightly tests, e.g. the restart test and some of the altered grid tests. Perhaps those pieces of code would provide ideas on how to implement separate behavior for nightly and manual runs.
However, we have a deadline on the TWP-ICE case, so don't bother with this until TWP-ICE is submitted, unless it is trivial to fix.
Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/24
```json
{
"status": "closed",
"changetime": "2009-09-02T20:37:37",
"description": "Some time ago, in order to test CLUBB's scalars, Brandon changed plotgen so that it outputs scalars in place of rtm and thlm. The nightly plots work great.\n\nHowever, if CLUBB is run manually without outputting scalars, and then plotgen is executed manually, then rtm, thlm, rtp2, and thlp2 are set to zero. For manual runs, typically we don't want to check scalars; we just want to plot standard versions of rtm, thlm, rtp2, and thlp2. I probably forgot to mention this earlier.\n\nIs it feasible to insert some nightly flags or re-arrange some code so that the nightly plots test the scalars, but the manual plots simply plot rtm, thlm, rtp2, and thlp2? I believe that this is what is done for other specialized nightly tests, e.g. the restart test and some of the altered grid tests. Perhaps those pieces of code would provide ideas on how to implement separate behavior for nightly and manual runs.\n\nHowever, we have a deadline on the TWP-ICE case, so don't bother with this until TWP-ICE is submitted, unless it is trivial to fix.",
"reporter": "vlarson@uwm.edu",
"cc": "",
"resolution": "Verified by V. Larson",
"_ts": "1251923857000000",
"component": "post_processing",
"summary": "GABLS2 rtm, rtp2, thlm, and thlp2 are set to zero when plotgen is run manually (but not for the nightly tests)",
"priority": "minor",
"keywords": "scalars, gabls2, nightly plots, rtm, thlm, rtp2, thlp2",
"time": "2009-05-13T14:26:46",
"milestone": "Plotgen 3.0",
"owner": "senkbeil@uwm.edu",
"type": "enhancement"
}
```
|
1.0
|
GABLS2 rtm, rtp2, thlm, and thlp2 are set to zero when plotgen is run manually (but not for the nightly tests) (Trac #24) - Some time ago, in order to test CLUBB's scalars, Brandon changed plotgen so that it outputs scalars in place of rtm and thlm. The nightly plots work great.
However, if CLUBB is run manually without outputting scalars, and then plotgen is executed manually, then rtm, thlm, rtp2, and thlp2 are set to zero. For manual runs, typically we don't want to check scalars; we just want to plot standard versions of rtm, thlm, rtp2, and thlp2. I probably forgot to mention this earlier.
Is it feasible to insert some nightly flags or re-arrange some code so that the nightly plots test the scalars, but the manual plots simply plot rtm, thlm, rtp2, and thlp2? I believe that this is what is done for other specialized nightly tests, e.g. the restart test and some of the altered grid tests. Perhaps those pieces of code would provide ideas on how to implement separate behavior for nightly and manual runs.
However, we have a deadline on the TWP-ICE case, so don't bother with this until TWP-ICE is submitted, unless it is trivial to fix.
Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/24
```json
{
"status": "closed",
"changetime": "2009-09-02T20:37:37",
"description": "Some time ago, in order to test CLUBB's scalars, Brandon changed plotgen so that it outputs scalars in place of rtm and thlm. The nightly plots work great.\n\nHowever, if CLUBB is run manually without outputting scalars, and then plotgen is executed manually, then rtm, thlm, rtp2, and thlp2 are set to zero. For manual runs, typically we don't want to check scalars; we just want to plot standard versions of rtm, thlm, rtp2, and thlp2. I probably forgot to mention this earlier.\n\nIs it feasible to insert some nightly flags or re-arrange some code so that the nightly plots test the scalars, but the manual plots simply plot rtm, thlm, rtp2, and thlp2? I believe that this is what is done for other specialized nightly tests, e.g. the restart test and some of the altered grid tests. Perhaps those pieces of code would provide ideas on how to implement separate behavior for nightly and manual runs.\n\nHowever, we have a deadline on the TWP-ICE case, so don't bother with this until TWP-ICE is submitted, unless it is trivial to fix.",
"reporter": "vlarson@uwm.edu",
"cc": "",
"resolution": "Verified by V. Larson",
"_ts": "1251923857000000",
"component": "post_processing",
"summary": "GABLS2 rtm, rtp2, thlm, and thlp2 are set to zero when plotgen is run manually (but not for the nightly tests)",
"priority": "minor",
"keywords": "scalars, gabls2, nightly plots, rtm, thlm, rtp2, thlp2",
"time": "2009-05-13T14:26:46",
"milestone": "Plotgen 3.0",
"owner": "senkbeil@uwm.edu",
"type": "enhancement"
}
```
|
process
|
rtm thlm and are set to zero when plotgen is run manually but not for the nightly tests trac some time ago in order to test clubb s scalars brandon changed plotgen so that it outputs scalars in place of rtm and thlm the nightly plots work great however if clubb is run manually without outputting scalars and then plotgen is executed manually then rtm thlm and are set to zero for manual runs typically we don t want to check scalars we just want to plot standard versions of rtm thlm and i probably forgot to mention this earlier is it feasible to insert some nightly flags or re arrange some code so that the nightly plots test the scalars but the manual plots simply plot rtm thlm and i believe that this is what is done for other specialized nightly tests e g the restart test and some of the altered grid tests perhaps those pieces of code would provide ideas on how to implement separate behavior for nightly and manual runs however we have a deadline on the twp ice case so don t bother with this until twp ice is submitted unless it is trivial to fix migrated from json status closed changetime description some time ago in order to test clubb s scalars brandon changed plotgen so that it outputs scalars in place of rtm and thlm the nightly plots work great n nhowever if clubb is run manually without outputting scalars and then plotgen is executed manually then rtm thlm and are set to zero for manual runs typically we don t want to check scalars we just want to plot standard versions of rtm thlm and i probably forgot to mention this earlier n nis it feasible to insert some nightly flags or re arrange some code so that the nightly plots test the scalars but the manual plots simply plot rtm thlm and i believe that this is what is done for other specialized nightly tests e g the restart test and some of the altered grid tests perhaps those pieces of code would provide ideas on how to implement separate behavior for nightly and manual runs n nhowever we have a deadline on the twp ice case so don t bother with this until twp ice is submitted unless it is trivial to fix reporter vlarson uwm edu cc resolution verified by v larson ts component post processing summary rtm thlm and are set to zero when plotgen is run manually but not for the nightly tests priority minor keywords scalars nightly plots rtm thlm time milestone plotgen owner senkbeil uwm edu type enhancement
| 1
|
96,728
| 10,961,381,741
|
IssuesEvent
|
2019-11-27 15:17:55
|
sunpy/sunpy
|
https://api.github.com/repos/sunpy/sunpy
|
opened
|
Database search() method missing returns documentation
|
Documentation database
|
See https://docs.sunpy.org/en/latest/api/sunpy.database.Database.html#sunpy.database.Database.search - the return type of `search()` is currently not documented.
|
1.0
|
Database search() method missing returns documentation - See https://docs.sunpy.org/en/latest/api/sunpy.database.Database.html#sunpy.database.Database.search - the return type of `search()` is currently not documented.
|
non_process
|
database search method missing returns documentation see the return type of search is currently not documented
| 0
|
15,482
| 19,689,298,411
|
IssuesEvent
|
2022-01-12 03:56:20
|
amor71/LiuAlgoTrader
|
https://api.github.com/repos/amor71/LiuAlgoTrader
|
closed
|
Multi User Support
|
enhancement in-process no-issue-activity
|
**Is your feature request related to a problem? Please describe.**
Currently, the framework supports a single user and a single `tradeplan.toml` file. This file is read each time the platform starts. It is quite convenient for a single user running trades for him/herself.
The required change address two related issues:
1. Support several users, running different strategies in parallel,
2. Move beyond the tradeplan.toml file allowing different plans for different users
**Describe the solution you'd like**
1. extend the Portfolio table to include additional details such as broker & external account id,
2. extend trader(s) to support trading on behalf of others,
3. extend DB model to allow setting up a trade-plan per user
|
1.0
|
Multi User Support - **Is your feature request related to a problem? Please describe.**
Currently, the framework supports a single user and a single `tradeplan.toml` file. This file is read each time the platform starts. It is quite convenient for a single user running trades for him/herself.
The required change address two related issues:
1. Support several users, running different strategies in parallel,
2. Move beyond the tradeplan.toml file allowing different plans for different users
**Describe the solution you'd like**
1. extend the Portfolio table to include additional details such as broker & external account id,
2. extend trader(s) to support trading on behalf of others,
3. extend DB model to allow setting up a trade-plan per user
|
process
|
multi user support is your feature request related to a problem please describe currently the framework supports a single user and a single tradeplan toml file this file is read each time the platform starts it is quite convenient for a single user running trades for him herself the required change address two related issues support several users running different strategies in parallel move beyond the tradeplan toml file allowing different plans for different users describe the solution you d like extend the portfolio table to include additional details such as broker external account id extend trader s to support trading on behalf of others extend db model to allow setting up a trade plan per user
| 1
|
28,048
| 12,758,050,268
|
IssuesEvent
|
2020-06-29 00:36:55
|
Azure/azure-rest-api-specs
|
https://api.github.com/repos/Azure/azure-rest-api-specs
|
closed
|
HTTP 500 - Cost Management API query usage
|
Cost Management Service Attention question
|
I am running the query usage example from this link:
https://docs.microsoft.com/en-us/rest/api/cost-management/query/usage
With the following URL:
```
https://management.azure.com/subscriptions/{subscription}/resourceGroups/{group}/providers/Microsoft.CostManagement/query?api-version=2019-11-01
```
And this body payload:
```
{
timeframe: "MonthToDate"
type: "Usage"
}
```
But I'm getting HTTP 500 error:
```
{
"error": {
"code": "500",
"message": "An error occurred during processing this request. Use this request id '2e416ae7-fbc0-4dd8-a8e5-4ea0ad843c11' for follow-up."
}
}
```
Would it be anything wrong that I'm doing?

|
1.0
|
HTTP 500 - Cost Management API query usage - I am running the query usage example from this link:
https://docs.microsoft.com/en-us/rest/api/cost-management/query/usage
With the following URL:
```
https://management.azure.com/subscriptions/{subscription}/resourceGroups/{group}/providers/Microsoft.CostManagement/query?api-version=2019-11-01
```
And this body payload:
```
{
timeframe: "MonthToDate"
type: "Usage"
}
```
But I'm getting HTTP 500 error:
```
{
"error": {
"code": "500",
"message": "An error occurred during processing this request. Use this request id '2e416ae7-fbc0-4dd8-a8e5-4ea0ad843c11' for follow-up."
}
}
```
Would it be anything wrong that I'm doing?

|
non_process
|
http cost management api query usage i am running the query usage example from this link with the following url and this body payload timeframe monthtodate type usage but i m getting http error error code message an error occurred during processing this request use this request id for follow up would it be anything wrong that i m doing
| 0
|
345,483
| 24,861,815,223
|
IssuesEvent
|
2022-10-27 08:52:02
|
facebook/docusaurus
|
https://api.github.com/repos/facebook/docusaurus
|
opened
|
Plugin-content-pages preset options, 'routeBasePath' can't be ''
|
documentation status: needs triage
|
### Have you read the Contributing Guidelines on issues?
- [X] I have read the [Contributing Guidelines on issues](https://github.com/facebook/docusaurus/blob/main/CONTRIBUTING.md#reporting-new-issues).
### Description
When I use default preset options
```shell
module.exports = {
presets: [
[
'@docusaurus/preset-classic',
{
pages: {
path: 'src/pages',
routeBasePath: '',
include: ['**/*.{js,jsx,ts,tsx,md,mdx}'],
exclude: [
'**/_*.{js,jsx,ts,tsx,md,mdx}',
'**/_*/**',
'**/*.test.{js,jsx,ts,tsx}',
'**/__tests__/**',
],
mdxPageComponent: '@theme/MDXPage',
remarkPlugins: [require('remark-math')],
rehypePlugins: [],
beforeDefaultRemarkPlugins: [],
beforeDefaultRehypePlugins: [],
},
},
],
],
};
```
It pointed out that `ValidationError: "routeBasePath" is not allowed to be empty`

### Self-service
- [X] I'd be willing to address this documentation request myself.
|
1.0
|
Plugin-content-pages preset options, 'routeBasePath' can't be '' - ### Have you read the Contributing Guidelines on issues?
- [X] I have read the [Contributing Guidelines on issues](https://github.com/facebook/docusaurus/blob/main/CONTRIBUTING.md#reporting-new-issues).
### Description
When I use default preset options
```shell
module.exports = {
presets: [
[
'@docusaurus/preset-classic',
{
pages: {
path: 'src/pages',
routeBasePath: '',
include: ['**/*.{js,jsx,ts,tsx,md,mdx}'],
exclude: [
'**/_*.{js,jsx,ts,tsx,md,mdx}',
'**/_*/**',
'**/*.test.{js,jsx,ts,tsx}',
'**/__tests__/**',
],
mdxPageComponent: '@theme/MDXPage',
remarkPlugins: [require('remark-math')],
rehypePlugins: [],
beforeDefaultRemarkPlugins: [],
beforeDefaultRehypePlugins: [],
},
},
],
],
};
```
It pointed out that `ValidationError: "routeBasePath" is not allowed to be empty`

### Self-service
- [X] I'd be willing to address this documentation request myself.
|
non_process
|
plugin content pages preset options routebasepath can t be have you read the contributing guidelines on issues i have read the description when i use default preset options shell module exports presets docusaurus preset classic pages path src pages routebasepath include exclude js jsx ts tsx md mdx test js jsx ts tsx tests mdxpagecomponent theme mdxpage remarkplugins rehypeplugins beforedefaultremarkplugins beforedefaultrehypeplugins it pointed out that validationerror routebasepath is not allowed to be empty self service i d be willing to address this documentation request myself
| 0
|
21,783
| 30,294,985,929
|
IssuesEvent
|
2023-07-09 18:45:49
|
The-Data-Alchemists-Manipal/MindWave
|
https://api.github.com/repos/The-Data-Alchemists-Manipal/MindWave
|
closed
|
Add Filter Application using Flask and OpenCV.
|
image-processing
|
### Is your feature request related to a problem? Please describe.
In this project, I have used Flask and OpenCV to add filters to the frames of the video.
This project combines the power of Flask and OpenCV to create a user-friendly application where users can apply various filters to video frames. It provides an interactive and intuitive interface for experimenting with different filters and enhancing the visual appeal of images.
### Describe the solution you'd like
The UI shows up the video and several filter options which can be applied to the video frame by the user.
### Describe alternatives you've considered
_No response_
### Additional context
_No response_
### Code of Conduct
- [X] I agree to follow this project's Code of Conduct
|
1.0
|
Add Filter Application using Flask and OpenCV. - ### Is your feature request related to a problem? Please describe.
In this project, I have used Flask and OpenCV to add filters to the frames of the video.
This project combines the power of Flask and OpenCV to create a user-friendly application where users can apply various filters to video frames. It provides an interactive and intuitive interface for experimenting with different filters and enhancing the visual appeal of images.
### Describe the solution you'd like
The UI shows up the video and several filter options which can be applied to the video frame by the user.
### Describe alternatives you've considered
_No response_
### Additional context
_No response_
### Code of Conduct
- [X] I agree to follow this project's Code of Conduct
|
process
|
add filter application using flask and opencv is your feature request related to a problem please describe in this project i have used flask and opencv to add filters to the frames of the video this project combines the power of flask and opencv to create a user friendly application where users can apply various filters to video frames it provides an interactive and intuitive interface for experimenting with different filters and enhancing the visual appeal of images describe the solution you d like the ui shows up the video and several filter options which can be applied to the video frame by the user describe alternatives you ve considered no response additional context no response code of conduct i agree to follow this project s code of conduct
| 1
|
17,282
| 23,086,837,790
|
IssuesEvent
|
2022-07-26 12:12:21
|
apache/arrow-rs
|
https://api.github.com/repos/apache/arrow-rs
|
closed
|
Release Arrow `19.0.0` (next release after `18.0.0`)
|
development-process
|
* Planned Release Candidate: 2022-07-22
* Planned Release and Publish to crates.io: 2022-07-25
Items:
- [x] Update changelog and readme: https://github.com/apache/arrow-rs/pull/2120
- [x] Create release candidate https://lists.apache.org/thread/txw3hlm06qjwczcbn20mjlgp7gf178fc
- [x] Release candidate approved https://lists.apache.org/thread/3stno7fw741bpxo81s0fph80cbcymztq
- [x] Release to crates.io
- [x] Draft update to DataFusion: https://github.com/apache/arrow-datafusion/pull/2955
See full list here:
https://github.com/apache/arrow-rs/compare/18.0.0...19.0.0
https://github.com/apache/arrow-rs/issues/1952
cc @viirya @jhorstmann @tustvold @sunchao @HaoYang670
|
1.0
|
Release Arrow `19.0.0` (next release after `18.0.0`) - * Planned Release Candidate: 2022-07-22
* Planned Release and Publish to crates.io: 2022-07-25
Items:
- [x] Update changelog and readme: https://github.com/apache/arrow-rs/pull/2120
- [x] Create release candidate https://lists.apache.org/thread/txw3hlm06qjwczcbn20mjlgp7gf178fc
- [x] Release candidate approved https://lists.apache.org/thread/3stno7fw741bpxo81s0fph80cbcymztq
- [x] Release to crates.io
- [x] Draft update to DataFusion: https://github.com/apache/arrow-datafusion/pull/2955
See full list here:
https://github.com/apache/arrow-rs/compare/18.0.0...19.0.0
https://github.com/apache/arrow-rs/issues/1952
cc @viirya @jhorstmann @tustvold @sunchao @HaoYang670
|
process
|
release arrow next release after planned release candidate planned release and publish to crates io items update changelog and readme create release candidate release candidate approved release to crates io draft update to datafusion see full list here cc viirya jhorstmann tustvold sunchao
| 1
|
20,847
| 27,626,441,121
|
IssuesEvent
|
2023-03-10 07:16:25
|
bazelbuild/bazel
|
https://api.github.com/repos/bazelbuild/bazel
|
closed
|
iOS platform constraints locations
|
P2 type: support / not a bug (process) team-Configurability
|
In order to use bazel toolchains with Apple platforms, we have to define the core set of platforms it supports. This is mostly done by https://github.com/bazelbuild/bazel/blob/96c8a9073807c9e97635ddafe2ed0365a9318d6f/tools/osx/crosstool/BUILD.toolchains
The issue is that since M1 macs were released there are now overlapping OS + CPU combos, like `ios` + `arm64` that need more specific information for whether or not the build is for a physical device, or the simulator. To support this, these custom constraints exist: https://github.com/bazelbuild/apple_support/blob/2ec0ef3eb2954371596fbb1a225fb244fcde31a9/constraints/BUILD#L12-L26
Since they live in apple_support today, and bazel does not take them into account, as far as I can tell this setup isn't able to do what we want. It seems to me that we either need to move these constraints into bazel, and annotate the toolchains here, or move the toolchain definitions out of bazel and into apple_support (or another location?).
I'm hoping we can use this issue to decide on this, which will help make some more progress on toolchain support for Apple platforms.
|
1.0
|
iOS platform constraints locations - In order to use bazel toolchains with Apple platforms, we have to define the core set of platforms it supports. This is mostly done by https://github.com/bazelbuild/bazel/blob/96c8a9073807c9e97635ddafe2ed0365a9318d6f/tools/osx/crosstool/BUILD.toolchains
The issue is that since M1 macs were released there are now overlapping OS + CPU combos, like `ios` + `arm64` that need more specific information for whether or not the build is for a physical device, or the simulator. To support this, these custom constraints exist: https://github.com/bazelbuild/apple_support/blob/2ec0ef3eb2954371596fbb1a225fb244fcde31a9/constraints/BUILD#L12-L26
Since they live in apple_support today, and bazel does not take them into account, as far as I can tell this setup isn't able to do what we want. It seems to me that we either need to move these constraints into bazel, and annotate the toolchains here, or move the toolchain definitions out of bazel and into apple_support (or another location?).
I'm hoping we can use this issue to decide on this, which will help make some more progress on toolchain support for Apple platforms.
|
process
|
ios platform constraints locations in order to use bazel toolchains with apple platforms we have to define the core set of platforms it supports this is mostly done by the issue is that since macs were released there are now overlapping os cpu combos like ios that need more specific information for whether or not the build is for a physical device or the simulator to support this these custom constraints exist since they live in apple support today and bazel does not take them into account as far as i can tell this setup isn t able to do what we want it seems to me that we either need to move these constraints into bazel and annotate the toolchains here or move the toolchain definitions out of bazel and into apple support or another location i m hoping we can use this issue to decide on this which will help make some more progress on toolchain support for apple platforms
| 1
|
18,997
| 24,993,561,547
|
IssuesEvent
|
2022-11-02 21:08:12
|
FreeCAD/FreeCAD
|
https://api.github.com/repos/FreeCAD/FreeCAD
|
closed
|
[Bug] FreeCAD Contribution policy and process is confusing.
|
Awaiting feedback Process
|
### Forums discussion
https://forum.freecadweb.org/viewtopic.php?f=10&t=72769
### Version
This issue is organizational and affects all versions and the general development process
### Issue description
The development process used by the FreeCAD project is not well defined. It's unclear to contributors how code contributions are to be structured, submitted, evaluated, and accepted. This can lead to several undesirable outcomes:
- A contributor spends considerable effort on a feature and then the feature is not accepted.
- Multiple conflicting solutions are implemented resulting in a confusing user experience
- Code contributions languish because no one knows who is responsible for accepting them.
- Potential contributors do not contribute and existing contributors leave the project in frustration.
A process should be established that clearly documents how code contributions will be handled. The process should be fair and transparent. The documented process should be put under git version control like any other part of the source so that it can be referenced, changed, and tracked over time.
### Anything else?
Much discussion has been had about the C4 process from the ZeroMQ project.
https://rfc.zeromq.org/spec/42/
When a process is accepted, the wiki including https://wiki.freecadweb.org/Developer_hub should be reviewed and revised to reflect the process.
|
1.0
|
[Bug] FreeCAD Contribution policy and process is confusing. - ### Forums discussion
https://forum.freecadweb.org/viewtopic.php?f=10&t=72769
### Version
This issue is organizational and affects all versions and the general development process
### Issue description
The development process used by the FreeCAD project is not well defined. It's unclear to contributors how code contributions are to be structured, submitted, evaluated, and accepted. This can lead to several undesirable outcomes:
- A contributor spends considerable effort on a feature and then the feature is not accepted.
- Multiple conflicting solutions are implemented resulting in a confusing user experience
- Code contributions languish because no one knows who is responsible for accepting them.
- Potential contributors do not contribute and existing contributors leave the project in frustration.
A process should be established that clearly documents how code contributions will be handled. The process should be fair and transparent. The documented process should be put under git version control like any other part of the source so that it can be referenced, changed, and tracked over time.
### Anything else?
Much discussion has been had about the C4 process from the ZeroMQ project.
https://rfc.zeromq.org/spec/42/
When a process is accepted, the wiki including https://wiki.freecadweb.org/Developer_hub should be reviewed and revised to reflect the process.
|
process
|
freecad contribution policy and process is confusing forums discussion version this issue is organizational and affects all versions and the general development process issue description the development process used by the freecad project is not well defined it s unclear to contributors how code contributions are to be structured submitted evaluated and accepted this can lead to several undesirable outcomes a contributor spends considerable effort on a feature and then the feature is not accepted multiple conflicting solutions are implemented resulting in a confusing user experience code contributions languish because no one knows who is responsible for accepting them potential contributors do not contribute and existing contributors leave the project in frustration a process should be established that clearly documents how code contributions will be handled the process should be fair and transparent the documented process should be put under git version control like any other part of the source so that it can be referenced changed and tracked over time anything else much discussion has been had about the process from the zeromq project when a process is accepted the wiki including should be reviewed and revised to reflect the process
| 1
|
369,415
| 25,844,220,453
|
IssuesEvent
|
2022-12-13 04:25:31
|
BiteSnail/ARPWithMFC
|
https://api.github.com/repos/BiteSnail/ARPWithMFC
|
closed
|
IP Layer 구현
|
documentation help wanted
|
## IP Routing table 필요 변수
`Destination IP` 4byte : unsigned char
`NetMask` 4byte : unsigned char
`GateWay` 4byte : unsigned char
`Flag` 1byte : unsigned char (bitwise 이용 1:Host, 2:Gateway, 4:Up)
`Interface` 1byte : unsigned char (index로 이용할 예정 1~255)
## 추가 함수
routing table에 아이템 추가
routing table에 아이템 삭제
... 기타 필요한 함수 추가 바람
---
만약 가능하다면 flag와 interface를 합칠 수 있을 듯.
1byte에서 왼쪽 4bit은 flag, 오른쪽 4bit은 0~15값을 가지는 interface 인덱스
| 0000| 0000 |
| -- | -- |
| flag| interface|
|
1.0
|
IP Layer 구현 - ## IP Routing table 필요 변수
`Destination IP` 4byte : unsigned char
`NetMask` 4byte : unsigned char
`GateWay` 4byte : unsigned char
`Flag` 1byte : unsigned char (bitwise 이용 1:Host, 2:Gateway, 4:Up)
`Interface` 1byte : unsigned char (index로 이용할 예정 1~255)
## 추가 함수
routing table에 아이템 추가
routing table에 아이템 삭제
... 기타 필요한 함수 추가 바람
---
만약 가능하다면 flag와 interface를 합칠 수 있을 듯.
1byte에서 왼쪽 4bit은 flag, 오른쪽 4bit은 0~15값을 가지는 interface 인덱스
| 0000| 0000 |
| -- | -- |
| flag| interface|
|
non_process
|
ip layer 구현 ip routing table 필요 변수 destination ip unsigned char netmask unsigned char gateway unsigned char flag unsigned char bitwise 이용 host gateway up interface unsigned char index로 이용할 예정 추가 함수 routing table에 아이템 추가 routing table에 아이템 삭제 기타 필요한 함수 추가 바람 만약 가능하다면 flag와 interface를 합칠 수 있을 듯 왼쪽 flag 오른쪽 가지는 interface 인덱스 flag interface
| 0
|
1,431
| 3,995,827,290
|
IssuesEvent
|
2016-05-10 16:43:09
|
PHPOffice/PHPWord
|
https://api.github.com/repos/PHPOffice/PHPWord
|
closed
|
Support arrays in TemplateProcessor.setValue method
|
Change Request Template Processor
|
To do:
- arrays in `$subject` to handle headers, footers and main document part at once.
- arrays in `$search` and in `$replacement` to avoid looping on client side.
Requested by @OAFCROB (see https://github.com/PHPOffice/PHPWord/issues/513#issuecomment-136964062 for the details).
|
1.0
|
Support arrays in TemplateProcessor.setValue method - To do:
- arrays in `$subject` to handle headers, footers and main document part at once.
- arrays in `$search` and in `$replacement` to avoid looping on client side.
Requested by @OAFCROB (see https://github.com/PHPOffice/PHPWord/issues/513#issuecomment-136964062 for the details).
|
process
|
support arrays in templateprocessor setvalue method to do arrays in subject to handle headers footers and main document part at once arrays in search and in replacement to avoid looping on client side requested by oafcrob see for the details
| 1
|
52,744
| 3,028,354,365
|
IssuesEvent
|
2015-08-04 04:15:18
|
GoogleCloudPlatform/kubernetes
|
https://api.github.com/repos/GoogleCloudPlatform/kubernetes
|
closed
|
Replacing a service fails trying to modify clusterip
|
area/usability component/kubectl priority/P2 team/CSI
|
It is confusing that this doesn't work, especially since clusterIP was not part of the initial manifest or the replacement.
```
$ cat svc.yaml
apiVersion: v1
kind: Service
metadata:
name: nginxsvc
spec:
type: NodePort
ports:
- port: 80
name: http
protocol: TCP
selector:
app: nginx
$ kubectl create -f svc.yaml
services/nginxsvc
$ cat svc.yaml
apiVersion: v1
kind: Service
metadata:
name: nginxsvc
spec:
type: NodePort
ports:
- port: 80
name: http
protocol: TCP
- port: 443
name: https
protocol: TCP
selector:
app: nginx
$ kubectl replace -f svc.yaml
Replace failedspec.clusterIP: invalid value '': field is immutable
$ kubectl delete -f svc.yaml; kubeclt create -f svc.yaml
services/nginxsvc
```
|
1.0
|
Replacing a service fails trying to modify clusterip - It is confusing that this doesn't work, especially since clusterIP was not part of the initial manifest or the replacement.
```
$ cat svc.yaml
apiVersion: v1
kind: Service
metadata:
name: nginxsvc
spec:
type: NodePort
ports:
- port: 80
name: http
protocol: TCP
selector:
app: nginx
$ kubectl create -f svc.yaml
services/nginxsvc
$ cat svc.yaml
apiVersion: v1
kind: Service
metadata:
name: nginxsvc
spec:
type: NodePort
ports:
- port: 80
name: http
protocol: TCP
- port: 443
name: https
protocol: TCP
selector:
app: nginx
$ kubectl replace -f svc.yaml
Replace failedspec.clusterIP: invalid value '': field is immutable
$ kubectl delete -f svc.yaml; kubeclt create -f svc.yaml
services/nginxsvc
```
|
non_process
|
replacing a service fails trying to modify clusterip it is confusing that this doesn t work especially since clusterip was not part of the initial manifest or the replacement cat svc yaml apiversion kind service metadata name nginxsvc spec type nodeport ports port name http protocol tcp selector app nginx kubectl create f svc yaml services nginxsvc cat svc yaml apiversion kind service metadata name nginxsvc spec type nodeport ports port name http protocol tcp port name https protocol tcp selector app nginx kubectl replace f svc yaml replace failedspec clusterip invalid value field is immutable kubectl delete f svc yaml kubeclt create f svc yaml services nginxsvc
| 0
|
283,620
| 8,721,302,906
|
IssuesEvent
|
2018-12-08 21:35:53
|
bounswe/bounswe2018group6
|
https://api.github.com/repos/bounswe/bounswe2018group6
|
closed
|
Check email server procedures
|
Back-End Priority: High Status: Blocked Type: Bug
|
It seems sign up confirmation emails are not sent. Need to be fixed.
|
1.0
|
Check email server procedures - It seems sign up confirmation emails are not sent. Need to be fixed.
|
non_process
|
check email server procedures it seems sign up confirmation emails are not sent need to be fixed
| 0
|
3,630
| 6,665,333,434
|
IssuesEvent
|
2017-10-03 00:27:43
|
IIIF/api
|
https://api.github.com/repos/IIIF/api
|
closed
|
How best to version 0.x.y to maintain history?
|
process
|
Sensibly, we don't track patch level versions and just edit in place.
However, in 0.9 for both search and auth, there were significant changes at "patch" levels which then disappear. As the process converges on a 1.0 spec, the last 0.9.y == 1.0 ... meaning there's just two copies of the same thing.
I think we should have stronger process for our 0.x.y specs, such that significant changes increment the minor version number. So we would currently be at 0.10 for auth, as we were happy with 0.9 using JSONP.
|
1.0
|
How best to version 0.x.y to maintain history? - Sensibly, we don't track patch level versions and just edit in place.
However, in 0.9 for both search and auth, there were significant changes at "patch" levels which then disappear. As the process converges on a 1.0 spec, the last 0.9.y == 1.0 ... meaning there's just two copies of the same thing.
I think we should have stronger process for our 0.x.y specs, such that significant changes increment the minor version number. So we would currently be at 0.10 for auth, as we were happy with 0.9 using JSONP.
|
process
|
how best to version x y to maintain history sensibly we don t track patch level versions and just edit in place however in for both search and auth there were significant changes at patch levels which then disappear as the process converges on a spec the last y meaning there s just two copies of the same thing i think we should have stronger process for our x y specs such that significant changes increment the minor version number so we would currently be at for auth as we were happy with using jsonp
| 1
|
4,245
| 7,187,149,121
|
IssuesEvent
|
2018-02-02 03:14:32
|
Great-Hill-Corporation/quickBlocks
|
https://api.github.com/repos/Great-Hill-Corporation/quickBlocks
|
closed
|
The 'which' parameter stored in the transaction cache is wrong
|
monitors-all status-inprocess type-bug
|
I store this, but I don't need it. I think I stored it so I wouldn't have to find it later by spinning through watches, but since I started coloring and nicknaming addresses, I spin through watches anyway. I would remove it as it makes duplicate processing in teh cacheMan incorrect and may result in double account if a transaction cache is merged with another one with a different 'which'. Only teh transaction matters. Coloring it does not need to be optimized.
|
1.0
|
The 'which' parameter stored in the transaction cache is wrong - I store this, but I don't need it. I think I stored it so I wouldn't have to find it later by spinning through watches, but since I started coloring and nicknaming addresses, I spin through watches anyway. I would remove it as it makes duplicate processing in teh cacheMan incorrect and may result in double account if a transaction cache is merged with another one with a different 'which'. Only teh transaction matters. Coloring it does not need to be optimized.
|
process
|
the which parameter stored in the transaction cache is wrong i store this but i don t need it i think i stored it so i wouldn t have to find it later by spinning through watches but since i started coloring and nicknaming addresses i spin through watches anyway i would remove it as it makes duplicate processing in teh cacheman incorrect and may result in double account if a transaction cache is merged with another one with a different which only teh transaction matters coloring it does not need to be optimized
| 1
|
87,479
| 17,273,033,504
|
IssuesEvent
|
2021-07-22 23:12:56
|
learnpack/learnpack
|
https://api.github.com/repos/learnpack/learnpack
|
opened
|
When opening one exercise with several files, only the last one gets opened
|
vscode plugin 👽
|
All this behavior is happening in `grading: incremental`
The way the the plugin works, if the exercise has 3 files to open, for example: index.html, index.js and style.css.
It will open index.html, but then when it opens index.js it will replace the same TextEditor with the content of index.js (removing the index.html that was there in the first place), and then it will do the same with style.css.
This behavior was ideal for one file exercises but long term is better to make sure that a `new` editor is opened instead of reusing the old one.
Note: when another exercise is opened (with all of its files) we need to make sure the previous files are closed to avoid overwhelming the user with too many files.
|
1.0
|
When opening one exercise with several files, only the last one gets opened - All this behavior is happening in `grading: incremental`
The way the the plugin works, if the exercise has 3 files to open, for example: index.html, index.js and style.css.
It will open index.html, but then when it opens index.js it will replace the same TextEditor with the content of index.js (removing the index.html that was there in the first place), and then it will do the same with style.css.
This behavior was ideal for one file exercises but long term is better to make sure that a `new` editor is opened instead of reusing the old one.
Note: when another exercise is opened (with all of its files) we need to make sure the previous files are closed to avoid overwhelming the user with too many files.
|
non_process
|
when opening one exercise with several files only the last one gets opened all this behavior is happening in grading incremental the way the the plugin works if the exercise has files to open for example index html index js and style css it will open index html but then when it opens index js it will replace the same texteditor with the content of index js removing the index html that was there in the first place and then it will do the same with style css this behavior was ideal for one file exercises but long term is better to make sure that a new editor is opened instead of reusing the old one note when another exercise is opened with all of its files we need to make sure the previous files are closed to avoid overwhelming the user with too many files
| 0
|
265,722
| 8,357,899,425
|
IssuesEvent
|
2018-10-02 23:34:52
|
StrangeLoopGames/EcoIssues
|
https://api.github.com/repos/StrangeLoopGames/EcoIssues
|
closed
|
Typo (need dot) in text
|
Localization Low Priority
|
> Flood fixing tool! Left click to remove the top water layer. (Target a wall touching water)
https://crowdin.com/translate/eco-by-strange-loop-games/24/en-ru#52135
maybe need dot at end? :)
@Nanders
|
1.0
|
Typo (need dot) in text - > Flood fixing tool! Left click to remove the top water layer. (Target a wall touching water)
https://crowdin.com/translate/eco-by-strange-loop-games/24/en-ru#52135
maybe need dot at end? :)
@Nanders
|
non_process
|
typo need dot in text flood fixing tool left click to remove the top water layer target a wall touching water maybe need dot at end nanders
| 0
|
18,804
| 24,703,988,100
|
IssuesEvent
|
2022-10-19 17:26:52
|
eosnetworkfoundation/devrel
|
https://api.github.com/repos/eosnetworkfoundation/devrel
|
closed
|
Define a "Definition of Done" document
|
Process
|
AC:
- brainstorm meeting, DevRel team
- Approval by stakeholder (Nathan)
- publish in DevRel repository1
|
1.0
|
Define a "Definition of Done" document - AC:
- brainstorm meeting, DevRel team
- Approval by stakeholder (Nathan)
- publish in DevRel repository1
|
process
|
define a definition of done document ac brainstorm meeting devrel team approval by stakeholder nathan publish in devrel
| 1
|
300,665
| 22,691,525,524
|
IssuesEvent
|
2022-07-04 21:12:39
|
balbasty/torch-interpol
|
https://api.github.com/repos/balbasty/torch-interpol
|
closed
|
More examples
|
documentation question
|
I need to interpolate a SDF of say 128 x 128 x 128, as well as calculate normals (gradients).
[Link](https://docs.scipy.org/doc/scipy/reference/generated/scipy.interpolate.griddata.html) to `interpolate` in `scipy`.
Could you please provide an easy-to-use example?
|
1.0
|
More examples - I need to interpolate a SDF of say 128 x 128 x 128, as well as calculate normals (gradients).
[Link](https://docs.scipy.org/doc/scipy/reference/generated/scipy.interpolate.griddata.html) to `interpolate` in `scipy`.
Could you please provide an easy-to-use example?
|
non_process
|
more examples i need to interpolate a sdf of say x x as well as calculate normals gradients to interpolate in scipy could you please provide an easy to use example
| 0
|
238
| 2,663,301,274
|
IssuesEvent
|
2015-03-20 03:29:56
|
luc-github/Repetier-Firmware-0.92
|
https://api.github.com/repos/luc-github/Repetier-Firmware-0.92
|
closed
|
G10 and G11 commands and M209, M355, M600, M360
|
enhancement Waiting to be processed
|
these commands are now available in repetier, need to importe them
|
1.0
|
G10 and G11 commands and M209, M355, M600, M360 - these commands are now available in repetier, need to importe them
|
process
|
and commands and these commands are now available in repetier need to importe them
| 1
|
10,511
| 13,283,605,171
|
IssuesEvent
|
2020-08-24 03:48:15
|
valavila/eCommerceSite
|
https://api.github.com/repos/valavila/eCommerceSite
|
closed
|
Add CI PipeLine
|
developer process
|
Add continuous integration pipline that will check to make sure code in a pull request complies successfully
|
1.0
|
Add CI PipeLine - Add continuous integration pipline that will check to make sure code in a pull request complies successfully
|
process
|
add ci pipeline add continuous integration pipline that will check to make sure code in a pull request complies successfully
| 1
|
6,160
| 9,039,026,699
|
IssuesEvent
|
2019-02-10 01:29:18
|
material-components/material-components-ios
|
https://api.github.com/repos/material-components/material-components-ios
|
closed
|
[BottomNavigation] Finalize BottomNavigationController API
|
[BottomNavigation] type:Process
|
This was filed as an internal issue. If you are a Googler, please visit [b/119188596](http://b/119188596) for more details.
<!-- Auto-generated content below, do not modify -->
---
#### Internal data
- Associated internal bug: [b/119188596](http://b/119188596)
|
1.0
|
[BottomNavigation] Finalize BottomNavigationController API - This was filed as an internal issue. If you are a Googler, please visit [b/119188596](http://b/119188596) for more details.
<!-- Auto-generated content below, do not modify -->
---
#### Internal data
- Associated internal bug: [b/119188596](http://b/119188596)
|
process
|
finalize bottomnavigationcontroller api this was filed as an internal issue if you are a googler please visit for more details internal data associated internal bug
| 1
|
15,412
| 19,599,552,476
|
IssuesEvent
|
2022-01-05 22:37:32
|
googleapis/gapic-generator-java
|
https://api.github.com/repos/googleapis/gapic-generator-java
|
opened
|
Run tests for "self-service" libraries in CI
|
type: process priority: p2
|
Doing so will help validating changes to the "self-service" library generation workflow (e.g., as in #876).
Theoretically, we should be able to generate some libraries against actual API service protos like the following (may need to change a few things in `WORKSPACE` or some `BUILD` files).
```
bazel build @com_google_googleapis//google/cloud/compute/v1:google-cloud-compute-v1-java
```
Then unpack `bazel-bin/external/com_google_googleapis/google/cloud/compute/v1/google-cloud-compute-v1-java.tar.gz` and run `./gradlew clean check publishToMavenLocal`.
`compute` is REST and all others are gRPC at the moment, so `compute` and another library are good candidates.
|
1.0
|
Run tests for "self-service" libraries in CI - Doing so will help validating changes to the "self-service" library generation workflow (e.g., as in #876).
Theoretically, we should be able to generate some libraries against actual API service protos like the following (may need to change a few things in `WORKSPACE` or some `BUILD` files).
```
bazel build @com_google_googleapis//google/cloud/compute/v1:google-cloud-compute-v1-java
```
Then unpack `bazel-bin/external/com_google_googleapis/google/cloud/compute/v1/google-cloud-compute-v1-java.tar.gz` and run `./gradlew clean check publishToMavenLocal`.
`compute` is REST and all others are gRPC at the moment, so `compute` and another library are good candidates.
|
process
|
run tests for self service libraries in ci doing so will help validating changes to the self service library generation workflow e g as in theoretically we should be able to generate some libraries against actual api service protos like the following may need to change a few things in workspace or some build files bazel build com google googleapis google cloud compute google cloud compute java then unpack bazel bin external com google googleapis google cloud compute google cloud compute java tar gz and run gradlew clean check publishtomavenlocal compute is rest and all others are grpc at the moment so compute and another library are good candidates
| 1
|
39,834
| 16,102,615,605
|
IssuesEvent
|
2021-04-27 11:18:25
|
microsoft/vscode-cpptools
|
https://api.github.com/repos/microsoft/vscode-cpptools
|
closed
|
Extension causes high cpu load
|
Language Service more info needed
|
- Issue Type: `Performance`
- Extension Name: `cpptools`
- Extension Version: `1.2.2-insiders2`
- OS Version: `Windows_NT x64 10.0.18362`
- VSCode version: `1.53.2`
:warning: Make sure to **attach** this file
[ms-vscode.cpptools-unresponsive.cpuprofile.txt](https://github.com/microsoft/vscode-cpptools/files/6026948/ms-vscode.cpptools-unresponsive.cpuprofile.txt)
from your *home*-directory:
:warning:`c:\Users\*******\AppData\Local\Temp\ms-vscode.cpptools-unresponsive.cpuprofile.txt`
Find more details here: https://github.com/microsoft/vscode/wiki/Explain-extension-causes-high-cpu-load
|
1.0
|
Extension causes high cpu load - - Issue Type: `Performance`
- Extension Name: `cpptools`
- Extension Version: `1.2.2-insiders2`
- OS Version: `Windows_NT x64 10.0.18362`
- VSCode version: `1.53.2`
:warning: Make sure to **attach** this file
[ms-vscode.cpptools-unresponsive.cpuprofile.txt](https://github.com/microsoft/vscode-cpptools/files/6026948/ms-vscode.cpptools-unresponsive.cpuprofile.txt)
from your *home*-directory:
:warning:`c:\Users\*******\AppData\Local\Temp\ms-vscode.cpptools-unresponsive.cpuprofile.txt`
Find more details here: https://github.com/microsoft/vscode/wiki/Explain-extension-causes-high-cpu-load
|
non_process
|
extension causes high cpu load issue type performance extension name cpptools extension version os version windows nt vscode version warning make sure to attach this file from your home directory warning c users appdata local temp ms vscode cpptools unresponsive cpuprofile txt find more details here
| 0
|
14,259
| 17,192,791,672
|
IssuesEvent
|
2021-07-16 13:24:11
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
Managed identity is generally available, please remove "Preview"
|
Pri2 automation/svc cxp doc-bug doc-enhancement process-automation/subsvc triaged
|
https://docs.microsoft.com/en-us/azure/automation/automation-security-overview - says MI is in preview.
I checked in portal and there is no preview warning when I enabled this feature.
https://azure.microsoft.com/en-au/updates/azure-automation-system-assigned-managed-identities/ -- for reference
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 8721e209-24ce-2170-6caa-ed12a7060080
* Version Independent ID: ac13f91d-460c-cbe9-4778-50d20765b252
* Content: [Azure Automation account authentication overview](https://docs.microsoft.com/en-us/azure/automation/automation-security-overview)
* Content Source: [articles/automation/automation-security-overview.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/automation-security-overview.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @MGoedtel
* Microsoft Alias: **magoedte**
|
1.0
|
Managed identity is generally available, please remove "Preview" - https://docs.microsoft.com/en-us/azure/automation/automation-security-overview - says MI is in preview.
I checked in portal and there is no preview warning when I enabled this feature.
https://azure.microsoft.com/en-au/updates/azure-automation-system-assigned-managed-identities/ -- for reference
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 8721e209-24ce-2170-6caa-ed12a7060080
* Version Independent ID: ac13f91d-460c-cbe9-4778-50d20765b252
* Content: [Azure Automation account authentication overview](https://docs.microsoft.com/en-us/azure/automation/automation-security-overview)
* Content Source: [articles/automation/automation-security-overview.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/automation-security-overview.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @MGoedtel
* Microsoft Alias: **magoedte**
|
process
|
managed identity is generally available please remove preview says mi is in preview i checked in portal and there is no preview warning when i enabled this feature for reference document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login mgoedtel microsoft alias magoedte
| 1
|
7,678
| 10,762,210,840
|
IssuesEvent
|
2019-10-31 22:52:56
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
opened
|
Processing "output type" should default to "auto"
|
Bug Processing
|
In QGIS 3/Processing/GRASS the parameter "output type" has no selection by default and is mandatory.
This is *very* inconvenient, up to the old LTR this parameter defaulted to "auto", which is the right choice.
Moreover there are Processing/GRASS modules where the output type is hardcoded and this parameter does not make sense anyway, i.e.
https://github.com/qgis/QGIS/blob/master/python/plugins/processing/algs/grass7/ext/v_net_centrality.py
using "auto" as default would also help create less confusion with this parameters.
|
1.0
|
Processing "output type" should default to "auto" - In QGIS 3/Processing/GRASS the parameter "output type" has no selection by default and is mandatory.
This is *very* inconvenient, up to the old LTR this parameter defaulted to "auto", which is the right choice.
Moreover there are Processing/GRASS modules where the output type is hardcoded and this parameter does not make sense anyway, i.e.
https://github.com/qgis/QGIS/blob/master/python/plugins/processing/algs/grass7/ext/v_net_centrality.py
using "auto" as default would also help create less confusion with this parameters.
|
process
|
processing output type should default to auto in qgis processing grass the parameter output type has no selection by default and is mandatory this is very inconvenient up to the old ltr this parameter defaulted to auto which is the right choice moreover there are processing grass modules where the output type is hardcoded and this parameter does not make sense anyway i e using auto as default would also help create less confusion with this parameters
| 1
|
14,400
| 17,455,867,444
|
IssuesEvent
|
2021-08-06 01:01:38
|
marysaray/eCommerceWebsiteExample
|
https://api.github.com/repos/marysaray/eCommerceWebsiteExample
|
closed
|
Add CI Pipeline
|
development process
|
Add continuous integration pipeline that will check and make sure the code in a pull request compiles successfully.
|
1.0
|
Add CI Pipeline - Add continuous integration pipeline that will check and make sure the code in a pull request compiles successfully.
|
process
|
add ci pipeline add continuous integration pipeline that will check and make sure the code in a pull request compiles successfully
| 1
|
10,332
| 13,162,980,294
|
IssuesEvent
|
2020-08-10 22:59:17
|
googleapis/google-cloud-ruby
|
https://api.github.com/repos/googleapis/google-cloud-ruby
|
closed
|
Migrate google-cloud-debugger to the microgenerator
|
type: process
|
Migrate google-cloud-debugger to the microgenerator. This involves the following steps:
* [x] Write synth file and generate `google-cloud-debugger-v2`
* [x] Make sure the new libraries are configured in kokoro
* [x] Release `google-cloud-debugger-v2`
* [ ] Switch `google-cloud-debugger` backend to the versioned gems. That is:
* Rip out synth and all the generated code
* Add `google-cloud-debugger-v2` as a dependency
* Update the veneer code to the microgenerator usage
* [ ] Release `google-cloud-debugger` update. Note that it should remain pre-1.0. We are not GA-ing this library.
I do not believe samples need to be updated, unless they invoke the low-level interface directly.
|
1.0
|
Migrate google-cloud-debugger to the microgenerator - Migrate google-cloud-debugger to the microgenerator. This involves the following steps:
* [x] Write synth file and generate `google-cloud-debugger-v2`
* [x] Make sure the new libraries are configured in kokoro
* [x] Release `google-cloud-debugger-v2`
* [ ] Switch `google-cloud-debugger` backend to the versioned gems. That is:
* Rip out synth and all the generated code
* Add `google-cloud-debugger-v2` as a dependency
* Update the veneer code to the microgenerator usage
* [ ] Release `google-cloud-debugger` update. Note that it should remain pre-1.0. We are not GA-ing this library.
I do not believe samples need to be updated, unless they invoke the low-level interface directly.
|
process
|
migrate google cloud debugger to the microgenerator migrate google cloud debugger to the microgenerator this involves the following steps write synth file and generate google cloud debugger make sure the new libraries are configured in kokoro release google cloud debugger switch google cloud debugger backend to the versioned gems that is rip out synth and all the generated code add google cloud debugger as a dependency update the veneer code to the microgenerator usage release google cloud debugger update note that it should remain pre we are not ga ing this library i do not believe samples need to be updated unless they invoke the low level interface directly
| 1
|
335,457
| 24,468,926,815
|
IssuesEvent
|
2022-10-07 17:42:21
|
EnceeFinancials/XYO-NODE
|
https://api.github.com/repos/EnceeFinancials/XYO-NODE
|
closed
|
[DOCUMENTATION]:
|
documentation help wanted
|
**Is your documentation request related to a problem? Please describe.**
A clear and concise description of what the problem is. Ex. I'm always frustrated when [...]
**Describe the documentation and format you would like**
A clear and concise description of what documentation you would like to see and what type for format.
Ex. Step-by-step, Paragraph explainer, screenshots, etc.
**Additional context**
Add any other context or screenshots about the document request here.
|
1.0
|
[DOCUMENTATION]: - **Is your documentation request related to a problem? Please describe.**
A clear and concise description of what the problem is. Ex. I'm always frustrated when [...]
**Describe the documentation and format you would like**
A clear and concise description of what documentation you would like to see and what type for format.
Ex. Step-by-step, Paragraph explainer, screenshots, etc.
**Additional context**
Add any other context or screenshots about the document request here.
|
non_process
|
is your documentation request related to a problem please describe a clear and concise description of what the problem is ex i m always frustrated when describe the documentation and format you would like a clear and concise description of what documentation you would like to see and what type for format ex step by step paragraph explainer screenshots etc additional context add any other context or screenshots about the document request here
| 0
|
326,714
| 24,098,467,673
|
IssuesEvent
|
2022-09-19 21:09:46
|
Open-Telecoms-Data/open-fibre-data-standard
|
https://api.github.com/repos/Open-Telecoms-Data/open-fibre-data-standard
|
closed
|
Draft documentation on 'what does OFDS data look like'
|
Non-normative documentation
|
In the primer documentation on the [Open Fibre Data Standard](https://github.com/Open-Telecoms-Data/open-fibre-data-standard/blob/main/docs/primer/openfibredatastandard.md) the section on 'what does OFDS data look like is not yet drafted. This should be completed once decisions on publication formats are finalised.
|
1.0
|
Draft documentation on 'what does OFDS data look like' - In the primer documentation on the [Open Fibre Data Standard](https://github.com/Open-Telecoms-Data/open-fibre-data-standard/blob/main/docs/primer/openfibredatastandard.md) the section on 'what does OFDS data look like is not yet drafted. This should be completed once decisions on publication formats are finalised.
|
non_process
|
draft documentation on what does ofds data look like in the primer documentation on the the section on what does ofds data look like is not yet drafted this should be completed once decisions on publication formats are finalised
| 0
|
5,372
| 8,202,496,346
|
IssuesEvent
|
2018-09-02 10:08:13
|
linnovate/root
|
https://api.github.com/repos/linnovate/root
|
closed
|
Folders : can't download file from Attachment Preview
|
Process bug
|
@abrahamos
upload file in folder.
click on the attachment preview.
click on download.
get a Restricted accses error.
|
1.0
|
Folders : can't download file from Attachment Preview - @abrahamos
upload file in folder.
click on the attachment preview.
click on download.
get a Restricted accses error.
|
process
|
folders can t download file from attachment preview abrahamos upload file in folder click on the attachment preview click on download get a restricted accses error
| 1
|
8,264
| 11,426,558,101
|
IssuesEvent
|
2020-02-03 22:13:05
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
"Random points along line" algorithm execution fails every two calls
|
Bug Feedback Processing
|
I ran the alg multiple times with the same parameters and concluded that it works every two calls and fails as well every two calls. The error log:
> QGIS version: 3.11.0-Master
QGIS code revision: c52a6e711a
Qt version: 5.11.2
GDAL version: 3.1.0dev
GEOS version: 3.8.0-CAPI-1.13.1
PROJ version: Rel. 7.0.0, March 1st, 2020
Processing algorithm…
Algorithm 'Random points along line' starting…
Input parameters:
{ 'INPUT' : '//serveur//Études/doss/candidatures.gpkg|layername=A_laisons_douces', 'MIN_DISTANCE' : 0, 'OUTPUT' : 'TEMPORARY_OUTPUT', 'POINTS_NUMBER' : 5 }
>
>Traceback (most recent call last):
File "C:/OSGEO4~1/apps/qgis-dev/./python/plugins\processing\algs\qgis\RandomPointsAlongLines.py", line 126, in processAlgorithm
f = next(source.getFeatures(request.setFilterFid(fid).setSubsetOfAttributes([])))
StopIteration
>
>Execution failed after 0.16 seconds
>
>Loading resulting layers
Algorithm 'Random points along line' finished
|
1.0
|
"Random points along line" algorithm execution fails every two calls - I ran the alg multiple times with the same parameters and concluded that it works every two calls and fails as well every two calls. The error log:
> QGIS version: 3.11.0-Master
QGIS code revision: c52a6e711a
Qt version: 5.11.2
GDAL version: 3.1.0dev
GEOS version: 3.8.0-CAPI-1.13.1
PROJ version: Rel. 7.0.0, March 1st, 2020
Processing algorithm…
Algorithm 'Random points along line' starting…
Input parameters:
{ 'INPUT' : '//serveur//Études/doss/candidatures.gpkg|layername=A_laisons_douces', 'MIN_DISTANCE' : 0, 'OUTPUT' : 'TEMPORARY_OUTPUT', 'POINTS_NUMBER' : 5 }
>
>Traceback (most recent call last):
File "C:/OSGEO4~1/apps/qgis-dev/./python/plugins\processing\algs\qgis\RandomPointsAlongLines.py", line 126, in processAlgorithm
f = next(source.getFeatures(request.setFilterFid(fid).setSubsetOfAttributes([])))
StopIteration
>
>Execution failed after 0.16 seconds
>
>Loading resulting layers
Algorithm 'Random points along line' finished
|
process
|
random points along line algorithm execution fails every two calls i ran the alg multiple times with the same parameters and concluded that it works every two calls and fails as well every two calls the error log qgis version master qgis code revision qt version gdal version geos version capi proj version rel march processing algorithm… algorithm random points along line starting… input parameters input serveur études doss candidatures gpkg layername a laisons douces min distance output temporary output points number traceback most recent call last file c apps qgis dev python plugins processing algs qgis randompointsalonglines py line in processalgorithm f next source getfeatures request setfilterfid fid setsubsetofattributes stopiteration execution failed after seconds loading resulting layers algorithm random points along line finished
| 1
|
4,798
| 7,695,110,669
|
IssuesEvent
|
2018-05-18 11:03:25
|
zero-os/0-robot
|
https://api.github.com/repos/zero-os/0-robot
|
closed
|
robot sometimes fail to start from flist
|
process_duplicate type_bug
|
I've seen this happening a few time on some node running the robot in a container
used flist is `https://hub.gig.tech/gig-autobuilder/zero-os-0-robot-autostart-0.5.1.flist`
```
In [1264]: sub.stream()
[Thu12 14:30] - GitFactory.py :316 :j.clients.git - INFO - None:pull:http://github.com/zero-os/0-templates.git ->/opt/code/github/zero-os/0-templates
[Thu12 14:30] - GitFactory.py :399 :j.clients.git - INFO - git clone http://github.com/zero-os/0-templates.git -> /opt/code/github/zero-os/0-templates
[Thu12 14:30] - GitFactory.py :419 :j.clients.git - INFO - mkdir -p /opt/code/github/zero-os;cd /opt/code/github/zero-os;git -c http.sslVerify=false clone -b master http://github.com/zero-os/0-templates.git /opt/code/github/zero-os/0-templates
Traceback (most recent call last):
File "/usr/local/bin/zrobot", line 6, in <module>
exec(compile(open(__file__).read(), __file__, 'exec'))
File "/opt/code/github/zero-os/0-robot/cmd/zrobot", line 25, in <module>
entry_point()
File "/opt/code/github/zero-os/0-robot/zerorobot/cli/server.py", line 66, in start
robot.set_config_repo(config_repo, config_key)
File "/opt/code/github/zero-os/0-robot/zerorobot/robot/robot.py", line 82, in set_config_repo
config_repo.init(path, key)
File "/opt/code/github/zero-os/0-robot/zerorobot/robot/config_repo.py", line 54, in init
j.tools.configmanager.init(silent=True, configpath=path, keypath=key)
File "/usr/local/lib/python3.5/dist-packages/JumpScale9/tools/configmanager/ConfigManager.py", line 432, in init
j.clients.sshkey.key_get(keypath, load=True)
File "/usr/local/lib/python3.5/dist-packages/JumpScale9/clients/sshkey/SSHKeys.py", line 25, in key_get
sshkey = self.get(instance)
File "/usr/local/lib/python3.5/dist-packages/JumpScale9/tools/configmanager/JSBaseClassConfigs.py", line 43, in get
return self._child_class(instance=instance, data=data, parent=self, interactive=interactive,**kwargs)
File "/usr/local/lib/python3.5/dist-packages/JumpScale9/clients/sshkey/SSHKey.py", line 23, in __init__
parent=parent, template=TEMPLATE, interactive=interactive)
File "/usr/local/lib/python3.5/dist-packages/JumpScale9/tools/configmanager/JSBaseClassConfig.py", line 16, in __init__
self._ui = j.tools.formbuilder.baseclass_get() # is the default class
File "/usr/lib/python3/dist-packages/js9.py", line 1709, in formbuilder
from JumpScale9.tools.formbuilder.FormBuilder import FormBuilderFactory as FormBuilderFactory
File "/usr/local/lib/python3.5/dist-packages/JumpScale9/tools/formbuilder/FormBuilder.py", line 3, in <module>
import npyscreen
File "/usr/local/lib/python3.5/dist-packages/npyscreen/__init__.py", line 5, in <module>
from .wgwidget import TEST_SETTINGS, ExhaustedTestInput, add_test_input_from_iterable, add_test_input_ch
File "/usr/local/lib/python3.5/dist-packages/npyscreen/wgwidget.py", line 5, in <module>
import curses
File "/usr/lib/python3.5/curses/__init__.py", line 13, in <module>
from _curses import *
ImportError: /lib/x86_64-linux-gnu/libncursesw.so.5: symbol LINES, version NCURSES_TINFO_5.0.19991023 not defined in file libtinfo.so.5 with link time reference
[Thu12 14:30] - ErrorHandler.py :257 :j.core.errorhandler - ERROR - /lib/x86_64-linux-gnu/libncursesw.so.5: symbol LINES, version NCURSES_TINFO_5.0.19991023 not defined in file libtinfo.so.5 with link time reference
```
|
1.0
|
robot sometimes fail to start from flist - I've seen this happening a few time on some node running the robot in a container
used flist is `https://hub.gig.tech/gig-autobuilder/zero-os-0-robot-autostart-0.5.1.flist`
```
In [1264]: sub.stream()
[Thu12 14:30] - GitFactory.py :316 :j.clients.git - INFO - None:pull:http://github.com/zero-os/0-templates.git ->/opt/code/github/zero-os/0-templates
[Thu12 14:30] - GitFactory.py :399 :j.clients.git - INFO - git clone http://github.com/zero-os/0-templates.git -> /opt/code/github/zero-os/0-templates
[Thu12 14:30] - GitFactory.py :419 :j.clients.git - INFO - mkdir -p /opt/code/github/zero-os;cd /opt/code/github/zero-os;git -c http.sslVerify=false clone -b master http://github.com/zero-os/0-templates.git /opt/code/github/zero-os/0-templates
Traceback (most recent call last):
File "/usr/local/bin/zrobot", line 6, in <module>
exec(compile(open(__file__).read(), __file__, 'exec'))
File "/opt/code/github/zero-os/0-robot/cmd/zrobot", line 25, in <module>
entry_point()
File "/opt/code/github/zero-os/0-robot/zerorobot/cli/server.py", line 66, in start
robot.set_config_repo(config_repo, config_key)
File "/opt/code/github/zero-os/0-robot/zerorobot/robot/robot.py", line 82, in set_config_repo
config_repo.init(path, key)
File "/opt/code/github/zero-os/0-robot/zerorobot/robot/config_repo.py", line 54, in init
j.tools.configmanager.init(silent=True, configpath=path, keypath=key)
File "/usr/local/lib/python3.5/dist-packages/JumpScale9/tools/configmanager/ConfigManager.py", line 432, in init
j.clients.sshkey.key_get(keypath, load=True)
File "/usr/local/lib/python3.5/dist-packages/JumpScale9/clients/sshkey/SSHKeys.py", line 25, in key_get
sshkey = self.get(instance)
File "/usr/local/lib/python3.5/dist-packages/JumpScale9/tools/configmanager/JSBaseClassConfigs.py", line 43, in get
return self._child_class(instance=instance, data=data, parent=self, interactive=interactive,**kwargs)
File "/usr/local/lib/python3.5/dist-packages/JumpScale9/clients/sshkey/SSHKey.py", line 23, in __init__
parent=parent, template=TEMPLATE, interactive=interactive)
File "/usr/local/lib/python3.5/dist-packages/JumpScale9/tools/configmanager/JSBaseClassConfig.py", line 16, in __init__
self._ui = j.tools.formbuilder.baseclass_get() # is the default class
File "/usr/lib/python3/dist-packages/js9.py", line 1709, in formbuilder
from JumpScale9.tools.formbuilder.FormBuilder import FormBuilderFactory as FormBuilderFactory
File "/usr/local/lib/python3.5/dist-packages/JumpScale9/tools/formbuilder/FormBuilder.py", line 3, in <module>
import npyscreen
File "/usr/local/lib/python3.5/dist-packages/npyscreen/__init__.py", line 5, in <module>
from .wgwidget import TEST_SETTINGS, ExhaustedTestInput, add_test_input_from_iterable, add_test_input_ch
File "/usr/local/lib/python3.5/dist-packages/npyscreen/wgwidget.py", line 5, in <module>
import curses
File "/usr/lib/python3.5/curses/__init__.py", line 13, in <module>
from _curses import *
ImportError: /lib/x86_64-linux-gnu/libncursesw.so.5: symbol LINES, version NCURSES_TINFO_5.0.19991023 not defined in file libtinfo.so.5 with link time reference
[Thu12 14:30] - ErrorHandler.py :257 :j.core.errorhandler - ERROR - /lib/x86_64-linux-gnu/libncursesw.so.5: symbol LINES, version NCURSES_TINFO_5.0.19991023 not defined in file libtinfo.so.5 with link time reference
```
|
process
|
robot sometimes fail to start from flist i ve seen this happening a few time on some node running the robot in a container used flist is in sub stream gitfactory py j clients git info none pull opt code github zero os templates gitfactory py j clients git info git clone opt code github zero os templates gitfactory py j clients git info mkdir p opt code github zero os cd opt code github zero os git c http sslverify false clone b master opt code github zero os templates traceback most recent call last file usr local bin zrobot line in exec compile open file read file exec file opt code github zero os robot cmd zrobot line in entry point file opt code github zero os robot zerorobot cli server py line in start robot set config repo config repo config key file opt code github zero os robot zerorobot robot robot py line in set config repo config repo init path key file opt code github zero os robot zerorobot robot config repo py line in init j tools configmanager init silent true configpath path keypath key file usr local lib dist packages tools configmanager configmanager py line in init j clients sshkey key get keypath load true file usr local lib dist packages clients sshkey sshkeys py line in key get sshkey self get instance file usr local lib dist packages tools configmanager jsbaseclassconfigs py line in get return self child class instance instance data data parent self interactive interactive kwargs file usr local lib dist packages clients sshkey sshkey py line in init parent parent template template interactive interactive file usr local lib dist packages tools configmanager jsbaseclassconfig py line in init self ui j tools formbuilder baseclass get is the default class file usr lib dist packages py line in formbuilder from tools formbuilder formbuilder import formbuilderfactory as formbuilderfactory file usr local lib dist packages tools formbuilder formbuilder py line in import npyscreen file usr local lib dist packages npyscreen init py line in from wgwidget import test settings exhaustedtestinput add test input from iterable add test input ch file usr local lib dist packages npyscreen wgwidget py line in import curses file usr lib curses init py line in from curses import importerror lib linux gnu libncursesw so symbol lines version ncurses tinfo not defined in file libtinfo so with link time reference errorhandler py j core errorhandler error lib linux gnu libncursesw so symbol lines version ncurses tinfo not defined in file libtinfo so with link time reference
| 1
|
2,081
| 4,896,463,930
|
IssuesEvent
|
2016-11-20 11:14:24
|
Alfresco/alfresco-ng2-components
|
https://api.github.com/repos/Alfresco/alfresco-ng2-components
|
closed
|
Can navigate around activit app while start process dialog is present
|
browser: firefox browser: safari bug comp: activiti-processList
|
1. Click start process
2. Navigate around Activit app
3. Navigate back to process list
**Expected result**
While dialog is present cannot navigate around app (behaviour in Chrome)
**Actual result**
While dialog is present can navigate around app (behaviour in Firefox and Safari)
|
1.0
|
Can navigate around activit app while start process dialog is present - 1. Click start process
2. Navigate around Activit app
3. Navigate back to process list
**Expected result**
While dialog is present cannot navigate around app (behaviour in Chrome)
**Actual result**
While dialog is present can navigate around app (behaviour in Firefox and Safari)
|
process
|
can navigate around activit app while start process dialog is present click start process navigate around activit app navigate back to process list expected result while dialog is present cannot navigate around app behaviour in chrome actual result while dialog is present can navigate around app behaviour in firefox and safari
| 1
|
47,690
| 13,066,100,017
|
IssuesEvent
|
2020-07-30 20:59:50
|
icecube-trac/tix2
|
https://api.github.com/repos/icecube-trac/tix2
|
closed
|
sim-services - missing sphinx docs (Trac #996)
|
Migrated from Trac combo simulation defect
|
`sim-services/resources/docs/sanity_checkers/sanity_checkers.rst` isn't picked up by sphinx on `make docs`
Migrated from https://code.icecube.wisc.edu/ticket/996
```json
{
"status": "closed",
"changetime": "2015-05-26T21:51:48",
"description": "`sim-services/resources/docs/sanity_checkers/sanity_checkers.rst` isn't picked up by sphinx on `make docs`",
"reporter": "nega",
"cc": "olivas",
"resolution": "fixed",
"_ts": "1432677108464162",
"component": "combo simulation",
"summary": "sim-services - missing sphinx docs",
"priority": "normal",
"keywords": "documentation",
"time": "2015-05-26T20:41:22",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
|
1.0
|
sim-services - missing sphinx docs (Trac #996) - `sim-services/resources/docs/sanity_checkers/sanity_checkers.rst` isn't picked up by sphinx on `make docs`
Migrated from https://code.icecube.wisc.edu/ticket/996
```json
{
"status": "closed",
"changetime": "2015-05-26T21:51:48",
"description": "`sim-services/resources/docs/sanity_checkers/sanity_checkers.rst` isn't picked up by sphinx on `make docs`",
"reporter": "nega",
"cc": "olivas",
"resolution": "fixed",
"_ts": "1432677108464162",
"component": "combo simulation",
"summary": "sim-services - missing sphinx docs",
"priority": "normal",
"keywords": "documentation",
"time": "2015-05-26T20:41:22",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
|
non_process
|
sim services missing sphinx docs trac sim services resources docs sanity checkers sanity checkers rst isn t picked up by sphinx on make docs migrated from json status closed changetime description sim services resources docs sanity checkers sanity checkers rst isn t picked up by sphinx on make docs reporter nega cc olivas resolution fixed ts component combo simulation summary sim services missing sphinx docs priority normal keywords documentation time milestone owner nega type defect
| 0
|
9,716
| 12,716,574,083
|
IssuesEvent
|
2020-06-24 02:21:24
|
OUDcollective/twenty20times
|
https://api.github.com/repos/OUDcollective/twenty20times
|
opened
|
Understanding the GitHub flow · GitHub Guides
|
workflow-process
|

## GIT FLOW
---
**Source URL**:
[https://guides.github.com/introduction/flow/](https://guides.github.com/introduction/flow/)
<table><tr><td><strong>Browser</strong></td><td>Chrome 84.0.4147.56</td></tr><tr><td><strong>OS</strong></td><td>Windows 10 64-bit</td></tr><tr><td><strong>Screen Size</strong></td><td>2560x1080</td></tr><tr><td><strong>Viewport Size</strong></td><td>2560x888</td></tr><tr><td><strong>Pixel Ratio</strong></td><td>@1x</td></tr><tr><td><strong>Zoom Level</strong></td><td>100%</td></tr></table>
|
1.0
|
Understanding the GitHub flow · GitHub Guides - 
## GIT FLOW
---
**Source URL**:
[https://guides.github.com/introduction/flow/](https://guides.github.com/introduction/flow/)
<table><tr><td><strong>Browser</strong></td><td>Chrome 84.0.4147.56</td></tr><tr><td><strong>OS</strong></td><td>Windows 10 64-bit</td></tr><tr><td><strong>Screen Size</strong></td><td>2560x1080</td></tr><tr><td><strong>Viewport Size</strong></td><td>2560x888</td></tr><tr><td><strong>Pixel Ratio</strong></td><td>@1x</td></tr><tr><td><strong>Zoom Level</strong></td><td>100%</td></tr></table>
|
process
|
understanding the github flow · github guides git flow source url browser chrome os windows bit screen size viewport size pixel ratio zoom level
| 1
|
106,587
| 16,700,442,292
|
IssuesEvent
|
2021-06-09 01:02:01
|
shaimael/cset
|
https://api.github.com/repos/shaimael/cset
|
opened
|
CVE-2021-25949 (High) detected in set-getter-0.1.0.tgz
|
security vulnerability
|
## CVE-2021-25949 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>set-getter-0.1.0.tgz</b></p></summary>
<p>Create nested getter properties and any intermediary dot notation (`'a.b.c'`) paths</p>
<p>Library home page: <a href="https://registry.npmjs.org/set-getter/-/set-getter-0.1.0.tgz">https://registry.npmjs.org/set-getter/-/set-getter-0.1.0.tgz</a></p>
<p>Path to dependency file: cset/CSETWebNg/package.json</p>
<p>Path to vulnerable library: cset/CSETWebNg/node_modules/set-getter/package.json</p>
<p>
Dependency Hierarchy:
- update-0.7.4.tgz (Root Library)
- lazy-cache-2.0.2.tgz
- :x: **set-getter-0.1.0.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Prototype pollution vulnerability in ‘set-getter’ version 0.1.0 allows an attacker to cause a denial of service and may lead to remote code execution.
<p>Publish Date: 2021-01-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-25949>CVE-2021-25949</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.2</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"set-getter","packageVersion":"0.1.0","packageFilePaths":["/CSETWebNg/package.json"],"isTransitiveDependency":true,"dependencyTree":"update:0.7.4;lazy-cache:2.0.2;set-getter:0.1.0","isMinimumFixVersionAvailable":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-25949","vulnerabilityDetails":"Prototype pollution vulnerability in ‘set-getter’ version 0.1.0 allows an attacker to cause a denial of service and may lead to remote code execution.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-25949","cvss3Severity":"high","cvss3Score":"8.2","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2021-25949 (High) detected in set-getter-0.1.0.tgz - ## CVE-2021-25949 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>set-getter-0.1.0.tgz</b></p></summary>
<p>Create nested getter properties and any intermediary dot notation (`'a.b.c'`) paths</p>
<p>Library home page: <a href="https://registry.npmjs.org/set-getter/-/set-getter-0.1.0.tgz">https://registry.npmjs.org/set-getter/-/set-getter-0.1.0.tgz</a></p>
<p>Path to dependency file: cset/CSETWebNg/package.json</p>
<p>Path to vulnerable library: cset/CSETWebNg/node_modules/set-getter/package.json</p>
<p>
Dependency Hierarchy:
- update-0.7.4.tgz (Root Library)
- lazy-cache-2.0.2.tgz
- :x: **set-getter-0.1.0.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Prototype pollution vulnerability in ‘set-getter’ version 0.1.0 allows an attacker to cause a denial of service and may lead to remote code execution.
<p>Publish Date: 2021-01-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-25949>CVE-2021-25949</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.2</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"set-getter","packageVersion":"0.1.0","packageFilePaths":["/CSETWebNg/package.json"],"isTransitiveDependency":true,"dependencyTree":"update:0.7.4;lazy-cache:2.0.2;set-getter:0.1.0","isMinimumFixVersionAvailable":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-25949","vulnerabilityDetails":"Prototype pollution vulnerability in ‘set-getter’ version 0.1.0 allows an attacker to cause a denial of service and may lead to remote code execution.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-25949","cvss3Severity":"high","cvss3Score":"8.2","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
|
non_process
|
cve high detected in set getter tgz cve high severity vulnerability vulnerable library set getter tgz create nested getter properties and any intermediary dot notation a b c paths library home page a href path to dependency file cset csetwebng package json path to vulnerable library cset csetwebng node modules set getter package json dependency hierarchy update tgz root library lazy cache tgz x set getter tgz vulnerable library found in base branch master vulnerability details prototype pollution vulnerability in ‘set getter’ version allows an attacker to cause a denial of service and may lead to remote code execution publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact high for more information on scores click a href isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree update lazy cache set getter isminimumfixversionavailable false basebranches vulnerabilityidentifier cve vulnerabilitydetails prototype pollution vulnerability in ‘set getter’ version allows an attacker to cause a denial of service and may lead to remote code execution vulnerabilityurl
| 0
|
8,671
| 11,802,944,557
|
IssuesEvent
|
2020-03-18 22:48:59
|
hashicorp/packer
|
https://api.github.com/repos/hashicorp/packer
|
closed
|
Packer drops preferred NIC types from Vagrant boxes
|
bug need-more-info post-processor/vagrant waiting-reply
|
Vagrant 2.2.2 now warns when some vulnerable NIC drivers are used, such as E1000. I am able to configure a more appropriate NIC at base box build time, with:
```json
"vboxmanage": [
["modifyvm", "{{.Name}}", "--nictype1", "82543GC"]
]
```
However, the Vagrant box produced by this Packer configuration completely forgets about this preference, resulting in the E1000 NIC being used when the box boots.
Can we fix Packer so that NIC types are better preserved in .BOX's?
Note: Vagrant also has a habit of dropping important configurations for further downstream box exports, so users will likely have to manually configure better NIC types there as well, with an `export.Vagrantfile` for use with `--vagrantfile` in `vagrant package`. But we could at least do a better job having Packer artifacts setup the right default configurations, out of the box so to speak.
|
1.0
|
Packer drops preferred NIC types from Vagrant boxes - Vagrant 2.2.2 now warns when some vulnerable NIC drivers are used, such as E1000. I am able to configure a more appropriate NIC at base box build time, with:
```json
"vboxmanage": [
["modifyvm", "{{.Name}}", "--nictype1", "82543GC"]
]
```
However, the Vagrant box produced by this Packer configuration completely forgets about this preference, resulting in the E1000 NIC being used when the box boots.
Can we fix Packer so that NIC types are better preserved in .BOX's?
Note: Vagrant also has a habit of dropping important configurations for further downstream box exports, so users will likely have to manually configure better NIC types there as well, with an `export.Vagrantfile` for use with `--vagrantfile` in `vagrant package`. But we could at least do a better job having Packer artifacts setup the right default configurations, out of the box so to speak.
|
process
|
packer drops preferred nic types from vagrant boxes vagrant now warns when some vulnerable nic drivers are used such as i am able to configure a more appropriate nic at base box build time with json vboxmanage however the vagrant box produced by this packer configuration completely forgets about this preference resulting in the nic being used when the box boots can we fix packer so that nic types are better preserved in box s note vagrant also has a habit of dropping important configurations for further downstream box exports so users will likely have to manually configure better nic types there as well with an export vagrantfile for use with vagrantfile in vagrant package but we could at least do a better job having packer artifacts setup the right default configurations out of the box so to speak
| 1
|
29,245
| 13,067,348,137
|
IssuesEvent
|
2020-07-31 00:10:25
|
microsoft/botframework-cli
|
https://api.github.com/repos/microsoft/botframework-cli
|
closed
|
update_cognitive_models.ps1 misbehave when Machine learning features are added
|
Bot Services R10 customer-replied-to customer-reported lu
|
#### What project is affected?
Skill template
#### What language is this in?
C#
#### What happens?
When a Machine learning feature is added to an intent on the LUIS authoring page, the execution of update_cognitive_models.ps1 (RemoteToLocal=true) does not generate the xxxLuis.cs file.
The cli tool ends with a successfull "> Update complete." message, no warning is presented.
Just the log file (update_cognitive_models_log.txt) reports the error:
```
AZ CLI passes minimum version. Current version is 2.8.0
[ERROR] line 19:0 - line 19:57: Features must be defined before assigned to an intent. No definition found for feature ""Application synonyms"" in usesFeature definition for intent "Application"
Invalid LUIS JSON file content.
```
#### What are the steps to reproduce this issue?
1. Create an empty Skill from the template.
2. Add a new intent on LUIS authoring for that skill.
3. Add a "Machine learning feature" to that intent.
4. Train and publish.
5. Execute update_cognitive_models.ps1 (RemoteToLocal=true)
#### What were you expecting to happen?
See in my project the xxxLuis.cs file updated with the new intent.
#### Can you share any logs, error output, etc.?
The json file is as follows (it seems a matter of order of the intent and feature definition in the json):
```
MySkill.json
{
"name": "myskillen-us_MySkill",
"versionId": "0.1",
"desc": "",
"culture": "en-us",
"intents": [
{
"name": "Application",
"features": [
{
"featureName": "Application synonyms",
"isRequired": false
}
]
},
{
"name": "None",
"features": []
},
{
"name": "Sample",
"features": []
}
],
"entities": [],
"closedLists": [],
"composites": [],
"hierarchicals": [],
"patternAnyEntities": [],
"regexEntities": [],
"prebuiltEntities": [],
"regexFeatures": [],
"phraselists": [
{
"activated": true,
"name": "Application synonyms",
"words": "Application,Web pages,Sites",
"mode": true,
"enabledForAllModels": false
}
],
"patterns": [],
"utterances": [
{
"text": "demo dialog",
"intent": "Sample",
"entities": []
},
{
"text": "goodbye",
"intent": "None",
"entities": []
},
{
"text": "hello",
"intent": "None",
"entities": []
},
{
"text": "hi",
"intent": "None",
"entities": []
},
{
"text": "list the applications i'm allowed to",
"intent": "Application",
"entities": []
},
{
"text": "logout",
"intent": "None",
"entities": []
},
{
"text": "may you list me my applications?",
"intent": "Application",
"entities": []
},
{
"text": "run dialog",
"intent": "Sample",
"entities": []
},
{
"text": "run sample dialog",
"intent": "Sample",
"entities": []
},
{
"text": "sample dialog",
"intent": "Sample",
"entities": []
},
{
"text": "test dialog",
"intent": "Sample",
"entities": []
},
{
"text": "to what application i've access?",
"intent": "Application",
"entities": []
},
{
"text": "what are my applications?",
"intent": "Application",
"entities": []
}
],
"luis_schema_version": "7.0.0",
"tokenizerVersion": "1.0.0",
"settings": []
}
```
#### Any screenshots or additional context?
No.
|
1.0
|
update_cognitive_models.ps1 misbehave when Machine learning features are added - #### What project is affected?
Skill template
#### What language is this in?
C#
#### What happens?
When a Machine learning feature is added to an intent on the LUIS authoring page, the execution of update_cognitive_models.ps1 (RemoteToLocal=true) does not generate the xxxLuis.cs file.
The cli tool ends with a successfull "> Update complete." message, no warning is presented.
Just the log file (update_cognitive_models_log.txt) reports the error:
```
AZ CLI passes minimum version. Current version is 2.8.0
[ERROR] line 19:0 - line 19:57: Features must be defined before assigned to an intent. No definition found for feature ""Application synonyms"" in usesFeature definition for intent "Application"
Invalid LUIS JSON file content.
```
#### What are the steps to reproduce this issue?
1. Create an empty Skill from the template.
2. Add a new intent on LUIS authoring for that skill.
3. Add a "Machine learning feature" to that intent.
4. Train and publish.
5. Execute update_cognitive_models.ps1 (RemoteToLocal=true)
#### What were you expecting to happen?
See in my project the xxxLuis.cs file updated with the new intent.
#### Can you share any logs, error output, etc.?
The json file is as follows (it seems a matter of order of the intent and feature definition in the json):
```
MySkill.json
{
"name": "myskillen-us_MySkill",
"versionId": "0.1",
"desc": "",
"culture": "en-us",
"intents": [
{
"name": "Application",
"features": [
{
"featureName": "Application synonyms",
"isRequired": false
}
]
},
{
"name": "None",
"features": []
},
{
"name": "Sample",
"features": []
}
],
"entities": [],
"closedLists": [],
"composites": [],
"hierarchicals": [],
"patternAnyEntities": [],
"regexEntities": [],
"prebuiltEntities": [],
"regexFeatures": [],
"phraselists": [
{
"activated": true,
"name": "Application synonyms",
"words": "Application,Web pages,Sites",
"mode": true,
"enabledForAllModels": false
}
],
"patterns": [],
"utterances": [
{
"text": "demo dialog",
"intent": "Sample",
"entities": []
},
{
"text": "goodbye",
"intent": "None",
"entities": []
},
{
"text": "hello",
"intent": "None",
"entities": []
},
{
"text": "hi",
"intent": "None",
"entities": []
},
{
"text": "list the applications i'm allowed to",
"intent": "Application",
"entities": []
},
{
"text": "logout",
"intent": "None",
"entities": []
},
{
"text": "may you list me my applications?",
"intent": "Application",
"entities": []
},
{
"text": "run dialog",
"intent": "Sample",
"entities": []
},
{
"text": "run sample dialog",
"intent": "Sample",
"entities": []
},
{
"text": "sample dialog",
"intent": "Sample",
"entities": []
},
{
"text": "test dialog",
"intent": "Sample",
"entities": []
},
{
"text": "to what application i've access?",
"intent": "Application",
"entities": []
},
{
"text": "what are my applications?",
"intent": "Application",
"entities": []
}
],
"luis_schema_version": "7.0.0",
"tokenizerVersion": "1.0.0",
"settings": []
}
```
#### Any screenshots or additional context?
No.
|
non_process
|
update cognitive models misbehave when machine learning features are added what project is affected skill template what language is this in c what happens when a machine learning feature is added to an intent on the luis authoring page the execution of update cognitive models remotetolocal true does not generate the xxxluis cs file the cli tool ends with a successfull update complete message no warning is presented just the log file update cognitive models log txt reports the error az cli passes minimum version current version is line line features must be defined before assigned to an intent no definition found for feature application synonyms in usesfeature definition for intent application invalid luis json file content what are the steps to reproduce this issue create an empty skill from the template add a new intent on luis authoring for that skill add a machine learning feature to that intent train and publish execute update cognitive models remotetolocal true what were you expecting to happen see in my project the xxxluis cs file updated with the new intent can you share any logs error output etc the json file is as follows it seems a matter of order of the intent and feature definition in the json myskill json name myskillen us myskill versionid desc culture en us intents name application features featurename application synonyms isrequired false name none features name sample features entities closedlists composites hierarchicals patternanyentities regexentities prebuiltentities regexfeatures phraselists activated true name application synonyms words application web pages sites mode true enabledforallmodels false patterns utterances text demo dialog intent sample entities text goodbye intent none entities text hello intent none entities text hi intent none entities text list the applications i m allowed to intent application entities text logout intent none entities text may you list me my applications intent application entities text run dialog intent sample entities text run sample dialog intent sample entities text sample dialog intent sample entities text test dialog intent sample entities text to what application i ve access intent application entities text what are my applications intent application entities luis schema version tokenizerversion settings any screenshots or additional context no
| 0
|
22,003
| 11,660,552,031
|
IssuesEvent
|
2020-03-03 03:44:05
|
cityofaustin/atd-data-tech
|
https://api.github.com/repos/cityofaustin/atd-data-tech
|
opened
|
TURP and Valet Permit Status Issue
|
Product: AMANDA Project: ATD AMANDA Backlog Service: Apps Type: Bug Report Workgroup: ROW migrated
|
Description: Fix the TURP (all subtypes) and Valet (annual and operator) processes so that there is no longer a need for a forced status change.
Request Date: 2018-11-29 16:26:00
Request ID: DTS18-103010
Status: Backlog
Customer Priority: 0
Level of Effort: Minor
DTS URL: https://atd.knack.com/dts#service-requests/view-issue-details/5c0067afd4673a0f23888e88
*Migrated from [atd-amanda #32](https://github.com/cityofaustin/atd-amanda/issues/32)*
|
1.0
|
TURP and Valet Permit Status Issue - Description: Fix the TURP (all subtypes) and Valet (annual and operator) processes so that there is no longer a need for a forced status change.
Request Date: 2018-11-29 16:26:00
Request ID: DTS18-103010
Status: Backlog
Customer Priority: 0
Level of Effort: Minor
DTS URL: https://atd.knack.com/dts#service-requests/view-issue-details/5c0067afd4673a0f23888e88
*Migrated from [atd-amanda #32](https://github.com/cityofaustin/atd-amanda/issues/32)*
|
non_process
|
turp and valet permit status issue description fix the turp all subtypes and valet annual and operator processes so that there is no longer a need for a forced status change request date request id status backlog customer priority level of effort minor dts url migrated from
| 0
|
9,044
| 12,130,108,013
|
IssuesEvent
|
2020-04-23 00:30:40
|
GoogleCloudPlatform/python-docs-samples
|
https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples
|
closed
|
remove gcp-devrel-py-tools from appengine/standard/storage/appengine-client/requirements-test.txt
|
priority: p2 remove-gcp-devrel-py-tools type: process
|
remove gcp-devrel-py-tools from appengine/standard/storage/appengine-client/requirements-test.txt
|
1.0
|
remove gcp-devrel-py-tools from appengine/standard/storage/appengine-client/requirements-test.txt - remove gcp-devrel-py-tools from appengine/standard/storage/appengine-client/requirements-test.txt
|
process
|
remove gcp devrel py tools from appengine standard storage appengine client requirements test txt remove gcp devrel py tools from appengine standard storage appengine client requirements test txt
| 1
|
8,415
| 6,540,697,483
|
IssuesEvent
|
2017-09-01 16:28:28
|
smith-chem-wisc/MetaMorpheus
|
https://api.github.com/repos/smith-chem-wisc/MetaMorpheus
|
closed
|
GeneratePeptideScore is slow
|
Performance
|
Removal of double counting using a HashSet is computationally demanding.
Implement and profile a replacement that checks if experimental peaks have overlapping tolerances and merges them into a single peak with an increased tolerance.
|
True
|
GeneratePeptideScore is slow - Removal of double counting using a HashSet is computationally demanding.
Implement and profile a replacement that checks if experimental peaks have overlapping tolerances and merges them into a single peak with an increased tolerance.
|
non_process
|
generatepeptidescore is slow removal of double counting using a hashset is computationally demanding implement and profile a replacement that checks if experimental peaks have overlapping tolerances and merges them into a single peak with an increased tolerance
| 0
|
19,353
| 25,483,439,195
|
IssuesEvent
|
2022-11-26 03:44:38
|
dotnet/runtime
|
https://api.github.com/repos/dotnet/runtime
|
closed
|
Process.GetProcessesByName() error on linux
|
area-System.Diagnostics.Process untriaged
|
### Description
Chrome Browser's process name is chrome, but `Process.GetProcessesByName("chrome") ` return empty array.
I need pass "chrome --enable-crashpad" to get the right result.
### Reproduction Steps
var processes = Process.GetProcessesByName("chrome --enable-crashpad");
### Expected behavior
var processes = Process.GetProcessesByName("chrome");
### Actual behavior
...
### Regression?
_No response_
### Known Workarounds
_No response_
### Configuration
_No response_
### Other information
_No response_
|
1.0
|
Process.GetProcessesByName() error on linux - ### Description
Chrome Browser's process name is chrome, but `Process.GetProcessesByName("chrome") ` return empty array.
I need pass "chrome --enable-crashpad" to get the right result.
### Reproduction Steps
var processes = Process.GetProcessesByName("chrome --enable-crashpad");
### Expected behavior
var processes = Process.GetProcessesByName("chrome");
### Actual behavior
...
### Regression?
_No response_
### Known Workarounds
_No response_
### Configuration
_No response_
### Other information
_No response_
|
process
|
process getprocessesbyname error on linux description chrome browser s process name is chrome but process getprocessesbyname chrome return empty array i need pass chrome enable crashpad to get the right result reproduction steps var processes process getprocessesbyname chrome enable crashpad expected behavior var processes process getprocessesbyname chrome actual behavior regression no response known workarounds no response configuration no response other information no response
| 1
|
37,131
| 9,963,062,323
|
IssuesEvent
|
2019-07-07 19:57:42
|
ToucanToco/vue-query-builder
|
https://api.github.com/repos/ToucanToco/vue-query-builder
|
closed
|
get rid of webpack
|
build tech
|
We use `rollup` to build and ship an ES6 module. We should therefore get rid of `webpack` since we don't want to keep two distinct package bundlers.
|
1.0
|
get rid of webpack - We use `rollup` to build and ship an ES6 module. We should therefore get rid of `webpack` since we don't want to keep two distinct package bundlers.
|
non_process
|
get rid of webpack we use rollup to build and ship an module we should therefore get rid of webpack since we don t want to keep two distinct package bundlers
| 0
|
169,721
| 26,850,312,665
|
IssuesEvent
|
2023-02-03 10:30:14
|
starplanter93/The_Garden_of_Musicsheet
|
https://api.github.com/repos/starplanter93/The_Garden_of_Musicsheet
|
opened
|
Feat: Header Organism 구현
|
Feat Design Test
|
## Description
Header에 포함되는 molecule, organism 구현
## Todo
- [x]
- [x]
|
1.0
|
Feat: Header Organism 구현 - ## Description
Header에 포함되는 molecule, organism 구현
## Todo
- [x]
- [x]
|
non_process
|
feat header organism 구현 description header에 포함되는 molecule organism 구현 todo
| 0
|
3,108
| 6,130,695,164
|
IssuesEvent
|
2017-06-24 08:02:59
|
kmycode/storycanvas-csharp
|
https://api.github.com/repos/kmycode/storycanvas-csharp
|
closed
|
ファイル保存・読み込みコードの全面見直し
|
correction priority-high processing
|
現行のコードでは、Streamの利用やファイル存在確認などがこんがらかっていて、読みづらいコードになっている。これでは、新しいクラウドサービスの実装がやりづらいばかりが、ファイルの保存などでバグが発生したら対応に時間がかかる。
いっそのこと、この部分のコードを全面的に見直し、見通しの良いコードにしておきたい
|
1.0
|
ファイル保存・読み込みコードの全面見直し - 現行のコードでは、Streamの利用やファイル存在確認などがこんがらかっていて、読みづらいコードになっている。これでは、新しいクラウドサービスの実装がやりづらいばかりが、ファイルの保存などでバグが発生したら対応に時間がかかる。
いっそのこと、この部分のコードを全面的に見直し、見通しの良いコードにしておきたい
|
process
|
ファイル保存・読み込みコードの全面見直し 現行のコードでは、streamの利用やファイル存在確認などがこんがらかっていて、読みづらいコードになっている。これでは、新しいクラウドサービスの実装がやりづらいばかりが、ファイルの保存などでバグが発生したら対応に時間がかかる。 いっそのこと、この部分のコードを全面的に見直し、見通しの良いコードにしておきたい
| 1
|
13,844
| 16,604,724,892
|
IssuesEvent
|
2021-06-02 01:23:26
|
ERFSEDS/Nova_Flight_Computer
|
https://api.github.com/repos/ERFSEDS/Nova_Flight_Computer
|
closed
|
Create device layer
|
Software:C/Cpp Software:Preprocessor prio:medium status:not-started type:feature
|
For which component is used use preprocessor definitions, this will change between boards but not while on one board
|
1.0
|
Create device layer - For which component is used use preprocessor definitions, this will change between boards but not while on one board
|
process
|
create device layer for which component is used use preprocessor definitions this will change between boards but not while on one board
| 1
|
59,287
| 11,956,304,409
|
IssuesEvent
|
2020-04-04 09:43:50
|
home-assistant/brands
|
https://api.github.com/repos/home-assistant/brands
|
closed
|
Yi Home Cameras is missing brand images
|
has-codeowner
|
## The problem
The Yi Home Cameras integration has missing brand images.
We recently started this Brands repository, to create a centralized storage of all brand-related images. These images are used on our website and the Home Assistant frontend.
The following images are missing and would ideally be added:
- `src/yi/logo.png`
- `src/yi/icon@2x.png`
- `src/yi/logo@2x.png`
For image specifications and requirements, please see [README.md](https://github.com/home-assistant/brands/blob/master/README.md).
## Additional information
For more information about this repository, read the [README.md](https://github.com/home-assistant/brands/blob/master/README.md) file of this repository. It contains information on how this repository works, and image specification and requirements.
## Codeowner mention
Hi there, @bachya! Mind taking a look at this issue as it is with an integration (yi) you are listed as a [codeowner](https://github.com/home-assistant/core/blob/dev/homeassistant/components/yi/manifest.json) for? Thanks!
Resolving this issue is not limited to codeowners! If you want to help us out, feel free to resolve this issue! Thanks already!
|
1.0
|
Yi Home Cameras is missing brand images -
## The problem
The Yi Home Cameras integration has missing brand images.
We recently started this Brands repository, to create a centralized storage of all brand-related images. These images are used on our website and the Home Assistant frontend.
The following images are missing and would ideally be added:
- `src/yi/logo.png`
- `src/yi/icon@2x.png`
- `src/yi/logo@2x.png`
For image specifications and requirements, please see [README.md](https://github.com/home-assistant/brands/blob/master/README.md).
## Additional information
For more information about this repository, read the [README.md](https://github.com/home-assistant/brands/blob/master/README.md) file of this repository. It contains information on how this repository works, and image specification and requirements.
## Codeowner mention
Hi there, @bachya! Mind taking a look at this issue as it is with an integration (yi) you are listed as a [codeowner](https://github.com/home-assistant/core/blob/dev/homeassistant/components/yi/manifest.json) for? Thanks!
Resolving this issue is not limited to codeowners! If you want to help us out, feel free to resolve this issue! Thanks already!
|
non_process
|
yi home cameras is missing brand images the problem the yi home cameras integration has missing brand images we recently started this brands repository to create a centralized storage of all brand related images these images are used on our website and the home assistant frontend the following images are missing and would ideally be added src yi logo png src yi icon png src yi logo png for image specifications and requirements please see additional information for more information about this repository read the file of this repository it contains information on how this repository works and image specification and requirements codeowner mention hi there bachya mind taking a look at this issue as it is with an integration yi you are listed as a for thanks resolving this issue is not limited to codeowners if you want to help us out feel free to resolve this issue thanks already
| 0
|
146,727
| 13,190,167,732
|
IssuesEvent
|
2020-08-13 09:43:00
|
riptideio/pymodbus
|
https://api.github.com/repos/riptideio/pymodbus
|
closed
|
Read RTU Holding Register through Serial Forwarder/TCP.
|
Documentation Update Required
|
I have Energy Meter connected through RTU and able to get holding registers data through simple RTU Code.
Now i want to make Convert this RTU to TCP through Forwarder. I want to send data to TCP which forwards the command to RTU and fetches data for me.
I have implement the Forwarder code just dont know how to fetch the holding register of RTU through it.
**Code for Simple RTU Read**
> import pymodbus
> from pymodbus.pdu import ModbusRequest
> from pymodbus.client.sync import ModbusSerialClient as ModbusClient
> #initialize a serial RTU client instance
> from pymodbus.transaction import ModbusRtuFramer
>
> #count= the number of registers to read
> #unit= the slave unit this request is targeting
> #address= the starting address to read from
>
> client = ModbusClient(method = 'rtu', port='/dev/ttyUSB0', baudrate= 9600)
>
> #Connect to the serial modbus server
> connection = client.connect()
> print(connection)
>
> #Starting add, num of reg to read, slave unit.
> read = client.read_holding_registers(address = 0x01,count =2, unit=1)
> data = read.registers
>
> print(data)
>
> #Closes the underlying socket connection
> client.close()
|
1.0
|
Read RTU Holding Register through Serial Forwarder/TCP. - I have Energy Meter connected through RTU and able to get holding registers data through simple RTU Code.
Now i want to make Convert this RTU to TCP through Forwarder. I want to send data to TCP which forwards the command to RTU and fetches data for me.
I have implement the Forwarder code just dont know how to fetch the holding register of RTU through it.
**Code for Simple RTU Read**
> import pymodbus
> from pymodbus.pdu import ModbusRequest
> from pymodbus.client.sync import ModbusSerialClient as ModbusClient
> #initialize a serial RTU client instance
> from pymodbus.transaction import ModbusRtuFramer
>
> #count= the number of registers to read
> #unit= the slave unit this request is targeting
> #address= the starting address to read from
>
> client = ModbusClient(method = 'rtu', port='/dev/ttyUSB0', baudrate= 9600)
>
> #Connect to the serial modbus server
> connection = client.connect()
> print(connection)
>
> #Starting add, num of reg to read, slave unit.
> read = client.read_holding_registers(address = 0x01,count =2, unit=1)
> data = read.registers
>
> print(data)
>
> #Closes the underlying socket connection
> client.close()
|
non_process
|
read rtu holding register through serial forwarder tcp i have energy meter connected through rtu and able to get holding registers data through simple rtu code now i want to make convert this rtu to tcp through forwarder i want to send data to tcp which forwards the command to rtu and fetches data for me i have implement the forwarder code just dont know how to fetch the holding register of rtu through it code for simple rtu read import pymodbus from pymodbus pdu import modbusrequest from pymodbus client sync import modbusserialclient as modbusclient initialize a serial rtu client instance from pymodbus transaction import modbusrtuframer count the number of registers to read unit the slave unit this request is targeting address the starting address to read from client modbusclient method rtu port dev baudrate connect to the serial modbus server connection client connect print connection starting add num of reg to read slave unit read client read holding registers address count unit data read registers print data closes the underlying socket connection client close
| 0
|
816,009
| 30,583,369,490
|
IssuesEvent
|
2023-07-21 11:29:01
|
Neural-Systems-at-UIO/RB-workbench
|
https://api.github.com/repos/Neural-Systems-at-UIO/RB-workbench
|
opened
|
WebAlign failed to load in Firefox and Chrome
|
High Priority
|
WebAlign failed to load in the RB-workbench.

|
1.0
|
WebAlign failed to load in Firefox and Chrome - WebAlign failed to load in the RB-workbench.

|
non_process
|
webalign failed to load in firefox and chrome webalign failed to load in the rb workbench
| 0
|
2,328
| 5,142,548,021
|
IssuesEvent
|
2017-01-12 13:39:39
|
jimbrown75/Permit-Vision-Enhancements
|
https://api.github.com/repos/jimbrown75/Permit-Vision-Enhancements
|
opened
|
During RA/JH, allow selection of Pre-requisite confirmaiton either once, or every issue
|
High Priority Should Fix Verified by PTW Process Lead
|
When creating a RA/JHA allow the risk assessor to select, via a radio button that all pre-requisites need to be confirmed by signature either at first issue or every issue.
|
1.0
|
During RA/JH, allow selection of Pre-requisite confirmaiton either once, or every issue - When creating a RA/JHA allow the risk assessor to select, via a radio button that all pre-requisites need to be confirmed by signature either at first issue or every issue.
|
process
|
during ra jh allow selection of pre requisite confirmaiton either once or every issue when creating a ra jha allow the risk assessor to select via a radio button that all pre requisites need to be confirmed by signature either at first issue or every issue
| 1
|
21,296
| 28,494,402,447
|
IssuesEvent
|
2023-04-18 13:20:41
|
Open-EO/openeo-api
|
https://api.github.com/repos/Open-EO/openeo-api
|
closed
|
minimum log_level for `GET /../logs`
|
data processing minor
|
#431 added option to specify minimum log_level at job submission time.
We are currently also playing with setting a minimum log_level when requesting the logs (e.g. `GET /jobs/{job_id}/logs)`.
For example see Open-EO/openeo-python-client#332, where we filter client side to only show error logs, but we've experienced that it would be better to do the filtering already back-end side (for performance reasons).
Can we add an optional `log_level` parameter to `GET /jobs/{job_id}/logs` and `GET /services/{service_id}/logs` for back-end side log filtering at retrieval time?
refs and related issues
- #329
- #431
- Open-EO/openeo-python-client#332
|
1.0
|
minimum log_level for `GET /../logs` - #431 added option to specify minimum log_level at job submission time.
We are currently also playing with setting a minimum log_level when requesting the logs (e.g. `GET /jobs/{job_id}/logs)`.
For example see Open-EO/openeo-python-client#332, where we filter client side to only show error logs, but we've experienced that it would be better to do the filtering already back-end side (for performance reasons).
Can we add an optional `log_level` parameter to `GET /jobs/{job_id}/logs` and `GET /services/{service_id}/logs` for back-end side log filtering at retrieval time?
refs and related issues
- #329
- #431
- Open-EO/openeo-python-client#332
|
process
|
minimum log level for get logs added option to specify minimum log level at job submission time we are currently also playing with setting a minimum log level when requesting the logs e g get jobs job id logs for example see open eo openeo python client where we filter client side to only show error logs but we ve experienced that it would be better to do the filtering already back end side for performance reasons can we add an optional log level parameter to get jobs job id logs and get services service id logs for back end side log filtering at retrieval time refs and related issues open eo openeo python client
| 1
|
11,913
| 14,700,197,933
|
IssuesEvent
|
2021-01-04 09:50:13
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
Processing/GRASS tool fails to generate output
|
Bug Feedback Processing stale
|
I found something that looked like a bug.
so, When I run the [i.pansharpen] command, it finally fails.
The output message is as follows.
[Read output layer The next layer was not generated correctly.
• C: /Users/user/AppData/Local/Temp/processing_SiwLfp/2d043cc5b43542f58a317c16a7f5c1f8/redoutput.tif
• C: /Users/user/AppData/Local/Temp/processing_SiwLfp/7278fd9f94164d6fbd5483d23b909f7a/greenoutput.tif
• C: /Users/user/AppData/Local/Temp/processing_SiwLfp/97e48a96f6724a079668047d048ef753/blueoutput.tif
Check the "Log Message Panel" in the QGIS main window to see more information about running the algorithm.]
And another GRASS command same too.
Calculations can be performed on QGIS316.
Something is wrong with QGIS310.
|
1.0
|
Processing/GRASS tool fails to generate output - I found something that looked like a bug.
so, When I run the [i.pansharpen] command, it finally fails.
The output message is as follows.
[Read output layer The next layer was not generated correctly.
• C: /Users/user/AppData/Local/Temp/processing_SiwLfp/2d043cc5b43542f58a317c16a7f5c1f8/redoutput.tif
• C: /Users/user/AppData/Local/Temp/processing_SiwLfp/7278fd9f94164d6fbd5483d23b909f7a/greenoutput.tif
• C: /Users/user/AppData/Local/Temp/processing_SiwLfp/97e48a96f6724a079668047d048ef753/blueoutput.tif
Check the "Log Message Panel" in the QGIS main window to see more information about running the algorithm.]
And another GRASS command same too.
Calculations can be performed on QGIS316.
Something is wrong with QGIS310.
|
process
|
processing grass tool fails to generate output i found something that looked like a bug so when i run the command it finally fails the output message is as follows read output layer the next layer was not generated correctly • c users user appdata local temp processing siwlfp redoutput tif • c users user appdata local temp processing siwlfp greenoutput tif • c users user appdata local temp processing siwlfp blueoutput tif check the log message panel in the qgis main window to see more information about running the algorithm and another grass command same too calculations can be performed on something is wrong with
| 1
|
9,439
| 12,425,101,115
|
IssuesEvent
|
2020-05-24 14:50:51
|
raxod502/straight.el
|
https://api.github.com/repos/raxod502/straight.el
|
opened
|
Shallow clones not supported properly in VC commands
|
bug git interactive lockfiles process buffer shallow clone vc
|
When I run `M-x straight-freeze-versions`, for example, I get a "Git failed" error if one of my repos is a shallow clone. The VC commands need a complete overhaul at some point.
|
1.0
|
Shallow clones not supported properly in VC commands - When I run `M-x straight-freeze-versions`, for example, I get a "Git failed" error if one of my repos is a shallow clone. The VC commands need a complete overhaul at some point.
|
process
|
shallow clones not supported properly in vc commands when i run m x straight freeze versions for example i get a git failed error if one of my repos is a shallow clone the vc commands need a complete overhaul at some point
| 1
|
5,801
| 8,641,448,974
|
IssuesEvent
|
2018-11-24 17:52:26
|
carloseduardov8/Viajato
|
https://api.github.com/repos/carloseduardov8/Viajato
|
closed
|
Implementar cancelamento de seleção
|
Priority:Normal Process:Implement Requirement
|
Usuário deve poder visualizar escolhas já feitas e cancelar aquelas que não deseja mais.
|
1.0
|
Implementar cancelamento de seleção - Usuário deve poder visualizar escolhas já feitas e cancelar aquelas que não deseja mais.
|
process
|
implementar cancelamento de seleção usuário deve poder visualizar escolhas já feitas e cancelar aquelas que não deseja mais
| 1
|
222,070
| 17,391,459,265
|
IssuesEvent
|
2021-08-02 07:58:15
|
xamarin/xamarin-macios
|
https://api.github.com/repos/xamarin/xamarin-macios
|
opened
|
Test builds sometimes fail with: Cannot write to the output file "[..]". The process cannot access the file '[...]' because it is being used by another process.
|
bug tests
|
We try to build tests in parallel whenever we can, but in certain cases we end up building the same project simultaneously in different processes, and this can cause errors like this:
> /Users/builder/azdo/_work/1/s/xamarin-macios/builds/downloads/dotnet-sdk-6.0.100-rc.1.21381.5-osx-x64/sdk/6.0.100-rc.1.21381.5/Microsoft.Common.CurrentVersion.targets(3262,5): error MSB3554: Cannot write to the output file "/Users/builder/azdo/_work/1/s/xamarin-macios/tests/EmbeddedResources/dotnet/macOS/obj/Debug/net6.0-macos/EmbeddedResources.Welcome.resources". The process cannot access the file '/Users/builder/azdo/_work/1/s/xamarin-macios/tests/EmbeddedResources/dotnet/macOS/obj/Debug/net6.0-macos/EmbeddedResources.Welcome.resources' because it is being used by another process.
[vsdrops_index-105.html.txt](https://github.com/xamarin/xamarin-macios/files/6915293/vsdrops_index-105.html.txt)
|
1.0
|
Test builds sometimes fail with: Cannot write to the output file "[..]". The process cannot access the file '[...]' because it is being used by another process. - We try to build tests in parallel whenever we can, but in certain cases we end up building the same project simultaneously in different processes, and this can cause errors like this:
> /Users/builder/azdo/_work/1/s/xamarin-macios/builds/downloads/dotnet-sdk-6.0.100-rc.1.21381.5-osx-x64/sdk/6.0.100-rc.1.21381.5/Microsoft.Common.CurrentVersion.targets(3262,5): error MSB3554: Cannot write to the output file "/Users/builder/azdo/_work/1/s/xamarin-macios/tests/EmbeddedResources/dotnet/macOS/obj/Debug/net6.0-macos/EmbeddedResources.Welcome.resources". The process cannot access the file '/Users/builder/azdo/_work/1/s/xamarin-macios/tests/EmbeddedResources/dotnet/macOS/obj/Debug/net6.0-macos/EmbeddedResources.Welcome.resources' because it is being used by another process.
[vsdrops_index-105.html.txt](https://github.com/xamarin/xamarin-macios/files/6915293/vsdrops_index-105.html.txt)
|
non_process
|
test builds sometimes fail with cannot write to the output file the process cannot access the file because it is being used by another process we try to build tests in parallel whenever we can but in certain cases we end up building the same project simultaneously in different processes and this can cause errors like this users builder azdo work s xamarin macios builds downloads dotnet sdk rc osx sdk rc microsoft common currentversion targets error cannot write to the output file users builder azdo work s xamarin macios tests embeddedresources dotnet macos obj debug macos embeddedresources welcome resources the process cannot access the file users builder azdo work s xamarin macios tests embeddedresources dotnet macos obj debug macos embeddedresources welcome resources because it is being used by another process
| 0
|
3,505
| 6,559,511,665
|
IssuesEvent
|
2017-09-07 04:31:20
|
amaster507/ifbmt
|
https://api.github.com/repos/amaster507/ifbmt
|
closed
|
Mass emailer
|
email enhancement GUI idea process wontfix
|
What if there was a way to refine a search list of churches and then send an email to all if then at once?
This may need to be reviewed or limited in use to not cause spam. We don't want this to end up being something that every preacher/church ignores because it is overused.
|
1.0
|
Mass emailer - What if there was a way to refine a search list of churches and then send an email to all if then at once?
This may need to be reviewed or limited in use to not cause spam. We don't want this to end up being something that every preacher/church ignores because it is overused.
|
process
|
mass emailer what if there was a way to refine a search list of churches and then send an email to all if then at once this may need to be reviewed or limited in use to not cause spam we don t want this to end up being something that every preacher church ignores because it is overused
| 1
|
50,435
| 10,510,304,422
|
IssuesEvent
|
2019-09-27 13:09:48
|
Samsung/TAU-Design-Editor
|
https://api.github.com/repos/Samsung/TAU-Design-Editor
|
closed
|
[VSCode] [Wearable] Image widget has strange image src set by default
|
1.0 bug in review vscode wearable
|
## Technical
[VSCode] [Wearable] Image widget has strange image src set by default
## Steps to reproduce
[W Sample TAU 1.1.4.zip](https://github.com/Samsung/TAU-Design-Editor/files/3590275/W.Sample.TAU.1.1.4.zip)
1. open attached sample project in DE
2. D&D image widget
3. Go to Attributes->Image-->SRC
## Expected behaviour
Nothing is set in src
## Actual behaviour
Strange "# is set as a source

|
1.0
|
[VSCode] [Wearable] Image widget has strange image src set by default - ## Technical
[VSCode] [Wearable] Image widget has strange image src set by default
## Steps to reproduce
[W Sample TAU 1.1.4.zip](https://github.com/Samsung/TAU-Design-Editor/files/3590275/W.Sample.TAU.1.1.4.zip)
1. open attached sample project in DE
2. D&D image widget
3. Go to Attributes->Image-->SRC
## Expected behaviour
Nothing is set in src
## Actual behaviour
Strange "# is set as a source

|
non_process
|
image widget has strange image src set by default technical image widget has strange image src set by default steps to reproduce open attached sample project in de d d image widget go to attributes image src expected behaviour nothing is set in src actual behaviour strange is set as a source
| 0
|
20,738
| 27,438,052,950
|
IssuesEvent
|
2023-03-02 09:03:37
|
camunda/issues
|
https://api.github.com/repos/camunda/issues
|
opened
|
BPMN Escalation Events
|
component:desktopModeler component:operate component:optimize component:webModeler component:zeebe-process-automation public feature-parity version:8.2-alpha4
|
> This is an epic internal-docs issue. It bundles all activities we conduct around a certain initiative. It will typically links to various child issues from various repositories and can be spread across multiple teams.
### Value Proposition Statement
Communicate from a subprocess to an upper process using BPMN Escalation Events
### User Problem
Escalation events are events which reference a named escalation. They are mostly used to communicate from a subprocess to an upper process. Unlike an error, an escalation event is non critical and execution continues at the location of throwing.
Currently this BPMN symbols is not supported by Camunda Platform 8.
### User Stories
**Modeler**
As a developer I can use Escalation Events in the Modeler.
**Zeebe**
As a developer I can deploy Esalation Events to Zeebe.
As a developer I can trust that Escalation Events are executed automatically by Zeebe according to BPMN spec.
**Operate / Optimize**
As a developer I can see Escalation Events in Operate and Optimize correctly and the import is not broken.
The technical properties of Escalation should be most likely similar to Camunda 7.
### Implementation Notes
<!-- Notes to consider for implementation, for example:
* In Cawemo we already have the capability to manage templates via the feature that we call “catalog”
* What we would build now is the ability to a) use this feature in the web modeler to create templates and b) when the context pad opens for defining the type of a task, the templates that decorate service tasks are shown
* We should clarify terminology (integrations vs. connectors vs. job workers vs. element templates.) Particularly “element templates” might not be a term that a user intuitively understands.
* See these high level wireframes to capture the idea -->
### Breakdown
**Zeebe**
- https://github.com/camunda/zeebe/issues/10682
**Modeler**
- [x] https://github.com/camunda/camunda-modeler/issues/3318
**Documentation**
- https://github.com/camunda/camunda-platform-docs/issues/1531
#### Discovery phase ##
<!-- Example: link to "Conduct customer interview with xyz" -->
#### Define phase ##
<!-- Consider: UI, UX, technical design, documentation design -->
<!-- Example: link to "Define User-Journey Flow" or "Define target architecture" -->
Design Planning
* Reviewed by design: {date}
* Designer assigned: {Yes, No Design Necessary, or No Designer Available}
* Assignee:
* Design Brief - {link to design brief }
* Research Brief - {link to research brief }
Design Deliverables
* {Deliverable Name} {Link to GH Issue}
#### Implement phase ##
**Optimize**
- [ ] https://jira.camunda.com/browse/OPT-6612
**Modeler**
- [x] https://github.com/camunda/camunda-modeler/issues/3318
**Operate**
- [x] https://github.com/camunda/operate/issues/3757
#### Validate phase ##
<!-- Example: link to "Evaluate usage data of last quarter" -->
### Links to additional collateral
|
1.0
|
BPMN Escalation Events - > This is an epic internal-docs issue. It bundles all activities we conduct around a certain initiative. It will typically links to various child issues from various repositories and can be spread across multiple teams.
### Value Proposition Statement
Communicate from a subprocess to an upper process using BPMN Escalation Events
### User Problem
Escalation events are events which reference a named escalation. They are mostly used to communicate from a subprocess to an upper process. Unlike an error, an escalation event is non critical and execution continues at the location of throwing.
Currently this BPMN symbols is not supported by Camunda Platform 8.
### User Stories
**Modeler**
As a developer I can use Escalation Events in the Modeler.
**Zeebe**
As a developer I can deploy Esalation Events to Zeebe.
As a developer I can trust that Escalation Events are executed automatically by Zeebe according to BPMN spec.
**Operate / Optimize**
As a developer I can see Escalation Events in Operate and Optimize correctly and the import is not broken.
The technical properties of Escalation should be most likely similar to Camunda 7.
### Implementation Notes
<!-- Notes to consider for implementation, for example:
* In Cawemo we already have the capability to manage templates via the feature that we call “catalog”
* What we would build now is the ability to a) use this feature in the web modeler to create templates and b) when the context pad opens for defining the type of a task, the templates that decorate service tasks are shown
* We should clarify terminology (integrations vs. connectors vs. job workers vs. element templates.) Particularly “element templates” might not be a term that a user intuitively understands.
* See these high level wireframes to capture the idea -->
### Breakdown
**Zeebe**
- https://github.com/camunda/zeebe/issues/10682
**Modeler**
- [x] https://github.com/camunda/camunda-modeler/issues/3318
**Documentation**
- https://github.com/camunda/camunda-platform-docs/issues/1531
#### Discovery phase ##
<!-- Example: link to "Conduct customer interview with xyz" -->
#### Define phase ##
<!-- Consider: UI, UX, technical design, documentation design -->
<!-- Example: link to "Define User-Journey Flow" or "Define target architecture" -->
Design Planning
* Reviewed by design: {date}
* Designer assigned: {Yes, No Design Necessary, or No Designer Available}
* Assignee:
* Design Brief - {link to design brief }
* Research Brief - {link to research brief }
Design Deliverables
* {Deliverable Name} {Link to GH Issue}
#### Implement phase ##
**Optimize**
- [ ] https://jira.camunda.com/browse/OPT-6612
**Modeler**
- [x] https://github.com/camunda/camunda-modeler/issues/3318
**Operate**
- [x] https://github.com/camunda/operate/issues/3757
#### Validate phase ##
<!-- Example: link to "Evaluate usage data of last quarter" -->
### Links to additional collateral
|
process
|
bpmn escalation events this is an epic internal docs issue it bundles all activities we conduct around a certain initiative it will typically links to various child issues from various repositories and can be spread across multiple teams value proposition statement communicate from a subprocess to an upper process using bpmn escalation events user problem escalation events are events which reference a named escalation they are mostly used to communicate from a subprocess to an upper process unlike an error an escalation event is non critical and execution continues at the location of throwing currently this bpmn symbols is not supported by camunda platform user stories modeler as a developer i can use escalation events in the modeler zeebe as a developer i can deploy esalation events to zeebe as a developer i can trust that escalation events are executed automatically by zeebe according to bpmn spec operate optimize as a developer i can see escalation events in operate and optimize correctly and the import is not broken the technical properties of escalation should be most likely similar to camunda implementation notes notes to consider for implementation for example in cawemo we already have the capability to manage templates via the feature that we call “catalog” what we would build now is the ability to a use this feature in the web modeler to create templates and b when the context pad opens for defining the type of a task the templates that decorate service tasks are shown we should clarify terminology integrations vs connectors vs job workers vs element templates particularly “element templates” might not be a term that a user intuitively understands see these high level wireframes to capture the idea breakdown zeebe modeler documentation discovery phase define phase design planning reviewed by design date designer assigned yes no design necessary or no designer available assignee design brief link to design brief research brief link to research brief design deliverables deliverable name link to gh issue implement phase optimize modeler operate validate phase links to additional collateral
| 1
|
12,002
| 3,561,583,960
|
IssuesEvent
|
2016-01-23 22:06:13
|
empowerhack/hub
|
https://api.github.com/repos/empowerhack/hub
|
closed
|
Create Spring boot project
|
type: documentation type: java
|
* [x] Create a Vanilla Spring boot project
* [x] Documentation on how to setup project locally
* [x] Add Contribution file to repo include:
* [x] include git flow usage & link to official docs
* [x] pull request, labels, template etc
|
1.0
|
Create Spring boot project - * [x] Create a Vanilla Spring boot project
* [x] Documentation on how to setup project locally
* [x] Add Contribution file to repo include:
* [x] include git flow usage & link to official docs
* [x] pull request, labels, template etc
|
non_process
|
create spring boot project create a vanilla spring boot project documentation on how to setup project locally add contribution file to repo include include git flow usage link to official docs pull request labels template etc
| 0
|
22,709
| 32,036,924,273
|
IssuesEvent
|
2023-09-22 16:01:32
|
GoogleCloudPlatform/alloydb-python-connector
|
https://api.github.com/repos/GoogleCloudPlatform/alloydb-python-connector
|
closed
|
Add test demonstating Auto IAM AuthN on direct path
|
priority: p1 type: docs type: process
|
### Description
Show how to leverage asyncpg and sqlalchemy to refresh an OAuth2 token on connect
### Potential Solution
_No response_
### Additional Details
_No response_
|
1.0
|
Add test demonstating Auto IAM AuthN on direct path - ### Description
Show how to leverage asyncpg and sqlalchemy to refresh an OAuth2 token on connect
### Potential Solution
_No response_
### Additional Details
_No response_
|
process
|
add test demonstating auto iam authn on direct path description show how to leverage asyncpg and sqlalchemy to refresh an token on connect potential solution no response additional details no response
| 1
|
14,129
| 17,023,119,883
|
IssuesEvent
|
2021-07-03 00:27:50
|
fweikert/bugs
|
https://api.github.com/repos/fweikert/bugs
|
opened
|
Status of Bazel 5.0.0-pre.20210623.2
|
P1 release team-XProduct type: process
|
- Expected release date: 2021-07-03
Task list:
- [ ] Pick release baseline: 8b453331163378071f1cfe0ae7c74d551c21b834 with cherrypick [223113c9](https://github.com/bazelbuild/bazel/commit/223113c9202e8f338b183d1736d97327d28241ea)
- [ ] Create release candidate: https://releases.bazel.build/5.0.0/rolling/5.0.0-pre.20210623.2rc1/index.html
- [ ] Post-submit: https://buildkite.com/bazel/bazel-bazel
- [ ] Push the release: https://releases.bazel.build/5.0.0/rolling/5.0.0-pre.20210623.2/index.html
- [ ] Update the [release page](https://github.com/bazelbuild/bazel/releases/)
|
1.0
|
Status of Bazel 5.0.0-pre.20210623.2 -
- Expected release date: 2021-07-03
Task list:
- [ ] Pick release baseline: 8b453331163378071f1cfe0ae7c74d551c21b834 with cherrypick [223113c9](https://github.com/bazelbuild/bazel/commit/223113c9202e8f338b183d1736d97327d28241ea)
- [ ] Create release candidate: https://releases.bazel.build/5.0.0/rolling/5.0.0-pre.20210623.2rc1/index.html
- [ ] Post-submit: https://buildkite.com/bazel/bazel-bazel
- [ ] Push the release: https://releases.bazel.build/5.0.0/rolling/5.0.0-pre.20210623.2/index.html
- [ ] Update the [release page](https://github.com/bazelbuild/bazel/releases/)
|
process
|
status of bazel pre expected release date task list pick release baseline with cherrypick create release candidate post submit push the release update the
| 1
|
648,971
| 21,214,613,183
|
IssuesEvent
|
2022-04-11 05:40:33
|
AY2122S2-CS2103T-W15-2/tp
|
https://api.github.com/repos/AY2122S2-CS2103T-W15-2/tp
|
closed
|
Add tests for Salary class
|
priority.High type.Task
|
Tests for the `salary` class are not implemented. Add tests to automate testing and minimise regression.
|
1.0
|
Add tests for Salary class - Tests for the `salary` class are not implemented. Add tests to automate testing and minimise regression.
|
non_process
|
add tests for salary class tests for the salary class are not implemented add tests to automate testing and minimise regression
| 0
|
285,468
| 24,668,603,186
|
IssuesEvent
|
2022-10-18 12:15:32
|
Kimthien94/Feedback-Online
|
https://api.github.com/repos/Kimthien94/Feedback-Online
|
opened
|
[BugID_45]_FUNC_Xóa topic_Không có thông báo xóa topic thành công
|
bug Fun_Feature Missing comestic low Intergration test open
|
Precondition:
- Admin đang ở màn Quản lý Topic
- Topic có Tên Topic = 'TopicA' không được sử dụng ở module khác (ex: Gán topic)
Steps:
1. Click [Xóa] button tại record 'TopicA'
2. Click [Có] button
Thực tế: Không hiển thị thông báo
Mong đợi: Hiển thị inline message màu xanh “Xóa Topic thành công” ở góc bên trái của page
Testcase ID = 20
|
1.0
|
[BugID_45]_FUNC_Xóa topic_Không có thông báo xóa topic thành công - Precondition:
- Admin đang ở màn Quản lý Topic
- Topic có Tên Topic = 'TopicA' không được sử dụng ở module khác (ex: Gán topic)
Steps:
1. Click [Xóa] button tại record 'TopicA'
2. Click [Có] button
Thực tế: Không hiển thị thông báo
Mong đợi: Hiển thị inline message màu xanh “Xóa Topic thành công” ở góc bên trái của page
Testcase ID = 20
|
non_process
|
func xóa topic không có thông báo xóa topic thành công precondition admin đang ở màn quản lý topic topic có tên topic topica không được sử dụng ở module khác ex gán topic steps click button tại record topica click button thực tế không hiển thị thông báo mong đợi hiển thị inline message màu xanh “xóa topic thành công” ở góc bên trái của page testcase id
| 0
|
228,669
| 25,239,602,904
|
IssuesEvent
|
2022-11-15 05:55:25
|
Satheesh575555/linux-3.0.35
|
https://api.github.com/repos/Satheesh575555/linux-3.0.35
|
opened
|
CVE-2016-4470 (Medium) detected in linuxlinux-3.0.40
|
security vulnerability
|
## CVE-2016-4470 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-3.0.40</b></p></summary>
<p>
<p>Apache Software Foundation (ASF)</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v3.0/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v3.0/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Satheesh575555/linux-3.0.35/commit/d886d5c33aadc1c4f116214d0060f5869b445fe1">d886d5c33aadc1c4f116214d0060f5869b445fe1</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/security/keys/key.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/security/keys/key.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/security/keys/key.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The key_reject_and_link function in security/keys/key.c in the Linux kernel through 4.6.3 does not ensure that a certain data structure is initialized, which allows local users to cause a denial of service (system crash) via vectors involving a crafted keyctl request2 command.
<p>Publish Date: 2016-06-27
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2016-4470>CVE-2016-4470</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.2</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2016-4470">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2016-4470</a></p>
<p>Release Date: 2016-06-27</p>
<p>Fix Resolution: v4.7-rc4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2016-4470 (Medium) detected in linuxlinux-3.0.40 - ## CVE-2016-4470 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-3.0.40</b></p></summary>
<p>
<p>Apache Software Foundation (ASF)</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v3.0/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v3.0/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Satheesh575555/linux-3.0.35/commit/d886d5c33aadc1c4f116214d0060f5869b445fe1">d886d5c33aadc1c4f116214d0060f5869b445fe1</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/security/keys/key.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/security/keys/key.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/security/keys/key.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The key_reject_and_link function in security/keys/key.c in the Linux kernel through 4.6.3 does not ensure that a certain data structure is initialized, which allows local users to cause a denial of service (system crash) via vectors involving a crafted keyctl request2 command.
<p>Publish Date: 2016-06-27
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2016-4470>CVE-2016-4470</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.2</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2016-4470">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2016-4470</a></p>
<p>Release Date: 2016-06-27</p>
<p>Fix Resolution: v4.7-rc4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in linuxlinux cve medium severity vulnerability vulnerable library linuxlinux apache software foundation asf library home page a href found in head commit a href found in base branch master vulnerable source files security keys key c security keys key c security keys key c vulnerability details the key reject and link function in security keys key c in the linux kernel through does not ensure that a certain data structure is initialized which allows local users to cause a denial of service system crash via vectors involving a crafted keyctl command publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
11,281
| 2,648,796,649
|
IssuesEvent
|
2015-03-14 08:15:35
|
STEllAR-GROUP/hpx
|
https://api.github.com/repos/STEllAR-GROUP/hpx
|
opened
|
Thread performance counters (pending queue counts) totals disagree with worker threads
|
affecting NMSU category: performance counters difficulty: easy type: defect
|
Command:
bin/1d_stencil_4 --nx=1000 --np=1000 -t2 \
--hpx:print-counter /threads{locality#*/total}/<thread counter> \
--hpx:print-counter /threads{locality#*/worker-thread#*}/<thread counter>
Counters where individual worker thread counts disagree with total include:
/count/pending-accesses
/count/pending-misses
/count/stolen-to-pending
/count/stolen-from-pending
Examples:
/count/pending-accesses:
OS_Threads,Execution_Time_sec,Points_per_Partition,Partitions,Time_Steps
2, 0.14836133, 1000, 1000, 45
/threads{locality#0/total}/count/pending-accesses,1,0.157037,[s],186465
/threads{locality#0/worker-thread#0}/count/pending-accesses,1,0.157106,[s],14337
/threads{locality#0/worker-thread#1}/count/pending-accesses,1,0.157077,[s],61692
/count/pending-misses:
OS_Threads,Execution_Time_sec,Points_per_Partition,Partitions,Time_Steps
2, 0.150560948, 1000, 1000, 45
/threads{locality#0/total}/count/pending-misses,1,0.159072,[s],127346
/threads{locality#0/worker-thread#0}/count/pending-misses,1,0.159071,[s],23505
/threads{locality#0/worker-thread#1}/count/pending-misses,1,0.159115,[s],39951
/count/pending-accesses:
OS_Threads,Execution_Time_sec,Points_per_Partition,Partitions,Time_Steps
2, 0.14886279, 1000, 1000, 45
/threads{locality#0/total}/count/pending-accesses,1,0.157788,[s],179778
/threads{locality#0/worker-thread#0}/count/pending-accesses,1,0.157897,[s],53157
/threads{locality#0/worker-thread#1}/count/pending-accesses,1,0.157833,[s],23033
/count/pending-misses:
OS_Threads,Execution_Time_sec,Points_per_Partition,Partitions,Time_Steps
2, 0.150670529, 1000, 1000, 45
/threads{locality#0/total}/count/pending-misses,1,0.160003,[s],122847
/threads{locality#0/worker-thread#0}/count/pending-misses,1,0.160061,[s],52292
/threads{locality#0/worker-thread#1}/count/pending-misses,1,0.160091,[s],8937
/count/stolen-to-pending:
OS_Threads,Execution_Time_sec,Points_per_Partition,Partitions,Time_Steps
2, 0.147684702, 1000, 1000, 45
/threads{locality#0/total}/count/stolen-to-pending,1,0.156292,[s],10580
/threads{locality#0/worker-thread#0}/count/stolen-to-pending,1,0.156333,[s],13
/threads{locality#0/worker-thread#1}/count/stolen-to-pending,1,0.156282,[s],29
/count/stolen-from-pending:
OS_Threads,Execution_Time_sec,Points_per_Partition,Partitions,Time_Steps
2, 0.150269219, 1000, 1000, 45
/threads{locality#0/total}/count/stolen-from-pending,1,0.159178,[s],9343
/threads{locality#0/worker-thread#0}/count/stolen-from-pending,1,0.159268,[s],20
/threads{locality#0/worker-thread#1}/count/stolen-from-pending,1,0.159230,[s],50
|
1.0
|
Thread performance counters (pending queue counts) totals disagree with worker threads - Command:
bin/1d_stencil_4 --nx=1000 --np=1000 -t2 \
--hpx:print-counter /threads{locality#*/total}/<thread counter> \
--hpx:print-counter /threads{locality#*/worker-thread#*}/<thread counter>
Counters where individual worker thread counts disagree with total include:
/count/pending-accesses
/count/pending-misses
/count/stolen-to-pending
/count/stolen-from-pending
Examples:
/count/pending-accesses:
OS_Threads,Execution_Time_sec,Points_per_Partition,Partitions,Time_Steps
2, 0.14836133, 1000, 1000, 45
/threads{locality#0/total}/count/pending-accesses,1,0.157037,[s],186465
/threads{locality#0/worker-thread#0}/count/pending-accesses,1,0.157106,[s],14337
/threads{locality#0/worker-thread#1}/count/pending-accesses,1,0.157077,[s],61692
/count/pending-misses:
OS_Threads,Execution_Time_sec,Points_per_Partition,Partitions,Time_Steps
2, 0.150560948, 1000, 1000, 45
/threads{locality#0/total}/count/pending-misses,1,0.159072,[s],127346
/threads{locality#0/worker-thread#0}/count/pending-misses,1,0.159071,[s],23505
/threads{locality#0/worker-thread#1}/count/pending-misses,1,0.159115,[s],39951
/count/pending-accesses:
OS_Threads,Execution_Time_sec,Points_per_Partition,Partitions,Time_Steps
2, 0.14886279, 1000, 1000, 45
/threads{locality#0/total}/count/pending-accesses,1,0.157788,[s],179778
/threads{locality#0/worker-thread#0}/count/pending-accesses,1,0.157897,[s],53157
/threads{locality#0/worker-thread#1}/count/pending-accesses,1,0.157833,[s],23033
/count/pending-misses:
OS_Threads,Execution_Time_sec,Points_per_Partition,Partitions,Time_Steps
2, 0.150670529, 1000, 1000, 45
/threads{locality#0/total}/count/pending-misses,1,0.160003,[s],122847
/threads{locality#0/worker-thread#0}/count/pending-misses,1,0.160061,[s],52292
/threads{locality#0/worker-thread#1}/count/pending-misses,1,0.160091,[s],8937
/count/stolen-to-pending:
OS_Threads,Execution_Time_sec,Points_per_Partition,Partitions,Time_Steps
2, 0.147684702, 1000, 1000, 45
/threads{locality#0/total}/count/stolen-to-pending,1,0.156292,[s],10580
/threads{locality#0/worker-thread#0}/count/stolen-to-pending,1,0.156333,[s],13
/threads{locality#0/worker-thread#1}/count/stolen-to-pending,1,0.156282,[s],29
/count/stolen-from-pending:
OS_Threads,Execution_Time_sec,Points_per_Partition,Partitions,Time_Steps
2, 0.150269219, 1000, 1000, 45
/threads{locality#0/total}/count/stolen-from-pending,1,0.159178,[s],9343
/threads{locality#0/worker-thread#0}/count/stolen-from-pending,1,0.159268,[s],20
/threads{locality#0/worker-thread#1}/count/stolen-from-pending,1,0.159230,[s],50
|
non_process
|
thread performance counters pending queue counts totals disagree with worker threads command bin stencil nx np hpx print counter threads locality total hpx print counter threads locality worker thread counters where individual worker thread counts disagree with total include count pending accesses count pending misses count stolen to pending count stolen from pending examples count pending accesses os threads execution time sec points per partition partitions time steps threads locality total count pending accesses threads locality worker thread count pending accesses threads locality worker thread count pending accesses count pending misses os threads execution time sec points per partition partitions time steps threads locality total count pending misses threads locality worker thread count pending misses threads locality worker thread count pending misses count pending accesses os threads execution time sec points per partition partitions time steps threads locality total count pending accesses threads locality worker thread count pending accesses threads locality worker thread count pending accesses count pending misses os threads execution time sec points per partition partitions time steps threads locality total count pending misses threads locality worker thread count pending misses threads locality worker thread count pending misses count stolen to pending os threads execution time sec points per partition partitions time steps threads locality total count stolen to pending threads locality worker thread count stolen to pending threads locality worker thread count stolen to pending count stolen from pending os threads execution time sec points per partition partitions time steps threads locality total count stolen from pending threads locality worker thread count stolen from pending threads locality worker thread count stolen from pending
| 0
|
317,897
| 27,273,527,664
|
IssuesEvent
|
2023-02-23 01:32:33
|
MPMG-DCC-UFMG/F01
|
https://api.github.com/repos/MPMG-DCC-UFMG/F01
|
closed
|
Teste de generalizacao para a tag Contratos - Dados dos Contratos - Rio Acima
|
generalization test development
|
DoD: Realizar o teste de Generalização do validador da tag Contratos - Dados dos Contratos para o Município de Rio Acima.
|
1.0
|
Teste de generalizacao para a tag Contratos - Dados dos Contratos - Rio Acima - DoD: Realizar o teste de Generalização do validador da tag Contratos - Dados dos Contratos para o Município de Rio Acima.
|
non_process
|
teste de generalizacao para a tag contratos dados dos contratos rio acima dod realizar o teste de generalização do validador da tag contratos dados dos contratos para o município de rio acima
| 0
|
24,567
| 7,526,599,311
|
IssuesEvent
|
2018-04-13 14:29:17
|
flyve-mdm/web-mdm-dashboard
|
https://api.github.com/repos/flyve-mdm/web-mdm-dashboard
|
opened
|
Build for translations
|
bug build
|
Hi, guys.
I already translated and reviewed all strings in the Transifex (Portuguese, French, Spanish).
Could you take a look what happen with the build?
### Observed Results

### Expected behavior:
|
1.0
|
Build for translations - Hi, guys.
I already translated and reviewed all strings in the Transifex (Portuguese, French, Spanish).
Could you take a look what happen with the build?
### Observed Results

### Expected behavior:
|
non_process
|
build for translations hi guys i already translated and reviewed all strings in the transifex portuguese french spanish could you take a look what happen with the build observed results expected behavior
| 0
|
41,431
| 8,971,856,255
|
IssuesEvent
|
2019-01-29 16:49:36
|
alechamed/decideEuropa-censo
|
https://api.github.com/repos/alechamed/decideEuropa-censo
|
opened
|
Automatización de despliegue en heroku
|
EST_NEW PRI_HIGH TEM_CODE TYP_ENHACEMENT
|
Voy a realizar la automatización para que cuando se haga un push en master se suba una copia a heroku automáticamente desde travis, ya que actualmente para que se despliegue hay que subirlo manualmente a heroku.
|
1.0
|
Automatización de despliegue en heroku - Voy a realizar la automatización para que cuando se haga un push en master se suba una copia a heroku automáticamente desde travis, ya que actualmente para que se despliegue hay que subirlo manualmente a heroku.
|
non_process
|
automatización de despliegue en heroku voy a realizar la automatización para que cuando se haga un push en master se suba una copia a heroku automáticamente desde travis ya que actualmente para que se despliegue hay que subirlo manualmente a heroku
| 0
|
4,700
| 7,542,642,519
|
IssuesEvent
|
2018-04-17 13:34:22
|
inviwo/inviwo
|
https://api.github.com/repos/inviwo/inviwo
|
closed
|
Composite Processor creation upon deserialization issue
|
P: Composite Processor T: Bug
|
It seems to be an issue when trying to deserialize a network with with a Composite Processor in it.
Notice it when trying to load a network with a composite processor and got deserialization errors.
This also seem to affect copy pasting of Composite Processors.
**Steps to reproduce**
1. Load the boron network
2. Select any set of processor and create a composite out of them (I selected the proxy+ entryexit + raycaster)
3. Copy and paste the create composite processor
Result:
```
Could not create Processor: "composite" of class "org.inviwo.CompositeProcessor" at line: 1. Reason: "org.inviwo.CompositeProcessor" Not found in factory.
```
|
1.0
|
Composite Processor creation upon deserialization issue - It seems to be an issue when trying to deserialize a network with with a Composite Processor in it.
Notice it when trying to load a network with a composite processor and got deserialization errors.
This also seem to affect copy pasting of Composite Processors.
**Steps to reproduce**
1. Load the boron network
2. Select any set of processor and create a composite out of them (I selected the proxy+ entryexit + raycaster)
3. Copy and paste the create composite processor
Result:
```
Could not create Processor: "composite" of class "org.inviwo.CompositeProcessor" at line: 1. Reason: "org.inviwo.CompositeProcessor" Not found in factory.
```
|
process
|
composite processor creation upon deserialization issue it seems to be an issue when trying to deserialize a network with with a composite processor in it notice it when trying to load a network with a composite processor and got deserialization errors this also seem to affect copy pasting of composite processors steps to reproduce load the boron network select any set of processor and create a composite out of them i selected the proxy entryexit raycaster copy and paste the create composite processor result could not create processor composite of class org inviwo compositeprocessor at line reason org inviwo compositeprocessor not found in factory
| 1
|
656,873
| 21,778,732,573
|
IssuesEvent
|
2022-05-13 16:17:57
|
proselog/proselog
|
https://api.github.com/repos/proselog/proselog
|
closed
|
Allow to become a subscriber
|
priority
|
- [x] add subscriber to our database
- [ ] add subscriber to mailgun mail list
|
1.0
|
Allow to become a subscriber - - [x] add subscriber to our database
- [ ] add subscriber to mailgun mail list
|
non_process
|
allow to become a subscriber add subscriber to our database add subscriber to mailgun mail list
| 0
|
64,857
| 12,500,425,079
|
IssuesEvent
|
2020-06-01 22:15:03
|
mjseok/OSS_team1
|
https://api.github.com/repos/mjseok/OSS_team1
|
closed
|
Functionalize_playGame
|
clean code
|
코드를 함수화하여 main을 최대한 간단하게 작성하고 라벨을 지우고 goto 구문도 없애기
게임실행하는 부분을 없애기( menu=1)선택했을 때의 부분
|
1.0
|
Functionalize_playGame - 코드를 함수화하여 main을 최대한 간단하게 작성하고 라벨을 지우고 goto 구문도 없애기
게임실행하는 부분을 없애기( menu=1)선택했을 때의 부분
|
non_process
|
functionalize playgame 코드를 함수화하여 main을 최대한 간단하게 작성하고 라벨을 지우고 goto 구문도 없애기 게임실행하는 부분을 없애기 menu 선택했을 때의 부분
| 0
|
266,764
| 28,435,915,665
|
IssuesEvent
|
2023-04-15 10:00:51
|
dotnet/aspnetcore
|
https://api.github.com/repos/dotnet/aspnetcore
|
closed
|
I configure the server to requiere client certificate but it accepts the request when client doesn't add a certificate
|
area-security :heavy_check_mark: Resolution: Answered Status: Resolved
|
### Is there an existing issue for this?
- [X] I have searched the existing issues
### Describe the bug
I would like to implement mTLS in a service, but when create the client without adding the certificate, the server accepts the request.
This is the code of my server, in the program.cs file of my ASP Core 7 application:
```
try
{
var builder = WebApplication.CreateBuilder(args);
builder.WebHost.ConfigureKestrel((context, options) =>
{
string miStrCertificado = File.ReadAllText("certificados/server.crt");
string miStrKey = File.ReadAllText("certificados/server.key");
X509Certificate2 miCertficadoX509 = X509Certificate2.CreateFromPem(miStrCertificado, miStrKey);
X509Certificate2 miCertificado2 = new X509Certificate2(miCertficadoX509.Export(X509ContentType.Pkcs12));
miCertficadoX509.Dispose();
options.ListenAnyIP(5001, listenOptions =>
{
listenOptions.Protocols = HttpProtocols.Http2;
listenOptions.UseHttps(miCertificado2);
});
options.ConfigureHttpsDefaults(miHttpsOptions =>
{
miHttpsOptions.ClientCertificateMode = ClientCertificateMode.RequireCertificate;
miHttpsOptions.ServerCertificate = miCertificado2;
});
});
builder.Services.AddHttpContextAccessor();
builder.Services.AddGrpc();
builder.Services.AddCodeFirstGrpc();
builder.Services.AddAuthentication(JwtBearerDefaults.AuthenticationScheme)
.AddJwtBearer(options =>
{
options.TokenValidationParameters = new TokenValidationParameters
{
//@#ESTUDIAR: ¿el string utilizado tiene que tener 64 caracteres? No parece pero parece ser que no
//vale cualquier longitud. Igual tiene que ser un string que codifique datos binarios en base 4.
IssuerSigningKey = new SymmetricSecurityKey(System.Text.Encoding.Default.GetBytes("ABCD")),
RequireExpirationTime = true,
RequireSignedTokens = true,
ClockSkew = TimeSpan.FromSeconds(10),
ValidateIssuer = false,
ValidateAudience = false,
ValidateLifetime = true,
ValidateIssuerSigningKey = true,
};
});
builder.Services.AddAuthorization();
var app = builder.Build();
app.UseRouting();
app.UseAuthentication();
app.UseAuthorization();
app.MapGrpcService<MyServiceLogin>();
app.MapGrpcService<MyMainService>();
await app.RunAsync();
}
catch (Exception ex)
{
Console.WriteLine(ex.Message);
}
```
This is how I create my client:
```
private ChannelCredentials CrearCredencialesDelCanal()
{
CallCredentials misCredencialesDeLlamada = CallCredentials.FromInterceptor((c, m) =>
{
m.Add(HeaderNames.Authorization, $"Bearer {_tokenJwt}");
return Task.CompletedTask;
});
return ChannelCredentials.Create(new SslCredentials(), misCredencialesDeLlamada);
}
private GrpcChannelOptions CrearOpcionesDelCanal()
{
HttpClientHandler miHttpHandler = new HttpClientHandler();
//Always return true
miHttpHandler.ServerCertificateCustomValidationCallback = ValidateServerCertificate;
HttpClient httpClient = new HttpClient(miHttpHandler);
GrpcChannelOptions misOpcionesDelCanal = new GrpcChannelOptions()
{
MaxReceiveMessageSize = 62914560,
MaxSendMessageSize = 62914560,
HttpClient = httpClient,
};
misOpcionesDelCanal.Credentials = CrearCredencialesDelCanal();
return misOpcionesDelCanal;
}
private GrpcChannel CrearCanal(string paramStrDireccion, int paramIntPuerto)
{
GrpcChannelOptions misOpciones = CrearOpcionesDelCanal();
try
{
Uri miUri = new Uri(paramStrDireccion + ":" + paramIntPuerto.ToString());
GrpcChannel miCanalGrpc = GrpcChannel.ForAddress(miUri, misOpciones);
OnCanalGrpcActualizado(miCanalGrpc);
return miCanalGrpc;
}
catch
{
throw new ArgumentException("Los datos del servidor no son correctos.");
}
}
```
In this example, when I create the HttClientHandler, I don't set the client certificate, but the server accepts the request and replies.
How could I set the server to requiere a client certificate and if it is not add or it is not a valid certificate, refuse the request?
Thanks.
### Expected Behavior
The server should to denied or not reply to a client when it doesn't add a valid certificate.
### Steps To Reproduce
_No response_
### Exceptions (if any)
_No response_
### .NET Version
_No response_
### Anything else?
_No response_
|
True
|
I configure the server to requiere client certificate but it accepts the request when client doesn't add a certificate - ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Describe the bug
I would like to implement mTLS in a service, but when create the client without adding the certificate, the server accepts the request.
This is the code of my server, in the program.cs file of my ASP Core 7 application:
```
try
{
var builder = WebApplication.CreateBuilder(args);
builder.WebHost.ConfigureKestrel((context, options) =>
{
string miStrCertificado = File.ReadAllText("certificados/server.crt");
string miStrKey = File.ReadAllText("certificados/server.key");
X509Certificate2 miCertficadoX509 = X509Certificate2.CreateFromPem(miStrCertificado, miStrKey);
X509Certificate2 miCertificado2 = new X509Certificate2(miCertficadoX509.Export(X509ContentType.Pkcs12));
miCertficadoX509.Dispose();
options.ListenAnyIP(5001, listenOptions =>
{
listenOptions.Protocols = HttpProtocols.Http2;
listenOptions.UseHttps(miCertificado2);
});
options.ConfigureHttpsDefaults(miHttpsOptions =>
{
miHttpsOptions.ClientCertificateMode = ClientCertificateMode.RequireCertificate;
miHttpsOptions.ServerCertificate = miCertificado2;
});
});
builder.Services.AddHttpContextAccessor();
builder.Services.AddGrpc();
builder.Services.AddCodeFirstGrpc();
builder.Services.AddAuthentication(JwtBearerDefaults.AuthenticationScheme)
.AddJwtBearer(options =>
{
options.TokenValidationParameters = new TokenValidationParameters
{
//@#ESTUDIAR: ¿el string utilizado tiene que tener 64 caracteres? No parece pero parece ser que no
//vale cualquier longitud. Igual tiene que ser un string que codifique datos binarios en base 4.
IssuerSigningKey = new SymmetricSecurityKey(System.Text.Encoding.Default.GetBytes("ABCD")),
RequireExpirationTime = true,
RequireSignedTokens = true,
ClockSkew = TimeSpan.FromSeconds(10),
ValidateIssuer = false,
ValidateAudience = false,
ValidateLifetime = true,
ValidateIssuerSigningKey = true,
};
});
builder.Services.AddAuthorization();
var app = builder.Build();
app.UseRouting();
app.UseAuthentication();
app.UseAuthorization();
app.MapGrpcService<MyServiceLogin>();
app.MapGrpcService<MyMainService>();
await app.RunAsync();
}
catch (Exception ex)
{
Console.WriteLine(ex.Message);
}
```
This is how I create my client:
```
private ChannelCredentials CrearCredencialesDelCanal()
{
CallCredentials misCredencialesDeLlamada = CallCredentials.FromInterceptor((c, m) =>
{
m.Add(HeaderNames.Authorization, $"Bearer {_tokenJwt}");
return Task.CompletedTask;
});
return ChannelCredentials.Create(new SslCredentials(), misCredencialesDeLlamada);
}
private GrpcChannelOptions CrearOpcionesDelCanal()
{
HttpClientHandler miHttpHandler = new HttpClientHandler();
//Always return true
miHttpHandler.ServerCertificateCustomValidationCallback = ValidateServerCertificate;
HttpClient httpClient = new HttpClient(miHttpHandler);
GrpcChannelOptions misOpcionesDelCanal = new GrpcChannelOptions()
{
MaxReceiveMessageSize = 62914560,
MaxSendMessageSize = 62914560,
HttpClient = httpClient,
};
misOpcionesDelCanal.Credentials = CrearCredencialesDelCanal();
return misOpcionesDelCanal;
}
private GrpcChannel CrearCanal(string paramStrDireccion, int paramIntPuerto)
{
GrpcChannelOptions misOpciones = CrearOpcionesDelCanal();
try
{
Uri miUri = new Uri(paramStrDireccion + ":" + paramIntPuerto.ToString());
GrpcChannel miCanalGrpc = GrpcChannel.ForAddress(miUri, misOpciones);
OnCanalGrpcActualizado(miCanalGrpc);
return miCanalGrpc;
}
catch
{
throw new ArgumentException("Los datos del servidor no son correctos.");
}
}
```
In this example, when I create the HttClientHandler, I don't set the client certificate, but the server accepts the request and replies.
How could I set the server to requiere a client certificate and if it is not add or it is not a valid certificate, refuse the request?
Thanks.
### Expected Behavior
The server should to denied or not reply to a client when it doesn't add a valid certificate.
### Steps To Reproduce
_No response_
### Exceptions (if any)
_No response_
### .NET Version
_No response_
### Anything else?
_No response_
|
non_process
|
i configure the server to requiere client certificate but it accepts the request when client doesn t add a certificate is there an existing issue for this i have searched the existing issues describe the bug i would like to implement mtls in a service but when create the client without adding the certificate the server accepts the request this is the code of my server in the program cs file of my asp core application try var builder webapplication createbuilder args builder webhost configurekestrel context options string mistrcertificado file readalltext certificados server crt string mistrkey file readalltext certificados server key createfrompem mistrcertificado mistrkey new export dispose options listenanyip listenoptions listenoptions protocols httpprotocols listenoptions usehttps options configurehttpsdefaults mihttpsoptions mihttpsoptions clientcertificatemode clientcertificatemode requirecertificate mihttpsoptions servercertificate builder services addhttpcontextaccessor builder services addgrpc builder services addcodefirstgrpc builder services addauthentication jwtbearerdefaults authenticationscheme addjwtbearer options options tokenvalidationparameters new tokenvalidationparameters estudiar ¿el string utilizado tiene que tener caracteres no parece pero parece ser que no vale cualquier longitud igual tiene que ser un string que codifique datos binarios en base issuersigningkey new symmetricsecuritykey system text encoding default getbytes abcd requireexpirationtime true requiresignedtokens true clockskew timespan fromseconds validateissuer false validateaudience false validatelifetime true validateissuersigningkey true builder services addauthorization var app builder build app userouting app useauthentication app useauthorization app mapgrpcservice app mapgrpcservice await app runasync catch exception ex console writeline ex message this is how i create my client private channelcredentials crearcredencialesdelcanal callcredentials miscredencialesdellamada callcredentials frominterceptor c m m add headernames authorization bearer tokenjwt return task completedtask return channelcredentials create new sslcredentials miscredencialesdellamada private grpcchanneloptions crearopcionesdelcanal httpclienthandler mihttphandler new httpclienthandler always return true mihttphandler servercertificatecustomvalidationcallback validateservercertificate httpclient httpclient new httpclient mihttphandler grpcchanneloptions misopcionesdelcanal new grpcchanneloptions maxreceivemessagesize maxsendmessagesize httpclient httpclient misopcionesdelcanal credentials crearcredencialesdelcanal return misopcionesdelcanal private grpcchannel crearcanal string paramstrdireccion int paramintpuerto grpcchanneloptions misopciones crearopcionesdelcanal try uri miuri new uri paramstrdireccion paramintpuerto tostring grpcchannel micanalgrpc grpcchannel foraddress miuri misopciones oncanalgrpcactualizado micanalgrpc return micanalgrpc catch throw new argumentexception los datos del servidor no son correctos in this example when i create the httclienthandler i don t set the client certificate but the server accepts the request and replies how could i set the server to requiere a client certificate and if it is not add or it is not a valid certificate refuse the request thanks expected behavior the server should to denied or not reply to a client when it doesn t add a valid certificate steps to reproduce no response exceptions if any no response net version no response anything else no response
| 0
|
137,742
| 30,745,626,555
|
IssuesEvent
|
2023-07-28 14:50:26
|
WordPress/openverse
|
https://api.github.com/repos/WordPress/openverse
|
closed
|
Baseline monitoring changes for Terraform and the unhealthy host count alarm
|
🟧 priority: high 🌟 goal: addition 💻 aspect: code 🧱 stack: infra
|
## Description
<!-- Describe the feature and how it solves the problem. -->
Project thread: https://github.com/WordPress/openverse/issues/2344
Implementation plan: https://docs.openverse.org/projects/proposals/monitoring/20230606_implementation_plan_ecs_alarms.html
1. Create the monitoring modules for frontend and API staging and production and move existing alarms into these
- Create a new `next/modules/monitoring` directory with a directory for each service/environment: `staging-frontend`, `production-frontend`, `staging-api`, etc.
- This includes moving the UptimeRobot configuration for each service as well as the database and Redis monitors
- Rename `service-monitors` to `service-uptime-robot` to clarify the module's purpose
- Also create the new SNS topic for the unstable alerts' notification channel
- This does not include moving the ECS service cloudwatch dashboard module, that should remain in the root modules
1. Create the unhealthy host count alarm for production and staging services
> **Note**
>
> The only thing that will be present in the staging monitoring modules for each service is the unhealthy host count alarm.
## Additional context
<!-- Add any other context about the feature here; or delete the section entirely. -->
This issue will remain open until the unhealthy host count alarm is stabilised. However, once the terraform configuration changes and the new monitoring modules are present, all other alarm issues in this milestone will be unblocked.
<!-- If you would like to work on this, please comment below separately. -->
|
1.0
|
Baseline monitoring changes for Terraform and the unhealthy host count alarm - ## Description
<!-- Describe the feature and how it solves the problem. -->
Project thread: https://github.com/WordPress/openverse/issues/2344
Implementation plan: https://docs.openverse.org/projects/proposals/monitoring/20230606_implementation_plan_ecs_alarms.html
1. Create the monitoring modules for frontend and API staging and production and move existing alarms into these
- Create a new `next/modules/monitoring` directory with a directory for each service/environment: `staging-frontend`, `production-frontend`, `staging-api`, etc.
- This includes moving the UptimeRobot configuration for each service as well as the database and Redis monitors
- Rename `service-monitors` to `service-uptime-robot` to clarify the module's purpose
- Also create the new SNS topic for the unstable alerts' notification channel
- This does not include moving the ECS service cloudwatch dashboard module, that should remain in the root modules
1. Create the unhealthy host count alarm for production and staging services
> **Note**
>
> The only thing that will be present in the staging monitoring modules for each service is the unhealthy host count alarm.
## Additional context
<!-- Add any other context about the feature here; or delete the section entirely. -->
This issue will remain open until the unhealthy host count alarm is stabilised. However, once the terraform configuration changes and the new monitoring modules are present, all other alarm issues in this milestone will be unblocked.
<!-- If you would like to work on this, please comment below separately. -->
|
non_process
|
baseline monitoring changes for terraform and the unhealthy host count alarm description project thread implementation plan create the monitoring modules for frontend and api staging and production and move existing alarms into these create a new next modules monitoring directory with a directory for each service environment staging frontend production frontend staging api etc this includes moving the uptimerobot configuration for each service as well as the database and redis monitors rename service monitors to service uptime robot to clarify the module s purpose also create the new sns topic for the unstable alerts notification channel this does not include moving the ecs service cloudwatch dashboard module that should remain in the root modules create the unhealthy host count alarm for production and staging services note the only thing that will be present in the staging monitoring modules for each service is the unhealthy host count alarm additional context this issue will remain open until the unhealthy host count alarm is stabilised however once the terraform configuration changes and the new monitoring modules are present all other alarm issues in this milestone will be unblocked
| 0
|
10,722
| 13,524,633,324
|
IssuesEvent
|
2020-09-15 11:51:46
|
gfx-rs/naga
|
https://api.github.com/repos/gfx-rs/naga
|
closed
|
Use typifier in the backends
|
area: back-end area: processing help wanted kind: refactor
|
Currently it's used in the frontends only, and it mutates the type arena.
GLSL and MSL backends would benefit from having this automated too, but we need to do something about the mutable arena.
|
1.0
|
Use typifier in the backends - Currently it's used in the frontends only, and it mutates the type arena.
GLSL and MSL backends would benefit from having this automated too, but we need to do something about the mutable arena.
|
process
|
use typifier in the backends currently it s used in the frontends only and it mutates the type arena glsl and msl backends would benefit from having this automated too but we need to do something about the mutable arena
| 1
|
1,142
| 3,631,346,238
|
IssuesEvent
|
2016-02-11 00:54:04
|
ParsePlatform/parse-server
|
https://api.github.com/repos/ParsePlatform/parse-server
|
closed
|
ParseServer does not delete Session object on logout
|
bug in-process
|
I have noticed that the behavior of hosted Parse is to create a new Session object when a user login and delete this object when this user logout from the same device.
When I tested the ParseServer version instead of the hosted Parse I noticed that this behavior is different, the Session object is created on login as expected, but on logout the object is not deleted, so if the user keeps logging in and logging out there will be as many Session objects on DB as the number of times the user login and logout.
Is this the expected behavior or is it a bug on ParseServer?
|
1.0
|
ParseServer does not delete Session object on logout - I have noticed that the behavior of hosted Parse is to create a new Session object when a user login and delete this object when this user logout from the same device.
When I tested the ParseServer version instead of the hosted Parse I noticed that this behavior is different, the Session object is created on login as expected, but on logout the object is not deleted, so if the user keeps logging in and logging out there will be as many Session objects on DB as the number of times the user login and logout.
Is this the expected behavior or is it a bug on ParseServer?
|
process
|
parseserver does not delete session object on logout i have noticed that the behavior of hosted parse is to create a new session object when a user login and delete this object when this user logout from the same device when i tested the parseserver version instead of the hosted parse i noticed that this behavior is different the session object is created on login as expected but on logout the object is not deleted so if the user keeps logging in and logging out there will be as many session objects on db as the number of times the user login and logout is this the expected behavior or is it a bug on parseserver
| 1
|
13,188
| 8,824,079,554
|
IssuesEvent
|
2019-01-02 15:50:46
|
arcus-azure/arcus.security
|
https://api.github.com/repos/arcus-azure/arcus.security
|
closed
|
Provide support for Managed identities for Azure resources authentication
|
azure-key-vault feature security
|
Provide support for Managed identities for Azure resources authentication, formerly known as Azure AD Managed Service Identity (MSI).
|
True
|
Provide support for Managed identities for Azure resources authentication - Provide support for Managed identities for Azure resources authentication, formerly known as Azure AD Managed Service Identity (MSI).
|
non_process
|
provide support for managed identities for azure resources authentication provide support for managed identities for azure resources authentication formerly known as azure ad managed service identity msi
| 0
|
11,679
| 14,537,160,807
|
IssuesEvent
|
2020-12-15 08:46:27
|
threefoldtech/js-sdk
|
https://api.github.com/repos/threefoldtech/js-sdk
|
closed
|
Unable to start threebot server after fresh install
|
process_wontfix type_bug
|
After installing a fresh js-sdk (following the instructions on the wiki), I am unable to start a threebot. It keeps failing on the fact that I don't have an identity (and to get one I presumably need to run the threebot first). Passing the `--development` flag, which claims no identity is needed, does not solve the issue.
```
threebot start --local --development
2020-11-20 23:27:38.978 | ERROR | jumpscale.tools.errorhandler.errorhandler:_handle_exception:93 - jumpscale.core.exceptions.exceptions.Value: No configured identity found
Traceback (most recent call last):
> File "/home/lee/.cache/pypoetry/virtualenvs/js-sdk-4phGcms2-py3.8/bin/threebot", line 5, in <module>
cli()
└ <Group cli>
File "/home/lee/.cache/pypoetry/virtualenvs/js-sdk-4phGcms2-py3.8/lib/python3.8/site-packages/click/core.py", line 829, in __call__
return self.main(*args, **kwargs)
│ │ │ └ {}
│ │ └ ()
│ └ <function BaseCommand.main at 0x7ff40a282ca0>
└ <Group cli>
File "/home/lee/.cache/pypoetry/virtualenvs/js-sdk-4phGcms2-py3.8/lib/python3.8/site-packages/click/core.py", line 782, in main
rv = self.invoke(ctx)
│ │ └ <click.core.Context object at 0x7ff40b1740a0>
│ └ <function MultiCommand.invoke at 0x7ff40a283a60>
└ <Group cli>
File "/home/lee/.cache/pypoetry/virtualenvs/js-sdk-4phGcms2-py3.8/lib/python3.8/site-packages/click/core.py", line 1259, in invoke
return _process_result(sub_ctx.command.invoke(sub_ctx))
│ │ │ │ └ <click.core.Context object at 0x7ff40b203580>
│ │ │ └ <function Command.invoke at 0x7ff40a283670>
│ │ └ <Command start>
│ └ <click.core.Context object at 0x7ff40b203580>
└ <function MultiCommand.invoke.<locals>._process_result at 0x7ff40b2301f0>
File "/home/lee/.cache/pypoetry/virtualenvs/js-sdk-4phGcms2-py3.8/lib/python3.8/site-packages/click/core.py", line 1066, in invoke
return ctx.invoke(self.callback, **ctx.params)
│ │ │ │ │ └ {'local': True, 'development': True, 'identity': None, 'domain': None, 'email': None, 'background': False}
│ │ │ │ └ <click.core.Context object at 0x7ff40b203580>
│ │ │ └ <function start at 0x7ff40712d790>
│ │ └ <Command start>
│ └ <function Context.invoke at 0x7ff40a282790>
└ <click.core.Context object at 0x7ff40b203580>
File "/home/lee/.cache/pypoetry/virtualenvs/js-sdk-4phGcms2-py3.8/lib/python3.8/site-packages/click/core.py", line 610, in invoke
return callback(*args, **kwargs)
│ │ └ {'local': True, 'development': True, 'identity': None, 'domain': None, 'email': None, 'background': False}
│ └ ()
└ <function start at 0x7ff40712d790>
File "/home/lee/js-sdk/jumpscale/entry_points/threebot.py", line 86, in start
create_wallets_if_not_exists()
└ <function create_wallets_if_not_exists at 0x7ff40712d820>
File "/home/lee/js-sdk/jumpscale/entry_points/threebot.py", line 277, in create_wallets_if_not_exists
if not test and "testnet" in j.core.identity.me.explorer_url:
│ │ └ <property object at 0x7ff40a288cc0>
│ └ <jumpscale.loader.J object at 0x7ff40a27c8b0>
└ False
File "/home/lee/js-sdk/jumpscale/core/identity/__init__.py", line 208, in me
raise Value("No configured identity found")
└ <class 'jumpscale.core.exceptions.exceptions.Value'>
jumpscale.core.exceptions.exceptions.Value: No configured identity found
```
|
1.0
|
Unable to start threebot server after fresh install - After installing a fresh js-sdk (following the instructions on the wiki), I am unable to start a threebot. It keeps failing on the fact that I don't have an identity (and to get one I presumably need to run the threebot first). Passing the `--development` flag, which claims no identity is needed, does not solve the issue.
```
threebot start --local --development
2020-11-20 23:27:38.978 | ERROR | jumpscale.tools.errorhandler.errorhandler:_handle_exception:93 - jumpscale.core.exceptions.exceptions.Value: No configured identity found
Traceback (most recent call last):
> File "/home/lee/.cache/pypoetry/virtualenvs/js-sdk-4phGcms2-py3.8/bin/threebot", line 5, in <module>
cli()
└ <Group cli>
File "/home/lee/.cache/pypoetry/virtualenvs/js-sdk-4phGcms2-py3.8/lib/python3.8/site-packages/click/core.py", line 829, in __call__
return self.main(*args, **kwargs)
│ │ │ └ {}
│ │ └ ()
│ └ <function BaseCommand.main at 0x7ff40a282ca0>
└ <Group cli>
File "/home/lee/.cache/pypoetry/virtualenvs/js-sdk-4phGcms2-py3.8/lib/python3.8/site-packages/click/core.py", line 782, in main
rv = self.invoke(ctx)
│ │ └ <click.core.Context object at 0x7ff40b1740a0>
│ └ <function MultiCommand.invoke at 0x7ff40a283a60>
└ <Group cli>
File "/home/lee/.cache/pypoetry/virtualenvs/js-sdk-4phGcms2-py3.8/lib/python3.8/site-packages/click/core.py", line 1259, in invoke
return _process_result(sub_ctx.command.invoke(sub_ctx))
│ │ │ │ └ <click.core.Context object at 0x7ff40b203580>
│ │ │ └ <function Command.invoke at 0x7ff40a283670>
│ │ └ <Command start>
│ └ <click.core.Context object at 0x7ff40b203580>
└ <function MultiCommand.invoke.<locals>._process_result at 0x7ff40b2301f0>
File "/home/lee/.cache/pypoetry/virtualenvs/js-sdk-4phGcms2-py3.8/lib/python3.8/site-packages/click/core.py", line 1066, in invoke
return ctx.invoke(self.callback, **ctx.params)
│ │ │ │ │ └ {'local': True, 'development': True, 'identity': None, 'domain': None, 'email': None, 'background': False}
│ │ │ │ └ <click.core.Context object at 0x7ff40b203580>
│ │ │ └ <function start at 0x7ff40712d790>
│ │ └ <Command start>
│ └ <function Context.invoke at 0x7ff40a282790>
└ <click.core.Context object at 0x7ff40b203580>
File "/home/lee/.cache/pypoetry/virtualenvs/js-sdk-4phGcms2-py3.8/lib/python3.8/site-packages/click/core.py", line 610, in invoke
return callback(*args, **kwargs)
│ │ └ {'local': True, 'development': True, 'identity': None, 'domain': None, 'email': None, 'background': False}
│ └ ()
└ <function start at 0x7ff40712d790>
File "/home/lee/js-sdk/jumpscale/entry_points/threebot.py", line 86, in start
create_wallets_if_not_exists()
└ <function create_wallets_if_not_exists at 0x7ff40712d820>
File "/home/lee/js-sdk/jumpscale/entry_points/threebot.py", line 277, in create_wallets_if_not_exists
if not test and "testnet" in j.core.identity.me.explorer_url:
│ │ └ <property object at 0x7ff40a288cc0>
│ └ <jumpscale.loader.J object at 0x7ff40a27c8b0>
└ False
File "/home/lee/js-sdk/jumpscale/core/identity/__init__.py", line 208, in me
raise Value("No configured identity found")
└ <class 'jumpscale.core.exceptions.exceptions.Value'>
jumpscale.core.exceptions.exceptions.Value: No configured identity found
```
|
process
|
unable to start threebot server after fresh install after installing a fresh js sdk following the instructions on the wiki i am unable to start a threebot it keeps failing on the fact that i don t have an identity and to get one i presumably need to run the threebot first passing the development flag which claims no identity is needed does not solve the issue threebot start local development error jumpscale tools errorhandler errorhandler handle exception jumpscale core exceptions exceptions value no configured identity found traceback most recent call last file home lee cache pypoetry virtualenvs js sdk bin threebot line in cli └ file home lee cache pypoetry virtualenvs js sdk lib site packages click core py line in call return self main args kwargs │ │ │ └ │ │ └ │ └ └ file home lee cache pypoetry virtualenvs js sdk lib site packages click core py line in main rv self invoke ctx │ │ └ │ └ └ file home lee cache pypoetry virtualenvs js sdk lib site packages click core py line in invoke return process result sub ctx command invoke sub ctx │ │ │ │ └ │ │ │ └ │ │ └ │ └ └ process result at file home lee cache pypoetry virtualenvs js sdk lib site packages click core py line in invoke return ctx invoke self callback ctx params │ │ │ │ │ └ local true development true identity none domain none email none background false │ │ │ │ └ │ │ │ └ │ │ └ │ └ └ file home lee cache pypoetry virtualenvs js sdk lib site packages click core py line in invoke return callback args kwargs │ │ └ local true development true identity none domain none email none background false │ └ └ file home lee js sdk jumpscale entry points threebot py line in start create wallets if not exists └ file home lee js sdk jumpscale entry points threebot py line in create wallets if not exists if not test and testnet in j core identity me explorer url │ │ └ │ └ └ false file home lee js sdk jumpscale core identity init py line in me raise value no configured identity found └ jumpscale core exceptions exceptions value no configured identity found
| 1
|
494,003
| 14,243,224,443
|
IssuesEvent
|
2020-11-19 03:51:08
|
jcr7467/UCLAbookstack
|
https://api.github.com/repos/jcr7467/UCLAbookstack
|
opened
|
Add Book Quality Feature
|
Priority - Medium new feature
|
We have filters for the subject areas, we also need a filter for the book quality!
|
1.0
|
Add Book Quality Feature - We have filters for the subject areas, we also need a filter for the book quality!
|
non_process
|
add book quality feature we have filters for the subject areas we also need a filter for the book quality
| 0
|
402,850
| 27,389,629,178
|
IssuesEvent
|
2023-02-28 15:28:44
|
weaveworks/weave-gitops
|
https://api.github.com/repos/weaveworks/weave-gitops
|
closed
|
docs: consider breaking out the CLI installations into a separate page
|
documentation
|
Personally I would also break out the CLI installations into a separate page
_Originally posted by @sympatheticmoose in https://github.com/weaveworks/weave-gitops/issues/3039#issuecomment-1324887551_
|
1.0
|
docs: consider breaking out the CLI installations into a separate page - Personally I would also break out the CLI installations into a separate page
_Originally posted by @sympatheticmoose in https://github.com/weaveworks/weave-gitops/issues/3039#issuecomment-1324887551_
|
non_process
|
docs consider breaking out the cli installations into a separate page personally i would also break out the cli installations into a separate page originally posted by sympatheticmoose in
| 0
|
2,663
| 5,436,410,569
|
IssuesEvent
|
2017-03-06 00:47:10
|
jlm2017/jlm-video-subtitles
|
https://api.github.com/repos/jlm2017/jlm-video-subtitles
|
reopened
|
[subtitles] [en] MONSIEUR HAMON, CHOISISSEZ !
|
Process: [6] Approved ⚑ English
|
# Video title
MONSIEUR HAMON, CHOISISSEZ !
# URL
https://www.youtube.com/watch?v=tQIsWprmxDA&t=
# Youtube subtitles language
Langue des sous-titres (Anglais)
# Duration
10:44
# Subtitles URL
https://www.youtube.com/timedtext_editor?lang=en&action_mde_edit_form=1&ref=player&tab=captions&v=tQIsWprmxDA&bl=vmp&ui=hd
|
1.0
|
[subtitles] [en] MONSIEUR HAMON, CHOISISSEZ ! - # Video title
MONSIEUR HAMON, CHOISISSEZ !
# URL
https://www.youtube.com/watch?v=tQIsWprmxDA&t=
# Youtube subtitles language
Langue des sous-titres (Anglais)
# Duration
10:44
# Subtitles URL
https://www.youtube.com/timedtext_editor?lang=en&action_mde_edit_form=1&ref=player&tab=captions&v=tQIsWprmxDA&bl=vmp&ui=hd
|
process
|
monsieur hamon choisissez video title monsieur hamon choisissez url youtube subtitles language langue des sous titres anglais duration subtitles url
| 1
|
43,880
| 13,040,724,635
|
IssuesEvent
|
2020-07-28 19:02:03
|
LevyForchh/clusterfuzz
|
https://api.github.com/repos/LevyForchh/clusterfuzz
|
opened
|
CVE-2019-12855 (High) detected in Twisted-14.0.0.tar.bz2
|
security vulnerability
|
## CVE-2019-12855 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Twisted-14.0.0.tar.bz2</b></p></summary>
<p>An asynchronous networking framework written in Python</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/76/38/cf8f81c1d7d84fec922d67f0d92bfa9fee59145d875d7263ceefa2bbbaf4/Twisted-14.0.0.tar.bz2">https://files.pythonhosted.org/packages/76/38/cf8f81c1d7d84fec922d67f0d92bfa9fee59145d875d7263ceefa2bbbaf4/Twisted-14.0.0.tar.bz2</a></p>
<p>Path to dependency file: /tmp/ws-ua_20200728185744_PPILLS/archiveExtraction_CNYBMU/20200728185744/ws-scm_depth_0/clusterfuzz/resources/platform/linux/peach/peach_mutator/peach_mutator/requirements.txt</p>
<p>Path to vulnerable library: _depth_0/clusterfuzz/resources/platform/linux/peach/peach_mutator/peach_mutator/requirements.txt,/clusterfuzz/src/local/butler/scripts,/clusterfuzz/src/python/bot/untrusted_runner/build,_depth_0/clusterfuzz/resources/platform/linux/peach/peach_mutator/peach_mutator/third_party/peach/requirements.txt,/clusterfuzz/src/appengine/handlers/cron/project,/clusterfuzz/src/python/bot/tasks</p>
<p>
Dependency Hierarchy:
- :x: **Twisted-14.0.0.tar.bz2** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/LevyForchh/clusterfuzz/commit/aefe4b0859891117218fba5984e5c3e753ea9597">aefe4b0859891117218fba5984e5c3e753ea9597</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In words.protocols.jabber.xmlstream in Twisted through 19.2.1, XMPP support did not verify certificates when used with TLS, allowing an attacker to MITM connections.
<p>Publish Date: 2019-06-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-12855>CVE-2019-12855</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Change files</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-65rm-h285-5cc5">https://github.com/advisories/GHSA-65rm-h285-5cc5</a></p>
<p>Release Date: 2019-12-19</p>
<p>Fix Resolution: Replace or update the following file: 19.7.0</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Python","packageName":"Twisted","packageVersion":"14.0.0","isTransitiveDependency":false,"dependencyTree":"Twisted:14.0.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"Replace or update the following file: 19.7.0"}],"vulnerabilityIdentifier":"CVE-2019-12855","vulnerabilityDetails":"In words.protocols.jabber.xmlstream in Twisted through 19.2.1, XMPP support did not verify certificates when used with TLS, allowing an attacker to MITM connections.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-12855","cvss3Severity":"high","cvss3Score":"7.4","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2019-12855 (High) detected in Twisted-14.0.0.tar.bz2 - ## CVE-2019-12855 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Twisted-14.0.0.tar.bz2</b></p></summary>
<p>An asynchronous networking framework written in Python</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/76/38/cf8f81c1d7d84fec922d67f0d92bfa9fee59145d875d7263ceefa2bbbaf4/Twisted-14.0.0.tar.bz2">https://files.pythonhosted.org/packages/76/38/cf8f81c1d7d84fec922d67f0d92bfa9fee59145d875d7263ceefa2bbbaf4/Twisted-14.0.0.tar.bz2</a></p>
<p>Path to dependency file: /tmp/ws-ua_20200728185744_PPILLS/archiveExtraction_CNYBMU/20200728185744/ws-scm_depth_0/clusterfuzz/resources/platform/linux/peach/peach_mutator/peach_mutator/requirements.txt</p>
<p>Path to vulnerable library: _depth_0/clusterfuzz/resources/platform/linux/peach/peach_mutator/peach_mutator/requirements.txt,/clusterfuzz/src/local/butler/scripts,/clusterfuzz/src/python/bot/untrusted_runner/build,_depth_0/clusterfuzz/resources/platform/linux/peach/peach_mutator/peach_mutator/third_party/peach/requirements.txt,/clusterfuzz/src/appengine/handlers/cron/project,/clusterfuzz/src/python/bot/tasks</p>
<p>
Dependency Hierarchy:
- :x: **Twisted-14.0.0.tar.bz2** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/LevyForchh/clusterfuzz/commit/aefe4b0859891117218fba5984e5c3e753ea9597">aefe4b0859891117218fba5984e5c3e753ea9597</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In words.protocols.jabber.xmlstream in Twisted through 19.2.1, XMPP support did not verify certificates when used with TLS, allowing an attacker to MITM connections.
<p>Publish Date: 2019-06-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-12855>CVE-2019-12855</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Change files</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-65rm-h285-5cc5">https://github.com/advisories/GHSA-65rm-h285-5cc5</a></p>
<p>Release Date: 2019-12-19</p>
<p>Fix Resolution: Replace or update the following file: 19.7.0</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Python","packageName":"Twisted","packageVersion":"14.0.0","isTransitiveDependency":false,"dependencyTree":"Twisted:14.0.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"Replace or update the following file: 19.7.0"}],"vulnerabilityIdentifier":"CVE-2019-12855","vulnerabilityDetails":"In words.protocols.jabber.xmlstream in Twisted through 19.2.1, XMPP support did not verify certificates when used with TLS, allowing an attacker to MITM connections.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-12855","cvss3Severity":"high","cvss3Score":"7.4","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
non_process
|
cve high detected in twisted tar cve high severity vulnerability vulnerable library twisted tar an asynchronous networking framework written in python library home page a href path to dependency file tmp ws ua ppills archiveextraction cnybmu ws scm depth clusterfuzz resources platform linux peach peach mutator peach mutator requirements txt path to vulnerable library depth clusterfuzz resources platform linux peach peach mutator peach mutator requirements txt clusterfuzz src local butler scripts clusterfuzz src python bot untrusted runner build depth clusterfuzz resources platform linux peach peach mutator peach mutator third party peach requirements txt clusterfuzz src appengine handlers cron project clusterfuzz src python bot tasks dependency hierarchy x twisted tar vulnerable library found in head commit a href vulnerability details in words protocols jabber xmlstream in twisted through xmpp support did not verify certificates when used with tls allowing an attacker to mitm connections publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact none for more information on scores click a href suggested fix type change files origin a href release date fix resolution replace or update the following file check this box to open an automated fix pr isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails in words protocols jabber xmlstream in twisted through xmpp support did not verify certificates when used with tls allowing an attacker to mitm connections vulnerabilityurl
| 0
|
22,676
| 31,899,966,307
|
IssuesEvent
|
2023-09-18 07:03:11
|
geneontology/go-ontology
|
https://api.github.com/repos/geneontology/go-ontology
|
reopened
|
pattern recognition receptor signaling pathway - NTR and reorganization
|
multi-species process
|
Hello,
This comes from the SIB-Virus group @pmasson55 and Chantal
1. pattern recognition receptor signaling pathway GO:0002221
Add as childen of GO:0002221:
- [x] 1.1 Create GO: endolysosomal pattern recognition receptor signaling pathway
1.1.1 Create GO: endolysosomal toll-like receptor signaling pathways
- [x] moved terms under the new term
1.1.1.1 GO:0034138 toll-like receptor 3 signaling pathway
1.1.1.2 GO:0034154 toll-like receptor 7 signaling pathway
1.1.1.3 GO:0034158 toll-like receptor 8 signaling pathway
1.1.1.4 GO:0034162 toll-like receptor 9 signaling pathway
1.1.1.5 GO:0034170 toll-like receptor 11 signaling pathway
1.1.1.6 GO:0034174 toll-like receptor 12 signaling pathway
1.1.1.7 GO:0034178 toll-like receptor 13 signaling pathway
-> endolysosomal DNA, rRNA, ssRNA sensors
1.2. GO:0002753: cytosolic pattern recognition receptor signaling pathway
1.2.1 GO:0039529: RIG-I signaling pathway
-> cytosolic ssRNA and dsRNA sensor
- [x] added ssRNA and dsRNA in definition
1.2.2 GO:0039530: MDA-5 signaling pathway
-> cytosolic dsRNA sensor
- [x] added dsRNA in definition
1.2.3 GO:0039585: PKR signal transduction
-> cytosolic dsRNA sensor
- [x] dsRNA was already in in definition
- [x] 1.2.4 Create GO: cGAS/STING pathway
-> cytosolic DNA sensor
1.3 GO:0002752: cell surface pattern recognition receptor signaling pathway
- [x] 1.3.1 Create GO: cell surface toll-like receptor signaling pathway
1.3.1.1 GO:0034130 toll-like receptor 1 signaling pathway
1.3.1.2 GO:0034134 toll-like receptor 2 signaling pathway
1.3.1.3 GO:0034142 toll-like receptor 4 signaling pathway
1.3.1.4 GO:0034146 toll-like receptor 5 signaling pathway
1.3.1.5 GO:0034150 toll-like receptor 6 signaling pathway
1.3.1.6 GO:0034166 toll-like receptor 10 signaling pathway
1.3.1.7 GO:0035681 toll-like receptor 15 signaling pathway
-> lipopeptides, dsRNA, ssRNA, LPS, flagellin sensors
GO:0002754 toll-like receptor signaling pathway will have both children “endolysosomal toll-like receptor signaling pathways” and “cell surface toll-like receptor signaling pathways”
Decide if we keep
GO:0039528 cytoplasmic pattern recognition receptor signaling pathway in response to virus
The other PRR signaling pathways (membrane, endolysosomal) do not have this granularity
List of terms negative/positive regulation:
Negative/positive regulation of endolysosomal pattern recognition receptor signaling pathway
Negative/positive regulation cytosolic pattern recognition receptor signaling pathway
Negative/positive regulation cell surface pattern recognition receptor signaling pathway
Liu, Xing & Wang, Qiang & Chen, Wei & Wang, Chen. (2013). Dynamic regulation of innate immunity by ubiquitin and ubiquitin-like proteins. Cytokine & growth factor reviews. 24. 10.1016/j.cytogfr.2013.07.002.
----
Any comments/suggestions ?
Thanks, Pascale
|
1.0
|
pattern recognition receptor signaling pathway - NTR and reorganization - Hello,
This comes from the SIB-Virus group @pmasson55 and Chantal
1. pattern recognition receptor signaling pathway GO:0002221
Add as childen of GO:0002221:
- [x] 1.1 Create GO: endolysosomal pattern recognition receptor signaling pathway
1.1.1 Create GO: endolysosomal toll-like receptor signaling pathways
- [x] moved terms under the new term
1.1.1.1 GO:0034138 toll-like receptor 3 signaling pathway
1.1.1.2 GO:0034154 toll-like receptor 7 signaling pathway
1.1.1.3 GO:0034158 toll-like receptor 8 signaling pathway
1.1.1.4 GO:0034162 toll-like receptor 9 signaling pathway
1.1.1.5 GO:0034170 toll-like receptor 11 signaling pathway
1.1.1.6 GO:0034174 toll-like receptor 12 signaling pathway
1.1.1.7 GO:0034178 toll-like receptor 13 signaling pathway
-> endolysosomal DNA, rRNA, ssRNA sensors
1.2. GO:0002753: cytosolic pattern recognition receptor signaling pathway
1.2.1 GO:0039529: RIG-I signaling pathway
-> cytosolic ssRNA and dsRNA sensor
- [x] added ssRNA and dsRNA in definition
1.2.2 GO:0039530: MDA-5 signaling pathway
-> cytosolic dsRNA sensor
- [x] added dsRNA in definition
1.2.3 GO:0039585: PKR signal transduction
-> cytosolic dsRNA sensor
- [x] dsRNA was already in in definition
- [x] 1.2.4 Create GO: cGAS/STING pathway
-> cytosolic DNA sensor
1.3 GO:0002752: cell surface pattern recognition receptor signaling pathway
- [x] 1.3.1 Create GO: cell surface toll-like receptor signaling pathway
1.3.1.1 GO:0034130 toll-like receptor 1 signaling pathway
1.3.1.2 GO:0034134 toll-like receptor 2 signaling pathway
1.3.1.3 GO:0034142 toll-like receptor 4 signaling pathway
1.3.1.4 GO:0034146 toll-like receptor 5 signaling pathway
1.3.1.5 GO:0034150 toll-like receptor 6 signaling pathway
1.3.1.6 GO:0034166 toll-like receptor 10 signaling pathway
1.3.1.7 GO:0035681 toll-like receptor 15 signaling pathway
-> lipopeptides, dsRNA, ssRNA, LPS, flagellin sensors
GO:0002754 toll-like receptor signaling pathway will have both children “endolysosomal toll-like receptor signaling pathways” and “cell surface toll-like receptor signaling pathways”
Decide if we keep
GO:0039528 cytoplasmic pattern recognition receptor signaling pathway in response to virus
The other PRR signaling pathways (membrane, endolysosomal) do not have this granularity
List of terms negative/positive regulation:
Negative/positive regulation of endolysosomal pattern recognition receptor signaling pathway
Negative/positive regulation cytosolic pattern recognition receptor signaling pathway
Negative/positive regulation cell surface pattern recognition receptor signaling pathway
Liu, Xing & Wang, Qiang & Chen, Wei & Wang, Chen. (2013). Dynamic regulation of innate immunity by ubiquitin and ubiquitin-like proteins. Cytokine & growth factor reviews. 24. 10.1016/j.cytogfr.2013.07.002.
----
Any comments/suggestions ?
Thanks, Pascale
|
process
|
pattern recognition receptor signaling pathway ntr and reorganization hello this comes from the sib virus group and chantal pattern recognition receptor signaling pathway go add as childen of go create go endolysosomal pattern recognition receptor signaling pathway create go endolysosomal toll like receptor signaling pathways moved terms under the new term go toll like receptor signaling pathway go toll like receptor signaling pathway go toll like receptor signaling pathway go toll like receptor signaling pathway go toll like receptor signaling pathway go toll like receptor signaling pathway go toll like receptor signaling pathway endolysosomal dna rrna ssrna sensors go cytosolic pattern recognition receptor signaling pathway go rig i signaling pathway cytosolic ssrna and dsrna sensor added ssrna and dsrna in definition go mda signaling pathway cytosolic dsrna sensor added dsrna in definition go pkr signal transduction cytosolic dsrna sensor dsrna was already in in definition create go cgas sting pathway cytosolic dna sensor go cell surface pattern recognition receptor signaling pathway create go cell surface toll like receptor signaling pathway go toll like receptor signaling pathway go toll like receptor signaling pathway go toll like receptor signaling pathway go toll like receptor signaling pathway go toll like receptor signaling pathway go toll like receptor signaling pathway go toll like receptor signaling pathway lipopeptides dsrna ssrna lps flagellin sensors go toll like receptor signaling pathway will have both children “endolysosomal toll like receptor signaling pathways” and “cell surface toll like receptor signaling pathways” decide if we keep go cytoplasmic pattern recognition receptor signaling pathway in response to virus the other prr signaling pathways membrane endolysosomal do not have this granularity list of terms negative positive regulation negative positive regulation of endolysosomal pattern recognition receptor signaling pathway negative positive regulation cytosolic pattern recognition receptor signaling pathway negative positive regulation cell surface pattern recognition receptor signaling pathway liu xing wang qiang chen wei wang chen dynamic regulation of innate immunity by ubiquitin and ubiquitin like proteins cytokine growth factor reviews j cytogfr any comments suggestions thanks pascale
| 1
|
7,014
| 2,596,372,515
|
IssuesEvent
|
2015-02-20 20:13:35
|
DoSomething/dosomething
|
https://api.github.com/repos/DoSomething/dosomething
|
closed
|
client side validation needed for accepted image file types
|
#campaign-template @fender priority-medium
|
error only comes after the whole form has been submitted, ideally this error should appear right after the user selects an invalid file:

related trello card: https://trello.com/c/dflw2sBj
|
1.0
|
client side validation needed for accepted image file types - error only comes after the whole form has been submitted, ideally this error should appear right after the user selects an invalid file:

related trello card: https://trello.com/c/dflw2sBj
|
non_process
|
client side validation needed for accepted image file types error only comes after the whole form has been submitted ideally this error should appear right after the user selects an invalid file related trello card
| 0
|
5,955
| 8,780,691,094
|
IssuesEvent
|
2018-12-19 18:04:56
|
googleapis/google-cloud-python
|
https://api.github.com/repos/googleapis/google-cloud-python
|
opened
|
BigQuery: system tests fail w/ 429 errors from GCS
|
api: bigquery testing type: process
|
BigQuery system tests which to create buckets / blobs need to be hardened against 429 responses from GCS.
|
1.0
|
BigQuery: system tests fail w/ 429 errors from GCS - BigQuery system tests which to create buckets / blobs need to be hardened against 429 responses from GCS.
|
process
|
bigquery system tests fail w errors from gcs bigquery system tests which to create buckets blobs need to be hardened against responses from gcs
| 1
|
259,683
| 19,610,233,061
|
IssuesEvent
|
2022-01-06 14:36:30
|
ViGEm/HidHide
|
https://api.github.com/repos/ViGEm/HidHide
|
closed
|
Cannot select Flight Simulator 2020 (Microsoft Store Edition)
|
documentation wontfix
|
Not sure if this is a bug with the Windows UMP apps but I cannot select Flight Simulator 2020.exe. It will give an error saying I don't have permissions. Tried the new beta Xbox app, which lets you relocate the folder outside the WindowsApp protected one, but the issue still occurs. Open to suggestions!
|
1.0
|
Cannot select Flight Simulator 2020 (Microsoft Store Edition) - Not sure if this is a bug with the Windows UMP apps but I cannot select Flight Simulator 2020.exe. It will give an error saying I don't have permissions. Tried the new beta Xbox app, which lets you relocate the folder outside the WindowsApp protected one, but the issue still occurs. Open to suggestions!
|
non_process
|
cannot select flight simulator microsoft store edition not sure if this is a bug with the windows ump apps but i cannot select flight simulator exe it will give an error saying i don t have permissions tried the new beta xbox app which lets you relocate the folder outside the windowsapp protected one but the issue still occurs open to suggestions
| 0
|
21,830
| 3,924,038,576
|
IssuesEvent
|
2016-04-22 13:54:30
|
ngageoint/hootenanny-ui
|
https://api.github.com/repos/ngageoint/hootenanny-ui
|
closed
|
Fix and run eslint tests
|
Category: Test Category: UI Priority: High Status: In Progress Type: Task
|
Following iD v1.9.2, update and run eslint tests for iD editor/Hoot-UI.
|
1.0
|
Fix and run eslint tests - Following iD v1.9.2, update and run eslint tests for iD editor/Hoot-UI.
|
non_process
|
fix and run eslint tests following id update and run eslint tests for id editor hoot ui
| 0
|
12,505
| 14,961,670,549
|
IssuesEvent
|
2021-01-27 08:09:22
|
beyondhb1079/s4us
|
https://api.github.com/repos/beyondhb1079/s4us
|
closed
|
About: Job Hernandez
|
process
|
Come up with 2-3 sentences about yourself and add yourself to the `team` array in the [About page](https://github.com/beyondhb1079/s4us/blob/main/src/pages/About.jsx#L11).
|
1.0
|
About: Job Hernandez - Come up with 2-3 sentences about yourself and add yourself to the `team` array in the [About page](https://github.com/beyondhb1079/s4us/blob/main/src/pages/About.jsx#L11).
|
process
|
about job hernandez come up with sentences about yourself and add yourself to the team array in the
| 1
|
531,924
| 15,527,450,774
|
IssuesEvent
|
2021-03-13 06:04:51
|
creativecommons/ccos-scripts
|
https://api.github.com/repos/creativecommons/ccos-scripts
|
closed
|
Update Tooling of the packages
|
good first issue help wanted ✨ goal: improvement 🟩 priority: low 🤖 aspect: dx
|
## Problem
The Python packages of the repository have older versions of black in `Pipfile` which makes various packages of the repositories.It can be a better approach to optimise these configurations by
- Following the Python Community Guidelines (Also see [Python Guidelines — Creative Commons Open Source](https://opensource.creativecommons.org/contributing-code/python-guidelines/) about adding Black and Flake8)
## Solution
The configuration could be optimised with that of the configs in the [creativecommons/cc-licenses](https://github.com/creativecommons/cc-licenses/blob/main/Pipfile)
## Resolution
<!-- Replace the [ ] with [x] to check the box. -->
- [x] I would be interested in resolving this bug.
|
1.0
|
Update Tooling of the packages - ## Problem
The Python packages of the repository have older versions of black in `Pipfile` which makes various packages of the repositories.It can be a better approach to optimise these configurations by
- Following the Python Community Guidelines (Also see [Python Guidelines — Creative Commons Open Source](https://opensource.creativecommons.org/contributing-code/python-guidelines/) about adding Black and Flake8)
## Solution
The configuration could be optimised with that of the configs in the [creativecommons/cc-licenses](https://github.com/creativecommons/cc-licenses/blob/main/Pipfile)
## Resolution
<!-- Replace the [ ] with [x] to check the box. -->
- [x] I would be interested in resolving this bug.
|
non_process
|
update tooling of the packages problem the python packages of the repository have older versions of black in pipfile which makes various packages of the repositories it can be a better approach to optimise these configurations by following the python community guidelines also see about adding black and solution the configuration could be optimised with that of the configs in the resolution i would be interested in resolving this bug
| 0
|
3,335
| 3,129,290,821
|
IssuesEvent
|
2015-09-09 00:02:37
|
spring-projects/spring-boot
|
https://api.github.com/repos/spring-projects/spring-boot
|
closed
|
Add checkstyle to build
|
build
|
The recent Eclipse update changed the way that code formatting is applied. We could potentially use [this plugin](https://github.com/revelc/formatter-maven-plugin) to automatically format the code from Maven rather than relying on Eclipse.
Unfortunately we also currently apply Eclipse cleanup save actions with the Maven plugin doesn't support.
If we apply checkstyle to the build we could enforce that the code style matches what would occur via the cleanup actions (adding `this.` etc). This would then allow us to use the Maven plugin to format the code.
As an added benefit checkstyle is also likely to help a lot when people contribute code.
|
1.0
|
Add checkstyle to build - The recent Eclipse update changed the way that code formatting is applied. We could potentially use [this plugin](https://github.com/revelc/formatter-maven-plugin) to automatically format the code from Maven rather than relying on Eclipse.
Unfortunately we also currently apply Eclipse cleanup save actions with the Maven plugin doesn't support.
If we apply checkstyle to the build we could enforce that the code style matches what would occur via the cleanup actions (adding `this.` etc). This would then allow us to use the Maven plugin to format the code.
As an added benefit checkstyle is also likely to help a lot when people contribute code.
|
non_process
|
add checkstyle to build the recent eclipse update changed the way that code formatting is applied we could potentially use to automatically format the code from maven rather than relying on eclipse unfortunately we also currently apply eclipse cleanup save actions with the maven plugin doesn t support if we apply checkstyle to the build we could enforce that the code style matches what would occur via the cleanup actions adding this etc this would then allow us to use the maven plugin to format the code as an added benefit checkstyle is also likely to help a lot when people contribute code
| 0
|
591,116
| 17,795,130,745
|
IssuesEvent
|
2021-08-31 21:03:17
|
magma/magma
|
https://api.github.com/repos/magma/magma
|
closed
|
Intra AGW Mobility
|
Epic priority: high
|
Impact: we have seen issues in the field are seeing issues with UEs and this is a required feature by multiple partners but we haven't been able to promote it yet. CBRS at FB also requesting this. This is S1 mobility, intra, within the same AGW, 2 enodes.
|
1.0
|
Intra AGW Mobility - Impact: we have seen issues in the field are seeing issues with UEs and this is a required feature by multiple partners but we haven't been able to promote it yet. CBRS at FB also requesting this. This is S1 mobility, intra, within the same AGW, 2 enodes.
|
non_process
|
intra agw mobility impact we have seen issues in the field are seeing issues with ues and this is a required feature by multiple partners but we haven t been able to promote it yet cbrs at fb also requesting this this is mobility intra within the same agw enodes
| 0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.