Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
22,346
31,022,451,519
IssuesEvent
2023-08-10 06:45:03
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
[Mirror] https://go.dev/dl/go1.20.7.linux-amd64.tar.gz
P2 type: process team-OSS mirror request
### Please list the URLs of the archives you'd like to mirror: https://go.dev/dl/go1.20.7.linux-amd64.tar.gz This is to upgrade go toolchain in https://github.com/bazelbuild/remote-apis-sdks/pull/471.
1.0
[Mirror] https://go.dev/dl/go1.20.7.linux-amd64.tar.gz - ### Please list the URLs of the archives you'd like to mirror: https://go.dev/dl/go1.20.7.linux-amd64.tar.gz This is to upgrade go toolchain in https://github.com/bazelbuild/remote-apis-sdks/pull/471.
process
please list the urls of the archives you d like to mirror this is to upgrade go toolchain in
1
54,813
13,453,801,575
IssuesEvent
2020-09-09 02:00:09
golang/go
https://api.github.com/repos/golang/go
closed
x/build: 2 darwin builders are missing
Builders NeedsInvestigation
From https://farmer.golang.org/#health: ``` # "macs" status: MacStadium Mac VMs # Notes: https://github.com/golang/build/tree/master/env/darwin/macstadium Warn: macstadium_host08a missing, not seen for 55h42m8s Warn: macstadium_host08b missing, not seen for 55h42m6s Warn: 2 machines missing, 10% of capacity Warn: makemac daemon: vm.destroy("mac_10_12_host08a") = govc vm.destroy ...: exit status 1, govc: Unable to communicate with the remote host, since it is disconnected. Warn: makemac daemon: vm.destroy("mac_10_12_host08b") = govc vm.destroy ...: exit status 1, govc: Unable to communicate with the remote host, since it is disconnected. ``` The machine hosting `macstadium_host08a` and `macstadium_host08b` is being unresponsive. I've taken some initial steps to bring it back up, but this might need more work. I'll resume tomorrow. /cc @toothrot @andybons
1.0
x/build: 2 darwin builders are missing - From https://farmer.golang.org/#health: ``` # "macs" status: MacStadium Mac VMs # Notes: https://github.com/golang/build/tree/master/env/darwin/macstadium Warn: macstadium_host08a missing, not seen for 55h42m8s Warn: macstadium_host08b missing, not seen for 55h42m6s Warn: 2 machines missing, 10% of capacity Warn: makemac daemon: vm.destroy("mac_10_12_host08a") = govc vm.destroy ...: exit status 1, govc: Unable to communicate with the remote host, since it is disconnected. Warn: makemac daemon: vm.destroy("mac_10_12_host08b") = govc vm.destroy ...: exit status 1, govc: Unable to communicate with the remote host, since it is disconnected. ``` The machine hosting `macstadium_host08a` and `macstadium_host08b` is being unresponsive. I've taken some initial steps to bring it back up, but this might need more work. I'll resume tomorrow. /cc @toothrot @andybons
non_process
x build darwin builders are missing from macs status macstadium mac vms notes warn macstadium missing not seen for warn macstadium missing not seen for warn machines missing of capacity warn makemac daemon vm destroy mac govc vm destroy exit status govc unable to communicate with the remote host since it is disconnected warn makemac daemon vm destroy mac govc vm destroy exit status govc unable to communicate with the remote host since it is disconnected the machine hosting macstadium and macstadium is being unresponsive i ve taken some initial steps to bring it back up but this might need more work i ll resume tomorrow cc toothrot andybons
0
103,083
8,877,190,488
IssuesEvent
2019-01-12 21:58:48
swe-ms-boun/2018fall-swe574-g2
https://api.github.com/repos/swe-ms-boun/2018fall-swe574-g2
closed
Selenium Web-Driver UI Test Automation Framework
test
Acceptance test cases should be automated by using selenium web driver when the UI is deployed.
1.0
Selenium Web-Driver UI Test Automation Framework - Acceptance test cases should be automated by using selenium web driver when the UI is deployed.
non_process
selenium web driver ui test automation framework acceptance test cases should be automated by using selenium web driver when the ui is deployed
0
5,317
8,130,529,603
IssuesEvent
2018-08-17 18:49:36
WalkthroughVR/Handbook
https://api.github.com/repos/WalkthroughVR/Handbook
opened
v0.8 - Automating the 360 connection point mapping. Having the 360 model automatically map where the last 360 was taken.
Photography Processing Automation
## Problem: - N/A ## Hypothesis/Goal: - N/A ## Solution/Requirements: - [ ] N/A ## What we should learn: - N/A ## Additional notes - N/A
1.0
v0.8 - Automating the 360 connection point mapping. Having the 360 model automatically map where the last 360 was taken. - ## Problem: - N/A ## Hypothesis/Goal: - N/A ## Solution/Requirements: - [ ] N/A ## What we should learn: - N/A ## Additional notes - N/A
process
automating the connection point mapping having the model automatically map where the last was taken problem n a hypothesis goal n a solution requirements n a what we should learn n a additional notes n a
1
20,435
27,098,849,907
IssuesEvent
2023-02-15 06:43:34
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
Move Python rules documentation from Bazel site to rules_python
P3 type: process team-Rules-Python stale
Blocked on bazelbuild/rules_python#202 and bazelbuild/rules_python#144, both of which will be fixed imminently. (But I don't think this will be prioritized anywhere near as soon as those.)
1.0
Move Python rules documentation from Bazel site to rules_python - Blocked on bazelbuild/rules_python#202 and bazelbuild/rules_python#144, both of which will be fixed imminently. (But I don't think this will be prioritized anywhere near as soon as those.)
process
move python rules documentation from bazel site to rules python blocked on bazelbuild rules python and bazelbuild rules python both of which will be fixed imminently but i don t think this will be prioritized anywhere near as soon as those
1
217,946
24,351,688,644
IssuesEvent
2022-10-03 01:09:57
benlazarine/cas-overlay
https://api.github.com/repos/benlazarine/cas-overlay
opened
CVE-2022-38751 (Medium) detected in snakeyaml-1.17.jar
security vulnerability
## CVE-2022-38751 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>snakeyaml-1.17.jar</b></p></summary> <p>YAML 1.1 parser and emitter for Java</p> <p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /root/.m2/repository/org/yaml/snakeyaml/1.17/snakeyaml-1.17.jar</p> <p> Dependency Hierarchy: - cas-server-support-oauth-webflow-5.3.7.jar (Root Library) - spring-boot-starter-websocket-1.5.18.RELEASE.jar - spring-boot-starter-1.5.18.RELEASE.jar - :x: **snakeyaml-1.17.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Using snakeYAML to parse untrusted YAML files may be vulnerable to Denial of Service attacks (DOS). If the parser is running on user supplied input, an attacker may supply content that causes the parser to crash by stackoverflow. <p>Publish Date: 2022-09-05 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-38751>CVE-2022-38751</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bugs.chromium.org/p/oss-fuzz/issues/detail?id=47039">https://bugs.chromium.org/p/oss-fuzz/issues/detail?id=47039</a></p> <p>Release Date: 2022-09-05</p> <p>Fix Resolution: org.yaml:snakeyaml:1.31</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-38751 (Medium) detected in snakeyaml-1.17.jar - ## CVE-2022-38751 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>snakeyaml-1.17.jar</b></p></summary> <p>YAML 1.1 parser and emitter for Java</p> <p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /root/.m2/repository/org/yaml/snakeyaml/1.17/snakeyaml-1.17.jar</p> <p> Dependency Hierarchy: - cas-server-support-oauth-webflow-5.3.7.jar (Root Library) - spring-boot-starter-websocket-1.5.18.RELEASE.jar - spring-boot-starter-1.5.18.RELEASE.jar - :x: **snakeyaml-1.17.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Using snakeYAML to parse untrusted YAML files may be vulnerable to Denial of Service attacks (DOS). If the parser is running on user supplied input, an attacker may supply content that causes the parser to crash by stackoverflow. <p>Publish Date: 2022-09-05 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-38751>CVE-2022-38751</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bugs.chromium.org/p/oss-fuzz/issues/detail?id=47039">https://bugs.chromium.org/p/oss-fuzz/issues/detail?id=47039</a></p> <p>Release Date: 2022-09-05</p> <p>Fix Resolution: org.yaml:snakeyaml:1.31</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in snakeyaml jar cve medium severity vulnerability vulnerable library snakeyaml jar yaml parser and emitter for java library home page a href path to dependency file pom xml path to vulnerable library root repository org yaml snakeyaml snakeyaml jar dependency hierarchy cas server support oauth webflow jar root library spring boot starter websocket release jar spring boot starter release jar x snakeyaml jar vulnerable library vulnerability details using snakeyaml to parse untrusted yaml files may be vulnerable to denial of service attacks dos if the parser is running on user supplied input an attacker may supply content that causes the parser to crash by stackoverflow publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org yaml snakeyaml step up your open source security game with mend
0
11,524
14,401,981,829
IssuesEvent
2020-12-03 14:23:28
pystatgen/sgkit
https://api.github.com/repos/pystatgen/sgkit
closed
png files from documentation should be git ignored?
bug process + tools
`make clean html` will create `png` files that are not git ignored (making git dirty). Files: ``` mydask.png order.png ```
1.0
png files from documentation should be git ignored? - `make clean html` will create `png` files that are not git ignored (making git dirty). Files: ``` mydask.png order.png ```
process
png files from documentation should be git ignored make clean html will create png files that are not git ignored making git dirty files mydask png order png
1
1,641
4,259,980,857
IssuesEvent
2016-07-11 13:02:18
e-government-ua/iBP
https://api.github.com/repos/e-government-ua/iBP
closed
Нетишин: раскрітие и уточнения в "Розірвання договору оренди земельної ділянки"
In process of testing in work test
Розірвання договору оренди земельної ділянки або припинення права постійного користування земельною ділянкою Зауваження: 1. Потрібна можливість завантажити сканкопії наступних документів: - копія довідки з податкової, що людина не має заборгованості по сплаті; - копія договору оренди. 2. Додати графу: підстава для розірвання.
1.0
Нетишин: раскрітие и уточнения в "Розірвання договору оренди земельної ділянки" - Розірвання договору оренди земельної ділянки або припинення права постійного користування земельною ділянкою Зауваження: 1. Потрібна можливість завантажити сканкопії наступних документів: - копія довідки з податкової, що людина не має заборгованості по сплаті; - копія договору оренди. 2. Додати графу: підстава для розірвання.
process
нетишин раскрітие и уточнения в розірвання договору оренди земельної ділянки розірвання договору оренди земельної ділянки або припинення права постійного користування земельною ділянкою зауваження потрібна можливість завантажити сканкопії наступних документів копія довідки з податкової що людина не має заборгованості по сплаті копія договору оренди додати графу підстава для розірвання
1
20,145
26,694,734,309
IssuesEvent
2023-01-27 09:21:47
UnitTestBot/UTBotJava
https://api.github.com/repos/UnitTestBot/UTBotJava
closed
`UtExecutionInstrumentation` does not respect timeout in concrete execution
ctg-bug comp-contest-estimator comp-instrumented-process
**Description** Currently `UtExecutionInstrumentation` measures only concrete execution, but not model construction for parameters and execution result. Sometimes model construction might take too much time and will hang Instrumentation process indefenitely. This occurs too often in Contest Estimator, but problem relates to plugin too. **To Reproduce** Steps to reproduce the behavior: 1. In `ContestEstimator.kt` set: - set `javaHome` to JDK8 - `timeLimit=120` - `projectFilter = listOf("guava-26.0")` - set `utbot-junit-contest/src/main/resources/classes/guava-26.0/list` to ``` com.google.common.primitives.Booleans com.google.common.primitives.Shorts ``` 2. Start `ContestEstimator` 3. Tests for class `Shorts` once in a 2-3 generations will not be generated at all **Expected behavior** Everything works **Actual behavior** Tests for `Short` does not generate, Instrumentation process hangs on some requrest.
1.0
`UtExecutionInstrumentation` does not respect timeout in concrete execution - **Description** Currently `UtExecutionInstrumentation` measures only concrete execution, but not model construction for parameters and execution result. Sometimes model construction might take too much time and will hang Instrumentation process indefenitely. This occurs too often in Contest Estimator, but problem relates to plugin too. **To Reproduce** Steps to reproduce the behavior: 1. In `ContestEstimator.kt` set: - set `javaHome` to JDK8 - `timeLimit=120` - `projectFilter = listOf("guava-26.0")` - set `utbot-junit-contest/src/main/resources/classes/guava-26.0/list` to ``` com.google.common.primitives.Booleans com.google.common.primitives.Shorts ``` 2. Start `ContestEstimator` 3. Tests for class `Shorts` once in a 2-3 generations will not be generated at all **Expected behavior** Everything works **Actual behavior** Tests for `Short` does not generate, Instrumentation process hangs on some requrest.
process
utexecutioninstrumentation does not respect timeout in concrete execution description currently utexecutioninstrumentation measures only concrete execution but not model construction for parameters and execution result sometimes model construction might take too much time and will hang instrumentation process indefenitely this occurs too often in contest estimator but problem relates to plugin too to reproduce steps to reproduce the behavior in contestestimator kt set set javahome to timelimit projectfilter listof guava set utbot junit contest src main resources classes guava list to com google common primitives booleans com google common primitives shorts start contestestimator tests for class shorts once in a generations will not be generated at all expected behavior everything works actual behavior tests for short does not generate instrumentation process hangs on some requrest
1
4,893
7,763,824,683
IssuesEvent
2018-06-01 17:58:28
StrikeNP/trac_test
https://api.github.com/repos/StrikeNP/trac_test
closed
GABLS2 rtm, rtp2, thlm, and thlp2 are set to zero when plotgen is run manually (but not for the nightly tests) (Trac #24)
Migrated from Trac enhancement post_processing senkbeil@uwm.edu
Some time ago, in order to test CLUBB's scalars, Brandon changed plotgen so that it outputs scalars in place of rtm and thlm. The nightly plots work great. However, if CLUBB is run manually without outputting scalars, and then plotgen is executed manually, then rtm, thlm, rtp2, and thlp2 are set to zero. For manual runs, typically we don't want to check scalars; we just want to plot standard versions of rtm, thlm, rtp2, and thlp2. I probably forgot to mention this earlier. Is it feasible to insert some nightly flags or re-arrange some code so that the nightly plots test the scalars, but the manual plots simply plot rtm, thlm, rtp2, and thlp2? I believe that this is what is done for other specialized nightly tests, e.g. the restart test and some of the altered grid tests. Perhaps those pieces of code would provide ideas on how to implement separate behavior for nightly and manual runs. However, we have a deadline on the TWP-ICE case, so don't bother with this until TWP-ICE is submitted, unless it is trivial to fix. Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/24 ```json { "status": "closed", "changetime": "2009-09-02T20:37:37", "description": "Some time ago, in order to test CLUBB's scalars, Brandon changed plotgen so that it outputs scalars in place of rtm and thlm. The nightly plots work great.\n\nHowever, if CLUBB is run manually without outputting scalars, and then plotgen is executed manually, then rtm, thlm, rtp2, and thlp2 are set to zero. For manual runs, typically we don't want to check scalars; we just want to plot standard versions of rtm, thlm, rtp2, and thlp2. I probably forgot to mention this earlier.\n\nIs it feasible to insert some nightly flags or re-arrange some code so that the nightly plots test the scalars, but the manual plots simply plot rtm, thlm, rtp2, and thlp2? I believe that this is what is done for other specialized nightly tests, e.g. the restart test and some of the altered grid tests. Perhaps those pieces of code would provide ideas on how to implement separate behavior for nightly and manual runs.\n\nHowever, we have a deadline on the TWP-ICE case, so don't bother with this until TWP-ICE is submitted, unless it is trivial to fix.", "reporter": "vlarson@uwm.edu", "cc": "", "resolution": "Verified by V. Larson", "_ts": "1251923857000000", "component": "post_processing", "summary": "GABLS2 rtm, rtp2, thlm, and thlp2 are set to zero when plotgen is run manually (but not for the nightly tests)", "priority": "minor", "keywords": "scalars, gabls2, nightly plots, rtm, thlm, rtp2, thlp2", "time": "2009-05-13T14:26:46", "milestone": "Plotgen 3.0", "owner": "senkbeil@uwm.edu", "type": "enhancement" } ```
1.0
GABLS2 rtm, rtp2, thlm, and thlp2 are set to zero when plotgen is run manually (but not for the nightly tests) (Trac #24) - Some time ago, in order to test CLUBB's scalars, Brandon changed plotgen so that it outputs scalars in place of rtm and thlm. The nightly plots work great. However, if CLUBB is run manually without outputting scalars, and then plotgen is executed manually, then rtm, thlm, rtp2, and thlp2 are set to zero. For manual runs, typically we don't want to check scalars; we just want to plot standard versions of rtm, thlm, rtp2, and thlp2. I probably forgot to mention this earlier. Is it feasible to insert some nightly flags or re-arrange some code so that the nightly plots test the scalars, but the manual plots simply plot rtm, thlm, rtp2, and thlp2? I believe that this is what is done for other specialized nightly tests, e.g. the restart test and some of the altered grid tests. Perhaps those pieces of code would provide ideas on how to implement separate behavior for nightly and manual runs. However, we have a deadline on the TWP-ICE case, so don't bother with this until TWP-ICE is submitted, unless it is trivial to fix. Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/24 ```json { "status": "closed", "changetime": "2009-09-02T20:37:37", "description": "Some time ago, in order to test CLUBB's scalars, Brandon changed plotgen so that it outputs scalars in place of rtm and thlm. The nightly plots work great.\n\nHowever, if CLUBB is run manually without outputting scalars, and then plotgen is executed manually, then rtm, thlm, rtp2, and thlp2 are set to zero. For manual runs, typically we don't want to check scalars; we just want to plot standard versions of rtm, thlm, rtp2, and thlp2. I probably forgot to mention this earlier.\n\nIs it feasible to insert some nightly flags or re-arrange some code so that the nightly plots test the scalars, but the manual plots simply plot rtm, thlm, rtp2, and thlp2? I believe that this is what is done for other specialized nightly tests, e.g. the restart test and some of the altered grid tests. Perhaps those pieces of code would provide ideas on how to implement separate behavior for nightly and manual runs.\n\nHowever, we have a deadline on the TWP-ICE case, so don't bother with this until TWP-ICE is submitted, unless it is trivial to fix.", "reporter": "vlarson@uwm.edu", "cc": "", "resolution": "Verified by V. Larson", "_ts": "1251923857000000", "component": "post_processing", "summary": "GABLS2 rtm, rtp2, thlm, and thlp2 are set to zero when plotgen is run manually (but not for the nightly tests)", "priority": "minor", "keywords": "scalars, gabls2, nightly plots, rtm, thlm, rtp2, thlp2", "time": "2009-05-13T14:26:46", "milestone": "Plotgen 3.0", "owner": "senkbeil@uwm.edu", "type": "enhancement" } ```
process
rtm thlm and are set to zero when plotgen is run manually but not for the nightly tests trac some time ago in order to test clubb s scalars brandon changed plotgen so that it outputs scalars in place of rtm and thlm the nightly plots work great however if clubb is run manually without outputting scalars and then plotgen is executed manually then rtm thlm and are set to zero for manual runs typically we don t want to check scalars we just want to plot standard versions of rtm thlm and i probably forgot to mention this earlier is it feasible to insert some nightly flags or re arrange some code so that the nightly plots test the scalars but the manual plots simply plot rtm thlm and i believe that this is what is done for other specialized nightly tests e g the restart test and some of the altered grid tests perhaps those pieces of code would provide ideas on how to implement separate behavior for nightly and manual runs however we have a deadline on the twp ice case so don t bother with this until twp ice is submitted unless it is trivial to fix migrated from json status closed changetime description some time ago in order to test clubb s scalars brandon changed plotgen so that it outputs scalars in place of rtm and thlm the nightly plots work great n nhowever if clubb is run manually without outputting scalars and then plotgen is executed manually then rtm thlm and are set to zero for manual runs typically we don t want to check scalars we just want to plot standard versions of rtm thlm and i probably forgot to mention this earlier n nis it feasible to insert some nightly flags or re arrange some code so that the nightly plots test the scalars but the manual plots simply plot rtm thlm and i believe that this is what is done for other specialized nightly tests e g the restart test and some of the altered grid tests perhaps those pieces of code would provide ideas on how to implement separate behavior for nightly and manual runs n nhowever we have a deadline on the twp ice case so don t bother with this until twp ice is submitted unless it is trivial to fix reporter vlarson uwm edu cc resolution verified by v larson ts component post processing summary rtm thlm and are set to zero when plotgen is run manually but not for the nightly tests priority minor keywords scalars nightly plots rtm thlm time milestone plotgen owner senkbeil uwm edu type enhancement
1
96,728
10,961,381,741
IssuesEvent
2019-11-27 15:17:55
sunpy/sunpy
https://api.github.com/repos/sunpy/sunpy
opened
Database search() method missing returns documentation
Documentation database
See https://docs.sunpy.org/en/latest/api/sunpy.database.Database.html#sunpy.database.Database.search - the return type of `search()` is currently not documented.
1.0
Database search() method missing returns documentation - See https://docs.sunpy.org/en/latest/api/sunpy.database.Database.html#sunpy.database.Database.search - the return type of `search()` is currently not documented.
non_process
database search method missing returns documentation see the return type of search is currently not documented
0
15,482
19,689,298,411
IssuesEvent
2022-01-12 03:56:20
amor71/LiuAlgoTrader
https://api.github.com/repos/amor71/LiuAlgoTrader
closed
Multi User Support
enhancement in-process no-issue-activity
**Is your feature request related to a problem? Please describe.** Currently, the framework supports a single user and a single `tradeplan.toml` file. This file is read each time the platform starts. It is quite convenient for a single user running trades for him/herself. The required change address two related issues: 1. Support several users, running different strategies in parallel, 2. Move beyond the tradeplan.toml file allowing different plans for different users **Describe the solution you'd like** 1. extend the Portfolio table to include additional details such as broker & external account id, 2. extend trader(s) to support trading on behalf of others, 3. extend DB model to allow setting up a trade-plan per user
1.0
Multi User Support - **Is your feature request related to a problem? Please describe.** Currently, the framework supports a single user and a single `tradeplan.toml` file. This file is read each time the platform starts. It is quite convenient for a single user running trades for him/herself. The required change address two related issues: 1. Support several users, running different strategies in parallel, 2. Move beyond the tradeplan.toml file allowing different plans for different users **Describe the solution you'd like** 1. extend the Portfolio table to include additional details such as broker & external account id, 2. extend trader(s) to support trading on behalf of others, 3. extend DB model to allow setting up a trade-plan per user
process
multi user support is your feature request related to a problem please describe currently the framework supports a single user and a single tradeplan toml file this file is read each time the platform starts it is quite convenient for a single user running trades for him herself the required change address two related issues support several users running different strategies in parallel move beyond the tradeplan toml file allowing different plans for different users describe the solution you d like extend the portfolio table to include additional details such as broker external account id extend trader s to support trading on behalf of others extend db model to allow setting up a trade plan per user
1
28,048
12,758,050,268
IssuesEvent
2020-06-29 00:36:55
Azure/azure-rest-api-specs
https://api.github.com/repos/Azure/azure-rest-api-specs
closed
HTTP 500 - Cost Management API query usage
Cost Management Service Attention question
I am running the query usage example from this link: https://docs.microsoft.com/en-us/rest/api/cost-management/query/usage With the following URL: ``` https://management.azure.com/subscriptions/{subscription}/resourceGroups/{group}/providers/Microsoft.CostManagement/query?api-version=2019-11-01 ``` And this body payload: ``` { timeframe: "MonthToDate" type: "Usage" } ``` But I'm getting HTTP 500 error: ``` { "error": { "code": "500", "message": "An error occurred during processing this request. Use this request id '2e416ae7-fbc0-4dd8-a8e5-4ea0ad843c11' for follow-up." } } ``` Would it be anything wrong that I'm doing? ![image](https://user-images.githubusercontent.com/6530769/82713806-b247cf00-9c62-11ea-9280-fa259f5ff4d9.png)
1.0
HTTP 500 - Cost Management API query usage - I am running the query usage example from this link: https://docs.microsoft.com/en-us/rest/api/cost-management/query/usage With the following URL: ``` https://management.azure.com/subscriptions/{subscription}/resourceGroups/{group}/providers/Microsoft.CostManagement/query?api-version=2019-11-01 ``` And this body payload: ``` { timeframe: "MonthToDate" type: "Usage" } ``` But I'm getting HTTP 500 error: ``` { "error": { "code": "500", "message": "An error occurred during processing this request. Use this request id '2e416ae7-fbc0-4dd8-a8e5-4ea0ad843c11' for follow-up." } } ``` Would it be anything wrong that I'm doing? ![image](https://user-images.githubusercontent.com/6530769/82713806-b247cf00-9c62-11ea-9280-fa259f5ff4d9.png)
non_process
http cost management api query usage i am running the query usage example from this link with the following url and this body payload timeframe monthtodate type usage but i m getting http error error code message an error occurred during processing this request use this request id for follow up would it be anything wrong that i m doing
0
345,483
24,861,815,223
IssuesEvent
2022-10-27 08:52:02
facebook/docusaurus
https://api.github.com/repos/facebook/docusaurus
opened
Plugin-content-pages preset options, 'routeBasePath' can't be ''
documentation status: needs triage
### Have you read the Contributing Guidelines on issues? - [X] I have read the [Contributing Guidelines on issues](https://github.com/facebook/docusaurus/blob/main/CONTRIBUTING.md#reporting-new-issues). ### Description When I use default preset options ```shell module.exports = { presets: [ [ '@docusaurus/preset-classic', { pages: { path: 'src/pages', routeBasePath: '', include: ['**/*.{js,jsx,ts,tsx,md,mdx}'], exclude: [ '**/_*.{js,jsx,ts,tsx,md,mdx}', '**/_*/**', '**/*.test.{js,jsx,ts,tsx}', '**/__tests__/**', ], mdxPageComponent: '@theme/MDXPage', remarkPlugins: [require('remark-math')], rehypePlugins: [], beforeDefaultRemarkPlugins: [], beforeDefaultRehypePlugins: [], }, }, ], ], }; ``` It pointed out that `ValidationError: "routeBasePath" is not allowed to be empty` ![image](https://user-images.githubusercontent.com/82041918/198238848-6cbca559-943a-471b-a608-5490716014cf.png) ### Self-service - [X] I'd be willing to address this documentation request myself.
1.0
Plugin-content-pages preset options, 'routeBasePath' can't be '' - ### Have you read the Contributing Guidelines on issues? - [X] I have read the [Contributing Guidelines on issues](https://github.com/facebook/docusaurus/blob/main/CONTRIBUTING.md#reporting-new-issues). ### Description When I use default preset options ```shell module.exports = { presets: [ [ '@docusaurus/preset-classic', { pages: { path: 'src/pages', routeBasePath: '', include: ['**/*.{js,jsx,ts,tsx,md,mdx}'], exclude: [ '**/_*.{js,jsx,ts,tsx,md,mdx}', '**/_*/**', '**/*.test.{js,jsx,ts,tsx}', '**/__tests__/**', ], mdxPageComponent: '@theme/MDXPage', remarkPlugins: [require('remark-math')], rehypePlugins: [], beforeDefaultRemarkPlugins: [], beforeDefaultRehypePlugins: [], }, }, ], ], }; ``` It pointed out that `ValidationError: "routeBasePath" is not allowed to be empty` ![image](https://user-images.githubusercontent.com/82041918/198238848-6cbca559-943a-471b-a608-5490716014cf.png) ### Self-service - [X] I'd be willing to address this documentation request myself.
non_process
plugin content pages preset options routebasepath can t be have you read the contributing guidelines on issues i have read the description when i use default preset options shell module exports presets docusaurus preset classic pages path src pages routebasepath include exclude js jsx ts tsx md mdx test js jsx ts tsx tests mdxpagecomponent theme mdxpage remarkplugins rehypeplugins beforedefaultremarkplugins beforedefaultrehypeplugins it pointed out that validationerror routebasepath is not allowed to be empty self service i d be willing to address this documentation request myself
0
21,783
30,294,985,929
IssuesEvent
2023-07-09 18:45:49
The-Data-Alchemists-Manipal/MindWave
https://api.github.com/repos/The-Data-Alchemists-Manipal/MindWave
closed
Add Filter Application using Flask and OpenCV.
image-processing
### Is your feature request related to a problem? Please describe. In this project, I have used Flask and OpenCV to add filters to the frames of the video. This project combines the power of Flask and OpenCV to create a user-friendly application where users can apply various filters to video frames. It provides an interactive and intuitive interface for experimenting with different filters and enhancing the visual appeal of images. ### Describe the solution you'd like The UI shows up the video and several filter options which can be applied to the video frame by the user. ### Describe alternatives you've considered _No response_ ### Additional context _No response_ ### Code of Conduct - [X] I agree to follow this project's Code of Conduct
1.0
Add Filter Application using Flask and OpenCV. - ### Is your feature request related to a problem? Please describe. In this project, I have used Flask and OpenCV to add filters to the frames of the video. This project combines the power of Flask and OpenCV to create a user-friendly application where users can apply various filters to video frames. It provides an interactive and intuitive interface for experimenting with different filters and enhancing the visual appeal of images. ### Describe the solution you'd like The UI shows up the video and several filter options which can be applied to the video frame by the user. ### Describe alternatives you've considered _No response_ ### Additional context _No response_ ### Code of Conduct - [X] I agree to follow this project's Code of Conduct
process
add filter application using flask and opencv is your feature request related to a problem please describe in this project i have used flask and opencv to add filters to the frames of the video this project combines the power of flask and opencv to create a user friendly application where users can apply various filters to video frames it provides an interactive and intuitive interface for experimenting with different filters and enhancing the visual appeal of images describe the solution you d like the ui shows up the video and several filter options which can be applied to the video frame by the user describe alternatives you ve considered no response additional context no response code of conduct i agree to follow this project s code of conduct
1
17,282
23,086,837,790
IssuesEvent
2022-07-26 12:12:21
apache/arrow-rs
https://api.github.com/repos/apache/arrow-rs
closed
Release Arrow `19.0.0` (next release after `18.0.0`)
development-process
* Planned Release Candidate: 2022-07-22 * Planned Release and Publish to crates.io: 2022-07-25 Items: - [x] Update changelog and readme: https://github.com/apache/arrow-rs/pull/2120 - [x] Create release candidate https://lists.apache.org/thread/txw3hlm06qjwczcbn20mjlgp7gf178fc - [x] Release candidate approved https://lists.apache.org/thread/3stno7fw741bpxo81s0fph80cbcymztq - [x] Release to crates.io - [x] Draft update to DataFusion: https://github.com/apache/arrow-datafusion/pull/2955 See full list here: https://github.com/apache/arrow-rs/compare/18.0.0...19.0.0 https://github.com/apache/arrow-rs/issues/1952 cc @viirya @jhorstmann @tustvold @sunchao @HaoYang670
1.0
Release Arrow `19.0.0` (next release after `18.0.0`) - * Planned Release Candidate: 2022-07-22 * Planned Release and Publish to crates.io: 2022-07-25 Items: - [x] Update changelog and readme: https://github.com/apache/arrow-rs/pull/2120 - [x] Create release candidate https://lists.apache.org/thread/txw3hlm06qjwczcbn20mjlgp7gf178fc - [x] Release candidate approved https://lists.apache.org/thread/3stno7fw741bpxo81s0fph80cbcymztq - [x] Release to crates.io - [x] Draft update to DataFusion: https://github.com/apache/arrow-datafusion/pull/2955 See full list here: https://github.com/apache/arrow-rs/compare/18.0.0...19.0.0 https://github.com/apache/arrow-rs/issues/1952 cc @viirya @jhorstmann @tustvold @sunchao @HaoYang670
process
release arrow next release after planned release candidate planned release and publish to crates io items update changelog and readme create release candidate release candidate approved release to crates io draft update to datafusion see full list here cc viirya jhorstmann tustvold sunchao
1
20,847
27,626,441,121
IssuesEvent
2023-03-10 07:16:25
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
iOS platform constraints locations
P2 type: support / not a bug (process) team-Configurability
In order to use bazel toolchains with Apple platforms, we have to define the core set of platforms it supports. This is mostly done by https://github.com/bazelbuild/bazel/blob/96c8a9073807c9e97635ddafe2ed0365a9318d6f/tools/osx/crosstool/BUILD.toolchains The issue is that since M1 macs were released there are now overlapping OS + CPU combos, like `ios` + `arm64` that need more specific information for whether or not the build is for a physical device, or the simulator. To support this, these custom constraints exist: https://github.com/bazelbuild/apple_support/blob/2ec0ef3eb2954371596fbb1a225fb244fcde31a9/constraints/BUILD#L12-L26 Since they live in apple_support today, and bazel does not take them into account, as far as I can tell this setup isn't able to do what we want. It seems to me that we either need to move these constraints into bazel, and annotate the toolchains here, or move the toolchain definitions out of bazel and into apple_support (or another location?). I'm hoping we can use this issue to decide on this, which will help make some more progress on toolchain support for Apple platforms.
1.0
iOS platform constraints locations - In order to use bazel toolchains with Apple platforms, we have to define the core set of platforms it supports. This is mostly done by https://github.com/bazelbuild/bazel/blob/96c8a9073807c9e97635ddafe2ed0365a9318d6f/tools/osx/crosstool/BUILD.toolchains The issue is that since M1 macs were released there are now overlapping OS + CPU combos, like `ios` + `arm64` that need more specific information for whether or not the build is for a physical device, or the simulator. To support this, these custom constraints exist: https://github.com/bazelbuild/apple_support/blob/2ec0ef3eb2954371596fbb1a225fb244fcde31a9/constraints/BUILD#L12-L26 Since they live in apple_support today, and bazel does not take them into account, as far as I can tell this setup isn't able to do what we want. It seems to me that we either need to move these constraints into bazel, and annotate the toolchains here, or move the toolchain definitions out of bazel and into apple_support (or another location?). I'm hoping we can use this issue to decide on this, which will help make some more progress on toolchain support for Apple platforms.
process
ios platform constraints locations in order to use bazel toolchains with apple platforms we have to define the core set of platforms it supports this is mostly done by the issue is that since macs were released there are now overlapping os cpu combos like ios that need more specific information for whether or not the build is for a physical device or the simulator to support this these custom constraints exist since they live in apple support today and bazel does not take them into account as far as i can tell this setup isn t able to do what we want it seems to me that we either need to move these constraints into bazel and annotate the toolchains here or move the toolchain definitions out of bazel and into apple support or another location i m hoping we can use this issue to decide on this which will help make some more progress on toolchain support for apple platforms
1
18,997
24,993,561,547
IssuesEvent
2022-11-02 21:08:12
FreeCAD/FreeCAD
https://api.github.com/repos/FreeCAD/FreeCAD
closed
[Bug] FreeCAD Contribution policy and process is confusing.
Awaiting feedback Process
### Forums discussion https://forum.freecadweb.org/viewtopic.php?f=10&t=72769 ### Version This issue is organizational and affects all versions and the general development process ### Issue description The development process used by the FreeCAD project is not well defined. It's unclear to contributors how code contributions are to be structured, submitted, evaluated, and accepted. This can lead to several undesirable outcomes: - A contributor spends considerable effort on a feature and then the feature is not accepted. - Multiple conflicting solutions are implemented resulting in a confusing user experience - Code contributions languish because no one knows who is responsible for accepting them. - Potential contributors do not contribute and existing contributors leave the project in frustration. A process should be established that clearly documents how code contributions will be handled. The process should be fair and transparent. The documented process should be put under git version control like any other part of the source so that it can be referenced, changed, and tracked over time. ### Anything else? Much discussion has been had about the C4 process from the ZeroMQ project. https://rfc.zeromq.org/spec/42/ When a process is accepted, the wiki including https://wiki.freecadweb.org/Developer_hub should be reviewed and revised to reflect the process.
1.0
[Bug] FreeCAD Contribution policy and process is confusing. - ### Forums discussion https://forum.freecadweb.org/viewtopic.php?f=10&t=72769 ### Version This issue is organizational and affects all versions and the general development process ### Issue description The development process used by the FreeCAD project is not well defined. It's unclear to contributors how code contributions are to be structured, submitted, evaluated, and accepted. This can lead to several undesirable outcomes: - A contributor spends considerable effort on a feature and then the feature is not accepted. - Multiple conflicting solutions are implemented resulting in a confusing user experience - Code contributions languish because no one knows who is responsible for accepting them. - Potential contributors do not contribute and existing contributors leave the project in frustration. A process should be established that clearly documents how code contributions will be handled. The process should be fair and transparent. The documented process should be put under git version control like any other part of the source so that it can be referenced, changed, and tracked over time. ### Anything else? Much discussion has been had about the C4 process from the ZeroMQ project. https://rfc.zeromq.org/spec/42/ When a process is accepted, the wiki including https://wiki.freecadweb.org/Developer_hub should be reviewed and revised to reflect the process.
process
freecad contribution policy and process is confusing forums discussion version this issue is organizational and affects all versions and the general development process issue description the development process used by the freecad project is not well defined it s unclear to contributors how code contributions are to be structured submitted evaluated and accepted this can lead to several undesirable outcomes a contributor spends considerable effort on a feature and then the feature is not accepted multiple conflicting solutions are implemented resulting in a confusing user experience code contributions languish because no one knows who is responsible for accepting them potential contributors do not contribute and existing contributors leave the project in frustration a process should be established that clearly documents how code contributions will be handled the process should be fair and transparent the documented process should be put under git version control like any other part of the source so that it can be referenced changed and tracked over time anything else much discussion has been had about the process from the zeromq project when a process is accepted the wiki including should be reviewed and revised to reflect the process
1
369,415
25,844,220,453
IssuesEvent
2022-12-13 04:25:31
BiteSnail/ARPWithMFC
https://api.github.com/repos/BiteSnail/ARPWithMFC
closed
IP Layer 구현
documentation help wanted
## IP Routing table 필요 변수 `Destination IP` 4byte : unsigned char `NetMask` 4byte : unsigned char `GateWay` 4byte : unsigned char `Flag` 1byte : unsigned char (bitwise 이용 1:Host, 2:Gateway, 4:Up) `Interface` 1byte : unsigned char (index로 이용할 예정 1~255) ## 추가 함수 routing table에 아이템 추가 routing table에 아이템 삭제 ... 기타 필요한 함수 추가 바람 --- 만약 가능하다면 flag와 interface를 합칠 수 있을 듯. 1byte에서 왼쪽 4bit은 flag, 오른쪽 4bit은 0~15값을 가지는 interface 인덱스 | 0000| 0000 | | -- | -- | | flag| interface|
1.0
IP Layer 구현 - ## IP Routing table 필요 변수 `Destination IP` 4byte : unsigned char `NetMask` 4byte : unsigned char `GateWay` 4byte : unsigned char `Flag` 1byte : unsigned char (bitwise 이용 1:Host, 2:Gateway, 4:Up) `Interface` 1byte : unsigned char (index로 이용할 예정 1~255) ## 추가 함수 routing table에 아이템 추가 routing table에 아이템 삭제 ... 기타 필요한 함수 추가 바람 --- 만약 가능하다면 flag와 interface를 합칠 수 있을 듯. 1byte에서 왼쪽 4bit은 flag, 오른쪽 4bit은 0~15값을 가지는 interface 인덱스 | 0000| 0000 | | -- | -- | | flag| interface|
non_process
ip layer 구현 ip routing table 필요 변수 destination ip unsigned char netmask unsigned char gateway unsigned char flag unsigned char bitwise 이용 host gateway up interface unsigned char index로 이용할 예정 추가 함수 routing table에 아이템 추가 routing table에 아이템 삭제 기타 필요한 함수 추가 바람 만약 가능하다면 flag와 interface를 합칠 수 있을 듯 왼쪽 flag 오른쪽 가지는 interface 인덱스 flag interface
0
1,431
3,995,827,290
IssuesEvent
2016-05-10 16:43:09
PHPOffice/PHPWord
https://api.github.com/repos/PHPOffice/PHPWord
closed
Support arrays in TemplateProcessor.setValue method
Change Request Template Processor
To do: - arrays in `$subject` to handle headers, footers and main document part at once. - arrays in `$search` and in `$replacement` to avoid looping on client side. Requested by @OAFCROB (see https://github.com/PHPOffice/PHPWord/issues/513#issuecomment-136964062 for the details).
1.0
Support arrays in TemplateProcessor.setValue method - To do: - arrays in `$subject` to handle headers, footers and main document part at once. - arrays in `$search` and in `$replacement` to avoid looping on client side. Requested by @OAFCROB (see https://github.com/PHPOffice/PHPWord/issues/513#issuecomment-136964062 for the details).
process
support arrays in templateprocessor setvalue method to do arrays in subject to handle headers footers and main document part at once arrays in search and in replacement to avoid looping on client side requested by oafcrob see for the details
1
52,744
3,028,354,365
IssuesEvent
2015-08-04 04:15:18
GoogleCloudPlatform/kubernetes
https://api.github.com/repos/GoogleCloudPlatform/kubernetes
closed
Replacing a service fails trying to modify clusterip
area/usability component/kubectl priority/P2 team/CSI
It is confusing that this doesn't work, especially since clusterIP was not part of the initial manifest or the replacement. ``` $ cat svc.yaml apiVersion: v1 kind: Service metadata: name: nginxsvc spec: type: NodePort ports: - port: 80 name: http protocol: TCP selector: app: nginx $ kubectl create -f svc.yaml services/nginxsvc $ cat svc.yaml apiVersion: v1 kind: Service metadata: name: nginxsvc spec: type: NodePort ports: - port: 80 name: http protocol: TCP - port: 443 name: https protocol: TCP selector: app: nginx $ kubectl replace -f svc.yaml Replace failedspec.clusterIP: invalid value '': field is immutable $ kubectl delete -f svc.yaml; kubeclt create -f svc.yaml services/nginxsvc ```
1.0
Replacing a service fails trying to modify clusterip - It is confusing that this doesn't work, especially since clusterIP was not part of the initial manifest or the replacement. ``` $ cat svc.yaml apiVersion: v1 kind: Service metadata: name: nginxsvc spec: type: NodePort ports: - port: 80 name: http protocol: TCP selector: app: nginx $ kubectl create -f svc.yaml services/nginxsvc $ cat svc.yaml apiVersion: v1 kind: Service metadata: name: nginxsvc spec: type: NodePort ports: - port: 80 name: http protocol: TCP - port: 443 name: https protocol: TCP selector: app: nginx $ kubectl replace -f svc.yaml Replace failedspec.clusterIP: invalid value '': field is immutable $ kubectl delete -f svc.yaml; kubeclt create -f svc.yaml services/nginxsvc ```
non_process
replacing a service fails trying to modify clusterip it is confusing that this doesn t work especially since clusterip was not part of the initial manifest or the replacement cat svc yaml apiversion kind service metadata name nginxsvc spec type nodeport ports port name http protocol tcp selector app nginx kubectl create f svc yaml services nginxsvc cat svc yaml apiversion kind service metadata name nginxsvc spec type nodeport ports port name http protocol tcp port name https protocol tcp selector app nginx kubectl replace f svc yaml replace failedspec clusterip invalid value field is immutable kubectl delete f svc yaml kubeclt create f svc yaml services nginxsvc
0
283,620
8,721,302,906
IssuesEvent
2018-12-08 21:35:53
bounswe/bounswe2018group6
https://api.github.com/repos/bounswe/bounswe2018group6
closed
Check email server procedures
Back-End Priority: High Status: Blocked Type: Bug
It seems sign up confirmation emails are not sent. Need to be fixed.
1.0
Check email server procedures - It seems sign up confirmation emails are not sent. Need to be fixed.
non_process
check email server procedures it seems sign up confirmation emails are not sent need to be fixed
0
3,630
6,665,333,434
IssuesEvent
2017-10-03 00:27:43
IIIF/api
https://api.github.com/repos/IIIF/api
closed
How best to version 0.x.y to maintain history?
process
Sensibly, we don't track patch level versions and just edit in place. However, in 0.9 for both search and auth, there were significant changes at "patch" levels which then disappear. As the process converges on a 1.0 spec, the last 0.9.y == 1.0 ... meaning there's just two copies of the same thing. I think we should have stronger process for our 0.x.y specs, such that significant changes increment the minor version number. So we would currently be at 0.10 for auth, as we were happy with 0.9 using JSONP.
1.0
How best to version 0.x.y to maintain history? - Sensibly, we don't track patch level versions and just edit in place. However, in 0.9 for both search and auth, there were significant changes at "patch" levels which then disappear. As the process converges on a 1.0 spec, the last 0.9.y == 1.0 ... meaning there's just two copies of the same thing. I think we should have stronger process for our 0.x.y specs, such that significant changes increment the minor version number. So we would currently be at 0.10 for auth, as we were happy with 0.9 using JSONP.
process
how best to version x y to maintain history sensibly we don t track patch level versions and just edit in place however in for both search and auth there were significant changes at patch levels which then disappear as the process converges on a spec the last y meaning there s just two copies of the same thing i think we should have stronger process for our x y specs such that significant changes increment the minor version number so we would currently be at for auth as we were happy with using jsonp
1
4,245
7,187,149,121
IssuesEvent
2018-02-02 03:14:32
Great-Hill-Corporation/quickBlocks
https://api.github.com/repos/Great-Hill-Corporation/quickBlocks
closed
The 'which' parameter stored in the transaction cache is wrong
monitors-all status-inprocess type-bug
I store this, but I don't need it. I think I stored it so I wouldn't have to find it later by spinning through watches, but since I started coloring and nicknaming addresses, I spin through watches anyway. I would remove it as it makes duplicate processing in teh cacheMan incorrect and may result in double account if a transaction cache is merged with another one with a different 'which'. Only teh transaction matters. Coloring it does not need to be optimized.
1.0
The 'which' parameter stored in the transaction cache is wrong - I store this, but I don't need it. I think I stored it so I wouldn't have to find it later by spinning through watches, but since I started coloring and nicknaming addresses, I spin through watches anyway. I would remove it as it makes duplicate processing in teh cacheMan incorrect and may result in double account if a transaction cache is merged with another one with a different 'which'. Only teh transaction matters. Coloring it does not need to be optimized.
process
the which parameter stored in the transaction cache is wrong i store this but i don t need it i think i stored it so i wouldn t have to find it later by spinning through watches but since i started coloring and nicknaming addresses i spin through watches anyway i would remove it as it makes duplicate processing in teh cacheman incorrect and may result in double account if a transaction cache is merged with another one with a different which only teh transaction matters coloring it does not need to be optimized
1
87,479
17,273,033,504
IssuesEvent
2021-07-22 23:12:56
learnpack/learnpack
https://api.github.com/repos/learnpack/learnpack
opened
When opening one exercise with several files, only the last one gets opened
vscode plugin 👽
All this behavior is happening in `grading: incremental` The way the the plugin works, if the exercise has 3 files to open, for example: index.html, index.js and style.css. It will open index.html, but then when it opens index.js it will replace the same TextEditor with the content of index.js (removing the index.html that was there in the first place), and then it will do the same with style.css. This behavior was ideal for one file exercises but long term is better to make sure that a `new` editor is opened instead of reusing the old one. Note: when another exercise is opened (with all of its files) we need to make sure the previous files are closed to avoid overwhelming the user with too many files.
1.0
When opening one exercise with several files, only the last one gets opened - All this behavior is happening in `grading: incremental` The way the the plugin works, if the exercise has 3 files to open, for example: index.html, index.js and style.css. It will open index.html, but then when it opens index.js it will replace the same TextEditor with the content of index.js (removing the index.html that was there in the first place), and then it will do the same with style.css. This behavior was ideal for one file exercises but long term is better to make sure that a `new` editor is opened instead of reusing the old one. Note: when another exercise is opened (with all of its files) we need to make sure the previous files are closed to avoid overwhelming the user with too many files.
non_process
when opening one exercise with several files only the last one gets opened all this behavior is happening in grading incremental the way the the plugin works if the exercise has files to open for example index html index js and style css it will open index html but then when it opens index js it will replace the same texteditor with the content of index js removing the index html that was there in the first place and then it will do the same with style css this behavior was ideal for one file exercises but long term is better to make sure that a new editor is opened instead of reusing the old one note when another exercise is opened with all of its files we need to make sure the previous files are closed to avoid overwhelming the user with too many files
0
265,722
8,357,899,425
IssuesEvent
2018-10-02 23:34:52
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
closed
Typo (need dot) in text
Localization Low Priority
> Flood fixing tool! Left click to remove the top water layer. (Target a wall touching water) https://crowdin.com/translate/eco-by-strange-loop-games/24/en-ru#52135 maybe need dot at end? :) @Nanders
1.0
Typo (need dot) in text - > Flood fixing tool! Left click to remove the top water layer. (Target a wall touching water) https://crowdin.com/translate/eco-by-strange-loop-games/24/en-ru#52135 maybe need dot at end? :) @Nanders
non_process
typo need dot in text flood fixing tool left click to remove the top water layer target a wall touching water maybe need dot at end nanders
0
18,804
24,703,988,100
IssuesEvent
2022-10-19 17:26:52
eosnetworkfoundation/devrel
https://api.github.com/repos/eosnetworkfoundation/devrel
closed
Define a "Definition of Done" document
Process
AC: - brainstorm meeting, DevRel team - Approval by stakeholder (Nathan) - publish in DevRel repository1
1.0
Define a "Definition of Done" document - AC: - brainstorm meeting, DevRel team - Approval by stakeholder (Nathan) - publish in DevRel repository1
process
define a definition of done document ac brainstorm meeting devrel team approval by stakeholder nathan publish in devrel
1
300,665
22,691,525,524
IssuesEvent
2022-07-04 21:12:39
balbasty/torch-interpol
https://api.github.com/repos/balbasty/torch-interpol
closed
More examples
documentation question
I need to interpolate a SDF of say 128 x 128 x 128, as well as calculate normals (gradients). [Link](https://docs.scipy.org/doc/scipy/reference/generated/scipy.interpolate.griddata.html) to `interpolate` in `scipy`. Could you please provide an easy-to-use example?
1.0
More examples - I need to interpolate a SDF of say 128 x 128 x 128, as well as calculate normals (gradients). [Link](https://docs.scipy.org/doc/scipy/reference/generated/scipy.interpolate.griddata.html) to `interpolate` in `scipy`. Could you please provide an easy-to-use example?
non_process
more examples i need to interpolate a sdf of say x x as well as calculate normals gradients to interpolate in scipy could you please provide an easy to use example
0
238
2,663,301,274
IssuesEvent
2015-03-20 03:29:56
luc-github/Repetier-Firmware-0.92
https://api.github.com/repos/luc-github/Repetier-Firmware-0.92
closed
G10 and G11 commands and M209, M355, M600, M360
enhancement Waiting to be processed
these commands are now available in repetier, need to importe them
1.0
G10 and G11 commands and M209, M355, M600, M360 - these commands are now available in repetier, need to importe them
process
and commands and these commands are now available in repetier need to importe them
1
10,511
13,283,605,171
IssuesEvent
2020-08-24 03:48:15
valavila/eCommerceSite
https://api.github.com/repos/valavila/eCommerceSite
closed
Add CI PipeLine
developer process
Add continuous integration pipline that will check to make sure code in a pull request complies successfully
1.0
Add CI PipeLine - Add continuous integration pipline that will check to make sure code in a pull request complies successfully
process
add ci pipeline add continuous integration pipline that will check to make sure code in a pull request complies successfully
1
6,160
9,039,026,699
IssuesEvent
2019-02-10 01:29:18
material-components/material-components-ios
https://api.github.com/repos/material-components/material-components-ios
closed
[BottomNavigation] Finalize BottomNavigationController API
[BottomNavigation] type:Process
This was filed as an internal issue. If you are a Googler, please visit [b/119188596](http://b/119188596) for more details. <!-- Auto-generated content below, do not modify --> --- #### Internal data - Associated internal bug: [b/119188596](http://b/119188596)
1.0
[BottomNavigation] Finalize BottomNavigationController API - This was filed as an internal issue. If you are a Googler, please visit [b/119188596](http://b/119188596) for more details. <!-- Auto-generated content below, do not modify --> --- #### Internal data - Associated internal bug: [b/119188596](http://b/119188596)
process
finalize bottomnavigationcontroller api this was filed as an internal issue if you are a googler please visit for more details internal data associated internal bug
1
15,412
19,599,552,476
IssuesEvent
2022-01-05 22:37:32
googleapis/gapic-generator-java
https://api.github.com/repos/googleapis/gapic-generator-java
opened
Run tests for "self-service" libraries in CI
type: process priority: p2
Doing so will help validating changes to the "self-service" library generation workflow (e.g., as in #876). Theoretically, we should be able to generate some libraries against actual API service protos like the following (may need to change a few things in `WORKSPACE` or some `BUILD` files). ``` bazel build @com_google_googleapis//google/cloud/compute/v1:google-cloud-compute-v1-java ``` Then unpack `bazel-bin/external/com_google_googleapis/google/cloud/compute/v1/google-cloud-compute-v1-java.tar.gz` and run `./gradlew clean check publishToMavenLocal`. `compute` is REST and all others are gRPC at the moment, so `compute` and another library are good candidates.
1.0
Run tests for "self-service" libraries in CI - Doing so will help validating changes to the "self-service" library generation workflow (e.g., as in #876). Theoretically, we should be able to generate some libraries against actual API service protos like the following (may need to change a few things in `WORKSPACE` or some `BUILD` files). ``` bazel build @com_google_googleapis//google/cloud/compute/v1:google-cloud-compute-v1-java ``` Then unpack `bazel-bin/external/com_google_googleapis/google/cloud/compute/v1/google-cloud-compute-v1-java.tar.gz` and run `./gradlew clean check publishToMavenLocal`. `compute` is REST and all others are gRPC at the moment, so `compute` and another library are good candidates.
process
run tests for self service libraries in ci doing so will help validating changes to the self service library generation workflow e g as in theoretically we should be able to generate some libraries against actual api service protos like the following may need to change a few things in workspace or some build files bazel build com google googleapis google cloud compute google cloud compute java then unpack bazel bin external com google googleapis google cloud compute google cloud compute java tar gz and run gradlew clean check publishtomavenlocal compute is rest and all others are grpc at the moment so compute and another library are good candidates
1
39,834
16,102,615,605
IssuesEvent
2021-04-27 11:18:25
microsoft/vscode-cpptools
https://api.github.com/repos/microsoft/vscode-cpptools
closed
Extension causes high cpu load
Language Service more info needed
- Issue Type: `Performance` - Extension Name: `cpptools` - Extension Version: `1.2.2-insiders2` - OS Version: `Windows_NT x64 10.0.18362` - VSCode version: `1.53.2` :warning: Make sure to **attach** this file [ms-vscode.cpptools-unresponsive.cpuprofile.txt](https://github.com/microsoft/vscode-cpptools/files/6026948/ms-vscode.cpptools-unresponsive.cpuprofile.txt) from your *home*-directory: :warning:`c:\Users\*******\AppData\Local\Temp\ms-vscode.cpptools-unresponsive.cpuprofile.txt` Find more details here: https://github.com/microsoft/vscode/wiki/Explain-extension-causes-high-cpu-load
1.0
Extension causes high cpu load - - Issue Type: `Performance` - Extension Name: `cpptools` - Extension Version: `1.2.2-insiders2` - OS Version: `Windows_NT x64 10.0.18362` - VSCode version: `1.53.2` :warning: Make sure to **attach** this file [ms-vscode.cpptools-unresponsive.cpuprofile.txt](https://github.com/microsoft/vscode-cpptools/files/6026948/ms-vscode.cpptools-unresponsive.cpuprofile.txt) from your *home*-directory: :warning:`c:\Users\*******\AppData\Local\Temp\ms-vscode.cpptools-unresponsive.cpuprofile.txt` Find more details here: https://github.com/microsoft/vscode/wiki/Explain-extension-causes-high-cpu-load
non_process
extension causes high cpu load issue type performance extension name cpptools extension version os version windows nt vscode version warning make sure to attach this file from your home directory warning c users appdata local temp ms vscode cpptools unresponsive cpuprofile txt find more details here
0
14,259
17,192,791,672
IssuesEvent
2021-07-16 13:24:11
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Managed identity is generally available, please remove "Preview"
Pri2 automation/svc cxp doc-bug doc-enhancement process-automation/subsvc triaged
https://docs.microsoft.com/en-us/azure/automation/automation-security-overview - says MI is in preview. I checked in portal and there is no preview warning when I enabled this feature. https://azure.microsoft.com/en-au/updates/azure-automation-system-assigned-managed-identities/ -- for reference --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 8721e209-24ce-2170-6caa-ed12a7060080 * Version Independent ID: ac13f91d-460c-cbe9-4778-50d20765b252 * Content: [Azure Automation account authentication overview](https://docs.microsoft.com/en-us/azure/automation/automation-security-overview) * Content Source: [articles/automation/automation-security-overview.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/automation-security-overview.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @MGoedtel * Microsoft Alias: **magoedte**
1.0
Managed identity is generally available, please remove "Preview" - https://docs.microsoft.com/en-us/azure/automation/automation-security-overview - says MI is in preview. I checked in portal and there is no preview warning when I enabled this feature. https://azure.microsoft.com/en-au/updates/azure-automation-system-assigned-managed-identities/ -- for reference --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 8721e209-24ce-2170-6caa-ed12a7060080 * Version Independent ID: ac13f91d-460c-cbe9-4778-50d20765b252 * Content: [Azure Automation account authentication overview](https://docs.microsoft.com/en-us/azure/automation/automation-security-overview) * Content Source: [articles/automation/automation-security-overview.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/automation-security-overview.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @MGoedtel * Microsoft Alias: **magoedte**
process
managed identity is generally available please remove preview says mi is in preview i checked in portal and there is no preview warning when i enabled this feature for reference document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login mgoedtel microsoft alias magoedte
1
7,678
10,762,210,840
IssuesEvent
2019-10-31 22:52:56
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
opened
Processing "output type" should default to "auto"
Bug Processing
In QGIS 3/Processing/GRASS the parameter "output type" has no selection by default and is mandatory. This is *very* inconvenient, up to the old LTR this parameter defaulted to "auto", which is the right choice. Moreover there are Processing/GRASS modules where the output type is hardcoded and this parameter does not make sense anyway, i.e. https://github.com/qgis/QGIS/blob/master/python/plugins/processing/algs/grass7/ext/v_net_centrality.py using "auto" as default would also help create less confusion with this parameters.
1.0
Processing "output type" should default to "auto" - In QGIS 3/Processing/GRASS the parameter "output type" has no selection by default and is mandatory. This is *very* inconvenient, up to the old LTR this parameter defaulted to "auto", which is the right choice. Moreover there are Processing/GRASS modules where the output type is hardcoded and this parameter does not make sense anyway, i.e. https://github.com/qgis/QGIS/blob/master/python/plugins/processing/algs/grass7/ext/v_net_centrality.py using "auto" as default would also help create less confusion with this parameters.
process
processing output type should default to auto in qgis processing grass the parameter output type has no selection by default and is mandatory this is very inconvenient up to the old ltr this parameter defaulted to auto which is the right choice moreover there are processing grass modules where the output type is hardcoded and this parameter does not make sense anyway i e using auto as default would also help create less confusion with this parameters
1
14,400
17,455,867,444
IssuesEvent
2021-08-06 01:01:38
marysaray/eCommerceWebsiteExample
https://api.github.com/repos/marysaray/eCommerceWebsiteExample
closed
Add CI Pipeline
development process
Add continuous integration pipeline that will check and make sure the code in a pull request compiles successfully.
1.0
Add CI Pipeline - Add continuous integration pipeline that will check and make sure the code in a pull request compiles successfully.
process
add ci pipeline add continuous integration pipeline that will check and make sure the code in a pull request compiles successfully
1
10,332
13,162,980,294
IssuesEvent
2020-08-10 22:59:17
googleapis/google-cloud-ruby
https://api.github.com/repos/googleapis/google-cloud-ruby
closed
Migrate google-cloud-debugger to the microgenerator
type: process
Migrate google-cloud-debugger to the microgenerator. This involves the following steps: * [x] Write synth file and generate `google-cloud-debugger-v2` * [x] Make sure the new libraries are configured in kokoro * [x] Release `google-cloud-debugger-v2` * [ ] Switch `google-cloud-debugger` backend to the versioned gems. That is: * Rip out synth and all the generated code * Add `google-cloud-debugger-v2` as a dependency * Update the veneer code to the microgenerator usage * [ ] Release `google-cloud-debugger` update. Note that it should remain pre-1.0. We are not GA-ing this library. I do not believe samples need to be updated, unless they invoke the low-level interface directly.
1.0
Migrate google-cloud-debugger to the microgenerator - Migrate google-cloud-debugger to the microgenerator. This involves the following steps: * [x] Write synth file and generate `google-cloud-debugger-v2` * [x] Make sure the new libraries are configured in kokoro * [x] Release `google-cloud-debugger-v2` * [ ] Switch `google-cloud-debugger` backend to the versioned gems. That is: * Rip out synth and all the generated code * Add `google-cloud-debugger-v2` as a dependency * Update the veneer code to the microgenerator usage * [ ] Release `google-cloud-debugger` update. Note that it should remain pre-1.0. We are not GA-ing this library. I do not believe samples need to be updated, unless they invoke the low-level interface directly.
process
migrate google cloud debugger to the microgenerator migrate google cloud debugger to the microgenerator this involves the following steps write synth file and generate google cloud debugger make sure the new libraries are configured in kokoro release google cloud debugger switch google cloud debugger backend to the versioned gems that is rip out synth and all the generated code add google cloud debugger as a dependency update the veneer code to the microgenerator usage release google cloud debugger update note that it should remain pre we are not ga ing this library i do not believe samples need to be updated unless they invoke the low level interface directly
1
335,457
24,468,926,815
IssuesEvent
2022-10-07 17:42:21
EnceeFinancials/XYO-NODE
https://api.github.com/repos/EnceeFinancials/XYO-NODE
closed
[DOCUMENTATION]:
documentation help wanted
**Is your documentation request related to a problem? Please describe.** A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] **Describe the documentation and format you would like** A clear and concise description of what documentation you would like to see and what type for format. Ex. Step-by-step, Paragraph explainer, screenshots, etc. **Additional context** Add any other context or screenshots about the document request here.
1.0
[DOCUMENTATION]: - **Is your documentation request related to a problem? Please describe.** A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] **Describe the documentation and format you would like** A clear and concise description of what documentation you would like to see and what type for format. Ex. Step-by-step, Paragraph explainer, screenshots, etc. **Additional context** Add any other context or screenshots about the document request here.
non_process
is your documentation request related to a problem please describe a clear and concise description of what the problem is ex i m always frustrated when describe the documentation and format you would like a clear and concise description of what documentation you would like to see and what type for format ex step by step paragraph explainer screenshots etc additional context add any other context or screenshots about the document request here
0
326,714
24,098,467,673
IssuesEvent
2022-09-19 21:09:46
Open-Telecoms-Data/open-fibre-data-standard
https://api.github.com/repos/Open-Telecoms-Data/open-fibre-data-standard
closed
Draft documentation on 'what does OFDS data look like'
Non-normative documentation
In the primer documentation on the [Open Fibre Data Standard](https://github.com/Open-Telecoms-Data/open-fibre-data-standard/blob/main/docs/primer/openfibredatastandard.md) the section on 'what does OFDS data look like is not yet drafted. This should be completed once decisions on publication formats are finalised.
1.0
Draft documentation on 'what does OFDS data look like' - In the primer documentation on the [Open Fibre Data Standard](https://github.com/Open-Telecoms-Data/open-fibre-data-standard/blob/main/docs/primer/openfibredatastandard.md) the section on 'what does OFDS data look like is not yet drafted. This should be completed once decisions on publication formats are finalised.
non_process
draft documentation on what does ofds data look like in the primer documentation on the the section on what does ofds data look like is not yet drafted this should be completed once decisions on publication formats are finalised
0
5,372
8,202,496,346
IssuesEvent
2018-09-02 10:08:13
linnovate/root
https://api.github.com/repos/linnovate/root
closed
Folders : can't download file from Attachment Preview
Process bug
@abrahamos upload file in folder. click on the attachment preview. click on download. get a Restricted accses error.
1.0
Folders : can't download file from Attachment Preview - @abrahamos upload file in folder. click on the attachment preview. click on download. get a Restricted accses error.
process
folders can t download file from attachment preview abrahamos upload file in folder click on the attachment preview click on download get a restricted accses error
1
8,264
11,426,558,101
IssuesEvent
2020-02-03 22:13:05
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
"Random points along line" algorithm execution fails every two calls
Bug Feedback Processing
I ran the alg multiple times with the same parameters and concluded that it works every two calls and fails as well every two calls. The error log: > QGIS version: 3.11.0-Master QGIS code revision: c52a6e711a Qt version: 5.11.2 GDAL version: 3.1.0dev GEOS version: 3.8.0-CAPI-1.13.1 PROJ version: Rel. 7.0.0, March 1st, 2020 Processing algorithm… Algorithm 'Random points along line' starting… Input parameters: { 'INPUT' : '//serveur//Études/doss/candidatures.gpkg|layername=A_laisons_douces', 'MIN_DISTANCE' : 0, 'OUTPUT' : 'TEMPORARY_OUTPUT', 'POINTS_NUMBER' : 5 } > >Traceback (most recent call last): File "C:/OSGEO4~1/apps/qgis-dev/./python/plugins\processing\algs\qgis\RandomPointsAlongLines.py", line 126, in processAlgorithm f = next(source.getFeatures(request.setFilterFid(fid).setSubsetOfAttributes([]))) StopIteration > >Execution failed after 0.16 seconds > >Loading resulting layers Algorithm 'Random points along line' finished
1.0
"Random points along line" algorithm execution fails every two calls - I ran the alg multiple times with the same parameters and concluded that it works every two calls and fails as well every two calls. The error log: > QGIS version: 3.11.0-Master QGIS code revision: c52a6e711a Qt version: 5.11.2 GDAL version: 3.1.0dev GEOS version: 3.8.0-CAPI-1.13.1 PROJ version: Rel. 7.0.0, March 1st, 2020 Processing algorithm… Algorithm 'Random points along line' starting… Input parameters: { 'INPUT' : '//serveur//Études/doss/candidatures.gpkg|layername=A_laisons_douces', 'MIN_DISTANCE' : 0, 'OUTPUT' : 'TEMPORARY_OUTPUT', 'POINTS_NUMBER' : 5 } > >Traceback (most recent call last): File "C:/OSGEO4~1/apps/qgis-dev/./python/plugins\processing\algs\qgis\RandomPointsAlongLines.py", line 126, in processAlgorithm f = next(source.getFeatures(request.setFilterFid(fid).setSubsetOfAttributes([]))) StopIteration > >Execution failed after 0.16 seconds > >Loading resulting layers Algorithm 'Random points along line' finished
process
random points along line algorithm execution fails every two calls i ran the alg multiple times with the same parameters and concluded that it works every two calls and fails as well every two calls the error log qgis version master qgis code revision qt version gdal version geos version capi proj version rel march processing algorithm… algorithm random points along line starting… input parameters input serveur études doss candidatures gpkg layername a laisons douces min distance output temporary output points number traceback most recent call last file c apps qgis dev python plugins processing algs qgis randompointsalonglines py line in processalgorithm f next source getfeatures request setfilterfid fid setsubsetofattributes stopiteration execution failed after seconds loading resulting layers algorithm random points along line finished
1
4,798
7,695,110,669
IssuesEvent
2018-05-18 11:03:25
zero-os/0-robot
https://api.github.com/repos/zero-os/0-robot
closed
robot sometimes fail to start from flist
process_duplicate type_bug
I've seen this happening a few time on some node running the robot in a container used flist is `https://hub.gig.tech/gig-autobuilder/zero-os-0-robot-autostart-0.5.1.flist` ``` In [1264]: sub.stream() [Thu12 14:30] - GitFactory.py :316 :j.clients.git - INFO - None:pull:http://github.com/zero-os/0-templates.git ->/opt/code/github/zero-os/0-templates [Thu12 14:30] - GitFactory.py :399 :j.clients.git - INFO - git clone http://github.com/zero-os/0-templates.git -> /opt/code/github/zero-os/0-templates [Thu12 14:30] - GitFactory.py :419 :j.clients.git - INFO - mkdir -p /opt/code/github/zero-os;cd /opt/code/github/zero-os;git -c http.sslVerify=false clone -b master http://github.com/zero-os/0-templates.git /opt/code/github/zero-os/0-templates Traceback (most recent call last): File "/usr/local/bin/zrobot", line 6, in <module> exec(compile(open(__file__).read(), __file__, 'exec')) File "/opt/code/github/zero-os/0-robot/cmd/zrobot", line 25, in <module> entry_point() File "/opt/code/github/zero-os/0-robot/zerorobot/cli/server.py", line 66, in start robot.set_config_repo(config_repo, config_key) File "/opt/code/github/zero-os/0-robot/zerorobot/robot/robot.py", line 82, in set_config_repo config_repo.init(path, key) File "/opt/code/github/zero-os/0-robot/zerorobot/robot/config_repo.py", line 54, in init j.tools.configmanager.init(silent=True, configpath=path, keypath=key) File "/usr/local/lib/python3.5/dist-packages/JumpScale9/tools/configmanager/ConfigManager.py", line 432, in init j.clients.sshkey.key_get(keypath, load=True) File "/usr/local/lib/python3.5/dist-packages/JumpScale9/clients/sshkey/SSHKeys.py", line 25, in key_get sshkey = self.get(instance) File "/usr/local/lib/python3.5/dist-packages/JumpScale9/tools/configmanager/JSBaseClassConfigs.py", line 43, in get return self._child_class(instance=instance, data=data, parent=self, interactive=interactive,**kwargs) File "/usr/local/lib/python3.5/dist-packages/JumpScale9/clients/sshkey/SSHKey.py", line 23, in __init__ parent=parent, template=TEMPLATE, interactive=interactive) File "/usr/local/lib/python3.5/dist-packages/JumpScale9/tools/configmanager/JSBaseClassConfig.py", line 16, in __init__ self._ui = j.tools.formbuilder.baseclass_get() # is the default class File "/usr/lib/python3/dist-packages/js9.py", line 1709, in formbuilder from JumpScale9.tools.formbuilder.FormBuilder import FormBuilderFactory as FormBuilderFactory File "/usr/local/lib/python3.5/dist-packages/JumpScale9/tools/formbuilder/FormBuilder.py", line 3, in <module> import npyscreen File "/usr/local/lib/python3.5/dist-packages/npyscreen/__init__.py", line 5, in <module> from .wgwidget import TEST_SETTINGS, ExhaustedTestInput, add_test_input_from_iterable, add_test_input_ch File "/usr/local/lib/python3.5/dist-packages/npyscreen/wgwidget.py", line 5, in <module> import curses File "/usr/lib/python3.5/curses/__init__.py", line 13, in <module> from _curses import * ImportError: /lib/x86_64-linux-gnu/libncursesw.so.5: symbol LINES, version NCURSES_TINFO_5.0.19991023 not defined in file libtinfo.so.5 with link time reference [Thu12 14:30] - ErrorHandler.py :257 :j.core.errorhandler - ERROR - /lib/x86_64-linux-gnu/libncursesw.so.5: symbol LINES, version NCURSES_TINFO_5.0.19991023 not defined in file libtinfo.so.5 with link time reference ```
1.0
robot sometimes fail to start from flist - I've seen this happening a few time on some node running the robot in a container used flist is `https://hub.gig.tech/gig-autobuilder/zero-os-0-robot-autostart-0.5.1.flist` ``` In [1264]: sub.stream() [Thu12 14:30] - GitFactory.py :316 :j.clients.git - INFO - None:pull:http://github.com/zero-os/0-templates.git ->/opt/code/github/zero-os/0-templates [Thu12 14:30] - GitFactory.py :399 :j.clients.git - INFO - git clone http://github.com/zero-os/0-templates.git -> /opt/code/github/zero-os/0-templates [Thu12 14:30] - GitFactory.py :419 :j.clients.git - INFO - mkdir -p /opt/code/github/zero-os;cd /opt/code/github/zero-os;git -c http.sslVerify=false clone -b master http://github.com/zero-os/0-templates.git /opt/code/github/zero-os/0-templates Traceback (most recent call last): File "/usr/local/bin/zrobot", line 6, in <module> exec(compile(open(__file__).read(), __file__, 'exec')) File "/opt/code/github/zero-os/0-robot/cmd/zrobot", line 25, in <module> entry_point() File "/opt/code/github/zero-os/0-robot/zerorobot/cli/server.py", line 66, in start robot.set_config_repo(config_repo, config_key) File "/opt/code/github/zero-os/0-robot/zerorobot/robot/robot.py", line 82, in set_config_repo config_repo.init(path, key) File "/opt/code/github/zero-os/0-robot/zerorobot/robot/config_repo.py", line 54, in init j.tools.configmanager.init(silent=True, configpath=path, keypath=key) File "/usr/local/lib/python3.5/dist-packages/JumpScale9/tools/configmanager/ConfigManager.py", line 432, in init j.clients.sshkey.key_get(keypath, load=True) File "/usr/local/lib/python3.5/dist-packages/JumpScale9/clients/sshkey/SSHKeys.py", line 25, in key_get sshkey = self.get(instance) File "/usr/local/lib/python3.5/dist-packages/JumpScale9/tools/configmanager/JSBaseClassConfigs.py", line 43, in get return self._child_class(instance=instance, data=data, parent=self, interactive=interactive,**kwargs) File "/usr/local/lib/python3.5/dist-packages/JumpScale9/clients/sshkey/SSHKey.py", line 23, in __init__ parent=parent, template=TEMPLATE, interactive=interactive) File "/usr/local/lib/python3.5/dist-packages/JumpScale9/tools/configmanager/JSBaseClassConfig.py", line 16, in __init__ self._ui = j.tools.formbuilder.baseclass_get() # is the default class File "/usr/lib/python3/dist-packages/js9.py", line 1709, in formbuilder from JumpScale9.tools.formbuilder.FormBuilder import FormBuilderFactory as FormBuilderFactory File "/usr/local/lib/python3.5/dist-packages/JumpScale9/tools/formbuilder/FormBuilder.py", line 3, in <module> import npyscreen File "/usr/local/lib/python3.5/dist-packages/npyscreen/__init__.py", line 5, in <module> from .wgwidget import TEST_SETTINGS, ExhaustedTestInput, add_test_input_from_iterable, add_test_input_ch File "/usr/local/lib/python3.5/dist-packages/npyscreen/wgwidget.py", line 5, in <module> import curses File "/usr/lib/python3.5/curses/__init__.py", line 13, in <module> from _curses import * ImportError: /lib/x86_64-linux-gnu/libncursesw.so.5: symbol LINES, version NCURSES_TINFO_5.0.19991023 not defined in file libtinfo.so.5 with link time reference [Thu12 14:30] - ErrorHandler.py :257 :j.core.errorhandler - ERROR - /lib/x86_64-linux-gnu/libncursesw.so.5: symbol LINES, version NCURSES_TINFO_5.0.19991023 not defined in file libtinfo.so.5 with link time reference ```
process
robot sometimes fail to start from flist i ve seen this happening a few time on some node running the robot in a container used flist is in sub stream gitfactory py j clients git info none pull opt code github zero os templates gitfactory py j clients git info git clone opt code github zero os templates gitfactory py j clients git info mkdir p opt code github zero os cd opt code github zero os git c http sslverify false clone b master opt code github zero os templates traceback most recent call last file usr local bin zrobot line in exec compile open file read file exec file opt code github zero os robot cmd zrobot line in entry point file opt code github zero os robot zerorobot cli server py line in start robot set config repo config repo config key file opt code github zero os robot zerorobot robot robot py line in set config repo config repo init path key file opt code github zero os robot zerorobot robot config repo py line in init j tools configmanager init silent true configpath path keypath key file usr local lib dist packages tools configmanager configmanager py line in init j clients sshkey key get keypath load true file usr local lib dist packages clients sshkey sshkeys py line in key get sshkey self get instance file usr local lib dist packages tools configmanager jsbaseclassconfigs py line in get return self child class instance instance data data parent self interactive interactive kwargs file usr local lib dist packages clients sshkey sshkey py line in init parent parent template template interactive interactive file usr local lib dist packages tools configmanager jsbaseclassconfig py line in init self ui j tools formbuilder baseclass get is the default class file usr lib dist packages py line in formbuilder from tools formbuilder formbuilder import formbuilderfactory as formbuilderfactory file usr local lib dist packages tools formbuilder formbuilder py line in import npyscreen file usr local lib dist packages npyscreen init py line in from wgwidget import test settings exhaustedtestinput add test input from iterable add test input ch file usr local lib dist packages npyscreen wgwidget py line in import curses file usr lib curses init py line in from curses import importerror lib linux gnu libncursesw so symbol lines version ncurses tinfo not defined in file libtinfo so with link time reference errorhandler py j core errorhandler error lib linux gnu libncursesw so symbol lines version ncurses tinfo not defined in file libtinfo so with link time reference
1
2,081
4,896,463,930
IssuesEvent
2016-11-20 11:14:24
Alfresco/alfresco-ng2-components
https://api.github.com/repos/Alfresco/alfresco-ng2-components
closed
Can navigate around activit app while start process dialog is present
browser: firefox browser: safari bug comp: activiti-processList
1. Click start process 2. Navigate around Activit app 3. Navigate back to process list **Expected result** While dialog is present cannot navigate around app (behaviour in Chrome) **Actual result** While dialog is present can navigate around app (behaviour in Firefox and Safari)
1.0
Can navigate around activit app while start process dialog is present - 1. Click start process 2. Navigate around Activit app 3. Navigate back to process list **Expected result** While dialog is present cannot navigate around app (behaviour in Chrome) **Actual result** While dialog is present can navigate around app (behaviour in Firefox and Safari)
process
can navigate around activit app while start process dialog is present click start process navigate around activit app navigate back to process list expected result while dialog is present cannot navigate around app behaviour in chrome actual result while dialog is present can navigate around app behaviour in firefox and safari
1
47,690
13,066,100,017
IssuesEvent
2020-07-30 20:59:50
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
closed
sim-services - missing sphinx docs (Trac #996)
Migrated from Trac combo simulation defect
`sim-services/resources/docs/sanity_checkers/sanity_checkers.rst` isn't picked up by sphinx on `make docs` Migrated from https://code.icecube.wisc.edu/ticket/996 ```json { "status": "closed", "changetime": "2015-05-26T21:51:48", "description": "`sim-services/resources/docs/sanity_checkers/sanity_checkers.rst` isn't picked up by sphinx on `make docs`", "reporter": "nega", "cc": "olivas", "resolution": "fixed", "_ts": "1432677108464162", "component": "combo simulation", "summary": "sim-services - missing sphinx docs", "priority": "normal", "keywords": "documentation", "time": "2015-05-26T20:41:22", "milestone": "", "owner": "nega", "type": "defect" } ```
1.0
sim-services - missing sphinx docs (Trac #996) - `sim-services/resources/docs/sanity_checkers/sanity_checkers.rst` isn't picked up by sphinx on `make docs` Migrated from https://code.icecube.wisc.edu/ticket/996 ```json { "status": "closed", "changetime": "2015-05-26T21:51:48", "description": "`sim-services/resources/docs/sanity_checkers/sanity_checkers.rst` isn't picked up by sphinx on `make docs`", "reporter": "nega", "cc": "olivas", "resolution": "fixed", "_ts": "1432677108464162", "component": "combo simulation", "summary": "sim-services - missing sphinx docs", "priority": "normal", "keywords": "documentation", "time": "2015-05-26T20:41:22", "milestone": "", "owner": "nega", "type": "defect" } ```
non_process
sim services missing sphinx docs trac sim services resources docs sanity checkers sanity checkers rst isn t picked up by sphinx on make docs migrated from json status closed changetime description sim services resources docs sanity checkers sanity checkers rst isn t picked up by sphinx on make docs reporter nega cc olivas resolution fixed ts component combo simulation summary sim services missing sphinx docs priority normal keywords documentation time milestone owner nega type defect
0
9,716
12,716,574,083
IssuesEvent
2020-06-24 02:21:24
OUDcollective/twenty20times
https://api.github.com/repos/OUDcollective/twenty20times
opened
Understanding the GitHub flow · GitHub Guides
workflow-process
![Screen Shot from awesomescreenshot.com](https://www.awesomescreenshot.com/api/v1/destination/image/show?ImageKey=tm-3919-15892-c3f432de651aff23e6f63fb3bc4005da) ## GIT FLOW --- **Source URL**: [https://guides.github.com/introduction/flow/](https://guides.github.com/introduction/flow/) <table><tr><td><strong>Browser</strong></td><td>Chrome 84.0.4147.56</td></tr><tr><td><strong>OS</strong></td><td>Windows 10 64-bit</td></tr><tr><td><strong>Screen Size</strong></td><td>2560x1080</td></tr><tr><td><strong>Viewport Size</strong></td><td>2560x888</td></tr><tr><td><strong>Pixel Ratio</strong></td><td>@1x</td></tr><tr><td><strong>Zoom Level</strong></td><td>100%</td></tr></table>
1.0
Understanding the GitHub flow · GitHub Guides - ![Screen Shot from awesomescreenshot.com](https://www.awesomescreenshot.com/api/v1/destination/image/show?ImageKey=tm-3919-15892-c3f432de651aff23e6f63fb3bc4005da) ## GIT FLOW --- **Source URL**: [https://guides.github.com/introduction/flow/](https://guides.github.com/introduction/flow/) <table><tr><td><strong>Browser</strong></td><td>Chrome 84.0.4147.56</td></tr><tr><td><strong>OS</strong></td><td>Windows 10 64-bit</td></tr><tr><td><strong>Screen Size</strong></td><td>2560x1080</td></tr><tr><td><strong>Viewport Size</strong></td><td>2560x888</td></tr><tr><td><strong>Pixel Ratio</strong></td><td>@1x</td></tr><tr><td><strong>Zoom Level</strong></td><td>100%</td></tr></table>
process
understanding the github flow · github guides git flow source url browser chrome os windows bit screen size viewport size pixel ratio zoom level
1
106,587
16,700,442,292
IssuesEvent
2021-06-09 01:02:01
shaimael/cset
https://api.github.com/repos/shaimael/cset
opened
CVE-2021-25949 (High) detected in set-getter-0.1.0.tgz
security vulnerability
## CVE-2021-25949 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>set-getter-0.1.0.tgz</b></p></summary> <p>Create nested getter properties and any intermediary dot notation (`'a.b.c'`) paths</p> <p>Library home page: <a href="https://registry.npmjs.org/set-getter/-/set-getter-0.1.0.tgz">https://registry.npmjs.org/set-getter/-/set-getter-0.1.0.tgz</a></p> <p>Path to dependency file: cset/CSETWebNg/package.json</p> <p>Path to vulnerable library: cset/CSETWebNg/node_modules/set-getter/package.json</p> <p> Dependency Hierarchy: - update-0.7.4.tgz (Root Library) - lazy-cache-2.0.2.tgz - :x: **set-getter-0.1.0.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Prototype pollution vulnerability in ‘set-getter’ version 0.1.0 allows an attacker to cause a denial of service and may lead to remote code execution. <p>Publish Date: 2021-01-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-25949>CVE-2021-25949</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.2</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"set-getter","packageVersion":"0.1.0","packageFilePaths":["/CSETWebNg/package.json"],"isTransitiveDependency":true,"dependencyTree":"update:0.7.4;lazy-cache:2.0.2;set-getter:0.1.0","isMinimumFixVersionAvailable":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-25949","vulnerabilityDetails":"Prototype pollution vulnerability in ‘set-getter’ version 0.1.0 allows an attacker to cause a denial of service and may lead to remote code execution.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-25949","cvss3Severity":"high","cvss3Score":"8.2","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
True
CVE-2021-25949 (High) detected in set-getter-0.1.0.tgz - ## CVE-2021-25949 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>set-getter-0.1.0.tgz</b></p></summary> <p>Create nested getter properties and any intermediary dot notation (`'a.b.c'`) paths</p> <p>Library home page: <a href="https://registry.npmjs.org/set-getter/-/set-getter-0.1.0.tgz">https://registry.npmjs.org/set-getter/-/set-getter-0.1.0.tgz</a></p> <p>Path to dependency file: cset/CSETWebNg/package.json</p> <p>Path to vulnerable library: cset/CSETWebNg/node_modules/set-getter/package.json</p> <p> Dependency Hierarchy: - update-0.7.4.tgz (Root Library) - lazy-cache-2.0.2.tgz - :x: **set-getter-0.1.0.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Prototype pollution vulnerability in ‘set-getter’ version 0.1.0 allows an attacker to cause a denial of service and may lead to remote code execution. <p>Publish Date: 2021-01-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-25949>CVE-2021-25949</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.2</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"set-getter","packageVersion":"0.1.0","packageFilePaths":["/CSETWebNg/package.json"],"isTransitiveDependency":true,"dependencyTree":"update:0.7.4;lazy-cache:2.0.2;set-getter:0.1.0","isMinimumFixVersionAvailable":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-25949","vulnerabilityDetails":"Prototype pollution vulnerability in ‘set-getter’ version 0.1.0 allows an attacker to cause a denial of service and may lead to remote code execution.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-25949","cvss3Severity":"high","cvss3Score":"8.2","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
non_process
cve high detected in set getter tgz cve high severity vulnerability vulnerable library set getter tgz create nested getter properties and any intermediary dot notation a b c paths library home page a href path to dependency file cset csetwebng package json path to vulnerable library cset csetwebng node modules set getter package json dependency hierarchy update tgz root library lazy cache tgz x set getter tgz vulnerable library found in base branch master vulnerability details prototype pollution vulnerability in ‘set getter’ version allows an attacker to cause a denial of service and may lead to remote code execution publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact high for more information on scores click a href isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree update lazy cache set getter isminimumfixversionavailable false basebranches vulnerabilityidentifier cve vulnerabilitydetails prototype pollution vulnerability in ‘set getter’ version allows an attacker to cause a denial of service and may lead to remote code execution vulnerabilityurl
0
8,671
11,802,944,557
IssuesEvent
2020-03-18 22:48:59
hashicorp/packer
https://api.github.com/repos/hashicorp/packer
closed
Packer drops preferred NIC types from Vagrant boxes
bug need-more-info post-processor/vagrant waiting-reply
Vagrant 2.2.2 now warns when some vulnerable NIC drivers are used, such as E1000. I am able to configure a more appropriate NIC at base box build time, with: ```json "vboxmanage": [ ["modifyvm", "{{.Name}}", "--nictype1", "82543GC"] ] ``` However, the Vagrant box produced by this Packer configuration completely forgets about this preference, resulting in the E1000 NIC being used when the box boots. Can we fix Packer so that NIC types are better preserved in .BOX's? Note: Vagrant also has a habit of dropping important configurations for further downstream box exports, so users will likely have to manually configure better NIC types there as well, with an `export.Vagrantfile` for use with `--vagrantfile` in `vagrant package`. But we could at least do a better job having Packer artifacts setup the right default configurations, out of the box so to speak.
1.0
Packer drops preferred NIC types from Vagrant boxes - Vagrant 2.2.2 now warns when some vulnerable NIC drivers are used, such as E1000. I am able to configure a more appropriate NIC at base box build time, with: ```json "vboxmanage": [ ["modifyvm", "{{.Name}}", "--nictype1", "82543GC"] ] ``` However, the Vagrant box produced by this Packer configuration completely forgets about this preference, resulting in the E1000 NIC being used when the box boots. Can we fix Packer so that NIC types are better preserved in .BOX's? Note: Vagrant also has a habit of dropping important configurations for further downstream box exports, so users will likely have to manually configure better NIC types there as well, with an `export.Vagrantfile` for use with `--vagrantfile` in `vagrant package`. But we could at least do a better job having Packer artifacts setup the right default configurations, out of the box so to speak.
process
packer drops preferred nic types from vagrant boxes vagrant now warns when some vulnerable nic drivers are used such as i am able to configure a more appropriate nic at base box build time with json vboxmanage however the vagrant box produced by this packer configuration completely forgets about this preference resulting in the nic being used when the box boots can we fix packer so that nic types are better preserved in box s note vagrant also has a habit of dropping important configurations for further downstream box exports so users will likely have to manually configure better nic types there as well with an export vagrantfile for use with vagrantfile in vagrant package but we could at least do a better job having packer artifacts setup the right default configurations out of the box so to speak
1
29,245
13,067,348,137
IssuesEvent
2020-07-31 00:10:25
microsoft/botframework-cli
https://api.github.com/repos/microsoft/botframework-cli
closed
update_cognitive_models.ps1 misbehave when Machine learning features are added
Bot Services R10 customer-replied-to customer-reported lu
#### What project is affected? Skill template #### What language is this in? C# #### What happens? When a Machine learning feature is added to an intent on the LUIS authoring page, the execution of update_cognitive_models.ps1 (RemoteToLocal=true) does not generate the xxxLuis.cs file. The cli tool ends with a successfull "> Update complete." message, no warning is presented. Just the log file (update_cognitive_models_log.txt) reports the error: ``` AZ CLI passes minimum version. Current version is 2.8.0 [ERROR] line 19:0 - line 19:57: Features must be defined before assigned to an intent. No definition found for feature ""Application synonyms"" in usesFeature definition for intent "Application" Invalid LUIS JSON file content. ``` #### What are the steps to reproduce this issue? 1. Create an empty Skill from the template. 2. Add a new intent on LUIS authoring for that skill. 3. Add a "Machine learning feature" to that intent. 4. Train and publish. 5. Execute update_cognitive_models.ps1 (RemoteToLocal=true) #### What were you expecting to happen? See in my project the xxxLuis.cs file updated with the new intent. #### Can you share any logs, error output, etc.? The json file is as follows (it seems a matter of order of the intent and feature definition in the json): ``` MySkill.json { "name": "myskillen-us_MySkill", "versionId": "0.1", "desc": "", "culture": "en-us", "intents": [ { "name": "Application", "features": [ { "featureName": "Application synonyms", "isRequired": false } ] }, { "name": "None", "features": [] }, { "name": "Sample", "features": [] } ], "entities": [], "closedLists": [], "composites": [], "hierarchicals": [], "patternAnyEntities": [], "regexEntities": [], "prebuiltEntities": [], "regexFeatures": [], "phraselists": [ { "activated": true, "name": "Application synonyms", "words": "Application,Web pages,Sites", "mode": true, "enabledForAllModels": false } ], "patterns": [], "utterances": [ { "text": "demo dialog", "intent": "Sample", "entities": [] }, { "text": "goodbye", "intent": "None", "entities": [] }, { "text": "hello", "intent": "None", "entities": [] }, { "text": "hi", "intent": "None", "entities": [] }, { "text": "list the applications i'm allowed to", "intent": "Application", "entities": [] }, { "text": "logout", "intent": "None", "entities": [] }, { "text": "may you list me my applications?", "intent": "Application", "entities": [] }, { "text": "run dialog", "intent": "Sample", "entities": [] }, { "text": "run sample dialog", "intent": "Sample", "entities": [] }, { "text": "sample dialog", "intent": "Sample", "entities": [] }, { "text": "test dialog", "intent": "Sample", "entities": [] }, { "text": "to what application i've access?", "intent": "Application", "entities": [] }, { "text": "what are my applications?", "intent": "Application", "entities": [] } ], "luis_schema_version": "7.0.0", "tokenizerVersion": "1.0.0", "settings": [] } ``` #### Any screenshots or additional context? No.
1.0
update_cognitive_models.ps1 misbehave when Machine learning features are added - #### What project is affected? Skill template #### What language is this in? C# #### What happens? When a Machine learning feature is added to an intent on the LUIS authoring page, the execution of update_cognitive_models.ps1 (RemoteToLocal=true) does not generate the xxxLuis.cs file. The cli tool ends with a successfull "> Update complete." message, no warning is presented. Just the log file (update_cognitive_models_log.txt) reports the error: ``` AZ CLI passes minimum version. Current version is 2.8.0 [ERROR] line 19:0 - line 19:57: Features must be defined before assigned to an intent. No definition found for feature ""Application synonyms"" in usesFeature definition for intent "Application" Invalid LUIS JSON file content. ``` #### What are the steps to reproduce this issue? 1. Create an empty Skill from the template. 2. Add a new intent on LUIS authoring for that skill. 3. Add a "Machine learning feature" to that intent. 4. Train and publish. 5. Execute update_cognitive_models.ps1 (RemoteToLocal=true) #### What were you expecting to happen? See in my project the xxxLuis.cs file updated with the new intent. #### Can you share any logs, error output, etc.? The json file is as follows (it seems a matter of order of the intent and feature definition in the json): ``` MySkill.json { "name": "myskillen-us_MySkill", "versionId": "0.1", "desc": "", "culture": "en-us", "intents": [ { "name": "Application", "features": [ { "featureName": "Application synonyms", "isRequired": false } ] }, { "name": "None", "features": [] }, { "name": "Sample", "features": [] } ], "entities": [], "closedLists": [], "composites": [], "hierarchicals": [], "patternAnyEntities": [], "regexEntities": [], "prebuiltEntities": [], "regexFeatures": [], "phraselists": [ { "activated": true, "name": "Application synonyms", "words": "Application,Web pages,Sites", "mode": true, "enabledForAllModels": false } ], "patterns": [], "utterances": [ { "text": "demo dialog", "intent": "Sample", "entities": [] }, { "text": "goodbye", "intent": "None", "entities": [] }, { "text": "hello", "intent": "None", "entities": [] }, { "text": "hi", "intent": "None", "entities": [] }, { "text": "list the applications i'm allowed to", "intent": "Application", "entities": [] }, { "text": "logout", "intent": "None", "entities": [] }, { "text": "may you list me my applications?", "intent": "Application", "entities": [] }, { "text": "run dialog", "intent": "Sample", "entities": [] }, { "text": "run sample dialog", "intent": "Sample", "entities": [] }, { "text": "sample dialog", "intent": "Sample", "entities": [] }, { "text": "test dialog", "intent": "Sample", "entities": [] }, { "text": "to what application i've access?", "intent": "Application", "entities": [] }, { "text": "what are my applications?", "intent": "Application", "entities": [] } ], "luis_schema_version": "7.0.0", "tokenizerVersion": "1.0.0", "settings": [] } ``` #### Any screenshots or additional context? No.
non_process
update cognitive models misbehave when machine learning features are added what project is affected skill template what language is this in c what happens when a machine learning feature is added to an intent on the luis authoring page the execution of update cognitive models remotetolocal true does not generate the xxxluis cs file the cli tool ends with a successfull update complete message no warning is presented just the log file update cognitive models log txt reports the error az cli passes minimum version current version is line line features must be defined before assigned to an intent no definition found for feature application synonyms in usesfeature definition for intent application invalid luis json file content what are the steps to reproduce this issue create an empty skill from the template add a new intent on luis authoring for that skill add a machine learning feature to that intent train and publish execute update cognitive models remotetolocal true what were you expecting to happen see in my project the xxxluis cs file updated with the new intent can you share any logs error output etc the json file is as follows it seems a matter of order of the intent and feature definition in the json myskill json name myskillen us myskill versionid desc culture en us intents name application features featurename application synonyms isrequired false name none features name sample features entities closedlists composites hierarchicals patternanyentities regexentities prebuiltentities regexfeatures phraselists activated true name application synonyms words application web pages sites mode true enabledforallmodels false patterns utterances text demo dialog intent sample entities text goodbye intent none entities text hello intent none entities text hi intent none entities text list the applications i m allowed to intent application entities text logout intent none entities text may you list me my applications intent application entities text run dialog intent sample entities text run sample dialog intent sample entities text sample dialog intent sample entities text test dialog intent sample entities text to what application i ve access intent application entities text what are my applications intent application entities luis schema version tokenizerversion settings any screenshots or additional context no
0
22,003
11,660,552,031
IssuesEvent
2020-03-03 03:44:05
cityofaustin/atd-data-tech
https://api.github.com/repos/cityofaustin/atd-data-tech
opened
TURP and Valet Permit Status Issue
Product: AMANDA Project: ATD AMANDA Backlog Service: Apps Type: Bug Report Workgroup: ROW migrated
Description: Fix the TURP (all subtypes) and Valet (annual and operator) processes so that there is no longer a need for a forced status change. Request Date: 2018-11-29 16:26:00 Request ID: DTS18-103010 Status: Backlog Customer Priority: 0 Level of Effort: Minor DTS URL: https://atd.knack.com/dts#service-requests/view-issue-details/5c0067afd4673a0f23888e88 *Migrated from [atd-amanda #32](https://github.com/cityofaustin/atd-amanda/issues/32)*
1.0
TURP and Valet Permit Status Issue - Description: Fix the TURP (all subtypes) and Valet (annual and operator) processes so that there is no longer a need for a forced status change. Request Date: 2018-11-29 16:26:00 Request ID: DTS18-103010 Status: Backlog Customer Priority: 0 Level of Effort: Minor DTS URL: https://atd.knack.com/dts#service-requests/view-issue-details/5c0067afd4673a0f23888e88 *Migrated from [atd-amanda #32](https://github.com/cityofaustin/atd-amanda/issues/32)*
non_process
turp and valet permit status issue description fix the turp all subtypes and valet annual and operator processes so that there is no longer a need for a forced status change request date request id status backlog customer priority level of effort minor dts url migrated from
0
9,044
12,130,108,013
IssuesEvent
2020-04-23 00:30:40
GoogleCloudPlatform/python-docs-samples
https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples
closed
remove gcp-devrel-py-tools from appengine/standard/storage/appengine-client/requirements-test.txt
priority: p2 remove-gcp-devrel-py-tools type: process
remove gcp-devrel-py-tools from appengine/standard/storage/appengine-client/requirements-test.txt
1.0
remove gcp-devrel-py-tools from appengine/standard/storage/appengine-client/requirements-test.txt - remove gcp-devrel-py-tools from appengine/standard/storage/appengine-client/requirements-test.txt
process
remove gcp devrel py tools from appengine standard storage appengine client requirements test txt remove gcp devrel py tools from appengine standard storage appengine client requirements test txt
1
8,415
6,540,697,483
IssuesEvent
2017-09-01 16:28:28
smith-chem-wisc/MetaMorpheus
https://api.github.com/repos/smith-chem-wisc/MetaMorpheus
closed
GeneratePeptideScore is slow
Performance
Removal of double counting using a HashSet is computationally demanding. Implement and profile a replacement that checks if experimental peaks have overlapping tolerances and merges them into a single peak with an increased tolerance.
True
GeneratePeptideScore is slow - Removal of double counting using a HashSet is computationally demanding. Implement and profile a replacement that checks if experimental peaks have overlapping tolerances and merges them into a single peak with an increased tolerance.
non_process
generatepeptidescore is slow removal of double counting using a hashset is computationally demanding implement and profile a replacement that checks if experimental peaks have overlapping tolerances and merges them into a single peak with an increased tolerance
0
19,353
25,483,439,195
IssuesEvent
2022-11-26 03:44:38
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
Process.GetProcessesByName() error on linux
area-System.Diagnostics.Process untriaged
### Description Chrome Browser's process name is chrome, but `Process.GetProcessesByName("chrome") ` return empty array. I need pass "chrome --enable-crashpad" to get the right result. ### Reproduction Steps var processes = Process.GetProcessesByName("chrome --enable-crashpad"); ### Expected behavior var processes = Process.GetProcessesByName("chrome"); ### Actual behavior ... ### Regression? _No response_ ### Known Workarounds _No response_ ### Configuration _No response_ ### Other information _No response_
1.0
Process.GetProcessesByName() error on linux - ### Description Chrome Browser's process name is chrome, but `Process.GetProcessesByName("chrome") ` return empty array. I need pass "chrome --enable-crashpad" to get the right result. ### Reproduction Steps var processes = Process.GetProcessesByName("chrome --enable-crashpad"); ### Expected behavior var processes = Process.GetProcessesByName("chrome"); ### Actual behavior ... ### Regression? _No response_ ### Known Workarounds _No response_ ### Configuration _No response_ ### Other information _No response_
process
process getprocessesbyname error on linux description chrome browser s process name is chrome but process getprocessesbyname chrome return empty array i need pass chrome enable crashpad to get the right result reproduction steps var processes process getprocessesbyname chrome enable crashpad expected behavior var processes process getprocessesbyname chrome actual behavior regression no response known workarounds no response configuration no response other information no response
1
37,131
9,963,062,323
IssuesEvent
2019-07-07 19:57:42
ToucanToco/vue-query-builder
https://api.github.com/repos/ToucanToco/vue-query-builder
closed
get rid of webpack
build tech
We use `rollup` to build and ship an ES6 module. We should therefore get rid of `webpack` since we don't want to keep two distinct package bundlers.
1.0
get rid of webpack - We use `rollup` to build and ship an ES6 module. We should therefore get rid of `webpack` since we don't want to keep two distinct package bundlers.
non_process
get rid of webpack we use rollup to build and ship an module we should therefore get rid of webpack since we don t want to keep two distinct package bundlers
0
169,721
26,850,312,665
IssuesEvent
2023-02-03 10:30:14
starplanter93/The_Garden_of_Musicsheet
https://api.github.com/repos/starplanter93/The_Garden_of_Musicsheet
opened
Feat: Header Organism 구현
Feat Design Test
## Description Header에 포함되는 molecule, organism 구현 ## Todo - [x] - [x]
1.0
Feat: Header Organism 구현 - ## Description Header에 포함되는 molecule, organism 구현 ## Todo - [x] - [x]
non_process
feat header organism 구현 description header에 포함되는 molecule organism 구현 todo
0
3,108
6,130,695,164
IssuesEvent
2017-06-24 08:02:59
kmycode/storycanvas-csharp
https://api.github.com/repos/kmycode/storycanvas-csharp
closed
ファイル保存・読み込みコードの全面見直し
correction priority-high processing
現行のコードでは、Streamの利用やファイル存在確認などがこんがらかっていて、読みづらいコードになっている。これでは、新しいクラウドサービスの実装がやりづらいばかりが、ファイルの保存などでバグが発生したら対応に時間がかかる。 いっそのこと、この部分のコードを全面的に見直し、見通しの良いコードにしておきたい
1.0
ファイル保存・読み込みコードの全面見直し - 現行のコードでは、Streamの利用やファイル存在確認などがこんがらかっていて、読みづらいコードになっている。これでは、新しいクラウドサービスの実装がやりづらいばかりが、ファイルの保存などでバグが発生したら対応に時間がかかる。 いっそのこと、この部分のコードを全面的に見直し、見通しの良いコードにしておきたい
process
ファイル保存・読み込みコードの全面見直し 現行のコードでは、streamの利用やファイル存在確認などがこんがらかっていて、読みづらいコードになっている。これでは、新しいクラウドサービスの実装がやりづらいばかりが、ファイルの保存などでバグが発生したら対応に時間がかかる。 いっそのこと、この部分のコードを全面的に見直し、見通しの良いコードにしておきたい
1
13,844
16,604,724,892
IssuesEvent
2021-06-02 01:23:26
ERFSEDS/Nova_Flight_Computer
https://api.github.com/repos/ERFSEDS/Nova_Flight_Computer
closed
Create device layer
Software:C/Cpp Software:Preprocessor prio:medium status:not-started type:feature
For which component is used use preprocessor definitions, this will change between boards but not while on one board
1.0
Create device layer - For which component is used use preprocessor definitions, this will change between boards but not while on one board
process
create device layer for which component is used use preprocessor definitions this will change between boards but not while on one board
1
59,287
11,956,304,409
IssuesEvent
2020-04-04 09:43:50
home-assistant/brands
https://api.github.com/repos/home-assistant/brands
closed
Yi Home Cameras is missing brand images
has-codeowner
## The problem The Yi Home Cameras integration has missing brand images. We recently started this Brands repository, to create a centralized storage of all brand-related images. These images are used on our website and the Home Assistant frontend. The following images are missing and would ideally be added: - `src/yi/logo.png` - `src/yi/icon@2x.png` - `src/yi/logo@2x.png` For image specifications and requirements, please see [README.md](https://github.com/home-assistant/brands/blob/master/README.md). ## Additional information For more information about this repository, read the [README.md](https://github.com/home-assistant/brands/blob/master/README.md) file of this repository. It contains information on how this repository works, and image specification and requirements. ## Codeowner mention Hi there, @bachya! Mind taking a look at this issue as it is with an integration (yi) you are listed as a [codeowner](https://github.com/home-assistant/core/blob/dev/homeassistant/components/yi/manifest.json) for? Thanks! Resolving this issue is not limited to codeowners! If you want to help us out, feel free to resolve this issue! Thanks already!
1.0
Yi Home Cameras is missing brand images - ## The problem The Yi Home Cameras integration has missing brand images. We recently started this Brands repository, to create a centralized storage of all brand-related images. These images are used on our website and the Home Assistant frontend. The following images are missing and would ideally be added: - `src/yi/logo.png` - `src/yi/icon@2x.png` - `src/yi/logo@2x.png` For image specifications and requirements, please see [README.md](https://github.com/home-assistant/brands/blob/master/README.md). ## Additional information For more information about this repository, read the [README.md](https://github.com/home-assistant/brands/blob/master/README.md) file of this repository. It contains information on how this repository works, and image specification and requirements. ## Codeowner mention Hi there, @bachya! Mind taking a look at this issue as it is with an integration (yi) you are listed as a [codeowner](https://github.com/home-assistant/core/blob/dev/homeassistant/components/yi/manifest.json) for? Thanks! Resolving this issue is not limited to codeowners! If you want to help us out, feel free to resolve this issue! Thanks already!
non_process
yi home cameras is missing brand images the problem the yi home cameras integration has missing brand images we recently started this brands repository to create a centralized storage of all brand related images these images are used on our website and the home assistant frontend the following images are missing and would ideally be added src yi logo png src yi icon png src yi logo png for image specifications and requirements please see additional information for more information about this repository read the file of this repository it contains information on how this repository works and image specification and requirements codeowner mention hi there bachya mind taking a look at this issue as it is with an integration yi you are listed as a for thanks resolving this issue is not limited to codeowners if you want to help us out feel free to resolve this issue thanks already
0
146,727
13,190,167,732
IssuesEvent
2020-08-13 09:43:00
riptideio/pymodbus
https://api.github.com/repos/riptideio/pymodbus
closed
Read RTU Holding Register through Serial Forwarder/TCP.
Documentation Update Required
I have Energy Meter connected through RTU and able to get holding registers data through simple RTU Code. Now i want to make Convert this RTU to TCP through Forwarder. I want to send data to TCP which forwards the command to RTU and fetches data for me. I have implement the Forwarder code just dont know how to fetch the holding register of RTU through it. **Code for Simple RTU Read** > import pymodbus > from pymodbus.pdu import ModbusRequest > from pymodbus.client.sync import ModbusSerialClient as ModbusClient > #initialize a serial RTU client instance > from pymodbus.transaction import ModbusRtuFramer > > #count= the number of registers to read > #unit= the slave unit this request is targeting > #address= the starting address to read from > > client = ModbusClient(method = 'rtu', port='/dev/ttyUSB0', baudrate= 9600) > > #Connect to the serial modbus server > connection = client.connect() > print(connection) > > #Starting add, num of reg to read, slave unit. > read = client.read_holding_registers(address = 0x01,count =2, unit=1) > data = read.registers > > print(data) > > #Closes the underlying socket connection > client.close()
1.0
Read RTU Holding Register through Serial Forwarder/TCP. - I have Energy Meter connected through RTU and able to get holding registers data through simple RTU Code. Now i want to make Convert this RTU to TCP through Forwarder. I want to send data to TCP which forwards the command to RTU and fetches data for me. I have implement the Forwarder code just dont know how to fetch the holding register of RTU through it. **Code for Simple RTU Read** > import pymodbus > from pymodbus.pdu import ModbusRequest > from pymodbus.client.sync import ModbusSerialClient as ModbusClient > #initialize a serial RTU client instance > from pymodbus.transaction import ModbusRtuFramer > > #count= the number of registers to read > #unit= the slave unit this request is targeting > #address= the starting address to read from > > client = ModbusClient(method = 'rtu', port='/dev/ttyUSB0', baudrate= 9600) > > #Connect to the serial modbus server > connection = client.connect() > print(connection) > > #Starting add, num of reg to read, slave unit. > read = client.read_holding_registers(address = 0x01,count =2, unit=1) > data = read.registers > > print(data) > > #Closes the underlying socket connection > client.close()
non_process
read rtu holding register through serial forwarder tcp i have energy meter connected through rtu and able to get holding registers data through simple rtu code now i want to make convert this rtu to tcp through forwarder i want to send data to tcp which forwards the command to rtu and fetches data for me i have implement the forwarder code just dont know how to fetch the holding register of rtu through it code for simple rtu read import pymodbus from pymodbus pdu import modbusrequest from pymodbus client sync import modbusserialclient as modbusclient initialize a serial rtu client instance from pymodbus transaction import modbusrtuframer count the number of registers to read unit the slave unit this request is targeting address the starting address to read from client modbusclient method rtu port dev baudrate connect to the serial modbus server connection client connect print connection starting add num of reg to read slave unit read client read holding registers address count unit data read registers print data closes the underlying socket connection client close
0
816,009
30,583,369,490
IssuesEvent
2023-07-21 11:29:01
Neural-Systems-at-UIO/RB-workbench
https://api.github.com/repos/Neural-Systems-at-UIO/RB-workbench
opened
WebAlign failed to load in Firefox and Chrome
High Priority
WebAlign failed to load in the RB-workbench. ![image](https://github.com/Neural-Systems-at-UIO/RB-workbench/assets/30626642/6da37a54-b4b0-4fc7-9012-1444859df504)
1.0
WebAlign failed to load in Firefox and Chrome - WebAlign failed to load in the RB-workbench. ![image](https://github.com/Neural-Systems-at-UIO/RB-workbench/assets/30626642/6da37a54-b4b0-4fc7-9012-1444859df504)
non_process
webalign failed to load in firefox and chrome webalign failed to load in the rb workbench
0
2,328
5,142,548,021
IssuesEvent
2017-01-12 13:39:39
jimbrown75/Permit-Vision-Enhancements
https://api.github.com/repos/jimbrown75/Permit-Vision-Enhancements
opened
During RA/JH, allow selection of Pre-requisite confirmaiton either once, or every issue
High Priority Should Fix Verified by PTW Process Lead
When creating a RA/JHA allow the risk assessor to select, via a radio button that all pre-requisites need to be confirmed by signature either at first issue or every issue.
1.0
During RA/JH, allow selection of Pre-requisite confirmaiton either once, or every issue - When creating a RA/JHA allow the risk assessor to select, via a radio button that all pre-requisites need to be confirmed by signature either at first issue or every issue.
process
during ra jh allow selection of pre requisite confirmaiton either once or every issue when creating a ra jha allow the risk assessor to select via a radio button that all pre requisites need to be confirmed by signature either at first issue or every issue
1
21,296
28,494,402,447
IssuesEvent
2023-04-18 13:20:41
Open-EO/openeo-api
https://api.github.com/repos/Open-EO/openeo-api
closed
minimum log_level for `GET /../logs`
data processing minor
#431 added option to specify minimum log_level at job submission time. We are currently also playing with setting a minimum log_level when requesting the logs (e.g. `GET /jobs/{job_id}/logs)`. For example see Open-EO/openeo-python-client#332, where we filter client side to only show error logs, but we've experienced that it would be better to do the filtering already back-end side (for performance reasons). Can we add an optional `log_level` parameter to `GET /jobs/{job_id}/logs` and `GET /services/{service_id}/logs` for back-end side log filtering at retrieval time? refs and related issues - #329 - #431 - Open-EO/openeo-python-client#332
1.0
minimum log_level for `GET /../logs` - #431 added option to specify minimum log_level at job submission time. We are currently also playing with setting a minimum log_level when requesting the logs (e.g. `GET /jobs/{job_id}/logs)`. For example see Open-EO/openeo-python-client#332, where we filter client side to only show error logs, but we've experienced that it would be better to do the filtering already back-end side (for performance reasons). Can we add an optional `log_level` parameter to `GET /jobs/{job_id}/logs` and `GET /services/{service_id}/logs` for back-end side log filtering at retrieval time? refs and related issues - #329 - #431 - Open-EO/openeo-python-client#332
process
minimum log level for get logs added option to specify minimum log level at job submission time we are currently also playing with setting a minimum log level when requesting the logs e g get jobs job id logs for example see open eo openeo python client where we filter client side to only show error logs but we ve experienced that it would be better to do the filtering already back end side for performance reasons can we add an optional log level parameter to get jobs job id logs and get services service id logs for back end side log filtering at retrieval time refs and related issues open eo openeo python client
1
11,913
14,700,197,933
IssuesEvent
2021-01-04 09:50:13
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Processing/GRASS tool fails to generate output
Bug Feedback Processing stale
I found something that looked like a bug. so, When I run the [i.pansharpen] command, it finally fails. The output message is as follows. [Read output layer The next layer was not generated correctly. • C: /Users/user/AppData/Local/Temp/processing_SiwLfp/2d043cc5b43542f58a317c16a7f5c1f8/redoutput.tif • C: /Users/user/AppData/Local/Temp/processing_SiwLfp/7278fd9f94164d6fbd5483d23b909f7a/greenoutput.tif • C: /Users/user/AppData/Local/Temp/processing_SiwLfp/97e48a96f6724a079668047d048ef753/blueoutput.tif Check the "Log Message Panel" in the QGIS main window to see more information about running the algorithm.] And another GRASS command same too. Calculations can be performed on QGIS316. Something is wrong with QGIS310.
1.0
Processing/GRASS tool fails to generate output - I found something that looked like a bug. so, When I run the [i.pansharpen] command, it finally fails. The output message is as follows. [Read output layer The next layer was not generated correctly. • C: /Users/user/AppData/Local/Temp/processing_SiwLfp/2d043cc5b43542f58a317c16a7f5c1f8/redoutput.tif • C: /Users/user/AppData/Local/Temp/processing_SiwLfp/7278fd9f94164d6fbd5483d23b909f7a/greenoutput.tif • C: /Users/user/AppData/Local/Temp/processing_SiwLfp/97e48a96f6724a079668047d048ef753/blueoutput.tif Check the "Log Message Panel" in the QGIS main window to see more information about running the algorithm.] And another GRASS command same too. Calculations can be performed on QGIS316. Something is wrong with QGIS310.
process
processing grass tool fails to generate output i found something that looked like a bug so when i run the command it finally fails the output message is as follows read output layer the next layer was not generated correctly • c users user appdata local temp processing siwlfp redoutput tif • c users user appdata local temp processing siwlfp greenoutput tif • c users user appdata local temp processing siwlfp blueoutput tif check the log message panel in the qgis main window to see more information about running the algorithm and another grass command same too calculations can be performed on something is wrong with
1
9,439
12,425,101,115
IssuesEvent
2020-05-24 14:50:51
raxod502/straight.el
https://api.github.com/repos/raxod502/straight.el
opened
Shallow clones not supported properly in VC commands
bug git interactive lockfiles process buffer shallow clone vc
When I run `M-x straight-freeze-versions`, for example, I get a "Git failed" error if one of my repos is a shallow clone. The VC commands need a complete overhaul at some point.
1.0
Shallow clones not supported properly in VC commands - When I run `M-x straight-freeze-versions`, for example, I get a "Git failed" error if one of my repos is a shallow clone. The VC commands need a complete overhaul at some point.
process
shallow clones not supported properly in vc commands when i run m x straight freeze versions for example i get a git failed error if one of my repos is a shallow clone the vc commands need a complete overhaul at some point
1
5,801
8,641,448,974
IssuesEvent
2018-11-24 17:52:26
carloseduardov8/Viajato
https://api.github.com/repos/carloseduardov8/Viajato
closed
Implementar cancelamento de seleção
Priority:Normal Process:Implement Requirement
Usuário deve poder visualizar escolhas já feitas e cancelar aquelas que não deseja mais.
1.0
Implementar cancelamento de seleção - Usuário deve poder visualizar escolhas já feitas e cancelar aquelas que não deseja mais.
process
implementar cancelamento de seleção usuário deve poder visualizar escolhas já feitas e cancelar aquelas que não deseja mais
1
222,070
17,391,459,265
IssuesEvent
2021-08-02 07:58:15
xamarin/xamarin-macios
https://api.github.com/repos/xamarin/xamarin-macios
opened
Test builds sometimes fail with: Cannot write to the output file "[..]". The process cannot access the file '[...]' because it is being used by another process.
bug tests
We try to build tests in parallel whenever we can, but in certain cases we end up building the same project simultaneously in different processes, and this can cause errors like this: > /Users/builder/azdo/_work/1/s/xamarin-macios/builds/downloads/dotnet-sdk-6.0.100-rc.1.21381.5-osx-x64/sdk/6.0.100-rc.1.21381.5/Microsoft.Common.CurrentVersion.targets(3262,5): error MSB3554: Cannot write to the output file "/Users/builder/azdo/_work/1/s/xamarin-macios/tests/EmbeddedResources/dotnet/macOS/obj/Debug/net6.0-macos/EmbeddedResources.Welcome.resources". The process cannot access the file '/Users/builder/azdo/_work/1/s/xamarin-macios/tests/EmbeddedResources/dotnet/macOS/obj/Debug/net6.0-macos/EmbeddedResources.Welcome.resources' because it is being used by another process. [vsdrops_index-105.html.txt](https://github.com/xamarin/xamarin-macios/files/6915293/vsdrops_index-105.html.txt)
1.0
Test builds sometimes fail with: Cannot write to the output file "[..]". The process cannot access the file '[...]' because it is being used by another process. - We try to build tests in parallel whenever we can, but in certain cases we end up building the same project simultaneously in different processes, and this can cause errors like this: > /Users/builder/azdo/_work/1/s/xamarin-macios/builds/downloads/dotnet-sdk-6.0.100-rc.1.21381.5-osx-x64/sdk/6.0.100-rc.1.21381.5/Microsoft.Common.CurrentVersion.targets(3262,5): error MSB3554: Cannot write to the output file "/Users/builder/azdo/_work/1/s/xamarin-macios/tests/EmbeddedResources/dotnet/macOS/obj/Debug/net6.0-macos/EmbeddedResources.Welcome.resources". The process cannot access the file '/Users/builder/azdo/_work/1/s/xamarin-macios/tests/EmbeddedResources/dotnet/macOS/obj/Debug/net6.0-macos/EmbeddedResources.Welcome.resources' because it is being used by another process. [vsdrops_index-105.html.txt](https://github.com/xamarin/xamarin-macios/files/6915293/vsdrops_index-105.html.txt)
non_process
test builds sometimes fail with cannot write to the output file the process cannot access the file because it is being used by another process we try to build tests in parallel whenever we can but in certain cases we end up building the same project simultaneously in different processes and this can cause errors like this users builder azdo work s xamarin macios builds downloads dotnet sdk rc osx sdk rc microsoft common currentversion targets error cannot write to the output file users builder azdo work s xamarin macios tests embeddedresources dotnet macos obj debug macos embeddedresources welcome resources the process cannot access the file users builder azdo work s xamarin macios tests embeddedresources dotnet macos obj debug macos embeddedresources welcome resources because it is being used by another process
0
3,505
6,559,511,665
IssuesEvent
2017-09-07 04:31:20
amaster507/ifbmt
https://api.github.com/repos/amaster507/ifbmt
closed
Mass emailer
email enhancement GUI idea process wontfix
What if there was a way to refine a search list of churches and then send an email to all if then at once? This may need to be reviewed or limited in use to not cause spam. We don't want this to end up being something that every preacher/church ignores because it is overused.
1.0
Mass emailer - What if there was a way to refine a search list of churches and then send an email to all if then at once? This may need to be reviewed or limited in use to not cause spam. We don't want this to end up being something that every preacher/church ignores because it is overused.
process
mass emailer what if there was a way to refine a search list of churches and then send an email to all if then at once this may need to be reviewed or limited in use to not cause spam we don t want this to end up being something that every preacher church ignores because it is overused
1
50,435
10,510,304,422
IssuesEvent
2019-09-27 13:09:48
Samsung/TAU-Design-Editor
https://api.github.com/repos/Samsung/TAU-Design-Editor
closed
[VSCode] [Wearable] Image widget has strange image src set by default
1.0 bug in review vscode wearable
## Technical [VSCode] [Wearable] Image widget has strange image src set by default ## Steps to reproduce [W Sample TAU 1.1.4.zip](https://github.com/Samsung/TAU-Design-Editor/files/3590275/W.Sample.TAU.1.1.4.zip) 1. open attached sample project in DE 2. D&D image widget 3. Go to Attributes->Image-->SRC ## Expected behaviour Nothing is set in src ## Actual behaviour Strange "# is set as a source ![image](https://user-images.githubusercontent.com/2246331/64528228-3beb8280-d308-11e9-9602-51c39855f1d8.png)
1.0
[VSCode] [Wearable] Image widget has strange image src set by default - ## Technical [VSCode] [Wearable] Image widget has strange image src set by default ## Steps to reproduce [W Sample TAU 1.1.4.zip](https://github.com/Samsung/TAU-Design-Editor/files/3590275/W.Sample.TAU.1.1.4.zip) 1. open attached sample project in DE 2. D&D image widget 3. Go to Attributes->Image-->SRC ## Expected behaviour Nothing is set in src ## Actual behaviour Strange "# is set as a source ![image](https://user-images.githubusercontent.com/2246331/64528228-3beb8280-d308-11e9-9602-51c39855f1d8.png)
non_process
image widget has strange image src set by default technical image widget has strange image src set by default steps to reproduce open attached sample project in de d d image widget go to attributes image src expected behaviour nothing is set in src actual behaviour strange is set as a source
0
20,738
27,438,052,950
IssuesEvent
2023-03-02 09:03:37
camunda/issues
https://api.github.com/repos/camunda/issues
opened
BPMN Escalation Events
component:desktopModeler component:operate component:optimize component:webModeler component:zeebe-process-automation public feature-parity version:8.2-alpha4
> This is an epic internal-docs issue. It bundles all activities we conduct around a certain initiative. It will typically links to various child issues from various repositories and can be spread across multiple teams. ### Value Proposition Statement Communicate from a subprocess to an upper process using BPMN Escalation Events ### User Problem Escalation events are events which reference a named escalation. They are mostly used to communicate from a subprocess to an upper process. Unlike an error, an escalation event is non critical and execution continues at the location of throwing. Currently this BPMN symbols is not supported by Camunda Platform 8. ### User Stories **Modeler** As a developer I can use Escalation Events in the Modeler. **Zeebe** As a developer I can deploy Esalation Events to Zeebe. As a developer I can trust that Escalation Events are executed automatically by Zeebe according to BPMN spec. **Operate / Optimize** As a developer I can see Escalation Events in Operate and Optimize correctly and the import is not broken. The technical properties of Escalation should be most likely similar to Camunda 7. ### Implementation Notes <!-- Notes to consider for implementation, for example: * In Cawemo we already have the capability to manage templates via the feature that we call “catalog” * What we would build now is the ability to a) use this feature in the web modeler to create templates and b) when the context pad opens for defining the type of a task, the templates that decorate service tasks are shown * We should clarify terminology (integrations vs. connectors vs. job workers vs. element templates.) Particularly “element templates” might not be a term that a user intuitively understands. * See these high level wireframes to capture the idea --> ### Breakdown **Zeebe** - https://github.com/camunda/zeebe/issues/10682 **Modeler** - [x] https://github.com/camunda/camunda-modeler/issues/3318 **Documentation** - https://github.com/camunda/camunda-platform-docs/issues/1531 #### Discovery phase ## <!-- Example: link to "Conduct customer interview with xyz" --> #### Define phase ## <!-- Consider: UI, UX, technical design, documentation design --> <!-- Example: link to "Define User-Journey Flow" or "Define target architecture" --> Design Planning * Reviewed by design: {date} * Designer assigned: {Yes, No Design Necessary, or No Designer Available} * Assignee: * Design Brief - {link to design brief } * Research Brief - {link to research brief } Design Deliverables * {Deliverable Name} {Link to GH Issue} #### Implement phase ## **Optimize** - [ ] https://jira.camunda.com/browse/OPT-6612 **Modeler** - [x] https://github.com/camunda/camunda-modeler/issues/3318 **Operate** - [x] https://github.com/camunda/operate/issues/3757 #### Validate phase ## <!-- Example: link to "Evaluate usage data of last quarter" --> ### Links to additional collateral
1.0
BPMN Escalation Events - > This is an epic internal-docs issue. It bundles all activities we conduct around a certain initiative. It will typically links to various child issues from various repositories and can be spread across multiple teams. ### Value Proposition Statement Communicate from a subprocess to an upper process using BPMN Escalation Events ### User Problem Escalation events are events which reference a named escalation. They are mostly used to communicate from a subprocess to an upper process. Unlike an error, an escalation event is non critical and execution continues at the location of throwing. Currently this BPMN symbols is not supported by Camunda Platform 8. ### User Stories **Modeler** As a developer I can use Escalation Events in the Modeler. **Zeebe** As a developer I can deploy Esalation Events to Zeebe. As a developer I can trust that Escalation Events are executed automatically by Zeebe according to BPMN spec. **Operate / Optimize** As a developer I can see Escalation Events in Operate and Optimize correctly and the import is not broken. The technical properties of Escalation should be most likely similar to Camunda 7. ### Implementation Notes <!-- Notes to consider for implementation, for example: * In Cawemo we already have the capability to manage templates via the feature that we call “catalog” * What we would build now is the ability to a) use this feature in the web modeler to create templates and b) when the context pad opens for defining the type of a task, the templates that decorate service tasks are shown * We should clarify terminology (integrations vs. connectors vs. job workers vs. element templates.) Particularly “element templates” might not be a term that a user intuitively understands. * See these high level wireframes to capture the idea --> ### Breakdown **Zeebe** - https://github.com/camunda/zeebe/issues/10682 **Modeler** - [x] https://github.com/camunda/camunda-modeler/issues/3318 **Documentation** - https://github.com/camunda/camunda-platform-docs/issues/1531 #### Discovery phase ## <!-- Example: link to "Conduct customer interview with xyz" --> #### Define phase ## <!-- Consider: UI, UX, technical design, documentation design --> <!-- Example: link to "Define User-Journey Flow" or "Define target architecture" --> Design Planning * Reviewed by design: {date} * Designer assigned: {Yes, No Design Necessary, or No Designer Available} * Assignee: * Design Brief - {link to design brief } * Research Brief - {link to research brief } Design Deliverables * {Deliverable Name} {Link to GH Issue} #### Implement phase ## **Optimize** - [ ] https://jira.camunda.com/browse/OPT-6612 **Modeler** - [x] https://github.com/camunda/camunda-modeler/issues/3318 **Operate** - [x] https://github.com/camunda/operate/issues/3757 #### Validate phase ## <!-- Example: link to "Evaluate usage data of last quarter" --> ### Links to additional collateral
process
bpmn escalation events this is an epic internal docs issue it bundles all activities we conduct around a certain initiative it will typically links to various child issues from various repositories and can be spread across multiple teams value proposition statement communicate from a subprocess to an upper process using bpmn escalation events user problem escalation events are events which reference a named escalation they are mostly used to communicate from a subprocess to an upper process unlike an error an escalation event is non critical and execution continues at the location of throwing currently this bpmn symbols is not supported by camunda platform user stories modeler as a developer i can use escalation events in the modeler zeebe as a developer i can deploy esalation events to zeebe as a developer i can trust that escalation events are executed automatically by zeebe according to bpmn spec operate optimize as a developer i can see escalation events in operate and optimize correctly and the import is not broken the technical properties of escalation should be most likely similar to camunda implementation notes notes to consider for implementation for example in cawemo we already have the capability to manage templates via the feature that we call “catalog” what we would build now is the ability to a use this feature in the web modeler to create templates and b when the context pad opens for defining the type of a task the templates that decorate service tasks are shown we should clarify terminology integrations vs connectors vs job workers vs element templates particularly “element templates” might not be a term that a user intuitively understands see these high level wireframes to capture the idea breakdown zeebe modeler documentation discovery phase define phase design planning reviewed by design date designer assigned yes no design necessary or no designer available assignee design brief link to design brief research brief link to research brief design deliverables deliverable name link to gh issue implement phase optimize modeler operate validate phase links to additional collateral
1
12,002
3,561,583,960
IssuesEvent
2016-01-23 22:06:13
empowerhack/hub
https://api.github.com/repos/empowerhack/hub
closed
Create Spring boot project
type: documentation type: java
* [x] Create a Vanilla Spring boot project * [x] Documentation on how to setup project locally * [x] Add Contribution file to repo include: * [x] include git flow usage & link to official docs * [x] pull request, labels, template etc
1.0
Create Spring boot project - * [x] Create a Vanilla Spring boot project * [x] Documentation on how to setup project locally * [x] Add Contribution file to repo include: * [x] include git flow usage & link to official docs * [x] pull request, labels, template etc
non_process
create spring boot project create a vanilla spring boot project documentation on how to setup project locally add contribution file to repo include include git flow usage link to official docs pull request labels template etc
0
22,709
32,036,924,273
IssuesEvent
2023-09-22 16:01:32
GoogleCloudPlatform/alloydb-python-connector
https://api.github.com/repos/GoogleCloudPlatform/alloydb-python-connector
closed
Add test demonstating Auto IAM AuthN on direct path
priority: p1 type: docs type: process
### Description Show how to leverage asyncpg and sqlalchemy to refresh an OAuth2 token on connect ### Potential Solution _No response_ ### Additional Details _No response_
1.0
Add test demonstating Auto IAM AuthN on direct path - ### Description Show how to leverage asyncpg and sqlalchemy to refresh an OAuth2 token on connect ### Potential Solution _No response_ ### Additional Details _No response_
process
add test demonstating auto iam authn on direct path description show how to leverage asyncpg and sqlalchemy to refresh an token on connect potential solution no response additional details no response
1
14,129
17,023,119,883
IssuesEvent
2021-07-03 00:27:50
fweikert/bugs
https://api.github.com/repos/fweikert/bugs
opened
Status of Bazel 5.0.0-pre.20210623.2
P1 release team-XProduct type: process
- Expected release date: 2021-07-03 Task list: - [ ] Pick release baseline: 8b453331163378071f1cfe0ae7c74d551c21b834 with cherrypick [223113c9](https://github.com/bazelbuild/bazel/commit/223113c9202e8f338b183d1736d97327d28241ea) - [ ] Create release candidate: https://releases.bazel.build/5.0.0/rolling/5.0.0-pre.20210623.2rc1/index.html - [ ] Post-submit: https://buildkite.com/bazel/bazel-bazel - [ ] Push the release: https://releases.bazel.build/5.0.0/rolling/5.0.0-pre.20210623.2/index.html - [ ] Update the [release page](https://github.com/bazelbuild/bazel/releases/)
1.0
Status of Bazel 5.0.0-pre.20210623.2 - - Expected release date: 2021-07-03 Task list: - [ ] Pick release baseline: 8b453331163378071f1cfe0ae7c74d551c21b834 with cherrypick [223113c9](https://github.com/bazelbuild/bazel/commit/223113c9202e8f338b183d1736d97327d28241ea) - [ ] Create release candidate: https://releases.bazel.build/5.0.0/rolling/5.0.0-pre.20210623.2rc1/index.html - [ ] Post-submit: https://buildkite.com/bazel/bazel-bazel - [ ] Push the release: https://releases.bazel.build/5.0.0/rolling/5.0.0-pre.20210623.2/index.html - [ ] Update the [release page](https://github.com/bazelbuild/bazel/releases/)
process
status of bazel pre expected release date task list pick release baseline with cherrypick create release candidate post submit push the release update the
1
648,971
21,214,613,183
IssuesEvent
2022-04-11 05:40:33
AY2122S2-CS2103T-W15-2/tp
https://api.github.com/repos/AY2122S2-CS2103T-W15-2/tp
closed
Add tests for Salary class
priority.High type.Task
Tests for the `salary` class are not implemented. Add tests to automate testing and minimise regression.
1.0
Add tests for Salary class - Tests for the `salary` class are not implemented. Add tests to automate testing and minimise regression.
non_process
add tests for salary class tests for the salary class are not implemented add tests to automate testing and minimise regression
0
285,468
24,668,603,186
IssuesEvent
2022-10-18 12:15:32
Kimthien94/Feedback-Online
https://api.github.com/repos/Kimthien94/Feedback-Online
opened
[BugID_45]_FUNC_Xóa topic_Không có thông báo xóa topic thành công
bug Fun_Feature Missing comestic low Intergration test open
Precondition: - Admin đang ở màn Quản lý Topic - Topic có Tên Topic = 'TopicA' không được sử dụng ở module khác (ex: Gán topic) Steps: 1. Click [Xóa] button tại record 'TopicA' 2. Click [Có] button Thực tế: Không hiển thị thông báo Mong đợi: Hiển thị inline message màu xanh “Xóa Topic thành công” ở góc bên trái của page Testcase ID = 20
1.0
[BugID_45]_FUNC_Xóa topic_Không có thông báo xóa topic thành công - Precondition: - Admin đang ở màn Quản lý Topic - Topic có Tên Topic = 'TopicA' không được sử dụng ở module khác (ex: Gán topic) Steps: 1. Click [Xóa] button tại record 'TopicA' 2. Click [Có] button Thực tế: Không hiển thị thông báo Mong đợi: Hiển thị inline message màu xanh “Xóa Topic thành công” ở góc bên trái của page Testcase ID = 20
non_process
func xóa topic không có thông báo xóa topic thành công precondition admin đang ở màn quản lý topic topic có tên topic topica không được sử dụng ở module khác ex gán topic steps click button tại record topica click button thực tế không hiển thị thông báo mong đợi hiển thị inline message màu xanh “xóa topic thành công” ở góc bên trái của page testcase id
0
228,669
25,239,602,904
IssuesEvent
2022-11-15 05:55:25
Satheesh575555/linux-3.0.35
https://api.github.com/repos/Satheesh575555/linux-3.0.35
opened
CVE-2016-4470 (Medium) detected in linuxlinux-3.0.40
security vulnerability
## CVE-2016-4470 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-3.0.40</b></p></summary> <p> <p>Apache Software Foundation (ASF)</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v3.0/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v3.0/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/Satheesh575555/linux-3.0.35/commit/d886d5c33aadc1c4f116214d0060f5869b445fe1">d886d5c33aadc1c4f116214d0060f5869b445fe1</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/security/keys/key.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/security/keys/key.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/security/keys/key.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The key_reject_and_link function in security/keys/key.c in the Linux kernel through 4.6.3 does not ensure that a certain data structure is initialized, which allows local users to cause a denial of service (system crash) via vectors involving a crafted keyctl request2 command. <p>Publish Date: 2016-06-27 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2016-4470>CVE-2016-4470</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.2</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2016-4470">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2016-4470</a></p> <p>Release Date: 2016-06-27</p> <p>Fix Resolution: v4.7-rc4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2016-4470 (Medium) detected in linuxlinux-3.0.40 - ## CVE-2016-4470 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-3.0.40</b></p></summary> <p> <p>Apache Software Foundation (ASF)</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v3.0/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v3.0/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/Satheesh575555/linux-3.0.35/commit/d886d5c33aadc1c4f116214d0060f5869b445fe1">d886d5c33aadc1c4f116214d0060f5869b445fe1</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/security/keys/key.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/security/keys/key.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/security/keys/key.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The key_reject_and_link function in security/keys/key.c in the Linux kernel through 4.6.3 does not ensure that a certain data structure is initialized, which allows local users to cause a denial of service (system crash) via vectors involving a crafted keyctl request2 command. <p>Publish Date: 2016-06-27 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2016-4470>CVE-2016-4470</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.2</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2016-4470">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2016-4470</a></p> <p>Release Date: 2016-06-27</p> <p>Fix Resolution: v4.7-rc4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in linuxlinux cve medium severity vulnerability vulnerable library linuxlinux apache software foundation asf library home page a href found in head commit a href found in base branch master vulnerable source files security keys key c security keys key c security keys key c vulnerability details the key reject and link function in security keys key c in the linux kernel through does not ensure that a certain data structure is initialized which allows local users to cause a denial of service system crash via vectors involving a crafted keyctl command publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
11,281
2,648,796,649
IssuesEvent
2015-03-14 08:15:35
STEllAR-GROUP/hpx
https://api.github.com/repos/STEllAR-GROUP/hpx
opened
Thread performance counters (pending queue counts) totals disagree with worker threads
affecting NMSU category: performance counters difficulty: easy type: defect
Command: bin/1d_stencil_4 --nx=1000 --np=1000 -t2 \ --hpx:print-counter /threads{locality#*/total}/<thread counter> \ --hpx:print-counter /threads{locality#*/worker-thread#*}/<thread counter> Counters where individual worker thread counts disagree with total include: /count/pending-accesses /count/pending-misses /count/stolen-to-pending /count/stolen-from-pending Examples: /count/pending-accesses: OS_Threads,Execution_Time_sec,Points_per_Partition,Partitions,Time_Steps 2, 0.14836133, 1000, 1000, 45 /threads{locality#0/total}/count/pending-accesses,1,0.157037,[s],186465 /threads{locality#0/worker-thread#0}/count/pending-accesses,1,0.157106,[s],14337 /threads{locality#0/worker-thread#1}/count/pending-accesses,1,0.157077,[s],61692 /count/pending-misses: OS_Threads,Execution_Time_sec,Points_per_Partition,Partitions,Time_Steps 2, 0.150560948, 1000, 1000, 45 /threads{locality#0/total}/count/pending-misses,1,0.159072,[s],127346 /threads{locality#0/worker-thread#0}/count/pending-misses,1,0.159071,[s],23505 /threads{locality#0/worker-thread#1}/count/pending-misses,1,0.159115,[s],39951 /count/pending-accesses: OS_Threads,Execution_Time_sec,Points_per_Partition,Partitions,Time_Steps 2, 0.14886279, 1000, 1000, 45 /threads{locality#0/total}/count/pending-accesses,1,0.157788,[s],179778 /threads{locality#0/worker-thread#0}/count/pending-accesses,1,0.157897,[s],53157 /threads{locality#0/worker-thread#1}/count/pending-accesses,1,0.157833,[s],23033 /count/pending-misses: OS_Threads,Execution_Time_sec,Points_per_Partition,Partitions,Time_Steps 2, 0.150670529, 1000, 1000, 45 /threads{locality#0/total}/count/pending-misses,1,0.160003,[s],122847 /threads{locality#0/worker-thread#0}/count/pending-misses,1,0.160061,[s],52292 /threads{locality#0/worker-thread#1}/count/pending-misses,1,0.160091,[s],8937 /count/stolen-to-pending: OS_Threads,Execution_Time_sec,Points_per_Partition,Partitions,Time_Steps 2, 0.147684702, 1000, 1000, 45 /threads{locality#0/total}/count/stolen-to-pending,1,0.156292,[s],10580 /threads{locality#0/worker-thread#0}/count/stolen-to-pending,1,0.156333,[s],13 /threads{locality#0/worker-thread#1}/count/stolen-to-pending,1,0.156282,[s],29 /count/stolen-from-pending: OS_Threads,Execution_Time_sec,Points_per_Partition,Partitions,Time_Steps 2, 0.150269219, 1000, 1000, 45 /threads{locality#0/total}/count/stolen-from-pending,1,0.159178,[s],9343 /threads{locality#0/worker-thread#0}/count/stolen-from-pending,1,0.159268,[s],20 /threads{locality#0/worker-thread#1}/count/stolen-from-pending,1,0.159230,[s],50
1.0
Thread performance counters (pending queue counts) totals disagree with worker threads - Command: bin/1d_stencil_4 --nx=1000 --np=1000 -t2 \ --hpx:print-counter /threads{locality#*/total}/<thread counter> \ --hpx:print-counter /threads{locality#*/worker-thread#*}/<thread counter> Counters where individual worker thread counts disagree with total include: /count/pending-accesses /count/pending-misses /count/stolen-to-pending /count/stolen-from-pending Examples: /count/pending-accesses: OS_Threads,Execution_Time_sec,Points_per_Partition,Partitions,Time_Steps 2, 0.14836133, 1000, 1000, 45 /threads{locality#0/total}/count/pending-accesses,1,0.157037,[s],186465 /threads{locality#0/worker-thread#0}/count/pending-accesses,1,0.157106,[s],14337 /threads{locality#0/worker-thread#1}/count/pending-accesses,1,0.157077,[s],61692 /count/pending-misses: OS_Threads,Execution_Time_sec,Points_per_Partition,Partitions,Time_Steps 2, 0.150560948, 1000, 1000, 45 /threads{locality#0/total}/count/pending-misses,1,0.159072,[s],127346 /threads{locality#0/worker-thread#0}/count/pending-misses,1,0.159071,[s],23505 /threads{locality#0/worker-thread#1}/count/pending-misses,1,0.159115,[s],39951 /count/pending-accesses: OS_Threads,Execution_Time_sec,Points_per_Partition,Partitions,Time_Steps 2, 0.14886279, 1000, 1000, 45 /threads{locality#0/total}/count/pending-accesses,1,0.157788,[s],179778 /threads{locality#0/worker-thread#0}/count/pending-accesses,1,0.157897,[s],53157 /threads{locality#0/worker-thread#1}/count/pending-accesses,1,0.157833,[s],23033 /count/pending-misses: OS_Threads,Execution_Time_sec,Points_per_Partition,Partitions,Time_Steps 2, 0.150670529, 1000, 1000, 45 /threads{locality#0/total}/count/pending-misses,1,0.160003,[s],122847 /threads{locality#0/worker-thread#0}/count/pending-misses,1,0.160061,[s],52292 /threads{locality#0/worker-thread#1}/count/pending-misses,1,0.160091,[s],8937 /count/stolen-to-pending: OS_Threads,Execution_Time_sec,Points_per_Partition,Partitions,Time_Steps 2, 0.147684702, 1000, 1000, 45 /threads{locality#0/total}/count/stolen-to-pending,1,0.156292,[s],10580 /threads{locality#0/worker-thread#0}/count/stolen-to-pending,1,0.156333,[s],13 /threads{locality#0/worker-thread#1}/count/stolen-to-pending,1,0.156282,[s],29 /count/stolen-from-pending: OS_Threads,Execution_Time_sec,Points_per_Partition,Partitions,Time_Steps 2, 0.150269219, 1000, 1000, 45 /threads{locality#0/total}/count/stolen-from-pending,1,0.159178,[s],9343 /threads{locality#0/worker-thread#0}/count/stolen-from-pending,1,0.159268,[s],20 /threads{locality#0/worker-thread#1}/count/stolen-from-pending,1,0.159230,[s],50
non_process
thread performance counters pending queue counts totals disagree with worker threads command bin stencil nx np hpx print counter threads locality total hpx print counter threads locality worker thread counters where individual worker thread counts disagree with total include count pending accesses count pending misses count stolen to pending count stolen from pending examples count pending accesses os threads execution time sec points per partition partitions time steps threads locality total count pending accesses threads locality worker thread count pending accesses threads locality worker thread count pending accesses count pending misses os threads execution time sec points per partition partitions time steps threads locality total count pending misses threads locality worker thread count pending misses threads locality worker thread count pending misses count pending accesses os threads execution time sec points per partition partitions time steps threads locality total count pending accesses threads locality worker thread count pending accesses threads locality worker thread count pending accesses count pending misses os threads execution time sec points per partition partitions time steps threads locality total count pending misses threads locality worker thread count pending misses threads locality worker thread count pending misses count stolen to pending os threads execution time sec points per partition partitions time steps threads locality total count stolen to pending threads locality worker thread count stolen to pending threads locality worker thread count stolen to pending count stolen from pending os threads execution time sec points per partition partitions time steps threads locality total count stolen from pending threads locality worker thread count stolen from pending threads locality worker thread count stolen from pending
0
317,897
27,273,527,664
IssuesEvent
2023-02-23 01:32:33
MPMG-DCC-UFMG/F01
https://api.github.com/repos/MPMG-DCC-UFMG/F01
closed
Teste de generalizacao para a tag Contratos - Dados dos Contratos - Rio Acima
generalization test development
DoD: Realizar o teste de Generalização do validador da tag Contratos - Dados dos Contratos para o Município de Rio Acima.
1.0
Teste de generalizacao para a tag Contratos - Dados dos Contratos - Rio Acima - DoD: Realizar o teste de Generalização do validador da tag Contratos - Dados dos Contratos para o Município de Rio Acima.
non_process
teste de generalizacao para a tag contratos dados dos contratos rio acima dod realizar o teste de generalização do validador da tag contratos dados dos contratos para o município de rio acima
0
24,567
7,526,599,311
IssuesEvent
2018-04-13 14:29:17
flyve-mdm/web-mdm-dashboard
https://api.github.com/repos/flyve-mdm/web-mdm-dashboard
opened
Build for translations
bug build
Hi, guys. I already translated and reviewed all strings in the Transifex (Portuguese, French, Spanish). Could you take a look what happen with the build? ### Observed Results ![image](https://user-images.githubusercontent.com/663460/38740325-94f370b0-3f37-11e8-8c98-e73dd3027f25.png) ### Expected behavior:
1.0
Build for translations - Hi, guys. I already translated and reviewed all strings in the Transifex (Portuguese, French, Spanish). Could you take a look what happen with the build? ### Observed Results ![image](https://user-images.githubusercontent.com/663460/38740325-94f370b0-3f37-11e8-8c98-e73dd3027f25.png) ### Expected behavior:
non_process
build for translations hi guys i already translated and reviewed all strings in the transifex portuguese french spanish could you take a look what happen with the build observed results expected behavior
0
41,431
8,971,856,255
IssuesEvent
2019-01-29 16:49:36
alechamed/decideEuropa-censo
https://api.github.com/repos/alechamed/decideEuropa-censo
opened
Automatización de despliegue en heroku
EST_NEW PRI_HIGH TEM_CODE TYP_ENHACEMENT
Voy a realizar la automatización para que cuando se haga un push en master se suba una copia a heroku automáticamente desde travis, ya que actualmente para que se despliegue hay que subirlo manualmente a heroku.
1.0
Automatización de despliegue en heroku - Voy a realizar la automatización para que cuando se haga un push en master se suba una copia a heroku automáticamente desde travis, ya que actualmente para que se despliegue hay que subirlo manualmente a heroku.
non_process
automatización de despliegue en heroku voy a realizar la automatización para que cuando se haga un push en master se suba una copia a heroku automáticamente desde travis ya que actualmente para que se despliegue hay que subirlo manualmente a heroku
0
4,700
7,542,642,519
IssuesEvent
2018-04-17 13:34:22
inviwo/inviwo
https://api.github.com/repos/inviwo/inviwo
closed
Composite Processor creation upon deserialization issue
P: Composite Processor T: Bug
It seems to be an issue when trying to deserialize a network with with a Composite Processor in it. Notice it when trying to load a network with a composite processor and got deserialization errors. This also seem to affect copy pasting of Composite Processors. **Steps to reproduce** 1. Load the boron network 2. Select any set of processor and create a composite out of them (I selected the proxy+ entryexit + raycaster) 3. Copy and paste the create composite processor Result: ``` Could not create Processor: "composite" of class "org.inviwo.CompositeProcessor" at line: 1. Reason: "org.inviwo.CompositeProcessor" Not found in factory. ```
1.0
Composite Processor creation upon deserialization issue - It seems to be an issue when trying to deserialize a network with with a Composite Processor in it. Notice it when trying to load a network with a composite processor and got deserialization errors. This also seem to affect copy pasting of Composite Processors. **Steps to reproduce** 1. Load the boron network 2. Select any set of processor and create a composite out of them (I selected the proxy+ entryexit + raycaster) 3. Copy and paste the create composite processor Result: ``` Could not create Processor: "composite" of class "org.inviwo.CompositeProcessor" at line: 1. Reason: "org.inviwo.CompositeProcessor" Not found in factory. ```
process
composite processor creation upon deserialization issue it seems to be an issue when trying to deserialize a network with with a composite processor in it notice it when trying to load a network with a composite processor and got deserialization errors this also seem to affect copy pasting of composite processors steps to reproduce load the boron network select any set of processor and create a composite out of them i selected the proxy entryexit raycaster copy and paste the create composite processor result could not create processor composite of class org inviwo compositeprocessor at line reason org inviwo compositeprocessor not found in factory
1
656,873
21,778,732,573
IssuesEvent
2022-05-13 16:17:57
proselog/proselog
https://api.github.com/repos/proselog/proselog
closed
Allow to become a subscriber
priority
- [x] add subscriber to our database - [ ] add subscriber to mailgun mail list
1.0
Allow to become a subscriber - - [x] add subscriber to our database - [ ] add subscriber to mailgun mail list
non_process
allow to become a subscriber add subscriber to our database add subscriber to mailgun mail list
0
64,857
12,500,425,079
IssuesEvent
2020-06-01 22:15:03
mjseok/OSS_team1
https://api.github.com/repos/mjseok/OSS_team1
closed
Functionalize_playGame
clean code
코드를 함수화하여 main을 최대한 간단하게 작성하고 라벨을 지우고 goto 구문도 없애기 게임실행하는 부분을 없애기( menu=1)선택했을 때의 부분
1.0
Functionalize_playGame - 코드를 함수화하여 main을 최대한 간단하게 작성하고 라벨을 지우고 goto 구문도 없애기 게임실행하는 부분을 없애기( menu=1)선택했을 때의 부분
non_process
functionalize playgame 코드를 함수화하여 main을 최대한 간단하게 작성하고 라벨을 지우고 goto 구문도 없애기 게임실행하는 부분을 없애기 menu 선택했을 때의 부분
0
266,764
28,435,915,665
IssuesEvent
2023-04-15 10:00:51
dotnet/aspnetcore
https://api.github.com/repos/dotnet/aspnetcore
closed
I configure the server to requiere client certificate but it accepts the request when client doesn't add a certificate
area-security :heavy_check_mark: Resolution: Answered Status: Resolved
### Is there an existing issue for this? - [X] I have searched the existing issues ### Describe the bug I would like to implement mTLS in a service, but when create the client without adding the certificate, the server accepts the request. This is the code of my server, in the program.cs file of my ASP Core 7 application: ``` try { var builder = WebApplication.CreateBuilder(args); builder.WebHost.ConfigureKestrel((context, options) => { string miStrCertificado = File.ReadAllText("certificados/server.crt"); string miStrKey = File.ReadAllText("certificados/server.key"); X509Certificate2 miCertficadoX509 = X509Certificate2.CreateFromPem(miStrCertificado, miStrKey); X509Certificate2 miCertificado2 = new X509Certificate2(miCertficadoX509.Export(X509ContentType.Pkcs12)); miCertficadoX509.Dispose(); options.ListenAnyIP(5001, listenOptions => { listenOptions.Protocols = HttpProtocols.Http2; listenOptions.UseHttps(miCertificado2); }); options.ConfigureHttpsDefaults(miHttpsOptions => { miHttpsOptions.ClientCertificateMode = ClientCertificateMode.RequireCertificate; miHttpsOptions.ServerCertificate = miCertificado2; }); }); builder.Services.AddHttpContextAccessor(); builder.Services.AddGrpc(); builder.Services.AddCodeFirstGrpc(); builder.Services.AddAuthentication(JwtBearerDefaults.AuthenticationScheme) .AddJwtBearer(options => { options.TokenValidationParameters = new TokenValidationParameters { //@#ESTUDIAR: ¿el string utilizado tiene que tener 64 caracteres? No parece pero parece ser que no //vale cualquier longitud. Igual tiene que ser un string que codifique datos binarios en base 4. IssuerSigningKey = new SymmetricSecurityKey(System.Text.Encoding.Default.GetBytes("ABCD")), RequireExpirationTime = true, RequireSignedTokens = true, ClockSkew = TimeSpan.FromSeconds(10), ValidateIssuer = false, ValidateAudience = false, ValidateLifetime = true, ValidateIssuerSigningKey = true, }; }); builder.Services.AddAuthorization(); var app = builder.Build(); app.UseRouting(); app.UseAuthentication(); app.UseAuthorization(); app.MapGrpcService<MyServiceLogin>(); app.MapGrpcService<MyMainService>(); await app.RunAsync(); } catch (Exception ex) { Console.WriteLine(ex.Message); } ``` This is how I create my client: ``` private ChannelCredentials CrearCredencialesDelCanal() { CallCredentials misCredencialesDeLlamada = CallCredentials.FromInterceptor((c, m) => { m.Add(HeaderNames.Authorization, $"Bearer {_tokenJwt}"); return Task.CompletedTask; }); return ChannelCredentials.Create(new SslCredentials(), misCredencialesDeLlamada); } private GrpcChannelOptions CrearOpcionesDelCanal() { HttpClientHandler miHttpHandler = new HttpClientHandler(); //Always return true miHttpHandler.ServerCertificateCustomValidationCallback = ValidateServerCertificate; HttpClient httpClient = new HttpClient(miHttpHandler); GrpcChannelOptions misOpcionesDelCanal = new GrpcChannelOptions() { MaxReceiveMessageSize = 62914560, MaxSendMessageSize = 62914560, HttpClient = httpClient, }; misOpcionesDelCanal.Credentials = CrearCredencialesDelCanal(); return misOpcionesDelCanal; } private GrpcChannel CrearCanal(string paramStrDireccion, int paramIntPuerto) { GrpcChannelOptions misOpciones = CrearOpcionesDelCanal(); try { Uri miUri = new Uri(paramStrDireccion + ":" + paramIntPuerto.ToString()); GrpcChannel miCanalGrpc = GrpcChannel.ForAddress(miUri, misOpciones); OnCanalGrpcActualizado(miCanalGrpc); return miCanalGrpc; } catch { throw new ArgumentException("Los datos del servidor no son correctos."); } } ``` In this example, when I create the HttClientHandler, I don't set the client certificate, but the server accepts the request and replies. How could I set the server to requiere a client certificate and if it is not add or it is not a valid certificate, refuse the request? Thanks. ### Expected Behavior The server should to denied or not reply to a client when it doesn't add a valid certificate. ### Steps To Reproduce _No response_ ### Exceptions (if any) _No response_ ### .NET Version _No response_ ### Anything else? _No response_
True
I configure the server to requiere client certificate but it accepts the request when client doesn't add a certificate - ### Is there an existing issue for this? - [X] I have searched the existing issues ### Describe the bug I would like to implement mTLS in a service, but when create the client without adding the certificate, the server accepts the request. This is the code of my server, in the program.cs file of my ASP Core 7 application: ``` try { var builder = WebApplication.CreateBuilder(args); builder.WebHost.ConfigureKestrel((context, options) => { string miStrCertificado = File.ReadAllText("certificados/server.crt"); string miStrKey = File.ReadAllText("certificados/server.key"); X509Certificate2 miCertficadoX509 = X509Certificate2.CreateFromPem(miStrCertificado, miStrKey); X509Certificate2 miCertificado2 = new X509Certificate2(miCertficadoX509.Export(X509ContentType.Pkcs12)); miCertficadoX509.Dispose(); options.ListenAnyIP(5001, listenOptions => { listenOptions.Protocols = HttpProtocols.Http2; listenOptions.UseHttps(miCertificado2); }); options.ConfigureHttpsDefaults(miHttpsOptions => { miHttpsOptions.ClientCertificateMode = ClientCertificateMode.RequireCertificate; miHttpsOptions.ServerCertificate = miCertificado2; }); }); builder.Services.AddHttpContextAccessor(); builder.Services.AddGrpc(); builder.Services.AddCodeFirstGrpc(); builder.Services.AddAuthentication(JwtBearerDefaults.AuthenticationScheme) .AddJwtBearer(options => { options.TokenValidationParameters = new TokenValidationParameters { //@#ESTUDIAR: ¿el string utilizado tiene que tener 64 caracteres? No parece pero parece ser que no //vale cualquier longitud. Igual tiene que ser un string que codifique datos binarios en base 4. IssuerSigningKey = new SymmetricSecurityKey(System.Text.Encoding.Default.GetBytes("ABCD")), RequireExpirationTime = true, RequireSignedTokens = true, ClockSkew = TimeSpan.FromSeconds(10), ValidateIssuer = false, ValidateAudience = false, ValidateLifetime = true, ValidateIssuerSigningKey = true, }; }); builder.Services.AddAuthorization(); var app = builder.Build(); app.UseRouting(); app.UseAuthentication(); app.UseAuthorization(); app.MapGrpcService<MyServiceLogin>(); app.MapGrpcService<MyMainService>(); await app.RunAsync(); } catch (Exception ex) { Console.WriteLine(ex.Message); } ``` This is how I create my client: ``` private ChannelCredentials CrearCredencialesDelCanal() { CallCredentials misCredencialesDeLlamada = CallCredentials.FromInterceptor((c, m) => { m.Add(HeaderNames.Authorization, $"Bearer {_tokenJwt}"); return Task.CompletedTask; }); return ChannelCredentials.Create(new SslCredentials(), misCredencialesDeLlamada); } private GrpcChannelOptions CrearOpcionesDelCanal() { HttpClientHandler miHttpHandler = new HttpClientHandler(); //Always return true miHttpHandler.ServerCertificateCustomValidationCallback = ValidateServerCertificate; HttpClient httpClient = new HttpClient(miHttpHandler); GrpcChannelOptions misOpcionesDelCanal = new GrpcChannelOptions() { MaxReceiveMessageSize = 62914560, MaxSendMessageSize = 62914560, HttpClient = httpClient, }; misOpcionesDelCanal.Credentials = CrearCredencialesDelCanal(); return misOpcionesDelCanal; } private GrpcChannel CrearCanal(string paramStrDireccion, int paramIntPuerto) { GrpcChannelOptions misOpciones = CrearOpcionesDelCanal(); try { Uri miUri = new Uri(paramStrDireccion + ":" + paramIntPuerto.ToString()); GrpcChannel miCanalGrpc = GrpcChannel.ForAddress(miUri, misOpciones); OnCanalGrpcActualizado(miCanalGrpc); return miCanalGrpc; } catch { throw new ArgumentException("Los datos del servidor no son correctos."); } } ``` In this example, when I create the HttClientHandler, I don't set the client certificate, but the server accepts the request and replies. How could I set the server to requiere a client certificate and if it is not add or it is not a valid certificate, refuse the request? Thanks. ### Expected Behavior The server should to denied or not reply to a client when it doesn't add a valid certificate. ### Steps To Reproduce _No response_ ### Exceptions (if any) _No response_ ### .NET Version _No response_ ### Anything else? _No response_
non_process
i configure the server to requiere client certificate but it accepts the request when client doesn t add a certificate is there an existing issue for this i have searched the existing issues describe the bug i would like to implement mtls in a service but when create the client without adding the certificate the server accepts the request this is the code of my server in the program cs file of my asp core application try var builder webapplication createbuilder args builder webhost configurekestrel context options string mistrcertificado file readalltext certificados server crt string mistrkey file readalltext certificados server key createfrompem mistrcertificado mistrkey new export dispose options listenanyip listenoptions listenoptions protocols httpprotocols listenoptions usehttps options configurehttpsdefaults mihttpsoptions mihttpsoptions clientcertificatemode clientcertificatemode requirecertificate mihttpsoptions servercertificate builder services addhttpcontextaccessor builder services addgrpc builder services addcodefirstgrpc builder services addauthentication jwtbearerdefaults authenticationscheme addjwtbearer options options tokenvalidationparameters new tokenvalidationparameters estudiar ¿el string utilizado tiene que tener caracteres no parece pero parece ser que no vale cualquier longitud igual tiene que ser un string que codifique datos binarios en base issuersigningkey new symmetricsecuritykey system text encoding default getbytes abcd requireexpirationtime true requiresignedtokens true clockskew timespan fromseconds validateissuer false validateaudience false validatelifetime true validateissuersigningkey true builder services addauthorization var app builder build app userouting app useauthentication app useauthorization app mapgrpcservice app mapgrpcservice await app runasync catch exception ex console writeline ex message this is how i create my client private channelcredentials crearcredencialesdelcanal callcredentials miscredencialesdellamada callcredentials frominterceptor c m m add headernames authorization bearer tokenjwt return task completedtask return channelcredentials create new sslcredentials miscredencialesdellamada private grpcchanneloptions crearopcionesdelcanal httpclienthandler mihttphandler new httpclienthandler always return true mihttphandler servercertificatecustomvalidationcallback validateservercertificate httpclient httpclient new httpclient mihttphandler grpcchanneloptions misopcionesdelcanal new grpcchanneloptions maxreceivemessagesize maxsendmessagesize httpclient httpclient misopcionesdelcanal credentials crearcredencialesdelcanal return misopcionesdelcanal private grpcchannel crearcanal string paramstrdireccion int paramintpuerto grpcchanneloptions misopciones crearopcionesdelcanal try uri miuri new uri paramstrdireccion paramintpuerto tostring grpcchannel micanalgrpc grpcchannel foraddress miuri misopciones oncanalgrpcactualizado micanalgrpc return micanalgrpc catch throw new argumentexception los datos del servidor no son correctos in this example when i create the httclienthandler i don t set the client certificate but the server accepts the request and replies how could i set the server to requiere a client certificate and if it is not add or it is not a valid certificate refuse the request thanks expected behavior the server should to denied or not reply to a client when it doesn t add a valid certificate steps to reproduce no response exceptions if any no response net version no response anything else no response
0
137,742
30,745,626,555
IssuesEvent
2023-07-28 14:50:26
WordPress/openverse
https://api.github.com/repos/WordPress/openverse
closed
Baseline monitoring changes for Terraform and the unhealthy host count alarm
🟧 priority: high 🌟 goal: addition 💻 aspect: code 🧱 stack: infra
## Description <!-- Describe the feature and how it solves the problem. --> Project thread: https://github.com/WordPress/openverse/issues/2344 Implementation plan: https://docs.openverse.org/projects/proposals/monitoring/20230606_implementation_plan_ecs_alarms.html 1. Create the monitoring modules for frontend and API staging and production and move existing alarms into these - Create a new `next/modules/monitoring` directory with a directory for each service/environment: `staging-frontend`, `production-frontend`, `staging-api`, etc. - This includes moving the UptimeRobot configuration for each service as well as the database and Redis monitors - Rename `service-monitors` to `service-uptime-robot` to clarify the module's purpose - Also create the new SNS topic for the unstable alerts' notification channel - This does not include moving the ECS service cloudwatch dashboard module, that should remain in the root modules 1. Create the unhealthy host count alarm for production and staging services > **Note** > > The only thing that will be present in the staging monitoring modules for each service is the unhealthy host count alarm. ## Additional context <!-- Add any other context about the feature here; or delete the section entirely. --> This issue will remain open until the unhealthy host count alarm is stabilised. However, once the terraform configuration changes and the new monitoring modules are present, all other alarm issues in this milestone will be unblocked. <!-- If you would like to work on this, please comment below separately. -->
1.0
Baseline monitoring changes for Terraform and the unhealthy host count alarm - ## Description <!-- Describe the feature and how it solves the problem. --> Project thread: https://github.com/WordPress/openverse/issues/2344 Implementation plan: https://docs.openverse.org/projects/proposals/monitoring/20230606_implementation_plan_ecs_alarms.html 1. Create the monitoring modules for frontend and API staging and production and move existing alarms into these - Create a new `next/modules/monitoring` directory with a directory for each service/environment: `staging-frontend`, `production-frontend`, `staging-api`, etc. - This includes moving the UptimeRobot configuration for each service as well as the database and Redis monitors - Rename `service-monitors` to `service-uptime-robot` to clarify the module's purpose - Also create the new SNS topic for the unstable alerts' notification channel - This does not include moving the ECS service cloudwatch dashboard module, that should remain in the root modules 1. Create the unhealthy host count alarm for production and staging services > **Note** > > The only thing that will be present in the staging monitoring modules for each service is the unhealthy host count alarm. ## Additional context <!-- Add any other context about the feature here; or delete the section entirely. --> This issue will remain open until the unhealthy host count alarm is stabilised. However, once the terraform configuration changes and the new monitoring modules are present, all other alarm issues in this milestone will be unblocked. <!-- If you would like to work on this, please comment below separately. -->
non_process
baseline monitoring changes for terraform and the unhealthy host count alarm description project thread implementation plan create the monitoring modules for frontend and api staging and production and move existing alarms into these create a new next modules monitoring directory with a directory for each service environment staging frontend production frontend staging api etc this includes moving the uptimerobot configuration for each service as well as the database and redis monitors rename service monitors to service uptime robot to clarify the module s purpose also create the new sns topic for the unstable alerts notification channel this does not include moving the ecs service cloudwatch dashboard module that should remain in the root modules create the unhealthy host count alarm for production and staging services note the only thing that will be present in the staging monitoring modules for each service is the unhealthy host count alarm additional context this issue will remain open until the unhealthy host count alarm is stabilised however once the terraform configuration changes and the new monitoring modules are present all other alarm issues in this milestone will be unblocked
0
10,722
13,524,633,324
IssuesEvent
2020-09-15 11:51:46
gfx-rs/naga
https://api.github.com/repos/gfx-rs/naga
closed
Use typifier in the backends
area: back-end area: processing help wanted kind: refactor
Currently it's used in the frontends only, and it mutates the type arena. GLSL and MSL backends would benefit from having this automated too, but we need to do something about the mutable arena.
1.0
Use typifier in the backends - Currently it's used in the frontends only, and it mutates the type arena. GLSL and MSL backends would benefit from having this automated too, but we need to do something about the mutable arena.
process
use typifier in the backends currently it s used in the frontends only and it mutates the type arena glsl and msl backends would benefit from having this automated too but we need to do something about the mutable arena
1
1,142
3,631,346,238
IssuesEvent
2016-02-11 00:54:04
ParsePlatform/parse-server
https://api.github.com/repos/ParsePlatform/parse-server
closed
ParseServer does not delete Session object on logout
bug in-process
I have noticed that the behavior of hosted Parse is to create a new Session object when a user login and delete this object when this user logout from the same device. When I tested the ParseServer version instead of the hosted Parse I noticed that this behavior is different, the Session object is created on login as expected, but on logout the object is not deleted, so if the user keeps logging in and logging out there will be as many Session objects on DB as the number of times the user login and logout. Is this the expected behavior or is it a bug on ParseServer?
1.0
ParseServer does not delete Session object on logout - I have noticed that the behavior of hosted Parse is to create a new Session object when a user login and delete this object when this user logout from the same device. When I tested the ParseServer version instead of the hosted Parse I noticed that this behavior is different, the Session object is created on login as expected, but on logout the object is not deleted, so if the user keeps logging in and logging out there will be as many Session objects on DB as the number of times the user login and logout. Is this the expected behavior or is it a bug on ParseServer?
process
parseserver does not delete session object on logout i have noticed that the behavior of hosted parse is to create a new session object when a user login and delete this object when this user logout from the same device when i tested the parseserver version instead of the hosted parse i noticed that this behavior is different the session object is created on login as expected but on logout the object is not deleted so if the user keeps logging in and logging out there will be as many session objects on db as the number of times the user login and logout is this the expected behavior or is it a bug on parseserver
1
13,188
8,824,079,554
IssuesEvent
2019-01-02 15:50:46
arcus-azure/arcus.security
https://api.github.com/repos/arcus-azure/arcus.security
closed
Provide support for Managed identities for Azure resources authentication
azure-key-vault feature security
Provide support for Managed identities for Azure resources authentication, formerly known as Azure AD Managed Service Identity (MSI).
True
Provide support for Managed identities for Azure resources authentication - Provide support for Managed identities for Azure resources authentication, formerly known as Azure AD Managed Service Identity (MSI).
non_process
provide support for managed identities for azure resources authentication provide support for managed identities for azure resources authentication formerly known as azure ad managed service identity msi
0
11,679
14,537,160,807
IssuesEvent
2020-12-15 08:46:27
threefoldtech/js-sdk
https://api.github.com/repos/threefoldtech/js-sdk
closed
Unable to start threebot server after fresh install
process_wontfix type_bug
After installing a fresh js-sdk (following the instructions on the wiki), I am unable to start a threebot. It keeps failing on the fact that I don't have an identity (and to get one I presumably need to run the threebot first). Passing the `--development` flag, which claims no identity is needed, does not solve the issue. ``` threebot start --local --development 2020-11-20 23:27:38.978 | ERROR | jumpscale.tools.errorhandler.errorhandler:_handle_exception:93 - jumpscale.core.exceptions.exceptions.Value: No configured identity found Traceback (most recent call last): > File "/home/lee/.cache/pypoetry/virtualenvs/js-sdk-4phGcms2-py3.8/bin/threebot", line 5, in <module> cli() └ <Group cli> File "/home/lee/.cache/pypoetry/virtualenvs/js-sdk-4phGcms2-py3.8/lib/python3.8/site-packages/click/core.py", line 829, in __call__ return self.main(*args, **kwargs) │ │ │ └ {} │ │ └ () │ └ <function BaseCommand.main at 0x7ff40a282ca0> └ <Group cli> File "/home/lee/.cache/pypoetry/virtualenvs/js-sdk-4phGcms2-py3.8/lib/python3.8/site-packages/click/core.py", line 782, in main rv = self.invoke(ctx) │ │ └ <click.core.Context object at 0x7ff40b1740a0> │ └ <function MultiCommand.invoke at 0x7ff40a283a60> └ <Group cli> File "/home/lee/.cache/pypoetry/virtualenvs/js-sdk-4phGcms2-py3.8/lib/python3.8/site-packages/click/core.py", line 1259, in invoke return _process_result(sub_ctx.command.invoke(sub_ctx)) │ │ │ │ └ <click.core.Context object at 0x7ff40b203580> │ │ │ └ <function Command.invoke at 0x7ff40a283670> │ │ └ <Command start> │ └ <click.core.Context object at 0x7ff40b203580> └ <function MultiCommand.invoke.<locals>._process_result at 0x7ff40b2301f0> File "/home/lee/.cache/pypoetry/virtualenvs/js-sdk-4phGcms2-py3.8/lib/python3.8/site-packages/click/core.py", line 1066, in invoke return ctx.invoke(self.callback, **ctx.params) │ │ │ │ │ └ {'local': True, 'development': True, 'identity': None, 'domain': None, 'email': None, 'background': False} │ │ │ │ └ <click.core.Context object at 0x7ff40b203580> │ │ │ └ <function start at 0x7ff40712d790> │ │ └ <Command start> │ └ <function Context.invoke at 0x7ff40a282790> └ <click.core.Context object at 0x7ff40b203580> File "/home/lee/.cache/pypoetry/virtualenvs/js-sdk-4phGcms2-py3.8/lib/python3.8/site-packages/click/core.py", line 610, in invoke return callback(*args, **kwargs) │ │ └ {'local': True, 'development': True, 'identity': None, 'domain': None, 'email': None, 'background': False} │ └ () └ <function start at 0x7ff40712d790> File "/home/lee/js-sdk/jumpscale/entry_points/threebot.py", line 86, in start create_wallets_if_not_exists() └ <function create_wallets_if_not_exists at 0x7ff40712d820> File "/home/lee/js-sdk/jumpscale/entry_points/threebot.py", line 277, in create_wallets_if_not_exists if not test and "testnet" in j.core.identity.me.explorer_url: │ │ └ <property object at 0x7ff40a288cc0> │ └ <jumpscale.loader.J object at 0x7ff40a27c8b0> └ False File "/home/lee/js-sdk/jumpscale/core/identity/__init__.py", line 208, in me raise Value("No configured identity found") └ <class 'jumpscale.core.exceptions.exceptions.Value'> jumpscale.core.exceptions.exceptions.Value: No configured identity found ```
1.0
Unable to start threebot server after fresh install - After installing a fresh js-sdk (following the instructions on the wiki), I am unable to start a threebot. It keeps failing on the fact that I don't have an identity (and to get one I presumably need to run the threebot first). Passing the `--development` flag, which claims no identity is needed, does not solve the issue. ``` threebot start --local --development 2020-11-20 23:27:38.978 | ERROR | jumpscale.tools.errorhandler.errorhandler:_handle_exception:93 - jumpscale.core.exceptions.exceptions.Value: No configured identity found Traceback (most recent call last): > File "/home/lee/.cache/pypoetry/virtualenvs/js-sdk-4phGcms2-py3.8/bin/threebot", line 5, in <module> cli() └ <Group cli> File "/home/lee/.cache/pypoetry/virtualenvs/js-sdk-4phGcms2-py3.8/lib/python3.8/site-packages/click/core.py", line 829, in __call__ return self.main(*args, **kwargs) │ │ │ └ {} │ │ └ () │ └ <function BaseCommand.main at 0x7ff40a282ca0> └ <Group cli> File "/home/lee/.cache/pypoetry/virtualenvs/js-sdk-4phGcms2-py3.8/lib/python3.8/site-packages/click/core.py", line 782, in main rv = self.invoke(ctx) │ │ └ <click.core.Context object at 0x7ff40b1740a0> │ └ <function MultiCommand.invoke at 0x7ff40a283a60> └ <Group cli> File "/home/lee/.cache/pypoetry/virtualenvs/js-sdk-4phGcms2-py3.8/lib/python3.8/site-packages/click/core.py", line 1259, in invoke return _process_result(sub_ctx.command.invoke(sub_ctx)) │ │ │ │ └ <click.core.Context object at 0x7ff40b203580> │ │ │ └ <function Command.invoke at 0x7ff40a283670> │ │ └ <Command start> │ └ <click.core.Context object at 0x7ff40b203580> └ <function MultiCommand.invoke.<locals>._process_result at 0x7ff40b2301f0> File "/home/lee/.cache/pypoetry/virtualenvs/js-sdk-4phGcms2-py3.8/lib/python3.8/site-packages/click/core.py", line 1066, in invoke return ctx.invoke(self.callback, **ctx.params) │ │ │ │ │ └ {'local': True, 'development': True, 'identity': None, 'domain': None, 'email': None, 'background': False} │ │ │ │ └ <click.core.Context object at 0x7ff40b203580> │ │ │ └ <function start at 0x7ff40712d790> │ │ └ <Command start> │ └ <function Context.invoke at 0x7ff40a282790> └ <click.core.Context object at 0x7ff40b203580> File "/home/lee/.cache/pypoetry/virtualenvs/js-sdk-4phGcms2-py3.8/lib/python3.8/site-packages/click/core.py", line 610, in invoke return callback(*args, **kwargs) │ │ └ {'local': True, 'development': True, 'identity': None, 'domain': None, 'email': None, 'background': False} │ └ () └ <function start at 0x7ff40712d790> File "/home/lee/js-sdk/jumpscale/entry_points/threebot.py", line 86, in start create_wallets_if_not_exists() └ <function create_wallets_if_not_exists at 0x7ff40712d820> File "/home/lee/js-sdk/jumpscale/entry_points/threebot.py", line 277, in create_wallets_if_not_exists if not test and "testnet" in j.core.identity.me.explorer_url: │ │ └ <property object at 0x7ff40a288cc0> │ └ <jumpscale.loader.J object at 0x7ff40a27c8b0> └ False File "/home/lee/js-sdk/jumpscale/core/identity/__init__.py", line 208, in me raise Value("No configured identity found") └ <class 'jumpscale.core.exceptions.exceptions.Value'> jumpscale.core.exceptions.exceptions.Value: No configured identity found ```
process
unable to start threebot server after fresh install after installing a fresh js sdk following the instructions on the wiki i am unable to start a threebot it keeps failing on the fact that i don t have an identity and to get one i presumably need to run the threebot first passing the development flag which claims no identity is needed does not solve the issue threebot start local development error jumpscale tools errorhandler errorhandler handle exception jumpscale core exceptions exceptions value no configured identity found traceback most recent call last file home lee cache pypoetry virtualenvs js sdk bin threebot line in cli └ file home lee cache pypoetry virtualenvs js sdk lib site packages click core py line in call return self main args kwargs │ │ │ └ │ │ └ │ └ └ file home lee cache pypoetry virtualenvs js sdk lib site packages click core py line in main rv self invoke ctx │ │ └ │ └ └ file home lee cache pypoetry virtualenvs js sdk lib site packages click core py line in invoke return process result sub ctx command invoke sub ctx │ │ │ │ └ │ │ │ └ │ │ └ │ └ └ process result at file home lee cache pypoetry virtualenvs js sdk lib site packages click core py line in invoke return ctx invoke self callback ctx params │ │ │ │ │ └ local true development true identity none domain none email none background false │ │ │ │ └ │ │ │ └ │ │ └ │ └ └ file home lee cache pypoetry virtualenvs js sdk lib site packages click core py line in invoke return callback args kwargs │ │ └ local true development true identity none domain none email none background false │ └ └ file home lee js sdk jumpscale entry points threebot py line in start create wallets if not exists └ file home lee js sdk jumpscale entry points threebot py line in create wallets if not exists if not test and testnet in j core identity me explorer url │ │ └ │ └ └ false file home lee js sdk jumpscale core identity init py line in me raise value no configured identity found └ jumpscale core exceptions exceptions value no configured identity found
1
494,003
14,243,224,443
IssuesEvent
2020-11-19 03:51:08
jcr7467/UCLAbookstack
https://api.github.com/repos/jcr7467/UCLAbookstack
opened
Add Book Quality Feature
Priority - Medium new feature
We have filters for the subject areas, we also need a filter for the book quality!
1.0
Add Book Quality Feature - We have filters for the subject areas, we also need a filter for the book quality!
non_process
add book quality feature we have filters for the subject areas we also need a filter for the book quality
0
402,850
27,389,629,178
IssuesEvent
2023-02-28 15:28:44
weaveworks/weave-gitops
https://api.github.com/repos/weaveworks/weave-gitops
closed
docs: consider breaking out the CLI installations into a separate page
documentation
Personally I would also break out the CLI installations into a separate page _Originally posted by @sympatheticmoose in https://github.com/weaveworks/weave-gitops/issues/3039#issuecomment-1324887551_
1.0
docs: consider breaking out the CLI installations into a separate page - Personally I would also break out the CLI installations into a separate page _Originally posted by @sympatheticmoose in https://github.com/weaveworks/weave-gitops/issues/3039#issuecomment-1324887551_
non_process
docs consider breaking out the cli installations into a separate page personally i would also break out the cli installations into a separate page originally posted by sympatheticmoose in
0
2,663
5,436,410,569
IssuesEvent
2017-03-06 00:47:10
jlm2017/jlm-video-subtitles
https://api.github.com/repos/jlm2017/jlm-video-subtitles
reopened
[subtitles] [en] MONSIEUR HAMON, CHOISISSEZ !
Process: [6] Approved ⚑ English
# Video title MONSIEUR HAMON, CHOISISSEZ ! # URL https://www.youtube.com/watch?v=tQIsWprmxDA&t= # Youtube subtitles language Langue des sous-titres (Anglais) # Duration 10:44 # Subtitles URL https://www.youtube.com/timedtext_editor?lang=en&action_mde_edit_form=1&ref=player&tab=captions&v=tQIsWprmxDA&bl=vmp&ui=hd
1.0
[subtitles] [en] MONSIEUR HAMON, CHOISISSEZ ! - # Video title MONSIEUR HAMON, CHOISISSEZ ! # URL https://www.youtube.com/watch?v=tQIsWprmxDA&t= # Youtube subtitles language Langue des sous-titres (Anglais) # Duration 10:44 # Subtitles URL https://www.youtube.com/timedtext_editor?lang=en&action_mde_edit_form=1&ref=player&tab=captions&v=tQIsWprmxDA&bl=vmp&ui=hd
process
monsieur hamon choisissez video title monsieur hamon choisissez url youtube subtitles language langue des sous titres anglais duration subtitles url
1
43,880
13,040,724,635
IssuesEvent
2020-07-28 19:02:03
LevyForchh/clusterfuzz
https://api.github.com/repos/LevyForchh/clusterfuzz
opened
CVE-2019-12855 (High) detected in Twisted-14.0.0.tar.bz2
security vulnerability
## CVE-2019-12855 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Twisted-14.0.0.tar.bz2</b></p></summary> <p>An asynchronous networking framework written in Python</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/76/38/cf8f81c1d7d84fec922d67f0d92bfa9fee59145d875d7263ceefa2bbbaf4/Twisted-14.0.0.tar.bz2">https://files.pythonhosted.org/packages/76/38/cf8f81c1d7d84fec922d67f0d92bfa9fee59145d875d7263ceefa2bbbaf4/Twisted-14.0.0.tar.bz2</a></p> <p>Path to dependency file: /tmp/ws-ua_20200728185744_PPILLS/archiveExtraction_CNYBMU/20200728185744/ws-scm_depth_0/clusterfuzz/resources/platform/linux/peach/peach_mutator/peach_mutator/requirements.txt</p> <p>Path to vulnerable library: _depth_0/clusterfuzz/resources/platform/linux/peach/peach_mutator/peach_mutator/requirements.txt,/clusterfuzz/src/local/butler/scripts,/clusterfuzz/src/python/bot/untrusted_runner/build,_depth_0/clusterfuzz/resources/platform/linux/peach/peach_mutator/peach_mutator/third_party/peach/requirements.txt,/clusterfuzz/src/appengine/handlers/cron/project,/clusterfuzz/src/python/bot/tasks</p> <p> Dependency Hierarchy: - :x: **Twisted-14.0.0.tar.bz2** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/LevyForchh/clusterfuzz/commit/aefe4b0859891117218fba5984e5c3e753ea9597">aefe4b0859891117218fba5984e5c3e753ea9597</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In words.protocols.jabber.xmlstream in Twisted through 19.2.1, XMPP support did not verify certificates when used with TLS, allowing an attacker to MITM connections. <p>Publish Date: 2019-06-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-12855>CVE-2019-12855</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Change files</p> <p>Origin: <a href="https://github.com/advisories/GHSA-65rm-h285-5cc5">https://github.com/advisories/GHSA-65rm-h285-5cc5</a></p> <p>Release Date: 2019-12-19</p> <p>Fix Resolution: Replace or update the following file: 19.7.0</p> </p> </details> <p></p> *** <!-- REMEDIATE-OPEN-PR-START --> - [ ] Check this box to open an automated fix PR <!-- REMEDIATE-OPEN-PR-END --> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Python","packageName":"Twisted","packageVersion":"14.0.0","isTransitiveDependency":false,"dependencyTree":"Twisted:14.0.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"Replace or update the following file: 19.7.0"}],"vulnerabilityIdentifier":"CVE-2019-12855","vulnerabilityDetails":"In words.protocols.jabber.xmlstream in Twisted through 19.2.1, XMPP support did not verify certificates when used with TLS, allowing an attacker to MITM connections.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-12855","cvss3Severity":"high","cvss3Score":"7.4","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2019-12855 (High) detected in Twisted-14.0.0.tar.bz2 - ## CVE-2019-12855 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Twisted-14.0.0.tar.bz2</b></p></summary> <p>An asynchronous networking framework written in Python</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/76/38/cf8f81c1d7d84fec922d67f0d92bfa9fee59145d875d7263ceefa2bbbaf4/Twisted-14.0.0.tar.bz2">https://files.pythonhosted.org/packages/76/38/cf8f81c1d7d84fec922d67f0d92bfa9fee59145d875d7263ceefa2bbbaf4/Twisted-14.0.0.tar.bz2</a></p> <p>Path to dependency file: /tmp/ws-ua_20200728185744_PPILLS/archiveExtraction_CNYBMU/20200728185744/ws-scm_depth_0/clusterfuzz/resources/platform/linux/peach/peach_mutator/peach_mutator/requirements.txt</p> <p>Path to vulnerable library: _depth_0/clusterfuzz/resources/platform/linux/peach/peach_mutator/peach_mutator/requirements.txt,/clusterfuzz/src/local/butler/scripts,/clusterfuzz/src/python/bot/untrusted_runner/build,_depth_0/clusterfuzz/resources/platform/linux/peach/peach_mutator/peach_mutator/third_party/peach/requirements.txt,/clusterfuzz/src/appengine/handlers/cron/project,/clusterfuzz/src/python/bot/tasks</p> <p> Dependency Hierarchy: - :x: **Twisted-14.0.0.tar.bz2** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/LevyForchh/clusterfuzz/commit/aefe4b0859891117218fba5984e5c3e753ea9597">aefe4b0859891117218fba5984e5c3e753ea9597</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In words.protocols.jabber.xmlstream in Twisted through 19.2.1, XMPP support did not verify certificates when used with TLS, allowing an attacker to MITM connections. <p>Publish Date: 2019-06-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-12855>CVE-2019-12855</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Change files</p> <p>Origin: <a href="https://github.com/advisories/GHSA-65rm-h285-5cc5">https://github.com/advisories/GHSA-65rm-h285-5cc5</a></p> <p>Release Date: 2019-12-19</p> <p>Fix Resolution: Replace or update the following file: 19.7.0</p> </p> </details> <p></p> *** <!-- REMEDIATE-OPEN-PR-START --> - [ ] Check this box to open an automated fix PR <!-- REMEDIATE-OPEN-PR-END --> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Python","packageName":"Twisted","packageVersion":"14.0.0","isTransitiveDependency":false,"dependencyTree":"Twisted:14.0.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"Replace or update the following file: 19.7.0"}],"vulnerabilityIdentifier":"CVE-2019-12855","vulnerabilityDetails":"In words.protocols.jabber.xmlstream in Twisted through 19.2.1, XMPP support did not verify certificates when used with TLS, allowing an attacker to MITM connections.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-12855","cvss3Severity":"high","cvss3Score":"7.4","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_process
cve high detected in twisted tar cve high severity vulnerability vulnerable library twisted tar an asynchronous networking framework written in python library home page a href path to dependency file tmp ws ua ppills archiveextraction cnybmu ws scm depth clusterfuzz resources platform linux peach peach mutator peach mutator requirements txt path to vulnerable library depth clusterfuzz resources platform linux peach peach mutator peach mutator requirements txt clusterfuzz src local butler scripts clusterfuzz src python bot untrusted runner build depth clusterfuzz resources platform linux peach peach mutator peach mutator third party peach requirements txt clusterfuzz src appengine handlers cron project clusterfuzz src python bot tasks dependency hierarchy x twisted tar vulnerable library found in head commit a href vulnerability details in words protocols jabber xmlstream in twisted through xmpp support did not verify certificates when used with tls allowing an attacker to mitm connections publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact none for more information on scores click a href suggested fix type change files origin a href release date fix resolution replace or update the following file check this box to open an automated fix pr isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails in words protocols jabber xmlstream in twisted through xmpp support did not verify certificates when used with tls allowing an attacker to mitm connections vulnerabilityurl
0
22,676
31,899,966,307
IssuesEvent
2023-09-18 07:03:11
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
reopened
pattern recognition receptor signaling pathway - NTR and reorganization
multi-species process
Hello, This comes from the SIB-Virus group @pmasson55 and Chantal 1. pattern recognition receptor signaling pathway GO:0002221 Add as childen of GO:0002221: - [x] 1.1 Create GO: endolysosomal pattern recognition receptor signaling pathway 1.1.1 Create GO: endolysosomal toll-like receptor signaling pathways - [x] moved terms under the new term 1.1.1.1 GO:0034138 toll-like receptor 3 signaling pathway 1.1.1.2 GO:0034154 toll-like receptor 7 signaling pathway 1.1.1.3 GO:0034158 toll-like receptor 8 signaling pathway 1.1.1.4 GO:0034162 toll-like receptor 9 signaling pathway 1.1.1.5 GO:0034170 toll-like receptor 11 signaling pathway 1.1.1.6 GO:0034174 toll-like receptor 12 signaling pathway 1.1.1.7 GO:0034178 toll-like receptor 13 signaling pathway -> endolysosomal DNA, rRNA, ssRNA sensors 1.2. GO:0002753: cytosolic pattern recognition receptor signaling pathway 1.2.1 GO:0039529: RIG-I signaling pathway -> cytosolic ssRNA and dsRNA sensor - [x] added ssRNA and dsRNA in definition 1.2.2 GO:0039530: MDA-5 signaling pathway -> cytosolic dsRNA sensor - [x] added dsRNA in definition 1.2.3 GO:0039585: PKR signal transduction -> cytosolic dsRNA sensor - [x] dsRNA was already in in definition - [x] 1.2.4 Create GO: cGAS/STING pathway -> cytosolic DNA sensor 1.3 GO:0002752: cell surface pattern recognition receptor signaling pathway - [x] 1.3.1 Create GO: cell surface toll-like receptor signaling pathway 1.3.1.1 GO:0034130 toll-like receptor 1 signaling pathway 1.3.1.2 GO:0034134 toll-like receptor 2 signaling pathway 1.3.1.3 GO:0034142 toll-like receptor 4 signaling pathway 1.3.1.4 GO:0034146 toll-like receptor 5 signaling pathway 1.3.1.5 GO:0034150 toll-like receptor 6 signaling pathway 1.3.1.6 GO:0034166 toll-like receptor 10 signaling pathway 1.3.1.7 GO:0035681 toll-like receptor 15 signaling pathway -> lipopeptides, dsRNA, ssRNA, LPS, flagellin sensors GO:0002754 toll-like receptor signaling pathway will have both children “endolysosomal toll-like receptor signaling pathways” and “cell surface toll-like receptor signaling pathways” Decide if we keep GO:0039528 cytoplasmic pattern recognition receptor signaling pathway in response to virus The other PRR signaling pathways (membrane, endolysosomal) do not have this granularity List of terms negative/positive regulation: Negative/positive regulation of endolysosomal pattern recognition receptor signaling pathway Negative/positive regulation cytosolic pattern recognition receptor signaling pathway Negative/positive regulation cell surface pattern recognition receptor signaling pathway Liu, Xing & Wang, Qiang & Chen, Wei & Wang, Chen. (2013). Dynamic regulation of innate immunity by ubiquitin and ubiquitin-like proteins. Cytokine & growth factor reviews. 24. 10.1016/j.cytogfr.2013.07.002. ---- Any comments/suggestions ? Thanks, Pascale
1.0
pattern recognition receptor signaling pathway - NTR and reorganization - Hello, This comes from the SIB-Virus group @pmasson55 and Chantal 1. pattern recognition receptor signaling pathway GO:0002221 Add as childen of GO:0002221: - [x] 1.1 Create GO: endolysosomal pattern recognition receptor signaling pathway 1.1.1 Create GO: endolysosomal toll-like receptor signaling pathways - [x] moved terms under the new term 1.1.1.1 GO:0034138 toll-like receptor 3 signaling pathway 1.1.1.2 GO:0034154 toll-like receptor 7 signaling pathway 1.1.1.3 GO:0034158 toll-like receptor 8 signaling pathway 1.1.1.4 GO:0034162 toll-like receptor 9 signaling pathway 1.1.1.5 GO:0034170 toll-like receptor 11 signaling pathway 1.1.1.6 GO:0034174 toll-like receptor 12 signaling pathway 1.1.1.7 GO:0034178 toll-like receptor 13 signaling pathway -> endolysosomal DNA, rRNA, ssRNA sensors 1.2. GO:0002753: cytosolic pattern recognition receptor signaling pathway 1.2.1 GO:0039529: RIG-I signaling pathway -> cytosolic ssRNA and dsRNA sensor - [x] added ssRNA and dsRNA in definition 1.2.2 GO:0039530: MDA-5 signaling pathway -> cytosolic dsRNA sensor - [x] added dsRNA in definition 1.2.3 GO:0039585: PKR signal transduction -> cytosolic dsRNA sensor - [x] dsRNA was already in in definition - [x] 1.2.4 Create GO: cGAS/STING pathway -> cytosolic DNA sensor 1.3 GO:0002752: cell surface pattern recognition receptor signaling pathway - [x] 1.3.1 Create GO: cell surface toll-like receptor signaling pathway 1.3.1.1 GO:0034130 toll-like receptor 1 signaling pathway 1.3.1.2 GO:0034134 toll-like receptor 2 signaling pathway 1.3.1.3 GO:0034142 toll-like receptor 4 signaling pathway 1.3.1.4 GO:0034146 toll-like receptor 5 signaling pathway 1.3.1.5 GO:0034150 toll-like receptor 6 signaling pathway 1.3.1.6 GO:0034166 toll-like receptor 10 signaling pathway 1.3.1.7 GO:0035681 toll-like receptor 15 signaling pathway -> lipopeptides, dsRNA, ssRNA, LPS, flagellin sensors GO:0002754 toll-like receptor signaling pathway will have both children “endolysosomal toll-like receptor signaling pathways” and “cell surface toll-like receptor signaling pathways” Decide if we keep GO:0039528 cytoplasmic pattern recognition receptor signaling pathway in response to virus The other PRR signaling pathways (membrane, endolysosomal) do not have this granularity List of terms negative/positive regulation: Negative/positive regulation of endolysosomal pattern recognition receptor signaling pathway Negative/positive regulation cytosolic pattern recognition receptor signaling pathway Negative/positive regulation cell surface pattern recognition receptor signaling pathway Liu, Xing & Wang, Qiang & Chen, Wei & Wang, Chen. (2013). Dynamic regulation of innate immunity by ubiquitin and ubiquitin-like proteins. Cytokine & growth factor reviews. 24. 10.1016/j.cytogfr.2013.07.002. ---- Any comments/suggestions ? Thanks, Pascale
process
pattern recognition receptor signaling pathway ntr and reorganization hello this comes from the sib virus group and chantal pattern recognition receptor signaling pathway go add as childen of go create go endolysosomal pattern recognition receptor signaling pathway create go endolysosomal toll like receptor signaling pathways moved terms under the new term go toll like receptor signaling pathway go toll like receptor signaling pathway go toll like receptor signaling pathway go toll like receptor signaling pathway go toll like receptor signaling pathway go toll like receptor signaling pathway go toll like receptor signaling pathway endolysosomal dna rrna ssrna sensors go cytosolic pattern recognition receptor signaling pathway go rig i signaling pathway cytosolic ssrna and dsrna sensor added ssrna and dsrna in definition go mda signaling pathway cytosolic dsrna sensor added dsrna in definition go pkr signal transduction cytosolic dsrna sensor dsrna was already in in definition create go cgas sting pathway cytosolic dna sensor go cell surface pattern recognition receptor signaling pathway create go cell surface toll like receptor signaling pathway go toll like receptor signaling pathway go toll like receptor signaling pathway go toll like receptor signaling pathway go toll like receptor signaling pathway go toll like receptor signaling pathway go toll like receptor signaling pathway go toll like receptor signaling pathway lipopeptides dsrna ssrna lps flagellin sensors go toll like receptor signaling pathway will have both children “endolysosomal toll like receptor signaling pathways” and “cell surface toll like receptor signaling pathways” decide if we keep go cytoplasmic pattern recognition receptor signaling pathway in response to virus the other prr signaling pathways membrane endolysosomal do not have this granularity list of terms negative positive regulation negative positive regulation of endolysosomal pattern recognition receptor signaling pathway negative positive regulation cytosolic pattern recognition receptor signaling pathway negative positive regulation cell surface pattern recognition receptor signaling pathway liu xing wang qiang chen wei wang chen dynamic regulation of innate immunity by ubiquitin and ubiquitin like proteins cytokine growth factor reviews j cytogfr any comments suggestions thanks pascale
1
7,014
2,596,372,515
IssuesEvent
2015-02-20 20:13:35
DoSomething/dosomething
https://api.github.com/repos/DoSomething/dosomething
closed
client side validation needed for accepted image file types
#campaign-template @fender priority-medium
error only comes after the whole form has been submitted, ideally this error should appear right after the user selects an invalid file: ![image](https://cloud.githubusercontent.com/assets/6330971/6042932/c1a546a0-ac56-11e4-909a-839d9bdfe79a.png) related trello card: https://trello.com/c/dflw2sBj
1.0
client side validation needed for accepted image file types - error only comes after the whole form has been submitted, ideally this error should appear right after the user selects an invalid file: ![image](https://cloud.githubusercontent.com/assets/6330971/6042932/c1a546a0-ac56-11e4-909a-839d9bdfe79a.png) related trello card: https://trello.com/c/dflw2sBj
non_process
client side validation needed for accepted image file types error only comes after the whole form has been submitted ideally this error should appear right after the user selects an invalid file related trello card
0
5,955
8,780,691,094
IssuesEvent
2018-12-19 18:04:56
googleapis/google-cloud-python
https://api.github.com/repos/googleapis/google-cloud-python
opened
BigQuery: system tests fail w/ 429 errors from GCS
api: bigquery testing type: process
BigQuery system tests which to create buckets / blobs need to be hardened against 429 responses from GCS.
1.0
BigQuery: system tests fail w/ 429 errors from GCS - BigQuery system tests which to create buckets / blobs need to be hardened against 429 responses from GCS.
process
bigquery system tests fail w errors from gcs bigquery system tests which to create buckets blobs need to be hardened against responses from gcs
1
259,683
19,610,233,061
IssuesEvent
2022-01-06 14:36:30
ViGEm/HidHide
https://api.github.com/repos/ViGEm/HidHide
closed
Cannot select Flight Simulator 2020 (Microsoft Store Edition)
documentation wontfix
Not sure if this is a bug with the Windows UMP apps but I cannot select Flight Simulator 2020.exe. It will give an error saying I don't have permissions. Tried the new beta Xbox app, which lets you relocate the folder outside the WindowsApp protected one, but the issue still occurs. Open to suggestions!
1.0
Cannot select Flight Simulator 2020 (Microsoft Store Edition) - Not sure if this is a bug with the Windows UMP apps but I cannot select Flight Simulator 2020.exe. It will give an error saying I don't have permissions. Tried the new beta Xbox app, which lets you relocate the folder outside the WindowsApp protected one, but the issue still occurs. Open to suggestions!
non_process
cannot select flight simulator microsoft store edition not sure if this is a bug with the windows ump apps but i cannot select flight simulator exe it will give an error saying i don t have permissions tried the new beta xbox app which lets you relocate the folder outside the windowsapp protected one but the issue still occurs open to suggestions
0
21,830
3,924,038,576
IssuesEvent
2016-04-22 13:54:30
ngageoint/hootenanny-ui
https://api.github.com/repos/ngageoint/hootenanny-ui
closed
Fix and run eslint tests
Category: Test Category: UI Priority: High Status: In Progress Type: Task
Following iD v1.9.2, update and run eslint tests for iD editor/Hoot-UI.
1.0
Fix and run eslint tests - Following iD v1.9.2, update and run eslint tests for iD editor/Hoot-UI.
non_process
fix and run eslint tests following id update and run eslint tests for id editor hoot ui
0
12,505
14,961,670,549
IssuesEvent
2021-01-27 08:09:22
beyondhb1079/s4us
https://api.github.com/repos/beyondhb1079/s4us
closed
About: Job Hernandez
process
Come up with 2-3 sentences about yourself and add yourself to the `team` array in the [About page](https://github.com/beyondhb1079/s4us/blob/main/src/pages/About.jsx#L11).
1.0
About: Job Hernandez - Come up with 2-3 sentences about yourself and add yourself to the `team` array in the [About page](https://github.com/beyondhb1079/s4us/blob/main/src/pages/About.jsx#L11).
process
about job hernandez come up with sentences about yourself and add yourself to the team array in the
1
531,924
15,527,450,774
IssuesEvent
2021-03-13 06:04:51
creativecommons/ccos-scripts
https://api.github.com/repos/creativecommons/ccos-scripts
closed
Update Tooling of the packages
good first issue help wanted ✨ goal: improvement 🟩 priority: low 🤖 aspect: dx
## Problem The Python packages of the repository have older versions of black in `Pipfile` which makes various packages of the repositories.It can be a better approach to optimise these configurations by - Following the Python Community Guidelines (Also see [Python Guidelines — Creative Commons Open Source](https://opensource.creativecommons.org/contributing-code/python-guidelines/) about adding Black and Flake8) ## Solution The configuration could be optimised with that of the configs in the [creativecommons/cc-licenses](https://github.com/creativecommons/cc-licenses/blob/main/Pipfile) ## Resolution <!-- Replace the [ ] with [x] to check the box. --> - [x] I would be interested in resolving this bug.
1.0
Update Tooling of the packages - ## Problem The Python packages of the repository have older versions of black in `Pipfile` which makes various packages of the repositories.It can be a better approach to optimise these configurations by - Following the Python Community Guidelines (Also see [Python Guidelines — Creative Commons Open Source](https://opensource.creativecommons.org/contributing-code/python-guidelines/) about adding Black and Flake8) ## Solution The configuration could be optimised with that of the configs in the [creativecommons/cc-licenses](https://github.com/creativecommons/cc-licenses/blob/main/Pipfile) ## Resolution <!-- Replace the [ ] with [x] to check the box. --> - [x] I would be interested in resolving this bug.
non_process
update tooling of the packages problem the python packages of the repository have older versions of black in pipfile which makes various packages of the repositories it can be a better approach to optimise these configurations by following the python community guidelines also see about adding black and solution the configuration could be optimised with that of the configs in the resolution i would be interested in resolving this bug
0
3,335
3,129,290,821
IssuesEvent
2015-09-09 00:02:37
spring-projects/spring-boot
https://api.github.com/repos/spring-projects/spring-boot
closed
Add checkstyle to build
build
The recent Eclipse update changed the way that code formatting is applied. We could potentially use [this plugin](https://github.com/revelc/formatter-maven-plugin) to automatically format the code from Maven rather than relying on Eclipse. Unfortunately we also currently apply Eclipse cleanup save actions with the Maven plugin doesn't support. If we apply checkstyle to the build we could enforce that the code style matches what would occur via the cleanup actions (adding `this.` etc). This would then allow us to use the Maven plugin to format the code. As an added benefit checkstyle is also likely to help a lot when people contribute code.
1.0
Add checkstyle to build - The recent Eclipse update changed the way that code formatting is applied. We could potentially use [this plugin](https://github.com/revelc/formatter-maven-plugin) to automatically format the code from Maven rather than relying on Eclipse. Unfortunately we also currently apply Eclipse cleanup save actions with the Maven plugin doesn't support. If we apply checkstyle to the build we could enforce that the code style matches what would occur via the cleanup actions (adding `this.` etc). This would then allow us to use the Maven plugin to format the code. As an added benefit checkstyle is also likely to help a lot when people contribute code.
non_process
add checkstyle to build the recent eclipse update changed the way that code formatting is applied we could potentially use to automatically format the code from maven rather than relying on eclipse unfortunately we also currently apply eclipse cleanup save actions with the maven plugin doesn t support if we apply checkstyle to the build we could enforce that the code style matches what would occur via the cleanup actions adding this etc this would then allow us to use the maven plugin to format the code as an added benefit checkstyle is also likely to help a lot when people contribute code
0
591,116
17,795,130,745
IssuesEvent
2021-08-31 21:03:17
magma/magma
https://api.github.com/repos/magma/magma
closed
Intra AGW Mobility
Epic priority: high
Impact: we have seen issues in the field are seeing issues with UEs and this is a required feature by multiple partners but we haven't been able to promote it yet. CBRS at FB also requesting this. This is S1 mobility, intra, within the same AGW, 2 enodes.
1.0
Intra AGW Mobility - Impact: we have seen issues in the field are seeing issues with UEs and this is a required feature by multiple partners but we haven't been able to promote it yet. CBRS at FB also requesting this. This is S1 mobility, intra, within the same AGW, 2 enodes.
non_process
intra agw mobility impact we have seen issues in the field are seeing issues with ues and this is a required feature by multiple partners but we haven t been able to promote it yet cbrs at fb also requesting this this is mobility intra within the same agw enodes
0