Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
79,527
| 28,359,648,139
|
IssuesEvent
|
2023-04-12 09:49:15
|
galasa-dev/projectmanagement
|
https://api.github.com/repos/galasa-dev/projectmanagement
|
opened
|
--class parameter of 'galasactl runs submit' not working
|
defect cli
|
## Describe the bug
The --class parameter of 'galasactl runs submit' should enable you to submit tests in the ecosystem, without first generating a test portfolio, however this does not seem to be working. The documentation specifies:
```
The --class parameter tells the tool which test class to run. The string is in the format of <osgi-bundle-id>/<fully-qualified-java-class>. All the test methods within the class will be run. You can use multiple such flags to test multiple classes.
```
## Steps to reproduce
1. Run `galasactl runs submit --bootstrap $bootstrap --class dev.galasa.inttests/dev.galasa.inttests.core.local.CoreLocalJava11Ubuntu --throttle 1 --poll 10 --progress 1 --log -`
2. The test will come back as UNKNOWN, and you should see this message in the log for the test run itself:
```
12/04/2023 09:44:20.239 ERROR d.g.f.TestRunner - Unable to load the test bundle dev.galasa.inttests
dev.galasa.framework.spi.FrameworkException: Unable to install bundle "dev.galasa.inttests" from OBR repository
at dev.galasa.framework.BundleManagement.loadBundle(BundleManagement.java:55) ~[-1681292658247:?]
at dev.galasa.framework.TestRunner.runTest(TestRunner.java:222) [-1681292658247:?]
at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:?]
at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[?:?]
at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?]
at java.lang.reflect.Method.invoke(Method.java:566) ~[?:?]
at dev.galasa.boot.felix.FelixFramework.runTest(FelixFramework.java:235) [boot.jar:?]
at dev.galasa.boot.Launcher.launch(Launcher.java:168) [boot.jar:?]
at dev.galasa.boot.Launcher.main(Launcher.java:122) [boot.jar:?]
Caused by: dev.galasa.framework.spi.FrameworkException: Unable to locate bundle "dev.galasa.inttests" in OBR repository
at dev.galasa.framework.BundleManagement.loadBundle(BundleManagement.java:50) ~[-1681292658247:?]
... 8 more
dev.galasa.framework.spi.FrameworkException: Unable to install bundle "dev.galasa.inttests" from OBR repository
at dev.galasa.framework.BundleManagement.loadBundle(BundleManagement.java:55) ~[-1681292658247:?]
at dev.galasa.framework.TestRunner.runTest(TestRunner.java:222) [-1681292658247:?]
at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:?]
at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[?:?]
at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?]
at java.lang.reflect.Method.invoke(Method.java:566) ~[?:?]
at dev.galasa.boot.felix.FelixFramework.runTest(FelixFramework.java:235) [boot.jar:?]
at dev.galasa.boot.Launcher.launch(Launcher.java:168) [boot.jar:?]
at dev.galasa.boot.Launcher.main(Launcher.java:122) [boot.jar:?]
Caused by: dev.galasa.framework.spi.FrameworkException: Unable to locate bundle "dev.galasa.inttests" in OBR repository
at dev.galasa.framework.BundleManagement.loadBundle(BundleManagement.java:50) ~[-1681292658247:?]
... 8 more
```
## Expected behavior
The test dev.galasa.inttests/dev.galasa.inttests.core.local.CoreLocalJava11Ubuntu exists in our ecosystem so galasactl should be able to find it and run it without issues.
## Additional context
N/A
|
1.0
|
--class parameter of 'galasactl runs submit' not working - ## Describe the bug
The --class parameter of 'galasactl runs submit' should enable you to submit tests in the ecosystem, without first generating a test portfolio, however this does not seem to be working. The documentation specifies:
```
The --class parameter tells the tool which test class to run. The string is in the format of <osgi-bundle-id>/<fully-qualified-java-class>. All the test methods within the class will be run. You can use multiple such flags to test multiple classes.
```
## Steps to reproduce
1. Run `galasactl runs submit --bootstrap $bootstrap --class dev.galasa.inttests/dev.galasa.inttests.core.local.CoreLocalJava11Ubuntu --throttle 1 --poll 10 --progress 1 --log -`
2. The test will come back as UNKNOWN, and you should see this message in the log for the test run itself:
```
12/04/2023 09:44:20.239 ERROR d.g.f.TestRunner - Unable to load the test bundle dev.galasa.inttests
dev.galasa.framework.spi.FrameworkException: Unable to install bundle "dev.galasa.inttests" from OBR repository
at dev.galasa.framework.BundleManagement.loadBundle(BundleManagement.java:55) ~[-1681292658247:?]
at dev.galasa.framework.TestRunner.runTest(TestRunner.java:222) [-1681292658247:?]
at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:?]
at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[?:?]
at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?]
at java.lang.reflect.Method.invoke(Method.java:566) ~[?:?]
at dev.galasa.boot.felix.FelixFramework.runTest(FelixFramework.java:235) [boot.jar:?]
at dev.galasa.boot.Launcher.launch(Launcher.java:168) [boot.jar:?]
at dev.galasa.boot.Launcher.main(Launcher.java:122) [boot.jar:?]
Caused by: dev.galasa.framework.spi.FrameworkException: Unable to locate bundle "dev.galasa.inttests" in OBR repository
at dev.galasa.framework.BundleManagement.loadBundle(BundleManagement.java:50) ~[-1681292658247:?]
... 8 more
dev.galasa.framework.spi.FrameworkException: Unable to install bundle "dev.galasa.inttests" from OBR repository
at dev.galasa.framework.BundleManagement.loadBundle(BundleManagement.java:55) ~[-1681292658247:?]
at dev.galasa.framework.TestRunner.runTest(TestRunner.java:222) [-1681292658247:?]
at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:?]
at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[?:?]
at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?]
at java.lang.reflect.Method.invoke(Method.java:566) ~[?:?]
at dev.galasa.boot.felix.FelixFramework.runTest(FelixFramework.java:235) [boot.jar:?]
at dev.galasa.boot.Launcher.launch(Launcher.java:168) [boot.jar:?]
at dev.galasa.boot.Launcher.main(Launcher.java:122) [boot.jar:?]
Caused by: dev.galasa.framework.spi.FrameworkException: Unable to locate bundle "dev.galasa.inttests" in OBR repository
at dev.galasa.framework.BundleManagement.loadBundle(BundleManagement.java:50) ~[-1681292658247:?]
... 8 more
```
## Expected behavior
The test dev.galasa.inttests/dev.galasa.inttests.core.local.CoreLocalJava11Ubuntu exists in our ecosystem so galasactl should be able to find it and run it without issues.
## Additional context
N/A
|
non_process
|
class parameter of galasactl runs submit not working describe the bug the class parameter of galasactl runs submit should enable you to submit tests in the ecosystem without first generating a test portfolio however this does not seem to be working the documentation specifies the class parameter tells the tool which test class to run the string is in the format of all the test methods within the class will be run you can use multiple such flags to test multiple classes steps to reproduce run galasactl runs submit bootstrap bootstrap class dev galasa inttests dev galasa inttests core local throttle poll progress log the test will come back as unknown and you should see this message in the log for the test run itself error d g f testrunner unable to load the test bundle dev galasa inttests dev galasa framework spi frameworkexception unable to install bundle dev galasa inttests from obr repository at dev galasa framework bundlemanagement loadbundle bundlemanagement java at dev galasa framework testrunner runtest testrunner java at jdk internal reflect nativemethodaccessorimpl native method at jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at dev galasa boot felix felixframework runtest felixframework java at dev galasa boot launcher launch launcher java at dev galasa boot launcher main launcher java caused by dev galasa framework spi frameworkexception unable to locate bundle dev galasa inttests in obr repository at dev galasa framework bundlemanagement loadbundle bundlemanagement java more dev galasa framework spi frameworkexception unable to install bundle dev galasa inttests from obr repository at dev galasa framework bundlemanagement loadbundle bundlemanagement java at dev galasa framework testrunner runtest testrunner java at jdk internal reflect nativemethodaccessorimpl native method at jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at dev galasa boot felix felixframework runtest felixframework java at dev galasa boot launcher launch launcher java at dev galasa boot launcher main launcher java caused by dev galasa framework spi frameworkexception unable to locate bundle dev galasa inttests in obr repository at dev galasa framework bundlemanagement loadbundle bundlemanagement java more expected behavior the test dev galasa inttests dev galasa inttests core local exists in our ecosystem so galasactl should be able to find it and run it without issues additional context n a
| 0
|
53,358
| 6,311,972,852
|
IssuesEvent
|
2017-07-24 00:23:20
|
JuliaLang/julia
|
https://api.github.com/repos/JuliaLang/julia
|
closed
|
Add test sets to tuple tests
|
Hacktoberfest intro issue test
|
There's actually one example already there! Another well commented file.
|
1.0
|
Add test sets to tuple tests - There's actually one example already there! Another well commented file.
|
non_process
|
add test sets to tuple tests there s actually one example already there another well commented file
| 0
|
21,811
| 30,316,498,787
|
IssuesEvent
|
2023-07-10 15:55:12
|
tdwg/dwc
|
https://api.github.com/repos/tdwg/dwc
|
closed
|
Change term - infraspecificEpithet update usage comments and examples
|
Term - change Class - Taxon non-normative Process - complete
|
## Term change
* Submitter: Ian Engelbrecht @ianengelbrecht
* Efficacy Justification (why is this change necessary?): Clarification for use of this term as well as dwc:scientificName.
* Demand Justification (if the change is semantic in nature, name at least two organizations that independently need this term): This is a non-normative change request, for clarification in the examples for this term. I expect it will clarify usage for many users.
* Stability Justification (what concerns are there that this might affect existing implementations?): There may be cases where the term is being used inconsistently with its intended purpose, the changes proposed will assist in improving consistency.
* Implications for dwciri: namespace (does this change affect a dwciri term version)?: None
Current Term definition: https://dwc.tdwg.org/list/#dwc_infraspecificEpithet
Proposed changes:
* Usage comments (recommendations regarding content, etc., not normative): **In botany, name strings in literature and identifications may have multiple infraspecific ranks. According to the Code (Art. 6.7 & Art. 24.1), valid names only have two epithets, with the lowest rank being the infraspecificEpithet. For example: the infraspecificEpithet in the string "Indigofera charlieriana subsp. sessilis var. scaberrima" is 'scaberrima' and the scientificName is 'Indigofera charlieriana var. scaberrima (Schinz) J.B.Gillett'. Use dwc:verbatimIdentification for the full name string used in an Identification.**
* Examples (not normative): **concolor (for scientificName "Puma concolor concolor (Linnaeus, 1771)"), oxyadenia (for scientificName "Quercus agrifolia var. oxyadenia (Torr.) J.T. Howell"), laxa (for scientificName "Cheilanthes hirta f. laxa (Kunze) W.Jacobsen & N.Jacobsen"), scaberrima (for scientificName "Indigofera charlieriana var. scaberrima (Schinz) J.B.Gillett")**.
Please see discussion at #241
|
1.0
|
Change term - infraspecificEpithet update usage comments and examples - ## Term change
* Submitter: Ian Engelbrecht @ianengelbrecht
* Efficacy Justification (why is this change necessary?): Clarification for use of this term as well as dwc:scientificName.
* Demand Justification (if the change is semantic in nature, name at least two organizations that independently need this term): This is a non-normative change request, for clarification in the examples for this term. I expect it will clarify usage for many users.
* Stability Justification (what concerns are there that this might affect existing implementations?): There may be cases where the term is being used inconsistently with its intended purpose, the changes proposed will assist in improving consistency.
* Implications for dwciri: namespace (does this change affect a dwciri term version)?: None
Current Term definition: https://dwc.tdwg.org/list/#dwc_infraspecificEpithet
Proposed changes:
* Usage comments (recommendations regarding content, etc., not normative): **In botany, name strings in literature and identifications may have multiple infraspecific ranks. According to the Code (Art. 6.7 & Art. 24.1), valid names only have two epithets, with the lowest rank being the infraspecificEpithet. For example: the infraspecificEpithet in the string "Indigofera charlieriana subsp. sessilis var. scaberrima" is 'scaberrima' and the scientificName is 'Indigofera charlieriana var. scaberrima (Schinz) J.B.Gillett'. Use dwc:verbatimIdentification for the full name string used in an Identification.**
* Examples (not normative): **concolor (for scientificName "Puma concolor concolor (Linnaeus, 1771)"), oxyadenia (for scientificName "Quercus agrifolia var. oxyadenia (Torr.) J.T. Howell"), laxa (for scientificName "Cheilanthes hirta f. laxa (Kunze) W.Jacobsen & N.Jacobsen"), scaberrima (for scientificName "Indigofera charlieriana var. scaberrima (Schinz) J.B.Gillett")**.
Please see discussion at #241
|
process
|
change term infraspecificepithet update usage comments and examples term change submitter ian engelbrecht ianengelbrecht efficacy justification why is this change necessary clarification for use of this term as well as dwc scientificname demand justification if the change is semantic in nature name at least two organizations that independently need this term this is a non normative change request for clarification in the examples for this term i expect it will clarify usage for many users stability justification what concerns are there that this might affect existing implementations there may be cases where the term is being used inconsistently with its intended purpose the changes proposed will assist in improving consistency implications for dwciri namespace does this change affect a dwciri term version none current term definition proposed changes usage comments recommendations regarding content etc not normative in botany name strings in literature and identifications may have multiple infraspecific ranks according to the code art art valid names only have two epithets with the lowest rank being the infraspecificepithet for example the infraspecificepithet in the string indigofera charlieriana subsp sessilis var scaberrima is scaberrima and the scientificname is indigofera charlieriana var scaberrima schinz j b gillett use dwc verbatimidentification for the full name string used in an identification examples not normative concolor for scientificname puma concolor concolor linnaeus oxyadenia for scientificname quercus agrifolia var oxyadenia torr j t howell laxa for scientificname cheilanthes hirta f laxa kunze w jacobsen n jacobsen scaberrima for scientificname indigofera charlieriana var scaberrima schinz j b gillett please see discussion at
| 1
|
421,953
| 12,264,088,995
|
IssuesEvent
|
2020-05-07 03:06:40
|
okTurtles/group-income-simple
|
https://api.github.com/repos/okTurtles/group-income-simple
|
opened
|
Cannot add more than one account per payment method
|
Kind:Bug Priority:Low
|
### Problem
When adding information for multiple accounts using the same payment method, only the last account is saved. This is a problem because the app loses information without informing the user.
1. Go to "Income Details", add multiple e.g. "PayPal" payments and click save.
<img width="558" alt="3 PayPal Accounts" src="https://user-images.githubusercontent.com/13082264/81250036-dfdf1800-8fd4-11ea-8d0b-9a47fe0ba0d6.png">
2. No error or alert is shown.
3. Re-open Income Details, to see only the last "PayPal" payment was saved.
<img width="569" alt="1 PayPal Saved" src="https://user-images.githubusercontent.com/13082264/81250078-f38a7e80-8fd4-11ea-93c6-a40253c60023.png">
(Note: This does not affect the "other" payment method.)
### Solution
Two possible solutions:
1. Support adding multiple accounts for the same payment method, because users may have multiple accounts or the platform might support multiple ways of identifying the account (e.g. PayPal email & link).
2. Show the user an error/alert indicating that only one account per method can be saved.
|
1.0
|
Cannot add more than one account per payment method - ### Problem
When adding information for multiple accounts using the same payment method, only the last account is saved. This is a problem because the app loses information without informing the user.
1. Go to "Income Details", add multiple e.g. "PayPal" payments and click save.
<img width="558" alt="3 PayPal Accounts" src="https://user-images.githubusercontent.com/13082264/81250036-dfdf1800-8fd4-11ea-8d0b-9a47fe0ba0d6.png">
2. No error or alert is shown.
3. Re-open Income Details, to see only the last "PayPal" payment was saved.
<img width="569" alt="1 PayPal Saved" src="https://user-images.githubusercontent.com/13082264/81250078-f38a7e80-8fd4-11ea-93c6-a40253c60023.png">
(Note: This does not affect the "other" payment method.)
### Solution
Two possible solutions:
1. Support adding multiple accounts for the same payment method, because users may have multiple accounts or the platform might support multiple ways of identifying the account (e.g. PayPal email & link).
2. Show the user an error/alert indicating that only one account per method can be saved.
|
non_process
|
cannot add more than one account per payment method problem when adding information for multiple accounts using the same payment method only the last account is saved this is a problem because the app loses information without informing the user go to income details add multiple e g paypal payments and click save img width alt paypal accounts src no error or alert is shown re open income details to see only the last paypal payment was saved img width alt paypal saved src note this does not affect the other payment method solution two possible solutions support adding multiple accounts for the same payment method because users may have multiple accounts or the platform might support multiple ways of identifying the account e g paypal email link show the user an error alert indicating that only one account per method can be saved
| 0
|
86,418
| 15,755,653,611
|
IssuesEvent
|
2021-03-31 02:09:37
|
attesch/PrestaShop
|
https://api.github.com/repos/attesch/PrestaShop
|
opened
|
CVE-2017-1000048 (High) detected in qs-2.3.3.tgz
|
security vulnerability
|
## CVE-2017-1000048 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>qs-2.3.3.tgz</b></p></summary>
<p>A querystring parser that supports nesting and arrays, with a depth limit</p>
<p>Library home page: <a href="https://registry.npmjs.org/qs/-/qs-2.3.3.tgz">https://registry.npmjs.org/qs/-/qs-2.3.3.tgz</a></p>
<p>
Dependency Hierarchy:
- webdriverio-3.4.0.tgz (Root Library)
- request-2.49.0.tgz
- :x: **qs-2.3.3.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
the web framework using ljharb's qs module older than v6.3.2, v6.2.3, v6.1.2, and v6.0.4 is vulnerable to a DoS. A malicious user can send a evil request to cause the web framework crash.
<p>Publish Date: 2017-07-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-1000048>CVE-2017-1000048</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-1000048">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-1000048</a></p>
<p>Release Date: 2017-07-17</p>
<p>Fix Resolution: qs - 6.0.4,6.1.2,6.2.3,6.3.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2017-1000048 (High) detected in qs-2.3.3.tgz - ## CVE-2017-1000048 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>qs-2.3.3.tgz</b></p></summary>
<p>A querystring parser that supports nesting and arrays, with a depth limit</p>
<p>Library home page: <a href="https://registry.npmjs.org/qs/-/qs-2.3.3.tgz">https://registry.npmjs.org/qs/-/qs-2.3.3.tgz</a></p>
<p>
Dependency Hierarchy:
- webdriverio-3.4.0.tgz (Root Library)
- request-2.49.0.tgz
- :x: **qs-2.3.3.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
the web framework using ljharb's qs module older than v6.3.2, v6.2.3, v6.1.2, and v6.0.4 is vulnerable to a DoS. A malicious user can send a evil request to cause the web framework crash.
<p>Publish Date: 2017-07-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-1000048>CVE-2017-1000048</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-1000048">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-1000048</a></p>
<p>Release Date: 2017-07-17</p>
<p>Fix Resolution: qs - 6.0.4,6.1.2,6.2.3,6.3.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in qs tgz cve high severity vulnerability vulnerable library qs tgz a querystring parser that supports nesting and arrays with a depth limit library home page a href dependency hierarchy webdriverio tgz root library request tgz x qs tgz vulnerable library vulnerability details the web framework using ljharb s qs module older than and is vulnerable to a dos a malicious user can send a evil request to cause the web framework crash publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution qs step up your open source security game with whitesource
| 0
|
2,953
| 5,948,393,562
|
IssuesEvent
|
2017-05-26 11:13:44
|
DevExpress/testcafe-hammerhead
|
https://api.github.com/repos/DevExpress/testcafe-hammerhead
|
closed
|
Hammerhead crashes if script file contains a sourcemap comment
|
SYSTEM: resource processing TYPE: bug
|
Maybe we need to remove the source map comment from the script.
|
1.0
|
Hammerhead crashes if script file contains a sourcemap comment - Maybe we need to remove the source map comment from the script.
|
process
|
hammerhead crashes if script file contains a sourcemap comment maybe we need to remove the source map comment from the script
| 1
|
247,690
| 20,987,634,287
|
IssuesEvent
|
2022-03-29 06:01:18
|
mozilla-mobile/focus-android
|
https://api.github.com/repos/mozilla-mobile/focus-android
|
closed
|
Intermittent UI test failure - notificationEraseAndOpenButtonTest
|
eng:ui-test eng:intermittent-test
|
### Firebase Test Run:
Flaky on 11/22
https://console.firebase.google.com/project/moz-focus-android/testlab/histories/bh.2189b040bbce6d5a/matrices/9180648678380244643/executions/bs.dacc7ab0ed2b1f57
### Stacktrace:
`androidx.test.uiautomator.UiObjectNotFoundException: UiSelector[DESCRIPTION=Erase and Open]
at androidx.test.uiautomator.UiObject.getBounds(UiObject.java:911)
at androidx.test.uiautomator.UiScrollable.ensureFullyVisible(UiScrollable.java:294)
at org.mozilla.focus.activity.robots.NotificationRobot.expandEraseBrowsingNotification(NotificationRobot.kt:46)
at org.mozilla.focus.activity.EraseBrowsingDataTest$notificationEraseAndOpenButtonTest$4.invoke(EraseBrowsingDataTest.kt:104)
at org.mozilla.focus.activity.EraseBrowsingDataTest$notificationEraseAndOpenButtonTest$4.invoke(EraseBrowsingDataTest.kt:102)
at org.mozilla.focus.activity.robots.NotificationRobotKt.notificationTray(NotificationRobot.kt:97)
at org.mozilla.focus.activity.EraseBrowsingDataTest.notificationEraseAndOpenButtonTest(EraseBrowsingDataTest.kt:102)`
### Build:
11/22 Main
|
2.0
|
Intermittent UI test failure - notificationEraseAndOpenButtonTest - ### Firebase Test Run:
Flaky on 11/22
https://console.firebase.google.com/project/moz-focus-android/testlab/histories/bh.2189b040bbce6d5a/matrices/9180648678380244643/executions/bs.dacc7ab0ed2b1f57
### Stacktrace:
`androidx.test.uiautomator.UiObjectNotFoundException: UiSelector[DESCRIPTION=Erase and Open]
at androidx.test.uiautomator.UiObject.getBounds(UiObject.java:911)
at androidx.test.uiautomator.UiScrollable.ensureFullyVisible(UiScrollable.java:294)
at org.mozilla.focus.activity.robots.NotificationRobot.expandEraseBrowsingNotification(NotificationRobot.kt:46)
at org.mozilla.focus.activity.EraseBrowsingDataTest$notificationEraseAndOpenButtonTest$4.invoke(EraseBrowsingDataTest.kt:104)
at org.mozilla.focus.activity.EraseBrowsingDataTest$notificationEraseAndOpenButtonTest$4.invoke(EraseBrowsingDataTest.kt:102)
at org.mozilla.focus.activity.robots.NotificationRobotKt.notificationTray(NotificationRobot.kt:97)
at org.mozilla.focus.activity.EraseBrowsingDataTest.notificationEraseAndOpenButtonTest(EraseBrowsingDataTest.kt:102)`
### Build:
11/22 Main
|
non_process
|
intermittent ui test failure notificationeraseandopenbuttontest firebase test run flaky on stacktrace androidx test uiautomator uiobjectnotfoundexception uiselector at androidx test uiautomator uiobject getbounds uiobject java at androidx test uiautomator uiscrollable ensurefullyvisible uiscrollable java at org mozilla focus activity robots notificationrobot expanderasebrowsingnotification notificationrobot kt at org mozilla focus activity erasebrowsingdatatest notificationeraseandopenbuttontest invoke erasebrowsingdatatest kt at org mozilla focus activity erasebrowsingdatatest notificationeraseandopenbuttontest invoke erasebrowsingdatatest kt at org mozilla focus activity robots notificationrobotkt notificationtray notificationrobot kt at org mozilla focus activity erasebrowsingdatatest notificationeraseandopenbuttontest erasebrowsingdatatest kt build main
| 0
|
26,472
| 4,726,145,487
|
IssuesEvent
|
2016-10-18 09:17:15
|
PowerDNS/pdns
|
https://api.github.com/repos/PowerDNS/pdns
|
closed
|
DNSSEC root
|
defect rec
|
Hey,
when a client queries a pdns recursor DNSSEC isn't done for the root.
http://pastebin.com/raw/x8zte4C3
another server where this does work: http://pastebin.com/raw/FYKexiDi
|
1.0
|
DNSSEC root - Hey,
when a client queries a pdns recursor DNSSEC isn't done for the root.
http://pastebin.com/raw/x8zte4C3
another server where this does work: http://pastebin.com/raw/FYKexiDi
|
non_process
|
dnssec root hey when a client queries a pdns recursor dnssec isn t done for the root another server where this does work
| 0
|
4,079
| 7,030,904,269
|
IssuesEvent
|
2017-12-26 13:38:11
|
ppetpadriew/turbo-erp-api
|
https://api.github.com/repos/ppetpadriew/turbo-erp-api
|
closed
|
Schema design for warehousing master data
|
in process
|
**Warehousing**
- [x] Warehouse Master Data parameter
- [x] Warehousing order parameter
- [x] Receipt Parameter
- [x] Warehouse Inspection parameter
- [x] Outbound Parameter
- [x] Adjustment Order parameter
- [x] Item Warehouse
- [x] Item Data By Warehouse
- [x] Item Inventory By Warehouse
- [x] Activities
- [x] Procedures
- [x] Activities by Procedure
- [x] Warehouse Order Type
- [x] Label Layout
- [x] Number group
- [x] First Free Number
---
- [x] Move all enum types to table
- [x] Prefix all tables with module code
- [x] Consider move `generate_lots_automatically` in `warehouse order type` to `activities`
|
1.0
|
Schema design for warehousing master data - **Warehousing**
- [x] Warehouse Master Data parameter
- [x] Warehousing order parameter
- [x] Receipt Parameter
- [x] Warehouse Inspection parameter
- [x] Outbound Parameter
- [x] Adjustment Order parameter
- [x] Item Warehouse
- [x] Item Data By Warehouse
- [x] Item Inventory By Warehouse
- [x] Activities
- [x] Procedures
- [x] Activities by Procedure
- [x] Warehouse Order Type
- [x] Label Layout
- [x] Number group
- [x] First Free Number
---
- [x] Move all enum types to table
- [x] Prefix all tables with module code
- [x] Consider move `generate_lots_automatically` in `warehouse order type` to `activities`
|
process
|
schema design for warehousing master data warehousing warehouse master data parameter warehousing order parameter receipt parameter warehouse inspection parameter outbound parameter adjustment order parameter item warehouse item data by warehouse item inventory by warehouse activities procedures activities by procedure warehouse order type label layout number group first free number move all enum types to table prefix all tables with module code consider move generate lots automatically in warehouse order type to activities
| 1
|
2,210
| 5,049,222,215
|
IssuesEvent
|
2016-12-20 15:20:26
|
cfpb/design-manual
|
https://api.github.com/repos/cfpb/design-manual
|
opened
|
12.19.2016 meeting notes
|
process and planning
|
Presentation with current updates and changes can be found here:
https://drive.google.com/open?id=0B8s6MQ9aYdR7SWNuZkhnel9qR1E
Questions/to dos
- Add fewd approval to approval process, currently Jimmy and Scott who are managing
- How will we follow content guidelines? We should bring in content strategists, incorporate them into this process.
- How will pull requests be managed? Who accepts them? Rara has 5 requests out…some of them rather old.
Started backlog grooming, will finish in our working groups this week
https://docs.google.com/document/d/1X_9CLppj8-B5bleasVcgmy22NiP_-D4UgEGzxoWaMeg/edit
|
1.0
|
12.19.2016 meeting notes - Presentation with current updates and changes can be found here:
https://drive.google.com/open?id=0B8s6MQ9aYdR7SWNuZkhnel9qR1E
Questions/to dos
- Add fewd approval to approval process, currently Jimmy and Scott who are managing
- How will we follow content guidelines? We should bring in content strategists, incorporate them into this process.
- How will pull requests be managed? Who accepts them? Rara has 5 requests out…some of them rather old.
Started backlog grooming, will finish in our working groups this week
https://docs.google.com/document/d/1X_9CLppj8-B5bleasVcgmy22NiP_-D4UgEGzxoWaMeg/edit
|
process
|
meeting notes presentation with current updates and changes can be found here questions to dos add fewd approval to approval process currently jimmy and scott who are managing how will we follow content guidelines we should bring in content strategists incorporate them into this process how will pull requests be managed who accepts them rara has requests out…some of them rather old started backlog grooming will finish in our working groups this week
| 1
|
25,638
| 18,944,893,402
|
IssuesEvent
|
2021-11-18 09:07:52
|
ExeClim/Isca
|
https://api.github.com/repos/ExeClim/Isca
|
opened
|
This module file was not generated by any release of this compiler.
|
infrastructure
|
Dear Sir,
we ran into problems while compiling the test-experiment for the first time. We tried reinstalling the model several times, but compilation stops due to the following error: “This module file was not generated by any release of this compiler”. Although we have checked, and netcdf in our server is compiled with this compiler.
I am attaching the failed compilation log file, as well as link to the file emps-gv with our configuration:
[held_suarez_test_case.log](https://github.com/ExeClim/Isca/files/7561233/held_suarez_test_case.log)
https://drive.google.com/file/d/1d5I3F0BdwMu3O3CXtiUseWaDUwt60HHw/view?usp=sharing
To install the model, we used the instructions from the site and did the following:
First of all, we make sure the conda is completely removed! That is, there is no mention in -bash_profile, .bashrc and there is no .conda folder and .condarc file in the home directory. Then we install newest version of it:
```
wget https://repo.anaconda.com/archive/Anaconda3-2021.05-Linux-x86_64.sh
chmod +x Anaconda3-2021.05-Linux-x86_64.sh
# Install into /storage/workstation/zyulyaeva/anaconda3
conda config --set auto_activate_base false
```
Then we install ISCA:
```
conda activate base
git clone https://github.com/ExeClim/Isca
cd Isca
conda env create -f ci/environment-py3.9.yml # устанавливаем окружение
conda activate isca_env № активируем окружение
cd src/extra/python/
pip install -e .
```
We write the paths to the folders in .bashrc, exit the server, log in again and compile test experiment:
```
conda activate isca_env
cd $GFDL_BASE/exp/test_cases/held_suarez
python held_suarez_test_case.py
```
And it stops with the aforementioned error.
Maybe you can tell us what are we doing wrong?
|
1.0
|
This module file was not generated by any release of this compiler. - Dear Sir,
we ran into problems while compiling the test-experiment for the first time. We tried reinstalling the model several times, but compilation stops due to the following error: “This module file was not generated by any release of this compiler”. Although we have checked, and netcdf in our server is compiled with this compiler.
I am attaching the failed compilation log file, as well as link to the file emps-gv with our configuration:
[held_suarez_test_case.log](https://github.com/ExeClim/Isca/files/7561233/held_suarez_test_case.log)
https://drive.google.com/file/d/1d5I3F0BdwMu3O3CXtiUseWaDUwt60HHw/view?usp=sharing
To install the model, we used the instructions from the site and did the following:
First of all, we make sure the conda is completely removed! That is, there is no mention in -bash_profile, .bashrc and there is no .conda folder and .condarc file in the home directory. Then we install newest version of it:
```
wget https://repo.anaconda.com/archive/Anaconda3-2021.05-Linux-x86_64.sh
chmod +x Anaconda3-2021.05-Linux-x86_64.sh
# Install into /storage/workstation/zyulyaeva/anaconda3
conda config --set auto_activate_base false
```
Then we install ISCA:
```
conda activate base
git clone https://github.com/ExeClim/Isca
cd Isca
conda env create -f ci/environment-py3.9.yml # устанавливаем окружение
conda activate isca_env № активируем окружение
cd src/extra/python/
pip install -e .
```
We write the paths to the folders in .bashrc, exit the server, log in again and compile test experiment:
```
conda activate isca_env
cd $GFDL_BASE/exp/test_cases/held_suarez
python held_suarez_test_case.py
```
And it stops with the aforementioned error.
Maybe you can tell us what are we doing wrong?
|
non_process
|
this module file was not generated by any release of this compiler dear sir we ran into problems while compiling the test experiment for the first time we tried reinstalling the model several times but compilation stops due to the following error “this module file was not generated by any release of this compiler” although we have checked and netcdf in our server is compiled with this compiler i am attaching the failed compilation log file as well as link to the file emps gv with our configuration to install the model we used the instructions from the site and did the following first of all we make sure the conda is completely removed that is there is no mention in bash profile bashrc and there is no conda folder and condarc file in the home directory then we install newest version of it wget chmod x linux sh install into storage workstation zyulyaeva conda config set auto activate base false then we install isca conda activate base git clone cd isca conda env create f ci environment yml устанавливаем окружение conda activate isca env № активируем окружение cd src extra python pip install e we write the paths to the folders in bashrc exit the server log in again and compile test experiment conda activate isca env cd gfdl base exp test cases held suarez python held suarez test case py and it stops with the aforementioned error maybe you can tell us what are we doing wrong
| 0
|
2,740
| 5,630,488,378
|
IssuesEvent
|
2017-04-05 12:25:20
|
openvstorage/accelio
|
https://api.github.com/repos/openvstorage/accelio
|
closed
|
Server logs "xio_tcp_management.c:987 [ERROR] - ports mismatch"
|
process_wontfix type_bug
|
Observed on multiple nodes; on certain nodes a burst (consisting of 690 individual messages for example!) of these is logged periodically (every 30 secs - location check interval of edge clients?), while on other nodes this is logged constantly.
|
1.0
|
Server logs "xio_tcp_management.c:987 [ERROR] - ports mismatch" - Observed on multiple nodes; on certain nodes a burst (consisting of 690 individual messages for example!) of these is logged periodically (every 30 secs - location check interval of edge clients?), while on other nodes this is logged constantly.
|
process
|
server logs xio tcp management c ports mismatch observed on multiple nodes on certain nodes a burst consisting of individual messages for example of these is logged periodically every secs location check interval of edge clients while on other nodes this is logged constantly
| 1
|
4,747
| 7,604,097,260
|
IssuesEvent
|
2018-04-29 21:15:59
|
frc1418/VictiScout
|
https://api.github.com/repos/frc1418/VictiScout
|
closed
|
Add strikethrough when hovering over added files (data processing)
|
aesthetic easy processing
|
Add a strikethrough and "clickable" cursor when hovering over an element to show it can be removed upon clicking.
|
1.0
|
Add strikethrough when hovering over added files (data processing) - Add a strikethrough and "clickable" cursor when hovering over an element to show it can be removed upon clicking.
|
process
|
add strikethrough when hovering over added files data processing add a strikethrough and clickable cursor when hovering over an element to show it can be removed upon clicking
| 1
|
22,642
| 31,895,826,622
|
IssuesEvent
|
2023-09-18 01:31:53
|
tdwg/dwc
|
https://api.github.com/repos/tdwg/dwc
|
closed
|
Change term - lithostratigraphicTerms
|
Term - change Class - GeologicalContext normative Task Group - Material Sample Process - complete
|
## Term change
* Submitter: [Material Sample Task Group](https://www.tdwg.org/community/osr/material-sample/)
* Efficacy Justification (why is this change necessary?): Create consistency of terms for material in Darwin Core.
* Demand Justification (if the change is semantic in nature, name at least two organizations that independently need this term): [Material Sample Task Group](https://www.tdwg.org/community/osr/material-sample/), which includes representatives of over 10 organizations.
* Stability Justification (what concerns are there that this might affect existing implementations?): None
* Implications for dwciri: namespace (does this change affect a dwciri term version)?: No
Current Term definition: https://dwc.tdwg.org/list/#dwc_lithostratigraphicTerms
Proposed attributes of the new term version (Please put actual changes to be implemented in **bold** and ~strikethrough~):
* Term name (in lowerCamelCase for properties, UpperCamelCase for classes): lithostratigraphicTerms
* Term label (English, not normative): Lithostratigraphic Terms
* Organized in Class (e.g., Occurrence, Event, Location, Taxon): Geological Context
* Definition of the term (normative): The combination of all litho-stratigraphic names for the rock from which the ~~cataloged item~~**dwc:MaterialEntity** was collected.
* Usage comments (recommendations regarding content, etc., not normative):
* Examples (not normative): Pleistocene-Weichselien
* Refines (identifier of the broader term this term refines; normative): None
* Replaces (identifier of the existing term that would be deprecated and replaced by this term; normative): None
* ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG; not normative): DataSets/DataSet/Units/Unit/Gathering/Stratigraphy/LithostratigraphicTerms
|
1.0
|
Change term - lithostratigraphicTerms - ## Term change
* Submitter: [Material Sample Task Group](https://www.tdwg.org/community/osr/material-sample/)
* Efficacy Justification (why is this change necessary?): Create consistency of terms for material in Darwin Core.
* Demand Justification (if the change is semantic in nature, name at least two organizations that independently need this term): [Material Sample Task Group](https://www.tdwg.org/community/osr/material-sample/), which includes representatives of over 10 organizations.
* Stability Justification (what concerns are there that this might affect existing implementations?): None
* Implications for dwciri: namespace (does this change affect a dwciri term version)?: No
Current Term definition: https://dwc.tdwg.org/list/#dwc_lithostratigraphicTerms
Proposed attributes of the new term version (Please put actual changes to be implemented in **bold** and ~strikethrough~):
* Term name (in lowerCamelCase for properties, UpperCamelCase for classes): lithostratigraphicTerms
* Term label (English, not normative): Lithostratigraphic Terms
* Organized in Class (e.g., Occurrence, Event, Location, Taxon): Geological Context
* Definition of the term (normative): The combination of all litho-stratigraphic names for the rock from which the ~~cataloged item~~**dwc:MaterialEntity** was collected.
* Usage comments (recommendations regarding content, etc., not normative):
* Examples (not normative): Pleistocene-Weichselien
* Refines (identifier of the broader term this term refines; normative): None
* Replaces (identifier of the existing term that would be deprecated and replaced by this term; normative): None
* ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG; not normative): DataSets/DataSet/Units/Unit/Gathering/Stratigraphy/LithostratigraphicTerms
|
process
|
change term lithostratigraphicterms term change submitter efficacy justification why is this change necessary create consistency of terms for material in darwin core demand justification if the change is semantic in nature name at least two organizations that independently need this term which includes representatives of over organizations stability justification what concerns are there that this might affect existing implementations none implications for dwciri namespace does this change affect a dwciri term version no current term definition proposed attributes of the new term version please put actual changes to be implemented in bold and strikethrough term name in lowercamelcase for properties uppercamelcase for classes lithostratigraphicterms term label english not normative lithostratigraphic terms organized in class e g occurrence event location taxon geological context definition of the term normative the combination of all litho stratigraphic names for the rock from which the cataloged item dwc materialentity was collected usage comments recommendations regarding content etc not normative examples not normative pleistocene weichselien refines identifier of the broader term this term refines normative none replaces identifier of the existing term that would be deprecated and replaced by this term normative none abcd xpath of the equivalent term in abcd or efg not normative datasets dataset units unit gathering stratigraphy lithostratigraphicterms
| 1
|
21,197
| 28,214,746,842
|
IssuesEvent
|
2023-04-05 08:07:28
|
googleapis/nodejs-cloud-rad
|
https://api.github.com/repos/googleapis/nodejs-cloud-rad
|
opened
|
Your .repo-metadata.json file has a problem 🤒
|
type: process repo-metadata: lint
|
You have a problem with your .repo-metadata.json file:
Result of scan 📈:
* must have required property 'library_type' in test/fixtures/nodejs-deploy/.repo-metadata.json
* release_level must be equal to one of the allowed values in test/fixtures/nodejs-deploy/.repo-metadata.json
☝️ Once you address these problems, you can close this issue.
### Need help?
* [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field.
* [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**.
* Reach out to **go/github-automation** if you have any questions.
|
1.0
|
Your .repo-metadata.json file has a problem 🤒 - You have a problem with your .repo-metadata.json file:
Result of scan 📈:
* must have required property 'library_type' in test/fixtures/nodejs-deploy/.repo-metadata.json
* release_level must be equal to one of the allowed values in test/fixtures/nodejs-deploy/.repo-metadata.json
☝️ Once you address these problems, you can close this issue.
### Need help?
* [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field.
* [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**.
* Reach out to **go/github-automation** if you have any questions.
|
process
|
your repo metadata json file has a problem 🤒 you have a problem with your repo metadata json file result of scan 📈 must have required property library type in test fixtures nodejs deploy repo metadata json release level must be equal to one of the allowed values in test fixtures nodejs deploy repo metadata json ☝️ once you address these problems you can close this issue need help lists valid options for each field for grpc libraries api shortname should match the subdomain of an api s hostname reach out to go github automation if you have any questions
| 1
|
166,949
| 14,086,580,286
|
IssuesEvent
|
2020-11-05 04:07:40
|
voxel51/fiftyone
|
https://api.github.com/repos/voxel51/fiftyone
|
closed
|
[BUG] Make zoo dataset list table horizontally scrollable
|
bug documentation
|
It would be nice if the output of `fiftyone zoo list` in the docs on file:///Users/Brian/dev/fiftyone/docs/build/html/user_guide/dataset_creation/zoo.html would be horizontally scrollable rather than line-wrapping. With the latter, it is hard to read...
<img width="781" alt="Screen Shot 2020-10-19 at 1 47 51 PM" src="https://user-images.githubusercontent.com/25985824/96492662-fdfea480-1211-11eb-8497-8bc97af6ecf8.png">
|
1.0
|
[BUG] Make zoo dataset list table horizontally scrollable - It would be nice if the output of `fiftyone zoo list` in the docs on file:///Users/Brian/dev/fiftyone/docs/build/html/user_guide/dataset_creation/zoo.html would be horizontally scrollable rather than line-wrapping. With the latter, it is hard to read...
<img width="781" alt="Screen Shot 2020-10-19 at 1 47 51 PM" src="https://user-images.githubusercontent.com/25985824/96492662-fdfea480-1211-11eb-8497-8bc97af6ecf8.png">
|
non_process
|
make zoo dataset list table horizontally scrollable it would be nice if the output of fiftyone zoo list in the docs on file users brian dev fiftyone docs build html user guide dataset creation zoo html would be horizontally scrollable rather than line wrapping with the latter it is hard to read img width alt screen shot at pm src
| 0
|
148,726
| 19,542,549,897
|
IssuesEvent
|
2022-01-01 07:01:59
|
chill-cats/bk_google_calendar
|
https://api.github.com/repos/chill-cats/bk_google_calendar
|
opened
|
CVE-2021-23436 (High) detected in immer-8.0.1.tgz
|
security vulnerability
|
## CVE-2021-23436 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>immer-8.0.1.tgz</b></p></summary>
<p>Create your next immutable state by mutating the current one</p>
<p>Library home page: <a href="https://registry.npmjs.org/immer/-/immer-8.0.1.tgz">https://registry.npmjs.org/immer/-/immer-8.0.1.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/immer/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-4.0.3.tgz (Root Library)
- react-dev-utils-11.0.4.tgz
- :x: **immer-8.0.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/chill-cats/bk_google_calendar/commit/6526b4eff0dbbd951531cb5b10a769d9a8543b8f">6526b4eff0dbbd951531cb5b10a769d9a8543b8f</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
This affects the package immer before 9.0.6. A type confusion vulnerability can lead to a bypass of CVE-2020-28477 when the user-provided keys used in the path parameter are arrays. In particular, this bypass is possible because the condition (p === "__proto__" || p === "constructor") in applyPatches_ returns false if p is ['__proto__'] (or ['constructor']). The === operator (strict equality operator) returns false if the operands have different type.
<p>Publish Date: 2021-09-01
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23436>CVE-2021-23436</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23436">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23436</a></p>
<p>Release Date: 2021-09-01</p>
<p>Fix Resolution: immer - 9.0.6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-23436 (High) detected in immer-8.0.1.tgz - ## CVE-2021-23436 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>immer-8.0.1.tgz</b></p></summary>
<p>Create your next immutable state by mutating the current one</p>
<p>Library home page: <a href="https://registry.npmjs.org/immer/-/immer-8.0.1.tgz">https://registry.npmjs.org/immer/-/immer-8.0.1.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/immer/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-4.0.3.tgz (Root Library)
- react-dev-utils-11.0.4.tgz
- :x: **immer-8.0.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/chill-cats/bk_google_calendar/commit/6526b4eff0dbbd951531cb5b10a769d9a8543b8f">6526b4eff0dbbd951531cb5b10a769d9a8543b8f</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
This affects the package immer before 9.0.6. A type confusion vulnerability can lead to a bypass of CVE-2020-28477 when the user-provided keys used in the path parameter are arrays. In particular, this bypass is possible because the condition (p === "__proto__" || p === "constructor") in applyPatches_ returns false if p is ['__proto__'] (or ['constructor']). The === operator (strict equality operator) returns false if the operands have different type.
<p>Publish Date: 2021-09-01
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23436>CVE-2021-23436</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23436">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23436</a></p>
<p>Release Date: 2021-09-01</p>
<p>Fix Resolution: immer - 9.0.6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in immer tgz cve high severity vulnerability vulnerable library immer tgz create your next immutable state by mutating the current one library home page a href path to dependency file package json path to vulnerable library node modules immer package json dependency hierarchy react scripts tgz root library react dev utils tgz x immer tgz vulnerable library found in head commit a href found in base branch main vulnerability details this affects the package immer before a type confusion vulnerability can lead to a bypass of cve when the user provided keys used in the path parameter are arrays in particular this bypass is possible because the condition p proto p constructor in applypatches returns false if p is or the operator strict equality operator returns false if the operands have different type publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution immer step up your open source security game with whitesource
| 0
|
156,439
| 12,310,358,634
|
IssuesEvent
|
2020-05-12 10:28:41
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
opened
|
roachtest: schemachange/invertedindex failed
|
C-test-failure O-roachtest O-robot branch-release-19.2 release-blocker
|
[(roachtest).schemachange/invertedindex failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=1935848&tab=buildLog) on [release-19.2@a06674c55934425114e99ae2f9751b948d5c1d93](https://github.com/cockroachdb/cockroach/commits/a06674c55934425114e99ae2f9751b948d5c1d93):
```
| 158.0s 0 0.0 336.0 0.0 0.0 0.0 0.0 write
| 159.0s 0 0.0 333.5 0.0 0.0 0.0 0.0 read
| 159.0s 0 0.0 333.9 0.0 0.0 0.0 0.0 write
| 160.0s 0 0.0 331.4 0.0 0.0 0.0 0.0 read
| 160.0s 0 0.0 331.8 0.0 0.0 0.0 0.0 write
| 161.0s 0 8.0 329.4 2.0 12.1 12.1 12.1 read
| 161.0s 0 8.0 329.8 5637.1 6442.5 6442.5 6442.5 write
| 162.0s 0 19.0 327.5 1.6 3.7 8.4 8.4 read
| 162.0s 0 22.0 327.9 9.4 8053.1 8053.1 8053.1 write
| 163.0s 0 28.0 325.7 1.4 3.7 5.8 5.8 read
| 163.0s 0 48.0 326.1 9.4 1677.7 2550.1 2550.1 write
| 164.0s 0 0.0 323.7 0.0 0.0 0.0 0.0 read
| 164.0s 0 3.0 324.2 7.3 12.6 12.6 12.6 write
| 165.0s 0 0.0 321.7 0.0 0.0 0.0 0.0 read
| 165.0s 0 0.0 322.2 0.0 0.0 0.0 0.0 write
Wraps: (5) exit status 30
Error types: (1) *withstack.withStack (2) *safedetails.withSafeDetails (3) *errutil.withMessage (4) *main.withCommandDetails (5) *exec.ExitError
cluster.go:2456,inverted_index.go:104,inverted_index.go:27,test_runner.go:753: monitor failure: monitor task failed: Goexit() was called
(1) attached stack trace
| main.(*monitor).WaitE
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2444
| main.(*monitor).Wait
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2452
| main.runSchemaChangeInvertedIndex
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/inverted_index.go:104
| main.registerSchemaChangeInvertedIndex.func1
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/inverted_index.go:27
| main.(*testRunner).runTest.func2
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/test_runner.go:753
Wraps: (2) monitor failure
Wraps: (3) attached stack trace
| main.(*monitor).wait.func2
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2500
Wraps: (4) monitor task failed
Wraps: (5) attached stack trace
| main.init
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2397
| runtime.doInit
| /usr/local/go/src/runtime/proc.go:5222
| runtime.main
| /usr/local/go/src/runtime/proc.go:190
| runtime.goexit
| /usr/local/go/src/runtime/asm_amd64.s:1357
Wraps: (6) Goexit() was called
Error types: (1) *withstack.withStack (2) *errutil.withMessage (3) *withstack.withStack (4) *errutil.withMessage (5) *withstack.withStack (6) *errors.errorString
Failed to find issue assignee:
couldn't find GitHub commits for user email lucy@cockroachlabs.com
```
<details><summary>More</summary><p>
Artifacts: [/schemachange/invertedindex](https://teamcity.cockroachdb.com/viewLog.html?buildId=1935848&tab=artifacts#/schemachange/invertedindex)
Related:
- #47649 roachtest: schemachange/invertedindex failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202004171855_v20.1.0-rc.2](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202004171855_v20.1.0-rc.2) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #46847 roachtest: schemachange/invertedindex failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-20.1](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-20.1) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #43111 roachtest: schemachange/invertedindex failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-master](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-master)
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Aschemachange%2Finvertedindex.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
|
2.0
|
roachtest: schemachange/invertedindex failed - [(roachtest).schemachange/invertedindex failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=1935848&tab=buildLog) on [release-19.2@a06674c55934425114e99ae2f9751b948d5c1d93](https://github.com/cockroachdb/cockroach/commits/a06674c55934425114e99ae2f9751b948d5c1d93):
```
| 158.0s 0 0.0 336.0 0.0 0.0 0.0 0.0 write
| 159.0s 0 0.0 333.5 0.0 0.0 0.0 0.0 read
| 159.0s 0 0.0 333.9 0.0 0.0 0.0 0.0 write
| 160.0s 0 0.0 331.4 0.0 0.0 0.0 0.0 read
| 160.0s 0 0.0 331.8 0.0 0.0 0.0 0.0 write
| 161.0s 0 8.0 329.4 2.0 12.1 12.1 12.1 read
| 161.0s 0 8.0 329.8 5637.1 6442.5 6442.5 6442.5 write
| 162.0s 0 19.0 327.5 1.6 3.7 8.4 8.4 read
| 162.0s 0 22.0 327.9 9.4 8053.1 8053.1 8053.1 write
| 163.0s 0 28.0 325.7 1.4 3.7 5.8 5.8 read
| 163.0s 0 48.0 326.1 9.4 1677.7 2550.1 2550.1 write
| 164.0s 0 0.0 323.7 0.0 0.0 0.0 0.0 read
| 164.0s 0 3.0 324.2 7.3 12.6 12.6 12.6 write
| 165.0s 0 0.0 321.7 0.0 0.0 0.0 0.0 read
| 165.0s 0 0.0 322.2 0.0 0.0 0.0 0.0 write
Wraps: (5) exit status 30
Error types: (1) *withstack.withStack (2) *safedetails.withSafeDetails (3) *errutil.withMessage (4) *main.withCommandDetails (5) *exec.ExitError
cluster.go:2456,inverted_index.go:104,inverted_index.go:27,test_runner.go:753: monitor failure: monitor task failed: Goexit() was called
(1) attached stack trace
| main.(*monitor).WaitE
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2444
| main.(*monitor).Wait
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2452
| main.runSchemaChangeInvertedIndex
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/inverted_index.go:104
| main.registerSchemaChangeInvertedIndex.func1
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/inverted_index.go:27
| main.(*testRunner).runTest.func2
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/test_runner.go:753
Wraps: (2) monitor failure
Wraps: (3) attached stack trace
| main.(*monitor).wait.func2
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2500
Wraps: (4) monitor task failed
Wraps: (5) attached stack trace
| main.init
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2397
| runtime.doInit
| /usr/local/go/src/runtime/proc.go:5222
| runtime.main
| /usr/local/go/src/runtime/proc.go:190
| runtime.goexit
| /usr/local/go/src/runtime/asm_amd64.s:1357
Wraps: (6) Goexit() was called
Error types: (1) *withstack.withStack (2) *errutil.withMessage (3) *withstack.withStack (4) *errutil.withMessage (5) *withstack.withStack (6) *errors.errorString
Failed to find issue assignee:
couldn't find GitHub commits for user email lucy@cockroachlabs.com
```
<details><summary>More</summary><p>
Artifacts: [/schemachange/invertedindex](https://teamcity.cockroachdb.com/viewLog.html?buildId=1935848&tab=artifacts#/schemachange/invertedindex)
Related:
- #47649 roachtest: schemachange/invertedindex failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202004171855_v20.1.0-rc.2](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202004171855_v20.1.0-rc.2) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #46847 roachtest: schemachange/invertedindex failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-20.1](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-20.1) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #43111 roachtest: schemachange/invertedindex failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-master](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-master)
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Aschemachange%2Finvertedindex.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
|
non_process
|
roachtest schemachange invertedindex failed on write read write read write read write read write read write read write read write wraps exit status error types withstack withstack safedetails withsafedetails errutil withmessage main withcommanddetails exec exiterror cluster go inverted index go inverted index go test runner go monitor failure monitor task failed goexit was called attached stack trace main monitor waite home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go main monitor wait home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go main runschemachangeinvertedindex home agent work go src github com cockroachdb cockroach pkg cmd roachtest inverted index go main registerschemachangeinvertedindex home agent work go src github com cockroachdb cockroach pkg cmd roachtest inverted index go main testrunner runtest home agent work go src github com cockroachdb cockroach pkg cmd roachtest test runner go wraps monitor failure wraps attached stack trace main monitor wait home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go wraps monitor task failed wraps attached stack trace main init home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go runtime doinit usr local go src runtime proc go runtime main usr local go src runtime proc go runtime goexit usr local go src runtime asm s wraps goexit was called error types withstack withstack errutil withmessage withstack withstack errutil withmessage withstack withstack errors errorstring failed to find issue assignee couldn t find github commits for user email lucy cockroachlabs com more artifacts related roachtest schemachange invertedindex failed roachtest schemachange invertedindex failed roachtest schemachange invertedindex failed powered by
| 0
|
2,322
| 4,480,056,334
|
IssuesEvent
|
2016-08-28 01:19:31
|
iosdevelopershq/Foobot
|
https://api.github.com/repos/iosdevelopershq/Foobot
|
opened
|
Announcements
|
New Service
|
We would like a series of ways to broadcast information to our users.
- New user messages: when a new user joins we want Foobot to be able to send them a message
- Timed messages: we would like Foobot to, at a definable interval, send messages to all/certain channels
- One off messages: We would like to have Foobot relay some information to specified users/channels
|
1.0
|
Announcements - We would like a series of ways to broadcast information to our users.
- New user messages: when a new user joins we want Foobot to be able to send them a message
- Timed messages: we would like Foobot to, at a definable interval, send messages to all/certain channels
- One off messages: We would like to have Foobot relay some information to specified users/channels
|
non_process
|
announcements we would like a series of ways to broadcast information to our users new user messages when a new user joins we want foobot to be able to send them a message timed messages we would like foobot to at a definable interval send messages to all certain channels one off messages we would like to have foobot relay some information to specified users channels
| 0
|
14,016
| 16,816,723,091
|
IssuesEvent
|
2021-06-17 08:16:51
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[PM] Responsive issues in Site participant registry
|
Bug P2 Participant manager Process: Fixed Process: Tested QA Process: Tested dev
|
The responsive issues in Site participant registry
1.Multiple buttons are getting displayed to add participant, decommission site where the width is small
2. UI issues

3. UI issue in tables[it should be fixed for 'All, New, Disabled, Invited tab]

|
3.0
|
[PM] Responsive issues in Site participant registry - The responsive issues in Site participant registry
1.Multiple buttons are getting displayed to add participant, decommission site where the width is small
2. UI issues

3. UI issue in tables[it should be fixed for 'All, New, Disabled, Invited tab]

|
process
|
responsive issues in site participant registry the responsive issues in site participant registry multiple buttons are getting displayed to add participant decommission site where the width is small ui issues ui issue in tables
| 1
|
6,567
| 9,653,208,771
|
IssuesEvent
|
2019-05-19 01:30:13
|
NottingHack/hms2
|
https://api.github.com/repos/NottingHack/hms2
|
closed
|
Automatic deployment
|
PR open Process
|
Look into using
https://github.com/orphans/git-deploy-laravel
for auto deployment of hms2
will depend on how we deal with release / develop branches
but staging (roomie) could deploy on an commit to master
and production (holly) could deploy only on tagging of a version
|
1.0
|
Automatic deployment - Look into using
https://github.com/orphans/git-deploy-laravel
for auto deployment of hms2
will depend on how we deal with release / develop branches
but staging (roomie) could deploy on an commit to master
and production (holly) could deploy only on tagging of a version
|
process
|
automatic deployment look into using for auto deployment of will depend on how we deal with release develop branches but staging roomie could deploy on an commit to master and production holly could deploy only on tagging of a version
| 1
|
32,958
| 6,145,724,124
|
IssuesEvent
|
2017-06-27 12:18:10
|
Darksonn/backblaze-b2-rs
|
https://api.github.com/repos/Darksonn/backblaze-b2-rs
|
closed
|
Links in modules are broken in some places
|
documentation
|
In the module level documentation for each module there are links to various structs that use relative urls. Since the doc for the module is embedded on the index, which is in another folder, this causes broken links.
|
1.0
|
Links in modules are broken in some places - In the module level documentation for each module there are links to various structs that use relative urls. Since the doc for the module is embedded on the index, which is in another folder, this causes broken links.
|
non_process
|
links in modules are broken in some places in the module level documentation for each module there are links to various structs that use relative urls since the doc for the module is embedded on the index which is in another folder this causes broken links
| 0
|
26,199
| 11,273,835,316
|
IssuesEvent
|
2020-01-14 17:17:34
|
GooseWSS/mashup-webcomponent
|
https://api.github.com/repos/GooseWSS/mashup-webcomponent
|
opened
|
WS-2019-0019 (Medium) detected in braces-1.8.5.tgz
|
security vulnerability
|
## WS-2019-0019 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>braces-1.8.5.tgz</b></p></summary>
<p>Fastest brace expansion for node.js, with the most complete support for the Bash 4.3 braces specification.</p>
<p>Library home page: <a href="https://registry.npmjs.org/braces/-/braces-1.8.5.tgz">https://registry.npmjs.org/braces/-/braces-1.8.5.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/mashup-webcomponent/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/mashup-webcomponent/node_modules/rollup-watch/node_modules/braces/package.json</p>
<p>
Dependency Hierarchy:
- rollup-watch-4.3.1.tgz (Root Library)
- chokidar-1.7.0.tgz
- anymatch-1.3.2.tgz
- micromatch-2.3.11.tgz
- :x: **braces-1.8.5.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/GooseWSS/mashup-webcomponent/commit/123baa18c9aab6936f0d7b713ce91f5902e2a41b">123baa18c9aab6936f0d7b713ce91f5902e2a41b</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Version of braces prior to 2.3.1 are vulnerable to Regular Expression Denial of Service (ReDoS). Untrusted input may cause catastrophic backtracking while matching regular expressions. This can cause the application to be unresponsive leading to Denial of Service.
<p>Publish Date: 2019-03-25
<p>URL: <a href=https://github.com/micromatch/braces/commit/abdafb0cae1e0c00f184abbadc692f4eaa98f451>WS-2019-0019</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/786">https://www.npmjs.com/advisories/786</a></p>
<p>Release Date: 2019-02-21</p>
<p>Fix Resolution: 2.3.1</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"braces","packageVersion":"1.8.5","isTransitiveDependency":true,"dependencyTree":"rollup-watch:4.3.1;chokidar:1.7.0;anymatch:1.3.2;micromatch:2.3.11;braces:1.8.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.3.1"}],"vulnerabilityIdentifier":"WS-2019-0019","vulnerabilityDetails":"Version of braces prior to 2.3.1 are vulnerable to Regular Expression Denial of Service (ReDoS). Untrusted input may cause catastrophic backtracking while matching regular expressions. This can cause the application to be unresponsive leading to Denial of Service.","vulnerabilityUrl":"https://github.com/micromatch/braces/commit/abdafb0cae1e0c00f184abbadc692f4eaa98f451","cvss2Severity":"medium","cvss2Score":"5.0","extraData":{}}</REMEDIATE> -->
|
True
|
WS-2019-0019 (Medium) detected in braces-1.8.5.tgz - ## WS-2019-0019 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>braces-1.8.5.tgz</b></p></summary>
<p>Fastest brace expansion for node.js, with the most complete support for the Bash 4.3 braces specification.</p>
<p>Library home page: <a href="https://registry.npmjs.org/braces/-/braces-1.8.5.tgz">https://registry.npmjs.org/braces/-/braces-1.8.5.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/mashup-webcomponent/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/mashup-webcomponent/node_modules/rollup-watch/node_modules/braces/package.json</p>
<p>
Dependency Hierarchy:
- rollup-watch-4.3.1.tgz (Root Library)
- chokidar-1.7.0.tgz
- anymatch-1.3.2.tgz
- micromatch-2.3.11.tgz
- :x: **braces-1.8.5.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/GooseWSS/mashup-webcomponent/commit/123baa18c9aab6936f0d7b713ce91f5902e2a41b">123baa18c9aab6936f0d7b713ce91f5902e2a41b</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Version of braces prior to 2.3.1 are vulnerable to Regular Expression Denial of Service (ReDoS). Untrusted input may cause catastrophic backtracking while matching regular expressions. This can cause the application to be unresponsive leading to Denial of Service.
<p>Publish Date: 2019-03-25
<p>URL: <a href=https://github.com/micromatch/braces/commit/abdafb0cae1e0c00f184abbadc692f4eaa98f451>WS-2019-0019</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/786">https://www.npmjs.com/advisories/786</a></p>
<p>Release Date: 2019-02-21</p>
<p>Fix Resolution: 2.3.1</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"braces","packageVersion":"1.8.5","isTransitiveDependency":true,"dependencyTree":"rollup-watch:4.3.1;chokidar:1.7.0;anymatch:1.3.2;micromatch:2.3.11;braces:1.8.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.3.1"}],"vulnerabilityIdentifier":"WS-2019-0019","vulnerabilityDetails":"Version of braces prior to 2.3.1 are vulnerable to Regular Expression Denial of Service (ReDoS). Untrusted input may cause catastrophic backtracking while matching regular expressions. This can cause the application to be unresponsive leading to Denial of Service.","vulnerabilityUrl":"https://github.com/micromatch/braces/commit/abdafb0cae1e0c00f184abbadc692f4eaa98f451","cvss2Severity":"medium","cvss2Score":"5.0","extraData":{}}</REMEDIATE> -->
|
non_process
|
ws medium detected in braces tgz ws medium severity vulnerability vulnerable library braces tgz fastest brace expansion for node js with the most complete support for the bash braces specification library home page a href path to dependency file tmp ws scm mashup webcomponent package json path to vulnerable library tmp ws scm mashup webcomponent node modules rollup watch node modules braces package json dependency hierarchy rollup watch tgz root library chokidar tgz anymatch tgz micromatch tgz x braces tgz vulnerable library found in head commit a href vulnerability details version of braces prior to are vulnerable to regular expression denial of service redos untrusted input may cause catastrophic backtracking while matching regular expressions this can cause the application to be unresponsive leading to denial of service publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier ws vulnerabilitydetails version of braces prior to are vulnerable to regular expression denial of service redos untrusted input may cause catastrophic backtracking while matching regular expressions this can cause the application to be unresponsive leading to denial of service vulnerabilityurl
| 0
|
4,667
| 7,497,316,528
|
IssuesEvent
|
2018-04-08 18:33:36
|
UnbFeelings/unb-feelings-GQA
|
https://api.github.com/repos/UnbFeelings/unb-feelings-GQA
|
reopened
|
Separar Processo de Trabalho (GC) de GQA
|
process wiki
|
- [ ] Separar as imagens do Processo;
- [x] Subir as imagens pra WIki;
|
1.0
|
Separar Processo de Trabalho (GC) de GQA - - [ ] Separar as imagens do Processo;
- [x] Subir as imagens pra WIki;
|
process
|
separar processo de trabalho gc de gqa separar as imagens do processo subir as imagens pra wiki
| 1
|
487,546
| 14,048,276,288
|
IssuesEvent
|
2020-11-02 08:36:35
|
AY2021S1-CS2113T-F12-1/tp
|
https://api.github.com/repos/AY2021S1-CS2113T-F12-1/tp
|
closed
|
[PE-D] User input command too long
|
feature.All priority.Low severity.Low
|
In the main section, there are only 3 available commands, maybe you can shorten the first two commands into 'A' and 'C'. Otherwise, it's quite time-consuming to type the two commands, especially when users need to type again when switching between the academic calendar and caculator.

<!--session: 1604047649081-a81dc1d6-73cc-435e-a5fa-0fbd0a778169-->
-------------
Labels: `severity.Low` `type.FeatureFlaw`
original: Jingming517/ped#5
|
1.0
|
[PE-D] User input command too long - In the main section, there are only 3 available commands, maybe you can shorten the first two commands into 'A' and 'C'. Otherwise, it's quite time-consuming to type the two commands, especially when users need to type again when switching between the academic calendar and caculator.

<!--session: 1604047649081-a81dc1d6-73cc-435e-a5fa-0fbd0a778169-->
-------------
Labels: `severity.Low` `type.FeatureFlaw`
original: Jingming517/ped#5
|
non_process
|
user input command too long in the main section there are only available commands maybe you can shorten the first two commands into a and c otherwise it s quite time consuming to type the two commands especially when users need to type again when switching between the academic calendar and caculator labels severity low type featureflaw original ped
| 0
|
15,764
| 3,974,493,266
|
IssuesEvent
|
2016-05-04 22:26:27
|
RobotLocomotion/drake
|
https://api.github.com/repos/RobotLocomotion/drake
|
closed
|
drake.mit.edu 404s
|
team: kitware type: documentation
|
Hi @RobotLocomotion/kitware,
When opening ``drake.mit.edu`` a 404 is returned. ``dig drake.mit.edu`` shows that it's CNAMEd to ``http://robotlocomotion.github.io/``, which works.
Is it possible that the domain [hasn't been associated properly with GitHub Pages](https://help.github.com/articles/using-a-custom-domain-with-github-pages/)?
- Wolfgang
Related issue: https://github.com/RobotLocomotion/drake/issues/2149
|
1.0
|
drake.mit.edu 404s - Hi @RobotLocomotion/kitware,
When opening ``drake.mit.edu`` a 404 is returned. ``dig drake.mit.edu`` shows that it's CNAMEd to ``http://robotlocomotion.github.io/``, which works.
Is it possible that the domain [hasn't been associated properly with GitHub Pages](https://help.github.com/articles/using-a-custom-domain-with-github-pages/)?
- Wolfgang
Related issue: https://github.com/RobotLocomotion/drake/issues/2149
|
non_process
|
drake mit edu hi robotlocomotion kitware when opening drake mit edu a is returned dig drake mit edu shows that it s cnamed to which works is it possible that the domain wolfgang related issue
| 0
|
662,582
| 22,144,864,994
|
IssuesEvent
|
2022-06-03 10:52:17
|
zephyrproject-rtos/zephyr
|
https://api.github.com/repos/zephyrproject-rtos/zephyr
|
closed
|
Bluetooth: bt_conn: Unable to allocate buffer within timeout
|
bug priority: medium area: Bluetooth
|
**Describe the bug**
A BLE peripheral can sometimes get into a state where it will no longer send notifications. It seems to occur if connections are repeatedly dropped due to range or antenna issues. It may happen if the connection is lost during discovery but that's a guess.
The following errors are generated:
[00:02:21.702,667] <wrn> bt_conn: Disconnected while allocating context
[00:02:29.443,542] <wrn> bt_conn: Unable to allocate buffer within timeout
[00:02:29.443,572] <err> bt_l2cap: Unable to allocate buffer for op 0x12
[00:02:52.495,300] <wrn> bt_conn: Unable to allocate buffer within timeout
[00:02:52.495,300] <err> bt_att: Unable to allocate buffer for op 0x07
Once it gets into this state only a reboot will fix it. A reset on the central side has no effect.
**To Reproduce**
Use the central_hr and peripheral_hr sample apps on the nrf52840dk_nrf52840 boards
west build samples/bluetooth/peripheral_hr --build-dir=./build/peripheral_hr -b nrf52840dk_nrf52840
west build samples/bluetooth/central_hr --build-dir=./build/central_hr -b nrf52840dk_nrf52840
For ease of desktop testing it's convenient to use:
CONFIG_BT_CTLR_TX_PWR_MINUS_40=y
on the central_hr device. This allows a disconnect by moving the boards a short distance apart.
Move the boards closer and farther from each other to trigger disconnect/reconnect events and eventually generate the erroneous state.
**Expected behavior**
A peripheral can cleanly reconnect after a disconnect.
**Impact**
Serious impact since the peripheral will be unable to communicate without a reboot.
**Logs and console output**
***central_hr:***
```
<snip>
<this is what is normal>
Scanning successfully started
[DEVICE]: E6:56:45:48:5D:4A (random), AD evt type 0, AD data len 11, RSSI -41
[AD]: 1 data_len 1
[AD]: 3 data_len 6
Connected: E6:56:45:48:5D:4A (random)
[ATTRIBUTE] handle 25
[ATTRIBUTE] handle 26
[ATTRIBUTE] handle 28
[SUBSCRIBED]
[NOTIFICATION] data 0x20005b9d length 2
[NOTIFICATION] data 0x20005b9d length 2
[NOTIFICATION] data 0x20005b9d length 2
[NOTIFICATION] data 0x20005b9d length 2
[NOTIFICATION] data 0x20005b9d length 2
[NOTIFICATION] data 0x20005b9d length 2
[NOTIFICATION] data 0x20005b9d length 2
[UNSUBSCRIBED]
Disconnected: E6:56:45:48:5D:4A (random) (reason 0x08)
Scanning successfully started
<snip>
<this is what eventually happens>
[DEVICE]: E6:56:45:48:5D:4A (random), AD evt type 0, AD data len 11, RSSI -18
[AD]: 1 data_len 1
[AD]: 3 data_len 6
Connected: E6:56:45:48:5D:4A (random)
Discover complete
[00:04:51.470,947] <err> bt_att: ATT Timeout
Disconnected: E6:56:45:48:5D:4A (random) (reason 0x22)
Scanning successfully started
```
**peripheral_hr:**
```
*** Booting Zephyr OS build zephyr-v3.0.0-1041-g3ab10f2d52a5 ***
Bluetooth initialized
Advertising successfully started
[00:00:00.257,202] <inf> bt_hci_core: HW Platform: Nordic Semiconductor (0x0002)
[00:00:00.257,232] <inf> bt_hci_core: HW Variant: nRF52x (0x0002)
[00:00:00.257,263] <inf> bt_hci_core: Firmware: Standard Bluetooth controller (0x00) Version 3.0 Build 99
[00:00:00.257,995] <inf> bt_hci_core: Identity: E6:56:45:48:5D:4A (random)
[00:00:00.258,026] <inf> bt_hci_core: HCI: version 5.3 (0x0c) revision 0x0000, manufacturer 0x05f1
[00:00:00.258,056] <inf> bt_hci_core: LMP: version 5.3 (0x0c) subver 0xffff
Connected
[00:00:03.341,705] <inf> hrs: HRS notifications enabled
<snip>
Disconnected (reason 0x08)
[00:02:10.680,572] <inf> hrs: HRS notifications disabled
Connected
[00:02:12.349,761] <inf> hrs: HRS notifications enabled
Disconnected (reason 0x08)
Connected
[00:02:21.700,195] <inf> hrs: HRS notifications disabled
[00:02:21.702,667] <wrn> bt_conn: Disconnected while allocating context
[00:02:29.443,542] <wrn> bt_conn: Unable to allocate buffer within timeout
[00:02:29.443,572] <err> bt_l2cap: Unable to allocate buffer for op 0x12
[00:02:52.495,300] <wrn> bt_conn: Unable to allocate buffer within timeout
[00:02:52.495,300] <err> bt_att: Unable to allocate buffer for op 0x07
Disconnected (reason 0x22)
Connected
[00:03:22.495,391] <wrn> bt_conn: Unable to allocate buffer within timeout
[00:03:22.495,422] <err> bt_att: Unable to allocate buffer for op 0x01
[00:03:29.817,321] <wrn> bt_conn: Unable to allocate buffer within timeout
[00:03:29.817,352] <err> bt_l2cap: Unable to allocate buffer for op 0x12
[00:03:52.869,079] <wrn> bt_conn: Unable to allocate buffer within timeout
[00:03:52.869,110] <err> bt_att: Unable to allocate buffer for op 0x07
```
**Environment (please complete the following information):**
- OS: Linux
- Toolchain: Zephyr SDK 13.2
- Commit SHA or Version used Zephyr 3.0: 3ab10f2d52a5f802e6dc8e2472a8563e70de69fc
|
1.0
|
Bluetooth: bt_conn: Unable to allocate buffer within timeout - **Describe the bug**
A BLE peripheral can sometimes get into a state where it will no longer send notifications. It seems to occur if connections are repeatedly dropped due to range or antenna issues. It may happen if the connection is lost during discovery but that's a guess.
The following errors are generated:
[00:02:21.702,667] <wrn> bt_conn: Disconnected while allocating context
[00:02:29.443,542] <wrn> bt_conn: Unable to allocate buffer within timeout
[00:02:29.443,572] <err> bt_l2cap: Unable to allocate buffer for op 0x12
[00:02:52.495,300] <wrn> bt_conn: Unable to allocate buffer within timeout
[00:02:52.495,300] <err> bt_att: Unable to allocate buffer for op 0x07
Once it gets into this state only a reboot will fix it. A reset on the central side has no effect.
**To Reproduce**
Use the central_hr and peripheral_hr sample apps on the nrf52840dk_nrf52840 boards
west build samples/bluetooth/peripheral_hr --build-dir=./build/peripheral_hr -b nrf52840dk_nrf52840
west build samples/bluetooth/central_hr --build-dir=./build/central_hr -b nrf52840dk_nrf52840
For ease of desktop testing it's convenient to use:
CONFIG_BT_CTLR_TX_PWR_MINUS_40=y
on the central_hr device. This allows a disconnect by moving the boards a short distance apart.
Move the boards closer and farther from each other to trigger disconnect/reconnect events and eventually generate the erroneous state.
**Expected behavior**
A peripheral can cleanly reconnect after a disconnect.
**Impact**
Serious impact since the peripheral will be unable to communicate without a reboot.
**Logs and console output**
***central_hr:***
```
<snip>
<this is what is normal>
Scanning successfully started
[DEVICE]: E6:56:45:48:5D:4A (random), AD evt type 0, AD data len 11, RSSI -41
[AD]: 1 data_len 1
[AD]: 3 data_len 6
Connected: E6:56:45:48:5D:4A (random)
[ATTRIBUTE] handle 25
[ATTRIBUTE] handle 26
[ATTRIBUTE] handle 28
[SUBSCRIBED]
[NOTIFICATION] data 0x20005b9d length 2
[NOTIFICATION] data 0x20005b9d length 2
[NOTIFICATION] data 0x20005b9d length 2
[NOTIFICATION] data 0x20005b9d length 2
[NOTIFICATION] data 0x20005b9d length 2
[NOTIFICATION] data 0x20005b9d length 2
[NOTIFICATION] data 0x20005b9d length 2
[UNSUBSCRIBED]
Disconnected: E6:56:45:48:5D:4A (random) (reason 0x08)
Scanning successfully started
<snip>
<this is what eventually happens>
[DEVICE]: E6:56:45:48:5D:4A (random), AD evt type 0, AD data len 11, RSSI -18
[AD]: 1 data_len 1
[AD]: 3 data_len 6
Connected: E6:56:45:48:5D:4A (random)
Discover complete
[00:04:51.470,947] <err> bt_att: ATT Timeout
Disconnected: E6:56:45:48:5D:4A (random) (reason 0x22)
Scanning successfully started
```
**peripheral_hr:**
```
*** Booting Zephyr OS build zephyr-v3.0.0-1041-g3ab10f2d52a5 ***
Bluetooth initialized
Advertising successfully started
[00:00:00.257,202] <inf> bt_hci_core: HW Platform: Nordic Semiconductor (0x0002)
[00:00:00.257,232] <inf> bt_hci_core: HW Variant: nRF52x (0x0002)
[00:00:00.257,263] <inf> bt_hci_core: Firmware: Standard Bluetooth controller (0x00) Version 3.0 Build 99
[00:00:00.257,995] <inf> bt_hci_core: Identity: E6:56:45:48:5D:4A (random)
[00:00:00.258,026] <inf> bt_hci_core: HCI: version 5.3 (0x0c) revision 0x0000, manufacturer 0x05f1
[00:00:00.258,056] <inf> bt_hci_core: LMP: version 5.3 (0x0c) subver 0xffff
Connected
[00:00:03.341,705] <inf> hrs: HRS notifications enabled
<snip>
Disconnected (reason 0x08)
[00:02:10.680,572] <inf> hrs: HRS notifications disabled
Connected
[00:02:12.349,761] <inf> hrs: HRS notifications enabled
Disconnected (reason 0x08)
Connected
[00:02:21.700,195] <inf> hrs: HRS notifications disabled
[00:02:21.702,667] <wrn> bt_conn: Disconnected while allocating context
[00:02:29.443,542] <wrn> bt_conn: Unable to allocate buffer within timeout
[00:02:29.443,572] <err> bt_l2cap: Unable to allocate buffer for op 0x12
[00:02:52.495,300] <wrn> bt_conn: Unable to allocate buffer within timeout
[00:02:52.495,300] <err> bt_att: Unable to allocate buffer for op 0x07
Disconnected (reason 0x22)
Connected
[00:03:22.495,391] <wrn> bt_conn: Unable to allocate buffer within timeout
[00:03:22.495,422] <err> bt_att: Unable to allocate buffer for op 0x01
[00:03:29.817,321] <wrn> bt_conn: Unable to allocate buffer within timeout
[00:03:29.817,352] <err> bt_l2cap: Unable to allocate buffer for op 0x12
[00:03:52.869,079] <wrn> bt_conn: Unable to allocate buffer within timeout
[00:03:52.869,110] <err> bt_att: Unable to allocate buffer for op 0x07
```
**Environment (please complete the following information):**
- OS: Linux
- Toolchain: Zephyr SDK 13.2
- Commit SHA or Version used Zephyr 3.0: 3ab10f2d52a5f802e6dc8e2472a8563e70de69fc
|
non_process
|
bluetooth bt conn unable to allocate buffer within timeout describe the bug a ble peripheral can sometimes get into a state where it will no longer send notifications it seems to occur if connections are repeatedly dropped due to range or antenna issues it may happen if the connection is lost during discovery but that s a guess the following errors are generated bt conn disconnected while allocating context bt conn unable to allocate buffer within timeout bt unable to allocate buffer for op bt conn unable to allocate buffer within timeout bt att unable to allocate buffer for op once it gets into this state only a reboot will fix it a reset on the central side has no effect to reproduce use the central hr and peripheral hr sample apps on the boards west build samples bluetooth peripheral hr build dir build peripheral hr b west build samples bluetooth central hr build dir build central hr b for ease of desktop testing it s convenient to use config bt ctlr tx pwr minus y on the central hr device this allows a disconnect by moving the boards a short distance apart move the boards closer and farther from each other to trigger disconnect reconnect events and eventually generate the erroneous state expected behavior a peripheral can cleanly reconnect after a disconnect impact serious impact since the peripheral will be unable to communicate without a reboot logs and console output central hr scanning successfully started random ad evt type ad data len rssi data len data len connected random handle handle handle data length data length data length data length data length data length data length disconnected random reason scanning successfully started random ad evt type ad data len rssi data len data len connected random discover complete bt att att timeout disconnected random reason scanning successfully started peripheral hr booting zephyr os build zephyr bluetooth initialized advertising successfully started bt hci core hw platform nordic semiconductor bt hci core hw variant bt hci core firmware standard bluetooth controller version build bt hci core identity random bt hci core hci version revision manufacturer bt hci core lmp version subver connected hrs hrs notifications enabled disconnected reason hrs hrs notifications disabled connected hrs hrs notifications enabled disconnected reason connected hrs hrs notifications disabled bt conn disconnected while allocating context bt conn unable to allocate buffer within timeout bt unable to allocate buffer for op bt conn unable to allocate buffer within timeout bt att unable to allocate buffer for op disconnected reason connected bt conn unable to allocate buffer within timeout bt att unable to allocate buffer for op bt conn unable to allocate buffer within timeout bt unable to allocate buffer for op bt conn unable to allocate buffer within timeout bt att unable to allocate buffer for op environment please complete the following information os linux toolchain zephyr sdk commit sha or version used zephyr
| 0
|
138,754
| 31,022,880,763
|
IssuesEvent
|
2023-08-10 07:05:25
|
llvm/llvm-project
|
https://api.github.com/repos/llvm/llvm-project
|
closed
|
llvm-17.0.0-rc1 32-bit win packaging fails to build lldb: "error: cannot compile this forwarded non-trivially copyable parameter yet"
|
clang:codegen release:backport
|
After #64279 I hit the next problem: clang-cl on the release/17.x branch can't build lldb for 32-bit x86:
```
FAILED: tools/lldb/tools/lldb-vscode/CMakeFiles/lldb-vscode.dir/lldb-vscode.cpp.obj
C:\src\llvm_package_17.0.0-rc1\build32_stage0\bin\clang-cl.exe /nologo -TP -DGTEST_HAS_RTTI=0 -DUNICODE -D_CRT_NONSTDC_NO_DEPRECATE -D_CRT_NONSTDC_NO_WARNINGS -D_CRT_SECURE_NO_DEPRECATE -D_CRT_SECURE_NO_WARNINGS -D_ENABLE_EXTENDED_ALIGNED_STORAGE -D_FILE_OFFSET_BITS=64 -D_HAS_EXCEPTIONS=0 -D_LARGEFILE_SOURCE -D_SCL_SECURE_NO_DEPRECATE -D_SCL_SECURE_NO_WARNINGS -D_UNICODE -D__STDC_CONSTANT_MACROS -D__STDC_FORMAT_MACROS -D__STDC_LIMIT_MACROS -Itools\lldb\tools\lldb-vscode -IC:\src\llvm_package_17.0.0-rc1\llvm-project\lldb\tools\lldb-vscode -IC:\src\llvm_package_17.0.0-rc1\llvm-project\lldb\include -Itools\lldb\include -Iinclude -IC:\src\llvm_package_17.0.0-rc1\llvm-project\llvm\include -IC:\Users\hwennborg\AppData\Local\Programs\Python\Python310-32\include -IC:\src\llvm_package_17.0.0-rc1\llvm-project\llvm\..\clang\include -Itools\lldb\..\clang\include -DLIBXML_STATIC /Zc:inline /Zc:__cplusplus /Oi /Brepro /bigobj /permissive- /W4 -Wextra -Wno-unused-parameter -Wwrite-strings -Wcast-qual -Wmissing-field-initializers -Wimplicit-fallthrough -Wcovered-switch-default -Wno-noexcept-type -Wnon-virtual-dtor -Wdelete-non-virtual-dtor -Wsuggest-override -Wstring-conversion -Wmisleading-indentation -Wctad-maybe-unsupported /Gw -Wno-deprecated-declarations -Wno-unknown-pragmas -Wno-strict-aliasing -Wno-deprecated-register -Wno-vla-extension /O2 /Ob2 /DNDEBUG -MD -wd4018 -wd4068 -wd4150 -wd4201 -wd4251 -wd4521 -wd4530 /EHs-c- /GR- -std:c++17 /showIncludes /Fotools\lldb\tools\lldb-vscode\CMakeFiles\lldb-vscode.dir\lldb-vscode.cpp.obj /Fdtools\lldb\tools\lldb-vscode\CMakeFiles\lldb-vscode.dir\ -c -- C:\src\llvm_package_17.0.0-rc1\llvm-project\lldb\tools\lldb-vscode\lldb-vscode.cpp
C:\src\llvm_package_17.0.0-rc1\llvm-project\lldb\tools\lldb-vscode\lldb-vscode.cpp(1610,65): error: cannot compile this forwarded non-trivially copyable parameter yet
1610 | [](llvm::Expected<llvm::json::Value> value) {
| ^
1 error generated.
```
The code was added in 227b2180eb2be94986d63c75c144f88be13fc52f and I believe the error is the long-standing #28673
Luckily, a fix for that landed shortly after the branch, in 27dab4d305acb6e0935e014c061c5317016ae2b3. So maybe we should just cherry-pick that?
|
1.0
|
llvm-17.0.0-rc1 32-bit win packaging fails to build lldb: "error: cannot compile this forwarded non-trivially copyable parameter yet" - After #64279 I hit the next problem: clang-cl on the release/17.x branch can't build lldb for 32-bit x86:
```
FAILED: tools/lldb/tools/lldb-vscode/CMakeFiles/lldb-vscode.dir/lldb-vscode.cpp.obj
C:\src\llvm_package_17.0.0-rc1\build32_stage0\bin\clang-cl.exe /nologo -TP -DGTEST_HAS_RTTI=0 -DUNICODE -D_CRT_NONSTDC_NO_DEPRECATE -D_CRT_NONSTDC_NO_WARNINGS -D_CRT_SECURE_NO_DEPRECATE -D_CRT_SECURE_NO_WARNINGS -D_ENABLE_EXTENDED_ALIGNED_STORAGE -D_FILE_OFFSET_BITS=64 -D_HAS_EXCEPTIONS=0 -D_LARGEFILE_SOURCE -D_SCL_SECURE_NO_DEPRECATE -D_SCL_SECURE_NO_WARNINGS -D_UNICODE -D__STDC_CONSTANT_MACROS -D__STDC_FORMAT_MACROS -D__STDC_LIMIT_MACROS -Itools\lldb\tools\lldb-vscode -IC:\src\llvm_package_17.0.0-rc1\llvm-project\lldb\tools\lldb-vscode -IC:\src\llvm_package_17.0.0-rc1\llvm-project\lldb\include -Itools\lldb\include -Iinclude -IC:\src\llvm_package_17.0.0-rc1\llvm-project\llvm\include -IC:\Users\hwennborg\AppData\Local\Programs\Python\Python310-32\include -IC:\src\llvm_package_17.0.0-rc1\llvm-project\llvm\..\clang\include -Itools\lldb\..\clang\include -DLIBXML_STATIC /Zc:inline /Zc:__cplusplus /Oi /Brepro /bigobj /permissive- /W4 -Wextra -Wno-unused-parameter -Wwrite-strings -Wcast-qual -Wmissing-field-initializers -Wimplicit-fallthrough -Wcovered-switch-default -Wno-noexcept-type -Wnon-virtual-dtor -Wdelete-non-virtual-dtor -Wsuggest-override -Wstring-conversion -Wmisleading-indentation -Wctad-maybe-unsupported /Gw -Wno-deprecated-declarations -Wno-unknown-pragmas -Wno-strict-aliasing -Wno-deprecated-register -Wno-vla-extension /O2 /Ob2 /DNDEBUG -MD -wd4018 -wd4068 -wd4150 -wd4201 -wd4251 -wd4521 -wd4530 /EHs-c- /GR- -std:c++17 /showIncludes /Fotools\lldb\tools\lldb-vscode\CMakeFiles\lldb-vscode.dir\lldb-vscode.cpp.obj /Fdtools\lldb\tools\lldb-vscode\CMakeFiles\lldb-vscode.dir\ -c -- C:\src\llvm_package_17.0.0-rc1\llvm-project\lldb\tools\lldb-vscode\lldb-vscode.cpp
C:\src\llvm_package_17.0.0-rc1\llvm-project\lldb\tools\lldb-vscode\lldb-vscode.cpp(1610,65): error: cannot compile this forwarded non-trivially copyable parameter yet
1610 | [](llvm::Expected<llvm::json::Value> value) {
| ^
1 error generated.
```
The code was added in 227b2180eb2be94986d63c75c144f88be13fc52f and I believe the error is the long-standing #28673
Luckily, a fix for that landed shortly after the branch, in 27dab4d305acb6e0935e014c061c5317016ae2b3. So maybe we should just cherry-pick that?
|
non_process
|
llvm bit win packaging fails to build lldb error cannot compile this forwarded non trivially copyable parameter yet after i hit the next problem clang cl on the release x branch can t build lldb for bit failed tools lldb tools lldb vscode cmakefiles lldb vscode dir lldb vscode cpp obj c src llvm package bin clang cl exe nologo tp dgtest has rtti dunicode d crt nonstdc no deprecate d crt nonstdc no warnings d crt secure no deprecate d crt secure no warnings d enable extended aligned storage d file offset bits d has exceptions d largefile source d scl secure no deprecate d scl secure no warnings d unicode d stdc constant macros d stdc format macros d stdc limit macros itools lldb tools lldb vscode ic src llvm package llvm project lldb tools lldb vscode ic src llvm package llvm project lldb include itools lldb include iinclude ic src llvm package llvm project llvm include ic users hwennborg appdata local programs python include ic src llvm package llvm project llvm clang include itools lldb clang include dlibxml static zc inline zc cplusplus oi brepro bigobj permissive wextra wno unused parameter wwrite strings wcast qual wmissing field initializers wimplicit fallthrough wcovered switch default wno noexcept type wnon virtual dtor wdelete non virtual dtor wsuggest override wstring conversion wmisleading indentation wctad maybe unsupported gw wno deprecated declarations wno unknown pragmas wno strict aliasing wno deprecated register wno vla extension dndebug md ehs c gr std c showincludes fotools lldb tools lldb vscode cmakefiles lldb vscode dir lldb vscode cpp obj fdtools lldb tools lldb vscode cmakefiles lldb vscode dir c c src llvm package llvm project lldb tools lldb vscode lldb vscode cpp c src llvm package llvm project lldb tools lldb vscode lldb vscode cpp error cannot compile this forwarded non trivially copyable parameter yet llvm expected value error generated the code was added in and i believe the error is the long standing luckily a fix for that landed shortly after the branch in so maybe we should just cherry pick that
| 0
|
12,312
| 14,877,103,466
|
IssuesEvent
|
2021-01-20 02:21:53
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
Optional QgsProcessingParameterExpression defaults to field if input has one field
|
Bug Processing
|
<!--
Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone.
If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix
Checklist before submitting
- [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists
- [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles).
- [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue
-->
**Describe the bug**
<!-- A clear and concise description of what the bug is. -->
A processing algorithm that uses an optional `QgsProcessingParameterExpression` defaults its value to input field if the input layer has one field only, instead of being left blank like when there are more fields or no fields.
**How to Reproduce**
Open a point layer that has a single field and load the c++ points to path algorithm. Both order expression and group expression defaults to the single field instead of being blank.
<!-- Steps, sample datasets and qgis project file to reproduce the behavior. Screencasts or screenshots welcome -->
**QGIS and OS versions**
Master on debian testing
<!-- In the QGIS Help menu -> About, click in the table, Ctrl+A and then Ctrl+C. Finally paste here -->
<!-- Add any other context about the problem here. -->
|
1.0
|
Optional QgsProcessingParameterExpression defaults to field if input has one field - <!--
Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone.
If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix
Checklist before submitting
- [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists
- [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles).
- [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue
-->
**Describe the bug**
<!-- A clear and concise description of what the bug is. -->
A processing algorithm that uses an optional `QgsProcessingParameterExpression` defaults its value to input field if the input layer has one field only, instead of being left blank like when there are more fields or no fields.
**How to Reproduce**
Open a point layer that has a single field and load the c++ points to path algorithm. Both order expression and group expression defaults to the single field instead of being blank.
<!-- Steps, sample datasets and qgis project file to reproduce the behavior. Screencasts or screenshots welcome -->
**QGIS and OS versions**
Master on debian testing
<!-- In the QGIS Help menu -> About, click in the table, Ctrl+A and then Ctrl+C. Finally paste here -->
<!-- Add any other context about the problem here. -->
|
process
|
optional qgsprocessingparameterexpression defaults to field if input has one field bug fixing and feature development is a community responsibility and not the responsibility of the qgis project alone if this bug report or feature request is high priority for you we suggest engaging a qgis developer or support organisation and financially sponsoring a fix checklist before submitting search through existing issue reports and gis stackexchange com to check whether the issue already exists test with a create a light and self contained sample dataset and project file which demonstrates the issue describe the bug a processing algorithm that uses an optional qgsprocessingparameterexpression defaults its value to input field if the input layer has one field only instead of being left blank like when there are more fields or no fields how to reproduce open a point layer that has a single field and load the c points to path algorithm both order expression and group expression defaults to the single field instead of being blank qgis and os versions master on debian testing about click in the table ctrl a and then ctrl c finally paste here
| 1
|
252,002
| 21,531,186,562
|
IssuesEvent
|
2022-04-29 01:03:55
|
noamteyssier/pyscreen
|
https://api.github.com/repos/noamteyssier/pyscreen
|
opened
|
Testing Suite for Enrichment Classes
|
good first issue Testing
|
This would include writing up test cases for the enrichment class and measuring that they can be initialized and run.
Could also prove that they are doing what they are supposed to by measuring their results empirically
|
1.0
|
Testing Suite for Enrichment Classes - This would include writing up test cases for the enrichment class and measuring that they can be initialized and run.
Could also prove that they are doing what they are supposed to by measuring their results empirically
|
non_process
|
testing suite for enrichment classes this would include writing up test cases for the enrichment class and measuring that they can be initialized and run could also prove that they are doing what they are supposed to by measuring their results empirically
| 0
|
8,084
| 11,255,571,054
|
IssuesEvent
|
2020-01-12 10:21:02
|
wml-frc/CJ-Vision
|
https://api.github.com/repos/wml-frc/CJ-Vision
|
closed
|
Program Issue when on the TinkerBoard, (Not running, or not connecting to the camera)
|
Camera Display Processing Threading bug
|
This is a major bug, Or it could be a non issue. It really depends if somethings really wrong or not. But i'm unable to get our previous test cases (e.g 2018 vision tracking) deployed and working. So i'm very confused. Could just be tinkerboard. I know these things have many issues.
|
1.0
|
Program Issue when on the TinkerBoard, (Not running, or not connecting to the camera) - This is a major bug, Or it could be a non issue. It really depends if somethings really wrong or not. But i'm unable to get our previous test cases (e.g 2018 vision tracking) deployed and working. So i'm very confused. Could just be tinkerboard. I know these things have many issues.
|
process
|
program issue when on the tinkerboard not running or not connecting to the camera this is a major bug or it could be a non issue it really depends if somethings really wrong or not but i m unable to get our previous test cases e g vision tracking deployed and working so i m very confused could just be tinkerboard i know these things have many issues
| 1
|
121,788
| 17,662,703,317
|
IssuesEvent
|
2021-08-21 21:03:11
|
ghc-dev/Roberto-Frey
|
https://api.github.com/repos/ghc-dev/Roberto-Frey
|
opened
|
CVE-2020-14365 (High) detected in ansible-2.9.9.tar.gz
|
security vulnerability
|
## CVE-2020-14365 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ansible-2.9.9.tar.gz</b></p></summary>
<p>Radically simple IT automation</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/00/5d/e10b83e0e6056dbd5b4809b451a191395175a57e3175ce04e35d9c5fc2a0/ansible-2.9.9.tar.gz">https://files.pythonhosted.org/packages/00/5d/e10b83e0e6056dbd5b4809b451a191395175a57e3175ce04e35d9c5fc2a0/ansible-2.9.9.tar.gz</a></p>
<p>Path to dependency file: Roberto-Frey/requirements.txt</p>
<p>Path to vulnerable library: /requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **ansible-2.9.9.tar.gz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/ghc-dev/Roberto-Frey/commit/8b8494ffdfae810ad106a2571f13708880109f9a">8b8494ffdfae810ad106a2571f13708880109f9a</a></p>
<p>Found in base branch: <b>feature_branch</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A flaw was found in the Ansible Engine, in ansible-engine 2.8.x before 2.8.15 and ansible-engine 2.9.x before 2.9.13, when installing packages using the dnf module. GPG signatures are ignored during installation even when disable_gpg_check is set to False, which is the default behavior. This flaw leads to malicious packages being installed on the system and arbitrary code executed via package installation scripts. The highest threat from this vulnerability is to integrity and system availability.
<p>Publish Date: 2020-09-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-14365>CVE-2020-14365</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=1869154">https://bugzilla.redhat.com/show_bug.cgi?id=1869154</a></p>
<p>Release Date: 2020-07-21</p>
<p>Fix Resolution: 2.8.15,2.9.13</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Python","packageName":"ansible","packageVersion":"2.9.9","packageFilePaths":["/requirements.txt"],"isTransitiveDependency":false,"dependencyTree":"ansible:2.9.9","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.8.15,2.9.13"}],"baseBranches":["feature_branch"],"vulnerabilityIdentifier":"CVE-2020-14365","vulnerabilityDetails":"A flaw was found in the Ansible Engine, in ansible-engine 2.8.x before 2.8.15 and ansible-engine 2.9.x before 2.9.13, when installing packages using the dnf module. GPG signatures are ignored during installation even when disable_gpg_check is set to False, which is the default behavior. This flaw leads to malicious packages being installed on the system and arbitrary code executed via package installation scripts. The highest threat from this vulnerability is to integrity and system availability.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-14365","cvss3Severity":"high","cvss3Score":"7.1","cvss3Metrics":{"A":"High","AC":"Low","PR":"Low","S":"Unchanged","C":"None","UI":"None","AV":"Local","I":"High"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2020-14365 (High) detected in ansible-2.9.9.tar.gz - ## CVE-2020-14365 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ansible-2.9.9.tar.gz</b></p></summary>
<p>Radically simple IT automation</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/00/5d/e10b83e0e6056dbd5b4809b451a191395175a57e3175ce04e35d9c5fc2a0/ansible-2.9.9.tar.gz">https://files.pythonhosted.org/packages/00/5d/e10b83e0e6056dbd5b4809b451a191395175a57e3175ce04e35d9c5fc2a0/ansible-2.9.9.tar.gz</a></p>
<p>Path to dependency file: Roberto-Frey/requirements.txt</p>
<p>Path to vulnerable library: /requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **ansible-2.9.9.tar.gz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/ghc-dev/Roberto-Frey/commit/8b8494ffdfae810ad106a2571f13708880109f9a">8b8494ffdfae810ad106a2571f13708880109f9a</a></p>
<p>Found in base branch: <b>feature_branch</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A flaw was found in the Ansible Engine, in ansible-engine 2.8.x before 2.8.15 and ansible-engine 2.9.x before 2.9.13, when installing packages using the dnf module. GPG signatures are ignored during installation even when disable_gpg_check is set to False, which is the default behavior. This flaw leads to malicious packages being installed on the system and arbitrary code executed via package installation scripts. The highest threat from this vulnerability is to integrity and system availability.
<p>Publish Date: 2020-09-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-14365>CVE-2020-14365</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=1869154">https://bugzilla.redhat.com/show_bug.cgi?id=1869154</a></p>
<p>Release Date: 2020-07-21</p>
<p>Fix Resolution: 2.8.15,2.9.13</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Python","packageName":"ansible","packageVersion":"2.9.9","packageFilePaths":["/requirements.txt"],"isTransitiveDependency":false,"dependencyTree":"ansible:2.9.9","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.8.15,2.9.13"}],"baseBranches":["feature_branch"],"vulnerabilityIdentifier":"CVE-2020-14365","vulnerabilityDetails":"A flaw was found in the Ansible Engine, in ansible-engine 2.8.x before 2.8.15 and ansible-engine 2.9.x before 2.9.13, when installing packages using the dnf module. GPG signatures are ignored during installation even when disable_gpg_check is set to False, which is the default behavior. This flaw leads to malicious packages being installed on the system and arbitrary code executed via package installation scripts. The highest threat from this vulnerability is to integrity and system availability.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-14365","cvss3Severity":"high","cvss3Score":"7.1","cvss3Metrics":{"A":"High","AC":"Low","PR":"Low","S":"Unchanged","C":"None","UI":"None","AV":"Local","I":"High"},"extraData":{}}</REMEDIATE> -->
|
non_process
|
cve high detected in ansible tar gz cve high severity vulnerability vulnerable library ansible tar gz radically simple it automation library home page a href path to dependency file roberto frey requirements txt path to vulnerable library requirements txt dependency hierarchy x ansible tar gz vulnerable library found in head commit a href found in base branch feature branch vulnerability details a flaw was found in the ansible engine in ansible engine x before and ansible engine x before when installing packages using the dnf module gpg signatures are ignored during installation even when disable gpg check is set to false which is the default behavior this flaw leads to malicious packages being installed on the system and arbitrary code executed via package installation scripts the highest threat from this vulnerability is to integrity and system availability publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree ansible isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails a flaw was found in the ansible engine in ansible engine x before and ansible engine x before when installing packages using the dnf module gpg signatures are ignored during installation even when disable gpg check is set to false which is the default behavior this flaw leads to malicious packages being installed on the system and arbitrary code executed via package installation scripts the highest threat from this vulnerability is to integrity and system availability vulnerabilityurl
| 0
|
76,086
| 14,568,262,476
|
IssuesEvent
|
2020-12-17 11:22:34
|
betagouv/mon-entreprise
|
https://api.github.com/repos/betagouv/mon-entreprise
|
opened
|
Utiliser Next.js pour le site publicodes
|
⚙️ publicodes 💭 réflexion 💻 technique
|
Next.js a beaucoup gagné en popularité et permet de gérer des fonctionnalités avancées (server-side rendering, code splitting par page — utile pour ne pas charger Monaco dès la page d’accueil, etc.) sans avoir à configurer WebPack manuellement.
J'avais déjà envisagé la possibilité d'utiliser Next.js pour mon-entreprise.fr mais nous avons quelques parties du code qui seraient difficile à porter (systèmes de traductions et d'iframe notamment).
En revanche le site publi.codes est bien plus simple et sera bientôt dans son propre dépôt (il faudra dissocier sa configuration de "build" de celle de mon-entreprise), il me semble souhaitable et simple à mettre en œuvre d'utiliser Next.js pour ce site. Cela résoudrait #1282 par la même occasion.
|
1.0
|
Utiliser Next.js pour le site publicodes - Next.js a beaucoup gagné en popularité et permet de gérer des fonctionnalités avancées (server-side rendering, code splitting par page — utile pour ne pas charger Monaco dès la page d’accueil, etc.) sans avoir à configurer WebPack manuellement.
J'avais déjà envisagé la possibilité d'utiliser Next.js pour mon-entreprise.fr mais nous avons quelques parties du code qui seraient difficile à porter (systèmes de traductions et d'iframe notamment).
En revanche le site publi.codes est bien plus simple et sera bientôt dans son propre dépôt (il faudra dissocier sa configuration de "build" de celle de mon-entreprise), il me semble souhaitable et simple à mettre en œuvre d'utiliser Next.js pour ce site. Cela résoudrait #1282 par la même occasion.
|
non_process
|
utiliser next js pour le site publicodes next js a beaucoup gagné en popularité et permet de gérer des fonctionnalités avancées server side rendering code splitting par page — utile pour ne pas charger monaco dès la page d’accueil etc sans avoir à configurer webpack manuellement j avais déjà envisagé la possibilité d utiliser next js pour mon entreprise fr mais nous avons quelques parties du code qui seraient difficile à porter systèmes de traductions et d iframe notamment en revanche le site publi codes est bien plus simple et sera bientôt dans son propre dépôt il faudra dissocier sa configuration de build de celle de mon entreprise il me semble souhaitable et simple à mettre en œuvre d utiliser next js pour ce site cela résoudrait par la même occasion
| 0
|
15,016
| 18,727,689,277
|
IssuesEvent
|
2021-11-03 17:57:49
|
microsoft/vscode
|
https://api.github.com/repos/microsoft/vscode
|
closed
|
Broken env on terminal startup in current Insiders version
|
*question *not-reproducible regression confirmation-pending terminal-process
|
<!-- ⚠️⚠️ Do Not Delete This! bug_report_template ⚠️⚠️ -->
<!-- Please read our Rules of Conduct: https://opensource.microsoft.com/codeofconduct/ -->
<!-- 🕮 Read our guide about submitting issues: https://github.com/microsoft/vscode/wiki/Submitting-Bugs-and-Suggestions -->
<!-- 🔎 Search existing issues to avoid creating duplicates. -->
<!-- 🧪 Test using the latest Insiders build to see if your issue has already been fixed: https://code.visualstudio.com/insiders/ -->
<!-- 💡 Instead of creating your report here, use 'Report Issue' from the 'Help' menu in VS Code to pre-fill useful information. -->
<!-- 🔧 Launch with `code --disable-extensions` to check. -->
Does this issue occur when all extensions are disabled?: Yes
<!-- 🪓 If you answered No above, use 'Help: Start Extension Bisect' from Command Palette to try to identify the cause. -->
<!-- 📣 Issues caused by an extension need to be reported directly to the extension publisher. The 'Help > Report Issue' dialog can assist with this. -->
Version: 1.62.0-insider (Universal)
Commit: 4bbec283c36a51cf80f9b77c7a81c140a76a363b
Date: 2021-11-02T08:57:34.388Z
Electron: 13.5.1
Chrome: 91.0.4472.164
Node.js: 14.16.0
V8: 9.1.269.39-electron.0
OS: Darwin arm64 21.1.0
Steps to Reproduce:
1. Have a `.zshenv` file that uses `#!/usr/bin/env bash`
2. Open a terminal in VS Code Insiders
3. See `env: bash: No such file or directory`
It seems when the terminal starts in the current Insiders version, the env seems to be broken/missing important pieces. This appears to be a regression from Stable, works as expected in 1.61.2
|
1.0
|
Broken env on terminal startup in current Insiders version - <!-- ⚠️⚠️ Do Not Delete This! bug_report_template ⚠️⚠️ -->
<!-- Please read our Rules of Conduct: https://opensource.microsoft.com/codeofconduct/ -->
<!-- 🕮 Read our guide about submitting issues: https://github.com/microsoft/vscode/wiki/Submitting-Bugs-and-Suggestions -->
<!-- 🔎 Search existing issues to avoid creating duplicates. -->
<!-- 🧪 Test using the latest Insiders build to see if your issue has already been fixed: https://code.visualstudio.com/insiders/ -->
<!-- 💡 Instead of creating your report here, use 'Report Issue' from the 'Help' menu in VS Code to pre-fill useful information. -->
<!-- 🔧 Launch with `code --disable-extensions` to check. -->
Does this issue occur when all extensions are disabled?: Yes
<!-- 🪓 If you answered No above, use 'Help: Start Extension Bisect' from Command Palette to try to identify the cause. -->
<!-- 📣 Issues caused by an extension need to be reported directly to the extension publisher. The 'Help > Report Issue' dialog can assist with this. -->
Version: 1.62.0-insider (Universal)
Commit: 4bbec283c36a51cf80f9b77c7a81c140a76a363b
Date: 2021-11-02T08:57:34.388Z
Electron: 13.5.1
Chrome: 91.0.4472.164
Node.js: 14.16.0
V8: 9.1.269.39-electron.0
OS: Darwin arm64 21.1.0
Steps to Reproduce:
1. Have a `.zshenv` file that uses `#!/usr/bin/env bash`
2. Open a terminal in VS Code Insiders
3. See `env: bash: No such file or directory`
It seems when the terminal starts in the current Insiders version, the env seems to be broken/missing important pieces. This appears to be a regression from Stable, works as expected in 1.61.2
|
process
|
broken env on terminal startup in current insiders version does this issue occur when all extensions are disabled yes report issue dialog can assist with this version insider universal commit date electron chrome node js electron os darwin steps to reproduce have a zshenv file that uses usr bin env bash open a terminal in vs code insiders see env bash no such file or directory it seems when the terminal starts in the current insiders version the env seems to be broken missing important pieces this appears to be a regression from stable works as expected in
| 1
|
20,510
| 11,456,923,490
|
IssuesEvent
|
2020-02-06 22:19:29
|
dotnet/fsharp
|
https://api.github.com/repos/dotnet/fsharp
|
closed
|
F# projects automatically add 'open' declarations outside of module
|
Area-IDE Language Service Feature Request
|
_This issue has been moved from [a ticket on Developer Community](https://developercommunity.visualstudio.com/content/problem/552356/f-projects-automatically-add-open-declarations-out.html)._
---
When using a namespace as well as a module name, automatically resolving a reference adds the open declaration under the namespace but above the module name.
This didn't happen in VS 2017, and it is undesirable as it obviously opens that reference for the entire namespace rather than just the module you are working in.
---
### Original Comments
#### Visual Studio Feedback System on 5/1/2019, 09:49 PM:
<p>We have directed your feedback to the appropriate engineering team for further evaluation. The team will review the feedback and notify you about the next steps.</p>
#### Ryan Palmer on 8/9/2019, 02:53 AM:
Hi, any news on this? I just got the latest major update and it is still happening. It really is a major pain. Thanks.
---
### Original Solutions
(no solutions)
|
1.0
|
F# projects automatically add 'open' declarations outside of module - _This issue has been moved from [a ticket on Developer Community](https://developercommunity.visualstudio.com/content/problem/552356/f-projects-automatically-add-open-declarations-out.html)._
---
When using a namespace as well as a module name, automatically resolving a reference adds the open declaration under the namespace but above the module name.
This didn't happen in VS 2017, and it is undesirable as it obviously opens that reference for the entire namespace rather than just the module you are working in.
---
### Original Comments
#### Visual Studio Feedback System on 5/1/2019, 09:49 PM:
<p>We have directed your feedback to the appropriate engineering team for further evaluation. The team will review the feedback and notify you about the next steps.</p>
#### Ryan Palmer on 8/9/2019, 02:53 AM:
Hi, any news on this? I just got the latest major update and it is still happening. It really is a major pain. Thanks.
---
### Original Solutions
(no solutions)
|
non_process
|
f projects automatically add open declarations outside of module this issue has been moved from when using a namespace as well as a module name automatically resolving a reference adds the open declaration under the namespace but above the module name this didn t happen in vs and it is undesirable as it obviously opens that reference for the entire namespace rather than just the module you are working in original comments visual studio feedback system on pm we have directed your feedback to the appropriate engineering team for further evaluation the team will review the feedback and notify you about the next steps ryan palmer on am hi any news on this i just got the latest major update and it is still happening it really is a major pain thanks original solutions no solutions
| 0
|
14,890
| 18,289,289,710
|
IssuesEvent
|
2021-10-05 13:42:04
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
PowerShell Runbook information out of date?
|
automation/svc triaged assigned-to-author doc-enhancement process-automation/subsvc Pri2
|
In the section "PowerShell runbooks", some of that information appears to be wrong; perhaps it's out of date?
>Runbooks can't use parallel processing to execute multiple actions in parallel.
I'm doing this; a parent runbook triggers multiple jobs of child runbooks. Perhaps this was intended to apply to runbooks you want to get information back from? Mine are just handing off work to the child runbooks, without getting any information back.
>You can include only PowerShell Workflow runbooks and graphical runbooks as child runbooks by using the Start-AzAutomationRunbook cmdlet, which creates a new job.
I've done this both inline (calling the runbook using "& .\Runbook-Name.ps1") and via the Start-AutomationRunbook internal cmdlet.
>Runbooks can't use the PowerShell #Requires statement, it is not supported in Azure sandbox or on Hybrid Runbook Workers and will cause the job to fail.
I tried to use this and had logged a support ticket with Microsoft; they told me it's not supported as stated, but it didn't cause my jobs to fail, it just ignored them.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 8081200f-2bf4-db58-c957-c8ab7af5f90b
* Version Independent ID: b135cf1a-c391-03e5-41e7-e13571351e91
* Content: [Azure Automation runbook types](https://docs.microsoft.com/en-us/azure/automation/automation-runbook-types#powershell-runbooks)
* Content Source: [articles/automation/automation-runbook-types.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/automation-runbook-types.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @MGoedtel
* Microsoft Alias: **magoedte**
|
1.0
|
PowerShell Runbook information out of date? - In the section "PowerShell runbooks", some of that information appears to be wrong; perhaps it's out of date?
>Runbooks can't use parallel processing to execute multiple actions in parallel.
I'm doing this; a parent runbook triggers multiple jobs of child runbooks. Perhaps this was intended to apply to runbooks you want to get information back from? Mine are just handing off work to the child runbooks, without getting any information back.
>You can include only PowerShell Workflow runbooks and graphical runbooks as child runbooks by using the Start-AzAutomationRunbook cmdlet, which creates a new job.
I've done this both inline (calling the runbook using "& .\Runbook-Name.ps1") and via the Start-AutomationRunbook internal cmdlet.
>Runbooks can't use the PowerShell #Requires statement, it is not supported in Azure sandbox or on Hybrid Runbook Workers and will cause the job to fail.
I tried to use this and had logged a support ticket with Microsoft; they told me it's not supported as stated, but it didn't cause my jobs to fail, it just ignored them.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 8081200f-2bf4-db58-c957-c8ab7af5f90b
* Version Independent ID: b135cf1a-c391-03e5-41e7-e13571351e91
* Content: [Azure Automation runbook types](https://docs.microsoft.com/en-us/azure/automation/automation-runbook-types#powershell-runbooks)
* Content Source: [articles/automation/automation-runbook-types.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/automation-runbook-types.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @MGoedtel
* Microsoft Alias: **magoedte**
|
process
|
powershell runbook information out of date in the section powershell runbooks some of that information appears to be wrong perhaps it s out of date runbooks can t use parallel processing to execute multiple actions in parallel i m doing this a parent runbook triggers multiple jobs of child runbooks perhaps this was intended to apply to runbooks you want to get information back from mine are just handing off work to the child runbooks without getting any information back you can include only powershell workflow runbooks and graphical runbooks as child runbooks by using the start azautomationrunbook cmdlet which creates a new job i ve done this both inline calling the runbook using runbook name and via the start automationrunbook internal cmdlet runbooks can t use the powershell requires statement it is not supported in azure sandbox or on hybrid runbook workers and will cause the job to fail i tried to use this and had logged a support ticket with microsoft they told me it s not supported as stated but it didn t cause my jobs to fail it just ignored them document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login mgoedtel microsoft alias magoedte
| 1
|
20,771
| 27,504,334,159
|
IssuesEvent
|
2023-03-06 01:09:30
|
VolumeFi/paloma
|
https://api.github.com/repos/VolumeFi/paloma
|
opened
|
Add artefact caching on github for Go builds & tests
|
enhancement ReleaseProcess
|
# Background
Currently, when running tests, go needs to download all dependencies. This takes about 1min and 40seconds on our CI.
# Proposed solution
Add [action/cache](https://github.com/actions/cache) in between multiple runs to speed up the tests.
|
1.0
|
Add artefact caching on github for Go builds & tests - # Background
Currently, when running tests, go needs to download all dependencies. This takes about 1min and 40seconds on our CI.
# Proposed solution
Add [action/cache](https://github.com/actions/cache) in between multiple runs to speed up the tests.
|
process
|
add artefact caching on github for go builds tests background currently when running tests go needs to download all dependencies this takes about and on our ci proposed solution add in between multiple runs to speed up the tests
| 1
|
12,856
| 15,240,636,175
|
IssuesEvent
|
2021-02-19 07:04:32
|
pingcap/tidb
|
https://api.github.com/repos/pingcap/tidb
|
closed
|
bit_xor on coprocessor returns wrong result
|
component/coprocessor severity/critical type/bug
|
## Bug Report
Please answer these questions before submitting your issue. Thanks!
### 1. Minimal reproduce step (Required)
<!-- a step by step guide for reproducing the bug. -->
```
CREATE TABLE `city` (`id` int(11) NOT NULL,
`detail` json DEFAULT NULL,
PRIMARY KEY (`id`)
) ENGINE=InnoDB DEFAULT CHARSET=utf8mb4 COLLATE=utf8mb4_bin;
insert into city values(1,'{"population": 100}');
insert into city values(2,'{"1": 0, "name": 1, "population": 100}');
insert into city values(3,'{"1": 0, "name": 1, "population": 100}');
insert into city values(4,'{"1": 0, "name": 2, "population": 100}');
insert into city values(5,'{"1": 0, "name": 2, "population": 200}');
SELECT (CAST(CRC32(CONCAT_WS(',', id,detail, CONCAT(ISNULL(id), ISNULL(detail))))AS UNSIGNED)) AS checksum FROM city;
SELECT BIT_XOR(CAST(CRC32(CONCAT_WS(',', id,detail, CONCAT(ISNULL(id), ISNULL(detail))))AS UNSIGNED)) AS checksum FROM city;
```
### 2. What did you expect to see? (Required)
the correct result is `2188846694` (as is on MySQL), but on v5.0 and v4.0.10, the result is `793141910`.
if we
```
insert into mysql.expr_pushdown_blacklist values ('bit_xor', 'tikv', 'dont pushdown bit_xor to tikv');
admin reload expr_pushdown_blacklist;
```
then tidb will give the correct result.
### 3. What did you see instead (Required)
### 4. What is your TiDB version? (Required)
v4.0.10 and v5.0.0-rc
<!-- Paste the output of SELECT tidb_version() -->
|
1.0
|
bit_xor on coprocessor returns wrong result - ## Bug Report
Please answer these questions before submitting your issue. Thanks!
### 1. Minimal reproduce step (Required)
<!-- a step by step guide for reproducing the bug. -->
```
CREATE TABLE `city` (`id` int(11) NOT NULL,
`detail` json DEFAULT NULL,
PRIMARY KEY (`id`)
) ENGINE=InnoDB DEFAULT CHARSET=utf8mb4 COLLATE=utf8mb4_bin;
insert into city values(1,'{"population": 100}');
insert into city values(2,'{"1": 0, "name": 1, "population": 100}');
insert into city values(3,'{"1": 0, "name": 1, "population": 100}');
insert into city values(4,'{"1": 0, "name": 2, "population": 100}');
insert into city values(5,'{"1": 0, "name": 2, "population": 200}');
SELECT (CAST(CRC32(CONCAT_WS(',', id,detail, CONCAT(ISNULL(id), ISNULL(detail))))AS UNSIGNED)) AS checksum FROM city;
SELECT BIT_XOR(CAST(CRC32(CONCAT_WS(',', id,detail, CONCAT(ISNULL(id), ISNULL(detail))))AS UNSIGNED)) AS checksum FROM city;
```
### 2. What did you expect to see? (Required)
the correct result is `2188846694` (as is on MySQL), but on v5.0 and v4.0.10, the result is `793141910`.
if we
```
insert into mysql.expr_pushdown_blacklist values ('bit_xor', 'tikv', 'dont pushdown bit_xor to tikv');
admin reload expr_pushdown_blacklist;
```
then tidb will give the correct result.
### 3. What did you see instead (Required)
### 4. What is your TiDB version? (Required)
v4.0.10 and v5.0.0-rc
<!-- Paste the output of SELECT tidb_version() -->
|
process
|
bit xor on coprocessor returns wrong result bug report please answer these questions before submitting your issue thanks minimal reproduce step required create table city id int not null detail json default null primary key id engine innodb default charset collate bin insert into city values population insert into city values name population insert into city values name population insert into city values name population insert into city values name population select cast concat ws id detail concat isnull id isnull detail as unsigned as checksum from city select bit xor cast concat ws id detail concat isnull id isnull detail as unsigned as checksum from city what did you expect to see required the correct result is as is on mysql but on and the result is if we insert into mysql expr pushdown blacklist values bit xor tikv dont pushdown bit xor to tikv admin reload expr pushdown blacklist then tidb will give the correct result what did you see instead required what is your tidb version required and rc
| 1
|
8,541
| 11,714,080,930
|
IssuesEvent
|
2020-03-09 11:35:46
|
kazuwjnlab/cvpaper
|
https://api.github.com/repos/kazuwjnlab/cvpaper
|
opened
|
[cvpaper] CVPR2019 #730 A Bayesian Perspective on the Deep Image Prior
|
Gaussian Process bayes deep image prior stochastic gradient Langevin dynamics
|
## \#730 [A Bayesian Perspective on the Deep Image Prior](http://openaccess.thecvf.com/content_CVPR_2019/papers/Cheng_A_Bayesian_Perspective_on_the_Deep_Image_Prior_CVPR_2019_paper.pdf)
Zezhou Cheng, Matheus Gadelha, Subhransu Maji, Daniel Sheldon
### どんな論文か?
畳み込みネットワークの出力をランダムな入力から生成可能にするという、deep image priorの技術が定常なガウス過程に近似することを示し、そこからカーネルを生成。stochastic gradient Langevin dynamicsを用いることで、事後推定を行い、ベイズ的視点で生成が可能であることを示した。
### 新規性
deep image priorにベイズの視点を取り入れることで、事前訓練が必要なく、ランダムにパラメータを初期化可能であるため、1枚の画像から画像のノイズ除去と画像修復ができるようになったことが新規。
### 結果
画像のノイズ除去と画像修復について定性的評価を行い、SGLD法が優れていることを検証。そして、deep image priorとそのガウス過程の事前推定と事後推定を比較した結果、DIPカーネルを持つGPが最も効率が良いことがわかった。SGLD法により、DIPはよりロバストで効率的になった。
### その他(なぜ通ったか?等)
ベイズ推定を手法として使っている論文が少なく、Deep Image Priorに使うということが、新規性が高かった。

Bayes, Deep Image Prior, stochastic gradient Langevin dynamics, Gaussian Process,
|
1.0
|
[cvpaper] CVPR2019 #730 A Bayesian Perspective on the Deep Image Prior - ## \#730 [A Bayesian Perspective on the Deep Image Prior](http://openaccess.thecvf.com/content_CVPR_2019/papers/Cheng_A_Bayesian_Perspective_on_the_Deep_Image_Prior_CVPR_2019_paper.pdf)
Zezhou Cheng, Matheus Gadelha, Subhransu Maji, Daniel Sheldon
### どんな論文か?
畳み込みネットワークの出力をランダムな入力から生成可能にするという、deep image priorの技術が定常なガウス過程に近似することを示し、そこからカーネルを生成。stochastic gradient Langevin dynamicsを用いることで、事後推定を行い、ベイズ的視点で生成が可能であることを示した。
### 新規性
deep image priorにベイズの視点を取り入れることで、事前訓練が必要なく、ランダムにパラメータを初期化可能であるため、1枚の画像から画像のノイズ除去と画像修復ができるようになったことが新規。
### 結果
画像のノイズ除去と画像修復について定性的評価を行い、SGLD法が優れていることを検証。そして、deep image priorとそのガウス過程の事前推定と事後推定を比較した結果、DIPカーネルを持つGPが最も効率が良いことがわかった。SGLD法により、DIPはよりロバストで効率的になった。
### その他(なぜ通ったか?等)
ベイズ推定を手法として使っている論文が少なく、Deep Image Priorに使うということが、新規性が高かった。

Bayes, Deep Image Prior, stochastic gradient Langevin dynamics, Gaussian Process,
|
process
|
a bayesian perspective on the deep image prior zezhou cheng matheus gadelha subhransu maji daniel sheldon どんな論文か? 畳み込みネットワークの出力をランダムな入力から生成可能にするという、deep image priorの技術が定常なガウス過程に近似することを示し、そこからカーネルを生成。stochastic gradient langevin dynamicsを用いることで、事後推定を行い、ベイズ的視点で生成が可能であることを示した。 新規性 deep image priorにベイズの視点を取り入れることで、事前訓練が必要なく、ランダムにパラメータを初期化可能であるため、 。 結果 画像のノイズ除去と画像修復について定性的評価を行い、sgld法が優れていることを検証。そして、deep image priorとそのガウス過程の事前推定と事後推定を比較した結果、dipカーネルを持つgpが最も効率が良いことがわかった。sgld法により、dipはよりロバストで効率的になった。 その他(なぜ通ったか?等) ベイズ推定を手法として使っている論文が少なく、deep image priorに使うということが、新規性が高かった。 bayes deep image prior stochastic gradient langevin dynamics gaussian process
| 1
|
159,613
| 13,768,515,656
|
IssuesEvent
|
2020-10-07 17:12:03
|
fga-eps-mds/2020-1-Ziguen
|
https://api.github.com/repos/fga-eps-mds/2020-1-Ziguen
|
closed
|
Estudar Gitflow
|
Estudo documentation
|
Entender quais como funciona o gitflow e entender os principais comandos do gitflow :P
- [x] Procurar boas referencias de tutoriais
- [x] Documentar o padrão gitflow a ser adotado pelo time
# Critério de Aceitação
- [x] A documentação estar disponível no repositório do time e revisada
|
1.0
|
Estudar Gitflow - Entender quais como funciona o gitflow e entender os principais comandos do gitflow :P
- [x] Procurar boas referencias de tutoriais
- [x] Documentar o padrão gitflow a ser adotado pelo time
# Critério de Aceitação
- [x] A documentação estar disponível no repositório do time e revisada
|
non_process
|
estudar gitflow entender quais como funciona o gitflow e entender os principais comandos do gitflow p procurar boas referencias de tutoriais documentar o padrão gitflow a ser adotado pelo time critério de aceitação a documentação estar disponível no repositório do time e revisada
| 0
|
58,742
| 3,090,896,310
|
IssuesEvent
|
2015-08-26 09:45:49
|
teracyhq/flask-boilerplate
|
https://api.github.com/repos/teracyhq/flask-boilerplate
|
opened
|
add support for @extract_args to avoid boilerplate
|
priority:major type:feature
|
from:
```
@route('', methods=['GET'])
@token_auth_required()
@permissions_required(admin_role_permission)
@marshal_with(_user_list_schema)
@paginated
@use_args(search_args) # TODO(hoate): add support for @extract_args to avoid boilerplate
def index(self, args):
filters, args = extract_filters(args)
args['filters'] = filters
return auth_datastore.find_users(**args), args
```
to:
```
@route('', methods=['GET'])
@token_auth_required()
@permissions_required(admin_role_permission)
@marshal_with(_user_list_schema)
@paginated
@extract_args(search_args)
def index(self, args):
return auth_datastore.find_users(**args), args
```
|
1.0
|
add support for @extract_args to avoid boilerplate - from:
```
@route('', methods=['GET'])
@token_auth_required()
@permissions_required(admin_role_permission)
@marshal_with(_user_list_schema)
@paginated
@use_args(search_args) # TODO(hoate): add support for @extract_args to avoid boilerplate
def index(self, args):
filters, args = extract_filters(args)
args['filters'] = filters
return auth_datastore.find_users(**args), args
```
to:
```
@route('', methods=['GET'])
@token_auth_required()
@permissions_required(admin_role_permission)
@marshal_with(_user_list_schema)
@paginated
@extract_args(search_args)
def index(self, args):
return auth_datastore.find_users(**args), args
```
|
non_process
|
add support for extract args to avoid boilerplate from route methods token auth required permissions required admin role permission marshal with user list schema paginated use args search args todo hoate add support for extract args to avoid boilerplate def index self args filters args extract filters args args filters return auth datastore find users args args to route methods token auth required permissions required admin role permission marshal with user list schema paginated extract args search args def index self args return auth datastore find users args args
| 0
|
10,253
| 13,107,665,652
|
IssuesEvent
|
2020-08-04 15:35:45
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
QGIS 3.12 Merge Vector Layers Box behind main screen.
|
Bug Feedback MacOS Processing
|
<!--
Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone.
If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix
Checklist before submitting
- [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists
- [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles).
- [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue
-->
**Describe the bug**
<!-- A clear and concise description of what the bug is. -->
**How to Reproduce**
<!-- Steps, sample datasets and qgis project file to reproduce the behavior. Screencasts or screenshots welcome
1. Go to '...'
2. Click on '....'
3. Scroll down to '....'
4. See error -->
**QGIS and OS versions**
<!-- In the QGIS Help menu -> About, click in the table, Ctrl+A and then Ctrl+C. Finally paste here -->
**Additional context**
<!-- Add any other context about the problem here. -->
|
1.0
|
QGIS 3.12 Merge Vector Layers Box behind main screen. - <!--
Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone.
If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix
Checklist before submitting
- [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists
- [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles).
- [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue
-->
**Describe the bug**
<!-- A clear and concise description of what the bug is. -->
**How to Reproduce**
<!-- Steps, sample datasets and qgis project file to reproduce the behavior. Screencasts or screenshots welcome
1. Go to '...'
2. Click on '....'
3. Scroll down to '....'
4. See error -->
**QGIS and OS versions**
<!-- In the QGIS Help menu -> About, click in the table, Ctrl+A and then Ctrl+C. Finally paste here -->
**Additional context**
<!-- Add any other context about the problem here. -->
|
process
|
qgis merge vector layers box behind main screen bug fixing and feature development is a community responsibility and not the responsibility of the qgis project alone if this bug report or feature request is high priority for you we suggest engaging a qgis developer or support organisation and financially sponsoring a fix checklist before submitting search through existing issue reports and gis stackexchange com to check whether the issue already exists test with a create a light and self contained sample dataset and project file which demonstrates the issue describe the bug how to reproduce steps sample datasets and qgis project file to reproduce the behavior screencasts or screenshots welcome go to click on scroll down to see error qgis and os versions about click in the table ctrl a and then ctrl c finally paste here additional context
| 1
|
9,863
| 12,877,162,002
|
IssuesEvent
|
2020-07-11 09:25:10
|
dita-ot/dita-ot
|
https://api.github.com/repos/dita-ot/dita-ot
|
closed
|
Null Exception Processing Referenced Subject Scheme Map
|
bug preprocess stale
|
When building output from a map that references a subjectScheme map, the subjectScheme map is not available for processing at the debug-filter stage, generating a null exception.
Tested with DITA-OT 2.2.4, 2.2.5, 2.3, html5 transtype. Tested when the base map is a standard map and a classification map. Same results with all tests.
Attached are sample build files with the log file and temp directory.
[admin-build-new.zip](https://github.com/dita-ot/dita-ot/files/291497/admin-build-new.zip)
|
1.0
|
Null Exception Processing Referenced Subject Scheme Map - When building output from a map that references a subjectScheme map, the subjectScheme map is not available for processing at the debug-filter stage, generating a null exception.
Tested with DITA-OT 2.2.4, 2.2.5, 2.3, html5 transtype. Tested when the base map is a standard map and a classification map. Same results with all tests.
Attached are sample build files with the log file and temp directory.
[admin-build-new.zip](https://github.com/dita-ot/dita-ot/files/291497/admin-build-new.zip)
|
process
|
null exception processing referenced subject scheme map when building output from a map that references a subjectscheme map the subjectscheme map is not available for processing at the debug filter stage generating a null exception tested with dita ot transtype tested when the base map is a standard map and a classification map same results with all tests attached are sample build files with the log file and temp directory
| 1
|
10,807
| 9,105,095,591
|
IssuesEvent
|
2019-02-20 19:52:15
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
ExecutionContext moved or removed with newer SDK version
|
app-service-web/svc cxp product-question triaged
|
Under **Binding types** > the **UseCore** section speaks of ExecutionContext which is bound to a dead github link.
Manually navigating the SDK repository gives off the impression it has been restructured (compared to the link) thus I'm guessing this property has been moved (or perhaps removed)?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: a4500b14-64b4-4d1b-aae3-f0fd4dde85f2
* Version Independent ID: 8fb75741-37b9-b45f-c0ca-9948ddaf1c06
* Content: [How to use the Azure WebJobs SDK](https://docs.microsoft.com/en-us/azure/app-service/webjobs-sdk-how-to#binding-types)
* Content Source: [articles/app-service/webjobs-sdk-how-to.md](https://github.com/Microsoft/azure-docs/blob/master/articles/app-service/webjobs-sdk-how-to.md)
* Service: **app-service-web**
* GitHub Login: @ggailey777
* Microsoft Alias: **glenga**
|
1.0
|
ExecutionContext moved or removed with newer SDK version - Under **Binding types** > the **UseCore** section speaks of ExecutionContext which is bound to a dead github link.
Manually navigating the SDK repository gives off the impression it has been restructured (compared to the link) thus I'm guessing this property has been moved (or perhaps removed)?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: a4500b14-64b4-4d1b-aae3-f0fd4dde85f2
* Version Independent ID: 8fb75741-37b9-b45f-c0ca-9948ddaf1c06
* Content: [How to use the Azure WebJobs SDK](https://docs.microsoft.com/en-us/azure/app-service/webjobs-sdk-how-to#binding-types)
* Content Source: [articles/app-service/webjobs-sdk-how-to.md](https://github.com/Microsoft/azure-docs/blob/master/articles/app-service/webjobs-sdk-how-to.md)
* Service: **app-service-web**
* GitHub Login: @ggailey777
* Microsoft Alias: **glenga**
|
non_process
|
executioncontext moved or removed with newer sdk version under binding types gt the usecore section speaks of executioncontext which is bound to a dead github link manually navigating the sdk repository gives off the impression it has been restructured compared to the link thus i m guessing this property has been moved or perhaps removed document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service app service web github login microsoft alias glenga
| 0
|
98,430
| 20,731,751,026
|
IssuesEvent
|
2022-03-14 10:05:04
|
dotnet/runtime
|
https://api.github.com/repos/dotnet/runtime
|
closed
|
Review workaround for TailCallStress mode
|
area-CodeGen-coreclr in-pr
|
[Cleanup Issue-URLs in Code · Issue #63902 · dotnet/runtime](https://github.com/dotnet/runtime/issues/63902) identified a workaround in place where it's unclear if the backing issue that necessitated the workaround has been resolved.
- The original issue was: [JIT\opt\OSR\tailrecursetry\tailrecursetry.cmd fails with COMPlus_FastTailCalls=0 · Issue #35687 · dotnet/runtime](https://github.com/dotnet/runtime/issues/35687)
- The PR that applied the workaround was: [Fix TailCallStress mode. by erozenfeld · Pull Request #40698 · dotnet/runtime](https://github.com/dotnet/runtime/pull/40698)
https://github.com/dotnet/runtime/blob/f04a24249835096eea1a1a66e4af03cfec5ed32b/src/tests/JIT/opt/OSR/tailrecursetry.csproj#L5-L9
The workaround should be evaluated to determine if it's still necessary. With the findings, we should either update the code to remove the workaround or update the code comment to reference an open issue for tracking when we can remove the workaround.
/cc @deeprobin
|
1.0
|
Review workaround for TailCallStress mode - [Cleanup Issue-URLs in Code · Issue #63902 · dotnet/runtime](https://github.com/dotnet/runtime/issues/63902) identified a workaround in place where it's unclear if the backing issue that necessitated the workaround has been resolved.
- The original issue was: [JIT\opt\OSR\tailrecursetry\tailrecursetry.cmd fails with COMPlus_FastTailCalls=0 · Issue #35687 · dotnet/runtime](https://github.com/dotnet/runtime/issues/35687)
- The PR that applied the workaround was: [Fix TailCallStress mode. by erozenfeld · Pull Request #40698 · dotnet/runtime](https://github.com/dotnet/runtime/pull/40698)
https://github.com/dotnet/runtime/blob/f04a24249835096eea1a1a66e4af03cfec5ed32b/src/tests/JIT/opt/OSR/tailrecursetry.csproj#L5-L9
The workaround should be evaluated to determine if it's still necessary. With the findings, we should either update the code to remove the workaround or update the code comment to reference an open issue for tracking when we can remove the workaround.
/cc @deeprobin
|
non_process
|
review workaround for tailcallstress mode identified a workaround in place where it s unclear if the backing issue that necessitated the workaround has been resolved the original issue was the pr that applied the workaround was the workaround should be evaluated to determine if it s still necessary with the findings we should either update the code to remove the workaround or update the code comment to reference an open issue for tracking when we can remove the workaround cc deeprobin
| 0
|
14,451
| 17,532,815,286
|
IssuesEvent
|
2021-08-12 01:01:43
|
CodeForPittsburgh/food-access-map-data
|
https://api.github.com/repos/CodeForPittsburgh/food-access-map-data
|
closed
|
Break up Data Source Prep Scripts into individual scripts in new "Data Source Prep" folder
|
data processing data workflow
|
~~Right now they follow a nomenclature of "DATE_prep_sources_AUTHOR_NAME", which is confusing. This should probably be changed to "prep_sources_SOURCE_NAME", to avoid confusion.~~
See title and most recent issues for what needs to be done next. @cgmoreno and @hellonewman offering to do this right now.
|
1.0
|
Break up Data Source Prep Scripts into individual scripts in new "Data Source Prep" folder - ~~Right now they follow a nomenclature of "DATE_prep_sources_AUTHOR_NAME", which is confusing. This should probably be changed to "prep_sources_SOURCE_NAME", to avoid confusion.~~
See title and most recent issues for what needs to be done next. @cgmoreno and @hellonewman offering to do this right now.
|
process
|
break up data source prep scripts into individual scripts in new data source prep folder right now they follow a nomenclature of date prep sources author name which is confusing this should probably be changed to prep sources source name to avoid confusion see title and most recent issues for what needs to be done next cgmoreno and hellonewman offering to do this right now
| 1
|
2,416
| 5,199,145,548
|
IssuesEvent
|
2017-01-23 20:03:46
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
test-stdio-closed failing on some AIX environments
|
aix process test
|
Reference: #8375 (FYI @gireeshpunathil @Trott @mhdawson)
Happening on current versions of node 4 and 6 (4.7.0/6.9.2) and on AIX 6.1TL09
The fix that was applied in https://github.com/nodejs/node/commit/e65a2d7ddcf30077b14cc07d54c7953fb0028155 does not appear to make the test pass universally on all AIX systems. I have a mix of some that it works for, and others which are still returning "42" which is consistent with the other platforms, so is failing this assertion:
assert.strictEqual(exitCode, common.isAix ? 126 : 42);
On the system that it works on, if I set the LANG variable to anything other the en_US I get the "42" result (not certain yet why that is making a difference) although on one of my "failing" machines which is giving 42 setting it to en_US doesn't resolve it, despite it being in the output of locale -a. We could, potentially, be seeing a timing issue where going down the path of a non-default locale is changing things, but I'm guessing so far at this point.
The AIX level on the "failing" box (the one defaulting to en_GB which is always giving me 42) is patched to a slightly later level AIX level - .6100-09-08-1642 vs 6100-09-07-1614. On my 6100-07 system (defaults to en_US) it's always returning 42 as well.
I get the feeling we may need to be able to trap both cases for this test to pass reliably (or change the exit(126) to exit(42), or understand why we're getting the differences that caused the initial patch to be required.
|
1.0
|
test-stdio-closed failing on some AIX environments - Reference: #8375 (FYI @gireeshpunathil @Trott @mhdawson)
Happening on current versions of node 4 and 6 (4.7.0/6.9.2) and on AIX 6.1TL09
The fix that was applied in https://github.com/nodejs/node/commit/e65a2d7ddcf30077b14cc07d54c7953fb0028155 does not appear to make the test pass universally on all AIX systems. I have a mix of some that it works for, and others which are still returning "42" which is consistent with the other platforms, so is failing this assertion:
assert.strictEqual(exitCode, common.isAix ? 126 : 42);
On the system that it works on, if I set the LANG variable to anything other the en_US I get the "42" result (not certain yet why that is making a difference) although on one of my "failing" machines which is giving 42 setting it to en_US doesn't resolve it, despite it being in the output of locale -a. We could, potentially, be seeing a timing issue where going down the path of a non-default locale is changing things, but I'm guessing so far at this point.
The AIX level on the "failing" box (the one defaulting to en_GB which is always giving me 42) is patched to a slightly later level AIX level - .6100-09-08-1642 vs 6100-09-07-1614. On my 6100-07 system (defaults to en_US) it's always returning 42 as well.
I get the feeling we may need to be able to trap both cases for this test to pass reliably (or change the exit(126) to exit(42), or understand why we're getting the differences that caused the initial patch to be required.
|
process
|
test stdio closed failing on some aix environments reference fyi gireeshpunathil trott mhdawson happening on current versions of node and and on aix the fix that was applied in does not appear to make the test pass universally on all aix systems i have a mix of some that it works for and others which are still returning which is consistent with the other platforms so is failing this assertion assert strictequal exitcode common isaix on the system that it works on if i set the lang variable to anything other the en us i get the result not certain yet why that is making a difference although on one of my failing machines which is giving setting it to en us doesn t resolve it despite it being in the output of locale a we could potentially be seeing a timing issue where going down the path of a non default locale is changing things but i m guessing so far at this point the aix level on the failing box the one defaulting to en gb which is always giving me is patched to a slightly later level aix level vs on my system defaults to en us it s always returning as well i get the feeling we may need to be able to trap both cases for this test to pass reliably or change the exit to exit or understand why we re getting the differences that caused the initial patch to be required
| 1
|
3,225
| 6,285,004,389
|
IssuesEvent
|
2017-07-19 09:14:30
|
EstefaniaGuardado/weightApp
|
https://api.github.com/repos/EstefaniaGuardado/weightApp
|
closed
|
Convert measures
|
in process task
|
Convert measures:
- Height Unit : Pounds - Kilos
- Weight Unit: Feets - Meters
|
1.0
|
Convert measures - Convert measures:
- Height Unit : Pounds - Kilos
- Weight Unit: Feets - Meters
|
process
|
convert measures convert measures height unit pounds kilos weight unit feets meters
| 1
|
16,091
| 20,258,284,274
|
IssuesEvent
|
2022-02-15 03:04:54
|
allinurl/goaccess
|
https://api.github.com/repos/allinurl/goaccess
|
closed
|
Per Minute not working with json output
|
log-processing command-line options
|
I am using the command with `--hour-spec=min` but I don't get any additional by minute reports.
I am on version 1.5.3 with Ubuntu 20.04.
I have tried with and without `--date-spec=hr`, but no difference other then the hours are not included, etc.
Is `--hour-spec=min` suppose to work with json output?
Any Help would be appreciated, thanks.
|
1.0
|
Per Minute not working with json output - I am using the command with `--hour-spec=min` but I don't get any additional by minute reports.
I am on version 1.5.3 with Ubuntu 20.04.
I have tried with and without `--date-spec=hr`, but no difference other then the hours are not included, etc.
Is `--hour-spec=min` suppose to work with json output?
Any Help would be appreciated, thanks.
|
process
|
per minute not working with json output i am using the command with hour spec min but i don t get any additional by minute reports i am on version with ubuntu i have tried with and without date spec hr but no difference other then the hours are not included etc is hour spec min suppose to work with json output any help would be appreciated thanks
| 1
|
167,700
| 20,726,284,018
|
IssuesEvent
|
2022-03-14 02:32:49
|
kapseliboi/webuild
|
https://api.github.com/repos/kapseliboi/webuild
|
opened
|
CVE-2018-3728 (High) detected in hoek-2.16.3.tgz
|
security vulnerability
|
## CVE-2018-3728 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hoek-2.16.3.tgz</b></p></summary>
<p>General purpose node utilities</p>
<p>Library home page: <a href="https://registry.npmjs.org/hoek/-/hoek-2.16.3.tgz">https://registry.npmjs.org/hoek/-/hoek-2.16.3.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/hoek/package.json</p>
<p>
Dependency Hierarchy:
- webuild-events-1.4.39.tgz (Root Library)
- ical-0.5.0.tgz
- request-2.68.0.tgz
- hawk-3.1.3.tgz
- :x: **hoek-2.16.3.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
hoek node module before 4.2.0 and 5.0.x before 5.0.3 suffers from a Modification of Assumed-Immutable Data (MAID) vulnerability via 'merge' and 'applyToDefaults' functions, which allows a malicious user to modify the prototype of "Object" via __proto__, causing the addition or modification of an existing property that will exist on all objects.
<p>Publish Date: 2018-03-30
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-3728>CVE-2018-3728</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-16082">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-16082</a></p>
<p>Release Date: 2018-03-30</p>
<p>Fix Resolution: 4.2.0,5.0.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2018-3728 (High) detected in hoek-2.16.3.tgz - ## CVE-2018-3728 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hoek-2.16.3.tgz</b></p></summary>
<p>General purpose node utilities</p>
<p>Library home page: <a href="https://registry.npmjs.org/hoek/-/hoek-2.16.3.tgz">https://registry.npmjs.org/hoek/-/hoek-2.16.3.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/hoek/package.json</p>
<p>
Dependency Hierarchy:
- webuild-events-1.4.39.tgz (Root Library)
- ical-0.5.0.tgz
- request-2.68.0.tgz
- hawk-3.1.3.tgz
- :x: **hoek-2.16.3.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
hoek node module before 4.2.0 and 5.0.x before 5.0.3 suffers from a Modification of Assumed-Immutable Data (MAID) vulnerability via 'merge' and 'applyToDefaults' functions, which allows a malicious user to modify the prototype of "Object" via __proto__, causing the addition or modification of an existing property that will exist on all objects.
<p>Publish Date: 2018-03-30
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-3728>CVE-2018-3728</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-16082">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-16082</a></p>
<p>Release Date: 2018-03-30</p>
<p>Fix Resolution: 4.2.0,5.0.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in hoek tgz cve high severity vulnerability vulnerable library hoek tgz general purpose node utilities library home page a href path to dependency file package json path to vulnerable library node modules hoek package json dependency hierarchy webuild events tgz root library ical tgz request tgz hawk tgz x hoek tgz vulnerable library found in base branch master vulnerability details hoek node module before and x before suffers from a modification of assumed immutable data maid vulnerability via merge and applytodefaults functions which allows a malicious user to modify the prototype of object via proto causing the addition or modification of an existing property that will exist on all objects publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
32,396
| 13,797,971,283
|
IssuesEvent
|
2020-10-09 23:32:27
|
Azure/azure-cli-extensions
|
https://api.github.com/repos/Azure/azure-cli-extensions
|
closed
|
Azure CLI Extension Subscription Doesn't Support Azure Government Offer Types
|
Service Attention Subscription customer-reported extension/subscription
|
- If the issue is to do with Azure CLI 2.0 in-particular, create an issue here at [Azure/azure-cli](https://github.com/Azure/azure-cli/issues)
### Extension name (the extension in question)
subscription
### Description of issue (in as much detail as possible)
Doesn't support the Azure Government Offer Type MS-AZR-USGOV-0017P
```
az --version
azure-cli 2.6.0
command-modules-nspkg 2.0.3
core 2.6.0
nspkg 3.0.4
telemetry 1.0.4
Extensions:
subscription 0.1.3
az account create --enrollment-account-object-id <id> --offer-type "MS-AZR-USGOV-0017P" --display-name "DEMO-USGOV-SUB-01" --owner-object-id <id> --debug
az_command_data_logger : command args: account create --enrollment-account-object-id {} --offer-type {} --display-name {} --owner-object-id {} --debug
cli.azure.cli.core.parser : az account create: 'MS-AZR-USGOV-0017P' is not a valid value for '--offer-type'. See 'az account create --help'.
az account create: 'MS-AZR-USGOV-0017P' is not a valid value for '--offer-type'. See 'az account create --help'.
recommendation_provider: version: 2.6.0, command: account create, parameters: ['--enrollment-account-object-id', '--offer-type', '--display-name', '--owner-object-id', '--debug'], extension: subscription
The most similar choice to 'MS-AZR-USGOV-0017P' is:
MS-AZR-0017P
```
-----
|
1.0
|
Azure CLI Extension Subscription Doesn't Support Azure Government Offer Types - - If the issue is to do with Azure CLI 2.0 in-particular, create an issue here at [Azure/azure-cli](https://github.com/Azure/azure-cli/issues)
### Extension name (the extension in question)
subscription
### Description of issue (in as much detail as possible)
Doesn't support the Azure Government Offer Type MS-AZR-USGOV-0017P
```
az --version
azure-cli 2.6.0
command-modules-nspkg 2.0.3
core 2.6.0
nspkg 3.0.4
telemetry 1.0.4
Extensions:
subscription 0.1.3
az account create --enrollment-account-object-id <id> --offer-type "MS-AZR-USGOV-0017P" --display-name "DEMO-USGOV-SUB-01" --owner-object-id <id> --debug
az_command_data_logger : command args: account create --enrollment-account-object-id {} --offer-type {} --display-name {} --owner-object-id {} --debug
cli.azure.cli.core.parser : az account create: 'MS-AZR-USGOV-0017P' is not a valid value for '--offer-type'. See 'az account create --help'.
az account create: 'MS-AZR-USGOV-0017P' is not a valid value for '--offer-type'. See 'az account create --help'.
recommendation_provider: version: 2.6.0, command: account create, parameters: ['--enrollment-account-object-id', '--offer-type', '--display-name', '--owner-object-id', '--debug'], extension: subscription
The most similar choice to 'MS-AZR-USGOV-0017P' is:
MS-AZR-0017P
```
-----
|
non_process
|
azure cli extension subscription doesn t support azure government offer types if the issue is to do with azure cli in particular create an issue here at extension name the extension in question subscription description of issue in as much detail as possible doesn t support the azure government offer type ms azr usgov az version azure cli command modules nspkg core nspkg telemetry extensions subscription az account create enrollment account object id offer type ms azr usgov display name demo usgov sub owner object id debug az command data logger command args account create enrollment account object id offer type display name owner object id debug cli azure cli core parser az account create ms azr usgov is not a valid value for offer type see az account create help az account create ms azr usgov is not a valid value for offer type see az account create help recommendation provider version command account create parameters extension subscription the most similar choice to ms azr usgov is ms azr
| 0
|
9,149
| 12,203,221,367
|
IssuesEvent
|
2020-04-30 10:13:45
|
MHRA/products
|
https://api.github.com/repos/MHRA/products
|
closed
|
AUTO BATCH PROCESS - Documentation and Governance
|
EPIC - Auto Batch Process :oncoming_automobile: Medium Priority :arrow_forward: TASK :rescue_worker_helmet:
|
### User want
As a business
I want to have descriptive and accurate documentation
So that anyone can come along and support the product
**Customer acceptance criteria**
**Technical acceptance criteria**
**Data acceptance criteria**
**Testing acceptance criteria**
**Size**
L
**Value**
**Effort**
### Exit Criteria met
- [ ] Backlog
- [ ] Discovery
- [ ] DUXD
- [ ] Development
- [ ] Quality Assurance
- [ ] Release and Validate
|
1.0
|
AUTO BATCH PROCESS - Documentation and Governance - ### User want
As a business
I want to have descriptive and accurate documentation
So that anyone can come along and support the product
**Customer acceptance criteria**
**Technical acceptance criteria**
**Data acceptance criteria**
**Testing acceptance criteria**
**Size**
L
**Value**
**Effort**
### Exit Criteria met
- [ ] Backlog
- [ ] Discovery
- [ ] DUXD
- [ ] Development
- [ ] Quality Assurance
- [ ] Release and Validate
|
process
|
auto batch process documentation and governance user want as a business i want to have descriptive and accurate documentation so that anyone can come along and support the product customer acceptance criteria technical acceptance criteria data acceptance criteria testing acceptance criteria size l value effort exit criteria met backlog discovery duxd development quality assurance release and validate
| 1
|
7,680
| 10,762,400,637
|
IssuesEvent
|
2019-10-31 23:32:44
|
dCentralizedSystems/customer-support
|
https://api.github.com/repos/dCentralizedSystems/customer-support
|
closed
|
Expose UI control in remote stream sample client UI, to mark start/stop of a task (job)
|
UI enhancement fleet management stream processing
|
User connects to Unit, through its WiFi, goes to dashboard web page, selects Active Collection, which makes the unit publish data at 10x the rate. When job or task is done, user (or robot task service) selects Passive Collection to reduce publish rate and data costs
|
1.0
|
Expose UI control in remote stream sample client UI, to mark start/stop of a task (job) - User connects to Unit, through its WiFi, goes to dashboard web page, selects Active Collection, which makes the unit publish data at 10x the rate. When job or task is done, user (or robot task service) selects Passive Collection to reduce publish rate and data costs
|
process
|
expose ui control in remote stream sample client ui to mark start stop of a task job user connects to unit through its wifi goes to dashboard web page selects active collection which makes the unit publish data at the rate when job or task is done user or robot task service selects passive collection to reduce publish rate and data costs
| 1
|
826,778
| 31,711,877,094
|
IssuesEvent
|
2023-09-09 11:29:41
|
simonSlamka/simtoon-journal
|
https://api.github.com/repos/simonSlamka/simtoon-journal
|
closed
|
Dietary Deviations
|
bug help wanted !!! BLUE PRIORITY !!!
|
Occasional deviations from a strict dietary plan, especially concerning added sugar and alcohol, are compromising health goals and doctor-instructed regulations. This problem presents itself in unplanned indulgences, poor nutritional choices, and failure to adhere to the dietary discipline, leading to potential health risks.
|
1.0
|
Dietary Deviations - Occasional deviations from a strict dietary plan, especially concerning added sugar and alcohol, are compromising health goals and doctor-instructed regulations. This problem presents itself in unplanned indulgences, poor nutritional choices, and failure to adhere to the dietary discipline, leading to potential health risks.
|
non_process
|
dietary deviations occasional deviations from a strict dietary plan especially concerning added sugar and alcohol are compromising health goals and doctor instructed regulations this problem presents itself in unplanned indulgences poor nutritional choices and failure to adhere to the dietary discipline leading to potential health risks
| 0
|
21,101
| 28,056,084,549
|
IssuesEvent
|
2023-03-29 09:27:38
|
camunda/issues
|
https://api.github.com/repos/camunda/issues
|
opened
|
Allow unsupported BPMN elements in non executable pools
|
component:zeebe-process-automation public kind:epic feature-parity version:8.2-alpha4
|
### Value Proposition Statement
Use unsupported BPMN elements in non-executable pools for documentation purposes.
### User Problem
Customers following our BPMN methodology described also in Real-Life BPMN might create collaboration diagrams that contain executable and non-executable Pools (for example a model like in https://camunda.com/bpmn/examples/#bpmn-examples-four-eyes-principle).
In https://github.com/camunda-community-hub/camunda-7-to-8-migration/issues/19 Stephan raised the problem that Zeebe rejects deployments containing unsupported elements, even if in a nonexecutable pool (e.g. Element: ConditionalEventDefinition_0lvmueg ERROR: Event definition of this type is not supported ).
### User Stories
Zeebe does not need to check elements in non-executable pools - as they are not executed and are ignored anyway.
This would allow using all BPMN elements for documentation purposes
|
1.0
|
Allow unsupported BPMN elements in non executable pools - ### Value Proposition Statement
Use unsupported BPMN elements in non-executable pools for documentation purposes.
### User Problem
Customers following our BPMN methodology described also in Real-Life BPMN might create collaboration diagrams that contain executable and non-executable Pools (for example a model like in https://camunda.com/bpmn/examples/#bpmn-examples-four-eyes-principle).
In https://github.com/camunda-community-hub/camunda-7-to-8-migration/issues/19 Stephan raised the problem that Zeebe rejects deployments containing unsupported elements, even if in a nonexecutable pool (e.g. Element: ConditionalEventDefinition_0lvmueg ERROR: Event definition of this type is not supported ).
### User Stories
Zeebe does not need to check elements in non-executable pools - as they are not executed and are ignored anyway.
This would allow using all BPMN elements for documentation purposes
|
process
|
allow unsupported bpmn elements in non executable pools value proposition statement use unsupported bpmn elements in non executable pools for documentation purposes user problem customers following our bpmn methodology described also in real life bpmn might create collaboration diagrams that contain executable and non executable pools for example a model like in in stephan raised the problem that zeebe rejects deployments containing unsupported elements even if in a nonexecutable pool e g element conditionaleventdefinition error event definition of this type is not supported user stories zeebe does not need to check elements in non executable pools as they are not executed and are ignored anyway this would allow using all bpmn elements for documentation purposes
| 1
|
76,476
| 9,449,689,640
|
IssuesEvent
|
2019-04-16 03:02:37
|
danbriechle/In-Context
|
https://api.github.com/repos/danbriechle/In-Context
|
closed
|
Create wire frames
|
Chore Design
|
Create wire frames of the User dashboard page, the project bucket index page and the project bucket show page.
|
1.0
|
Create wire frames - Create wire frames of the User dashboard page, the project bucket index page and the project bucket show page.
|
non_process
|
create wire frames create wire frames of the user dashboard page the project bucket index page and the project bucket show page
| 0
|
400,827
| 11,781,271,897
|
IssuesEvent
|
2020-03-16 22:00:32
|
kubernetes-sigs/cluster-api
|
https://api.github.com/repos/kubernetes-sigs/cluster-api
|
closed
|
Support variants of image tags in KCP for kube-proxy, coredns etc...
|
kind/bug lifecycle/active priority/critical-urgent
|
**What steps did you take and what happened:**
[A clear and concise description on how to REPRODUCE the bug.]
Setting kcp.spec.version to 1.17.3 results in a kube-proxy daemonset being applied with a "1.17.3" tag.
**What did you expect to happen:**
Tag being "v1.17.3"
**Anything else you would like to add:**
[Miscellaneous information that will assist in solving the issue.]
This is currently occuring due to the reconciliation of the kube-proxy daemonset due to upgrades.
There are additional issues when dealing with version numbers that may be semver compliant, e.g. "v1.17.3+build1", but do not contain OCI compatible characters for tags.
Finally, the KCP controller should account for independent build tags for different components, e.g. API server could be at build1, and kube-proxy is at build2.
**Environment:**
- Cluster-api version: 0.3.0
- Minikube/KIND version: N/A
- Kubernetes version: (use `kubectl version`): N/A
- OS (e.g. from `/etc/os-release`): N/A
/kind bug
/lifecycle active
|
1.0
|
Support variants of image tags in KCP for kube-proxy, coredns etc... - **What steps did you take and what happened:**
[A clear and concise description on how to REPRODUCE the bug.]
Setting kcp.spec.version to 1.17.3 results in a kube-proxy daemonset being applied with a "1.17.3" tag.
**What did you expect to happen:**
Tag being "v1.17.3"
**Anything else you would like to add:**
[Miscellaneous information that will assist in solving the issue.]
This is currently occuring due to the reconciliation of the kube-proxy daemonset due to upgrades.
There are additional issues when dealing with version numbers that may be semver compliant, e.g. "v1.17.3+build1", but do not contain OCI compatible characters for tags.
Finally, the KCP controller should account for independent build tags for different components, e.g. API server could be at build1, and kube-proxy is at build2.
**Environment:**
- Cluster-api version: 0.3.0
- Minikube/KIND version: N/A
- Kubernetes version: (use `kubectl version`): N/A
- OS (e.g. from `/etc/os-release`): N/A
/kind bug
/lifecycle active
|
non_process
|
support variants of image tags in kcp for kube proxy coredns etc what steps did you take and what happened setting kcp spec version to results in a kube proxy daemonset being applied with a tag what did you expect to happen tag being anything else you would like to add this is currently occuring due to the reconciliation of the kube proxy daemonset due to upgrades there are additional issues when dealing with version numbers that may be semver compliant e g but do not contain oci compatible characters for tags finally the kcp controller should account for independent build tags for different components e g api server could be at and kube proxy is at environment cluster api version minikube kind version n a kubernetes version use kubectl version n a os e g from etc os release n a kind bug lifecycle active
| 0
|
446,481
| 12,857,804,945
|
IssuesEvent
|
2020-07-09 09:55:03
|
kuzzleio/kuzzle-admin-console
|
https://api.github.com/repos/kuzzleio/kuzzle-admin-console
|
closed
|
Crash if adminExists route is not allowed
|
4-dev bug priority-blocking
|
As title says...
Kuzzle version: 2.3.2
Node.js version:
Admin Console Version: 4
|
1.0
|
Crash if adminExists route is not allowed - As title says...
Kuzzle version: 2.3.2
Node.js version:
Admin Console Version: 4
|
non_process
|
crash if adminexists route is not allowed as title says kuzzle version node js version admin console version
| 0
|
407,702
| 27,626,477,965
|
IssuesEvent
|
2023-03-10 07:18:40
|
Rodez-IUT/blue-checkyourmood-blue1
|
https://api.github.com/repos/Rodez-IUT/blue-checkyourmood-blue1
|
opened
|
AUD001 - Livrable audit
|
documentation
|
À l’issue de l’audit, un document contiendra la restitution des résultats de l’audit.
• Le document aura pour nom "SAE-S4-Audit-CheckYourMood-V1-<Noms membres équipe>".
• Le document sera structuré en 4 parties en plus de l’introduction et de la conclusion :
1. Évaluation
2. Plan d’amélioration
3. Choix de la configuration des machines virtuelles
4. Démonstration de la faisabilité
Le document, en tant que livrable produit, sera accessible en ligne avec accès en mode
suggestion à toutes les parties prenantes
|
1.0
|
AUD001 - Livrable audit - À l’issue de l’audit, un document contiendra la restitution des résultats de l’audit.
• Le document aura pour nom "SAE-S4-Audit-CheckYourMood-V1-<Noms membres équipe>".
• Le document sera structuré en 4 parties en plus de l’introduction et de la conclusion :
1. Évaluation
2. Plan d’amélioration
3. Choix de la configuration des machines virtuelles
4. Démonstration de la faisabilité
Le document, en tant que livrable produit, sera accessible en ligne avec accès en mode
suggestion à toutes les parties prenantes
|
non_process
|
livrable audit à l’issue de l’audit un document contiendra la restitution des résultats de l’audit • le document aura pour nom sae audit checkyourmood • le document sera structuré en parties en plus de l’introduction et de la conclusion évaluation plan d’amélioration choix de la configuration des machines virtuelles démonstration de la faisabilité le document en tant que livrable produit sera accessible en ligne avec accès en mode suggestion à toutes les parties prenantes
| 0
|
73,091
| 3,407,034,876
|
IssuesEvent
|
2015-12-04 00:05:08
|
tangrams/eraser-map
|
https://api.github.com/repos/tangrams/eraser-map
|
opened
|
Building strokes are too strong
|
bug priority 1
|
Combo of stroke weight & stroke color makes it hard to read the street labels.
|
1.0
|
Building strokes are too strong - Combo of stroke weight & stroke color makes it hard to read the street labels.
|
non_process
|
building strokes are too strong combo of stroke weight stroke color makes it hard to read the street labels
| 0
|
340,479
| 30,518,794,227
|
IssuesEvent
|
2023-07-19 06:27:21
|
confidential-containers/cloud-api-adaptor
|
https://api.github.com/repos/confidential-containers/cloud-api-adaptor
|
closed
|
Remove repeated code from the Azure provisioner
|
CI provider/azure e2e-test
|
There is only one config change between the two. There is no need to repeat the config:
https://github.com/confidential-containers/cloud-api-adaptor/blob/1d84e3fce7456f70dc1370703e9eff29a1aed4e0/test/provisioner/provision_azure.go#L224-L272
|
1.0
|
Remove repeated code from the Azure provisioner - There is only one config change between the two. There is no need to repeat the config:
https://github.com/confidential-containers/cloud-api-adaptor/blob/1d84e3fce7456f70dc1370703e9eff29a1aed4e0/test/provisioner/provision_azure.go#L224-L272
|
non_process
|
remove repeated code from the azure provisioner there is only one config change between the two there is no need to repeat the config
| 0
|
38,416
| 2,847,035,194
|
IssuesEvent
|
2015-05-29 14:54:14
|
HellscreamWoW/Tracker
|
https://api.github.com/repos/HellscreamWoW/Tracker
|
closed
|
The Grol'dom Militia
|
Priority-Normal Type-Creature Type-Quest
|
This quest and the NPC are missing from the Grol'dom farm.
Quest Details:
http://www.wowhead.com/quest=13973/the-groldom-militia
|
1.0
|
The Grol'dom Militia - This quest and the NPC are missing from the Grol'dom farm.
Quest Details:
http://www.wowhead.com/quest=13973/the-groldom-militia
|
non_process
|
the grol dom militia this quest and the npc are missing from the grol dom farm quest details
| 0
|
4,490
| 7,345,952,022
|
IssuesEvent
|
2018-03-07 19:05:03
|
UKHomeOffice/dq-aws-transition
|
https://api.github.com/repos/UKHomeOffice/dq-aws-transition
|
closed
|
Add data-transfer job for ACL data to S3 archive
|
DQ Data Ingest DQ Tranche 1 Production SSM processing
|
- [x] Add data-transfer job for ACL data to S3 archive
|
1.0
|
Add data-transfer job for ACL data to S3 archive - - [x] Add data-transfer job for ACL data to S3 archive
|
process
|
add data transfer job for acl data to archive add data transfer job for acl data to archive
| 1
|
115,821
| 11,888,407,126
|
IssuesEvent
|
2020-03-28 08:15:46
|
RoboJackets/igvc-software
|
https://api.github.com/repos/RoboJackets/igvc-software
|
opened
|
Documentation for costmap layers
|
area ➤ mapping level ➤ easy type ➤ documentation
|
With the modifications to the local and global costmap over the year, they've grown increasingly complex. We need comments for every layer and their important parameters.
AC: Layers for the local and global costmap have header documentation and comments next to parameters.
|
1.0
|
Documentation for costmap layers - With the modifications to the local and global costmap over the year, they've grown increasingly complex. We need comments for every layer and their important parameters.
AC: Layers for the local and global costmap have header documentation and comments next to parameters.
|
non_process
|
documentation for costmap layers with the modifications to the local and global costmap over the year they ve grown increasingly complex we need comments for every layer and their important parameters ac layers for the local and global costmap have header documentation and comments next to parameters
| 0
|
12,243
| 14,743,988,296
|
IssuesEvent
|
2021-01-07 14:41:44
|
kdjstudios/SABillingGitlab
|
https://api.github.com/repos/kdjstudios/SABillingGitlab
|
closed
|
****URGENT**** Late fees in SAB
|
anc-process anp-urgent ant-bug ant-enhancement
|
In GitLab by @kdjstudios on Dec 23, 2019, 14:21
**Submitted by:** "Alina King" <alina.king@answernet.com>
**Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2019-12-23-27297
**Server:** Internal
**Client/Site:** 061
**Account:** 4343
**Issue:**
I am currently working on my billing now and I pulled the Service Code Summary for the late fees.
It is showing many of my late fees with high amounts and this is not correct.
For example account Avera & Linn 061-4343: Their monthly rate is $120 and they paid by CC on 12/3/19 (full balance). They are showing a late fee of $90.37.
Could this please be looked at right away as I can not continue billing until this is corrected.
I have attached the
|
1.0
|
****URGENT**** Late fees in SAB - In GitLab by @kdjstudios on Dec 23, 2019, 14:21
**Submitted by:** "Alina King" <alina.king@answernet.com>
**Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2019-12-23-27297
**Server:** Internal
**Client/Site:** 061
**Account:** 4343
**Issue:**
I am currently working on my billing now and I pulled the Service Code Summary for the late fees.
It is showing many of my late fees with high amounts and this is not correct.
For example account Avera & Linn 061-4343: Their monthly rate is $120 and they paid by CC on 12/3/19 (full balance). They are showing a late fee of $90.37.
Could this please be looked at right away as I can not continue billing until this is corrected.
I have attached the
|
process
|
urgent late fees in sab in gitlab by kdjstudios on dec submitted by alina king helpdesk server internal client site account issue i am currently working on my billing now and i pulled the service code summary for the late fees it is showing many of my late fees with high amounts and this is not correct for example account avera linn their monthly rate is and they paid by cc on full balance they are showing a late fee of could this please be looked at right away as i can not continue billing until this is corrected i have attached the
| 1
|
13,304
| 15,778,741,771
|
IssuesEvent
|
2021-04-01 08:01:17
|
didi/mpx
|
https://api.github.com/repos/didi/mpx
|
closed
|
[Bug report] 支付宝端externalClasses无法多层传递
|
processing
|
**问题描述**
mpx.conf.js
`externalClasses=["custom-class", "info-class"]`
页面样式
`.info-tag {xxx}`
```html
<!-- 页面 -->
<A info-class="info-tag"></A>
```
```html
<!-- 组件A -->
<B custom-class="info-class"></B>
<!-- externalClasses: [‘info-class’] -->
```
```html
<!-- 组件B -->
<view class="custom-class"></view>
<!-- externalClasses: [‘custom-class’] -->
```
组件B中的view无法应用页面中的`.info-tag`类
我现在处理方式是
组件B增加
`class@ali="info-class"`
微信里面不需要这样处理,不知道是否有其他解决办法?能做到类似微信里那种效果吗
|
1.0
|
[Bug report] 支付宝端externalClasses无法多层传递 - **问题描述**
mpx.conf.js
`externalClasses=["custom-class", "info-class"]`
页面样式
`.info-tag {xxx}`
```html
<!-- 页面 -->
<A info-class="info-tag"></A>
```
```html
<!-- 组件A -->
<B custom-class="info-class"></B>
<!-- externalClasses: [‘info-class’] -->
```
```html
<!-- 组件B -->
<view class="custom-class"></view>
<!-- externalClasses: [‘custom-class’] -->
```
组件B中的view无法应用页面中的`.info-tag`类
我现在处理方式是
组件B增加
`class@ali="info-class"`
微信里面不需要这样处理,不知道是否有其他解决办法?能做到类似微信里那种效果吗
|
process
|
支付宝端externalclasses无法多层传递 问题描述 mpx conf js externalclasses 页面样式 info tag xxx html html html 组件b中的view无法应用页面中的 info tag 类 我现在处理方式是 组件b增加 class ali info class 微信里面不需要这样处理,不知道是否有其他解决办法 能做到类似微信里那种效果吗
| 1
|
9,136
| 12,203,182,124
|
IssuesEvent
|
2020-04-30 10:09:47
|
MHRA/products
|
https://api.github.com/repos/MHRA/products
|
closed
|
SFTP error "File could not be opened"
|
BUG :bug: EPIC - Auto Batch Process :oncoming_automobile:
|
**Describe the bug**
Receiving a "File could not be opened" error when retrieving a file using SFTP from Sentinel.
This could be for a number of reasons:
- Inadequate permissions to access/open file
- File does not exist in the requested directory
- There's some form of lock on the file
**To Reproduce**
Only reproducible when the batch process is running from Sentinel.
**Expected behavior**
File can be retrieved without errors.
**Screenshots**
N/A
**Additional context**
N/A
|
1.0
|
SFTP error "File could not be opened" - **Describe the bug**
Receiving a "File could not be opened" error when retrieving a file using SFTP from Sentinel.
This could be for a number of reasons:
- Inadequate permissions to access/open file
- File does not exist in the requested directory
- There's some form of lock on the file
**To Reproduce**
Only reproducible when the batch process is running from Sentinel.
**Expected behavior**
File can be retrieved without errors.
**Screenshots**
N/A
**Additional context**
N/A
|
process
|
sftp error file could not be opened describe the bug receiving a file could not be opened error when retrieving a file using sftp from sentinel this could be for a number of reasons inadequate permissions to access open file file does not exist in the requested directory there s some form of lock on the file to reproduce only reproducible when the batch process is running from sentinel expected behavior file can be retrieved without errors screenshots n a additional context n a
| 1
|
27,324
| 12,540,633,285
|
IssuesEvent
|
2020-06-05 10:44:20
|
terraform-providers/terraform-provider-azurerm
|
https://api.github.com/repos/terraform-providers/terraform-provider-azurerm
|
closed
|
Support for Azure Function endpoints on EventGrid Subscriptions
|
enhancement service/event-grid
|
<!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Description
Azure Functions are a supported endpoint for EventGrid subscription handling, as [described in the documentation here.](https://docs.microsoft.com/en-us/azure/event-grid/handler-functions)
Currently, `azurerm_eventgrid_event_subscription` does not support adding such a function as the endpoint.
### New or Affected Resource(s)
<!--- Please list the new or affected resources and data sources. --->
* azurerm_function_app
* azurerm_eventgrid_event_subscription
### Potential Terraform Configuration
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
```hcl
resource "azurerm_resource_group" "example" {
name = "exampleResourceGroup"
location = "West US 2"
}
resource "azurerm_eventgrid_event_subscription" "example" {
name = "example"
scope = azurerm_resource_group.example.id
function_endpoint {
function_app_id = azurerm_function_app.example.id
function_name = "ExampleFunction"
}
```
### References
<!---
Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests
Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation? For example:
* https://azure.microsoft.com/en-us/roadmap/virtual-network-service-endpoint-for-azure-cosmos-db/
--->
* #0000
|
1.0
|
Support for Azure Function endpoints on EventGrid Subscriptions - <!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Description
Azure Functions are a supported endpoint for EventGrid subscription handling, as [described in the documentation here.](https://docs.microsoft.com/en-us/azure/event-grid/handler-functions)
Currently, `azurerm_eventgrid_event_subscription` does not support adding such a function as the endpoint.
### New or Affected Resource(s)
<!--- Please list the new or affected resources and data sources. --->
* azurerm_function_app
* azurerm_eventgrid_event_subscription
### Potential Terraform Configuration
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
```hcl
resource "azurerm_resource_group" "example" {
name = "exampleResourceGroup"
location = "West US 2"
}
resource "azurerm_eventgrid_event_subscription" "example" {
name = "example"
scope = azurerm_resource_group.example.id
function_endpoint {
function_app_id = azurerm_function_app.example.id
function_name = "ExampleFunction"
}
```
### References
<!---
Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests
Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation? For example:
* https://azure.microsoft.com/en-us/roadmap/virtual-network-service-endpoint-for-azure-cosmos-db/
--->
* #0000
|
non_process
|
support for azure function endpoints on eventgrid subscriptions community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment description azure functions are a supported endpoint for eventgrid subscription handling as currently azurerm eventgrid event subscription does not support adding such a function as the endpoint new or affected resource s azurerm function app azurerm eventgrid event subscription potential terraform configuration hcl resource azurerm resource group example name exampleresourcegroup location west us resource azurerm eventgrid event subscription example name example scope azurerm resource group example id function endpoint function app id azurerm function app example id function name examplefunction references information about referencing github issues are there any other github issues open or closed or pull requests that should be linked here vendor blog posts or documentation for example
| 0
|
477
| 2,911,385,732
|
IssuesEvent
|
2015-06-22 09:13:47
|
haskell-distributed/distributed-process
|
https://api.github.com/repos/haskell-distributed/distributed-process
|
closed
|
SimpleLocalnet should remove peers that no longer exist
|
distributed-process-simplelocalnet Feature Request
|
This is not a *major* issue, because even if we do, there is still no guarantee that a peer still exists after it has been returned by `findPeers`.
|
1.0
|
SimpleLocalnet should remove peers that no longer exist - This is not a *major* issue, because even if we do, there is still no guarantee that a peer still exists after it has been returned by `findPeers`.
|
process
|
simplelocalnet should remove peers that no longer exist this is not a major issue because even if we do there is still no guarantee that a peer still exists after it has been returned by findpeers
| 1
|
1,525
| 4,117,548,716
|
IssuesEvent
|
2016-06-08 07:58:31
|
ProgrammingLife2016/PL2-2016
|
https://api.github.com/repos/ProgrammingLife2016/PL2-2016
|
opened
|
Mutations are sometimes hidden away
|
data processing discussion
|
While browsing the application, I came across the following structure. I was wondering whether the straight vertical edges between nodes showed something interesting



I'm not sure whether we need to change this. The edge in a large zoom does show a small mutation at that position.
Thoughts would be nice.
|
1.0
|
Mutations are sometimes hidden away - While browsing the application, I came across the following structure. I was wondering whether the straight vertical edges between nodes showed something interesting



I'm not sure whether we need to change this. The edge in a large zoom does show a small mutation at that position.
Thoughts would be nice.
|
process
|
mutations are sometimes hidden away while browsing the application i came across the following structure i was wondering whether the straight vertical edges between nodes showed something interesting i m not sure whether we need to change this the edge in a large zoom does show a small mutation at that position thoughts would be nice
| 1
|
14,543
| 17,658,456,385
|
IssuesEvent
|
2021-08-21 02:39:09
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
Join attributes by location on large dataset: consumes all memory and crashes
|
Feedback stale Processing Bug Crash/Data Corruption
|
<!--
Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone.
If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix
Checklist before submitting
- [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists
- [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles).
- [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue
-->
**Describe the bug**
<When joining attributes by location for two layers that are larger (172,217 complex polygons (dense nodes) and 5,638,125 points, QGIS appears to increase memory usage till all memory is consumed then crashes.>
**How to Reproduce**
1. Added polygon layer (172,217 shapes)
2. Added point layer (5,638,125 points)
3. Memory consumed in Task Manager: 295.6 MB
4. Select Join attributes by location from Processing Tools panel
5. Parameters: discard unjoined records, one-to-many, create temporary layer for joined and unjoined records
6. Run
I cannot upload the data as I work for the Government and the data is protected.
I can try to describe for you the complexity of the data:
- the point data is simple with about a six attributes of type INT and FLOAT
- the polygon data is more complex meaning the feature geometries have a dense node distribution. The layer represents coverage of radio waves. The density of the nodes is required to capture the variations in signal strength coming from radio wave emmissions.
- the data spans the width of North America and is about the height of Canada or the US.
**QGIS and OS versions**
QGIS version | 3.16.4-Hannover | QGIS code revision | 654e76b3cb
-- | -- | -- | --
Compiled against Qt | 5.11.2 | Running against Qt | 5.11.2
Compiled against GDAL/OGR | 3.1.4 | Running against GDAL/OGR | 3.1.4
Compiled against GEOS | 3.8.1-CAPI-1.13.3 | Running against GEOS | 3.8.1-CAPI-1.13.3
Compiled against SQLite | 3.29.0 | Running against SQLite | 3.29.0
PostgreSQL Client Version | 11.5 | SpatiaLite Version | 4.3.0
QWT Version | 6.1.3 | QScintilla2 Version | 2.10.8
Compiled against PROJ | 6.3.2 | Running against PROJ | Rel. 6.3.2, May 1st, 2020
OS Version | Windows 10 (10.0)
Active python plugins | DissolveWithStats; GroupStats; latlontools; SpreadsheetLayers; db_manager; processing
**Additional context**
<!-- Add any other context about the problem here. -->
We are making great strides bringing QGIS into our department and replacing MapInfo as the go-to GIS client. Our work very directly impacts the day-to-day lives of millions of people (points represent houses and buildings. I am hoping this one thing can be escalated and fixed so that we can continue demonstrating the value of QGIS. Many thanks.
|
1.0
|
Join attributes by location on large dataset: consumes all memory and crashes - <!--
Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone.
If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix
Checklist before submitting
- [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists
- [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles).
- [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue
-->
**Describe the bug**
<When joining attributes by location for two layers that are larger (172,217 complex polygons (dense nodes) and 5,638,125 points, QGIS appears to increase memory usage till all memory is consumed then crashes.>
**How to Reproduce**
1. Added polygon layer (172,217 shapes)
2. Added point layer (5,638,125 points)
3. Memory consumed in Task Manager: 295.6 MB
4. Select Join attributes by location from Processing Tools panel
5. Parameters: discard unjoined records, one-to-many, create temporary layer for joined and unjoined records
6. Run
I cannot upload the data as I work for the Government and the data is protected.
I can try to describe for you the complexity of the data:
- the point data is simple with about a six attributes of type INT and FLOAT
- the polygon data is more complex meaning the feature geometries have a dense node distribution. The layer represents coverage of radio waves. The density of the nodes is required to capture the variations in signal strength coming from radio wave emmissions.
- the data spans the width of North America and is about the height of Canada or the US.
**QGIS and OS versions**
QGIS version | 3.16.4-Hannover | QGIS code revision | 654e76b3cb
-- | -- | -- | --
Compiled against Qt | 5.11.2 | Running against Qt | 5.11.2
Compiled against GDAL/OGR | 3.1.4 | Running against GDAL/OGR | 3.1.4
Compiled against GEOS | 3.8.1-CAPI-1.13.3 | Running against GEOS | 3.8.1-CAPI-1.13.3
Compiled against SQLite | 3.29.0 | Running against SQLite | 3.29.0
PostgreSQL Client Version | 11.5 | SpatiaLite Version | 4.3.0
QWT Version | 6.1.3 | QScintilla2 Version | 2.10.8
Compiled against PROJ | 6.3.2 | Running against PROJ | Rel. 6.3.2, May 1st, 2020
OS Version | Windows 10 (10.0)
Active python plugins | DissolveWithStats; GroupStats; latlontools; SpreadsheetLayers; db_manager; processing
**Additional context**
<!-- Add any other context about the problem here. -->
We are making great strides bringing QGIS into our department and replacing MapInfo as the go-to GIS client. Our work very directly impacts the day-to-day lives of millions of people (points represent houses and buildings. I am hoping this one thing can be escalated and fixed so that we can continue demonstrating the value of QGIS. Many thanks.
|
process
|
join attributes by location on large dataset consumes all memory and crashes bug fixing and feature development is a community responsibility and not the responsibility of the qgis project alone if this bug report or feature request is high priority for you we suggest engaging a qgis developer or support organisation and financially sponsoring a fix checklist before submitting search through existing issue reports and gis stackexchange com to check whether the issue already exists test with a create a light and self contained sample dataset and project file which demonstrates the issue describe the bug how to reproduce added polygon layer shapes added point layer points memory consumed in task manager mb select join attributes by location from processing tools panel parameters discard unjoined records one to many create temporary layer for joined and unjoined records run i cannot upload the data as i work for the government and the data is protected i can try to describe for you the complexity of the data the point data is simple with about a six attributes of type int and float the polygon data is more complex meaning the feature geometries have a dense node distribution the layer represents coverage of radio waves the density of the nodes is required to capture the variations in signal strength coming from radio wave emmissions the data spans the width of north america and is about the height of canada or the us qgis and os versions qgis version hannover qgis code revision compiled against qt running against qt compiled against gdal ogr running against gdal ogr compiled against geos capi running against geos capi compiled against sqlite running against sqlite postgresql client version spatialite version qwt version version compiled against proj running against proj rel may os version windows active python plugins dissolvewithstats groupstats latlontools spreadsheetlayers db manager processing additional context we are making great strides bringing qgis into our department and replacing mapinfo as the go to gis client our work very directly impacts the day to day lives of millions of people points represent houses and buildings i am hoping this one thing can be escalated and fixed so that we can continue demonstrating the value of qgis many thanks
| 1
|
21,672
| 30,119,613,706
|
IssuesEvent
|
2023-06-30 14:13:07
|
eclipse-theia/theia
|
https://api.github.com/repos/eclipse-theia/theia
|
closed
|
terminal: send `\r` to submit commands on Windows shells
|
bug terminal debug process
|
### Description
Follow-up of https://github.com/eclipse-theia/theia/pull/6836.
I originally did not understand why in some circumstances I had to send `\n` twice for commands to be evaluated, but I now realize it is because shells seems to expect `\r` instead.
### Reproduction Steps
See https://github.com/eclipse-theia/theia/issues/7141#issuecomment-609030444
**OS and Theia version:** Windows 10, Theia 1.0.0
|
1.0
|
terminal: send `\r` to submit commands on Windows shells - ### Description
Follow-up of https://github.com/eclipse-theia/theia/pull/6836.
I originally did not understand why in some circumstances I had to send `\n` twice for commands to be evaluated, but I now realize it is because shells seems to expect `\r` instead.
### Reproduction Steps
See https://github.com/eclipse-theia/theia/issues/7141#issuecomment-609030444
**OS and Theia version:** Windows 10, Theia 1.0.0
|
process
|
terminal send r to submit commands on windows shells description follow up of i originally did not understand why in some circumstances i had to send n twice for commands to be evaluated but i now realize it is because shells seems to expect r instead reproduction steps see os and theia version windows theia
| 1
|
18,006
| 24,024,188,998
|
IssuesEvent
|
2022-09-15 10:04:51
|
geneontology/go-ontology
|
https://api.github.com/repos/geneontology/go-ontology
|
closed
|
R-loop disassembly
|
cell cycle and DNA processes
|
Hi GO,
I would need a term to annotate R-loop disassembly
Thanks
Sylvain
Category: Process
Definition:
A R-loop disassembly process that results in the disassembly of R-loops. R-loops are three-stranded nucleic acid structures consisting of an RNA:DNA heteroduplex and a “looped-out” non-template strand. Aberrant formation and persistence of R-loops block transcription elongation and cause DNA damage, mechanisms that resolve R-loops are essential for genome stability.
is_a DNA metabolic process
Part_of GO:0006261 DNA-dependent DNA replication
Xref.
PMID: 28790157
|
1.0
|
R-loop disassembly - Hi GO,
I would need a term to annotate R-loop disassembly
Thanks
Sylvain
Category: Process
Definition:
A R-loop disassembly process that results in the disassembly of R-loops. R-loops are three-stranded nucleic acid structures consisting of an RNA:DNA heteroduplex and a “looped-out” non-template strand. Aberrant formation and persistence of R-loops block transcription elongation and cause DNA damage, mechanisms that resolve R-loops are essential for genome stability.
is_a DNA metabolic process
Part_of GO:0006261 DNA-dependent DNA replication
Xref.
PMID: 28790157
|
process
|
r loop disassembly hi go i would need a term to annotate r loop disassembly thanks sylvain category process definition a r loop disassembly process that results in the disassembly of r loops r loops are three stranded nucleic acid structures consisting of an rna dna heteroduplex and a “looped out” non template strand aberrant formation and persistence of r loops block transcription elongation and cause dna damage mechanisms that resolve r loops are essential for genome stability is a dna metabolic process part of go dna dependent dna replication xref pmid
| 1
|
11,214
| 13,979,430,305
|
IssuesEvent
|
2020-10-27 00:06:35
|
firebase/quickstart-unity
|
https://api.github.com/repos/firebase/quickstart-unity
|
closed
|
Invalid bitcode version
|
api: core type: process
|
<!-- DO NOT DELETE
validate_template=true
template_path=.github/ISSUE_TEMPLATE/issue.md
-->
### [REQUIRED] Please fill in the following fields:
* Unity editor version: 2020.1.3f1
* Firebase Unity SDK version: 6.15.2
* Source you installed the SDK: UPM
* Problematic Firebase Component: ---
* Other Firebase Components in use: Auth, FCM, Crashlytics, Realtime DB
* Additional SDKs you are using: Google Analytics for Firebase/Google App Bundle/Google In-App Review
* Platform you are using the Unity editor on: Mac
* Platform you are targeting: iOS
* Scripting Runtime: IL2CPP
### [REQUIRED] Please describe the issue here:
Stuck when archiving project from xCode v 11.3.1, building from xCode works fine, but when archiving to install the app on a device this error shows up.
`ld: could not reparse object file in bitcode bundle: 'Invalid bitcode version (Producer: '1103.0.32.29.0_0' Reader: '1100.0.33.17_0')', using libLTO version 'LLVM version 11.0.0, (clang-1100.0.33.17)' for architecture arm64`
#### Steps to reproduce:
Have you been able to reproduce this issue with just the Firebase Unity quickstarts (this GitHub project)? Haven't tried
What's the issue repro rate? 100%
What happened? How can we make the problem occur?
1 - Build the game on Unity making sure to select xCode v 11.3.1
2 - Once the build is done and since we are using Firebase, we followed the steps mentioned in the Firebase dashboard when adding an iOS app and installed the pod via `pod install`, that successfully installed the dependencies
3 - Open xCode project by double clicking on the _Unity-iPhone.xcworkspace_ file created
4 - Build for Generic device
5 - Archive -> When this process is close to ending it will crash generating the trace posted above.
Notice that, after googling for this issue I ended up reading this post [here](https://forum.unity.com/threads/ld-could-not-reparse-bitcode-bundle-on-ios-builds.669451/) which as far as I understand is telling me that the project has a library that has been generated using a more recent xCode build that the one I have, even though I don't quite understand the build number nomenclature here, I understand I am the "reader" here, hence I should have the same version than the producer, right?
|
1.0
|
Invalid bitcode version - <!-- DO NOT DELETE
validate_template=true
template_path=.github/ISSUE_TEMPLATE/issue.md
-->
### [REQUIRED] Please fill in the following fields:
* Unity editor version: 2020.1.3f1
* Firebase Unity SDK version: 6.15.2
* Source you installed the SDK: UPM
* Problematic Firebase Component: ---
* Other Firebase Components in use: Auth, FCM, Crashlytics, Realtime DB
* Additional SDKs you are using: Google Analytics for Firebase/Google App Bundle/Google In-App Review
* Platform you are using the Unity editor on: Mac
* Platform you are targeting: iOS
* Scripting Runtime: IL2CPP
### [REQUIRED] Please describe the issue here:
Stuck when archiving project from xCode v 11.3.1, building from xCode works fine, but when archiving to install the app on a device this error shows up.
`ld: could not reparse object file in bitcode bundle: 'Invalid bitcode version (Producer: '1103.0.32.29.0_0' Reader: '1100.0.33.17_0')', using libLTO version 'LLVM version 11.0.0, (clang-1100.0.33.17)' for architecture arm64`
#### Steps to reproduce:
Have you been able to reproduce this issue with just the Firebase Unity quickstarts (this GitHub project)? Haven't tried
What's the issue repro rate? 100%
What happened? How can we make the problem occur?
1 - Build the game on Unity making sure to select xCode v 11.3.1
2 - Once the build is done and since we are using Firebase, we followed the steps mentioned in the Firebase dashboard when adding an iOS app and installed the pod via `pod install`, that successfully installed the dependencies
3 - Open xCode project by double clicking on the _Unity-iPhone.xcworkspace_ file created
4 - Build for Generic device
5 - Archive -> When this process is close to ending it will crash generating the trace posted above.
Notice that, after googling for this issue I ended up reading this post [here](https://forum.unity.com/threads/ld-could-not-reparse-bitcode-bundle-on-ios-builds.669451/) which as far as I understand is telling me that the project has a library that has been generated using a more recent xCode build that the one I have, even though I don't quite understand the build number nomenclature here, I understand I am the "reader" here, hence I should have the same version than the producer, right?
|
process
|
invalid bitcode version do not delete validate template true template path github issue template issue md please fill in the following fields unity editor version firebase unity sdk version source you installed the sdk upm problematic firebase component other firebase components in use auth fcm crashlytics realtime db additional sdks you are using google analytics for firebase google app bundle google in app review platform you are using the unity editor on mac platform you are targeting ios scripting runtime please describe the issue here stuck when archiving project from xcode v building from xcode works fine but when archiving to install the app on a device this error shows up ld could not reparse object file in bitcode bundle invalid bitcode version producer reader using liblto version llvm version clang for architecture steps to reproduce have you been able to reproduce this issue with just the firebase unity quickstarts this github project haven t tried what s the issue repro rate what happened how can we make the problem occur build the game on unity making sure to select xcode v once the build is done and since we are using firebase we followed the steps mentioned in the firebase dashboard when adding an ios app and installed the pod via pod install that successfully installed the dependencies open xcode project by double clicking on the unity iphone xcworkspace file created build for generic device archive when this process is close to ending it will crash generating the trace posted above notice that after googling for this issue i ended up reading this post which as far as i understand is telling me that the project has a library that has been generated using a more recent xcode build that the one i have even though i don t quite understand the build number nomenclature here i understand i am the reader here hence i should have the same version than the producer right
| 1
|
18,617
| 24,579,387,549
|
IssuesEvent
|
2022-10-13 14:35:01
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[Consent API] Issue related to consentArtifacts get method API.
|
Bug P0 Process: Fixed Process: Tested QA Process: Tested dev
|
Getting permission denied error when executed consentArtifacts get method API in the google cloud console.
**Note:** Issue is not happening frequently.

|
3.0
|
[Consent API] Issue related to consentArtifacts get method API. - Getting permission denied error when executed consentArtifacts get method API in the google cloud console.
**Note:** Issue is not happening frequently.

|
process
|
issue related to consentartifacts get method api getting permission denied error when executed consentartifacts get method api in the google cloud console note issue is not happening frequently
| 1
|
291,827
| 25,178,224,285
|
IssuesEvent
|
2022-11-11 11:12:34
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
opened
|
roachtest: knex failed
|
C-test-failure O-robot O-roachtest release-blocker branch-release-22.2.0
|
roachtest.knex [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/7448384?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/7448384?buildTab=artifacts#/knex) on release-22.2.0 @ [234c9295cc02150f919cfa96b09ee2fa07b68ace](https://github.com/cockroachdb/cockroach/commits/234c9295cc02150f919cfa96b09ee2fa07b68ace):
```
test artifacts and logs in: /artifacts/knex/run_1
knex.go:115,knex.go:126,test_runner.go:930: Non-zero exit code: 2
```
<p>Parameters: <code>ROACHTEST_cloud=gce</code>
, <code>ROACHTEST_cpu=4</code>
, <code>ROACHTEST_encrypted=false</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
<details><summary>Same failure on other branches</summary>
<p>
- #91657 roachtest: knex failed [C-test-failure O-roachtest O-robot T-sql-experience branch-release-22.1 release-blocker]
- #90129 roachtest: knex failed [C-test-failure O-roachtest O-robot T-sql-experience branch-master]
</p>
</details>
/cc @cockroachdb/sql-experience
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*knex.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
2.0
|
roachtest: knex failed - roachtest.knex [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/7448384?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/7448384?buildTab=artifacts#/knex) on release-22.2.0 @ [234c9295cc02150f919cfa96b09ee2fa07b68ace](https://github.com/cockroachdb/cockroach/commits/234c9295cc02150f919cfa96b09ee2fa07b68ace):
```
test artifacts and logs in: /artifacts/knex/run_1
knex.go:115,knex.go:126,test_runner.go:930: Non-zero exit code: 2
```
<p>Parameters: <code>ROACHTEST_cloud=gce</code>
, <code>ROACHTEST_cpu=4</code>
, <code>ROACHTEST_encrypted=false</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
<details><summary>Same failure on other branches</summary>
<p>
- #91657 roachtest: knex failed [C-test-failure O-roachtest O-robot T-sql-experience branch-release-22.1 release-blocker]
- #90129 roachtest: knex failed [C-test-failure O-roachtest O-robot T-sql-experience branch-master]
</p>
</details>
/cc @cockroachdb/sql-experience
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*knex.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
non_process
|
roachtest knex failed roachtest knex with on release test artifacts and logs in artifacts knex run knex go knex go test runner go non zero exit code parameters roachtest cloud gce roachtest cpu roachtest encrypted false roachtest ssd help see see same failure on other branches roachtest knex failed roachtest knex failed cc cockroachdb sql experience
| 0
|
10,011
| 13,043,873,265
|
IssuesEvent
|
2020-07-29 02:54:54
|
tikv/tikv
|
https://api.github.com/repos/tikv/tikv
|
closed
|
UCP: Migrate scalar function `FindInSet` from TiDB
|
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
|
## Description
Port the scalar function `FindInSet` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @lonng
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
2.0
|
UCP: Migrate scalar function `FindInSet` from TiDB -
## Description
Port the scalar function `FindInSet` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @lonng
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
process
|
ucp migrate scalar function findinset from tidb description port the scalar function findinset from tidb to coprocessor score mentor s lonng recommended skills rust programming learning materials already implemented expressions ported from tidb
| 1
|
181,945
| 21,664,469,719
|
IssuesEvent
|
2022-05-07 01:27:21
|
eldorplus/portfolio
|
https://api.github.com/repos/eldorplus/portfolio
|
closed
|
CVE-2015-9251 (Medium) detected in jquery-2.1.4.min.js, jquery-1.7.1.min.js - autoclosed
|
security vulnerability
|
## CVE-2015-9251 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-2.1.4.min.js</b>, <b>jquery-1.7.1.min.js</b></p></summary>
<p>
<details><summary><b>jquery-2.1.4.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.4/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.4/jquery.min.js</a></p>
<p>Path to dependency file: /tmp/ws-scm/portfolio/node_modules/js-base64/test/index.html</p>
<p>Path to vulnerable library: /portfolio/node_modules/js-base64/test/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-2.1.4.min.js** (Vulnerable Library)
</details>
<details><summary><b>jquery-1.7.1.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js</a></p>
<p>Path to dependency file: /tmp/ws-scm/portfolio/node_modules/sockjs/examples/hapi/html/index.html</p>
<p>Path to vulnerable library: /portfolio/node_modules/sockjs/examples/hapi/html/index.html,/portfolio/node_modules/sockjs/examples/express-3.x/index.html,/portfolio/node_modules/sockjs/examples/multiplex/index.html,/portfolio/node_modules/vm-browserify/example/run/index.html,/portfolio/node_modules/sockjs/examples/echo/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.7.1.min.js** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/eldorplus/portfolio/commit/cd12ed4db80ee20275bd47147f3f8844df75f8c7">cd12ed4db80ee20275bd47147f3f8844df75f8c7</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed.
<p>Publish Date: 2018-01-18
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2015-9251>CVE-2015-9251</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p>
<p>Release Date: 2018-01-18</p>
<p>Fix Resolution: jQuery - v3.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2015-9251 (Medium) detected in jquery-2.1.4.min.js, jquery-1.7.1.min.js - autoclosed - ## CVE-2015-9251 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-2.1.4.min.js</b>, <b>jquery-1.7.1.min.js</b></p></summary>
<p>
<details><summary><b>jquery-2.1.4.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.4/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.4/jquery.min.js</a></p>
<p>Path to dependency file: /tmp/ws-scm/portfolio/node_modules/js-base64/test/index.html</p>
<p>Path to vulnerable library: /portfolio/node_modules/js-base64/test/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-2.1.4.min.js** (Vulnerable Library)
</details>
<details><summary><b>jquery-1.7.1.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js</a></p>
<p>Path to dependency file: /tmp/ws-scm/portfolio/node_modules/sockjs/examples/hapi/html/index.html</p>
<p>Path to vulnerable library: /portfolio/node_modules/sockjs/examples/hapi/html/index.html,/portfolio/node_modules/sockjs/examples/express-3.x/index.html,/portfolio/node_modules/sockjs/examples/multiplex/index.html,/portfolio/node_modules/vm-browserify/example/run/index.html,/portfolio/node_modules/sockjs/examples/echo/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.7.1.min.js** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/eldorplus/portfolio/commit/cd12ed4db80ee20275bd47147f3f8844df75f8c7">cd12ed4db80ee20275bd47147f3f8844df75f8c7</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed.
<p>Publish Date: 2018-01-18
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2015-9251>CVE-2015-9251</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p>
<p>Release Date: 2018-01-18</p>
<p>Fix Resolution: jQuery - v3.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in jquery min js jquery min js autoclosed cve medium severity vulnerability vulnerable libraries jquery min js jquery min js jquery min js javascript library for dom operations library home page a href path to dependency file tmp ws scm portfolio node modules js test index html path to vulnerable library portfolio node modules js test index html dependency hierarchy x jquery min js vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file tmp ws scm portfolio node modules sockjs examples hapi html index html path to vulnerable library portfolio node modules sockjs examples hapi html index html portfolio node modules sockjs examples express x index html portfolio node modules sockjs examples multiplex index html portfolio node modules vm browserify example run index html portfolio node modules sockjs examples echo index html dependency hierarchy x jquery min js vulnerable library found in head commit a href vulnerability details jquery before is vulnerable to cross site scripting xss attacks when a cross domain ajax request is performed without the datatype option causing text javascript responses to be executed publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with whitesource
| 0
|
320,021
| 9,763,685,744
|
IssuesEvent
|
2019-06-05 14:18:54
|
PecanProject/pecan
|
https://api.github.com/repos/PecanProject/pecan
|
closed
|
Meta-analysis AUTO broken ... again
|
Priority: 03 - High Type: Bug
|
Meta-analysis seems to be running for every run again. Need to add debugging this so it only runs when needed back on the list. I'm seriously tempted to just remove the MA block from the default run specification so that users have to turn it on during advanced edit.
|
1.0
|
Meta-analysis AUTO broken ... again - Meta-analysis seems to be running for every run again. Need to add debugging this so it only runs when needed back on the list. I'm seriously tempted to just remove the MA block from the default run specification so that users have to turn it on during advanced edit.
|
non_process
|
meta analysis auto broken again meta analysis seems to be running for every run again need to add debugging this so it only runs when needed back on the list i m seriously tempted to just remove the ma block from the default run specification so that users have to turn it on during advanced edit
| 0
|
451,392
| 13,034,658,361
|
IssuesEvent
|
2020-07-28 09:03:56
|
StrangeLoopGames/EcoIssues
|
https://api.github.com/repos/StrangeLoopGames/EcoIssues
|
opened
|
[0.9.0 staging-1686] Crafting UI glitches
|
Category: UI Priority: Medium
|
- [ ] 1. Dark stripe and light stripe have different width:

- [ ] 2. Sometimes string color don't update if you levelup or take specialty (String color remains the same, but the order is changed):

Just open some craft table and start to learn/specialize/levelup skills
|
1.0
|
[0.9.0 staging-1686] Crafting UI glitches - - [ ] 1. Dark stripe and light stripe have different width:

- [ ] 2. Sometimes string color don't update if you levelup or take specialty (String color remains the same, but the order is changed):

Just open some craft table and start to learn/specialize/levelup skills
|
non_process
|
crafting ui glitches dark stripe and light stripe have different width sometimes string color don t update if you levelup or take specialty string color remains the same but the order is changed just open some craft table and start to learn specialize levelup skills
| 0
|
10,176
| 13,044,162,788
|
IssuesEvent
|
2020-07-29 03:47:36
|
tikv/tikv
|
https://api.github.com/repos/tikv/tikv
|
closed
|
UCP: Migrate scalar function `RoundWithFracDec` from TiDB
|
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
|
## Description
Port the scalar function `RoundWithFracDec` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @breeswish
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
2.0
|
UCP: Migrate scalar function `RoundWithFracDec` from TiDB -
## Description
Port the scalar function `RoundWithFracDec` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @breeswish
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
process
|
ucp migrate scalar function roundwithfracdec from tidb description port the scalar function roundwithfracdec from tidb to coprocessor score mentor s breeswish recommended skills rust programming learning materials already implemented expressions ported from tidb
| 1
|
17,371
| 23,197,812,461
|
IssuesEvent
|
2022-08-01 18:10:43
|
vectordotdev/vector
|
https://api.github.com/repos/vectordotdev/vector
|
closed
|
Filter blank messages
|
source: file source: syslog source: socket source: stdin type: enhancement meta: idea needs: approval domain: processing domain: filtering
|
We should provide a configuration option to drop messages that only contain white-space. This is pretty common and this type of data just adds noise. This could be:
1. A global setting
2. A per source setting
3. A transformer that must be installed as part of the pipeline
|
1.0
|
Filter blank messages - We should provide a configuration option to drop messages that only contain white-space. This is pretty common and this type of data just adds noise. This could be:
1. A global setting
2. A per source setting
3. A transformer that must be installed as part of the pipeline
|
process
|
filter blank messages we should provide a configuration option to drop messages that only contain white space this is pretty common and this type of data just adds noise this could be a global setting a per source setting a transformer that must be installed as part of the pipeline
| 1
|
110,567
| 23,957,245,327
|
IssuesEvent
|
2022-09-12 15:52:19
|
Regalis11/Barotrauma
|
https://api.github.com/repos/Regalis11/Barotrauma
|
closed
|
Husk infection overwrites water prankster
|
Bug Code Unstable
|
### Disclaimers
- [X] I have searched the issue tracker to check if the issue has already been reported.
- [ ] My issue happened while using mods.
### What happened?
Was playing in Multiplayer (reproduced in Single Player Campaign) as assistant. Had "Water Prankster" talent as well as Husk Genes / Infection (Think of it as a poor mans perfect Thresher Genes). Was sunk in the Abyss due to Endworm. Swam out with nothing but Clown Suit/Mask and Sonar Beacon to try and distract it. Blew up to pressure immediately upon exiting the sub. Expected Water Prankster pressure immunity to allow me to exit sub safely.
Testing in Single Player and had character approx 1000m below the outpost, 400m below the bottom of the map. Screenshots show Clown Power active and character safely in the Abyss. 2nd Screenshot shows clown power still active yet character died. Character was not moved between 2 screenshots.


### Reproduction steps
1. Start Campaign (Single/Multi) as Assistant.
2. Obtain/Wear the Clown Suit/Mask
3. Obtain 2 Levels and select "Water Prankster" as Tier 2 talent
4. Obtain Husk Genetic Material + Gene Splicer. Use on self
5. Swim to "Abyss" (Testing was done at approx 400m below the map). Ensure Clown Power is present and no Dive Suit worn
6. Note no negative effects
7. Infect self with Husk Infection
8. Upon Husk Infection reaching "Heavy" stage (where you can press F to infect creatures) character immediately implodes due to pressure
### Bug prevalence
Happens every time I play
### Version
0.18.15.1 / 0.18.15.2 on macOS
### -
_No response_
### Which operating system did you encounter this bug on?
Windows
### Relevant error messages and crash reports
_No response_
|
1.0
|
Husk infection overwrites water prankster - ### Disclaimers
- [X] I have searched the issue tracker to check if the issue has already been reported.
- [ ] My issue happened while using mods.
### What happened?
Was playing in Multiplayer (reproduced in Single Player Campaign) as assistant. Had "Water Prankster" talent as well as Husk Genes / Infection (Think of it as a poor mans perfect Thresher Genes). Was sunk in the Abyss due to Endworm. Swam out with nothing but Clown Suit/Mask and Sonar Beacon to try and distract it. Blew up to pressure immediately upon exiting the sub. Expected Water Prankster pressure immunity to allow me to exit sub safely.
Testing in Single Player and had character approx 1000m below the outpost, 400m below the bottom of the map. Screenshots show Clown Power active and character safely in the Abyss. 2nd Screenshot shows clown power still active yet character died. Character was not moved between 2 screenshots.


### Reproduction steps
1. Start Campaign (Single/Multi) as Assistant.
2. Obtain/Wear the Clown Suit/Mask
3. Obtain 2 Levels and select "Water Prankster" as Tier 2 talent
4. Obtain Husk Genetic Material + Gene Splicer. Use on self
5. Swim to "Abyss" (Testing was done at approx 400m below the map). Ensure Clown Power is present and no Dive Suit worn
6. Note no negative effects
7. Infect self with Husk Infection
8. Upon Husk Infection reaching "Heavy" stage (where you can press F to infect creatures) character immediately implodes due to pressure
### Bug prevalence
Happens every time I play
### Version
0.18.15.1 / 0.18.15.2 on macOS
### -
_No response_
### Which operating system did you encounter this bug on?
Windows
### Relevant error messages and crash reports
_No response_
|
non_process
|
husk infection overwrites water prankster disclaimers i have searched the issue tracker to check if the issue has already been reported my issue happened while using mods what happened was playing in multiplayer reproduced in single player campaign as assistant had water prankster talent as well as husk genes infection think of it as a poor mans perfect thresher genes was sunk in the abyss due to endworm swam out with nothing but clown suit mask and sonar beacon to try and distract it blew up to pressure immediately upon exiting the sub expected water prankster pressure immunity to allow me to exit sub safely testing in single player and had character approx below the outpost below the bottom of the map screenshots show clown power active and character safely in the abyss screenshot shows clown power still active yet character died character was not moved between screenshots reproduction steps start campaign single multi as assistant obtain wear the clown suit mask obtain levels and select water prankster as tier talent obtain husk genetic material gene splicer use on self swim to abyss testing was done at approx below the map ensure clown power is present and no dive suit worn note no negative effects infect self with husk infection upon husk infection reaching heavy stage where you can press f to infect creatures character immediately implodes due to pressure bug prevalence happens every time i play version on macos no response which operating system did you encounter this bug on windows relevant error messages and crash reports no response
| 0
|
138,913
| 20,742,997,597
|
IssuesEvent
|
2022-03-14 19:36:58
|
EscolaDeSaudePublica/DesignLab
|
https://api.github.com/repos/EscolaDeSaudePublica/DesignLab
|
closed
|
3. Apoiar a Construção da Política da Pessoa com Deficiência (item 3.9.2)
|
Oficina Design Sem Projeto Definido Prioridade Design: Alta
|
## **Objetivo**
**Como** designer
**Quero** apoiar a construção da Política Estadual das Pessoas com Deficiência
**Para** nortear a elaboração do Plano Estadual das Políticas para Pessoas com Deficiência do Ceará (2022-2023)
## **Contexto**
- A Secretaria Executiva de Políticas de Saúde - SEPOS/SESA, solicitou à Superintendência da ESP/CE, por meio do ofício 10/2022, apoio do DesignLab/Felicilab para facilitação na construção da Política Estadual das Pessoas com Deficiência. O pedido partiu da experiência na I Oficina de Design de Serviços Públicos, realizada com a Sepos em Ago/2021. Foi definido que a Adins será a responsável pela condução do processo e o DesignLab entrará como apoio.
## **Escopo**
- [x] 3.9.2 Realizar oficina - para lideranças do seguimento (10/03)
## Observações
Épico: [Oficinas de Design para Políticas Públicas #225](https://github.com/EscolaDeSaudePublica/DesignLab/issues/225)
Pessoa de Contato: Luciana Donato da Sepos.

|
2.0
|
3. Apoiar a Construção da Política da Pessoa com Deficiência (item 3.9.2) - ## **Objetivo**
**Como** designer
**Quero** apoiar a construção da Política Estadual das Pessoas com Deficiência
**Para** nortear a elaboração do Plano Estadual das Políticas para Pessoas com Deficiência do Ceará (2022-2023)
## **Contexto**
- A Secretaria Executiva de Políticas de Saúde - SEPOS/SESA, solicitou à Superintendência da ESP/CE, por meio do ofício 10/2022, apoio do DesignLab/Felicilab para facilitação na construção da Política Estadual das Pessoas com Deficiência. O pedido partiu da experiência na I Oficina de Design de Serviços Públicos, realizada com a Sepos em Ago/2021. Foi definido que a Adins será a responsável pela condução do processo e o DesignLab entrará como apoio.
## **Escopo**
- [x] 3.9.2 Realizar oficina - para lideranças do seguimento (10/03)
## Observações
Épico: [Oficinas de Design para Políticas Públicas #225](https://github.com/EscolaDeSaudePublica/DesignLab/issues/225)
Pessoa de Contato: Luciana Donato da Sepos.

|
non_process
|
apoiar a construção da política da pessoa com deficiência item objetivo como designer quero apoiar a construção da política estadual das pessoas com deficiência para nortear a elaboração do plano estadual das políticas para pessoas com deficiência do ceará contexto a secretaria executiva de políticas de saúde sepos sesa solicitou à superintendência da esp ce por meio do ofício apoio do designlab felicilab para facilitação na construção da política estadual das pessoas com deficiência o pedido partiu da experiência na i oficina de design de serviços públicos realizada com a sepos em ago foi definido que a adins será a responsável pela condução do processo e o designlab entrará como apoio escopo realizar oficina para lideranças do seguimento observações épico pessoa de contato luciana donato da sepos
| 0
|
85,211
| 10,432,118,042
|
IssuesEvent
|
2019-09-17 10:32:23
|
lemzoo/flask-rabbitmq
|
https://api.github.com/repos/lemzoo/flask-rabbitmq
|
closed
|
Update the README content
|
documentation
|
The readme file is not updated. Please do not refer it. I will make the update as soon as possible
|
1.0
|
Update the README content - The readme file is not updated. Please do not refer it. I will make the update as soon as possible
|
non_process
|
update the readme content the readme file is not updated please do not refer it i will make the update as soon as possible
| 0
|
21,320
| 28,807,870,850
|
IssuesEvent
|
2023-05-03 00:17:41
|
NCAR/ucomp-pipeline
|
https://api.github.com/repos/NCAR/ucomp-pipeline
|
opened
|
Fix post angle finding
|
process
|
The post angle varies wildly instead of converging to the correct value.
|
1.0
|
Fix post angle finding - The post angle varies wildly instead of converging to the correct value.
|
process
|
fix post angle finding the post angle varies wildly instead of converging to the correct value
| 1
|
17,170
| 22,745,004,716
|
IssuesEvent
|
2022-07-07 08:24:32
|
2i2c-org/infrastructure
|
https://api.github.com/repos/2i2c-org/infrastructure
|
closed
|
Cloud usage monitoring and alerting infrastructure and process
|
type: enhancement :label: team-process :label: optimization :label: monitoring
|
### Description of problem and opportunity to address it
**Problem description**
In #908 we ran into a case where a user was abusing the JupyterHub for crypto mining. This resulted in a lot of stress and high costs for the hub's community. Part of the problem was that we did not detect the mining activity for several weeks. This activity was basically:
- The steady creation of new users on the hub
- Each user maxing out their CPU and never shutting down their session
**Proposed solution**
We should create a mechanism for automatically monitoring statistics around hub usage, and triggering notifications that suggest something nefarious is happening. Ideally, this would be a single process for all of our clusters, not one process for each cluster.
We need a quick way to:
1. Keep an eye on all these projects *in one place*
2. Have automated alerts for abnormal costs
3. Do rounds of cost optimizations
**What's the value and who would benefit**
This would allow us to minimize the risk of abuse if somebody *did* try to use a hub for the wrong purposes. It would give our team more confidence that something isn't happening without us knowing about it, and would give communities more confidence that they won't have an unexpected spike in their cloud bill.
### Implementation guide and constraints
A rough idea of what to try:
- Set up a Grafana dashboard that aggregates activity across all of our clusters (this will be tricky because the Prometheus instances are *private* for our clusters, not public like the Binder ones).
- Define a few metrics that are particularly useful for identifying abuse and problematic abnormal behavior. For example, here are two images from the openscapes grafana that were particularly useful:
- **Users over time**
<img src="https://user-images.githubusercontent.com/1839645/148128329-1d8bd83c-3a2e-4b83-91ae-91ff33769bd5.png" width=400 />
- **CPU usage histogram over time**
<img src="https://user-images.githubusercontent.com/1839645/148128328-58d190e3-99ce-4590-984d-258e2c46cd79.png" width=400 />
- And noting `5xx` errors from user pods in general is a good indication that something is wrong.
- Define some thresholds for these metrics, and create a reporting mechanism to ping `support@2i2c.org` when it thinks something problematic is going on.
### Issues where we have been bitten by this
- https://github.com/2i2c-org/infrastructure/issues/1081
- https://github.com/2i2c-org/infrastructure/issues/908
### Updates and ongoing work
#### 2022-01-06
@GeorgianaElena is going to work on these things for **one week**:
- [x] How complex will it be to aggregate feeds from each cluster's Prometheus?
- [x] See this HackMD for a brief analysis: https://hackmd.io/HqE3RgjtTBq1MuofvAiLlQ?view
See https://github.com/2i2c-org/infrastructure/issues/328#issuecomment-1006806575 for more details!
#### 2022-01-19
Some meeting notes around here: https://github.com/2i2c-org/infrastructure/issues/328#issuecomment-1016635397
We agreed that the best way forward is to start by implementing option 1 from the HackMD above, which is to follow the mybinder.org model of **one Grafana with multiple data sources**.
Our next steps here are to:
- [x] Follow @yuvipanda's advice about setting up ingress/auth for a Grafana instance that pulls from each Prometheus instance in all clusters: https://github.com/2i2c-org/infrastructure/issues/328#issuecomment-1016635397 ➡️ https://github.com/2i2c-org/infrastructure/pull/1091
- [x] Decide how much complexity this will add to our setup, and whether this justifies looking at a different option
- [ ] If it works, create a few dashboards in Grafana that we can use for reporting
#### 2022-03-30
*From https://github.com/2i2c-org/infrastructure/issues/328#issuecomment-1082542191*
- [x] Write a script that'll read all the encrypted grafana secrets, and put them in the centralized grafana as data sources via the grafana API ➡️ https://github.com/2i2c-org/infrastructure/pull/1215
- [x] Update the upstream jupyterhub/grafana-dashboard repo to support multiple datasources, via a datasource template variable. ➡️ https://github.com/jupyterhub/grafana-dashboards/pull/37
- [x] Deploy support charts in the few clusters where we don't currently have them deployed! I think that's meom-ige and farallon? We will need to tune their resource requests to match the smaller clusters.
- [x] Split support dashboards -> https://github.com/jupyterhub/grafana-dashboards/pull/38
- [x] Support dashboards across all clusters -> https://github.com/jupyterhub/grafana-dashboards/pull/39
|
1.0
|
Cloud usage monitoring and alerting infrastructure and process - ### Description of problem and opportunity to address it
**Problem description**
In #908 we ran into a case where a user was abusing the JupyterHub for crypto mining. This resulted in a lot of stress and high costs for the hub's community. Part of the problem was that we did not detect the mining activity for several weeks. This activity was basically:
- The steady creation of new users on the hub
- Each user maxing out their CPU and never shutting down their session
**Proposed solution**
We should create a mechanism for automatically monitoring statistics around hub usage, and triggering notifications that suggest something nefarious is happening. Ideally, this would be a single process for all of our clusters, not one process for each cluster.
We need a quick way to:
1. Keep an eye on all these projects *in one place*
2. Have automated alerts for abnormal costs
3. Do rounds of cost optimizations
**What's the value and who would benefit**
This would allow us to minimize the risk of abuse if somebody *did* try to use a hub for the wrong purposes. It would give our team more confidence that something isn't happening without us knowing about it, and would give communities more confidence that they won't have an unexpected spike in their cloud bill.
### Implementation guide and constraints
A rough idea of what to try:
- Set up a Grafana dashboard that aggregates activity across all of our clusters (this will be tricky because the Prometheus instances are *private* for our clusters, not public like the Binder ones).
- Define a few metrics that are particularly useful for identifying abuse and problematic abnormal behavior. For example, here are two images from the openscapes grafana that were particularly useful:
- **Users over time**
<img src="https://user-images.githubusercontent.com/1839645/148128329-1d8bd83c-3a2e-4b83-91ae-91ff33769bd5.png" width=400 />
- **CPU usage histogram over time**
<img src="https://user-images.githubusercontent.com/1839645/148128328-58d190e3-99ce-4590-984d-258e2c46cd79.png" width=400 />
- And noting `5xx` errors from user pods in general is a good indication that something is wrong.
- Define some thresholds for these metrics, and create a reporting mechanism to ping `support@2i2c.org` when it thinks something problematic is going on.
### Issues where we have been bitten by this
- https://github.com/2i2c-org/infrastructure/issues/1081
- https://github.com/2i2c-org/infrastructure/issues/908
### Updates and ongoing work
#### 2022-01-06
@GeorgianaElena is going to work on these things for **one week**:
- [x] How complex will it be to aggregate feeds from each cluster's Prometheus?
- [x] See this HackMD for a brief analysis: https://hackmd.io/HqE3RgjtTBq1MuofvAiLlQ?view
See https://github.com/2i2c-org/infrastructure/issues/328#issuecomment-1006806575 for more details!
#### 2022-01-19
Some meeting notes around here: https://github.com/2i2c-org/infrastructure/issues/328#issuecomment-1016635397
We agreed that the best way forward is to start by implementing option 1 from the HackMD above, which is to follow the mybinder.org model of **one Grafana with multiple data sources**.
Our next steps here are to:
- [x] Follow @yuvipanda's advice about setting up ingress/auth for a Grafana instance that pulls from each Prometheus instance in all clusters: https://github.com/2i2c-org/infrastructure/issues/328#issuecomment-1016635397 ➡️ https://github.com/2i2c-org/infrastructure/pull/1091
- [x] Decide how much complexity this will add to our setup, and whether this justifies looking at a different option
- [ ] If it works, create a few dashboards in Grafana that we can use for reporting
#### 2022-03-30
*From https://github.com/2i2c-org/infrastructure/issues/328#issuecomment-1082542191*
- [x] Write a script that'll read all the encrypted grafana secrets, and put them in the centralized grafana as data sources via the grafana API ➡️ https://github.com/2i2c-org/infrastructure/pull/1215
- [x] Update the upstream jupyterhub/grafana-dashboard repo to support multiple datasources, via a datasource template variable. ➡️ https://github.com/jupyterhub/grafana-dashboards/pull/37
- [x] Deploy support charts in the few clusters where we don't currently have them deployed! I think that's meom-ige and farallon? We will need to tune their resource requests to match the smaller clusters.
- [x] Split support dashboards -> https://github.com/jupyterhub/grafana-dashboards/pull/38
- [x] Support dashboards across all clusters -> https://github.com/jupyterhub/grafana-dashboards/pull/39
|
process
|
cloud usage monitoring and alerting infrastructure and process description of problem and opportunity to address it problem description in we ran into a case where a user was abusing the jupyterhub for crypto mining this resulted in a lot of stress and high costs for the hub s community part of the problem was that we did not detect the mining activity for several weeks this activity was basically the steady creation of new users on the hub each user maxing out their cpu and never shutting down their session proposed solution we should create a mechanism for automatically monitoring statistics around hub usage and triggering notifications that suggest something nefarious is happening ideally this would be a single process for all of our clusters not one process for each cluster we need a quick way to keep an eye on all these projects in one place have automated alerts for abnormal costs do rounds of cost optimizations what s the value and who would benefit this would allow us to minimize the risk of abuse if somebody did try to use a hub for the wrong purposes it would give our team more confidence that something isn t happening without us knowing about it and would give communities more confidence that they won t have an unexpected spike in their cloud bill implementation guide and constraints a rough idea of what to try set up a grafana dashboard that aggregates activity across all of our clusters this will be tricky because the prometheus instances are private for our clusters not public like the binder ones define a few metrics that are particularly useful for identifying abuse and problematic abnormal behavior for example here are two images from the openscapes grafana that were particularly useful users over time cpu usage histogram over time and noting errors from user pods in general is a good indication that something is wrong define some thresholds for these metrics and create a reporting mechanism to ping support org when it thinks something problematic is going on issues where we have been bitten by this updates and ongoing work georgianaelena is going to work on these things for one week how complex will it be to aggregate feeds from each cluster s prometheus see this hackmd for a brief analysis see for more details some meeting notes around here we agreed that the best way forward is to start by implementing option from the hackmd above which is to follow the mybinder org model of one grafana with multiple data sources our next steps here are to follow yuvipanda s advice about setting up ingress auth for a grafana instance that pulls from each prometheus instance in all clusters ➡️ decide how much complexity this will add to our setup and whether this justifies looking at a different option if it works create a few dashboards in grafana that we can use for reporting from write a script that ll read all the encrypted grafana secrets and put them in the centralized grafana as data sources via the grafana api ➡️ update the upstream jupyterhub grafana dashboard repo to support multiple datasources via a datasource template variable ➡️ deploy support charts in the few clusters where we don t currently have them deployed i think that s meom ige and farallon we will need to tune their resource requests to match the smaller clusters split support dashboards support dashboards across all clusters
| 1
|
22,236
| 30,784,785,937
|
IssuesEvent
|
2023-07-31 12:34:09
|
NationalSecurityAgency/ghidra
|
https://api.github.com/repos/NationalSecurityAgency/ghidra
|
closed
|
x64: failure to disassemble LIDT instructions
|
Type: Bug Feature: Processor/x86 Status: Internal
|
**Describe the bug**
The following are several (de-duped) byte patterns that Ghidra fails to disassemble on an x64 program, with corresponding disassembly from Capstone/IDA
```
0f 01 9d 98 01 00 00
lidt [rbp + 0x198]
0f 01 9f f6 00 00 00
lidt [rdi + 0xf6]
0f 01 59 66
lidt [rcx + 0x66]
0f 01 5d f7
lidt [rbp - 9]
```



|
1.0
|
x64: failure to disassemble LIDT instructions - **Describe the bug**
The following are several (de-duped) byte patterns that Ghidra fails to disassemble on an x64 program, with corresponding disassembly from Capstone/IDA
```
0f 01 9d 98 01 00 00
lidt [rbp + 0x198]
0f 01 9f f6 00 00 00
lidt [rdi + 0xf6]
0f 01 59 66
lidt [rcx + 0x66]
0f 01 5d f7
lidt [rbp - 9]
```



|
process
|
failure to disassemble lidt instructions describe the bug the following are several de duped byte patterns that ghidra fails to disassemble on an program with corresponding disassembly from capstone ida lidt lidt lidt lidt
| 1
|
168,319
| 26,630,636,511
|
IssuesEvent
|
2023-01-24 17:33:14
|
revoltchat/revite
|
https://api.github.com/repos/revoltchat/revite
|
closed
|
Improve memory usage / performance of message renderer.
|
improvement performance needs design
|
There seems to be some sort of memory leak coming from the message renderer. The code isn't the greatest, maybe we can use a MobX Observable array.
|
1.0
|
Improve memory usage / performance of message renderer. - There seems to be some sort of memory leak coming from the message renderer. The code isn't the greatest, maybe we can use a MobX Observable array.
|
non_process
|
improve memory usage performance of message renderer there seems to be some sort of memory leak coming from the message renderer the code isn t the greatest maybe we can use a mobx observable array
| 0
|
33,209
| 4,818,373,968
|
IssuesEvent
|
2016-11-04 16:11:44
|
dotnet/corefx
|
https://api.github.com/repos/dotnet/corefx
|
opened
|
System.Net.Sockets.Async.Performance.Tests threw ObjectDisposedException
|
area-System.Net test-run-core
|
No results generated.
https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20161103.04/workItem/System.Net.Sockets.Async.Performance.Tests/details/ef1c1a72-3184-4848-9894-303a848b3635
```2016-11-03 15:14:31,896: INFO: proc(65): run_and_log_output: Output: Starting: System.Net.Sockets.Async.Performance.Tests
2016-11-03 15:14:32,255: INFO: proc(65): run_and_log_output: Output:
2016-11-03 15:14:32,255: INFO: proc(65): run_and_log_output: Output: Unhandled Exception: System.ObjectDisposedException: Cannot access a disposed object.
2016-11-03 15:14:32,256: INFO: proc(65): run_and_log_output: Output: Object name: 'System.Net.Sockets.Socket'.
2016-11-03 15:14:32,256: INFO: proc(65): run_and_log_output: Output: at System.Net.Sockets.Socket.ReceiveAsync(SocketAsyncEventArgs e)
2016-11-03 15:14:32,256: INFO: proc(65): run_and_log_output: Output: at System.Net.Sockets.Performance.Tests.SocketTestClientAsync.Receive(Action`2 onReceiveCallback)
2016-11-03 15:14:32,256: INFO: proc(65): run_and_log_output: Output: at System.Threading.ExecutionContext.Run(ExecutionContext executionContext, ContextCallback callback, Object state)
2016-11-03 15:14:32,256: INFO: proc(65): run_and_log_output: Output: at System.Net.Sockets.MultipleConnectAsync.Succeed()
2016-11-03 15:14:32,256: INFO: proc(65): run_and_log_output: Output: at System.Net.Sockets.MultipleConnectAsync.InternalConnectCallback(Object sender, SocketAsyncEventArgs args)
2016-11-03 15:14:32,257: INFO: proc(65): run_and_log_output: Output: at System.Threading.ExecutionContext.Run(ExecutionContext executionContext, ContextCallback callback, Object state)
2016-11-03 15:14:32,257: INFO: proc(65): run_and_log_output: Output: at System.Threading.ExecutionContext.Run(ExecutionContext executionContext, ContextCallback callback, Object state)
2016-11-03 15:14:32,257: INFO: proc(65): run_and_log_output: Output: at System.Threading.QueueUserWorkItemCallbackDefaultContext.System.Threading.IThreadPoolWorkItem.ExecuteWorkItem()
2016-11-03 15:14:32,257: INFO: proc(65): run_and_log_output: Output: at System.Threading.ThreadPoolWorkQueue.Dispatch()
2016-11-03 15:14:32,259: INFO: proc(65): run_and_log_output: Output: /Users/dotnet-bot/dotnetbuild/work/ef1c1a72-3184-4848-9894-303a848b3635/Work/4effeedc-b714-487c-9ce0-32e9cb61ae1b/Unzip/RunTests.sh: line 284: 29083 Abort trap: 6 ./corerun xunit.console.netcore.exe System.Net.Sockets.Async.Performance.Tests.dll -xml testResults.xml -notrait Benchmark=true -notrait category=nonnetcoreapp1.1tests -notrait category=requireselevation -notrait category=failing -notrait category=nonosxtests
```
|
1.0
|
System.Net.Sockets.Async.Performance.Tests threw ObjectDisposedException - No results generated.
https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20161103.04/workItem/System.Net.Sockets.Async.Performance.Tests/details/ef1c1a72-3184-4848-9894-303a848b3635
```2016-11-03 15:14:31,896: INFO: proc(65): run_and_log_output: Output: Starting: System.Net.Sockets.Async.Performance.Tests
2016-11-03 15:14:32,255: INFO: proc(65): run_and_log_output: Output:
2016-11-03 15:14:32,255: INFO: proc(65): run_and_log_output: Output: Unhandled Exception: System.ObjectDisposedException: Cannot access a disposed object.
2016-11-03 15:14:32,256: INFO: proc(65): run_and_log_output: Output: Object name: 'System.Net.Sockets.Socket'.
2016-11-03 15:14:32,256: INFO: proc(65): run_and_log_output: Output: at System.Net.Sockets.Socket.ReceiveAsync(SocketAsyncEventArgs e)
2016-11-03 15:14:32,256: INFO: proc(65): run_and_log_output: Output: at System.Net.Sockets.Performance.Tests.SocketTestClientAsync.Receive(Action`2 onReceiveCallback)
2016-11-03 15:14:32,256: INFO: proc(65): run_and_log_output: Output: at System.Threading.ExecutionContext.Run(ExecutionContext executionContext, ContextCallback callback, Object state)
2016-11-03 15:14:32,256: INFO: proc(65): run_and_log_output: Output: at System.Net.Sockets.MultipleConnectAsync.Succeed()
2016-11-03 15:14:32,256: INFO: proc(65): run_and_log_output: Output: at System.Net.Sockets.MultipleConnectAsync.InternalConnectCallback(Object sender, SocketAsyncEventArgs args)
2016-11-03 15:14:32,257: INFO: proc(65): run_and_log_output: Output: at System.Threading.ExecutionContext.Run(ExecutionContext executionContext, ContextCallback callback, Object state)
2016-11-03 15:14:32,257: INFO: proc(65): run_and_log_output: Output: at System.Threading.ExecutionContext.Run(ExecutionContext executionContext, ContextCallback callback, Object state)
2016-11-03 15:14:32,257: INFO: proc(65): run_and_log_output: Output: at System.Threading.QueueUserWorkItemCallbackDefaultContext.System.Threading.IThreadPoolWorkItem.ExecuteWorkItem()
2016-11-03 15:14:32,257: INFO: proc(65): run_and_log_output: Output: at System.Threading.ThreadPoolWorkQueue.Dispatch()
2016-11-03 15:14:32,259: INFO: proc(65): run_and_log_output: Output: /Users/dotnet-bot/dotnetbuild/work/ef1c1a72-3184-4848-9894-303a848b3635/Work/4effeedc-b714-487c-9ce0-32e9cb61ae1b/Unzip/RunTests.sh: line 284: 29083 Abort trap: 6 ./corerun xunit.console.netcore.exe System.Net.Sockets.Async.Performance.Tests.dll -xml testResults.xml -notrait Benchmark=true -notrait category=nonnetcoreapp1.1tests -notrait category=requireselevation -notrait category=failing -notrait category=nonosxtests
```
|
non_process
|
system net sockets async performance tests threw objectdisposedexception no results generated info proc run and log output output starting system net sockets async performance tests info proc run and log output output info proc run and log output output unhandled exception system objectdisposedexception cannot access a disposed object info proc run and log output output object name system net sockets socket info proc run and log output output at system net sockets socket receiveasync socketasynceventargs e info proc run and log output output at system net sockets performance tests sockettestclientasync receive action onreceivecallback info proc run and log output output at system threading executioncontext run executioncontext executioncontext contextcallback callback object state info proc run and log output output at system net sockets multipleconnectasync succeed info proc run and log output output at system net sockets multipleconnectasync internalconnectcallback object sender socketasynceventargs args info proc run and log output output at system threading executioncontext run executioncontext executioncontext contextcallback callback object state info proc run and log output output at system threading executioncontext run executioncontext executioncontext contextcallback callback object state info proc run and log output output at system threading queueuserworkitemcallbackdefaultcontext system threading ithreadpoolworkitem executeworkitem info proc run and log output output at system threading threadpoolworkqueue dispatch info proc run and log output output users dotnet bot dotnetbuild work work unzip runtests sh line abort trap corerun xunit console netcore exe system net sockets async performance tests dll xml testresults xml notrait benchmark true notrait category notrait category requireselevation notrait category failing notrait category nonosxtests
| 0
|
14,984
| 18,524,076,376
|
IssuesEvent
|
2021-10-20 18:13:13
|
googleapis/python-bigtable
|
https://api.github.com/repos/googleapis/python-bigtable
|
closed
|
'test_cluster_create' systest flakes with 'TimeoutError'
|
api: bigtable type: process
|
From [this failed systest build](https://source.cloud.google.com/results/invocations/924fbb8c-f387-482d-8767-c16abd2de9e2/targets/cloud-devrel%2Fclient-libraries%2Fpython%2Fgoogleapis%2Fpython-bigtable%2Fpresubmit%2Fsystem-3.8/log):
```python
_____________________________ test_cluster_create ______________________________
target = functools.partial(<bound method PollingFuture._done_or_raise of <google.api_core.operation.Operation object at 0x7f4d5c321d30>>)
predicate = <function if_exception_type.<locals>.if_exception_type_predicate at 0x7f4d5ef43ee0>
sleep_generator = <generator object exponential_sleep_generator at 0x7f4d5c1b1660>
deadline = 30, on_error = None
def retry_target(target, predicate, sleep_generator, deadline, on_error=None):
"""Call a function and retry if it fails.
This is the lowest-level retry helper. Generally, you'll use the
higher-level retry helper :class:`Retry`.
Args:
target(Callable): The function to call and retry. This must be a
nullary function - apply arguments with `functools.partial`.
predicate (Callable[Exception]): A callable used to determine if an
exception raised by the target should be considered retryable.
It should return True to retry or False otherwise.
sleep_generator (Iterable[float]): An infinite iterator that determines
how long to sleep between retries.
deadline (float): How long to keep retrying the target. The last sleep
period is shortened as necessary, so that the last retry runs at
``deadline`` (and not considerably beyond it).
on_error (Callable[Exception]): A function to call while processing a
retryable exception. Any error raised by this function will *not*
be caught.
Returns:
Any: the return value of the target function.
Raises:
google.api_core.RetryError: If the deadline is exceeded while retrying.
ValueError: If the sleep generator stops yielding values.
Exception: If the target raises a method that isn't retryable.
"""
if deadline is not None:
deadline_datetime = datetime_helpers.utcnow() + datetime.timedelta(
seconds=deadline
)
else:
deadline_datetime = None
last_exc = None
for sleep in sleep_generator:
try:
> return target()
.nox/system-3-8/lib/python3.8/site-packages/google/api_core/retry.py:190:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <google.api_core.operation.Operation object at 0x7f4d5c321d30>
retry = <google.api_core.retry.Retry object at 0x7f4d5f0bad60>
def _done_or_raise(self, retry=DEFAULT_RETRY):
"""Check if the future is done and raise if it's not."""
kwargs = {} if retry is DEFAULT_RETRY else {"retry": retry}
if not self.done(**kwargs):
> raise _OperationNotComplete()
E google.api_core.future.polling._OperationNotComplete
.nox/system-3-8/lib/python3.8/site-packages/google/api_core/future/polling.py:89: _OperationNotComplete
The above exception was the direct cause of the following exception:
self = <google.api_core.operation.Operation object at 0x7f4d5c321d30>
timeout = 30, retry = <google.api_core.retry.Retry object at 0x7f4d5f0bad60>
def _blocking_poll(self, timeout=None, retry=DEFAULT_RETRY):
"""Poll and wait for the Future to be resolved.
Args:
timeout (int):
How long (in seconds) to wait for the operation to complete.
If None, wait indefinitely.
"""
if self._result_set:
return
retry_ = self._retry.with_deadline(timeout)
try:
kwargs = {} if retry is DEFAULT_RETRY else {"retry": retry}
> retry_(self._done_or_raise)(**kwargs)
.nox/system-3-8/lib/python3.8/site-packages/google/api_core/future/polling.py:110:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
args = (), kwargs = {}
target = functools.partial(<bound method PollingFuture._done_or_raise of <google.api_core.operation.Operation object at 0x7f4d5c321d30>>)
sleep_generator = <generator object exponential_sleep_generator at 0x7f4d5c1b1660>
@functools.wraps(func)
def retry_wrapped_func(*args, **kwargs):
"""A wrapper that calls target function with retry."""
target = functools.partial(func, *args, **kwargs)
sleep_generator = exponential_sleep_generator(
self._initial, self._maximum, multiplier=self._multiplier
)
> return retry_target(
target,
self._predicate,
sleep_generator,
self._deadline,
on_error=on_error,
)
.nox/system-3-8/lib/python3.8/site-packages/google/api_core/retry.py:283:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
target = functools.partial(<bound method PollingFuture._done_or_raise of <google.api_core.operation.Operation object at 0x7f4d5c321d30>>)
predicate = <function if_exception_type.<locals>.if_exception_type_predicate at 0x7f4d5ef43ee0>
sleep_generator = <generator object exponential_sleep_generator at 0x7f4d5c1b1660>
deadline = 30, on_error = None
def retry_target(target, predicate, sleep_generator, deadline, on_error=None):
"""Call a function and retry if it fails.
This is the lowest-level retry helper. Generally, you'll use the
higher-level retry helper :class:`Retry`.
Args:
target(Callable): The function to call and retry. This must be a
nullary function - apply arguments with `functools.partial`.
predicate (Callable[Exception]): A callable used to determine if an
exception raised by the target should be considered retryable.
It should return True to retry or False otherwise.
sleep_generator (Iterable[float]): An infinite iterator that determines
how long to sleep between retries.
deadline (float): How long to keep retrying the target. The last sleep
period is shortened as necessary, so that the last retry runs at
``deadline`` (and not considerably beyond it).
on_error (Callable[Exception]): A function to call while processing a
retryable exception. Any error raised by this function will *not*
be caught.
Returns:
Any: the return value of the target function.
Raises:
google.api_core.RetryError: If the deadline is exceeded while retrying.
ValueError: If the sleep generator stops yielding values.
Exception: If the target raises a method that isn't retryable.
"""
if deadline is not None:
deadline_datetime = datetime_helpers.utcnow() + datetime.timedelta(
seconds=deadline
)
else:
deadline_datetime = None
last_exc = None
for sleep in sleep_generator:
try:
return target()
# pylint: disable=broad-except
# This function explicitly must deal with broad exceptions.
except Exception as exc:
if not predicate(exc):
raise
last_exc = exc
if on_error is not None:
on_error(exc)
now = datetime_helpers.utcnow()
if deadline_datetime is not None:
if deadline_datetime <= now:
> raise exceptions.RetryError(
"Deadline of {:.1f}s exceeded while calling {}".format(
deadline, target
),
last_exc,
) from last_exc
E google.api_core.exceptions.RetryError: Deadline of 30.0s exceeded while calling functools.partial(<bound method PollingFuture._done_or_raise of <google.api_core.operation.Operation object at 0x7f4d5c321d30>>), last exception:
.nox/system-3-8/lib/python3.8/site-packages/google/api_core/retry.py:205: RetryError
During handling of the above exception, another exception occurred:
admin_instance_populated = <google.cloud.bigtable.instance.Instance object at 0x7f4d5eef10a0>
admin_instance_id = 'g-c-p-1634667466967', skip_on_emulator = None
def test_cluster_create(
admin_instance_populated, admin_instance_id, skip_on_emulator,
):
alt_cluster_id = f"{admin_instance_id}-c2"
alt_location_id = "us-central1-f"
serve_nodes = 2
cluster_2 = admin_instance_populated.cluster(
alt_cluster_id,
location_id=alt_location_id,
serve_nodes=serve_nodes,
default_storage_type=(enums.StorageType.SSD),
)
operation = cluster_2.create()
> operation.result(timeout=30) # Ensure the operation completes.
tests/system/test_instance_admin.py:576:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
.nox/system-3-8/lib/python3.8/site-packages/google/api_core/future/polling.py:132: in result
self._blocking_poll(timeout=timeout, **kwargs)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <google.api_core.operation.Operation object at 0x7f4d5c321d30>
timeout = 30, retry = <google.api_core.retry.Retry object at 0x7f4d5f0bad60>
def _blocking_poll(self, timeout=None, retry=DEFAULT_RETRY):
"""Poll and wait for the Future to be resolved.
Args:
timeout (int):
How long (in seconds) to wait for the operation to complete.
If None, wait indefinitely.
"""
if self._result_set:
return
retry_ = self._retry.with_deadline(timeout)
try:
kwargs = {} if retry is DEFAULT_RETRY else {"retry": retry}
retry_(self._done_or_raise)(**kwargs)
except exceptions.RetryError:
> raise concurrent.futures.TimeoutError(
"Operation did not complete within the designated " "timeout."
)
E concurrent.futures._base.TimeoutError: Operation did not complete within the designated timeout.
.nox/system-3-8/lib/python3.8/site-packages/google/api_core/future/polling.py:112: TimeoutError
```
|
1.0
|
'test_cluster_create' systest flakes with 'TimeoutError' - From [this failed systest build](https://source.cloud.google.com/results/invocations/924fbb8c-f387-482d-8767-c16abd2de9e2/targets/cloud-devrel%2Fclient-libraries%2Fpython%2Fgoogleapis%2Fpython-bigtable%2Fpresubmit%2Fsystem-3.8/log):
```python
_____________________________ test_cluster_create ______________________________
target = functools.partial(<bound method PollingFuture._done_or_raise of <google.api_core.operation.Operation object at 0x7f4d5c321d30>>)
predicate = <function if_exception_type.<locals>.if_exception_type_predicate at 0x7f4d5ef43ee0>
sleep_generator = <generator object exponential_sleep_generator at 0x7f4d5c1b1660>
deadline = 30, on_error = None
def retry_target(target, predicate, sleep_generator, deadline, on_error=None):
"""Call a function and retry if it fails.
This is the lowest-level retry helper. Generally, you'll use the
higher-level retry helper :class:`Retry`.
Args:
target(Callable): The function to call and retry. This must be a
nullary function - apply arguments with `functools.partial`.
predicate (Callable[Exception]): A callable used to determine if an
exception raised by the target should be considered retryable.
It should return True to retry or False otherwise.
sleep_generator (Iterable[float]): An infinite iterator that determines
how long to sleep between retries.
deadline (float): How long to keep retrying the target. The last sleep
period is shortened as necessary, so that the last retry runs at
``deadline`` (and not considerably beyond it).
on_error (Callable[Exception]): A function to call while processing a
retryable exception. Any error raised by this function will *not*
be caught.
Returns:
Any: the return value of the target function.
Raises:
google.api_core.RetryError: If the deadline is exceeded while retrying.
ValueError: If the sleep generator stops yielding values.
Exception: If the target raises a method that isn't retryable.
"""
if deadline is not None:
deadline_datetime = datetime_helpers.utcnow() + datetime.timedelta(
seconds=deadline
)
else:
deadline_datetime = None
last_exc = None
for sleep in sleep_generator:
try:
> return target()
.nox/system-3-8/lib/python3.8/site-packages/google/api_core/retry.py:190:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <google.api_core.operation.Operation object at 0x7f4d5c321d30>
retry = <google.api_core.retry.Retry object at 0x7f4d5f0bad60>
def _done_or_raise(self, retry=DEFAULT_RETRY):
"""Check if the future is done and raise if it's not."""
kwargs = {} if retry is DEFAULT_RETRY else {"retry": retry}
if not self.done(**kwargs):
> raise _OperationNotComplete()
E google.api_core.future.polling._OperationNotComplete
.nox/system-3-8/lib/python3.8/site-packages/google/api_core/future/polling.py:89: _OperationNotComplete
The above exception was the direct cause of the following exception:
self = <google.api_core.operation.Operation object at 0x7f4d5c321d30>
timeout = 30, retry = <google.api_core.retry.Retry object at 0x7f4d5f0bad60>
def _blocking_poll(self, timeout=None, retry=DEFAULT_RETRY):
"""Poll and wait for the Future to be resolved.
Args:
timeout (int):
How long (in seconds) to wait for the operation to complete.
If None, wait indefinitely.
"""
if self._result_set:
return
retry_ = self._retry.with_deadline(timeout)
try:
kwargs = {} if retry is DEFAULT_RETRY else {"retry": retry}
> retry_(self._done_or_raise)(**kwargs)
.nox/system-3-8/lib/python3.8/site-packages/google/api_core/future/polling.py:110:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
args = (), kwargs = {}
target = functools.partial(<bound method PollingFuture._done_or_raise of <google.api_core.operation.Operation object at 0x7f4d5c321d30>>)
sleep_generator = <generator object exponential_sleep_generator at 0x7f4d5c1b1660>
@functools.wraps(func)
def retry_wrapped_func(*args, **kwargs):
"""A wrapper that calls target function with retry."""
target = functools.partial(func, *args, **kwargs)
sleep_generator = exponential_sleep_generator(
self._initial, self._maximum, multiplier=self._multiplier
)
> return retry_target(
target,
self._predicate,
sleep_generator,
self._deadline,
on_error=on_error,
)
.nox/system-3-8/lib/python3.8/site-packages/google/api_core/retry.py:283:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
target = functools.partial(<bound method PollingFuture._done_or_raise of <google.api_core.operation.Operation object at 0x7f4d5c321d30>>)
predicate = <function if_exception_type.<locals>.if_exception_type_predicate at 0x7f4d5ef43ee0>
sleep_generator = <generator object exponential_sleep_generator at 0x7f4d5c1b1660>
deadline = 30, on_error = None
def retry_target(target, predicate, sleep_generator, deadline, on_error=None):
"""Call a function and retry if it fails.
This is the lowest-level retry helper. Generally, you'll use the
higher-level retry helper :class:`Retry`.
Args:
target(Callable): The function to call and retry. This must be a
nullary function - apply arguments with `functools.partial`.
predicate (Callable[Exception]): A callable used to determine if an
exception raised by the target should be considered retryable.
It should return True to retry or False otherwise.
sleep_generator (Iterable[float]): An infinite iterator that determines
how long to sleep between retries.
deadline (float): How long to keep retrying the target. The last sleep
period is shortened as necessary, so that the last retry runs at
``deadline`` (and not considerably beyond it).
on_error (Callable[Exception]): A function to call while processing a
retryable exception. Any error raised by this function will *not*
be caught.
Returns:
Any: the return value of the target function.
Raises:
google.api_core.RetryError: If the deadline is exceeded while retrying.
ValueError: If the sleep generator stops yielding values.
Exception: If the target raises a method that isn't retryable.
"""
if deadline is not None:
deadline_datetime = datetime_helpers.utcnow() + datetime.timedelta(
seconds=deadline
)
else:
deadline_datetime = None
last_exc = None
for sleep in sleep_generator:
try:
return target()
# pylint: disable=broad-except
# This function explicitly must deal with broad exceptions.
except Exception as exc:
if not predicate(exc):
raise
last_exc = exc
if on_error is not None:
on_error(exc)
now = datetime_helpers.utcnow()
if deadline_datetime is not None:
if deadline_datetime <= now:
> raise exceptions.RetryError(
"Deadline of {:.1f}s exceeded while calling {}".format(
deadline, target
),
last_exc,
) from last_exc
E google.api_core.exceptions.RetryError: Deadline of 30.0s exceeded while calling functools.partial(<bound method PollingFuture._done_or_raise of <google.api_core.operation.Operation object at 0x7f4d5c321d30>>), last exception:
.nox/system-3-8/lib/python3.8/site-packages/google/api_core/retry.py:205: RetryError
During handling of the above exception, another exception occurred:
admin_instance_populated = <google.cloud.bigtable.instance.Instance object at 0x7f4d5eef10a0>
admin_instance_id = 'g-c-p-1634667466967', skip_on_emulator = None
def test_cluster_create(
admin_instance_populated, admin_instance_id, skip_on_emulator,
):
alt_cluster_id = f"{admin_instance_id}-c2"
alt_location_id = "us-central1-f"
serve_nodes = 2
cluster_2 = admin_instance_populated.cluster(
alt_cluster_id,
location_id=alt_location_id,
serve_nodes=serve_nodes,
default_storage_type=(enums.StorageType.SSD),
)
operation = cluster_2.create()
> operation.result(timeout=30) # Ensure the operation completes.
tests/system/test_instance_admin.py:576:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
.nox/system-3-8/lib/python3.8/site-packages/google/api_core/future/polling.py:132: in result
self._blocking_poll(timeout=timeout, **kwargs)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <google.api_core.operation.Operation object at 0x7f4d5c321d30>
timeout = 30, retry = <google.api_core.retry.Retry object at 0x7f4d5f0bad60>
def _blocking_poll(self, timeout=None, retry=DEFAULT_RETRY):
"""Poll and wait for the Future to be resolved.
Args:
timeout (int):
How long (in seconds) to wait for the operation to complete.
If None, wait indefinitely.
"""
if self._result_set:
return
retry_ = self._retry.with_deadline(timeout)
try:
kwargs = {} if retry is DEFAULT_RETRY else {"retry": retry}
retry_(self._done_or_raise)(**kwargs)
except exceptions.RetryError:
> raise concurrent.futures.TimeoutError(
"Operation did not complete within the designated " "timeout."
)
E concurrent.futures._base.TimeoutError: Operation did not complete within the designated timeout.
.nox/system-3-8/lib/python3.8/site-packages/google/api_core/future/polling.py:112: TimeoutError
```
|
process
|
test cluster create systest flakes with timeouterror from python test cluster create target functools partial predicate if exception type predicate at sleep generator deadline on error none def retry target target predicate sleep generator deadline on error none call a function and retry if it fails this is the lowest level retry helper generally you ll use the higher level retry helper class retry args target callable the function to call and retry this must be a nullary function apply arguments with functools partial predicate callable a callable used to determine if an exception raised by the target should be considered retryable it should return true to retry or false otherwise sleep generator iterable an infinite iterator that determines how long to sleep between retries deadline float how long to keep retrying the target the last sleep period is shortened as necessary so that the last retry runs at deadline and not considerably beyond it on error callable a function to call while processing a retryable exception any error raised by this function will not be caught returns any the return value of the target function raises google api core retryerror if the deadline is exceeded while retrying valueerror if the sleep generator stops yielding values exception if the target raises a method that isn t retryable if deadline is not none deadline datetime datetime helpers utcnow datetime timedelta seconds deadline else deadline datetime none last exc none for sleep in sleep generator try return target nox system lib site packages google api core retry py self retry def done or raise self retry default retry check if the future is done and raise if it s not kwargs if retry is default retry else retry retry if not self done kwargs raise operationnotcomplete e google api core future polling operationnotcomplete nox system lib site packages google api core future polling py operationnotcomplete the above exception was the direct cause of the following exception self timeout retry def blocking poll self timeout none retry default retry poll and wait for the future to be resolved args timeout int how long in seconds to wait for the operation to complete if none wait indefinitely if self result set return retry self retry with deadline timeout try kwargs if retry is default retry else retry retry retry self done or raise kwargs nox system lib site packages google api core future polling py args kwargs target functools partial sleep generator functools wraps func def retry wrapped func args kwargs a wrapper that calls target function with retry target functools partial func args kwargs sleep generator exponential sleep generator self initial self maximum multiplier self multiplier return retry target target self predicate sleep generator self deadline on error on error nox system lib site packages google api core retry py target functools partial predicate if exception type predicate at sleep generator deadline on error none def retry target target predicate sleep generator deadline on error none call a function and retry if it fails this is the lowest level retry helper generally you ll use the higher level retry helper class retry args target callable the function to call and retry this must be a nullary function apply arguments with functools partial predicate callable a callable used to determine if an exception raised by the target should be considered retryable it should return true to retry or false otherwise sleep generator iterable an infinite iterator that determines how long to sleep between retries deadline float how long to keep retrying the target the last sleep period is shortened as necessary so that the last retry runs at deadline and not considerably beyond it on error callable a function to call while processing a retryable exception any error raised by this function will not be caught returns any the return value of the target function raises google api core retryerror if the deadline is exceeded while retrying valueerror if the sleep generator stops yielding values exception if the target raises a method that isn t retryable if deadline is not none deadline datetime datetime helpers utcnow datetime timedelta seconds deadline else deadline datetime none last exc none for sleep in sleep generator try return target pylint disable broad except this function explicitly must deal with broad exceptions except exception as exc if not predicate exc raise last exc exc if on error is not none on error exc now datetime helpers utcnow if deadline datetime is not none if deadline datetime now raise exceptions retryerror deadline of s exceeded while calling format deadline target last exc from last exc e google api core exceptions retryerror deadline of exceeded while calling functools partial last exception nox system lib site packages google api core retry py retryerror during handling of the above exception another exception occurred admin instance populated admin instance id g c p skip on emulator none def test cluster create admin instance populated admin instance id skip on emulator alt cluster id f admin instance id alt location id us f serve nodes cluster admin instance populated cluster alt cluster id location id alt location id serve nodes serve nodes default storage type enums storagetype ssd operation cluster create operation result timeout ensure the operation completes tests system test instance admin py nox system lib site packages google api core future polling py in result self blocking poll timeout timeout kwargs self timeout retry def blocking poll self timeout none retry default retry poll and wait for the future to be resolved args timeout int how long in seconds to wait for the operation to complete if none wait indefinitely if self result set return retry self retry with deadline timeout try kwargs if retry is default retry else retry retry retry self done or raise kwargs except exceptions retryerror raise concurrent futures timeouterror operation did not complete within the designated timeout e concurrent futures base timeouterror operation did not complete within the designated timeout nox system lib site packages google api core future polling py timeouterror
| 1
|
382,176
| 26,487,494,804
|
IssuesEvent
|
2023-01-17 19:20:39
|
RalphHightower/RalphHightower
|
https://api.github.com/repos/RalphHightower/RalphHightower
|
closed
|
MurdaughAlexJudgementDay.md
|
documentation
|
**What page should this be added to?**<br>
MurdaughAlexJudgementDay.md
**What section/heading should this be added to?**<br>
H1
**Include the Markdown text that is to be added below:**<br>
# A Southern Gothic Tale of Money, Mystery, and Murder: Judgement Day
**Describe alternatives you've considered**<br>
Bookmarks in browsers are not portable.
**Additional context**<br>
Add any other context or screenshots about the feature request here.
|
1.0
|
MurdaughAlexJudgementDay.md - **What page should this be added to?**<br>
MurdaughAlexJudgementDay.md
**What section/heading should this be added to?**<br>
H1
**Include the Markdown text that is to be added below:**<br>
# A Southern Gothic Tale of Money, Mystery, and Murder: Judgement Day
**Describe alternatives you've considered**<br>
Bookmarks in browsers are not portable.
**Additional context**<br>
Add any other context or screenshots about the feature request here.
|
non_process
|
murdaughalexjudgementday md what page should this be added to murdaughalexjudgementday md what section heading should this be added to include the markdown text that is to be added below a southern gothic tale of money mystery and murder judgement day describe alternatives you ve considered bookmarks in browsers are not portable additional context add any other context or screenshots about the feature request here
| 0
|
13,875
| 16,639,515,235
|
IssuesEvent
|
2021-06-04 06:34:35
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
Let child_process 'cwd' arguments accept a URL
|
child_process feature request
|
**Is your feature request related to a problem? Please describe.**
It would be nice if child_process functionality, such as `exec`, could accept a `URL` ([docs](https://nodejs.org/api/url.html#url_class_url)) for the `cwd` argument. For example the File System `writeFile` [accepts a URL](https://nodejs.org/api/fs.html#fs_fspromises_writefile_file_data_options) as its `file` argument already. So it would be nice for the child_process module to do so as well for path arguments.
This is especially nice with ECMAScript modules since e.g. the old `__dirname` does not exist anymore, but we are [forced to use](https://nodejs.org/api/esm.html#esm_no_filename_or_dirname) `URL` to get a file path to the current file.
**Describe the solution you'd like**
```js
// Using ESM Module syntax:
import { exec } from 'child_process/promises';
try {
const { stdout } = await exec(
'sysctl -n net.ipv4.ip_local_port_range',
{ encoding: 'utf8', cwd: new URL('../../', import.meta.url) } // without URL support we would have to have used .pathname here
);
console.log('successfully executed the child process command');
} catch (error) {
console.error('there was an error:', error.message);
}
```
|
1.0
|
Let child_process 'cwd' arguments accept a URL - **Is your feature request related to a problem? Please describe.**
It would be nice if child_process functionality, such as `exec`, could accept a `URL` ([docs](https://nodejs.org/api/url.html#url_class_url)) for the `cwd` argument. For example the File System `writeFile` [accepts a URL](https://nodejs.org/api/fs.html#fs_fspromises_writefile_file_data_options) as its `file` argument already. So it would be nice for the child_process module to do so as well for path arguments.
This is especially nice with ECMAScript modules since e.g. the old `__dirname` does not exist anymore, but we are [forced to use](https://nodejs.org/api/esm.html#esm_no_filename_or_dirname) `URL` to get a file path to the current file.
**Describe the solution you'd like**
```js
// Using ESM Module syntax:
import { exec } from 'child_process/promises';
try {
const { stdout } = await exec(
'sysctl -n net.ipv4.ip_local_port_range',
{ encoding: 'utf8', cwd: new URL('../../', import.meta.url) } // without URL support we would have to have used .pathname here
);
console.log('successfully executed the child process command');
} catch (error) {
console.error('there was an error:', error.message);
}
```
|
process
|
let child process cwd arguments accept a url is your feature request related to a problem please describe it would be nice if child process functionality such as exec could accept a url for the cwd argument for example the file system writefile as its file argument already so it would be nice for the child process module to do so as well for path arguments this is especially nice with ecmascript modules since e g the old dirname does not exist anymore but we are url to get a file path to the current file describe the solution you d like js using esm module syntax import exec from child process promises try const stdout await exec sysctl n net ip local port range encoding cwd new url import meta url without url support we would have to have used pathname here console log successfully executed the child process command catch error console error there was an error error message
| 1
|
188,677
| 15,167,609,824
|
IssuesEvent
|
2021-02-12 18:03:28
|
aws-greengrass/aws-greengrass-nucleus
|
https://api.github.com/repos/aws-greengrass/aws-greengrass-nucleus
|
closed
|
Installation doesn't work from behind a proxy
|
bug documentation
|
Hi,
I'm trying to run the installation from behind a corporate proxy.
- From this box, I am only allowed to reach the internet through an http proxy.
- Also name resolution happens through the proxy.
- I managed to make SQS work though the proxy
Using a custom init-config and all the operations that are using the aws sdk are working, but it fails when it's trying to download the rootCA. That part of the install script is using the java.net and java.nio and apparently they don't use the proxy: https://github.com/aws-greengrass/aws-greengrass-nucleus/blob/4da4e9b7feed8ef7b4e2037fa24add3dc67c3137/src/main/java/com/aws/greengrass/easysetup/DeviceProvisioningHelper.java#L285
I've also tried to specify `-Djava.net.useSystemProxies=true` but still it doesn't work. Here is the log:
```
sudo -E java -Djava.net.useSystemProxies=true -Droot="/greengrass/v2" -Dlog.store=FILE -jar ./GreengrassCore/lib/Greengrass.jar --aws-region eu-central-1 --thing-name GreengrassQuickStartCore-177876a9adf --thing-group-name GreengrassQuickStartGroup --component-default-user ggc_user:ggc_group --provision true --setup-system-service true --deploy-dev-tools true --init-config /tmp/effectiveConfig.yaml
Added ggc_user to ggc_group
Provisioning AWS IoT resources for the device with IoT Thing Name: [GreengrassQuickStartCore-177876a9adf]...
Found IoT policy "GreengrassV2IoTThingPolicy", reusing it
Creating keys and certificate...
Attaching policy to certificate...
Creating IoT Thing "GreengrassQuickStartCore-177876a9adf"...
Attaching certificate to IoT thing...
Successfully provisioned AWS IoT resources for the device with IoT Thing Name: [GreengrassQuickStartCore-177876a9adf]!
Adding IoT Thing [GreengrassQuickStartCore-177876a9adf] into Thing Group: [GreengrassQuickStartGroup]...
IoT Thing Group "GreengrassQuickStartGroup" already existed, reusing it
Successfully added Thing into Thing Group: [GreengrassQuickStartGroup]
Setting up resources for aws.greengrass.TokenExchangeService ...
Attaching TES role policy to IoT thing...
IAM policy named "GreengrassV2TokenExchangeRoleAccess" already exists. Please attach it to the IAM role if not already
Configuring Nucleus with provisioned resource details...
Downloading Root CA from "https://www.amazontrust.com/repository/AmazonRootCA1.pem"
Error while trying to setup Greengrass Nucleus
java.net.UnknownHostException: www.amazontrust.com
at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:184)
at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:392)
at java.net.Socket.connect(Socket.java:607)
at sun.security.ssl.SSLSocketImpl.connect(SSLSocketImpl.java:284)
at sun.security.ssl.BaseSSLSocketImpl.connect(BaseSSLSocketImpl.java:173)
at sun.net.NetworkClient.doConnect(NetworkClient.java:180)
at sun.net.www.http.HttpClient.openServer(HttpClient.java:463)
at sun.net.www.http.HttpClient.openServer(HttpClient.java:558)
at sun.net.www.protocol.https.HttpsClient.<init>(HttpsClient.java:264)
at sun.net.www.protocol.https.HttpsClient.New(HttpsClient.java:367)
at sun.net.www.protocol.https.AbstractDelegateHttpsURLConnection.getNewHttpClient(AbstractDelegateHttpsURLConnection.java:191)
at sun.net.www.protocol.http.HttpURLConnection.plainConnect0(HttpURLConnection.java:1162)
at sun.net.www.protocol.http.HttpURLConnection.plainConnect(HttpURLConnection.java:1056)
at sun.net.www.protocol.https.AbstractDelegateHttpsURLConnection.connect(AbstractDelegateHttpsURLConnection.java:177)
at sun.net.www.protocol.http.HttpURLConnection.getInputStream0(HttpURLConnection.java:1570)
at sun.net.www.protocol.http.HttpURLConnection.getInputStream(HttpURLConnection.java:1498)
at sun.net.www.protocol.https.HttpsURLConnectionImpl.getInputStream(HttpsURLConnectionImpl.java:268)
at java.net.URL.openStream(URL.java:1068)
at com.aws.greengrass.easysetup.DeviceProvisioningHelper.downloadFileFromURL(DeviceProvisioningHelper.java:286)
at com.aws.greengrass.easysetup.DeviceProvisioningHelper.downloadRootCAToFile(DeviceProvisioningHelper.java:278)
at com.aws.greengrass.easysetup.DeviceProvisioningHelper.updateKernelConfigWithIotConfiguration(DeviceProvisioningHelper.java:310)
at com.aws.greengrass.easysetup.GreengrassSetup.provision(GreengrassSetup.java:425)
at com.aws.greengrass.easysetup.GreengrassSetup.performSetup(GreengrassSetup.java:294)
at com.aws.greengrass.easysetup.GreengrassSetup.main(GreengrassSetup.java:242)
```
|
1.0
|
Installation doesn't work from behind a proxy - Hi,
I'm trying to run the installation from behind a corporate proxy.
- From this box, I am only allowed to reach the internet through an http proxy.
- Also name resolution happens through the proxy.
- I managed to make SQS work though the proxy
Using a custom init-config and all the operations that are using the aws sdk are working, but it fails when it's trying to download the rootCA. That part of the install script is using the java.net and java.nio and apparently they don't use the proxy: https://github.com/aws-greengrass/aws-greengrass-nucleus/blob/4da4e9b7feed8ef7b4e2037fa24add3dc67c3137/src/main/java/com/aws/greengrass/easysetup/DeviceProvisioningHelper.java#L285
I've also tried to specify `-Djava.net.useSystemProxies=true` but still it doesn't work. Here is the log:
```
sudo -E java -Djava.net.useSystemProxies=true -Droot="/greengrass/v2" -Dlog.store=FILE -jar ./GreengrassCore/lib/Greengrass.jar --aws-region eu-central-1 --thing-name GreengrassQuickStartCore-177876a9adf --thing-group-name GreengrassQuickStartGroup --component-default-user ggc_user:ggc_group --provision true --setup-system-service true --deploy-dev-tools true --init-config /tmp/effectiveConfig.yaml
Added ggc_user to ggc_group
Provisioning AWS IoT resources for the device with IoT Thing Name: [GreengrassQuickStartCore-177876a9adf]...
Found IoT policy "GreengrassV2IoTThingPolicy", reusing it
Creating keys and certificate...
Attaching policy to certificate...
Creating IoT Thing "GreengrassQuickStartCore-177876a9adf"...
Attaching certificate to IoT thing...
Successfully provisioned AWS IoT resources for the device with IoT Thing Name: [GreengrassQuickStartCore-177876a9adf]!
Adding IoT Thing [GreengrassQuickStartCore-177876a9adf] into Thing Group: [GreengrassQuickStartGroup]...
IoT Thing Group "GreengrassQuickStartGroup" already existed, reusing it
Successfully added Thing into Thing Group: [GreengrassQuickStartGroup]
Setting up resources for aws.greengrass.TokenExchangeService ...
Attaching TES role policy to IoT thing...
IAM policy named "GreengrassV2TokenExchangeRoleAccess" already exists. Please attach it to the IAM role if not already
Configuring Nucleus with provisioned resource details...
Downloading Root CA from "https://www.amazontrust.com/repository/AmazonRootCA1.pem"
Error while trying to setup Greengrass Nucleus
java.net.UnknownHostException: www.amazontrust.com
at java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:184)
at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:392)
at java.net.Socket.connect(Socket.java:607)
at sun.security.ssl.SSLSocketImpl.connect(SSLSocketImpl.java:284)
at sun.security.ssl.BaseSSLSocketImpl.connect(BaseSSLSocketImpl.java:173)
at sun.net.NetworkClient.doConnect(NetworkClient.java:180)
at sun.net.www.http.HttpClient.openServer(HttpClient.java:463)
at sun.net.www.http.HttpClient.openServer(HttpClient.java:558)
at sun.net.www.protocol.https.HttpsClient.<init>(HttpsClient.java:264)
at sun.net.www.protocol.https.HttpsClient.New(HttpsClient.java:367)
at sun.net.www.protocol.https.AbstractDelegateHttpsURLConnection.getNewHttpClient(AbstractDelegateHttpsURLConnection.java:191)
at sun.net.www.protocol.http.HttpURLConnection.plainConnect0(HttpURLConnection.java:1162)
at sun.net.www.protocol.http.HttpURLConnection.plainConnect(HttpURLConnection.java:1056)
at sun.net.www.protocol.https.AbstractDelegateHttpsURLConnection.connect(AbstractDelegateHttpsURLConnection.java:177)
at sun.net.www.protocol.http.HttpURLConnection.getInputStream0(HttpURLConnection.java:1570)
at sun.net.www.protocol.http.HttpURLConnection.getInputStream(HttpURLConnection.java:1498)
at sun.net.www.protocol.https.HttpsURLConnectionImpl.getInputStream(HttpsURLConnectionImpl.java:268)
at java.net.URL.openStream(URL.java:1068)
at com.aws.greengrass.easysetup.DeviceProvisioningHelper.downloadFileFromURL(DeviceProvisioningHelper.java:286)
at com.aws.greengrass.easysetup.DeviceProvisioningHelper.downloadRootCAToFile(DeviceProvisioningHelper.java:278)
at com.aws.greengrass.easysetup.DeviceProvisioningHelper.updateKernelConfigWithIotConfiguration(DeviceProvisioningHelper.java:310)
at com.aws.greengrass.easysetup.GreengrassSetup.provision(GreengrassSetup.java:425)
at com.aws.greengrass.easysetup.GreengrassSetup.performSetup(GreengrassSetup.java:294)
at com.aws.greengrass.easysetup.GreengrassSetup.main(GreengrassSetup.java:242)
```
|
non_process
|
installation doesn t work from behind a proxy hi i m trying to run the installation from behind a corporate proxy from this box i am only allowed to reach the internet through an http proxy also name resolution happens through the proxy i managed to make sqs work though the proxy using a custom init config and all the operations that are using the aws sdk are working but it fails when it s trying to download the rootca that part of the install script is using the java net and java nio and apparently they don t use the proxy i ve also tried to specify djava net usesystemproxies true but still it doesn t work here is the log sudo e java djava net usesystemproxies true droot greengrass dlog store file jar greengrasscore lib greengrass jar aws region eu central thing name greengrassquickstartcore thing group name greengrassquickstartgroup component default user ggc user ggc group provision true setup system service true deploy dev tools true init config tmp effectiveconfig yaml added ggc user to ggc group provisioning aws iot resources for the device with iot thing name found iot policy reusing it creating keys and certificate attaching policy to certificate creating iot thing greengrassquickstartcore attaching certificate to iot thing successfully provisioned aws iot resources for the device with iot thing name adding iot thing into thing group iot thing group greengrassquickstartgroup already existed reusing it successfully added thing into thing group setting up resources for aws greengrass tokenexchangeservice attaching tes role policy to iot thing iam policy named already exists please attach it to the iam role if not already configuring nucleus with provisioned resource details downloading root ca from error while trying to setup greengrass nucleus java net unknownhostexception at java net abstractplainsocketimpl connect abstractplainsocketimpl java at java net sockssocketimpl connect sockssocketimpl java at java net socket connect socket java at sun security ssl sslsocketimpl connect sslsocketimpl java at sun security ssl basesslsocketimpl connect basesslsocketimpl java at sun net networkclient doconnect networkclient java at sun net at sun net at sun net at sun net at sun net at sun net at sun net at sun net at sun net at sun net at sun net at java net url openstream url java at com aws greengrass easysetup deviceprovisioninghelper downloadfilefromurl deviceprovisioninghelper java at com aws greengrass easysetup deviceprovisioninghelper downloadrootcatofile deviceprovisioninghelper java at com aws greengrass easysetup deviceprovisioninghelper updatekernelconfigwithiotconfiguration deviceprovisioninghelper java at com aws greengrass easysetup greengrasssetup provision greengrasssetup java at com aws greengrass easysetup greengrasssetup performsetup greengrasssetup java at com aws greengrass easysetup greengrasssetup main greengrasssetup java
| 0
|
7,623
| 10,727,943,972
|
IssuesEvent
|
2019-10-28 12:55:51
|
linnovate/root
|
https://api.github.com/repos/linnovate/root
|
opened
|
upload a file in document the name change
|
2.0.8 Internal Test Process bug
|
when you upload document the entity name also changes
the user should be given the option to choose a name
|
1.0
|
upload a file in document the name change - when you upload document the entity name also changes
the user should be given the option to choose a name
|
process
|
upload a file in document the name change when you upload document the entity name also changes the user should be given the option to choose a name
| 1
|
20,558
| 27,217,832,065
|
IssuesEvent
|
2023-02-21 00:39:31
|
cse442-at-ub/project_s23-iweatherify
|
https://api.github.com/repos/cse442-at-ub/project_s23-iweatherify
|
opened
|
Think of possible ideas and document them
|
Processing Task Sprint 1
|
**Task Tests**
*Test 1*
1) Open document https://docs.google.com/document/d/1fQDM2_rvD49LgCHpRX-fh-yX0UqBG1fgJzAu0D9KsRQ/edit?usp=sharing
2) Verify that document has more than three ideas and that each idea gives at least a brief description of what it is about and what features it envisions being implemented
|
1.0
|
Think of possible ideas and document them - **Task Tests**
*Test 1*
1) Open document https://docs.google.com/document/d/1fQDM2_rvD49LgCHpRX-fh-yX0UqBG1fgJzAu0D9KsRQ/edit?usp=sharing
2) Verify that document has more than three ideas and that each idea gives at least a brief description of what it is about and what features it envisions being implemented
|
process
|
think of possible ideas and document them task tests test open document verify that document has more than three ideas and that each idea gives at least a brief description of what it is about and what features it envisions being implemented
| 1
|
97,202
| 8,651,573,090
|
IssuesEvent
|
2018-11-27 03:50:27
|
humera987/FXLabs-Test-Automation
|
https://api.github.com/repos/humera987/FXLabs-Test-Automation
|
closed
|
projecttest16 : ApiV1OrgsIdUsersGetQueryParamPageNegativeNumber
|
projecttest16
|
Project : projecttest16
Job : UAT
Env : UAT
Region : US_WEST
Result : fail
Status Code : 404
Headers : {X-Content-Type-Options=[nosniff], X-XSS-Protection=[1; mode=block], Cache-Control=[no-cache, no-store, max-age=0, must-revalidate], Pragma=[no-cache], Expires=[0], X-Frame-Options=[DENY], Set-Cookie=[SESSION=M2Q0ZGQyZjQtZmI1Ny00YjE0LWFiYzAtOWU1ZGJlMTI4MGYy; Path=/; HttpOnly], Content-Type=[application/json;charset=UTF-8], Transfer-Encoding=[chunked], Date=[Fri, 16 Nov 2018 05:40:26 GMT]}
Endpoint : http://13.56.210.25/api/v1/api/v1/orgs/pGhqMAS5/users?page=-1
Request :
Response :
{
"timestamp" : "2018-11-16T05:40:26.148+0000",
"status" : 404,
"error" : "Not Found",
"message" : "No message available",
"path" : "/api/v1/api/v1/orgs/pGhqMAS5/users"
}
Logs :
Assertion [@StatusCode != 401] resolved-to [404 != 401] result [Passed]Assertion [@StatusCode != 404] resolved-to [404 != 404] result [Failed]
--- FX Bot ---
|
1.0
|
projecttest16 : ApiV1OrgsIdUsersGetQueryParamPageNegativeNumber - Project : projecttest16
Job : UAT
Env : UAT
Region : US_WEST
Result : fail
Status Code : 404
Headers : {X-Content-Type-Options=[nosniff], X-XSS-Protection=[1; mode=block], Cache-Control=[no-cache, no-store, max-age=0, must-revalidate], Pragma=[no-cache], Expires=[0], X-Frame-Options=[DENY], Set-Cookie=[SESSION=M2Q0ZGQyZjQtZmI1Ny00YjE0LWFiYzAtOWU1ZGJlMTI4MGYy; Path=/; HttpOnly], Content-Type=[application/json;charset=UTF-8], Transfer-Encoding=[chunked], Date=[Fri, 16 Nov 2018 05:40:26 GMT]}
Endpoint : http://13.56.210.25/api/v1/api/v1/orgs/pGhqMAS5/users?page=-1
Request :
Response :
{
"timestamp" : "2018-11-16T05:40:26.148+0000",
"status" : 404,
"error" : "Not Found",
"message" : "No message available",
"path" : "/api/v1/api/v1/orgs/pGhqMAS5/users"
}
Logs :
Assertion [@StatusCode != 401] resolved-to [404 != 401] result [Passed]Assertion [@StatusCode != 404] resolved-to [404 != 404] result [Failed]
--- FX Bot ---
|
non_process
|
project job uat env uat region us west result fail status code headers x content type options x xss protection cache control pragma expires x frame options set cookie content type transfer encoding date endpoint request response timestamp status error not found message no message available path api api orgs users logs assertion resolved to result assertion resolved to result fx bot
| 0
|
255,642
| 21,941,720,004
|
IssuesEvent
|
2022-05-23 18:51:31
|
gonmarfer2/Acme-Toolkits
|
https://api.github.com/repos/gonmarfer2/Acme-Toolkits
|
closed
|
Task-096/T: Operations by inventors on patronage reports (7)
|
testing D04
|
Create both positive and negative test cases for the following functionality:
- [ ] Create a patronage report. They cannot be updated or deleted; the system must require confirmation before creating them.
|
1.0
|
Task-096/T: Operations by inventors on patronage reports (7) - Create both positive and negative test cases for the following functionality:
- [ ] Create a patronage report. They cannot be updated or deleted; the system must require confirmation before creating them.
|
non_process
|
task t operations by inventors on patronage reports create both positive and negative test cases for the following functionality create a patronage report they cannot be updated or deleted the system must require confirmation before creating them
| 0
|
7,015
| 10,166,052,256
|
IssuesEvent
|
2019-08-07 15:03:06
|
pwittchen/ReactiveNetwork
|
https://api.github.com/repos/pwittchen/ReactiveNetwork
|
closed
|
Release 3.0.4
|
release process
|
**Release notes**:
- fixed bug #330 - State CONNECTED sometimes is not returned when wifi is turned off while having mobile internet connection (Android 9)
- fixed bug #307 - Mobile data connection is not active
- switched default protocol from `http` to `https` in `WalledGardenInternetObservingStrategy` - solves #323
- added nopen for static code analysis - solves #322
- bumped project dependencies
- updated docs
**Things to do**:
- [x] bump library version
- [x] update javadocs
- [x] perform release by running `release.sh` script
- [x] update `CHANGELOG.md` with date after maven sync
- [x] update docs
- [x] create GitHub release
|
1.0
|
Release 3.0.4 - **Release notes**:
- fixed bug #330 - State CONNECTED sometimes is not returned when wifi is turned off while having mobile internet connection (Android 9)
- fixed bug #307 - Mobile data connection is not active
- switched default protocol from `http` to `https` in `WalledGardenInternetObservingStrategy` - solves #323
- added nopen for static code analysis - solves #322
- bumped project dependencies
- updated docs
**Things to do**:
- [x] bump library version
- [x] update javadocs
- [x] perform release by running `release.sh` script
- [x] update `CHANGELOG.md` with date after maven sync
- [x] update docs
- [x] create GitHub release
|
process
|
release release notes fixed bug state connected sometimes is not returned when wifi is turned off while having mobile internet connection android fixed bug mobile data connection is not active switched default protocol from http to https in walledgardeninternetobservingstrategy solves added nopen for static code analysis solves bumped project dependencies updated docs things to do bump library version update javadocs perform release by running release sh script update changelog md with date after maven sync update docs create github release
| 1
|
553,435
| 16,371,949,899
|
IssuesEvent
|
2021-05-15 10:01:17
|
center-for-knowledge-communication/mathspring
|
https://api.github.com/repos/center-for-knowledge-communication/mathspring
|
closed
|
Teacher Tools - Class Summary Per Student Per Topic report request
|
Super High Priority Teacher Tools enhancement
|
Please freeze the 'Student Name' and 'username' columns on the Class Summary Per Student Per Topic report the same way as in the Class Summary Per Student Per Problem report. Ty
<img width="1073" alt="Screen Shot 2021-03-08 at 1 20 26 PM" src="https://user-images.githubusercontent.com/6811862/110514570-6ffa5c00-80d5-11eb-9567-11aa890b71d2.png">

|
1.0
|
Teacher Tools - Class Summary Per Student Per Topic report request - Please freeze the 'Student Name' and 'username' columns on the Class Summary Per Student Per Topic report the same way as in the Class Summary Per Student Per Problem report. Ty
<img width="1073" alt="Screen Shot 2021-03-08 at 1 20 26 PM" src="https://user-images.githubusercontent.com/6811862/110514570-6ffa5c00-80d5-11eb-9567-11aa890b71d2.png">

|
non_process
|
teacher tools class summary per student per topic report request please freeze the student name and username columns on the class summary per student per topic report the same way as in the class summary per student per problem report ty img width alt screen shot at pm src
| 0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.