Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
248,179
| 7,928,251,009
|
IssuesEvent
|
2018-07-06 10:55:27
|
python/mypy
|
https://api.github.com/repos/python/mypy
|
closed
|
allow overloads to be distinguished by contained type of a generic
|
false-positive feature priority-1-normal topic-overloads
|
This is a follow-up to the discussion beginning at https://github.com/python/typing/issues/253#issuecomment-235652393
Mypy's choice to not distinguish overload signatures by contained types rules out some natural API choices, e.g.
```
@overload
def map(ids: List[int]) -> Dict[int, MyObj]: ...
@overload
def map(ids: List[str]) -> Dict[str, MyObj]: ...
```
We've run into this limitation several times already at Instagram, and there aren't any nice workarounds for it. From a user point of view, the "empty list" rationale doesn't feel right; it seems like it should be possible to only error in specific calls with an unknown contained type.
There is some more discussion in the above-linked thread of how to handle this. In any case, I think it deserves tracking as a mypy limitation.
|
1.0
|
allow overloads to be distinguished by contained type of a generic - This is a follow-up to the discussion beginning at https://github.com/python/typing/issues/253#issuecomment-235652393
Mypy's choice to not distinguish overload signatures by contained types rules out some natural API choices, e.g.
```
@overload
def map(ids: List[int]) -> Dict[int, MyObj]: ...
@overload
def map(ids: List[str]) -> Dict[str, MyObj]: ...
```
We've run into this limitation several times already at Instagram, and there aren't any nice workarounds for it. From a user point of view, the "empty list" rationale doesn't feel right; it seems like it should be possible to only error in specific calls with an unknown contained type.
There is some more discussion in the above-linked thread of how to handle this. In any case, I think it deserves tracking as a mypy limitation.
|
non_process
|
allow overloads to be distinguished by contained type of a generic this is a follow up to the discussion beginning at mypy s choice to not distinguish overload signatures by contained types rules out some natural api choices e g overload def map ids list dict overload def map ids list dict we ve run into this limitation several times already at instagram and there aren t any nice workarounds for it from a user point of view the empty list rationale doesn t feel right it seems like it should be possible to only error in specific calls with an unknown contained type there is some more discussion in the above linked thread of how to handle this in any case i think it deserves tracking as a mypy limitation
| 0
|
285,492
| 31,154,691,209
|
IssuesEvent
|
2023-08-16 12:25:17
|
Trinadh465/linux-4.1.15_CVE-2018-5873
|
https://api.github.com/repos/Trinadh465/linux-4.1.15_CVE-2018-5873
|
opened
|
CVE-2022-3564 (High) detected in linuxlinux-4.1.52
|
Mend: dependency security vulnerability
|
## CVE-2022-3564 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.1.52</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Trinadh465/linux-4.1.15_CVE-2018-5873/commit/32145daf0c96b012284199f23418243e0168269f">32145daf0c96b012284199f23418243e0168269f</a></p>
<p>Found in base branch: <b>main</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/bluetooth/l2cap_core.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/bluetooth/l2cap_core.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
A vulnerability classified as critical was found in Linux Kernel. Affected by this vulnerability is the function l2cap_reassemble_sdu of the file net/bluetooth/l2cap_core.c of the component Bluetooth. The manipulation leads to use after free. It is recommended to apply a patch to fix this issue. The associated identifier of this vulnerability is VDB-211087.
<p>Publish Date: 2022-10-17
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-3564>CVE-2022-3564</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Adjacent
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-3564">https://www.linuxkernelcves.com/cves/CVE-2022-3564</a></p>
<p>Release Date: 2022-10-17</p>
<p>Fix Resolution: v4.9.333,v4.14.299,v4.19.265,v5.4.224,v5.10.154,v5.15.78,v6.0.8,v6.1-rc4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2022-3564 (High) detected in linuxlinux-4.1.52 - ## CVE-2022-3564 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.1.52</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Trinadh465/linux-4.1.15_CVE-2018-5873/commit/32145daf0c96b012284199f23418243e0168269f">32145daf0c96b012284199f23418243e0168269f</a></p>
<p>Found in base branch: <b>main</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/bluetooth/l2cap_core.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/bluetooth/l2cap_core.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
A vulnerability classified as critical was found in Linux Kernel. Affected by this vulnerability is the function l2cap_reassemble_sdu of the file net/bluetooth/l2cap_core.c of the component Bluetooth. The manipulation leads to use after free. It is recommended to apply a patch to fix this issue. The associated identifier of this vulnerability is VDB-211087.
<p>Publish Date: 2022-10-17
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-3564>CVE-2022-3564</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Adjacent
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-3564">https://www.linuxkernelcves.com/cves/CVE-2022-3564</a></p>
<p>Release Date: 2022-10-17</p>
<p>Fix Resolution: v4.9.333,v4.14.299,v4.19.265,v5.4.224,v5.10.154,v5.15.78,v6.0.8,v6.1-rc4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in linuxlinux cve high severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch main vulnerable source files net bluetooth core c net bluetooth core c vulnerability details a vulnerability classified as critical was found in linux kernel affected by this vulnerability is the function reassemble sdu of the file net bluetooth core c of the component bluetooth the manipulation leads to use after free it is recommended to apply a patch to fix this issue the associated identifier of this vulnerability is vdb publish date url a href cvss score details base score metrics exploitability metrics attack vector adjacent attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
43,667
| 23,326,840,960
|
IssuesEvent
|
2022-08-08 22:18:00
|
dotnet/runtime
|
https://api.github.com/repos/dotnet/runtime
|
closed
|
Regressions in System.Tests.Perf_UInt64
|
area-System.Runtime tenet-performance tenet-performance-benchmarks refs/heads/main RunKind=micro Windows 10.0.19041 Regression CoreClr arm64
|
### Run Information
Architecture | arm64
-- | --
OS | Windows 10.0.19041
Baseline | [b92de6bf0351280cd36221f3232b2964a4e61e88](https://github.com/dotnet/runtime/commit/b92de6bf0351280cd36221f3232b2964a4e61e88)
Compare | [d4a9ade2dfbee1ef532e7793ea9c330c51b5c028](https://github.com/dotnet/runtime/commit/d4a9ade2dfbee1ef532e7793ea9c330c51b5c028)
Diff | [Diff](https://github.com/dotnet/runtime/compare/b92de6bf0351280cd36221f3232b2964a4e61e88...d4a9ade2dfbee1ef532e7793ea9c330c51b5c028)
### Regressions in System.Collections.ContainsKeyFalse<String, String>
Benchmark | Baseline | Test | Test/Base | Test Quality | Edge Detector | Baseline IR | Compare IR | IR Ratio | Baseline ETL | Compare ETL
-- | -- | -- | -- | -- | -- | -- | -- | -- | -- | --
[SortedList - Duration of single invocation](<https://pvscmdupload.blob.core.windows.net/reports/allTestHistory/refs/heads/main_arm64_Windows 10.0.19041/System.Collections.ContainsKeyFalse(String%2c%20String).SortedList(Size%3a%20512).html>) | 348.38 μs | 384.99 μs | 1.11 | 0.03 | False | | |
_1.png>)
[Test Report](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/06_30_2022/refs/heads/main_arm64_Windows%2010.0.19041_Regression/System.Collections.ContainsKeyFalse(String,%20String).html>)
### Repro
```cmd
git clone https://github.com/dotnet/performance.git
py .\performance\scripts\benchmarks_ci.py -f net6.0 --filter 'System.Collections.ContainsKeyFalse<String, String>*'
```
<details>
### Payloads
[Baseline](<https://helixdi107v0xdeko0k025g8.blob.core.windows.net/helix-job-f2155b21-97f4-43d1-8ea0-e50a2087f79d0f29de5213c4333af/e46ff89c-808b-4db0-af40-ae793cbb2013.zip?sv=2021-06-08&se=2022-07-24T14%3A05%3A13Z&sr=c&sp=rl&sig=cE1HJSPlysJjjb4SJi%2F%2B1OppkzGCW2pl5nOyJPfjV7Q%3D>)
[Compare](<https://helixdi107v0xdeko0k025g8.blob.core.windows.net/helix-job-2e5ad7fb-ee28-49b3-b168-19d3fb9ca43426a03d146a6421b8d/9785ffd5-9cf7-4ecf-a7be-7bffbaa3ec43.zip?sv=2021-06-08&se=2022-07-25T04%3A32%3A22Z&sr=c&sp=rl&sig=MGAYaVRv7WcZ5GCBKQiHYoQJGKrhexYXQjEGs8Sv64Q%3D>)
### Histogram
#### System.Collections.ContainsKeyFalse<String, String>.SortedList(Size: 512)
```log
```
### Description of detection logic
```IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsRegressionWindowed: Marked as regression because 384.9944918699187 > 366.0416569444444.
IsChangePoint: Marked as a change because one of 6/3/2022 8:17:54 AM, 6/6/2022 3:15:19 PM, 6/24/2022 5:32:42 PM, 6/30/2022 12:21:07 PM falls between 6/21/2022 9:29:01 PM and 6/30/2022 12:21:07 PM.
IsRegressionStdDev: Marked as regression because -18.592276578561822 (T) = (0 -386239.64188622445) / Math.Sqrt((827784.6548322291 / (25)) + (73255526.55193321 / (18))) is less than -2.019540970439573 = MathNet.Numerics.Distributions.StudentT.InvCDF(0, 1, (25) + (18) - 2, .025) and -0.10803726811218585 = (348580.00989829405 - 386239.64188622445) / 348580.00989829405 is less than -0.05.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsChangeEdgeDetector: Marked not as a regression because Edge Detector said so.
```
### Docs
[Profiling workflow for dotnet/runtime repository](https://github.com/dotnet/performance/blob/master/docs/profiling-workflow-dotnet-runtime.md)
[Benchmarking workflow for dotnet/runtime repository](https://github.com/dotnet/performance/blob/master/docs/benchmarking-workflow-dotnet-runtime.md)
</details>
### Run Information
Architecture | arm64
-- | --
OS | Windows 10.0.19041
Baseline | [b92de6bf0351280cd36221f3232b2964a4e61e88](https://github.com/dotnet/runtime/commit/b92de6bf0351280cd36221f3232b2964a4e61e88)
Compare | [d4a9ade2dfbee1ef532e7793ea9c330c51b5c028](https://github.com/dotnet/runtime/commit/d4a9ade2dfbee1ef532e7793ea9c330c51b5c028)
Diff | [Diff](https://github.com/dotnet/runtime/compare/b92de6bf0351280cd36221f3232b2964a4e61e88...d4a9ade2dfbee1ef532e7793ea9c330c51b5c028)
### Regressions in System.Tests.Perf_UInt64
Benchmark | Baseline | Test | Test/Base | Test Quality | Edge Detector | Baseline IR | Compare IR | IR Ratio | Baseline ETL | Compare ETL
-- | -- | -- | -- | -- | -- | -- | -- | -- | -- | --
[TryParseHex - Duration of single invocation](<https://pvscmdupload.blob.core.windows.net/reports/allTestHistory/refs/heads/main_arm64_Windows 10.0.19041/System.Tests.Perf_UInt64.TryParseHex(value%3a%20%220%22).html>) | 4.52 ns | 5.85 ns | 1.30 | 0.35 | False | | |

[Test Report](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/06_30_2022/refs/heads/main_arm64_Windows%2010.0.19041_Regression/System.Tests.Perf_UInt64.html>)
### Repro
```cmd
git clone https://github.com/dotnet/performance.git
py .\performance\scripts\benchmarks_ci.py -f net6.0 --filter 'System.Tests.Perf_UInt64*'
```
<details>
### Payloads
[Baseline](<https://helixdi107v0xdeko0k025g8.blob.core.windows.net/helix-job-f2155b21-97f4-43d1-8ea0-e50a2087f79d0f29de5213c4333af/e46ff89c-808b-4db0-af40-ae793cbb2013.zip?sv=2021-06-08&se=2022-07-24T14%3A05%3A13Z&sr=c&sp=rl&sig=cE1HJSPlysJjjb4SJi%2F%2B1OppkzGCW2pl5nOyJPfjV7Q%3D>)
[Compare](<https://helixdi107v0xdeko0k025g8.blob.core.windows.net/helix-job-2e5ad7fb-ee28-49b3-b168-19d3fb9ca43426a03d146a6421b8d/9785ffd5-9cf7-4ecf-a7be-7bffbaa3ec43.zip?sv=2021-06-08&se=2022-07-25T04%3A32%3A22Z&sr=c&sp=rl&sig=MGAYaVRv7WcZ5GCBKQiHYoQJGKrhexYXQjEGs8Sv64Q%3D>)
### Histogram
#### System.Tests.Perf_UInt64.TryParseHex(value: "0")
```log
```
### Description of detection logic
```IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsRegressionWindowed: Marked as regression because 5.851650109031447 > 5.099529018869719.
IsChangePoint: Marked as a change because one of 3/7/2022 12:49:56 AM, 3/7/2022 10:45:01 PM, 3/12/2022 3:27:04 PM, 3/13/2022 11:50:20 AM, 6/24/2022 5:32:42 PM, 6/30/2022 12:21:07 PM falls between 6/21/2022 9:29:01 PM and 6/30/2022 12:21:07 PM.
IsRegressionStdDev: Marked as regression because -22.045017915913398 (T) = (0 -5.800006004212765) / Math.Sqrt((0.03388441533165363 / (24)) + (0.012577560483055143 / (18))) is less than -2.0210753903043583 = MathNet.Numerics.Distributions.StudentT.InvCDF(0, 1, (24) + (18) - 2, .025) and -0.2115580874843213 = (4.7872289939113815 - 5.800006004212765) / 4.7872289939113815 is less than -0.05.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsChangeEdgeDetector: Marked not as a regression because Edge Detector said so.
```
### Docs
[Profiling workflow for dotnet/runtime repository](https://github.com/dotnet/performance/blob/master/docs/profiling-workflow-dotnet-runtime.md)
[Benchmarking workflow for dotnet/runtime repository](https://github.com/dotnet/performance/blob/master/docs/benchmarking-workflow-dotnet-runtime.md)
</details>
### Run Information
Architecture | arm64
-- | --
OS | Windows 10.0.19041
Baseline | [b92de6bf0351280cd36221f3232b2964a4e61e88](https://github.com/dotnet/runtime/commit/b92de6bf0351280cd36221f3232b2964a4e61e88)
Compare | [d4a9ade2dfbee1ef532e7793ea9c330c51b5c028](https://github.com/dotnet/runtime/commit/d4a9ade2dfbee1ef532e7793ea9c330c51b5c028)
Diff | [Diff](https://github.com/dotnet/runtime/compare/b92de6bf0351280cd36221f3232b2964a4e61e88...d4a9ade2dfbee1ef532e7793ea9c330c51b5c028)
### Regressions in System.Reflection.Attributes
Benchmark | Baseline | Test | Test/Base | Test Quality | Edge Detector | Baseline IR | Compare IR | IR Ratio | Baseline ETL | Compare ETL
-- | -- | -- | -- | -- | -- | -- | -- | -- | -- | --
[IsDefinedMethodOverrideMiss - Duration of single invocation](<https://pvscmdupload.blob.core.windows.net/reports/allTestHistory/refs/heads/main_arm64_Windows 10.0.19041/System.Reflection.Attributes.IsDefinedMethodOverrideMiss.html>) | 660.32 ns | 694.28 ns | 1.05 | 0.18 | False | | |
[IsDefinedClassMissInherit - Duration of single invocation](<https://pvscmdupload.blob.core.windows.net/reports/allTestHistory/refs/heads/main_arm64_Windows 10.0.19041/System.Reflection.Attributes.IsDefinedClassMissInherit.html>) | 1.11 μs | 1.21 μs | 1.09 | 0.13 | False | | |
[IsDefinedMethodOverrideHit - Duration of single invocation](<https://pvscmdupload.blob.core.windows.net/reports/allTestHistory/refs/heads/main_arm64_Windows 10.0.19041/System.Reflection.Attributes.IsDefinedMethodOverrideHit.html>) | 654.04 ns | 705.29 ns | 1.08 | 0.15 | False | | |
[IsDefinedMethodBaseHitInherit - Duration of single invocation](<https://pvscmdupload.blob.core.windows.net/reports/allTestHistory/refs/heads/main_arm64_Windows 10.0.19041/System.Reflection.Attributes.IsDefinedMethodBaseHitInherit.html>) | 660.85 ns | 701.58 ns | 1.06 | 0.16 | False | | |
[IsDefinedClassHit - Duration of single invocation](<https://pvscmdupload.blob.core.windows.net/reports/allTestHistory/refs/heads/main_arm64_Windows 10.0.19041/System.Reflection.Attributes.IsDefinedClassHit.html>) | 606.11 ns | 673.82 ns | 1.11 | 0.19 | False | | |





[Test Report](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/06_30_2022/refs/heads/main_arm64_Windows%2010.0.19041_Regression/System.Reflection.Attributes.html>)
### Repro
```cmd
git clone https://github.com/dotnet/performance.git
py .\performance\scripts\benchmarks_ci.py -f net6.0 --filter 'System.Reflection.Attributes*'
```
<details>
### Payloads
[Baseline](<https://helixdi107v0xdeko0k025g8.blob.core.windows.net/helix-job-f2155b21-97f4-43d1-8ea0-e50a2087f79d0f29de5213c4333af/e46ff89c-808b-4db0-af40-ae793cbb2013.zip?sv=2021-06-08&se=2022-07-24T14%3A05%3A13Z&sr=c&sp=rl&sig=cE1HJSPlysJjjb4SJi%2F%2B1OppkzGCW2pl5nOyJPfjV7Q%3D>)
[Compare](<https://helixdi107v0xdeko0k025g8.blob.core.windows.net/helix-job-2e5ad7fb-ee28-49b3-b168-19d3fb9ca43426a03d146a6421b8d/9785ffd5-9cf7-4ecf-a7be-7bffbaa3ec43.zip?sv=2021-06-08&se=2022-07-25T04%3A32%3A22Z&sr=c&sp=rl&sig=MGAYaVRv7WcZ5GCBKQiHYoQJGKrhexYXQjEGs8Sv64Q%3D>)
### Histogram
#### System.Reflection.Attributes.IsDefinedMethodOverrideMiss
```log
```
### Description of detection logic
```IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsRegressionWindowed: Marked as regression because 694.281147282633 > 668.6631138733229.
IsChangePoint: Marked as a change because one of 3/17/2022 6:54:53 PM, 5/9/2022 6:32:22 AM, 6/24/2022 5:32:42 PM, 6/30/2022 12:21:07 PM falls between 6/21/2022 9:29:01 PM and 6/30/2022 12:21:07 PM.
IsRegressionStdDev: Marked as regression because -8.451246977120563 (T) = (0 -708.0806406154733) / Math.Sqrt((495.03377189893916 / (25)) + (567.5766624145348 / (18))) is less than -2.019540970439573 = MathNet.Numerics.Distributions.StudentT.InvCDF(0, 1, (25) + (18) - 2, .025) and -0.09351064462495992 = (647.5297191627457 - 708.0806406154733) / 647.5297191627457 is less than -0.05.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsChangeEdgeDetector: Marked not as a regression because Edge Detector said so.
```#### System.Reflection.Attributes.IsDefinedClassMissInherit
```log
```
### Description of detection logic
```IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsRegressionWindowed: Marked as regression because 1.2112155248820418 > 1.13694964638556.
IsChangePoint: Marked as a change because one of 5/9/2022 6:32:22 AM, 6/24/2022 5:32:42 PM, 6/30/2022 12:21:07 PM falls between 6/21/2022 9:29:01 PM and 6/30/2022 12:21:07 PM.
IsRegressionStdDev: Marked as regression because -7.745527387390561 (T) = (0 -1213.9608926620335) / Math.Sqrt((1954.7879162499878 / (24)) + (1039.2474975435828 / (18))) is less than -2.0210753903043583 = MathNet.Numerics.Distributions.StudentT.InvCDF(0, 1, (24) + (18) - 2, .025) and -0.08140105517635117 = (1122.581568467274 - 1213.9608926620335) / 1122.581568467274 is less than -0.05.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsChangeEdgeDetector: Marked not as a regression because Edge Detector said so.
```#### System.Reflection.Attributes.IsDefinedMethodOverrideHit
```log
```
### Description of detection logic
```IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsRegressionWindowed: Marked as regression because 705.2855175216763 > 693.5954236978833.
IsChangePoint: Marked as a change because one of 3/17/2022 6:54:53 PM, 5/9/2022 6:32:22 AM, 6/24/2022 5:32:42 PM, 6/30/2022 12:21:07 PM falls between 6/21/2022 9:29:01 PM and 6/30/2022 12:21:07 PM.
IsRegressionStdDev: Marked as regression because -6.1016749220948405 (T) = (0 -709.1377152585584) / Math.Sqrt((647.7044138314708 / (25)) + (994.2426132288017 / (18))) is less than -2.019540970439573 = MathNet.Numerics.Distributions.StudentT.InvCDF(0, 1, (25) + (18) - 2, .025) and -0.08402020507970037 = (654.1738907960857 - 709.1377152585584) / 654.1738907960857 is less than -0.05.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsChangeEdgeDetector: Marked not as a regression because Edge Detector said so.
```#### System.Reflection.Attributes.IsDefinedMethodBaseHitInherit
```log
```
### Description of detection logic
```IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsRegressionWindowed: Marked as regression because 701.575127475433 > 696.2912043948237.
IsChangePoint: Marked as a change because one of 3/17/2022 6:54:53 PM, 5/9/2022 6:32:22 AM, 6/24/2022 5:32:42 PM, 6/30/2022 12:21:07 PM falls between 6/21/2022 9:29:01 PM and 6/30/2022 12:21:07 PM.
IsRegressionStdDev: Marked as regression because -6.276517735042958 (T) = (0 -709.118067717889) / Math.Sqrt((788.1459480483514 / (24)) + (809.2232784828515 / (18))) is less than -2.0210753903043583 = MathNet.Numerics.Distributions.StudentT.InvCDF(0, 1, (24) + (18) - 2, .025) and -0.08468007639528252 = (653.7578067023238 - 709.118067717889) / 653.7578067023238 is less than -0.05.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsChangeEdgeDetector: Marked not as a regression because Edge Detector said so.
```#### System.Reflection.Attributes.IsDefinedClassHit
```log
```
### Description of detection logic
```IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsRegressionWindowed: Marked as regression because 673.8237605760315 > 640.9646320332743.
IsChangePoint: Marked as a change because one of 5/9/2022 6:32:22 AM, 6/22/2022 11:10:17 PM, 6/30/2022 12:21:07 PM falls between 6/21/2022 9:29:01 PM and 6/30/2022 12:21:07 PM.
IsRegressionStdDev: Marked as regression because -6.410882182847598 (T) = (0 -669.5355188242203) / Math.Sqrt((867.0339451611512 / (17)) + (709.075798026562 / (26))) is less than -2.019540970439573 = MathNet.Numerics.Distributions.StudentT.InvCDF(0, 1, (17) + (26) - 2, .025) and -0.09255424522176141 = (612.8167290113098 - 669.5355188242203) / 612.8167290113098 is less than -0.05.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsChangeEdgeDetector: Marked not as a regression because Edge Detector said so.
```
### Docs
[Profiling workflow for dotnet/runtime repository](https://github.com/dotnet/performance/blob/master/docs/profiling-workflow-dotnet-runtime.md)
[Benchmarking workflow for dotnet/runtime repository](https://github.com/dotnet/performance/blob/master/docs/benchmarking-workflow-dotnet-runtime.md)
</details>
|
True
|
Regressions in System.Tests.Perf_UInt64 - ### Run Information
Architecture | arm64
-- | --
OS | Windows 10.0.19041
Baseline | [b92de6bf0351280cd36221f3232b2964a4e61e88](https://github.com/dotnet/runtime/commit/b92de6bf0351280cd36221f3232b2964a4e61e88)
Compare | [d4a9ade2dfbee1ef532e7793ea9c330c51b5c028](https://github.com/dotnet/runtime/commit/d4a9ade2dfbee1ef532e7793ea9c330c51b5c028)
Diff | [Diff](https://github.com/dotnet/runtime/compare/b92de6bf0351280cd36221f3232b2964a4e61e88...d4a9ade2dfbee1ef532e7793ea9c330c51b5c028)
### Regressions in System.Collections.ContainsKeyFalse<String, String>
Benchmark | Baseline | Test | Test/Base | Test Quality | Edge Detector | Baseline IR | Compare IR | IR Ratio | Baseline ETL | Compare ETL
-- | -- | -- | -- | -- | -- | -- | -- | -- | -- | --
[SortedList - Duration of single invocation](<https://pvscmdupload.blob.core.windows.net/reports/allTestHistory/refs/heads/main_arm64_Windows 10.0.19041/System.Collections.ContainsKeyFalse(String%2c%20String).SortedList(Size%3a%20512).html>) | 348.38 μs | 384.99 μs | 1.11 | 0.03 | False | | |
_1.png>)
[Test Report](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/06_30_2022/refs/heads/main_arm64_Windows%2010.0.19041_Regression/System.Collections.ContainsKeyFalse(String,%20String).html>)
### Repro
```cmd
git clone https://github.com/dotnet/performance.git
py .\performance\scripts\benchmarks_ci.py -f net6.0 --filter 'System.Collections.ContainsKeyFalse<String, String>*'
```
<details>
### Payloads
[Baseline](<https://helixdi107v0xdeko0k025g8.blob.core.windows.net/helix-job-f2155b21-97f4-43d1-8ea0-e50a2087f79d0f29de5213c4333af/e46ff89c-808b-4db0-af40-ae793cbb2013.zip?sv=2021-06-08&se=2022-07-24T14%3A05%3A13Z&sr=c&sp=rl&sig=cE1HJSPlysJjjb4SJi%2F%2B1OppkzGCW2pl5nOyJPfjV7Q%3D>)
[Compare](<https://helixdi107v0xdeko0k025g8.blob.core.windows.net/helix-job-2e5ad7fb-ee28-49b3-b168-19d3fb9ca43426a03d146a6421b8d/9785ffd5-9cf7-4ecf-a7be-7bffbaa3ec43.zip?sv=2021-06-08&se=2022-07-25T04%3A32%3A22Z&sr=c&sp=rl&sig=MGAYaVRv7WcZ5GCBKQiHYoQJGKrhexYXQjEGs8Sv64Q%3D>)
### Histogram
#### System.Collections.ContainsKeyFalse<String, String>.SortedList(Size: 512)
```log
```
### Description of detection logic
```IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsRegressionWindowed: Marked as regression because 384.9944918699187 > 366.0416569444444.
IsChangePoint: Marked as a change because one of 6/3/2022 8:17:54 AM, 6/6/2022 3:15:19 PM, 6/24/2022 5:32:42 PM, 6/30/2022 12:21:07 PM falls between 6/21/2022 9:29:01 PM and 6/30/2022 12:21:07 PM.
IsRegressionStdDev: Marked as regression because -18.592276578561822 (T) = (0 -386239.64188622445) / Math.Sqrt((827784.6548322291 / (25)) + (73255526.55193321 / (18))) is less than -2.019540970439573 = MathNet.Numerics.Distributions.StudentT.InvCDF(0, 1, (25) + (18) - 2, .025) and -0.10803726811218585 = (348580.00989829405 - 386239.64188622445) / 348580.00989829405 is less than -0.05.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsChangeEdgeDetector: Marked not as a regression because Edge Detector said so.
```
### Docs
[Profiling workflow for dotnet/runtime repository](https://github.com/dotnet/performance/blob/master/docs/profiling-workflow-dotnet-runtime.md)
[Benchmarking workflow for dotnet/runtime repository](https://github.com/dotnet/performance/blob/master/docs/benchmarking-workflow-dotnet-runtime.md)
</details>
### Run Information
Architecture | arm64
-- | --
OS | Windows 10.0.19041
Baseline | [b92de6bf0351280cd36221f3232b2964a4e61e88](https://github.com/dotnet/runtime/commit/b92de6bf0351280cd36221f3232b2964a4e61e88)
Compare | [d4a9ade2dfbee1ef532e7793ea9c330c51b5c028](https://github.com/dotnet/runtime/commit/d4a9ade2dfbee1ef532e7793ea9c330c51b5c028)
Diff | [Diff](https://github.com/dotnet/runtime/compare/b92de6bf0351280cd36221f3232b2964a4e61e88...d4a9ade2dfbee1ef532e7793ea9c330c51b5c028)
### Regressions in System.Tests.Perf_UInt64
Benchmark | Baseline | Test | Test/Base | Test Quality | Edge Detector | Baseline IR | Compare IR | IR Ratio | Baseline ETL | Compare ETL
-- | -- | -- | -- | -- | -- | -- | -- | -- | -- | --
[TryParseHex - Duration of single invocation](<https://pvscmdupload.blob.core.windows.net/reports/allTestHistory/refs/heads/main_arm64_Windows 10.0.19041/System.Tests.Perf_UInt64.TryParseHex(value%3a%20%220%22).html>) | 4.52 ns | 5.85 ns | 1.30 | 0.35 | False | | |

[Test Report](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/06_30_2022/refs/heads/main_arm64_Windows%2010.0.19041_Regression/System.Tests.Perf_UInt64.html>)
### Repro
```cmd
git clone https://github.com/dotnet/performance.git
py .\performance\scripts\benchmarks_ci.py -f net6.0 --filter 'System.Tests.Perf_UInt64*'
```
<details>
### Payloads
[Baseline](<https://helixdi107v0xdeko0k025g8.blob.core.windows.net/helix-job-f2155b21-97f4-43d1-8ea0-e50a2087f79d0f29de5213c4333af/e46ff89c-808b-4db0-af40-ae793cbb2013.zip?sv=2021-06-08&se=2022-07-24T14%3A05%3A13Z&sr=c&sp=rl&sig=cE1HJSPlysJjjb4SJi%2F%2B1OppkzGCW2pl5nOyJPfjV7Q%3D>)
[Compare](<https://helixdi107v0xdeko0k025g8.blob.core.windows.net/helix-job-2e5ad7fb-ee28-49b3-b168-19d3fb9ca43426a03d146a6421b8d/9785ffd5-9cf7-4ecf-a7be-7bffbaa3ec43.zip?sv=2021-06-08&se=2022-07-25T04%3A32%3A22Z&sr=c&sp=rl&sig=MGAYaVRv7WcZ5GCBKQiHYoQJGKrhexYXQjEGs8Sv64Q%3D>)
### Histogram
#### System.Tests.Perf_UInt64.TryParseHex(value: "0")
```log
```
### Description of detection logic
```IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsRegressionWindowed: Marked as regression because 5.851650109031447 > 5.099529018869719.
IsChangePoint: Marked as a change because one of 3/7/2022 12:49:56 AM, 3/7/2022 10:45:01 PM, 3/12/2022 3:27:04 PM, 3/13/2022 11:50:20 AM, 6/24/2022 5:32:42 PM, 6/30/2022 12:21:07 PM falls between 6/21/2022 9:29:01 PM and 6/30/2022 12:21:07 PM.
IsRegressionStdDev: Marked as regression because -22.045017915913398 (T) = (0 -5.800006004212765) / Math.Sqrt((0.03388441533165363 / (24)) + (0.012577560483055143 / (18))) is less than -2.0210753903043583 = MathNet.Numerics.Distributions.StudentT.InvCDF(0, 1, (24) + (18) - 2, .025) and -0.2115580874843213 = (4.7872289939113815 - 5.800006004212765) / 4.7872289939113815 is less than -0.05.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsChangeEdgeDetector: Marked not as a regression because Edge Detector said so.
```
### Docs
[Profiling workflow for dotnet/runtime repository](https://github.com/dotnet/performance/blob/master/docs/profiling-workflow-dotnet-runtime.md)
[Benchmarking workflow for dotnet/runtime repository](https://github.com/dotnet/performance/blob/master/docs/benchmarking-workflow-dotnet-runtime.md)
</details>
### Run Information
Architecture | arm64
-- | --
OS | Windows 10.0.19041
Baseline | [b92de6bf0351280cd36221f3232b2964a4e61e88](https://github.com/dotnet/runtime/commit/b92de6bf0351280cd36221f3232b2964a4e61e88)
Compare | [d4a9ade2dfbee1ef532e7793ea9c330c51b5c028](https://github.com/dotnet/runtime/commit/d4a9ade2dfbee1ef532e7793ea9c330c51b5c028)
Diff | [Diff](https://github.com/dotnet/runtime/compare/b92de6bf0351280cd36221f3232b2964a4e61e88...d4a9ade2dfbee1ef532e7793ea9c330c51b5c028)
### Regressions in System.Reflection.Attributes
Benchmark | Baseline | Test | Test/Base | Test Quality | Edge Detector | Baseline IR | Compare IR | IR Ratio | Baseline ETL | Compare ETL
-- | -- | -- | -- | -- | -- | -- | -- | -- | -- | --
[IsDefinedMethodOverrideMiss - Duration of single invocation](<https://pvscmdupload.blob.core.windows.net/reports/allTestHistory/refs/heads/main_arm64_Windows 10.0.19041/System.Reflection.Attributes.IsDefinedMethodOverrideMiss.html>) | 660.32 ns | 694.28 ns | 1.05 | 0.18 | False | | |
[IsDefinedClassMissInherit - Duration of single invocation](<https://pvscmdupload.blob.core.windows.net/reports/allTestHistory/refs/heads/main_arm64_Windows 10.0.19041/System.Reflection.Attributes.IsDefinedClassMissInherit.html>) | 1.11 μs | 1.21 μs | 1.09 | 0.13 | False | | |
[IsDefinedMethodOverrideHit - Duration of single invocation](<https://pvscmdupload.blob.core.windows.net/reports/allTestHistory/refs/heads/main_arm64_Windows 10.0.19041/System.Reflection.Attributes.IsDefinedMethodOverrideHit.html>) | 654.04 ns | 705.29 ns | 1.08 | 0.15 | False | | |
[IsDefinedMethodBaseHitInherit - Duration of single invocation](<https://pvscmdupload.blob.core.windows.net/reports/allTestHistory/refs/heads/main_arm64_Windows 10.0.19041/System.Reflection.Attributes.IsDefinedMethodBaseHitInherit.html>) | 660.85 ns | 701.58 ns | 1.06 | 0.16 | False | | |
[IsDefinedClassHit - Duration of single invocation](<https://pvscmdupload.blob.core.windows.net/reports/allTestHistory/refs/heads/main_arm64_Windows 10.0.19041/System.Reflection.Attributes.IsDefinedClassHit.html>) | 606.11 ns | 673.82 ns | 1.11 | 0.19 | False | | |





[Test Report](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/06_30_2022/refs/heads/main_arm64_Windows%2010.0.19041_Regression/System.Reflection.Attributes.html>)
### Repro
```cmd
git clone https://github.com/dotnet/performance.git
py .\performance\scripts\benchmarks_ci.py -f net6.0 --filter 'System.Reflection.Attributes*'
```
<details>
### Payloads
[Baseline](<https://helixdi107v0xdeko0k025g8.blob.core.windows.net/helix-job-f2155b21-97f4-43d1-8ea0-e50a2087f79d0f29de5213c4333af/e46ff89c-808b-4db0-af40-ae793cbb2013.zip?sv=2021-06-08&se=2022-07-24T14%3A05%3A13Z&sr=c&sp=rl&sig=cE1HJSPlysJjjb4SJi%2F%2B1OppkzGCW2pl5nOyJPfjV7Q%3D>)
[Compare](<https://helixdi107v0xdeko0k025g8.blob.core.windows.net/helix-job-2e5ad7fb-ee28-49b3-b168-19d3fb9ca43426a03d146a6421b8d/9785ffd5-9cf7-4ecf-a7be-7bffbaa3ec43.zip?sv=2021-06-08&se=2022-07-25T04%3A32%3A22Z&sr=c&sp=rl&sig=MGAYaVRv7WcZ5GCBKQiHYoQJGKrhexYXQjEGs8Sv64Q%3D>)
### Histogram
#### System.Reflection.Attributes.IsDefinedMethodOverrideMiss
```log
```
### Description of detection logic
```IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsRegressionWindowed: Marked as regression because 694.281147282633 > 668.6631138733229.
IsChangePoint: Marked as a change because one of 3/17/2022 6:54:53 PM, 5/9/2022 6:32:22 AM, 6/24/2022 5:32:42 PM, 6/30/2022 12:21:07 PM falls between 6/21/2022 9:29:01 PM and 6/30/2022 12:21:07 PM.
IsRegressionStdDev: Marked as regression because -8.451246977120563 (T) = (0 -708.0806406154733) / Math.Sqrt((495.03377189893916 / (25)) + (567.5766624145348 / (18))) is less than -2.019540970439573 = MathNet.Numerics.Distributions.StudentT.InvCDF(0, 1, (25) + (18) - 2, .025) and -0.09351064462495992 = (647.5297191627457 - 708.0806406154733) / 647.5297191627457 is less than -0.05.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsChangeEdgeDetector: Marked not as a regression because Edge Detector said so.
```#### System.Reflection.Attributes.IsDefinedClassMissInherit
```log
```
### Description of detection logic
```IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsRegressionWindowed: Marked as regression because 1.2112155248820418 > 1.13694964638556.
IsChangePoint: Marked as a change because one of 5/9/2022 6:32:22 AM, 6/24/2022 5:32:42 PM, 6/30/2022 12:21:07 PM falls between 6/21/2022 9:29:01 PM and 6/30/2022 12:21:07 PM.
IsRegressionStdDev: Marked as regression because -7.745527387390561 (T) = (0 -1213.9608926620335) / Math.Sqrt((1954.7879162499878 / (24)) + (1039.2474975435828 / (18))) is less than -2.0210753903043583 = MathNet.Numerics.Distributions.StudentT.InvCDF(0, 1, (24) + (18) - 2, .025) and -0.08140105517635117 = (1122.581568467274 - 1213.9608926620335) / 1122.581568467274 is less than -0.05.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsChangeEdgeDetector: Marked not as a regression because Edge Detector said so.
```#### System.Reflection.Attributes.IsDefinedMethodOverrideHit
```log
```
### Description of detection logic
```IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsRegressionWindowed: Marked as regression because 705.2855175216763 > 693.5954236978833.
IsChangePoint: Marked as a change because one of 3/17/2022 6:54:53 PM, 5/9/2022 6:32:22 AM, 6/24/2022 5:32:42 PM, 6/30/2022 12:21:07 PM falls between 6/21/2022 9:29:01 PM and 6/30/2022 12:21:07 PM.
IsRegressionStdDev: Marked as regression because -6.1016749220948405 (T) = (0 -709.1377152585584) / Math.Sqrt((647.7044138314708 / (25)) + (994.2426132288017 / (18))) is less than -2.019540970439573 = MathNet.Numerics.Distributions.StudentT.InvCDF(0, 1, (25) + (18) - 2, .025) and -0.08402020507970037 = (654.1738907960857 - 709.1377152585584) / 654.1738907960857 is less than -0.05.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsChangeEdgeDetector: Marked not as a regression because Edge Detector said so.
```#### System.Reflection.Attributes.IsDefinedMethodBaseHitInherit
```log
```
### Description of detection logic
```IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsRegressionWindowed: Marked as regression because 701.575127475433 > 696.2912043948237.
IsChangePoint: Marked as a change because one of 3/17/2022 6:54:53 PM, 5/9/2022 6:32:22 AM, 6/24/2022 5:32:42 PM, 6/30/2022 12:21:07 PM falls between 6/21/2022 9:29:01 PM and 6/30/2022 12:21:07 PM.
IsRegressionStdDev: Marked as regression because -6.276517735042958 (T) = (0 -709.118067717889) / Math.Sqrt((788.1459480483514 / (24)) + (809.2232784828515 / (18))) is less than -2.0210753903043583 = MathNet.Numerics.Distributions.StudentT.InvCDF(0, 1, (24) + (18) - 2, .025) and -0.08468007639528252 = (653.7578067023238 - 709.118067717889) / 653.7578067023238 is less than -0.05.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsChangeEdgeDetector: Marked not as a regression because Edge Detector said so.
```#### System.Reflection.Attributes.IsDefinedClassHit
```log
```
### Description of detection logic
```IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsRegressionBase: Marked as regression because the compare was 5% greater than the baseline, and the value was not too small.
IsRegressionChecked: Marked as regression because the three check build points were 0.05 greater than the baseline.
IsRegressionWindowed: Marked as regression because 673.8237605760315 > 640.9646320332743.
IsChangePoint: Marked as a change because one of 5/9/2022 6:32:22 AM, 6/22/2022 11:10:17 PM, 6/30/2022 12:21:07 PM falls between 6/21/2022 9:29:01 PM and 6/30/2022 12:21:07 PM.
IsRegressionStdDev: Marked as regression because -6.410882182847598 (T) = (0 -669.5355188242203) / Math.Sqrt((867.0339451611512 / (17)) + (709.075798026562 / (26))) is less than -2.019540970439573 = MathNet.Numerics.Distributions.StudentT.InvCDF(0, 1, (17) + (26) - 2, .025) and -0.09255424522176141 = (612.8167290113098 - 669.5355188242203) / 612.8167290113098 is less than -0.05.
IsImprovementBase: Marked as not an improvement because the compare was not 5% less than the baseline, or the value was too small.
IsChangeEdgeDetector: Marked not as a regression because Edge Detector said so.
```
### Docs
[Profiling workflow for dotnet/runtime repository](https://github.com/dotnet/performance/blob/master/docs/profiling-workflow-dotnet-runtime.md)
[Benchmarking workflow for dotnet/runtime repository](https://github.com/dotnet/performance/blob/master/docs/benchmarking-workflow-dotnet-runtime.md)
</details>
|
non_process
|
regressions in system tests perf run information architecture os windows baseline compare diff regressions in system collections containskeyfalse lt string string gt benchmark baseline test test base test quality edge detector baseline ir compare ir ir ratio baseline etl compare etl μs μs false repro cmd git clone py performance scripts benchmarks ci py f filter system collections containskeyfalse lt string string gt payloads histogram system collections containskeyfalse lt string string gt sortedlist size log description of detection logic isregressionbase marked as regression because the compare was greater than the baseline and the value was not too small isregressionchecked marked as regression because the three check build points were greater than the baseline isimprovementbase marked as not an improvement because the compare was not less than the baseline or the value was too small isregressionbase marked as regression because the compare was greater than the baseline and the value was not too small isregressionchecked marked as regression because the three check build points were greater than the baseline isregressionwindowed marked as regression because ischangepoint marked as a change because one of am pm pm pm falls between pm and pm isregressionstddev marked as regression because t math sqrt is less than mathnet numerics distributions studentt invcdf and is less than isimprovementbase marked as not an improvement because the compare was not less than the baseline or the value was too small ischangeedgedetector marked not as a regression because edge detector said so docs run information architecture os windows baseline compare diff regressions in system tests perf benchmark baseline test test base test quality edge detector baseline ir compare ir ir ratio baseline etl compare etl ns ns false repro cmd git clone py performance scripts benchmarks ci py f filter system tests perf payloads histogram system tests perf tryparsehex value log description of detection logic isregressionbase marked as regression because the compare was greater than the baseline and the value was not too small isregressionchecked marked as regression because the three check build points were greater than the baseline isimprovementbase marked as not an improvement because the compare was not less than the baseline or the value was too small isregressionbase marked as regression because the compare was greater than the baseline and the value was not too small isregressionchecked marked as regression because the three check build points were greater than the baseline isregressionwindowed marked as regression because ischangepoint marked as a change because one of am pm pm am pm pm falls between pm and pm isregressionstddev marked as regression because t math sqrt is less than mathnet numerics distributions studentt invcdf and is less than isimprovementbase marked as not an improvement because the compare was not less than the baseline or the value was too small ischangeedgedetector marked not as a regression because edge detector said so docs run information architecture os windows baseline compare diff regressions in system reflection attributes benchmark baseline test test base test quality edge detector baseline ir compare ir ir ratio baseline etl compare etl ns ns false μs μs false ns ns false ns ns false ns ns false repro cmd git clone py performance scripts benchmarks ci py f filter system reflection attributes payloads histogram system reflection attributes isdefinedmethodoverridemiss log description of detection logic isregressionbase marked as regression because the compare was greater than the baseline and the value was not too small isregressionchecked marked as regression because the three check build points were greater than the baseline isimprovementbase marked as not an improvement because the compare was not less than the baseline or the value was too small isregressionbase marked as regression because the compare was greater than the baseline and the value was not too small isregressionchecked marked as regression because the three check build points were greater than the baseline isregressionwindowed marked as regression because ischangepoint marked as a change because one of pm am pm pm falls between pm and pm isregressionstddev marked as regression because t math sqrt is less than mathnet numerics distributions studentt invcdf and is less than isimprovementbase marked as not an improvement because the compare was not less than the baseline or the value was too small ischangeedgedetector marked not as a regression because edge detector said so system reflection attributes isdefinedclassmissinherit log description of detection logic isregressionbase marked as regression because the compare was greater than the baseline and the value was not too small isregressionchecked marked as regression because the three check build points were greater than the baseline isimprovementbase marked as not an improvement because the compare was not less than the baseline or the value was too small isregressionbase marked as regression because the compare was greater than the baseline and the value was not too small isregressionchecked marked as regression because the three check build points were greater than the baseline isregressionwindowed marked as regression because ischangepoint marked as a change because one of am pm pm falls between pm and pm isregressionstddev marked as regression because t math sqrt is less than mathnet numerics distributions studentt invcdf and is less than isimprovementbase marked as not an improvement because the compare was not less than the baseline or the value was too small ischangeedgedetector marked not as a regression because edge detector said so system reflection attributes isdefinedmethodoverridehit log description of detection logic isregressionbase marked as regression because the compare was greater than the baseline and the value was not too small isregressionchecked marked as regression because the three check build points were greater than the baseline isimprovementbase marked as not an improvement because the compare was not less than the baseline or the value was too small isregressionbase marked as regression because the compare was greater than the baseline and the value was not too small isregressionchecked marked as regression because the three check build points were greater than the baseline isregressionwindowed marked as regression because ischangepoint marked as a change because one of pm am pm pm falls between pm and pm isregressionstddev marked as regression because t math sqrt is less than mathnet numerics distributions studentt invcdf and is less than isimprovementbase marked as not an improvement because the compare was not less than the baseline or the value was too small ischangeedgedetector marked not as a regression because edge detector said so system reflection attributes isdefinedmethodbasehitinherit log description of detection logic isregressionbase marked as regression because the compare was greater than the baseline and the value was not too small isregressionchecked marked as regression because the three check build points were greater than the baseline isimprovementbase marked as not an improvement because the compare was not less than the baseline or the value was too small isregressionbase marked as regression because the compare was greater than the baseline and the value was not too small isregressionchecked marked as regression because the three check build points were greater than the baseline isregressionwindowed marked as regression because ischangepoint marked as a change because one of pm am pm pm falls between pm and pm isregressionstddev marked as regression because t math sqrt is less than mathnet numerics distributions studentt invcdf and is less than isimprovementbase marked as not an improvement because the compare was not less than the baseline or the value was too small ischangeedgedetector marked not as a regression because edge detector said so system reflection attributes isdefinedclasshit log description of detection logic isregressionbase marked as regression because the compare was greater than the baseline and the value was not too small isregressionchecked marked as regression because the three check build points were greater than the baseline isimprovementbase marked as not an improvement because the compare was not less than the baseline or the value was too small isregressionbase marked as regression because the compare was greater than the baseline and the value was not too small isregressionchecked marked as regression because the three check build points were greater than the baseline isregressionwindowed marked as regression because ischangepoint marked as a change because one of am pm pm falls between pm and pm isregressionstddev marked as regression because t math sqrt is less than mathnet numerics distributions studentt invcdf and is less than isimprovementbase marked as not an improvement because the compare was not less than the baseline or the value was too small ischangeedgedetector marked not as a regression because edge detector said so docs
| 0
|
363,973
| 25,476,838,130
|
IssuesEvent
|
2022-11-25 15:16:00
|
rune004/mkdocs
|
https://api.github.com/repos/rune004/mkdocs
|
opened
|
Upload Ansible Documentation
|
documentation
|
Upload all Ansible Documentation needed for deployment via a Ubuntu controller.
|
1.0
|
Upload Ansible Documentation - Upload all Ansible Documentation needed for deployment via a Ubuntu controller.
|
non_process
|
upload ansible documentation upload all ansible documentation needed for deployment via a ubuntu controller
| 0
|
5,600
| 8,460,166,563
|
IssuesEvent
|
2018-10-22 18:02:40
|
aspnet/IISIntegration
|
https://api.github.com/repos/aspnet/IISIntegration
|
closed
|
ANCM TCP packet loss
|
investigate out-of-process
|
I have a setup when multiple sites (asp.net core, asp.net, python) are hosted in IIS together with [Shibboleth Service Provider](https://wiki.shibboleth.net/confluence/display/SP3/Home) which acts before request reaches the actual site code and passes authenticated user information to applications via request headers. It works fine until one of the headers get bigger (over 16 KB of base64 encoded text). Once it happens leading part of the request seem to do be dropped on transport level.
I've setup a wireshark to watch TCP packets going from IIS to Kestrel. When header values are small enough packets represent valid HTTP requests. When header got bigger the request seems to begin with part of long header value. First request line is missing, most of header definitions are missing. The message begins with truncated value of long header (which I've checked is really part of what should be send in that header) and ends with header value tail and possible other smaller header which might go after.
It seems like some buffer or memory overflow issue inside ANCM because it happens for ASP NET Core sites only. Similar setup works perfectly for ASP.NET and python sites for the same users.
Also, the same setup (ASP.NET Core sites) works perfectly on one machine but behaves as described above on another. Both machines are Windows 10 using same versions of IIS and same netcore hosting bundle (2.1.2).
Is there some setting I am missing ? What could go wrong ?
|
1.0
|
ANCM TCP packet loss - I have a setup when multiple sites (asp.net core, asp.net, python) are hosted in IIS together with [Shibboleth Service Provider](https://wiki.shibboleth.net/confluence/display/SP3/Home) which acts before request reaches the actual site code and passes authenticated user information to applications via request headers. It works fine until one of the headers get bigger (over 16 KB of base64 encoded text). Once it happens leading part of the request seem to do be dropped on transport level.
I've setup a wireshark to watch TCP packets going from IIS to Kestrel. When header values are small enough packets represent valid HTTP requests. When header got bigger the request seems to begin with part of long header value. First request line is missing, most of header definitions are missing. The message begins with truncated value of long header (which I've checked is really part of what should be send in that header) and ends with header value tail and possible other smaller header which might go after.
It seems like some buffer or memory overflow issue inside ANCM because it happens for ASP NET Core sites only. Similar setup works perfectly for ASP.NET and python sites for the same users.
Also, the same setup (ASP.NET Core sites) works perfectly on one machine but behaves as described above on another. Both machines are Windows 10 using same versions of IIS and same netcore hosting bundle (2.1.2).
Is there some setting I am missing ? What could go wrong ?
|
process
|
ancm tcp packet loss i have a setup when multiple sites asp net core asp net python are hosted in iis together with which acts before request reaches the actual site code and passes authenticated user information to applications via request headers it works fine until one of the headers get bigger over kb of encoded text once it happens leading part of the request seem to do be dropped on transport level i ve setup a wireshark to watch tcp packets going from iis to kestrel when header values are small enough packets represent valid http requests when header got bigger the request seems to begin with part of long header value first request line is missing most of header definitions are missing the message begins with truncated value of long header which i ve checked is really part of what should be send in that header and ends with header value tail and possible other smaller header which might go after it seems like some buffer or memory overflow issue inside ancm because it happens for asp net core sites only similar setup works perfectly for asp net and python sites for the same users also the same setup asp net core sites works perfectly on one machine but behaves as described above on another both machines are windows using same versions of iis and same netcore hosting bundle is there some setting i am missing what could go wrong
| 1
|
15,018
| 18,732,508,482
|
IssuesEvent
|
2021-11-04 00:20:37
|
woowacourse/prolog
|
https://api.github.com/repos/woowacourse/prolog
|
closed
|
신규 멤버 온보딩을 진행한다.
|
process
|
## 신규 멤버 온보딩
- 기존 인수 테스트를 모두 실행시켜보면서 기존 기능을 이해하기
- 인수 테스트 실행을 비롯한 로컬 개발 환경 설정은 멘토가 하루 날잡고 함께 도와주기
- 개선이 필요하거나 부족한 부분이 있는 인수 테스트가 있으면 개선해서 PR올리기(멘토가 리뷰 할 예정)
## 체크
- [x] 바다
- [x] 수리
- [x] 코다
|
1.0
|
신규 멤버 온보딩을 진행한다. - ## 신규 멤버 온보딩
- 기존 인수 테스트를 모두 실행시켜보면서 기존 기능을 이해하기
- 인수 테스트 실행을 비롯한 로컬 개발 환경 설정은 멘토가 하루 날잡고 함께 도와주기
- 개선이 필요하거나 부족한 부분이 있는 인수 테스트가 있으면 개선해서 PR올리기(멘토가 리뷰 할 예정)
## 체크
- [x] 바다
- [x] 수리
- [x] 코다
|
process
|
신규 멤버 온보딩을 진행한다 신규 멤버 온보딩 기존 인수 테스트를 모두 실행시켜보면서 기존 기능을 이해하기 인수 테스트 실행을 비롯한 로컬 개발 환경 설정은 멘토가 하루 날잡고 함께 도와주기 개선이 필요하거나 부족한 부분이 있는 인수 테스트가 있으면 개선해서 pr올리기 멘토가 리뷰 할 예정 체크 바다 수리 코다
| 1
|
159,697
| 20,085,894,240
|
IssuesEvent
|
2022-02-05 01:08:15
|
AkshayMukkavilli/Tensorflow
|
https://api.github.com/repos/AkshayMukkavilli/Tensorflow
|
opened
|
CVE-2021-41221 (High) detected in tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl
|
security vulnerability
|
## CVE-2021-41221 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary>
<p>TensorFlow is an open source machine learning framework for everyone.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</a></p>
<p>Path to dependency file: /Tensorflow/src/requirements.txt</p>
<p>Path to vulnerable library: /teSource-ArchiveExtractor_5ea86033-7612-4210-97f3-8edb65806ddf/20190525011619_2843/20190525011537_depth_0/2/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64/tensorflow-1.13.1.data/purelib/tensorflow</p>
<p>
Dependency Hierarchy:
- :x: **tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
TensorFlow is an open source platform for machine learning. In affected versions the shape inference code for the `Cudnn*` operations in TensorFlow can be tricked into accessing invalid memory, via a heap buffer overflow. This occurs because the ranks of the `input`, `input_h` and `input_c` parameters are not validated, but code assumes they have certain values. The fix will be included in TensorFlow 2.7.0. We will also cherrypick this commit on TensorFlow 2.6.1, TensorFlow 2.5.2, and TensorFlow 2.4.4, as these are also affected and still in supported range.
<p>Publish Date: 2021-11-05
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-41221>CVE-2021-41221</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-4f99-p9c2-3j8x">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-4f99-p9c2-3j8x</a></p>
<p>Release Date: 2021-11-05</p>
<p>Fix Resolution: tensorflow - 2.4.4, 2.5.2, 2.6.1, 2.7.0;tensorflow-cpu - 2.4.4, 2.5.2, 2.6.1, 2.7.0;tensorflow-gpu - 2.4.4, 2.5.2, 2.6.1, 2.7.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-41221 (High) detected in tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl - ## CVE-2021-41221 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary>
<p>TensorFlow is an open source machine learning framework for everyone.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</a></p>
<p>Path to dependency file: /Tensorflow/src/requirements.txt</p>
<p>Path to vulnerable library: /teSource-ArchiveExtractor_5ea86033-7612-4210-97f3-8edb65806ddf/20190525011619_2843/20190525011537_depth_0/2/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64/tensorflow-1.13.1.data/purelib/tensorflow</p>
<p>
Dependency Hierarchy:
- :x: **tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
TensorFlow is an open source platform for machine learning. In affected versions the shape inference code for the `Cudnn*` operations in TensorFlow can be tricked into accessing invalid memory, via a heap buffer overflow. This occurs because the ranks of the `input`, `input_h` and `input_c` parameters are not validated, but code assumes they have certain values. The fix will be included in TensorFlow 2.7.0. We will also cherrypick this commit on TensorFlow 2.6.1, TensorFlow 2.5.2, and TensorFlow 2.4.4, as these are also affected and still in supported range.
<p>Publish Date: 2021-11-05
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-41221>CVE-2021-41221</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-4f99-p9c2-3j8x">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-4f99-p9c2-3j8x</a></p>
<p>Release Date: 2021-11-05</p>
<p>Fix Resolution: tensorflow - 2.4.4, 2.5.2, 2.6.1, 2.7.0;tensorflow-cpu - 2.4.4, 2.5.2, 2.6.1, 2.7.0;tensorflow-gpu - 2.4.4, 2.5.2, 2.6.1, 2.7.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in tensorflow whl cve high severity vulnerability vulnerable library tensorflow whl tensorflow is an open source machine learning framework for everyone library home page a href path to dependency file tensorflow src requirements txt path to vulnerable library tesource archiveextractor depth tensorflow tensorflow data purelib tensorflow dependency hierarchy x tensorflow whl vulnerable library vulnerability details tensorflow is an open source platform for machine learning in affected versions the shape inference code for the cudnn operations in tensorflow can be tricked into accessing invalid memory via a heap buffer overflow this occurs because the ranks of the input input h and input c parameters are not validated but code assumes they have certain values the fix will be included in tensorflow we will also cherrypick this commit on tensorflow tensorflow and tensorflow as these are also affected and still in supported range publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tensorflow tensorflow cpu tensorflow gpu step up your open source security game with whitesource
| 0
|
10,723
| 4,817,936,749
|
IssuesEvent
|
2016-11-04 15:05:07
|
uProxy/uproxy
|
https://api.github.com/repos/uProxy/uproxy
|
closed
|
choose an API documentation generator
|
C:BuildProcess C:Code-Cleanup P3 T:Obsolete
|
Right now, we have a mish-mash of comment styles. The style guide specifies C++-style comments but we have tons of C-style comments, too, some of them with JSDoc-style annotations. At minimum, let's:
1. investigate API documentation generators for TypeScript (maybe JS generators are fine)
2. investigate what code completion tools exist for TypeScript
|
1.0
|
choose an API documentation generator - Right now, we have a mish-mash of comment styles. The style guide specifies C++-style comments but we have tons of C-style comments, too, some of them with JSDoc-style annotations. At minimum, let's:
1. investigate API documentation generators for TypeScript (maybe JS generators are fine)
2. investigate what code completion tools exist for TypeScript
|
non_process
|
choose an api documentation generator right now we have a mish mash of comment styles the style guide specifies c style comments but we have tons of c style comments too some of them with jsdoc style annotations at minimum let s investigate api documentation generators for typescript maybe js generators are fine investigate what code completion tools exist for typescript
| 0
|
338,294
| 10,227,138,006
|
IssuesEvent
|
2019-08-16 19:51:49
|
onaio/reveal-frontend
|
https://api.github.com/repos/onaio/reveal-frontend
|
closed
|
Debug Focus Area Profile Page
|
Priority: High bug has pr
|
This page seems to have a number of problems that we need to investigate:
- [x] The map sometimes loads, sometimes does not load and sometimes loads then starts reloading
- [x] The list of focus areas on this page should be filtered by the focus area (jurisdiction) id. This is not the case
- Some of the queries sent to Superset result in weird 400 errors (see #350)

|
1.0
|
Debug Focus Area Profile Page - This page seems to have a number of problems that we need to investigate:
- [x] The map sometimes loads, sometimes does not load and sometimes loads then starts reloading
- [x] The list of focus areas on this page should be filtered by the focus area (jurisdiction) id. This is not the case
- Some of the queries sent to Superset result in weird 400 errors (see #350)

|
non_process
|
debug focus area profile page this page seems to have a number of problems that we need to investigate the map sometimes loads sometimes does not load and sometimes loads then starts reloading the list of focus areas on this page should be filtered by the focus area jurisdiction id this is not the case some of the queries sent to superset result in weird errors see
| 0
|
143,599
| 11,570,438,180
|
IssuesEvent
|
2020-02-20 19:30:55
|
CBICA/CaPTk
|
https://api.github.com/repos/CBICA/CaPTk
|
closed
|
CaPTk 1.7.6 GUI Failure
|
Testathon-Feb-2020 wontfix
|
**Describe the bug**
Failed to load CaPTk GUI from cluster. Crashes immediately upon execution of 'captk' in CL. Have tried to unload captk before loading the new version as well.
**To Reproduce**
Steps to reproduce the behavior:
1. On the cluster
2. "Module load captk/1.7.6"
3. "captk"
4. See error
**Expected behavior**
The CaPTk GUI should open.
**Screenshots**
This is the error I am receiving:
Checking for compatible OpenGL - this will happen only once.
WARNING: Trying to run CaPTk GUI using software rendering - this might not work on all systems and in those cases, only the CLI will be available.
[0220/112554.064064:WARNING:stack_trace_posix.cc(699)] Failed to open file: /scratch/desaish/#9175046 (deleted)
Error: No such file or directory
ApplicationPreferences::DisplayPreferences()
font = ""
theme = ""
ApplicationPreferences::DisplayPreferences()
font = "Sans Serif,9,-1,5,50,0,0,0,0,0"
theme = "Dark"
ApplicationPreferences::DisplayPreferences()
font = "Sans Serif,9,-1,5,50,0,0,0,0,0"
theme = "Dark"
p11-kit: couldn't list directory: /etc/pki/ca-trust/source/anchors: Permission denied
The X11 connection broke: I/O error (code 1)
XIO: fatal IO error 0 (Success) on X server "170.166.98.52:17.0"
after 720 requests (720 known processed) with 0 events remaining.
**CaPTk Version**
New 1.7.6
**Desktop (please complete the following information):**
Windows 10 Enterprise
Version 1809
OS Build 17763.973
**Additional context**
I have tried going through the captk path which I believe is: /cbica/software/lab/captk/centos7/1.7.6/bin
From here I am able to access specific CaPTk software like FeatureExtraction/BreastTexturePipeline but so far have not been able to access the GUI for 1.7.6
|
1.0
|
CaPTk 1.7.6 GUI Failure - **Describe the bug**
Failed to load CaPTk GUI from cluster. Crashes immediately upon execution of 'captk' in CL. Have tried to unload captk before loading the new version as well.
**To Reproduce**
Steps to reproduce the behavior:
1. On the cluster
2. "Module load captk/1.7.6"
3. "captk"
4. See error
**Expected behavior**
The CaPTk GUI should open.
**Screenshots**
This is the error I am receiving:
Checking for compatible OpenGL - this will happen only once.
WARNING: Trying to run CaPTk GUI using software rendering - this might not work on all systems and in those cases, only the CLI will be available.
[0220/112554.064064:WARNING:stack_trace_posix.cc(699)] Failed to open file: /scratch/desaish/#9175046 (deleted)
Error: No such file or directory
ApplicationPreferences::DisplayPreferences()
font = ""
theme = ""
ApplicationPreferences::DisplayPreferences()
font = "Sans Serif,9,-1,5,50,0,0,0,0,0"
theme = "Dark"
ApplicationPreferences::DisplayPreferences()
font = "Sans Serif,9,-1,5,50,0,0,0,0,0"
theme = "Dark"
p11-kit: couldn't list directory: /etc/pki/ca-trust/source/anchors: Permission denied
The X11 connection broke: I/O error (code 1)
XIO: fatal IO error 0 (Success) on X server "170.166.98.52:17.0"
after 720 requests (720 known processed) with 0 events remaining.
**CaPTk Version**
New 1.7.6
**Desktop (please complete the following information):**
Windows 10 Enterprise
Version 1809
OS Build 17763.973
**Additional context**
I have tried going through the captk path which I believe is: /cbica/software/lab/captk/centos7/1.7.6/bin
From here I am able to access specific CaPTk software like FeatureExtraction/BreastTexturePipeline but so far have not been able to access the GUI for 1.7.6
|
non_process
|
captk gui failure describe the bug failed to load captk gui from cluster crashes immediately upon execution of captk in cl have tried to unload captk before loading the new version as well to reproduce steps to reproduce the behavior on the cluster module load captk captk see error expected behavior the captk gui should open screenshots this is the error i am receiving checking for compatible opengl this will happen only once warning trying to run captk gui using software rendering this might not work on all systems and in those cases only the cli will be available failed to open file scratch desaish deleted error no such file or directory applicationpreferences displaypreferences font theme applicationpreferences displaypreferences font sans serif theme dark applicationpreferences displaypreferences font sans serif theme dark kit couldn t list directory etc pki ca trust source anchors permission denied the connection broke i o error code xio fatal io error success on x server after requests known processed with events remaining captk version new desktop please complete the following information windows enterprise version os build additional context i have tried going through the captk path which i believe is cbica software lab captk bin from here i am able to access specific captk software like featureextraction breasttexturepipeline but so far have not been able to access the gui for
| 0
|
3,831
| 6,802,428,844
|
IssuesEvent
|
2017-11-02 20:10:31
|
gratipay/inside.gratipay.com
|
https://api.github.com/repos/gratipay/inside.gratipay.com
|
closed
|
Clarify terms around notification
|
Governance & Process
|
Reticketed from https://github.com/gratipay/inside.gratipay.com/issues/204 via https://github.com/gratipay/gratipay.com/pull/4117#issuecomment-262336157.
>> Regarding the “without notice to you” provisions in §11.4 and §12.1, I would **very** much like to be informed of it, just as under §17.2.
>
>I think these clauses mean "without prior notice to you." We certainly notify people of things after the fact, but it's Gratipay's discretion what to keep or remove in any context. "With prior notice" would imply that we're asking permission, which is not what will happen under pending legal action (copyright notice, etc.).
>
>>On that topic, it would be really nice if you could add a possibility for community involvement in the process of changing the ToS & PP, similar to the one in §4.5 (when a Team is being estabilshed).
>
>I think this is implied?
(Note that section numbers will change when https://github.com/gratipay/gratipay.com/pull/4117 lands.)
|
1.0
|
Clarify terms around notification - Reticketed from https://github.com/gratipay/inside.gratipay.com/issues/204 via https://github.com/gratipay/gratipay.com/pull/4117#issuecomment-262336157.
>> Regarding the “without notice to you” provisions in §11.4 and §12.1, I would **very** much like to be informed of it, just as under §17.2.
>
>I think these clauses mean "without prior notice to you." We certainly notify people of things after the fact, but it's Gratipay's discretion what to keep or remove in any context. "With prior notice" would imply that we're asking permission, which is not what will happen under pending legal action (copyright notice, etc.).
>
>>On that topic, it would be really nice if you could add a possibility for community involvement in the process of changing the ToS & PP, similar to the one in §4.5 (when a Team is being estabilshed).
>
>I think this is implied?
(Note that section numbers will change when https://github.com/gratipay/gratipay.com/pull/4117 lands.)
|
process
|
clarify terms around notification reticketed from via regarding the “without notice to you” provisions in § and § i would very much like to be informed of it just as under § i think these clauses mean without prior notice to you we certainly notify people of things after the fact but it s gratipay s discretion what to keep or remove in any context with prior notice would imply that we re asking permission which is not what will happen under pending legal action copyright notice etc on that topic it would be really nice if you could add a possibility for community involvement in the process of changing the tos pp similar to the one in § when a team is being estabilshed i think this is implied note that section numbers will change when lands
| 1
|
227,401
| 17,381,920,919
|
IssuesEvent
|
2021-07-31 22:24:55
|
moja-global/FLINT-UI
|
https://api.github.com/repos/moja-global/FLINT-UI
|
closed
|
Reference Vue 2 style guide in README
|
documentation good first issue
|
The Vue docs are amazing. Let's make sure our UI developers know we follow them by adding a section to our README on coding style and providing a link:
~https://v3.vuejs.org/style-guide/~
EDIT: see comment from @waridrox
|
1.0
|
Reference Vue 2 style guide in README - The Vue docs are amazing. Let's make sure our UI developers know we follow them by adding a section to our README on coding style and providing a link:
~https://v3.vuejs.org/style-guide/~
EDIT: see comment from @waridrox
|
non_process
|
reference vue style guide in readme the vue docs are amazing let s make sure our ui developers know we follow them by adding a section to our readme on coding style and providing a link edit see comment from waridrox
| 0
|
253,536
| 19,122,574,283
|
IssuesEvent
|
2021-12-01 01:15:01
|
gingerchicken/como-client
|
https://api.github.com/repos/gingerchicken/como-client
|
opened
|
Update README to new Java
|
documentation
|
I am pretty sure Java 17 is now required to build the mod, maybe update the README to say this.
|
1.0
|
Update README to new Java - I am pretty sure Java 17 is now required to build the mod, maybe update the README to say this.
|
non_process
|
update readme to new java i am pretty sure java is now required to build the mod maybe update the readme to say this
| 0
|
651,117
| 21,465,876,185
|
IssuesEvent
|
2022-04-26 03:37:00
|
ballerina-platform/ballerina-extended-library
|
https://api.github.com/repos/ballerina-platform/ballerina-extended-library
|
opened
|
[Improvement]: Provide pagination support for List and Query operations in Salesforce connector
|
Priority/High Type/Improvement Team/Connector Component/Connector
|
### Connector Name
module/salesforce (Salesforce)
### Suggested improvement
The pagination of Salesforce responses are not handled internally in the Salesforce connector. In the current approach users can call the `getQueryResult` operation to get the first page only. Then he has to call the `getNextQueryResult` operation by passing the `nextRecordsUrl` obtained from the previous call as a parameter and obtain the subsequent pages.
The fix will contain implementing Ballerina stream implementer to handle the second call on demand internally without user intervention.
### Related issues
_No response_
|
1.0
|
[Improvement]: Provide pagination support for List and Query operations in Salesforce connector - ### Connector Name
module/salesforce (Salesforce)
### Suggested improvement
The pagination of Salesforce responses are not handled internally in the Salesforce connector. In the current approach users can call the `getQueryResult` operation to get the first page only. Then he has to call the `getNextQueryResult` operation by passing the `nextRecordsUrl` obtained from the previous call as a parameter and obtain the subsequent pages.
The fix will contain implementing Ballerina stream implementer to handle the second call on demand internally without user intervention.
### Related issues
_No response_
|
non_process
|
provide pagination support for list and query operations in salesforce connector connector name module salesforce salesforce suggested improvement the pagination of salesforce responses are not handled internally in the salesforce connector in the current approach users can call the getqueryresult operation to get the first page only then he has to call the getnextqueryresult operation by passing the nextrecordsurl obtained from the previous call as a parameter and obtain the subsequent pages the fix will contain implementing ballerina stream implementer to handle the second call on demand internally without user intervention related issues no response
| 0
|
210,268
| 7,187,423,728
|
IssuesEvent
|
2018-02-02 05:10:35
|
wso2/testgrid
|
https://api.github.com/repos/wso2/testgrid
|
opened
|
TestGrid dev dashboard in apps.wso2 is pointing to the wrong URL
|
Priority/Highest Severity/Critical Type/Bug
|
**Description:**
TestGrid dev dashboard in apps.wso2 is pointing to the wrong URL (`https://testgrid-live-dev.private.wso2.com/testgrid/dashboard/api/acs`). This should be (`https://testgrid-live-dev.private.wso2.com/testgrid/dashboard/`)
|
1.0
|
TestGrid dev dashboard in apps.wso2 is pointing to the wrong URL - **Description:**
TestGrid dev dashboard in apps.wso2 is pointing to the wrong URL (`https://testgrid-live-dev.private.wso2.com/testgrid/dashboard/api/acs`). This should be (`https://testgrid-live-dev.private.wso2.com/testgrid/dashboard/`)
|
non_process
|
testgrid dev dashboard in apps is pointing to the wrong url description testgrid dev dashboard in apps is pointing to the wrong url this should be
| 0
|
82,494
| 7,842,843,588
|
IssuesEvent
|
2018-06-19 02:01:42
|
DynamoRIO/dynamorio
|
https://api.github.com/repos/DynamoRIO/dynamorio
|
closed
|
tool.drcachesim.TLB-threads failing non-deterministically on Travis
|
Bug-AppFail Component-Tests
|
It terminates early but at 90s it may be a timeout (though there's no timeout message).
https://api.travis-ci.org/jobs/174603662/log.txt?deansi=true
```
211:
211: Finished computing current solution distance in mode 0.
211: Mode changed to 0.
211:
211: Started iteration 10 of the computation...
216/248 Test #211: code_api|tool.drcachesim.TLB-threads .............................***Failed Required regular expression not found.Regex=[^
```
|
1.0
|
tool.drcachesim.TLB-threads failing non-deterministically on Travis - It terminates early but at 90s it may be a timeout (though there's no timeout message).
https://api.travis-ci.org/jobs/174603662/log.txt?deansi=true
```
211:
211: Finished computing current solution distance in mode 0.
211: Mode changed to 0.
211:
211: Started iteration 10 of the computation...
216/248 Test #211: code_api|tool.drcachesim.TLB-threads .............................***Failed Required regular expression not found.Regex=[^
```
|
non_process
|
tool drcachesim tlb threads failing non deterministically on travis it terminates early but at it may be a timeout though there s no timeout message finished computing current solution distance in mode mode changed to started iteration of the computation test code api tool drcachesim tlb threads failed required regular expression not found regex
| 0
|
576,964
| 17,100,067,530
|
IssuesEvent
|
2021-07-09 09:56:34
|
PlaceOS/driver
|
https://api.github.com/repos/PlaceOS/driver
|
closed
|
Refactor build process
|
priority: high type: enhancement
|
Drivers are currently compiled via setting an environment variable that points to the driver's entry point and compiling a common `build.cr`.
The most intuitive process would be...
- `require "placeos-driver"` like any other library, and write your driver code below.
- Point the compiler at the file the driver was written in and compile it like any other application.
|
1.0
|
Refactor build process - Drivers are currently compiled via setting an environment variable that points to the driver's entry point and compiling a common `build.cr`.
The most intuitive process would be...
- `require "placeos-driver"` like any other library, and write your driver code below.
- Point the compiler at the file the driver was written in and compile it like any other application.
|
non_process
|
refactor build process drivers are currently compiled via setting an environment variable that points to the driver s entry point and compiling a common build cr the most intuitive process would be require placeos driver like any other library and write your driver code below point the compiler at the file the driver was written in and compile it like any other application
| 0
|
46,534
| 6,021,258,811
|
IssuesEvent
|
2017-06-07 18:17:25
|
18F/omb-eregs
|
https://api.github.com/repos/18F/omb-eregs
|
closed
|
Review user research to date and create/make public a synthesized documentation of it
|
design
|
verbs
- [X] read through existing research documentation and notes
- [X] write up a synthesis doc (user archetypes)
- [x] have team (esp. Micah and Nicole) review and add comments/clarifications
- [x] publish on GitHub wiki
|
1.0
|
Review user research to date and create/make public a synthesized documentation of it - verbs
- [X] read through existing research documentation and notes
- [X] write up a synthesis doc (user archetypes)
- [x] have team (esp. Micah and Nicole) review and add comments/clarifications
- [x] publish on GitHub wiki
|
non_process
|
review user research to date and create make public a synthesized documentation of it verbs read through existing research documentation and notes write up a synthesis doc user archetypes have team esp micah and nicole review and add comments clarifications publish on github wiki
| 0
|
2,763
| 5,695,995,600
|
IssuesEvent
|
2017-04-16 05:59:56
|
AllenFang/react-bootstrap-table
|
https://api.github.com/repos/AllenFang/react-bootstrap-table
|
closed
|
Autocomplete input field in insert Modal
|
duplicate enhancement inprocess
|
I would like to have an input text field in my Insert Modal with autocomplete functionality (given a set of values). I don't really want to implement my custom Modal - the perfect solution would be if the file Editor.js had an additional case for type 'autocomplete' or something similar.What I mean by autocomplete is something like this: https://www.w3schools.com/tags/tryit.asp?filename=tryhtml5_datalist
Since I am new to webdevelopment and javascript in general, I wanted to ask advice if it can be done with the current react-bootstrap-table and how. Obviously I don't want to overwrite the original Editor.js file in my node_modules directory and extend it there.
If it is not possible at the moment, could you maybe add such functionality, AllenFang?
|
1.0
|
Autocomplete input field in insert Modal - I would like to have an input text field in my Insert Modal with autocomplete functionality (given a set of values). I don't really want to implement my custom Modal - the perfect solution would be if the file Editor.js had an additional case for type 'autocomplete' or something similar.What I mean by autocomplete is something like this: https://www.w3schools.com/tags/tryit.asp?filename=tryhtml5_datalist
Since I am new to webdevelopment and javascript in general, I wanted to ask advice if it can be done with the current react-bootstrap-table and how. Obviously I don't want to overwrite the original Editor.js file in my node_modules directory and extend it there.
If it is not possible at the moment, could you maybe add such functionality, AllenFang?
|
process
|
autocomplete input field in insert modal i would like to have an input text field in my insert modal with autocomplete functionality given a set of values i don t really want to implement my custom modal the perfect solution would be if the file editor js had an additional case for type autocomplete or something similar what i mean by autocomplete is something like this since i am new to webdevelopment and javascript in general i wanted to ask advice if it can be done with the current react bootstrap table and how obviously i don t want to overwrite the original editor js file in my node modules directory and extend it there if it is not possible at the moment could you maybe add such functionality allenfang
| 1
|
431,245
| 30,224,462,884
|
IssuesEvent
|
2023-07-05 22:29:24
|
rancher-sandbox/rancher-desktop
|
https://api.github.com/repos/rancher-sandbox/rancher-desktop
|
closed
|
Document information from original Upgrade Responder client epic in forked Upgrade Responder repo
|
kind/documentation-devhelp
|
In [the original Upgrade Responder client epic](https://github.com/rancher-sandbox/rancher-desktop/issues/3925) there is a lot of information about how Upgrade Responder works. We need to document this properly so that anybody looking to learn about it doesn't have to figure out how it should be from reading a long issue.
|
1.0
|
Document information from original Upgrade Responder client epic in forked Upgrade Responder repo - In [the original Upgrade Responder client epic](https://github.com/rancher-sandbox/rancher-desktop/issues/3925) there is a lot of information about how Upgrade Responder works. We need to document this properly so that anybody looking to learn about it doesn't have to figure out how it should be from reading a long issue.
|
non_process
|
document information from original upgrade responder client epic in forked upgrade responder repo in there is a lot of information about how upgrade responder works we need to document this properly so that anybody looking to learn about it doesn t have to figure out how it should be from reading a long issue
| 0
|
48,606
| 10,264,289,065
|
IssuesEvent
|
2019-08-22 16:01:27
|
cityofaustin/techstack
|
https://api.github.com/repos/cityofaustin/techstack
|
reopened
|
Footer copy changes
|
Copy; not code Janis Alpha Team: Dev
|
**Acceptance criteria**
- [ ] In English, lowercase "project site"
- [ ] In English and Spanish, "austintexas.gov" should all be lowercase. (We only capitalize a URL if it begins a sentence.)
- [ ] In Spanish, sentence case for the following sentence: "Encuentre más información en nuestro sitio del proyecto"
- [ ] In Spanish, project site hyperlink should only be applied to "sitio del proyecto," not the full sentence


|
1.0
|
Footer copy changes - **Acceptance criteria**
- [ ] In English, lowercase "project site"
- [ ] In English and Spanish, "austintexas.gov" should all be lowercase. (We only capitalize a URL if it begins a sentence.)
- [ ] In Spanish, sentence case for the following sentence: "Encuentre más información en nuestro sitio del proyecto"
- [ ] In Spanish, project site hyperlink should only be applied to "sitio del proyecto," not the full sentence


|
non_process
|
footer copy changes acceptance criteria in english lowercase project site in english and spanish austintexas gov should all be lowercase we only capitalize a url if it begins a sentence in spanish sentence case for the following sentence encuentre más información en nuestro sitio del proyecto in spanish project site hyperlink should only be applied to sitio del proyecto not the full sentence
| 0
|
9,812
| 12,824,241,624
|
IssuesEvent
|
2020-07-06 13:10:53
|
keep-network/keep-core
|
https://api.github.com/repos/keep-network/keep-core
|
opened
|
Stake delegation system tests
|
process & client team
|
Having top-ups implemented we need to execute system tests covering all possible stake delegation scenarios.
Depending on the progress on #1898, tests should be performed using KEEP token dashboard or directly against smart contracts.
- [ ] Grantee can delegate a stake
- [ ] Managed grantee can delegate a stake
- [ ] Owner of liquid tokens can delegate a stake
- [ ] Stake delegation can be canceled within the initialization period by a grantee directly on the staking contract, tokens are deposited to the escrow
- [ ] Stake delegation can be canceled within the initialization period by a managed grantee directly on the staking contract, tokens are deposited to the escrow
- [ ] Stake delegation can be canceled within the initialization period by an operator directly on the staking contract, tokens are returned to the owner's account
- [ ] Stake can be undelegated by grantee directly on the staking contract
- [ ] Stake can be undelegated by a managed grantee directly on the staking contract
- [ ] Stake can be undelegated by liquid tokens owner directly on the staking contract
- [ ] Tokens staked from the grant are deposited in the escrow after recovering them from the delegation
- [ ] Staked, liquid tokens are sent back to the owner after recovering them from the delegation
- [ ] Tokens can be withdrawn from escrow according to grant's unlocking schedule
- [ ] It is possible to top-up existing delegation from tokens deposited in the escrow
- [ ] It is possible to top-up existing delegation from tokens from a grant
- [ ] It is possible to top-up existing delegation from liquid tokens
- [ ] Tokens deposited in the escrow can be redelegated to a new operator
|
1.0
|
Stake delegation system tests - Having top-ups implemented we need to execute system tests covering all possible stake delegation scenarios.
Depending on the progress on #1898, tests should be performed using KEEP token dashboard or directly against smart contracts.
- [ ] Grantee can delegate a stake
- [ ] Managed grantee can delegate a stake
- [ ] Owner of liquid tokens can delegate a stake
- [ ] Stake delegation can be canceled within the initialization period by a grantee directly on the staking contract, tokens are deposited to the escrow
- [ ] Stake delegation can be canceled within the initialization period by a managed grantee directly on the staking contract, tokens are deposited to the escrow
- [ ] Stake delegation can be canceled within the initialization period by an operator directly on the staking contract, tokens are returned to the owner's account
- [ ] Stake can be undelegated by grantee directly on the staking contract
- [ ] Stake can be undelegated by a managed grantee directly on the staking contract
- [ ] Stake can be undelegated by liquid tokens owner directly on the staking contract
- [ ] Tokens staked from the grant are deposited in the escrow after recovering them from the delegation
- [ ] Staked, liquid tokens are sent back to the owner after recovering them from the delegation
- [ ] Tokens can be withdrawn from escrow according to grant's unlocking schedule
- [ ] It is possible to top-up existing delegation from tokens deposited in the escrow
- [ ] It is possible to top-up existing delegation from tokens from a grant
- [ ] It is possible to top-up existing delegation from liquid tokens
- [ ] Tokens deposited in the escrow can be redelegated to a new operator
|
process
|
stake delegation system tests having top ups implemented we need to execute system tests covering all possible stake delegation scenarios depending on the progress on tests should be performed using keep token dashboard or directly against smart contracts grantee can delegate a stake managed grantee can delegate a stake owner of liquid tokens can delegate a stake stake delegation can be canceled within the initialization period by a grantee directly on the staking contract tokens are deposited to the escrow stake delegation can be canceled within the initialization period by a managed grantee directly on the staking contract tokens are deposited to the escrow stake delegation can be canceled within the initialization period by an operator directly on the staking contract tokens are returned to the owner s account stake can be undelegated by grantee directly on the staking contract stake can be undelegated by a managed grantee directly on the staking contract stake can be undelegated by liquid tokens owner directly on the staking contract tokens staked from the grant are deposited in the escrow after recovering them from the delegation staked liquid tokens are sent back to the owner after recovering them from the delegation tokens can be withdrawn from escrow according to grant s unlocking schedule it is possible to top up existing delegation from tokens deposited in the escrow it is possible to top up existing delegation from tokens from a grant it is possible to top up existing delegation from liquid tokens tokens deposited in the escrow can be redelegated to a new operator
| 1
|
15,992
| 4,003,910,388
|
IssuesEvent
|
2016-05-12 03:36:31
|
abstratt/textuml
|
https://api.github.com/repos/abstratt/textuml
|
closed
|
Provide documentation for the collection/grouping operations in TextUML
|
documentation
|
See abstratt/cloudfier#85.
|
1.0
|
Provide documentation for the collection/grouping operations in TextUML - See abstratt/cloudfier#85.
|
non_process
|
provide documentation for the collection grouping operations in textuml see abstratt cloudfier
| 0
|
8,378
| 7,372,892,294
|
IssuesEvent
|
2018-03-13 15:50:41
|
OpenLiberty/open-liberty
|
https://api.github.com/repos/OpenLiberty/open-liberty
|
closed
|
When building lib.index.cache, preserve original jar names
|
springboot team:OSGi Infrastructure
|
The lib.index.cache currently renames the jars to sha-1 hashes of their contents. This is a little confusing when the jar names appear in error message. Preserve the original jar names in a folder named after the hash.
|
1.0
|
When building lib.index.cache, preserve original jar names - The lib.index.cache currently renames the jars to sha-1 hashes of their contents. This is a little confusing when the jar names appear in error message. Preserve the original jar names in a folder named after the hash.
|
non_process
|
when building lib index cache preserve original jar names the lib index cache currently renames the jars to sha hashes of their contents this is a little confusing when the jar names appear in error message preserve the original jar names in a folder named after the hash
| 0
|
346,525
| 24,886,957,300
|
IssuesEvent
|
2022-10-28 08:37:03
|
Yongbeom-Kim/ped
|
https://api.github.com/repos/Yongbeom-Kim/ped
|
opened
|
Inconsistency in attendance command documentation
|
severity.Low type.DocumentationBug
|
The documentation for attendance is not updated, as seen below:

<!--session: 1666944979154-bbc9b0eb-75d6-4d6c-be94-8f557b314ad1-->
<!--Version: Web v3.4.4-->
|
1.0
|
Inconsistency in attendance command documentation - The documentation for attendance is not updated, as seen below:

<!--session: 1666944979154-bbc9b0eb-75d6-4d6c-be94-8f557b314ad1-->
<!--Version: Web v3.4.4-->
|
non_process
|
inconsistency in attendance command documentation the documentation for attendance is not updated as seen below
| 0
|
16,835
| 22,068,051,199
|
IssuesEvent
|
2022-05-31 06:40:16
|
googleapis/google-cloud-dotnet
|
https://api.github.com/repos/googleapis/google-cloud-dotnet
|
opened
|
Translation snippet tests are failing
|
type: process
|
(This is probably just a matter of changing expectations again.)
|
1.0
|
Translation snippet tests are failing - (This is probably just a matter of changing expectations again.)
|
process
|
translation snippet tests are failing this is probably just a matter of changing expectations again
| 1
|
64,266
| 8,721,361,429
|
IssuesEvent
|
2018-12-08 22:14:26
|
facebook/create-react-app
|
https://api.github.com/repos/facebook/create-react-app
|
closed
|
Unclear default code splitting in cra v2
|
tag: documentation
|
<!--
PLEASE READ THE FIRST SECTION :-)
-->
### Is this a bug report?
If very unclear documentation is a bug then yes.
A newly created and built app with `Create react app v2` creates three js chucks by default. But there is not a single line in neither index.js nor in App.js that does code splitting according to this guide
https://reactjs.org/docs/code-splitting.html
I could not find any info, not in release notes, nor anywhere else why there are three bundles now without any dynamic import statements, what is the point and benefits of having three bundles even for 120K app, and what to do if I want only one js but want to keep other features of cra v2.
There is a ticket to disable a code splitting (https://github.com/facebook/create-react-app/issues/5306) but could you please elaborate a bit why it there in the default app in the first place if neither line of code asks for that.
|
1.0
|
Unclear default code splitting in cra v2 - <!--
PLEASE READ THE FIRST SECTION :-)
-->
### Is this a bug report?
If very unclear documentation is a bug then yes.
A newly created and built app with `Create react app v2` creates three js chucks by default. But there is not a single line in neither index.js nor in App.js that does code splitting according to this guide
https://reactjs.org/docs/code-splitting.html
I could not find any info, not in release notes, nor anywhere else why there are three bundles now without any dynamic import statements, what is the point and benefits of having three bundles even for 120K app, and what to do if I want only one js but want to keep other features of cra v2.
There is a ticket to disable a code splitting (https://github.com/facebook/create-react-app/issues/5306) but could you please elaborate a bit why it there in the default app in the first place if neither line of code asks for that.
|
non_process
|
unclear default code splitting in cra please read the first section is this a bug report if very unclear documentation is a bug then yes a newly created and built app with create react app creates three js chucks by default but there is not a single line in neither index js nor in app js that does code splitting according to this guide i could not find any info not in release notes nor anywhere else why there are three bundles now without any dynamic import statements what is the point and benefits of having three bundles even for app and what to do if i want only one js but want to keep other features of cra there is a ticket to disable a code splitting but could you please elaborate a bit why it there in the default app in the first place if neither line of code asks for that
| 0
|
42,396
| 11,013,587,047
|
IssuesEvent
|
2019-12-04 20:48:08
|
cliffparnitzky/CheckedEmail
|
https://api.github.com/repos/cliffparnitzky/CheckedEmail
|
closed
|
After completing the registration module, the mail address in the backend will not be displayed
|
Defect
|
Ich verwende die Erweiterung für die Überprüfung der Mail-Adresse bei der Mitgliederregistrierung (Frontend-Modul).
Das funktioniert soweit wunderbar. Leider wird die im Frontend bei der Registrierung eingegebene Mailadresse aber im Backend nicht angezeigt. In der Tabelle tl_member ist sie aber korrekt gespeichert.
Contao 4.8 In den Einstellungen ist E-Mail Feld mit Bestätigung Mitglieder aktiviert
|
1.0
|
After completing the registration module, the mail address in the backend will not be displayed - Ich verwende die Erweiterung für die Überprüfung der Mail-Adresse bei der Mitgliederregistrierung (Frontend-Modul).
Das funktioniert soweit wunderbar. Leider wird die im Frontend bei der Registrierung eingegebene Mailadresse aber im Backend nicht angezeigt. In der Tabelle tl_member ist sie aber korrekt gespeichert.
Contao 4.8 In den Einstellungen ist E-Mail Feld mit Bestätigung Mitglieder aktiviert
|
non_process
|
after completing the registration module the mail address in the backend will not be displayed ich verwende die erweiterung für die überprüfung der mail adresse bei der mitgliederregistrierung frontend modul das funktioniert soweit wunderbar leider wird die im frontend bei der registrierung eingegebene mailadresse aber im backend nicht angezeigt in der tabelle tl member ist sie aber korrekt gespeichert contao in den einstellungen ist e mail feld mit bestätigung mitglieder aktiviert
| 0
|
21,029
| 27,969,934,531
|
IssuesEvent
|
2023-03-25 00:19:49
|
darktable-org/darktable
|
https://api.github.com/repos/darktable-org/darktable
|
closed
|
Raster masks segfault if ROI size is changed
|
understood: unclear scope: image processing no-issue-activity
|
1. Define a crop in crop or crop and rotate or perspective modules,
2. Define a parametric mask in some module,
3. Re-use that mask as a raster mask later in pipe,
4. Disable the cropping module,
5. Witness :
```
Thread 14 "worker res 0" received signal SIGSEGV, Segmentation fault.
[Switching to Thread 0x7fffb37fe640 (LWP 120649)]
dt_iop_image_scaled_copy (buf=buf@entry=0x7ffee1b3e040, src=src@entry=0x7fff6585a040, scale=scale@entry=0.5, width=width@entry=1365, height=height@entry=1982, ch=ch@entry=1)
at src/common/imagebuf.c:220
220 buf[k] = scale * src[k];
```
`dt_iop_image_scaled_copy` is only called from `src/develop/blend.c` to copy raster masks. My guess is when the ROI size is changed, the raster mask buffer is not flushed and resized accordingly.
|
1.0
|
Raster masks segfault if ROI size is changed - 1. Define a crop in crop or crop and rotate or perspective modules,
2. Define a parametric mask in some module,
3. Re-use that mask as a raster mask later in pipe,
4. Disable the cropping module,
5. Witness :
```
Thread 14 "worker res 0" received signal SIGSEGV, Segmentation fault.
[Switching to Thread 0x7fffb37fe640 (LWP 120649)]
dt_iop_image_scaled_copy (buf=buf@entry=0x7ffee1b3e040, src=src@entry=0x7fff6585a040, scale=scale@entry=0.5, width=width@entry=1365, height=height@entry=1982, ch=ch@entry=1)
at src/common/imagebuf.c:220
220 buf[k] = scale * src[k];
```
`dt_iop_image_scaled_copy` is only called from `src/develop/blend.c` to copy raster masks. My guess is when the ROI size is changed, the raster mask buffer is not flushed and resized accordingly.
|
process
|
raster masks segfault if roi size is changed define a crop in crop or crop and rotate or perspective modules define a parametric mask in some module re use that mask as a raster mask later in pipe disable the cropping module witness thread worker res received signal sigsegv segmentation fault dt iop image scaled copy buf buf entry src src entry scale scale entry width width entry height height entry ch ch entry at src common imagebuf c buf scale src dt iop image scaled copy is only called from src develop blend c to copy raster masks my guess is when the roi size is changed the raster mask buffer is not flushed and resized accordingly
| 1
|
8,799
| 11,908,254,855
|
IssuesEvent
|
2020-03-31 00:25:15
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
processing doesn't output a shapefile's encoding
|
Bug Feedback Processing
|
Author Name: **Tobias Wendorff** (Tobias Wendorff)
Original Redmine Issue: [20556](https://issues.qgis.org/issues/20556)
Affected QGIS version: 3.5(master)
Redmine category:processing/core
---
The tools in "processing", like "split vector layer" don't write any information about encoding of shapefiles.
Postprocessing the files is very hard because you either have to know where they come from and which charset they might be in.
---
Related issue(s): #28434 (relates)
Redmine related issue(s): [20614](https://issues.qgis.org/issues/20614)
---
|
1.0
|
processing doesn't output a shapefile's encoding - Author Name: **Tobias Wendorff** (Tobias Wendorff)
Original Redmine Issue: [20556](https://issues.qgis.org/issues/20556)
Affected QGIS version: 3.5(master)
Redmine category:processing/core
---
The tools in "processing", like "split vector layer" don't write any information about encoding of shapefiles.
Postprocessing the files is very hard because you either have to know where they come from and which charset they might be in.
---
Related issue(s): #28434 (relates)
Redmine related issue(s): [20614](https://issues.qgis.org/issues/20614)
---
|
process
|
processing doesn t output a shapefile s encoding author name tobias wendorff tobias wendorff original redmine issue affected qgis version master redmine category processing core the tools in processing like split vector layer don t write any information about encoding of shapefiles postprocessing the files is very hard because you either have to know where they come from and which charset they might be in related issue s relates redmine related issue s
| 1
|
1,441
| 4,007,034,570
|
IssuesEvent
|
2016-05-12 16:43:53
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
Intermittent failures in test-process-exec-argv
|
process test
|
* **Version**: 6.0 (built from source)
* **Platform**: Seen on s390 but believe could occur on any
* **Subsystem**: process
We see this test fail intermittently (about 0.4% of the time on one of our machines) with:
<PRE>
680 - test-process-exec-argv.js
not ok 680 test-process-exec-argv.js
# undefined:1
#
#
#
# SyntaxError: Unexpected end of JSON input
# at Object.parse (native)
# at ChildProcess.<anonymous> (/build/jenkins/new/workspace/ibmnode-v6-test-linux/ARCH/s390/DISTRO/rhel7/GCC/gcc48/OS/linux/SUITE/node/labels/bvt/node/test/parallel/test-process-exec-argv.js:19:33)
# at emitTwo (events.js:106:13)
# at ChildProcess.emit (events.js:191:7)
# at Process.ChildProcess._handle.onexit (internal/child_process.js:204:12)
</PRE>
We believe the cause is related to the issue being discussed here: https://github.com/nodejs/node/issues/6456
Opening this issue as the outcome of https://github.com/nodejs/node/issues/6456 may or may not resolve the failing test depending on what's decided as the right way forward so we need to track this failure to make sure the test is either fixed by what we do for 6456 or that the test is updated to remove the dependency on the process exit timing if what we do for 6456 does not resolve the issue
|
1.0
|
Intermittent failures in test-process-exec-argv - * **Version**: 6.0 (built from source)
* **Platform**: Seen on s390 but believe could occur on any
* **Subsystem**: process
We see this test fail intermittently (about 0.4% of the time on one of our machines) with:
<PRE>
680 - test-process-exec-argv.js
not ok 680 test-process-exec-argv.js
# undefined:1
#
#
#
# SyntaxError: Unexpected end of JSON input
# at Object.parse (native)
# at ChildProcess.<anonymous> (/build/jenkins/new/workspace/ibmnode-v6-test-linux/ARCH/s390/DISTRO/rhel7/GCC/gcc48/OS/linux/SUITE/node/labels/bvt/node/test/parallel/test-process-exec-argv.js:19:33)
# at emitTwo (events.js:106:13)
# at ChildProcess.emit (events.js:191:7)
# at Process.ChildProcess._handle.onexit (internal/child_process.js:204:12)
</PRE>
We believe the cause is related to the issue being discussed here: https://github.com/nodejs/node/issues/6456
Opening this issue as the outcome of https://github.com/nodejs/node/issues/6456 may or may not resolve the failing test depending on what's decided as the right way forward so we need to track this failure to make sure the test is either fixed by what we do for 6456 or that the test is updated to remove the dependency on the process exit timing if what we do for 6456 does not resolve the issue
|
process
|
intermittent failures in test process exec argv version built from source platform seen on but believe could occur on any subsystem process we see this test fail intermittently about of the time on one of our machines with test process exec argv js not ok test process exec argv js undefined syntaxerror unexpected end of json input at object parse native at childprocess build jenkins new workspace ibmnode test linux arch distro gcc os linux suite node labels bvt node test parallel test process exec argv js at emittwo events js at childprocess emit events js at process childprocess handle onexit internal child process js we believe the cause is related to the issue being discussed here opening this issue as the outcome of may or may not resolve the failing test depending on what s decided as the right way forward so we need to track this failure to make sure the test is either fixed by what we do for or that the test is updated to remove the dependency on the process exit timing if what we do for does not resolve the issue
| 1
|
71,911
| 18,925,034,242
|
IssuesEvent
|
2021-11-17 08:37:09
|
firoorg/firo
|
https://api.github.com/repos/firoorg/firo
|
closed
|
'make cov' is broken
|
build-system Priority 2
|
## Actual behaviour
```
zcoin-builder@ac9b1304543f:~/zcoin$ make cov
/usr/bin/lcov --gcov-tool=/usr/bin/gcov -c -i -d /home/zcoin-builder/zcoin/src -o baseline.info
Capturing coverage data from /home/zcoin-builder/zcoin/src
Found gcov version: 4.8.4
Scanning /home/zcoin-builder/zcoin/src for .gcno files ...
Found 509 graph files in /home/zcoin-builder/zcoin/src
Processing src/compat/libbitcoin_util_a-glibc_sanity.gcno
Processing src/compat/libbitcoin_util_a-glibcxx_sanity.gcno
Processing src/compat/libbitcoin_util_a-strnlen.gcno
Processing src/consensus/.libs/libbitcoinconsensus_la-merkle.gcno
Processing src/consensus/libbitcoinconsensus_la-merkle.gcno
Processing src/consensus/libbitcoin_consensus_a-merkle.gcno
Processing src/crypto/MerkleTreeProof/libbitcoin_util_a-mtp.gcno
geninfo: Unsuccessful stat on filename containing newline at /usr/bin/geninfo line 2554.
geninfo: Unsuccessful stat on filename containing newline at /usr/bin/geninfo line 2554.
geninfo: Unsuccessful stat on filename containing newline at /usr/bin/geninfo line 2554.
geninfo: Unsuccessful stat on filename containing newline at /usr/bin/geninfo line 2554.
geninfo: Unsuccessful stat on filename containing newline at /usr/bin/geninfo line 2554.
geninfo: Unsuccessful stat on filename containing newline at /usr/bin/geninfo line 2554.
geninfo: Unsuccessful stat on filename containing newline at /usr/bin/geninfo line 2554.
geninfo: WARNING: could not open
geninfo: WARNING: could not open /home/zcoin-builder/zcoin/src
geninfo: WARNING: could not open /home/zcoin-builder/zcoin/src/
geninfo: WARNING: could not open /home/zcoin-builder/zcoin/src/
geninfo: WARNING: could not open /home/zcoin-builder/zcoin/src/
geninfo: WARNING: could not open /usr/include/bits/uintn-identity.h
geninfo: WARNING: could not open /usr/include/boost/system/detail/generic_category.hpp
geninfo: WARNING: could not open /usr/include/boost/system/detail/std_interoperability.hpp
geninfo: WARNING: could not open /usr/include/boost/system/detail/system_category_posix.hpp
geninfo: WARNING: could not open /usr/include/c++/9.1.0/bits/allocator.h
geninfo: WARNING: could not open /usr/include/c++/9.1.0/bits/basic_string.h
geninfo: WARNING: could not open /usr/include/c++/9.1.0/bits/char_traits.h
geninfo: WARNING: could not open /usr/include/c++/9.1.0/bits/deque.tcc
geninfo: WARNING: could not open /usr/include/c++/9.1.0/bits/move.h
geninfo: WARNING: could not open /usr/include/c++/9.1.0/bits/shared_ptr.h
geninfo: WARNING: could not open /usr/include/c++/9.1.0/bits/shared_ptr_base.h
geninfo: WARNING: could not open /usr/include/c++/9.1.0/bits/stl_algobase.h
geninfo: WARNING: could not open /usr/include/c++/9.1.0/bits/stl_deque.h
geninfo: WARNING: could not open /usr/include/c++/9.1.0/bits/stl_function.h
geninfo: WARNING: could not open /usr/include/c++/9.1.0/bits/stl_iterator.h
geninfo: WARNING: could not open /usr/include/c++/9.1.0/bits/stl_map.h
geninfo: WARNING: could not open /usr/include/c++/9.1.0/bits/stl_pair.h
geninfo: WARNING: could not open /usr/include/c++/9.1.0/bits/stl_tree.h
geninfo: WARNING: could not open /usr/include/c++/9.1.0/bits/stl_vector.h
geninfo: WARNING: could not open /usr/include/c++/9.1.0/bits/unique_ptr.h
geninfo: WARNING: could not open /usr/include/c++/9.1.0/ext/alloc_traits.h
geninfo: WARNING: could not open /usr/include/c++/9.1.0/ext/atomicity.h
geninfo: WARNING: could not open /usr/include/c++/9.1.0/ext/new_allocator.h
geninfo: WARNING: could not open /usr/include/c++/9.1.0/limits
geninfo: WARNING: could not open /usr/include/c++/9.1.0/new
geninfo: WARNING: could not open /usr/include/c++/9.1.0/system_error
geninfo: WARNING: could not open /usr/include/c++/9.1.0/typeinfo
geninfo: WARNING: could not open /usr/include/c++/9.1.0/x86_64-pc-linux-gnu/bits/gthr-default.h
geninfo: WARNING: some exclusion markers may be ignored
Processing src/crypto/MerkleTreeProof/libbitcoin_util_a-merkle-tree.gcno
geninfo: ERROR: /home/zcoin-builder/zcoin/src/crypto/MerkleTreeProof/libbitcoin_util_a-merkle-tree.gcno: reached unexpected end of file
make: *** [baseline.info] Error 255
rm baseline.info
```
## Desired behaviour
It should success without error.
## Steps to reproduce
Follow https://github.com/bitcoin/bitcoin/blob/master/doc/developer-notes.md#compiling-for-test-coverage
|
1.0
|
'make cov' is broken - ## Actual behaviour
```
zcoin-builder@ac9b1304543f:~/zcoin$ make cov
/usr/bin/lcov --gcov-tool=/usr/bin/gcov -c -i -d /home/zcoin-builder/zcoin/src -o baseline.info
Capturing coverage data from /home/zcoin-builder/zcoin/src
Found gcov version: 4.8.4
Scanning /home/zcoin-builder/zcoin/src for .gcno files ...
Found 509 graph files in /home/zcoin-builder/zcoin/src
Processing src/compat/libbitcoin_util_a-glibc_sanity.gcno
Processing src/compat/libbitcoin_util_a-glibcxx_sanity.gcno
Processing src/compat/libbitcoin_util_a-strnlen.gcno
Processing src/consensus/.libs/libbitcoinconsensus_la-merkle.gcno
Processing src/consensus/libbitcoinconsensus_la-merkle.gcno
Processing src/consensus/libbitcoin_consensus_a-merkle.gcno
Processing src/crypto/MerkleTreeProof/libbitcoin_util_a-mtp.gcno
geninfo: Unsuccessful stat on filename containing newline at /usr/bin/geninfo line 2554.
geninfo: Unsuccessful stat on filename containing newline at /usr/bin/geninfo line 2554.
geninfo: Unsuccessful stat on filename containing newline at /usr/bin/geninfo line 2554.
geninfo: Unsuccessful stat on filename containing newline at /usr/bin/geninfo line 2554.
geninfo: Unsuccessful stat on filename containing newline at /usr/bin/geninfo line 2554.
geninfo: Unsuccessful stat on filename containing newline at /usr/bin/geninfo line 2554.
geninfo: Unsuccessful stat on filename containing newline at /usr/bin/geninfo line 2554.
geninfo: WARNING: could not open
geninfo: WARNING: could not open /home/zcoin-builder/zcoin/src
geninfo: WARNING: could not open /home/zcoin-builder/zcoin/src/
geninfo: WARNING: could not open /home/zcoin-builder/zcoin/src/
geninfo: WARNING: could not open /home/zcoin-builder/zcoin/src/
geninfo: WARNING: could not open /usr/include/bits/uintn-identity.h
geninfo: WARNING: could not open /usr/include/boost/system/detail/generic_category.hpp
geninfo: WARNING: could not open /usr/include/boost/system/detail/std_interoperability.hpp
geninfo: WARNING: could not open /usr/include/boost/system/detail/system_category_posix.hpp
geninfo: WARNING: could not open /usr/include/c++/9.1.0/bits/allocator.h
geninfo: WARNING: could not open /usr/include/c++/9.1.0/bits/basic_string.h
geninfo: WARNING: could not open /usr/include/c++/9.1.0/bits/char_traits.h
geninfo: WARNING: could not open /usr/include/c++/9.1.0/bits/deque.tcc
geninfo: WARNING: could not open /usr/include/c++/9.1.0/bits/move.h
geninfo: WARNING: could not open /usr/include/c++/9.1.0/bits/shared_ptr.h
geninfo: WARNING: could not open /usr/include/c++/9.1.0/bits/shared_ptr_base.h
geninfo: WARNING: could not open /usr/include/c++/9.1.0/bits/stl_algobase.h
geninfo: WARNING: could not open /usr/include/c++/9.1.0/bits/stl_deque.h
geninfo: WARNING: could not open /usr/include/c++/9.1.0/bits/stl_function.h
geninfo: WARNING: could not open /usr/include/c++/9.1.0/bits/stl_iterator.h
geninfo: WARNING: could not open /usr/include/c++/9.1.0/bits/stl_map.h
geninfo: WARNING: could not open /usr/include/c++/9.1.0/bits/stl_pair.h
geninfo: WARNING: could not open /usr/include/c++/9.1.0/bits/stl_tree.h
geninfo: WARNING: could not open /usr/include/c++/9.1.0/bits/stl_vector.h
geninfo: WARNING: could not open /usr/include/c++/9.1.0/bits/unique_ptr.h
geninfo: WARNING: could not open /usr/include/c++/9.1.0/ext/alloc_traits.h
geninfo: WARNING: could not open /usr/include/c++/9.1.0/ext/atomicity.h
geninfo: WARNING: could not open /usr/include/c++/9.1.0/ext/new_allocator.h
geninfo: WARNING: could not open /usr/include/c++/9.1.0/limits
geninfo: WARNING: could not open /usr/include/c++/9.1.0/new
geninfo: WARNING: could not open /usr/include/c++/9.1.0/system_error
geninfo: WARNING: could not open /usr/include/c++/9.1.0/typeinfo
geninfo: WARNING: could not open /usr/include/c++/9.1.0/x86_64-pc-linux-gnu/bits/gthr-default.h
geninfo: WARNING: some exclusion markers may be ignored
Processing src/crypto/MerkleTreeProof/libbitcoin_util_a-merkle-tree.gcno
geninfo: ERROR: /home/zcoin-builder/zcoin/src/crypto/MerkleTreeProof/libbitcoin_util_a-merkle-tree.gcno: reached unexpected end of file
make: *** [baseline.info] Error 255
rm baseline.info
```
## Desired behaviour
It should success without error.
## Steps to reproduce
Follow https://github.com/bitcoin/bitcoin/blob/master/doc/developer-notes.md#compiling-for-test-coverage
|
non_process
|
make cov is broken actual behaviour zcoin builder zcoin make cov usr bin lcov gcov tool usr bin gcov c i d home zcoin builder zcoin src o baseline info capturing coverage data from home zcoin builder zcoin src found gcov version scanning home zcoin builder zcoin src for gcno files found graph files in home zcoin builder zcoin src processing src compat libbitcoin util a glibc sanity gcno processing src compat libbitcoin util a glibcxx sanity gcno processing src compat libbitcoin util a strnlen gcno processing src consensus libs libbitcoinconsensus la merkle gcno processing src consensus libbitcoinconsensus la merkle gcno processing src consensus libbitcoin consensus a merkle gcno processing src crypto merkletreeproof libbitcoin util a mtp gcno geninfo unsuccessful stat on filename containing newline at usr bin geninfo line geninfo unsuccessful stat on filename containing newline at usr bin geninfo line geninfo unsuccessful stat on filename containing newline at usr bin geninfo line geninfo unsuccessful stat on filename containing newline at usr bin geninfo line geninfo unsuccessful stat on filename containing newline at usr bin geninfo line geninfo unsuccessful stat on filename containing newline at usr bin geninfo line geninfo unsuccessful stat on filename containing newline at usr bin geninfo line geninfo warning could not open geninfo warning could not open home zcoin builder zcoin src geninfo warning could not open home zcoin builder zcoin src geninfo warning could not open home zcoin builder zcoin src geninfo warning could not open home zcoin builder zcoin src geninfo warning could not open usr include bits uintn identity h geninfo warning could not open usr include boost system detail generic category hpp geninfo warning could not open usr include boost system detail std interoperability hpp geninfo warning could not open usr include boost system detail system category posix hpp geninfo warning could not open usr include c bits allocator h geninfo warning could not open usr include c bits basic string h geninfo warning could not open usr include c bits char traits h geninfo warning could not open usr include c bits deque tcc geninfo warning could not open usr include c bits move h geninfo warning could not open usr include c bits shared ptr h geninfo warning could not open usr include c bits shared ptr base h geninfo warning could not open usr include c bits stl algobase h geninfo warning could not open usr include c bits stl deque h geninfo warning could not open usr include c bits stl function h geninfo warning could not open usr include c bits stl iterator h geninfo warning could not open usr include c bits stl map h geninfo warning could not open usr include c bits stl pair h geninfo warning could not open usr include c bits stl tree h geninfo warning could not open usr include c bits stl vector h geninfo warning could not open usr include c bits unique ptr h geninfo warning could not open usr include c ext alloc traits h geninfo warning could not open usr include c ext atomicity h geninfo warning could not open usr include c ext new allocator h geninfo warning could not open usr include c limits geninfo warning could not open usr include c new geninfo warning could not open usr include c system error geninfo warning could not open usr include c typeinfo geninfo warning could not open usr include c pc linux gnu bits gthr default h geninfo warning some exclusion markers may be ignored processing src crypto merkletreeproof libbitcoin util a merkle tree gcno geninfo error home zcoin builder zcoin src crypto merkletreeproof libbitcoin util a merkle tree gcno reached unexpected end of file make error rm baseline info desired behaviour it should success without error steps to reproduce follow
| 0
|
9,634
| 12,598,493,856
|
IssuesEvent
|
2020-06-11 03:04:18
|
googleapis/java-spanner
|
https://api.github.com/repos/googleapis/java-spanner
|
opened
|
SpannerRetryHelperTest.testExceptionWithRetryInfo test failure
|
type: process
|
In #251 (after moving to GitHub actions), we consistently get this test failure on Java 8 Windows.
```
[ERROR] Failures:
[ERROR] SpannerRetryHelperTest.testExceptionWithRetryInfo:218 expected to be true
[INFO]
[ERROR] Tests run: 3581, Failures: 1, Errors: 0, Skipped: 0
```
I will merge in #251 to cleanup the README but it would be good if we can investigate this test failure.
|
1.0
|
SpannerRetryHelperTest.testExceptionWithRetryInfo test failure - In #251 (after moving to GitHub actions), we consistently get this test failure on Java 8 Windows.
```
[ERROR] Failures:
[ERROR] SpannerRetryHelperTest.testExceptionWithRetryInfo:218 expected to be true
[INFO]
[ERROR] Tests run: 3581, Failures: 1, Errors: 0, Skipped: 0
```
I will merge in #251 to cleanup the README but it would be good if we can investigate this test failure.
|
process
|
spannerretryhelpertest testexceptionwithretryinfo test failure in after moving to github actions we consistently get this test failure on java windows failures spannerretryhelpertest testexceptionwithretryinfo expected to be true tests run failures errors skipped i will merge in to cleanup the readme but it would be good if we can investigate this test failure
| 1
|
285,343
| 21,514,745,822
|
IssuesEvent
|
2022-04-28 08:51:41
|
SciTools/iris
|
https://api.github.com/repos/SciTools/iris
|
closed
|
Iris API param not rendering
|
Type: Documentation
|
## 📚 Documentation
<!-- See https://scitools-iris.readthedocs.io/en/latest/ -->
<!-- Describe the issue or provide a suggestion for improving the Iris documentation -->
Whilst reviewing the dsocs for another task I noticed there are a few occurrences of `:param:` that is not rendering correctly due to alignment and maybe other syntax issues.
* https://scitools-iris.readthedocs.io/en/latest/generated/api/iris/fileformats/pp.html?highlight=%3Aparam%3A#iris.fileformats.pp.save
* https://scitools-iris.readthedocs.io/en/latest/generated/api/iris/fileformats/nimrod_load_rules.html?highlight=%3Aparam%3A#iris.fileformats.nimrod_load_rules.run
* https://scitools-iris.readthedocs.io/en/latest/generated/api/iris/fileformats/dot.html?highlight=%3Aparam%3A#iris.fileformats.dot.save
There maybe other occurrences too, not dug any further.
|
1.0
|
Iris API param not rendering - ## 📚 Documentation
<!-- See https://scitools-iris.readthedocs.io/en/latest/ -->
<!-- Describe the issue or provide a suggestion for improving the Iris documentation -->
Whilst reviewing the dsocs for another task I noticed there are a few occurrences of `:param:` that is not rendering correctly due to alignment and maybe other syntax issues.
* https://scitools-iris.readthedocs.io/en/latest/generated/api/iris/fileformats/pp.html?highlight=%3Aparam%3A#iris.fileformats.pp.save
* https://scitools-iris.readthedocs.io/en/latest/generated/api/iris/fileformats/nimrod_load_rules.html?highlight=%3Aparam%3A#iris.fileformats.nimrod_load_rules.run
* https://scitools-iris.readthedocs.io/en/latest/generated/api/iris/fileformats/dot.html?highlight=%3Aparam%3A#iris.fileformats.dot.save
There maybe other occurrences too, not dug any further.
|
non_process
|
iris api param not rendering 📚 documentation whilst reviewing the dsocs for another task i noticed there are a few occurrences of param that is not rendering correctly due to alignment and maybe other syntax issues there maybe other occurrences too not dug any further
| 0
|
9,904
| 12,908,536,151
|
IssuesEvent
|
2020-07-15 07:37:02
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
Get-AutomationConnection does not support
|
Pri2 automation/svc cxp process-automation/subsvc product-question triaged
|
[Enter feedback here]
Get-AutomationConnection does not support in PS 7.0 Az model.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 8dcc44c4-9f64-6b0c-e045-0e5e3c567970
* Version Independent ID: 6456a6aa-bdeb-651c-0c83-a84d4541f735
* Content: [Create a PowerShell runbook in Azure Automation](https://docs.microsoft.com/en-us/azure/automation/learn/automation-tutorial-runbook-textual-powershell)
* Content Source: [articles/automation/learn/automation-tutorial-runbook-textual-powershell.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/learn/automation-tutorial-runbook-textual-powershell.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @MGoedtel
* Microsoft Alias: **magoedte**
|
1.0
|
Get-AutomationConnection does not support -
[Enter feedback here]
Get-AutomationConnection does not support in PS 7.0 Az model.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 8dcc44c4-9f64-6b0c-e045-0e5e3c567970
* Version Independent ID: 6456a6aa-bdeb-651c-0c83-a84d4541f735
* Content: [Create a PowerShell runbook in Azure Automation](https://docs.microsoft.com/en-us/azure/automation/learn/automation-tutorial-runbook-textual-powershell)
* Content Source: [articles/automation/learn/automation-tutorial-runbook-textual-powershell.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/learn/automation-tutorial-runbook-textual-powershell.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @MGoedtel
* Microsoft Alias: **magoedte**
|
process
|
get automationconnection does not support get automationconnection does not support in ps az model document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id bdeb content content source service automation sub service process automation github login mgoedtel microsoft alias magoedte
| 1
|
65,374
| 14,724,298,341
|
IssuesEvent
|
2021-01-06 02:15:23
|
jamestiotio/orientation2020
|
https://api.github.com/repos/jamestiotio/orientation2020
|
opened
|
WS-2020-0208 (Medium) detected in highlight.js-9.18.5.tgz
|
security vulnerability
|
## WS-2020-0208 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>highlight.js-9.18.5.tgz</b></p></summary>
<p>Syntax highlighting with language autodetection.</p>
<p>Library home page: <a href="https://registry.npmjs.org/highlight.js/-/highlight.js-9.18.5.tgz">https://registry.npmjs.org/highlight.js/-/highlight.js-9.18.5.tgz</a></p>
<p>Path to dependency file: orientation2020/package.json</p>
<p>Path to vulnerable library: orientation2020/node_modules/highlight.js/package.json</p>
<p>
Dependency Hierarchy:
- cli-service-4.4.4.tgz (Root Library)
- cli-highlight-2.1.4.tgz
- :x: **highlight.js-9.18.5.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/jamestiotio/orientation2020/commit/ada06bc81bd11faad921ad6cce71a2a1d31b339c">ada06bc81bd11faad921ad6cce71a2a1d31b339c</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
If are you are using Highlight.js to highlight user-provided data you are possibly vulnerable. On the client-side (in a browser or Electron environment) risks could include lengthy freezes or crashes... On the server-side infinite freezes could occur... effectively preventing users from accessing your app or service (ie, Denial of Service). This is an issue with grammars shipped with the parser (and potentially 3rd party grammars also), not the parser itself. If you are using Highlight.js with any of the following grammars you are vulnerable. If you are using highlightAuto to detect the language (and have any of these grammars registered) you are vulnerable.
<p>Publish Date: 2020-12-04
<p>URL: <a href=https://github.com/highlightjs/highlight.js/commit/373b9d862401162e832ce77305e49b859e110f9c>WS-2020-0208</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/highlightjs/highlight.js/tree/10.4.1">https://github.com/highlightjs/highlight.js/tree/10.4.1</a></p>
<p>Release Date: 2020-12-04</p>
<p>Fix Resolution: 10.4.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
WS-2020-0208 (Medium) detected in highlight.js-9.18.5.tgz - ## WS-2020-0208 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>highlight.js-9.18.5.tgz</b></p></summary>
<p>Syntax highlighting with language autodetection.</p>
<p>Library home page: <a href="https://registry.npmjs.org/highlight.js/-/highlight.js-9.18.5.tgz">https://registry.npmjs.org/highlight.js/-/highlight.js-9.18.5.tgz</a></p>
<p>Path to dependency file: orientation2020/package.json</p>
<p>Path to vulnerable library: orientation2020/node_modules/highlight.js/package.json</p>
<p>
Dependency Hierarchy:
- cli-service-4.4.4.tgz (Root Library)
- cli-highlight-2.1.4.tgz
- :x: **highlight.js-9.18.5.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/jamestiotio/orientation2020/commit/ada06bc81bd11faad921ad6cce71a2a1d31b339c">ada06bc81bd11faad921ad6cce71a2a1d31b339c</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
If are you are using Highlight.js to highlight user-provided data you are possibly vulnerable. On the client-side (in a browser or Electron environment) risks could include lengthy freezes or crashes... On the server-side infinite freezes could occur... effectively preventing users from accessing your app or service (ie, Denial of Service). This is an issue with grammars shipped with the parser (and potentially 3rd party grammars also), not the parser itself. If you are using Highlight.js with any of the following grammars you are vulnerable. If you are using highlightAuto to detect the language (and have any of these grammars registered) you are vulnerable.
<p>Publish Date: 2020-12-04
<p>URL: <a href=https://github.com/highlightjs/highlight.js/commit/373b9d862401162e832ce77305e49b859e110f9c>WS-2020-0208</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/highlightjs/highlight.js/tree/10.4.1">https://github.com/highlightjs/highlight.js/tree/10.4.1</a></p>
<p>Release Date: 2020-12-04</p>
<p>Fix Resolution: 10.4.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
ws medium detected in highlight js tgz ws medium severity vulnerability vulnerable library highlight js tgz syntax highlighting with language autodetection library home page a href path to dependency file package json path to vulnerable library node modules highlight js package json dependency hierarchy cli service tgz root library cli highlight tgz x highlight js tgz vulnerable library found in head commit a href found in base branch master vulnerability details if are you are using highlight js to highlight user provided data you are possibly vulnerable on the client side in a browser or electron environment risks could include lengthy freezes or crashes on the server side infinite freezes could occur effectively preventing users from accessing your app or service ie denial of service this is an issue with grammars shipped with the parser and potentially party grammars also not the parser itself if you are using highlight js with any of the following grammars you are vulnerable if you are using highlightauto to detect the language and have any of these grammars registered you are vulnerable publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
22,615
| 31,841,521,961
|
IssuesEvent
|
2023-09-14 16:39:39
|
darktable-org/darktable
|
https://api.github.com/repos/darktable-org/darktable
|
closed
|
Rotation and perspective : crop not properly taken into account
|
priority: high reproduce: confirmed scope: image processing bug: pending
|
With the manual guides (manually draw - pencil - or manually define - square guide) and using the "largest area" mode the crop is not properly done with leaving the module:
All ok when doing the fit action:

No crop when collapsing module:

@jenshannoschwalm : Probably due to your latest changes handling the overlay & crop.
|
1.0
|
Rotation and perspective : crop not properly taken into account - With the manual guides (manually draw - pencil - or manually define - square guide) and using the "largest area" mode the crop is not properly done with leaving the module:
All ok when doing the fit action:

No crop when collapsing module:

@jenshannoschwalm : Probably due to your latest changes handling the overlay & crop.
|
process
|
rotation and perspective crop not properly taken into account with the manual guides manually draw pencil or manually define square guide and using the largest area mode the crop is not properly done with leaving the module all ok when doing the fit action no crop when collapsing module jenshannoschwalm probably due to your latest changes handling the overlay crop
| 1
|
11,613
| 14,479,138,761
|
IssuesEvent
|
2020-12-10 09:24:34
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
SecretValueText has been removed
|
automation/svc cxp doc-enhancement process-automation/subsvc triaged
|
SecretValueText has been removed from PSKeyVaultSecret as of Az 5.0.0. I had to replace
$secret = (Get-AzKeyVaultSecret -VaultName $VaultName -Name "SendGridAPIKey").SecretValueText
with
$secret = (Get-AzKeyVaultSecret -VaultName $VaultName -Name "SendGridAPIKey").SecretValue
$ssPtr = [System.Runtime.InteropServices.Marshal]::SecureStringToBSTR($secret)
try {
$SENDGRID_API_KEY = [System.Runtime.InteropServices.Marshal]::PtrToStringBSTR($ssPtr)
} finally {
[System.Runtime.InteropServices.Marshal]::ZeroFreeBSTR($ssPtr)
}
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: f3d18144-dd03-c659-f0a4-c0ab13d3d343
* Version Independent ID: ecdcbda2-259f-4992-57d8-c448a3934011
* Content: [Send an email from an Azure Automation runbook](https://docs.microsoft.com/en-us/azure/automation/automation-send-email)
* Content Source: [articles/automation/automation-send-email.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/automation-send-email.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @MGoedtel
* Microsoft Alias: **magoedte**
|
1.0
|
SecretValueText has been removed -
SecretValueText has been removed from PSKeyVaultSecret as of Az 5.0.0. I had to replace
$secret = (Get-AzKeyVaultSecret -VaultName $VaultName -Name "SendGridAPIKey").SecretValueText
with
$secret = (Get-AzKeyVaultSecret -VaultName $VaultName -Name "SendGridAPIKey").SecretValue
$ssPtr = [System.Runtime.InteropServices.Marshal]::SecureStringToBSTR($secret)
try {
$SENDGRID_API_KEY = [System.Runtime.InteropServices.Marshal]::PtrToStringBSTR($ssPtr)
} finally {
[System.Runtime.InteropServices.Marshal]::ZeroFreeBSTR($ssPtr)
}
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: f3d18144-dd03-c659-f0a4-c0ab13d3d343
* Version Independent ID: ecdcbda2-259f-4992-57d8-c448a3934011
* Content: [Send an email from an Azure Automation runbook](https://docs.microsoft.com/en-us/azure/automation/automation-send-email)
* Content Source: [articles/automation/automation-send-email.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/automation-send-email.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @MGoedtel
* Microsoft Alias: **magoedte**
|
process
|
secretvaluetext has been removed secretvaluetext has been removed from pskeyvaultsecret as of az i had to replace secret get azkeyvaultsecret vaultname vaultname name sendgridapikey secretvaluetext with secret get azkeyvaultsecret vaultname vaultname name sendgridapikey secretvalue ssptr securestringtobstr secret try sendgrid api key ptrtostringbstr ssptr finally zerofreebstr ssptr document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login mgoedtel microsoft alias magoedte
| 1
|
246,393
| 7,895,189,336
|
IssuesEvent
|
2018-06-29 01:39:10
|
aowen87/BAR
|
https://api.github.com/repos/aowen87/BAR
|
closed
|
Viewer windows pop-up automatically
|
Likelihood: 3 - Occasional OS: Windows 7 Priority: Normal Severity: 3 - Major Irritation Support Group: Any bug version: 2.10.0
|
While in Jose Milovich's office he showed me how Viewer windows will automatically pop-up when using VisIt on Windows 7 (Enterprise, 64-bit). From my observation, it seem to happen when he used the scroll wheel on his mouse. He said it also happens on his Windows box at home (not sure if he uses a mouse with the scroll wheel).
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. The following information
could not be accurately captured in the new ticket:
Original author: Kevin Griffin
Original creation: 04/11/2016 07:19 pm
Original update: 06/07/2016 08:13 pm
Ticket number: 2593
|
1.0
|
Viewer windows pop-up automatically - While in Jose Milovich's office he showed me how Viewer windows will automatically pop-up when using VisIt on Windows 7 (Enterprise, 64-bit). From my observation, it seem to happen when he used the scroll wheel on his mouse. He said it also happens on his Windows box at home (not sure if he uses a mouse with the scroll wheel).
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. The following information
could not be accurately captured in the new ticket:
Original author: Kevin Griffin
Original creation: 04/11/2016 07:19 pm
Original update: 06/07/2016 08:13 pm
Ticket number: 2593
|
non_process
|
viewer windows pop up automatically while in jose milovich s office he showed me how viewer windows will automatically pop up when using visit on windows enterprise bit from my observation it seem to happen when he used the scroll wheel on his mouse he said it also happens on his windows box at home not sure if he uses a mouse with the scroll wheel redmine migration this ticket was migrated from redmine the following information could not be accurately captured in the new ticket original author kevin griffin original creation pm original update pm ticket number
| 0
|
15,689
| 3,331,669,171
|
IssuesEvent
|
2015-11-11 16:45:55
|
owncloud/core
|
https://api.github.com/repos/owncloud/core
|
closed
|
Use product name in "save to ownCloud" button
|
design enhancement needs info sev4-low sharing
|
currently when a theme is used that changes the product name the button to save a public share still says "ownCloud"
|
1.0
|
Use product name in "save to ownCloud" button - currently when a theme is used that changes the product name the button to save a public share still says "ownCloud"
|
non_process
|
use product name in save to owncloud button currently when a theme is used that changes the product name the button to save a public share still says owncloud
| 0
|
10,317
| 13,159,912,372
|
IssuesEvent
|
2020-08-10 16:38:23
|
cncf/cnf-conformance
|
https://api.github.com/repos/cncf/cnf-conformance
|
closed
|
[Process] Add "good first issues" to issues and documentation
|
2 pts process sprint13
|
### [Process] Add "good first issues" to issues and documentation
Tasks:
- [x] Brainstorm on good first issues for a new contributor
- [x] Triage first set of issues
- [x] add new issues to Github
- [x] add "good first issues" label to issues
- [x] Add comment suggesting updates as needed for:
- [x] the [README](https://github.com/cncf/cnf-conformance/blob/master/README.md)
- [x] the [contributing guide](https://github.com/cncf/cnf-conformance/blob/master/CONTRIBUTING.md)
- [x] Tag 1 or more people to peer review
|
1.0
|
[Process] Add "good first issues" to issues and documentation - ### [Process] Add "good first issues" to issues and documentation
Tasks:
- [x] Brainstorm on good first issues for a new contributor
- [x] Triage first set of issues
- [x] add new issues to Github
- [x] add "good first issues" label to issues
- [x] Add comment suggesting updates as needed for:
- [x] the [README](https://github.com/cncf/cnf-conformance/blob/master/README.md)
- [x] the [contributing guide](https://github.com/cncf/cnf-conformance/blob/master/CONTRIBUTING.md)
- [x] Tag 1 or more people to peer review
|
process
|
add good first issues to issues and documentation add good first issues to issues and documentation tasks brainstorm on good first issues for a new contributor triage first set of issues add new issues to github add good first issues label to issues add comment suggesting updates as needed for the the tag or more people to peer review
| 1
|
19,764
| 26,139,113,375
|
IssuesEvent
|
2022-12-29 15:52:33
|
MicrosoftDocs/windows-dev-docs
|
https://api.github.com/repos/MicrosoftDocs/windows-dev-docs
|
closed
|
Is there a code somewhere
|
uwp/prod processes-and-threading/tech Pri2
|
Things on this page get confusing mostly with the names of each product. I'm not sure what this article seems to consider the service project to be. It is contained in an app to run in proc but I think the containing app is referred to as the service which it's not. Making it worse the article describes getting the package family name from the package manifest fie in the service project but the project type this article calls to create for the service doesn't have a manifest file. So I'm assuming the package family name you need is of the containing UWP project through which you'll make the call?
-
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: edde9dbc-6e04-69cf-206e-123792666abf
* Version Independent ID: 9894e78f-3270-9485-4769-11050669b805
* Content: [Create and consume an app service - UWP applications](https://docs.microsoft.com/en-us/windows/uwp/launch-resume/how-to-create-and-consume-an-app-service)
* Content Source: [windows-apps-src/launch-resume/how-to-create-and-consume-an-app-service.md](https://github.com/MicrosoftDocs/windows-uwp/blob/docs/windows-apps-src/launch-resume/how-to-create-and-consume-an-app-service.md)
* Product: **uwp**
* Technology: **processes-and-threading**
* GitHub Login: @alvinashcraft
* Microsoft Alias: **aashcraft**
|
1.0
|
Is there a code somewhere - Things on this page get confusing mostly with the names of each product. I'm not sure what this article seems to consider the service project to be. It is contained in an app to run in proc but I think the containing app is referred to as the service which it's not. Making it worse the article describes getting the package family name from the package manifest fie in the service project but the project type this article calls to create for the service doesn't have a manifest file. So I'm assuming the package family name you need is of the containing UWP project through which you'll make the call?
-
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: edde9dbc-6e04-69cf-206e-123792666abf
* Version Independent ID: 9894e78f-3270-9485-4769-11050669b805
* Content: [Create and consume an app service - UWP applications](https://docs.microsoft.com/en-us/windows/uwp/launch-resume/how-to-create-and-consume-an-app-service)
* Content Source: [windows-apps-src/launch-resume/how-to-create-and-consume-an-app-service.md](https://github.com/MicrosoftDocs/windows-uwp/blob/docs/windows-apps-src/launch-resume/how-to-create-and-consume-an-app-service.md)
* Product: **uwp**
* Technology: **processes-and-threading**
* GitHub Login: @alvinashcraft
* Microsoft Alias: **aashcraft**
|
process
|
is there a code somewhere things on this page get confusing mostly with the names of each product i m not sure what this article seems to consider the service project to be it is contained in an app to run in proc but i think the containing app is referred to as the service which it s not making it worse the article describes getting the package family name from the package manifest fie in the service project but the project type this article calls to create for the service doesn t have a manifest file so i m assuming the package family name you need is of the containing uwp project through which you ll make the call document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product uwp technology processes and threading github login alvinashcraft microsoft alias aashcraft
| 1
|
9,262
| 12,294,715,001
|
IssuesEvent
|
2020-05-11 01:15:56
|
allinurl/goaccess
|
https://api.github.com/repos/allinurl/goaccess
|
closed
|
On-disk storage crash in GoAccess
|
bug duplicate log-processing on-disk waiting reply
|
```
[root@sites ~]# goaccess -f /var/log/httpd/access_log -o /Sites/usdsites/bryan/report.html --real-time-html --log-format=COMBINED --ws-url=sites-mgmt.sandiego.edu --keep-db-files --load-from-disk
Unable to open fifo out: No such device or address.
Accepted: 168 10.80.20.81
Active: 1
CLOSE
Active: 0
Accepted: 168 10.80.20.81
Active: 1
==28326== GoAccess 1.0.2 crashed by Sig 11
==28326==
==28326== VALUES AT CRASH POINT
==28326==
==28326== Line number: 6591016
==28326== Offset: 0
==28326== Invalid data: 110691
==28326== Piping: 0
==28326== Response size: 141502306226 bytes
==28326==
==28326== STACK TRACE:
==28326==
==28326== 0 goaccess(sigsegv_handler+0x17a) [0x40a6ea]
==28326== 1 /lib64/libc.so.6() [0x3a62c32660]
==28326== 2 /lib64/libc.so.6(memcpy+0x286) [0x3a62c89a86]
==28326== 3 /usr/lib64/libtokyocabinet.so.8() [0x3a6484032c]
==28326== 4 /usr/lib64/libtokyocabinet.so.8(tcbdbput+0xbf) [0x3a6484409f]
==28326== 5 /usr/lib64/libtokyocabinet.so.8(tcadbput+0x1f4) [0x3a64869024]
==28326== 6 goaccess() [0x426f62]
==28326== 7 goaccess() [0x418be2]
==28326== 8 goaccess(parse_log+0x14a) [0x41901a]
==28326== 9 goaccess(main+0x1f3) [0x40ff33]
==28326== 10 /lib64/libc.so.6(__libc_start_main+0xfd) [0x3a62c1ed1d]
==28326== 11 goaccess() [0x407cf9]
==28326==
==28326== Please report it by opening an issue on GitHub:
==28326== https://github.com/allinurl/goaccess/issues
```
Not sure what else you will need.
Thanks
Bryan
|
1.0
|
On-disk storage crash in GoAccess - ```
[root@sites ~]# goaccess -f /var/log/httpd/access_log -o /Sites/usdsites/bryan/report.html --real-time-html --log-format=COMBINED --ws-url=sites-mgmt.sandiego.edu --keep-db-files --load-from-disk
Unable to open fifo out: No such device or address.
Accepted: 168 10.80.20.81
Active: 1
CLOSE
Active: 0
Accepted: 168 10.80.20.81
Active: 1
==28326== GoAccess 1.0.2 crashed by Sig 11
==28326==
==28326== VALUES AT CRASH POINT
==28326==
==28326== Line number: 6591016
==28326== Offset: 0
==28326== Invalid data: 110691
==28326== Piping: 0
==28326== Response size: 141502306226 bytes
==28326==
==28326== STACK TRACE:
==28326==
==28326== 0 goaccess(sigsegv_handler+0x17a) [0x40a6ea]
==28326== 1 /lib64/libc.so.6() [0x3a62c32660]
==28326== 2 /lib64/libc.so.6(memcpy+0x286) [0x3a62c89a86]
==28326== 3 /usr/lib64/libtokyocabinet.so.8() [0x3a6484032c]
==28326== 4 /usr/lib64/libtokyocabinet.so.8(tcbdbput+0xbf) [0x3a6484409f]
==28326== 5 /usr/lib64/libtokyocabinet.so.8(tcadbput+0x1f4) [0x3a64869024]
==28326== 6 goaccess() [0x426f62]
==28326== 7 goaccess() [0x418be2]
==28326== 8 goaccess(parse_log+0x14a) [0x41901a]
==28326== 9 goaccess(main+0x1f3) [0x40ff33]
==28326== 10 /lib64/libc.so.6(__libc_start_main+0xfd) [0x3a62c1ed1d]
==28326== 11 goaccess() [0x407cf9]
==28326==
==28326== Please report it by opening an issue on GitHub:
==28326== https://github.com/allinurl/goaccess/issues
```
Not sure what else you will need.
Thanks
Bryan
|
process
|
on disk storage crash in goaccess goaccess f var log httpd access log o sites usdsites bryan report html real time html log format combined ws url sites mgmt sandiego edu keep db files load from disk unable to open fifo out no such device or address accepted active close active accepted active goaccess crashed by sig values at crash point line number offset invalid data piping response size bytes stack trace goaccess sigsegv handler libc so libc so memcpy usr libtokyocabinet so usr libtokyocabinet so tcbdbput usr libtokyocabinet so tcadbput goaccess goaccess goaccess parse log goaccess main libc so libc start main goaccess please report it by opening an issue on github not sure what else you will need thanks bryan
| 1
|
17,568
| 2,615,147,331
|
IssuesEvent
|
2015-03-01 06:23:34
|
chrsmith/html5rocks
|
https://api.github.com/repos/chrsmith/html5rocks
|
closed
|
Including lovelywebapp on the front page
|
auto-migrated lovelywebapp Milestone-6 MovedFrom4 Priority-P1 Type-Feature
|
```
What steps will reproduce the problem?
1.
2.
3.
What is the expected output? What do you see instead?
Please use labels and text to provide additional information.
```
Original issue reported on code.google.com by `v...@google.com` on 24 Sep 2010 at 5:29
|
1.0
|
Including lovelywebapp on the front page - ```
What steps will reproduce the problem?
1.
2.
3.
What is the expected output? What do you see instead?
Please use labels and text to provide additional information.
```
Original issue reported on code.google.com by `v...@google.com` on 24 Sep 2010 at 5:29
|
non_process
|
including lovelywebapp on the front page what steps will reproduce the problem what is the expected output what do you see instead please use labels and text to provide additional information original issue reported on code google com by v google com on sep at
| 0
|
16,516
| 21,527,181,569
|
IssuesEvent
|
2022-04-28 19:43:17
|
oasis-tcs/csaf
|
https://api.github.com/repos/oasis-tcs/csaf
|
opened
|
Register `signature` and `hash` as link rel types for ROLIE
|
csaf 2.0 external oasis_tc_process
|
In the ROLIE feed, we suggest to explicitly mention the signature files and hash files. We should register those link rel types.
|
1.0
|
Register `signature` and `hash` as link rel types for ROLIE - In the ROLIE feed, we suggest to explicitly mention the signature files and hash files. We should register those link rel types.
|
process
|
register signature and hash as link rel types for rolie in the rolie feed we suggest to explicitly mention the signature files and hash files we should register those link rel types
| 1
|
20,195
| 26,768,721,412
|
IssuesEvent
|
2023-01-31 12:37:14
|
corona-warn-app/cwa-wishlist
|
https://api.github.com/repos/corona-warn-app/cwa-wishlist
|
closed
|
Allow pasting of TeleTAN
|
enhancement mirrored-to-jira Test/Share process
|
## Current implementation
When entering a Tele Tan it is not possible to past a TAN copied from somewhere.
## Suggested Enhancement
Use prompt to paste copied TeleTAN into entry field.
## Expected Benefits
This will reduce human error and increase convenience.
---
Internal Tracking ID: EXPOSUREAPP-4684
|
1.0
|
Allow pasting of TeleTAN - ## Current implementation
When entering a Tele Tan it is not possible to past a TAN copied from somewhere.
## Suggested Enhancement
Use prompt to paste copied TeleTAN into entry field.
## Expected Benefits
This will reduce human error and increase convenience.
---
Internal Tracking ID: EXPOSUREAPP-4684
|
process
|
allow pasting of teletan current implementation when entering a tele tan it is not possible to past a tan copied from somewhere suggested enhancement use prompt to paste copied teletan into entry field expected benefits this will reduce human error and increase convenience internal tracking id exposureapp
| 1
|
2,742
| 5,635,499,708
|
IssuesEvent
|
2017-04-06 00:58:48
|
metabase/metabase
|
https://api.github.com/repos/metabase/metabase
|
closed
|
Named aggregations in BigQuery can't have spaces/special characters
|
Bug Database/BigQuery Limitation Priority/P1 Query Processor
|
Trying to name an aggregation "hello world" results in this error:
> Invalid field name "hello world". Fields must contain only letters, numbers, and underscores, start with a letter or underscore, and be at most 128 characters long.
|
1.0
|
Named aggregations in BigQuery can't have spaces/special characters - Trying to name an aggregation "hello world" results in this error:
> Invalid field name "hello world". Fields must contain only letters, numbers, and underscores, start with a letter or underscore, and be at most 128 characters long.
|
process
|
named aggregations in bigquery can t have spaces special characters trying to name an aggregation hello world results in this error invalid field name hello world fields must contain only letters numbers and underscores start with a letter or underscore and be at most characters long
| 1
|
2,347
| 5,154,920,832
|
IssuesEvent
|
2017-01-15 05:46:21
|
rubberduck-vba/Rubberduck
|
https://api.github.com/repos/rubberduck-vba/Rubberduck
|
closed
|
Undeclared variables don't have annotations
|
bug parse-tree-processing
|
Undeclared variables are "declared" on-the-spot in the resolver's "identifier references" pass, so that when an identifier can't be resolved it's assumed to be a local variable, and so subsequent references to that identifier in the procedure scope will resolve to that newly created `Declaration`.
As a result, this annotation has no effect whatsoever:
````vb
Sub Test()
'@Ignore UndeclaredVariable
a = 42
MsgBox a
End Sub
````
Because the `a` local variable declaration gets created *after* the annotations are associated to (actual, in-code) declarations, the on-the-fly declaration doesn't see it:

As a result, an undeclared variable can't be ignored. And.. well, that's a problem, right?
|
1.0
|
Undeclared variables don't have annotations - Undeclared variables are "declared" on-the-spot in the resolver's "identifier references" pass, so that when an identifier can't be resolved it's assumed to be a local variable, and so subsequent references to that identifier in the procedure scope will resolve to that newly created `Declaration`.
As a result, this annotation has no effect whatsoever:
````vb
Sub Test()
'@Ignore UndeclaredVariable
a = 42
MsgBox a
End Sub
````
Because the `a` local variable declaration gets created *after* the annotations are associated to (actual, in-code) declarations, the on-the-fly declaration doesn't see it:

As a result, an undeclared variable can't be ignored. And.. well, that's a problem, right?
|
process
|
undeclared variables don t have annotations undeclared variables are declared on the spot in the resolver s identifier references pass so that when an identifier can t be resolved it s assumed to be a local variable and so subsequent references to that identifier in the procedure scope will resolve to that newly created declaration as a result this annotation has no effect whatsoever vb sub test ignore undeclaredvariable a msgbox a end sub because the a local variable declaration gets created after the annotations are associated to actual in code declarations the on the fly declaration doesn t see it as a result an undeclared variable can t be ignored and well that s a problem right
| 1
|
158,177
| 20,009,258,961
|
IssuesEvent
|
2022-02-01 02:55:48
|
snowdensb/sonar-xanitizer
|
https://api.github.com/repos/snowdensb/sonar-xanitizer
|
opened
|
CVE-2022-23437 (High) detected in xercesImpl-2.11.0.jar
|
security vulnerability
|
## CVE-2022-23437 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xercesImpl-2.11.0.jar</b></p></summary>
<p></p>
<p>Library home page: <a href="http://archive.apache.org/dist/drill/drill-1.17.0/apache-drill-1.17.0.tar.gz">http://archive.apache.org/dist/drill/drill-1.17.0/apache-drill-1.17.0.tar.gz</a></p>
<p>Path to vulnerable library: /src/test/resources/webgoat/WEB-INF/lib/xercesImpl-2.11.0.jar</p>
<p>
Dependency Hierarchy:
- :x: **xercesImpl-2.11.0.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
There's a vulnerability within the Apache Xerces Java (XercesJ) XML parser when handling specially crafted XML document payloads. This causes, the XercesJ XML parser to wait in an infinite loop, which may sometimes consume system resources for prolonged duration. This vulnerability is present within XercesJ version 2.12.1 and the previous versions.
<p>Publish Date: 2022-01-24
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-23437>CVE-2022-23437</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-h65f-jvqw-m9fj">https://github.com/advisories/GHSA-h65f-jvqw-m9fj</a></p>
<p>Release Date: 2022-01-24</p>
<p>Fix Resolution: xerces:xercesImpl:2.12.2</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java Archive","packageName":"xercesImpl-2.11","packageVersion":"2.11.0","packageFilePaths":[null],"isTransitiveDependency":false,"dependencyTree":"xercesImpl-2.11:2.11.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"xerces:xercesImpl:2.12.2","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2022-23437","vulnerabilityDetails":"There\u0027s a vulnerability within the Apache Xerces Java (XercesJ) XML parser when handling specially crafted XML document payloads. This causes, the XercesJ XML parser to wait in an infinite loop, which may sometimes consume system resources for prolonged duration. This vulnerability is present within XercesJ version 2.12.1 and the previous versions.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-23437","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2022-23437 (High) detected in xercesImpl-2.11.0.jar - ## CVE-2022-23437 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xercesImpl-2.11.0.jar</b></p></summary>
<p></p>
<p>Library home page: <a href="http://archive.apache.org/dist/drill/drill-1.17.0/apache-drill-1.17.0.tar.gz">http://archive.apache.org/dist/drill/drill-1.17.0/apache-drill-1.17.0.tar.gz</a></p>
<p>Path to vulnerable library: /src/test/resources/webgoat/WEB-INF/lib/xercesImpl-2.11.0.jar</p>
<p>
Dependency Hierarchy:
- :x: **xercesImpl-2.11.0.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
There's a vulnerability within the Apache Xerces Java (XercesJ) XML parser when handling specially crafted XML document payloads. This causes, the XercesJ XML parser to wait in an infinite loop, which may sometimes consume system resources for prolonged duration. This vulnerability is present within XercesJ version 2.12.1 and the previous versions.
<p>Publish Date: 2022-01-24
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-23437>CVE-2022-23437</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-h65f-jvqw-m9fj">https://github.com/advisories/GHSA-h65f-jvqw-m9fj</a></p>
<p>Release Date: 2022-01-24</p>
<p>Fix Resolution: xerces:xercesImpl:2.12.2</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java Archive","packageName":"xercesImpl-2.11","packageVersion":"2.11.0","packageFilePaths":[null],"isTransitiveDependency":false,"dependencyTree":"xercesImpl-2.11:2.11.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"xerces:xercesImpl:2.12.2","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2022-23437","vulnerabilityDetails":"There\u0027s a vulnerability within the Apache Xerces Java (XercesJ) XML parser when handling specially crafted XML document payloads. This causes, the XercesJ XML parser to wait in an infinite loop, which may sometimes consume system resources for prolonged duration. This vulnerability is present within XercesJ version 2.12.1 and the previous versions.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-23437","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
non_process
|
cve high detected in xercesimpl jar cve high severity vulnerability vulnerable library xercesimpl jar library home page a href path to vulnerable library src test resources webgoat web inf lib xercesimpl jar dependency hierarchy x xercesimpl jar vulnerable library found in base branch master vulnerability details there s a vulnerability within the apache xerces java xercesj xml parser when handling specially crafted xml document payloads this causes the xercesj xml parser to wait in an infinite loop which may sometimes consume system resources for prolonged duration this vulnerability is present within xercesj version and the previous versions publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution xerces xercesimpl isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree xercesimpl isminimumfixversionavailable true minimumfixversion xerces xercesimpl isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails there a vulnerability within the apache xerces java xercesj xml parser when handling specially crafted xml document payloads this causes the xercesj xml parser to wait in an infinite loop which may sometimes consume system resources for prolonged duration this vulnerability is present within xercesj version and the previous versions vulnerabilityurl
| 0
|
263,739
| 19,974,672,078
|
IssuesEvent
|
2022-01-29 00:15:47
|
carbonplan/cmip6-downscaling
|
https://api.github.com/repos/carbonplan/cmip6-downscaling
|
opened
|
Use tutorial: How to override configuration variables
|
documentation
|
### Intro
The `cmip6_downscaling` python package uses the [donfig](https://donfig.readthedocs.io/en/latest/) package for configuration. The config schema as well as defaults are laid out in [config.py](https://github.com/carbonplan/cmip6-downscaling/blob/main/cmip6_downscaling/config.py). These defaults can be overwritten by [specifying a yaml file](https://donfig.readthedocs.io/en/latest/configuration.html#specify-configuration) setting [environment variables](https://donfig.readthedocs.io/en/latest/configuration.html#environment-variables) or using the [`config.set()` method](https://donfig.readthedocs.io/en/latest/configuration.html#directly-within-python).
#### YAML configuration
Donfig allows yaml configuration files in three locations:
>1. The ~/.config/mypkg directory in the user’s home directory
>2. The {sys.prefix}/etc/mypkg directory local to Python
>3. The root directory (specified by the MYPKG_ROOT_CONFIG environment variable or /etc/mypkg/ by default)
In this example, we are using option #3.
Steps:
1. Create the directory ~/.config/cmip6_downscaling/{choose_a_name}.yaml (ex. cloud.yaml)
2. Populate the .yaml file. Note, the schema of the yaml file must reflect the schema of the `config.py` file.
In this case, we are modifying the default values for `storage.intermediate.uri`
```yaml
storage:
intermediate:
uri: "az://flow-outputs/intermediate"
```
|
1.0
|
Use tutorial: How to override configuration variables - ### Intro
The `cmip6_downscaling` python package uses the [donfig](https://donfig.readthedocs.io/en/latest/) package for configuration. The config schema as well as defaults are laid out in [config.py](https://github.com/carbonplan/cmip6-downscaling/blob/main/cmip6_downscaling/config.py). These defaults can be overwritten by [specifying a yaml file](https://donfig.readthedocs.io/en/latest/configuration.html#specify-configuration) setting [environment variables](https://donfig.readthedocs.io/en/latest/configuration.html#environment-variables) or using the [`config.set()` method](https://donfig.readthedocs.io/en/latest/configuration.html#directly-within-python).
#### YAML configuration
Donfig allows yaml configuration files in three locations:
>1. The ~/.config/mypkg directory in the user’s home directory
>2. The {sys.prefix}/etc/mypkg directory local to Python
>3. The root directory (specified by the MYPKG_ROOT_CONFIG environment variable or /etc/mypkg/ by default)
In this example, we are using option #3.
Steps:
1. Create the directory ~/.config/cmip6_downscaling/{choose_a_name}.yaml (ex. cloud.yaml)
2. Populate the .yaml file. Note, the schema of the yaml file must reflect the schema of the `config.py` file.
In this case, we are modifying the default values for `storage.intermediate.uri`
```yaml
storage:
intermediate:
uri: "az://flow-outputs/intermediate"
```
|
non_process
|
use tutorial how to override configuration variables intro the downscaling python package uses the package for configuration the config schema as well as defaults are laid out in these defaults can be overwritten by setting or using the yaml configuration donfig allows yaml configuration files in three locations the config mypkg directory in the user’s home directory the sys prefix etc mypkg directory local to python the root directory specified by the mypkg root config environment variable or etc mypkg by default in this example we are using option steps create the directory config downscaling choose a name yaml ex cloud yaml populate the yaml file note the schema of the yaml file must reflect the schema of the config py file in this case we are modifying the default values for storage intermediate uri yaml storage intermediate uri az flow outputs intermediate
| 0
|
181,493
| 14,045,106,434
|
IssuesEvent
|
2020-11-02 00:14:07
|
Elle624/FitLit-Refactor
|
https://api.github.com/repos/Elle624/FitLit-Refactor
|
closed
|
Add Chai Spy for domUpdate
|
Priority 1 js testing
|
As a developer, I should be able to test if functions are being called in domUpdate object
|
1.0
|
Add Chai Spy for domUpdate - As a developer, I should be able to test if functions are being called in domUpdate object
|
non_process
|
add chai spy for domupdate as a developer i should be able to test if functions are being called in domupdate object
| 0
|
344,561
| 24,818,839,568
|
IssuesEvent
|
2022-10-25 14:57:43
|
joshuadavidthomas/dotfiles
|
https://api.github.com/repos/joshuadavidthomas/dotfiles
|
opened
|
Add better documentation about what the installation script does
|
documentation
|
Should include applications installed
|
1.0
|
Add better documentation about what the installation script does - Should include applications installed
|
non_process
|
add better documentation about what the installation script does should include applications installed
| 0
|
247,522
| 20,984,019,236
|
IssuesEvent
|
2022-03-28 23:42:28
|
rstudio/rstudio
|
https://api.github.com/repos/rstudio/rstudio
|
closed
|
git diff (Review Changes) window fails to load
|
bug electron test
|
For Electron. It looks like it's trying to query `window.desktopInfo`, but that object isn't populated in windows other than the main window.
|
1.0
|
git diff (Review Changes) window fails to load - For Electron. It looks like it's trying to query `window.desktopInfo`, but that object isn't populated in windows other than the main window.
|
non_process
|
git diff review changes window fails to load for electron it looks like it s trying to query window desktopinfo but that object isn t populated in windows other than the main window
| 0
|
805,232
| 29,513,207,081
|
IssuesEvent
|
2023-06-04 07:04:50
|
slynch8/10x
|
https://api.github.com/repos/slynch8/10x
|
opened
|
updating subscription requires a restart
|
bug Priority 2
|
Potential minor bug if you designed 10x to pick up new subscription details right after updating them without the need to restart 10x:
My first cancelled subscription expired today and after starting 10x I got a window where I could put new subscription details.
I put the new subscription details and clicked ok.
After that there was another window saying subscription is invalid, it had single button "update subscription".
Clicking the button opened browser to paddle subscription management webpage of the old cancelled subscription(see screenshot). All I could do is to close the window, but after starting 10x again all is fine.
|
1.0
|
updating subscription requires a restart - Potential minor bug if you designed 10x to pick up new subscription details right after updating them without the need to restart 10x:
My first cancelled subscription expired today and after starting 10x I got a window where I could put new subscription details.
I put the new subscription details and clicked ok.
After that there was another window saying subscription is invalid, it had single button "update subscription".
Clicking the button opened browser to paddle subscription management webpage of the old cancelled subscription(see screenshot). All I could do is to close the window, but after starting 10x again all is fine.
|
non_process
|
updating subscription requires a restart potential minor bug if you designed to pick up new subscription details right after updating them without the need to restart my first cancelled subscription expired today and after starting i got a window where i could put new subscription details i put the new subscription details and clicked ok after that there was another window saying subscription is invalid it had single button update subscription clicking the button opened browser to paddle subscription management webpage of the old cancelled subscription see screenshot all i could do is to close the window but after starting again all is fine
| 0
|
21,649
| 30,083,638,391
|
IssuesEvent
|
2023-06-29 06:51:23
|
geneontology/go-ontology
|
https://api.github.com/repos/geneontology/go-ontology
|
closed
|
GO:0039644 'suppression by virus of host NF-kappaB cascade' missing parent
|
multi-species process
|
GO:0039644 'suppression by virus of host NF-kappaB cascade'
is not under modulation/perturbation of host signaling
<img width="780" alt="image" src="https://github.com/geneontology/go-ontology/assets/4782928/8e134ee0-1969-4964-8b7c-1f6d1761e7a2">
|
1.0
|
GO:0039644 'suppression by virus of host NF-kappaB cascade' missing parent - GO:0039644 'suppression by virus of host NF-kappaB cascade'
is not under modulation/perturbation of host signaling
<img width="780" alt="image" src="https://github.com/geneontology/go-ontology/assets/4782928/8e134ee0-1969-4964-8b7c-1f6d1761e7a2">
|
process
|
go suppression by virus of host nf kappab cascade missing parent go suppression by virus of host nf kappab cascade is not under modulation perturbation of host signaling img width alt image src
| 1
|
39,840
| 5,252,293,475
|
IssuesEvent
|
2017-02-02 03:39:34
|
semperfiwebdesign/all-in-one-seo-pack
|
https://api.github.com/repos/semperfiwebdesign/all-in-one-seo-pack
|
closed
|
Social Meta Module - Upload image is not working and is causing an error in console
|
Bug Needs Testing Priority - High
|
Reported here - https://wordpress.org/support/topic/upload-image-is-not-working/
AIOSEOP->Social Meta->Upload image is not working and is causing an error in console:
aioseop_module.js?ver=2.3.11.3:27 Uncaught TypeError: Cannot read property 'frames' of undefined
at HTMLInputElement.<anonymous> (aioseop_module.js?ver=2.3.11.3:27)
at HTMLInputElement.dispatch (load-scripts.php?c=0&load[]=jquery-core,jquery-migrate,utils,jquery-ui-core,jquery-ui-widget,jquery…:3)
at HTMLInputElement.r.handle (load-scripts.php?c=0&load[]=jquery-core,jquery-migrate,utils,jquery-ui-core,jquery-ui-widget,jquery…:3)
|
1.0
|
Social Meta Module - Upload image is not working and is causing an error in console - Reported here - https://wordpress.org/support/topic/upload-image-is-not-working/
AIOSEOP->Social Meta->Upload image is not working and is causing an error in console:
aioseop_module.js?ver=2.3.11.3:27 Uncaught TypeError: Cannot read property 'frames' of undefined
at HTMLInputElement.<anonymous> (aioseop_module.js?ver=2.3.11.3:27)
at HTMLInputElement.dispatch (load-scripts.php?c=0&load[]=jquery-core,jquery-migrate,utils,jquery-ui-core,jquery-ui-widget,jquery…:3)
at HTMLInputElement.r.handle (load-scripts.php?c=0&load[]=jquery-core,jquery-migrate,utils,jquery-ui-core,jquery-ui-widget,jquery…:3)
|
non_process
|
social meta module upload image is not working and is causing an error in console reported here aioseop social meta upload image is not working and is causing an error in console aioseop module js ver uncaught typeerror cannot read property frames of undefined at htmlinputelement aioseop module js ver at htmlinputelement dispatch load scripts php c load jquery core jquery migrate utils jquery ui core jquery ui widget jquery… at htmlinputelement r handle load scripts php c load jquery core jquery migrate utils jquery ui core jquery ui widget jquery…
| 0
|
5,898
| 8,712,869,239
|
IssuesEvent
|
2018-12-06 23:52:58
|
googleapis/nodejs-resource
|
https://api.github.com/repos/googleapis/nodejs-resource
|
closed
|
Resource is not a constructor error
|
type: process
|
Thanks for stopping by to let us know something could be better!
Please run down the following list and make sure you've tried the usual "quick
fixes":
- Search the issues already opened: https://github.com/googleapis/nodejs-resource/issues
- Search the issues on our "catch-all" repository: https://github.com/GoogleCloudPlatform/google-cloud-node
- Search StackOverflow: http://stackoverflow.com/questions/tagged/google-cloud-platform+node.js
If you are still having issues, please be sure to include as much information as
possible:
#### Environment details
- OS: OSX
- Node.js version: v10.10.0
- npm version: 6.4.1
- `@google-cloud/resource` version: 0.8.1
#### Steps to reproduce
1. Follow example documentation, receive error about "Resource is not a constructor.
2. https://github.com/googleapis/nodejs-resource/blob/master/samples/projects.js#L21
3 . The line above should read `const Resource = require('@google-cloud/resource');`. Notice the curly braces have been removed.
Following these steps will guarantee the quickest resolution possible.
Thanks!
|
1.0
|
Resource is not a constructor error - Thanks for stopping by to let us know something could be better!
Please run down the following list and make sure you've tried the usual "quick
fixes":
- Search the issues already opened: https://github.com/googleapis/nodejs-resource/issues
- Search the issues on our "catch-all" repository: https://github.com/GoogleCloudPlatform/google-cloud-node
- Search StackOverflow: http://stackoverflow.com/questions/tagged/google-cloud-platform+node.js
If you are still having issues, please be sure to include as much information as
possible:
#### Environment details
- OS: OSX
- Node.js version: v10.10.0
- npm version: 6.4.1
- `@google-cloud/resource` version: 0.8.1
#### Steps to reproduce
1. Follow example documentation, receive error about "Resource is not a constructor.
2. https://github.com/googleapis/nodejs-resource/blob/master/samples/projects.js#L21
3 . The line above should read `const Resource = require('@google-cloud/resource');`. Notice the curly braces have been removed.
Following these steps will guarantee the quickest resolution possible.
Thanks!
|
process
|
resource is not a constructor error thanks for stopping by to let us know something could be better please run down the following list and make sure you ve tried the usual quick fixes search the issues already opened search the issues on our catch all repository search stackoverflow if you are still having issues please be sure to include as much information as possible environment details os osx node js version npm version google cloud resource version steps to reproduce follow example documentation receive error about resource is not a constructor the line above should read const resource require google cloud resource notice the curly braces have been removed following these steps will guarantee the quickest resolution possible thanks
| 1
|
233,863
| 19,073,952,526
|
IssuesEvent
|
2021-11-27 12:18:51
|
JukeBot-Org/JukeBot
|
https://api.github.com/repos/JukeBot-Org/JukeBot
|
opened
|
!queue fails to account for a queue with >=10 tracks
|
type: bug branch: testing priority: low
|
If there is more than 10 tracks in a guild's queue when invoking the `!guild` command, the queue view shifts over. As expected, of course, but not as intended.

|
1.0
|
!queue fails to account for a queue with >=10 tracks - If there is more than 10 tracks in a guild's queue when invoking the `!guild` command, the queue view shifts over. As expected, of course, but not as intended.

|
non_process
|
queue fails to account for a queue with tracks if there is more than tracks in a guild s queue when invoking the guild command the queue view shifts over as expected of course but not as intended
| 0
|
289,631
| 31,991,439,442
|
IssuesEvent
|
2023-09-21 06:13:25
|
pazhanivel07/frameworks_av_AOSP_10_r33_CVE-2023-35687_CVE-2023-35679
|
https://api.github.com/repos/pazhanivel07/frameworks_av_AOSP_10_r33_CVE-2023-35687_CVE-2023-35679
|
opened
|
CVE-2020-0113 (Medium) detected in avandroid-10.0.0_r33
|
Mend: dependency security vulnerability
|
## CVE-2020-0113 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>avandroid-10.0.0_r33</b></p></summary>
<p>
<p>Library home page: <a href=https://android.googlesource.com/platform/frameworks/av>https://android.googlesource.com/platform/frameworks/av</a></p>
<p>Found in HEAD commit: <a href="https://github.com/pazhanivel07/frameworks_av_AOSP_10_r33_CVE-2023-35687_CVE-2023-35679/commit/f6c68d1f56234bf68c91ea85f355b0612cd36a2b">f6c68d1f56234bf68c91ea85f355b0612cd36a2b</a></p>
<p>Found in base branch: <b>main</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/services/camera/libcameraservice/device3/Camera3Device.cpp</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
In sendCaptureResult of Camera3OutputUtils.cpp, there is a possible out of bounds read due to a use after free. This could lead to local information disclosure with no additional execution privileges needed. User interaction is not needed for exploitation.Product: AndroidVersions: Android-10 Android-9Android ID: A-150944913
<p>Publish Date: 2020-06-10
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-0113>CVE-2020-0113</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://source.android.com/security/bulletin/2020-06-01">https://source.android.com/security/bulletin/2020-06-01</a></p>
<p>Release Date: 2020-06-02</p>
<p>Fix Resolution: android-10.0.0_r37,android-9.0.0_r56</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-0113 (Medium) detected in avandroid-10.0.0_r33 - ## CVE-2020-0113 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>avandroid-10.0.0_r33</b></p></summary>
<p>
<p>Library home page: <a href=https://android.googlesource.com/platform/frameworks/av>https://android.googlesource.com/platform/frameworks/av</a></p>
<p>Found in HEAD commit: <a href="https://github.com/pazhanivel07/frameworks_av_AOSP_10_r33_CVE-2023-35687_CVE-2023-35679/commit/f6c68d1f56234bf68c91ea85f355b0612cd36a2b">f6c68d1f56234bf68c91ea85f355b0612cd36a2b</a></p>
<p>Found in base branch: <b>main</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/services/camera/libcameraservice/device3/Camera3Device.cpp</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
In sendCaptureResult of Camera3OutputUtils.cpp, there is a possible out of bounds read due to a use after free. This could lead to local information disclosure with no additional execution privileges needed. User interaction is not needed for exploitation.Product: AndroidVersions: Android-10 Android-9Android ID: A-150944913
<p>Publish Date: 2020-06-10
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-0113>CVE-2020-0113</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://source.android.com/security/bulletin/2020-06-01">https://source.android.com/security/bulletin/2020-06-01</a></p>
<p>Release Date: 2020-06-02</p>
<p>Fix Resolution: android-10.0.0_r37,android-9.0.0_r56</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in avandroid cve medium severity vulnerability vulnerable library avandroid library home page a href found in head commit a href found in base branch main vulnerable source files services camera libcameraservice cpp vulnerability details in sendcaptureresult of cpp there is a possible out of bounds read due to a use after free this could lead to local information disclosure with no additional execution privileges needed user interaction is not needed for exploitation product androidversions android android id a publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution android android step up your open source security game with mend
| 0
|
17,450
| 23,269,180,506
|
IssuesEvent
|
2022-08-04 20:42:17
|
googleapis/google-cloud-ruby
|
https://api.github.com/repos/googleapis/google-cloud-ruby
|
closed
|
Solvable failures in the nightly samples tests
|
type: process samples
|
See https://source.cloud.google.com/results/invocations/7741f5cc-ca8e-4043-b484-4e7dae441929/log
* kms: appears to be due to the samples defining toplevel methods that are interfering with the clients. We should probably rename the sample methods (e.g. by appending `_sample` to them).
* monitoring: appears to be due to changes in the responses. We probably just need to update the expectations.
* firestore: this is probably due to the helical-zone project not being set up for firestore. we need to update the tests to use a different project.
|
1.0
|
Solvable failures in the nightly samples tests - See https://source.cloud.google.com/results/invocations/7741f5cc-ca8e-4043-b484-4e7dae441929/log
* kms: appears to be due to the samples defining toplevel methods that are interfering with the clients. We should probably rename the sample methods (e.g. by appending `_sample` to them).
* monitoring: appears to be due to changes in the responses. We probably just need to update the expectations.
* firestore: this is probably due to the helical-zone project not being set up for firestore. we need to update the tests to use a different project.
|
process
|
solvable failures in the nightly samples tests see kms appears to be due to the samples defining toplevel methods that are interfering with the clients we should probably rename the sample methods e g by appending sample to them monitoring appears to be due to changes in the responses we probably just need to update the expectations firestore this is probably due to the helical zone project not being set up for firestore we need to update the tests to use a different project
| 1
|
215,507
| 16,675,840,311
|
IssuesEvent
|
2021-06-07 16:04:14
|
sass/libsass
|
https://api.github.com/repos/sass/libsass
|
closed
|
url() should be able to contain an exclamation mark
|
Bug - Confirmed Compatibility - P3 Dart Backport Done Dev - Test Written
|
The [CSS spec](https://drafts.csswg.org/css-syntax-3/#url-token-diagram) says that a `url()` can contain `!`, so Sass should allow that too. Note that we don't allow *everything* that CSS does, because we want to be able to parse `url($foo)` as a URL taking a variable, but `!` is unlikely to be used in a SassScript expression.
See https://github.com/sass/dart-sass/issues/646
|
1.0
|
url() should be able to contain an exclamation mark - The [CSS spec](https://drafts.csswg.org/css-syntax-3/#url-token-diagram) says that a `url()` can contain `!`, so Sass should allow that too. Note that we don't allow *everything* that CSS does, because we want to be able to parse `url($foo)` as a URL taking a variable, but `!` is unlikely to be used in a SassScript expression.
See https://github.com/sass/dart-sass/issues/646
|
non_process
|
url should be able to contain an exclamation mark the says that a url can contain so sass should allow that too note that we don t allow everything that css does because we want to be able to parse url foo as a url taking a variable but is unlikely to be used in a sassscript expression see
| 0
|
18,332
| 24,450,529,156
|
IssuesEvent
|
2022-10-06 22:23:20
|
microsoft/react-native-windows
|
https://api.github.com/repos/microsoft/react-native-windows
|
closed
|
0.70 Release Status
|
enhancement Area: Release Process
|
## Checklist
**Before Preview**
- [x] Draft GitHub release notes from commit log (chiaramooney)
- [x] Promote canary build to preview using [wiki instructions](https://github.com/microsoft/react-native-windows/wiki/How-to-promote-a-release) (chiaramooney)
- [x] Push build to stable branch (chiaramooney)
- [x] Enable CI schedule for new branch of [CI pipeline](https://dev.azure.com/ms/react-native-windows/_apps/hub/ms.vss-ciworkflow.build-ci-hub?_a=edit-build-definition&id=468&view=Tab_Triggers)
- [x] Update [dashboard @ms](https://dev.azure.com/ms/react-native-windows/_dashboards/dashboard/28deb05d-f5bb-43e6-8aa9-36ad5e5476fb) with an entry for `CI ${version}`
- [x] Add release schedule for the new stable branch of [publish pipline](https://dev.azure.com/microsoft/ReactNative/_apps/hub/ms.vss-ciworkflow.build-ci-hub?_a=edit-build-definition&id=63081&view=Tab_Triggers)
- [x] Update [dashboard @microsoft](https://dev.azure.com/microsoft/ReactNative/_dashboards/dashboard/8ea77a11-83e2-493a-873f-9cd4562f213d) with an entry for `Publish ${version}`
- [x] Update GitHub release notes to use manually curated notes instead of a changelog (chiaramooney)
- [x] Post release notes internally (chiaramooney)
-----
**After Preview**
- [x] Move most issues targeting current release (chrisglein)
- [x] Test updated gallery app using [wiki instructions](https://github.com/microsoft/react-native-gallery/wiki/Manual-Validation-Steps-for-RNW-Release) (@chiaramooney)
- [x] Check [CI Runs](https://github.com/microsoft/react-native-windows-samples/actions?query=workflow:*(Upgrade)) for Upgrading Sample Apps (@jonthysell )
- [x] Snap Hermes-Windows release (@mganandraj)
- [x] Do a pass on API Docs using [wiki instructions](https://github.com/microsoft/react-native-windows/wiki/API-documentation#validating-api-docs-for-a-release) (@chiaramooney)
- [x] Integrate any applicable patch/prerelease releases for React Native (@chiaramooney )
-----
**Before Release**
- [x] Ensure doc issues are addressed (chiaramooney)
- [x] Promote `latest` build to `legacy` using [wiki instructions](https://github.com/microsoft/react-native-windows/wiki/How-to-promote-a-release) (chiaramooney)
-----
**Release**
- [x] Update preview release notes with any changes from cherry-picked PRs (chiaramooney)
- [x] Update samples (@TatianaKapos)
- [x] Update React Native Gallery and Publish (chiaramooney)
- [x] Promote `preview` build to `latest` using [wiki instructions](https://github.com/microsoft/react-native-windows/wiki/How-to-promote-a-release) (chiaramooney)
- [x] Update GitHub release notes to use manually curated notes instead of a changelog (chiaramooney)
- [x] Update website (chiaramooney)
- [x] Send out internal release announcement (chiaramooney)
- [x] Update CI to use /apiVersion 0.XX (chiaramooney) -- After Website Updated
- [x] Ensure Accessibility is tested through updated Gallery (@AgneLukoseviciute)
|
1.0
|
0.70 Release Status - ## Checklist
**Before Preview**
- [x] Draft GitHub release notes from commit log (chiaramooney)
- [x] Promote canary build to preview using [wiki instructions](https://github.com/microsoft/react-native-windows/wiki/How-to-promote-a-release) (chiaramooney)
- [x] Push build to stable branch (chiaramooney)
- [x] Enable CI schedule for new branch of [CI pipeline](https://dev.azure.com/ms/react-native-windows/_apps/hub/ms.vss-ciworkflow.build-ci-hub?_a=edit-build-definition&id=468&view=Tab_Triggers)
- [x] Update [dashboard @ms](https://dev.azure.com/ms/react-native-windows/_dashboards/dashboard/28deb05d-f5bb-43e6-8aa9-36ad5e5476fb) with an entry for `CI ${version}`
- [x] Add release schedule for the new stable branch of [publish pipline](https://dev.azure.com/microsoft/ReactNative/_apps/hub/ms.vss-ciworkflow.build-ci-hub?_a=edit-build-definition&id=63081&view=Tab_Triggers)
- [x] Update [dashboard @microsoft](https://dev.azure.com/microsoft/ReactNative/_dashboards/dashboard/8ea77a11-83e2-493a-873f-9cd4562f213d) with an entry for `Publish ${version}`
- [x] Update GitHub release notes to use manually curated notes instead of a changelog (chiaramooney)
- [x] Post release notes internally (chiaramooney)
-----
**After Preview**
- [x] Move most issues targeting current release (chrisglein)
- [x] Test updated gallery app using [wiki instructions](https://github.com/microsoft/react-native-gallery/wiki/Manual-Validation-Steps-for-RNW-Release) (@chiaramooney)
- [x] Check [CI Runs](https://github.com/microsoft/react-native-windows-samples/actions?query=workflow:*(Upgrade)) for Upgrading Sample Apps (@jonthysell )
- [x] Snap Hermes-Windows release (@mganandraj)
- [x] Do a pass on API Docs using [wiki instructions](https://github.com/microsoft/react-native-windows/wiki/API-documentation#validating-api-docs-for-a-release) (@chiaramooney)
- [x] Integrate any applicable patch/prerelease releases for React Native (@chiaramooney )
-----
**Before Release**
- [x] Ensure doc issues are addressed (chiaramooney)
- [x] Promote `latest` build to `legacy` using [wiki instructions](https://github.com/microsoft/react-native-windows/wiki/How-to-promote-a-release) (chiaramooney)
-----
**Release**
- [x] Update preview release notes with any changes from cherry-picked PRs (chiaramooney)
- [x] Update samples (@TatianaKapos)
- [x] Update React Native Gallery and Publish (chiaramooney)
- [x] Promote `preview` build to `latest` using [wiki instructions](https://github.com/microsoft/react-native-windows/wiki/How-to-promote-a-release) (chiaramooney)
- [x] Update GitHub release notes to use manually curated notes instead of a changelog (chiaramooney)
- [x] Update website (chiaramooney)
- [x] Send out internal release announcement (chiaramooney)
- [x] Update CI to use /apiVersion 0.XX (chiaramooney) -- After Website Updated
- [x] Ensure Accessibility is tested through updated Gallery (@AgneLukoseviciute)
|
process
|
release status checklist before preview draft github release notes from commit log chiaramooney promote canary build to preview using chiaramooney push build to stable branch chiaramooney enable ci schedule for new branch of update with an entry for ci version add release schedule for the new stable branch of update with an entry for publish version update github release notes to use manually curated notes instead of a changelog chiaramooney post release notes internally chiaramooney after preview move most issues targeting current release chrisglein test updated gallery app using chiaramooney check for upgrading sample apps jonthysell snap hermes windows release mganandraj do a pass on api docs using chiaramooney integrate any applicable patch prerelease releases for react native chiaramooney before release ensure doc issues are addressed chiaramooney promote latest build to legacy using chiaramooney release update preview release notes with any changes from cherry picked prs chiaramooney update samples tatianakapos update react native gallery and publish chiaramooney promote preview build to latest using chiaramooney update github release notes to use manually curated notes instead of a changelog chiaramooney update website chiaramooney send out internal release announcement chiaramooney update ci to use apiversion xx chiaramooney after website updated ensure accessibility is tested through updated gallery agnelukoseviciute
| 1
|
35,236
| 6,424,123,274
|
IssuesEvent
|
2017-08-09 12:51:02
|
RobertLucian/GoPiGo3
|
https://api.github.com/repos/RobertLucian/GoPiGo3
|
opened
|
`DistanceSensor.read_mm` giving false readings
|
bug documentation
|
The `read_mm` method of the `DistanceSensor` class seems to be returning values up to `3000` instead of going up to `8000`.
Also, when the sensor is out of range, the method doesn't return `8190` as the documentation says.
|
1.0
|
`DistanceSensor.read_mm` giving false readings - The `read_mm` method of the `DistanceSensor` class seems to be returning values up to `3000` instead of going up to `8000`.
Also, when the sensor is out of range, the method doesn't return `8190` as the documentation says.
|
non_process
|
distancesensor read mm giving false readings the read mm method of the distancesensor class seems to be returning values up to instead of going up to also when the sensor is out of range the method doesn t return as the documentation says
| 0
|
21,827
| 30,318,214,467
|
IssuesEvent
|
2023-07-10 17:07:09
|
tdwg/dwc
|
https://api.github.com/repos/tdwg/dwc
|
closed
|
Move terms from Occurrence to MaterialSample
|
Term - change Class - Occurrence Class - MaterialSample Process - need templated change request non-normative Task Group - Material Sample
|
Was
https://code.google.com/p/darwincore/issues/detail?id=236 and
https://code.google.com/p/darwincore/issues/detail?id=239 and
https://code.google.com/p/darwincore/issues/detail?id=241
When the MaterialSample proposal was ratified, no consideration had been given to which existing terms might be organized in this new class. Three terms (preparations, associatedSequences, and disposition) make sense to "belong in" this class.
==Term Change Recommendation==
Submitter: John Wieczorek
Justification: Previous to the adoption of the MaterialSample class, there was no better class in which to arrange preparations than Occurrence. MaterialSample is the appropriate class in which to arrange preparations.
Term Name: preparations
Reference: http://rs.tdwg.org/dwc/terms/index.htm#preparations
Version: preparations-2009-04-24
Status: deprecated
IsReplacedBy: preparations-2014-10-22
Term Name: preparations
Version: preparations-2014-10-22
Status: recommended
Replaces: preparations-2009-04-24
Class: http://rs.tdwg.org/dwc/terms/MaterialSample
==Term Change Recommendation==
Submitter: John Wieczorek
Justification: Previous to the adoption of the MaterialSample class, there was no better class in which to arrange associatedSequences than Occurrence. MaterialSample is the appropriate class in which to arrange associatedSequences.
Term Name: associatedSequences
Reference: http://rs.tdwg.org/dwc/terms/index.htm#associatedSequences
Version: associatedSequences-2009-04-24
Status: deprecated
Definition: A list (concatenated and separated) of identifiers (publication, global unique identifier, URI) of genetic sequence information associated with the Occurrence.
IsReplacedBy: associatedSequences-2014-10-22
Term Name: associatedSequences
Version: associatedSequences-2014-10-22
Status: recommended
Definition: A list (concatenated and separated) of identifiers (publication, global unique identifier, URI) of genetic sequence information associated with the MaterialSample.
Replaces: associatedSequences-2009-04-24
Class: http://rs.tdwg.org/dwc/terms/MaterialSample
==Term Change Recommendation==
Submitter: John Wieczorek
Justification: Previous to the adoption of the MaterialSample class, there was no better class in which to arrange disposition than Occurrence. MaterialSample is the appropriate class in which to arrange disposition.
Term Name: disposition
Reference: http://rs.tdwg.org/dwc/terms/index.htm#disposition
Version: disposition-2009-04-24
Status: deprecated
IsReplacedBy: disposition-2014-10-22
Term Name: disposition
Version: disposition-2014-10-22
Status: recommended
Replaces: disposition-2009-04-24
Class: http://rs.tdwg.org/dwc/terms/MaterialSample
|
1.0
|
Move terms from Occurrence to MaterialSample - Was
https://code.google.com/p/darwincore/issues/detail?id=236 and
https://code.google.com/p/darwincore/issues/detail?id=239 and
https://code.google.com/p/darwincore/issues/detail?id=241
When the MaterialSample proposal was ratified, no consideration had been given to which existing terms might be organized in this new class. Three terms (preparations, associatedSequences, and disposition) make sense to "belong in" this class.
==Term Change Recommendation==
Submitter: John Wieczorek
Justification: Previous to the adoption of the MaterialSample class, there was no better class in which to arrange preparations than Occurrence. MaterialSample is the appropriate class in which to arrange preparations.
Term Name: preparations
Reference: http://rs.tdwg.org/dwc/terms/index.htm#preparations
Version: preparations-2009-04-24
Status: deprecated
IsReplacedBy: preparations-2014-10-22
Term Name: preparations
Version: preparations-2014-10-22
Status: recommended
Replaces: preparations-2009-04-24
Class: http://rs.tdwg.org/dwc/terms/MaterialSample
==Term Change Recommendation==
Submitter: John Wieczorek
Justification: Previous to the adoption of the MaterialSample class, there was no better class in which to arrange associatedSequences than Occurrence. MaterialSample is the appropriate class in which to arrange associatedSequences.
Term Name: associatedSequences
Reference: http://rs.tdwg.org/dwc/terms/index.htm#associatedSequences
Version: associatedSequences-2009-04-24
Status: deprecated
Definition: A list (concatenated and separated) of identifiers (publication, global unique identifier, URI) of genetic sequence information associated with the Occurrence.
IsReplacedBy: associatedSequences-2014-10-22
Term Name: associatedSequences
Version: associatedSequences-2014-10-22
Status: recommended
Definition: A list (concatenated and separated) of identifiers (publication, global unique identifier, URI) of genetic sequence information associated with the MaterialSample.
Replaces: associatedSequences-2009-04-24
Class: http://rs.tdwg.org/dwc/terms/MaterialSample
==Term Change Recommendation==
Submitter: John Wieczorek
Justification: Previous to the adoption of the MaterialSample class, there was no better class in which to arrange disposition than Occurrence. MaterialSample is the appropriate class in which to arrange disposition.
Term Name: disposition
Reference: http://rs.tdwg.org/dwc/terms/index.htm#disposition
Version: disposition-2009-04-24
Status: deprecated
IsReplacedBy: disposition-2014-10-22
Term Name: disposition
Version: disposition-2014-10-22
Status: recommended
Replaces: disposition-2009-04-24
Class: http://rs.tdwg.org/dwc/terms/MaterialSample
|
process
|
move terms from occurrence to materialsample was and and when the materialsample proposal was ratified no consideration had been given to which existing terms might be organized in this new class three terms preparations associatedsequences and disposition make sense to belong in this class term change recommendation submitter john wieczorek justification previous to the adoption of the materialsample class there was no better class in which to arrange preparations than occurrence materialsample is the appropriate class in which to arrange preparations term name preparations reference version preparations status deprecated isreplacedby preparations term name preparations version preparations status recommended replaces preparations class term change recommendation submitter john wieczorek justification previous to the adoption of the materialsample class there was no better class in which to arrange associatedsequences than occurrence materialsample is the appropriate class in which to arrange associatedsequences term name associatedsequences reference version associatedsequences status deprecated definition a list concatenated and separated of identifiers publication global unique identifier uri of genetic sequence information associated with the occurrence isreplacedby associatedsequences term name associatedsequences version associatedsequences status recommended definition a list concatenated and separated of identifiers publication global unique identifier uri of genetic sequence information associated with the materialsample replaces associatedsequences class term change recommendation submitter john wieczorek justification previous to the adoption of the materialsample class there was no better class in which to arrange disposition than occurrence materialsample is the appropriate class in which to arrange disposition term name disposition reference version disposition status deprecated isreplacedby disposition term name disposition version disposition status recommended replaces disposition class
| 1
|
148,405
| 11,851,639,125
|
IssuesEvent
|
2020-03-24 18:26:42
|
hashgraph/hedera-mirror-node
|
https://api.github.com/repos/hashgraph/hedera-mirror-node
|
closed
|
REST API integration tests for all parameters
|
P2 enhancement rest test
|
**Problem**
We keep having regressions in the REST API because we don't have good test coverage. We have an integration test framework but it is not filled out with specs for all URLs and combinations of parameters.
In addition - the way we setup integration test data, execute integration tests, and compare results, is not intuitive/clear.
**Solution**
- Refactor the integration_test.js such that each test:
- ~clears out the database of data before each test (see cleanup.sql), using jest's `beforeEach`~
- ~specifies a test data file to use to setup database data at the beginning of the test (see integration_test_data.json and related logic in integration.test.js). The data setup files can be moved into a new `data` folder.~
- ~`specs` can specify which data file they will use in the spec's json (`data_setup_file`: _path_)~
- ~run the test against the target code (reqToSql or calling the REST API)~
- ~compare the results to a result file and output/assert differences~
- Add spec files to `hedera-mirror-rest/__tests__/specs` for every URL and parameters we support. Also add negative test cases.
**Alternatives**
**Additional Context**
integration.test.js contains tests for both the end-to-end REST API-through-DB and reqToSql-through-DB. Don't split the file up into multiple files without ensuring that the tests do not run in parallel, as there is only 1 integration test DB.
|
1.0
|
REST API integration tests for all parameters - **Problem**
We keep having regressions in the REST API because we don't have good test coverage. We have an integration test framework but it is not filled out with specs for all URLs and combinations of parameters.
In addition - the way we setup integration test data, execute integration tests, and compare results, is not intuitive/clear.
**Solution**
- Refactor the integration_test.js such that each test:
- ~clears out the database of data before each test (see cleanup.sql), using jest's `beforeEach`~
- ~specifies a test data file to use to setup database data at the beginning of the test (see integration_test_data.json and related logic in integration.test.js). The data setup files can be moved into a new `data` folder.~
- ~`specs` can specify which data file they will use in the spec's json (`data_setup_file`: _path_)~
- ~run the test against the target code (reqToSql or calling the REST API)~
- ~compare the results to a result file and output/assert differences~
- Add spec files to `hedera-mirror-rest/__tests__/specs` for every URL and parameters we support. Also add negative test cases.
**Alternatives**
**Additional Context**
integration.test.js contains tests for both the end-to-end REST API-through-DB and reqToSql-through-DB. Don't split the file up into multiple files without ensuring that the tests do not run in parallel, as there is only 1 integration test DB.
|
non_process
|
rest api integration tests for all parameters problem we keep having regressions in the rest api because we don t have good test coverage we have an integration test framework but it is not filled out with specs for all urls and combinations of parameters in addition the way we setup integration test data execute integration tests and compare results is not intuitive clear solution refactor the integration test js such that each test clears out the database of data before each test see cleanup sql using jest s beforeeach specifies a test data file to use to setup database data at the beginning of the test see integration test data json and related logic in integration test js the data setup files can be moved into a new data folder specs can specify which data file they will use in the spec s json data setup file path run the test against the target code reqtosql or calling the rest api compare the results to a result file and output assert differences add spec files to hedera mirror rest tests specs for every url and parameters we support also add negative test cases alternatives additional context integration test js contains tests for both the end to end rest api through db and reqtosql through db don t split the file up into multiple files without ensuring that the tests do not run in parallel as there is only integration test db
| 0
|
284,001
| 30,913,580,599
|
IssuesEvent
|
2023-08-05 02:18:11
|
hshivhare67/kernel_v4.19.72
|
https://api.github.com/repos/hshivhare67/kernel_v4.19.72
|
reopened
|
CVE-2020-24490 (Medium) detected in linuxlinux-4.19.282
|
Mend: dependency security vulnerability
|
## CVE-2020-24490 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.282</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/hshivhare67/kernel_v4.19.72/commit/139c4e073703974ca0b05255c4cff6dcd52a8e31">139c4e073703974ca0b05255c4cff6dcd52a8e31</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/bluetooth/hci_event.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
Improper buffer restrictions in BlueZ may allow an unauthenticated user to potentially enable denial of service via adjacent access. This affects all Linux kernel versions that support BlueZ.
<p>Publish Date: 2021-02-02
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-24490>CVE-2020-24490</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Adjacent
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2020-24490">https://www.linuxkernelcves.com/cves/CVE-2020-24490</a></p>
<p>Release Date: 2021-02-02</p>
<p>Fix Resolution: v4.4.240, v4.9.240, v4.14.202, v4.19.137, v5.4.56, v5.7.13, v5.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-24490 (Medium) detected in linuxlinux-4.19.282 - ## CVE-2020-24490 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.282</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/hshivhare67/kernel_v4.19.72/commit/139c4e073703974ca0b05255c4cff6dcd52a8e31">139c4e073703974ca0b05255c4cff6dcd52a8e31</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/bluetooth/hci_event.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
Improper buffer restrictions in BlueZ may allow an unauthenticated user to potentially enable denial of service via adjacent access. This affects all Linux kernel versions that support BlueZ.
<p>Publish Date: 2021-02-02
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-24490>CVE-2020-24490</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Adjacent
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2020-24490">https://www.linuxkernelcves.com/cves/CVE-2020-24490</a></p>
<p>Release Date: 2021-02-02</p>
<p>Fix Resolution: v4.4.240, v4.9.240, v4.14.202, v4.19.137, v5.4.56, v5.7.13, v5.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in linuxlinux cve medium severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch master vulnerable source files net bluetooth hci event c vulnerability details improper buffer restrictions in bluez may allow an unauthenticated user to potentially enable denial of service via adjacent access this affects all linux kernel versions that support bluez publish date url a href cvss score details base score metrics exploitability metrics attack vector adjacent attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
10,056
| 13,044,161,745
|
IssuesEvent
|
2020-07-29 03:47:25
|
tikv/tikv
|
https://api.github.com/repos/tikv/tikv
|
closed
|
UCP: Migrate scalar function `SubDateDatetimeDecimal` from TiDB
|
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
|
## Description
Port the scalar function `SubDateDatetimeDecimal` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @sticnarf
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
2.0
|
UCP: Migrate scalar function `SubDateDatetimeDecimal` from TiDB -
## Description
Port the scalar function `SubDateDatetimeDecimal` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @sticnarf
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
process
|
ucp migrate scalar function subdatedatetimedecimal from tidb description port the scalar function subdatedatetimedecimal from tidb to coprocessor score mentor s sticnarf recommended skills rust programming learning materials already implemented expressions ported from tidb
| 1
|
11,093
| 13,936,770,098
|
IssuesEvent
|
2020-10-22 13:21:52
|
prisma/prisma-engines
|
https://api.github.com/repos/prisma/prisma-engines
|
opened
|
Write a script to extract the `UserFacingError` metadata for all user facing errors and return it as JSON
|
kind/docs process/candidate team/engines topic: error
|
The script should use the `UserFacingError` trait to get information about each user facing error and put that information a JSON file for other tools to generate readable docs from it.
The only awkward bit is that we need a central list of all the errors in the user-facing-errors crate. Exhaustiveness can be enforced through a trait (make UserFacingError a subtrait of the trait implemented by being in that list) if we want to get fancy.
The JSON structure should be coordinated with @mhwelander — maybe something like:
```json
{
"user_facing_errors": [
{
"name": "DatabaseAccessDenied",
"docs": "General comment and context about the error - these are the doc comments on the rust type".
"error_code": "P1010",
"message": "User `{database_user}` was denied access on the database `{database_name}`",
"metadata": [{"name": "database_user", "docs": "Database user name"},{"name":"database_name"}],
},
"etc...",
],
}
```
Ideally, this would run in CI.
|
1.0
|
Write a script to extract the `UserFacingError` metadata for all user facing errors and return it as JSON - The script should use the `UserFacingError` trait to get information about each user facing error and put that information a JSON file for other tools to generate readable docs from it.
The only awkward bit is that we need a central list of all the errors in the user-facing-errors crate. Exhaustiveness can be enforced through a trait (make UserFacingError a subtrait of the trait implemented by being in that list) if we want to get fancy.
The JSON structure should be coordinated with @mhwelander — maybe something like:
```json
{
"user_facing_errors": [
{
"name": "DatabaseAccessDenied",
"docs": "General comment and context about the error - these are the doc comments on the rust type".
"error_code": "P1010",
"message": "User `{database_user}` was denied access on the database `{database_name}`",
"metadata": [{"name": "database_user", "docs": "Database user name"},{"name":"database_name"}],
},
"etc...",
],
}
```
Ideally, this would run in CI.
|
process
|
write a script to extract the userfacingerror metadata for all user facing errors and return it as json the script should use the userfacingerror trait to get information about each user facing error and put that information a json file for other tools to generate readable docs from it the only awkward bit is that we need a central list of all the errors in the user facing errors crate exhaustiveness can be enforced through a trait make userfacingerror a subtrait of the trait implemented by being in that list if we want to get fancy the json structure should be coordinated with mhwelander — maybe something like json user facing errors name databaseaccessdenied docs general comment and context about the error these are the doc comments on the rust type error code message user database user was denied access on the database database name metadata etc ideally this would run in ci
| 1
|
1,868
| 4,697,449,435
|
IssuesEvent
|
2016-10-12 09:24:13
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
degrading performance after using child_process
|
child_process confirmed-bug lts-watch-v4.x os x performance
|
After running a child process ( using exec or spawn ) I have found that the performance of my node.js application decreases by a factor of 10. Below is a contrived example and output.
```
var exec = require('child_process').exec;
function runExpensiveOperation(times) {
while(times > 0) {
console.time('expensiveOperation');
var str = 'lorem';
for ( var i=0;i< 10000000; i++) {
// string concatenation
str = str.length < 1000 ? str + str : '';
// math operation
i * i * i;
}
console.timeEnd('expensiveOperation');
times--;
}
}
console.log('PRE EXEC');
runExpensiveOperation(10);
exec('echo "hello"');
console.log('POST EXEC');
runExpensiveOperation(10);
```
Output:
```
PRE EXEC
expensiveOperation: 66.458ms
expensiveOperation: 65.735ms
expensiveOperation: 69.237ms
expensiveOperation: 65.269ms
expensiveOperation: 69.133ms
expensiveOperation: 65.639ms
expensiveOperation: 67.944ms
expensiveOperation: 63.595ms
expensiveOperation: 64.153ms
expensiveOperation: 65.093ms
POST EXEC
expensiveOperation: 715.861ms
expensiveOperation: 739.671ms
expensiveOperation: 714.546ms
expensiveOperation: 714.845ms
expensiveOperation: 745.719ms
expensiveOperation: 743.240ms
expensiveOperation: 716.481ms
expensiveOperation: 732.916ms
expensiveOperation: 736.576ms
expensiveOperation: 742.416ms
```
In addition, this problem only occurs if the string concatenation AND math operation are run in the expensiveOperation - if either are commented out then there is no issue.
* **Version**: 5.8.0
* **Platform**: Darwin Kernel Version 15.3.0: Thu Dec 10 18:40:58 PST 2015; root:xnu-3248.30.4~1/RELEASE_X86_64 x86_64 ( Macbook Air OS X El Capitan )
* **Subsystem**: child_process
|
1.0
|
degrading performance after using child_process - After running a child process ( using exec or spawn ) I have found that the performance of my node.js application decreases by a factor of 10. Below is a contrived example and output.
```
var exec = require('child_process').exec;
function runExpensiveOperation(times) {
while(times > 0) {
console.time('expensiveOperation');
var str = 'lorem';
for ( var i=0;i< 10000000; i++) {
// string concatenation
str = str.length < 1000 ? str + str : '';
// math operation
i * i * i;
}
console.timeEnd('expensiveOperation');
times--;
}
}
console.log('PRE EXEC');
runExpensiveOperation(10);
exec('echo "hello"');
console.log('POST EXEC');
runExpensiveOperation(10);
```
Output:
```
PRE EXEC
expensiveOperation: 66.458ms
expensiveOperation: 65.735ms
expensiveOperation: 69.237ms
expensiveOperation: 65.269ms
expensiveOperation: 69.133ms
expensiveOperation: 65.639ms
expensiveOperation: 67.944ms
expensiveOperation: 63.595ms
expensiveOperation: 64.153ms
expensiveOperation: 65.093ms
POST EXEC
expensiveOperation: 715.861ms
expensiveOperation: 739.671ms
expensiveOperation: 714.546ms
expensiveOperation: 714.845ms
expensiveOperation: 745.719ms
expensiveOperation: 743.240ms
expensiveOperation: 716.481ms
expensiveOperation: 732.916ms
expensiveOperation: 736.576ms
expensiveOperation: 742.416ms
```
In addition, this problem only occurs if the string concatenation AND math operation are run in the expensiveOperation - if either are commented out then there is no issue.
* **Version**: 5.8.0
* **Platform**: Darwin Kernel Version 15.3.0: Thu Dec 10 18:40:58 PST 2015; root:xnu-3248.30.4~1/RELEASE_X86_64 x86_64 ( Macbook Air OS X El Capitan )
* **Subsystem**: child_process
|
process
|
degrading performance after using child process after running a child process using exec or spawn i have found that the performance of my node js application decreases by a factor of below is a contrived example and output var exec require child process exec function runexpensiveoperation times while times console time expensiveoperation var str lorem for var i i i string concatenation str str length str str math operation i i i console timeend expensiveoperation times console log pre exec runexpensiveoperation exec echo hello console log post exec runexpensiveoperation output pre exec expensiveoperation expensiveoperation expensiveoperation expensiveoperation expensiveoperation expensiveoperation expensiveoperation expensiveoperation expensiveoperation expensiveoperation post exec expensiveoperation expensiveoperation expensiveoperation expensiveoperation expensiveoperation expensiveoperation expensiveoperation expensiveoperation expensiveoperation expensiveoperation in addition this problem only occurs if the string concatenation and math operation are run in the expensiveoperation if either are commented out then there is no issue version platform darwin kernel version thu dec pst root xnu release macbook air os x el capitan subsystem child process
| 1
|
37,841
| 18,792,476,583
|
IssuesEvent
|
2021-11-08 18:16:21
|
rust-lang/rustup
|
https://api.github.com/repos/rust-lang/rustup
|
reopened
|
Modify `rustup-init.sh` to support retrying the download of `rustup-init`
|
enhancement help wanted E-easy performance E-mentor
|
<!-- Thanks for filing an 🙋 enhancement request 😄! -->
**Describe the problem you are trying to solve**
<!-- A clear and concise description of the problem this enhancement request is trying to solve. -->
I am a newbie trying to install rustup. it is having difficulties on downloading required packages from static.rust-lang.org because the connection constantly drops randomly, it tries to reconnect 3 times all starting from zero (does the host not have resume capability!?, I dont know) then just fails the install progress. each reconnection fails at different percentages, one was at 15% while another at 95%.
**Describe the solution you'd like**
<!-- A clear and concise description of what you want to happen. -->
Instead of halting the progress, inform them and prompt users if they want to try for more.
Also, check beforehand that if the host has resume capability and again inform users about that. currently I am not sure who to blame for starting over the download progress.
**Notes**
<!-- Any additional context or information you feel may be relevant to the issue. -->
as of now, I could not complete the installation and unfortunately lost appetite towards rust for now. I hope problems would be solved till next time.
<!-- TRIAGEBOT_START -->
<!-- TRIAGEBOT_ASSIGN_START -->
<!-- TRIAGEBOT_ASSIGN_DATA_START$${"user":"facklambda"}$$TRIAGEBOT_ASSIGN_DATA_END -->
<!-- TRIAGEBOT_ASSIGN_END -->
<!-- TRIAGEBOT_END -->
|
True
|
Modify `rustup-init.sh` to support retrying the download of `rustup-init` - <!-- Thanks for filing an 🙋 enhancement request 😄! -->
**Describe the problem you are trying to solve**
<!-- A clear and concise description of the problem this enhancement request is trying to solve. -->
I am a newbie trying to install rustup. it is having difficulties on downloading required packages from static.rust-lang.org because the connection constantly drops randomly, it tries to reconnect 3 times all starting from zero (does the host not have resume capability!?, I dont know) then just fails the install progress. each reconnection fails at different percentages, one was at 15% while another at 95%.
**Describe the solution you'd like**
<!-- A clear and concise description of what you want to happen. -->
Instead of halting the progress, inform them and prompt users if they want to try for more.
Also, check beforehand that if the host has resume capability and again inform users about that. currently I am not sure who to blame for starting over the download progress.
**Notes**
<!-- Any additional context or information you feel may be relevant to the issue. -->
as of now, I could not complete the installation and unfortunately lost appetite towards rust for now. I hope problems would be solved till next time.
<!-- TRIAGEBOT_START -->
<!-- TRIAGEBOT_ASSIGN_START -->
<!-- TRIAGEBOT_ASSIGN_DATA_START$${"user":"facklambda"}$$TRIAGEBOT_ASSIGN_DATA_END -->
<!-- TRIAGEBOT_ASSIGN_END -->
<!-- TRIAGEBOT_END -->
|
non_process
|
modify rustup init sh to support retrying the download of rustup init describe the problem you are trying to solve i am a newbie trying to install rustup it is having difficulties on downloading required packages from static rust lang org because the connection constantly drops randomly it tries to reconnect times all starting from zero does the host not have resume capability i dont know then just fails the install progress each reconnection fails at different percentages one was at while another at describe the solution you d like instead of halting the progress inform them and prompt users if they want to try for more also check beforehand that if the host has resume capability and again inform users about that currently i am not sure who to blame for starting over the download progress notes as of now i could not complete the installation and unfortunately lost appetite towards rust for now i hope problems would be solved till next time
| 0
|
47,581
| 25,081,770,866
|
IssuesEvent
|
2022-11-07 19:59:34
|
benwbrum/fromthepage
|
https://api.github.com/repos/benwbrum/fromthepage
|
closed
|
user visit screen performance
|
performance
|
The admin-side screen displaying a list of user visits has terrible performance because we are fetching the deeds for each visit, but there is no index on `deeds.visit_id` so the DB has to do a full table scan, multiple times per screen.
|
True
|
user visit screen performance - The admin-side screen displaying a list of user visits has terrible performance because we are fetching the deeds for each visit, but there is no index on `deeds.visit_id` so the DB has to do a full table scan, multiple times per screen.
|
non_process
|
user visit screen performance the admin side screen displaying a list of user visits has terrible performance because we are fetching the deeds for each visit but there is no index on deeds visit id so the db has to do a full table scan multiple times per screen
| 0
|
12,365
| 14,892,885,511
|
IssuesEvent
|
2021-01-21 03:59:30
|
microsoft/react-native-windows
|
https://api.github.com/repos/microsoft/react-native-windows
|
closed
|
Automatic Release Note Creation Failed in 0.64-stable Branch
|
Area: Release Process bug
|
The task to create automated release notes failed for the first publish of the 0.64-stable branch. Beachball did correctly push Git tags, but the changelog creation script doesn't seem to locally find a new local tag without release notes. https://dev.azure.com/ms/react-native-windows/_build/results?buildId=125090&view=logs&j=bd0c9a5d-cce5-5ddc-8479-a97ca35bf82d&t=885808f2-8cab-58f0-ee91-26bf630acda8
|
1.0
|
Automatic Release Note Creation Failed in 0.64-stable Branch - The task to create automated release notes failed for the first publish of the 0.64-stable branch. Beachball did correctly push Git tags, but the changelog creation script doesn't seem to locally find a new local tag without release notes. https://dev.azure.com/ms/react-native-windows/_build/results?buildId=125090&view=logs&j=bd0c9a5d-cce5-5ddc-8479-a97ca35bf82d&t=885808f2-8cab-58f0-ee91-26bf630acda8
|
process
|
automatic release note creation failed in stable branch the task to create automated release notes failed for the first publish of the stable branch beachball did correctly push git tags but the changelog creation script doesn t seem to locally find a new local tag without release notes
| 1
|
182,445
| 6,670,244,344
|
IssuesEvent
|
2017-10-03 22:33:48
|
Cloud-CV/EvalAI
|
https://api.github.com/repos/Cloud-CV/EvalAI
|
opened
|
Add instructions to update settings.dev.py in Docker Installation section
|
docs easy_to_fix enhancement Hacktoberfest priority-high v2
|
## Deliverables
- [ ] Add a section to add instructions about renaming `dev.sample.py` to `dev.py` in the docker installation section in docs. See https://github.com/Cloud-CV/EvalAI/blob/master/README.md (4th point) for the text to add.
|
1.0
|
Add instructions to update settings.dev.py in Docker Installation section - ## Deliverables
- [ ] Add a section to add instructions about renaming `dev.sample.py` to `dev.py` in the docker installation section in docs. See https://github.com/Cloud-CV/EvalAI/blob/master/README.md (4th point) for the text to add.
|
non_process
|
add instructions to update settings dev py in docker installation section deliverables add a section to add instructions about renaming dev sample py to dev py in the docker installation section in docs see point for the text to add
| 0
|
15,439
| 19,655,787,602
|
IssuesEvent
|
2022-01-10 12:22:06
|
googleapis/google-cloud-dotnet
|
https://api.github.com/repos/googleapis/google-cloud-dotnet
|
closed
|
Compute V1 acceptance tests fail with dotnet test command in net461
|
priority: p2 type: process api: compute
|
AFAIK this bug only appears when running tests with `dotnet test` command with `net461` framework.
It does NOT appear:
- when running tests with `dotnet test` command with `netcoreapp2.1` framework.
- when running tests with `xunit.console.exe` (I used `~/.nuget/packages/xunit.runner.console/2.4.1/tools/net461/xunit.console.exe` specifically)
- when running tests with either built-in runner or the ReSharper runner from Visual Studio
- when I copied the test code (using `TranscodingTest.ListZones_MaxResults`) sans asserts into a new console application and ran it with `dotnet run`.
At this point I think there is some issue specifically when using `dotnet test`/`xUnit.net VSTest Adapter v2.4.1 (64-bit Desktop .NET 4.0.30319.42000) `
NB: when running from Visual Studio with the built-in test runner, it reports using ` xUnit.net VSTest Adapter v2.4.1 (64-bit Desktop .NET 4.0.30319.42000)` as well.
Repro steps (using windows console):
```
> set GOOGLE_APPLICATION_CREDENTIALS=C:\Path\To\Credentials\adc.json
> set TEST_PROJECT=my-project
> set TEST_PROJECT_LOCATION=us-central1
> cd C:\src\google-cloud-dotnet\apis\Google.Cloud.Compute.V1\Google.Cloud.Compute.V1.IntegrationTests
> dotnet build -c Debug --framework net461
<..>
Build succeeded.
0 Warning(s)
0 Error(s)
> dotnet test --no-build --logger:"console;verbosity=detailed" --filter:"FullyQualifiedName~TranscodingTest"
```
Result:
```
Test run for C:\src\google-cloud-dotnet\apis\Google.Cloud.Compute.V1\Google.Cloud.Compute.V1.IntegrationTests\bin\Debug\net461\Google.Cloud.Compute.V1.IntegrationTests.dll(.NETFramework,Version=v4.6.1) Microsoft (R) Test Execution Command Line Tool Version 16.7.1
Copyright (c) Microsoft Corporation. All rights reserved. Starting test execution, please wait... A total of 1 test files matched the specified pattern.
C:\src\google-cloud-dotnet\apis\Google.Cloud.Compute.V1\Google.Cloud.Compute.V1.IntegrationTests\bin\Debug\net461\Google.Cloud.Compute.V1.IntegrationTests.dll
[xUnit.net 00:00:00.00] xUnit.net VSTest Adapter v2.4.1 (64-bit Desktop .NET 4.0.30319.42000)
[xUnit.net 00:00:00.87] Discovering: Google.Cloud.Compute.V1.IntegrationTests
[xUnit.net 00:00:00.94] Discovered: Google.Cloud.Compute.V1.IntegrationTests
[xUnit.net 00:00:00.95] Starting: Google.Cloud.Compute.V1.IntegrationTests
[xUnit.net 00:00:03.98] Google.Cloud.Compute.V1.IntegrationTests.TranscodingTest.ListZones_MaxResults [FAIL]
[xUnit.net 00:00:03.98] Grpc.Core.RpcException : Status(StatusCode="Internal", Detail="")
[xUnit.net 00:00:03.98] Stack Trace:
<...>
```
Expected (result output for building and running tests with `netcoreapp2.1` framwork):
```
C:\src\google-cloud-dotnet\apis\Google.Cloud.Compute.V1\Google.Cloud.Compute.V1.IntegrationTests>dotnet test --no-build --logger:"console;verbosity=detailed" --filter:"FullyQualifiedName~TranscodingTest" Test run for C:\src\google-cloud-dotnet\apis\Google.Cloud.Compute.V1\Google.Cloud.Compute.V1.IntegrationTests\bin\Debug\netcoreapp2.1\Google.Cloud.Compute.V1.IntegrationTests.dll(.NETCoreApp,Version=v2.1) Microsoft (R) Test Execution Command Line Tool Version 16.7.1
Copyright (c) Microsoft Corporation. All rights reserved. Starting test execution, please wait... A total of 1 test files matched the specified pattern.
C:\src\google-cloud-dotnet\apis\Google.Cloud.Compute.V1\Google.Cloud.Compute.V1.IntegrationTests\bin\Debug\netcoreapp2.1\Google.Cloud.Compute.V1.IntegrationTests.dll
[xUnit.net 00:00:00.00] xUnit.net VSTest Adapter v2.4.1 (64-bit .NET Core 4.6.30015.01)
[xUnit.net 00:00:00.67] Discovering: Google.Cloud.Compute.V1.IntegrationTests
[xUnit.net 00:00:00.71] Discovered: Google.Cloud.Compute.V1.IntegrationTests
[xUnit.net 00:00:00.72] Starting: Google.Cloud.Compute.V1.IntegrationTests
[xUnit.net 00:00:02.57] Finished: Google.Cloud.Compute.V1.IntegrationTests
√ Google.Cloud.Compute.V1.IntegrationTests.TranscodingTest.ListZones_MaxResults [1s 762ms]
Standard Output Messages:
ListZones_MaxResults STARTS!
ListZones_MaxResults ENDS! Test Run Successful.
Total tests: 1
Passed: 1
Total time: 3.3102 Seconds
```
|
1.0
|
Compute V1 acceptance tests fail with dotnet test command in net461 - AFAIK this bug only appears when running tests with `dotnet test` command with `net461` framework.
It does NOT appear:
- when running tests with `dotnet test` command with `netcoreapp2.1` framework.
- when running tests with `xunit.console.exe` (I used `~/.nuget/packages/xunit.runner.console/2.4.1/tools/net461/xunit.console.exe` specifically)
- when running tests with either built-in runner or the ReSharper runner from Visual Studio
- when I copied the test code (using `TranscodingTest.ListZones_MaxResults`) sans asserts into a new console application and ran it with `dotnet run`.
At this point I think there is some issue specifically when using `dotnet test`/`xUnit.net VSTest Adapter v2.4.1 (64-bit Desktop .NET 4.0.30319.42000) `
NB: when running from Visual Studio with the built-in test runner, it reports using ` xUnit.net VSTest Adapter v2.4.1 (64-bit Desktop .NET 4.0.30319.42000)` as well.
Repro steps (using windows console):
```
> set GOOGLE_APPLICATION_CREDENTIALS=C:\Path\To\Credentials\adc.json
> set TEST_PROJECT=my-project
> set TEST_PROJECT_LOCATION=us-central1
> cd C:\src\google-cloud-dotnet\apis\Google.Cloud.Compute.V1\Google.Cloud.Compute.V1.IntegrationTests
> dotnet build -c Debug --framework net461
<..>
Build succeeded.
0 Warning(s)
0 Error(s)
> dotnet test --no-build --logger:"console;verbosity=detailed" --filter:"FullyQualifiedName~TranscodingTest"
```
Result:
```
Test run for C:\src\google-cloud-dotnet\apis\Google.Cloud.Compute.V1\Google.Cloud.Compute.V1.IntegrationTests\bin\Debug\net461\Google.Cloud.Compute.V1.IntegrationTests.dll(.NETFramework,Version=v4.6.1) Microsoft (R) Test Execution Command Line Tool Version 16.7.1
Copyright (c) Microsoft Corporation. All rights reserved. Starting test execution, please wait... A total of 1 test files matched the specified pattern.
C:\src\google-cloud-dotnet\apis\Google.Cloud.Compute.V1\Google.Cloud.Compute.V1.IntegrationTests\bin\Debug\net461\Google.Cloud.Compute.V1.IntegrationTests.dll
[xUnit.net 00:00:00.00] xUnit.net VSTest Adapter v2.4.1 (64-bit Desktop .NET 4.0.30319.42000)
[xUnit.net 00:00:00.87] Discovering: Google.Cloud.Compute.V1.IntegrationTests
[xUnit.net 00:00:00.94] Discovered: Google.Cloud.Compute.V1.IntegrationTests
[xUnit.net 00:00:00.95] Starting: Google.Cloud.Compute.V1.IntegrationTests
[xUnit.net 00:00:03.98] Google.Cloud.Compute.V1.IntegrationTests.TranscodingTest.ListZones_MaxResults [FAIL]
[xUnit.net 00:00:03.98] Grpc.Core.RpcException : Status(StatusCode="Internal", Detail="")
[xUnit.net 00:00:03.98] Stack Trace:
<...>
```
Expected (result output for building and running tests with `netcoreapp2.1` framwork):
```
C:\src\google-cloud-dotnet\apis\Google.Cloud.Compute.V1\Google.Cloud.Compute.V1.IntegrationTests>dotnet test --no-build --logger:"console;verbosity=detailed" --filter:"FullyQualifiedName~TranscodingTest" Test run for C:\src\google-cloud-dotnet\apis\Google.Cloud.Compute.V1\Google.Cloud.Compute.V1.IntegrationTests\bin\Debug\netcoreapp2.1\Google.Cloud.Compute.V1.IntegrationTests.dll(.NETCoreApp,Version=v2.1) Microsoft (R) Test Execution Command Line Tool Version 16.7.1
Copyright (c) Microsoft Corporation. All rights reserved. Starting test execution, please wait... A total of 1 test files matched the specified pattern.
C:\src\google-cloud-dotnet\apis\Google.Cloud.Compute.V1\Google.Cloud.Compute.V1.IntegrationTests\bin\Debug\netcoreapp2.1\Google.Cloud.Compute.V1.IntegrationTests.dll
[xUnit.net 00:00:00.00] xUnit.net VSTest Adapter v2.4.1 (64-bit .NET Core 4.6.30015.01)
[xUnit.net 00:00:00.67] Discovering: Google.Cloud.Compute.V1.IntegrationTests
[xUnit.net 00:00:00.71] Discovered: Google.Cloud.Compute.V1.IntegrationTests
[xUnit.net 00:00:00.72] Starting: Google.Cloud.Compute.V1.IntegrationTests
[xUnit.net 00:00:02.57] Finished: Google.Cloud.Compute.V1.IntegrationTests
√ Google.Cloud.Compute.V1.IntegrationTests.TranscodingTest.ListZones_MaxResults [1s 762ms]
Standard Output Messages:
ListZones_MaxResults STARTS!
ListZones_MaxResults ENDS! Test Run Successful.
Total tests: 1
Passed: 1
Total time: 3.3102 Seconds
```
|
process
|
compute acceptance tests fail with dotnet test command in afaik this bug only appears when running tests with dotnet test command with framework it does not appear when running tests with dotnet test command with framework when running tests with xunit console exe i used nuget packages xunit runner console tools xunit console exe specifically when running tests with either built in runner or the resharper runner from visual studio when i copied the test code using transcodingtest listzones maxresults sans asserts into a new console application and ran it with dotnet run at this point i think there is some issue specifically when using dotnet test xunit net vstest adapter bit desktop net nb when running from visual studio with the built in test runner it reports using xunit net vstest adapter bit desktop net as well repro steps using windows console set google application credentials c path to credentials adc json set test project my project set test project location us cd c src google cloud dotnet apis google cloud compute google cloud compute integrationtests dotnet build c debug framework build succeeded warning s error s dotnet test no build logger console verbosity detailed filter fullyqualifiedname transcodingtest result test run for c src google cloud dotnet apis google cloud compute google cloud compute integrationtests bin debug google cloud compute integrationtests dll netframework version microsoft r test execution command line tool version copyright c microsoft corporation all rights reserved starting test execution please wait a total of test files matched the specified pattern c src google cloud dotnet apis google cloud compute google cloud compute integrationtests bin debug google cloud compute integrationtests dll xunit net vstest adapter bit desktop net discovering google cloud compute integrationtests discovered google cloud compute integrationtests starting google cloud compute integrationtests google cloud compute integrationtests transcodingtest listzones maxresults grpc core rpcexception status statuscode internal detail stack trace expected result output for building and running tests with framwork c src google cloud dotnet apis google cloud compute google cloud compute integrationtests dotnet test no build logger console verbosity detailed filter fullyqualifiedname transcodingtest test run for c src google cloud dotnet apis google cloud compute google cloud compute integrationtests bin debug google cloud compute integrationtests dll netcoreapp version microsoft r test execution command line tool version copyright c microsoft corporation all rights reserved starting test execution please wait a total of test files matched the specified pattern c src google cloud dotnet apis google cloud compute google cloud compute integrationtests bin debug google cloud compute integrationtests dll xunit net vstest adapter bit net core discovering google cloud compute integrationtests discovered google cloud compute integrationtests starting google cloud compute integrationtests finished google cloud compute integrationtests √ google cloud compute integrationtests transcodingtest listzones maxresults standard output messages listzones maxresults starts listzones maxresults ends test run successful total tests passed total time seconds
| 1
|
10,586
| 27,119,783,632
|
IssuesEvent
|
2023-02-15 21:44:43
|
Azure/azure-sdk
|
https://api.github.com/repos/Azure/azure-sdk
|
closed
|
Board Review: Updates to Azure Digital Twins Data Plane SDKs (.NET, Java, Python, JS)
|
architecture board-review
|
**Intended to be a Small, Targeted Change**
Thank you for submitting this review request. Thorough review of your client library ensures that your APIs are consistent with the guidelines and the consumers of your client library have a consistently good experience when using Azure.
**The Architecture Board reviews [Track 2 libraries](https://azure.github.io/azure-sdk/general_introduction.html) only.** If your library does not meet this requirement, please reach out to [Architecture Board](adparch@microsoft.com) before creating the issue.
Please reference our [review process guidelines](https://azure.github.io/azure-sdk/policies_reviewprocess.html) to understand what is being asked for in the issue template.
To ensure consistency, all Tier-1 languages (C#, TypeScript, Java, Python) will generally be reviewed together. In expansive libraries, we will pair dynamic languages (Python, TypeScript) together, and strongly typed languages (C#, Java) together in separate meetings.
For Tier-2 languages (C, C++, Go, Android, iOS), the review will be on an as-needed basis.
**Before submitting, ensure you adjust the title of the issue appropriately.**
**Note that the required material must be included before a meeting can be scheduled.**
## Contacts and Timeline
* Responsible service team: Azure Digital Twins
* Main contacts: @sjiherzig
* Expected code complete date: 06/10/2022
* Expected release date: 06/30/2022
## About the Service
* Link to documentation introducing/describing the service: https://docs.microsoft.com/en-us/azure/digital-twins/
* Link to the service REST APIs: https://github.com/Azure/azure-rest-api-specs/pull/18790
* Link to GitHub issue for previous review sessions, if applicable:
## About the client library
* Name of the client library: Azure.DigitalTwins.Core
* Languages for this review: .NET, Java, Python, JS
## Artifacts required (per language)
Please read through “API Review” section [here](https://azure.github.io/azure-sdk/policies_reviewprocess.html) to understand how these artifacts are generated. **It is critical that these artifacts are present and are in the right format. If not, the language architects cannot review them with the SDK Team’s API review tool.**
### .NET
* APIView Link: https://apiview.dev/Assemblies/Review/291bf8a4879947cdb38fbc47c7fb7224?diffRevisionId=b455b58f8fe541d1a969f878ddbe1a26&doc=False&diffOnly=False&revisionId=a62a2453f71942cc84d00f18232548d5
* Link to Champion Scenarios/Quickstart samples: https://github.com/Azure/azure-sdk-for-net/pull/28785/files#diff-4714b93c926ef479d35cd67bc673fe86c575ff7166535bb067a947ca1652cfdc
PR: https://github.com/Azure/azure-sdk-for-net/pull/28785
### Java
* APIView Link: https://apiview.dev/Assemblies/Review/7ae2d85820a04a178fa7e0d5baa5d6d1?diffRevisionId=ecc65f047d9a46019f09f91c18f81c48&doc=False&diffOnly=False&revisionId=ab406df00d284c3d8079e4c9b168f3b9
* Link to Champion Scenarios/Quickstart samples: https://github.com/Azure/azure-sdk-for-java/pull/29420/files#diff-c2f7c604be6e61a4d4bb3799d4ee0d1dde06815814954d004a3dda3093a45306
PR: https://github.com/Azure/azure-sdk-for-java/pull/29420
### Python
* APIView Link: https://apiview.dev/Assemblies/Review/4f65c4fd0fe646d389f9cf7beee006d3?diffRevisionId=76081d5dbad7467881a6a98f814e1e3f&doc=False&diffOnly=False&revisionId=a66c60cdb1d44426bd059fc9b63a27d8
* Link to Champion Scenarios/Quickstart samples:
PR: https://github.com/Azure/azure-sdk-for-python/pull/24603
### TypeScript
* APIView Link: https://apiview.dev/Assemblies/Review/3acd8e0438b04a229a35cabb0e57edc8
* Link to Champion Scenarios/Quickstart samples:
Note: I could not find an API view artifact for the current stable release 1.0.3.
PR: https://github.com/Azure/azure-sdk-for-js/pull/22010
For all other languages, send a request to the Architecture Board to discuss the best format on individual basis.
## Thank you!
|
1.0
|
Board Review: Updates to Azure Digital Twins Data Plane SDKs (.NET, Java, Python, JS) - **Intended to be a Small, Targeted Change**
Thank you for submitting this review request. Thorough review of your client library ensures that your APIs are consistent with the guidelines and the consumers of your client library have a consistently good experience when using Azure.
**The Architecture Board reviews [Track 2 libraries](https://azure.github.io/azure-sdk/general_introduction.html) only.** If your library does not meet this requirement, please reach out to [Architecture Board](adparch@microsoft.com) before creating the issue.
Please reference our [review process guidelines](https://azure.github.io/azure-sdk/policies_reviewprocess.html) to understand what is being asked for in the issue template.
To ensure consistency, all Tier-1 languages (C#, TypeScript, Java, Python) will generally be reviewed together. In expansive libraries, we will pair dynamic languages (Python, TypeScript) together, and strongly typed languages (C#, Java) together in separate meetings.
For Tier-2 languages (C, C++, Go, Android, iOS), the review will be on an as-needed basis.
**Before submitting, ensure you adjust the title of the issue appropriately.**
**Note that the required material must be included before a meeting can be scheduled.**
## Contacts and Timeline
* Responsible service team: Azure Digital Twins
* Main contacts: @sjiherzig
* Expected code complete date: 06/10/2022
* Expected release date: 06/30/2022
## About the Service
* Link to documentation introducing/describing the service: https://docs.microsoft.com/en-us/azure/digital-twins/
* Link to the service REST APIs: https://github.com/Azure/azure-rest-api-specs/pull/18790
* Link to GitHub issue for previous review sessions, if applicable:
## About the client library
* Name of the client library: Azure.DigitalTwins.Core
* Languages for this review: .NET, Java, Python, JS
## Artifacts required (per language)
Please read through “API Review” section [here](https://azure.github.io/azure-sdk/policies_reviewprocess.html) to understand how these artifacts are generated. **It is critical that these artifacts are present and are in the right format. If not, the language architects cannot review them with the SDK Team’s API review tool.**
### .NET
* APIView Link: https://apiview.dev/Assemblies/Review/291bf8a4879947cdb38fbc47c7fb7224?diffRevisionId=b455b58f8fe541d1a969f878ddbe1a26&doc=False&diffOnly=False&revisionId=a62a2453f71942cc84d00f18232548d5
* Link to Champion Scenarios/Quickstart samples: https://github.com/Azure/azure-sdk-for-net/pull/28785/files#diff-4714b93c926ef479d35cd67bc673fe86c575ff7166535bb067a947ca1652cfdc
PR: https://github.com/Azure/azure-sdk-for-net/pull/28785
### Java
* APIView Link: https://apiview.dev/Assemblies/Review/7ae2d85820a04a178fa7e0d5baa5d6d1?diffRevisionId=ecc65f047d9a46019f09f91c18f81c48&doc=False&diffOnly=False&revisionId=ab406df00d284c3d8079e4c9b168f3b9
* Link to Champion Scenarios/Quickstart samples: https://github.com/Azure/azure-sdk-for-java/pull/29420/files#diff-c2f7c604be6e61a4d4bb3799d4ee0d1dde06815814954d004a3dda3093a45306
PR: https://github.com/Azure/azure-sdk-for-java/pull/29420
### Python
* APIView Link: https://apiview.dev/Assemblies/Review/4f65c4fd0fe646d389f9cf7beee006d3?diffRevisionId=76081d5dbad7467881a6a98f814e1e3f&doc=False&diffOnly=False&revisionId=a66c60cdb1d44426bd059fc9b63a27d8
* Link to Champion Scenarios/Quickstart samples:
PR: https://github.com/Azure/azure-sdk-for-python/pull/24603
### TypeScript
* APIView Link: https://apiview.dev/Assemblies/Review/3acd8e0438b04a229a35cabb0e57edc8
* Link to Champion Scenarios/Quickstart samples:
Note: I could not find an API view artifact for the current stable release 1.0.3.
PR: https://github.com/Azure/azure-sdk-for-js/pull/22010
For all other languages, send a request to the Architecture Board to discuss the best format on individual basis.
## Thank you!
|
non_process
|
board review updates to azure digital twins data plane sdks net java python js intended to be a small targeted change thank you for submitting this review request thorough review of your client library ensures that your apis are consistent with the guidelines and the consumers of your client library have a consistently good experience when using azure the architecture board reviews only if your library does not meet this requirement please reach out to adparch microsoft com before creating the issue please reference our to understand what is being asked for in the issue template to ensure consistency all tier languages c typescript java python will generally be reviewed together in expansive libraries we will pair dynamic languages python typescript together and strongly typed languages c java together in separate meetings for tier languages c c go android ios the review will be on an as needed basis before submitting ensure you adjust the title of the issue appropriately note that the required material must be included before a meeting can be scheduled contacts and timeline responsible service team azure digital twins main contacts sjiherzig expected code complete date expected release date about the service link to documentation introducing describing the service link to the service rest apis link to github issue for previous review sessions if applicable about the client library name of the client library azure digitaltwins core languages for this review net java python js artifacts required per language please read through “api review” section to understand how these artifacts are generated it is critical that these artifacts are present and are in the right format if not the language architects cannot review them with the sdk team’s api review tool net apiview link link to champion scenarios quickstart samples pr java apiview link link to champion scenarios quickstart samples pr python apiview link link to champion scenarios quickstart samples pr typescript apiview link link to champion scenarios quickstart samples note i could not find an api view artifact for the current stable release pr for all other languages send a request to the architecture board to discuss the best format on individual basis thank you
| 0
|
17,574
| 23,386,824,474
|
IssuesEvent
|
2022-08-11 14:23:44
|
apache/arrow-datafusion
|
https://api.github.com/repos/apache/arrow-datafusion
|
closed
|
Improve process for making changes to both `arrow-datafusion` and `arrow-ballista`
|
enhancement development-process
|
**Is your feature request related to a problem or challenge? Please describe what you are trying to do.**
As a follow-on to https://github.com/apache/arrow-datafusion/pull/2582 it would be nice if we could make the process a little easier and more automated.
**Describe the solution you'd like**
Perhaps we can implement a mechanism where we can say "Depends on ballista branch X" in the PR description and have the scripts pull that version?
**Describe alternatives you've considered**
Live with the current solution.
**Additional context**
None
|
1.0
|
Improve process for making changes to both `arrow-datafusion` and `arrow-ballista` - **Is your feature request related to a problem or challenge? Please describe what you are trying to do.**
As a follow-on to https://github.com/apache/arrow-datafusion/pull/2582 it would be nice if we could make the process a little easier and more automated.
**Describe the solution you'd like**
Perhaps we can implement a mechanism where we can say "Depends on ballista branch X" in the PR description and have the scripts pull that version?
**Describe alternatives you've considered**
Live with the current solution.
**Additional context**
None
|
process
|
improve process for making changes to both arrow datafusion and arrow ballista is your feature request related to a problem or challenge please describe what you are trying to do as a follow on to it would be nice if we could make the process a little easier and more automated describe the solution you d like perhaps we can implement a mechanism where we can say depends on ballista branch x in the pr description and have the scripts pull that version describe alternatives you ve considered live with the current solution additional context none
| 1
|
8,709
| 11,850,973,921
|
IssuesEvent
|
2020-03-24 17:25:02
|
googleapis/python-pubsub
|
https://api.github.com/repos/googleapis/python-pubsub
|
closed
|
Add test for 504 DeadlineExceeded for synchronous pull without published messages
|
api: pubsub testing type: process
|
Related to #47 and the [pull request](https://github.com/googleapis/python-pubsub/pull/43) that will fix it.
We should cover the fix with a system test to detect if a regression occurs again.
|
1.0
|
Add test for 504 DeadlineExceeded for synchronous pull without published messages - Related to #47 and the [pull request](https://github.com/googleapis/python-pubsub/pull/43) that will fix it.
We should cover the fix with a system test to detect if a regression occurs again.
|
process
|
add test for deadlineexceeded for synchronous pull without published messages related to and the that will fix it we should cover the fix with a system test to detect if a regression occurs again
| 1
|
1,002
| 3,004,566,972
|
IssuesEvent
|
2015-07-26 02:45:43
|
DarthGandalf/kvirc-issues
|
https://api.github.com/repos/DarthGandalf/kvirc-issues
|
closed
|
security issue in all versions
|
C: IRC Protocol P: Security R: fixed T: Bug V: 4.0.2 Insomnia
|
**Reported by unic0rn on 25 Jul 2010 19:00:58 UTC**
remote command execution.
/ctcp nickname DCC GET\rQUIT\r
/ctcp nickname DCC GET\rPRIVMSG\40#channel\40:epic\40fail\r
whole channel can be exploited at once. need i say more?
when testing, make sure whole dcc thing is enclosed in 0x01. tested on 3.x and 4.x, including current svn build.
---
Migrated from: https://svn.kvirc.de/kvirc/ticket/858
|
True
|
security issue in all versions - **Reported by unic0rn on 25 Jul 2010 19:00:58 UTC**
remote command execution.
/ctcp nickname DCC GET\rQUIT\r
/ctcp nickname DCC GET\rPRIVMSG\40#channel\40:epic\40fail\r
whole channel can be exploited at once. need i say more?
when testing, make sure whole dcc thing is enclosed in 0x01. tested on 3.x and 4.x, including current svn build.
---
Migrated from: https://svn.kvirc.de/kvirc/ticket/858
|
non_process
|
security issue in all versions reported by on jul utc remote command execution ctcp nickname dcc get rquit r ctcp nickname dcc get rprivmsg channel epic r whole channel can be exploited at once need i say more when testing make sure whole dcc thing is enclosed in tested on x and x including current svn build migrated from
| 0
|
500,855
| 14,516,292,442
|
IssuesEvent
|
2020-12-13 15:27:26
|
Aam-Digital/ndb-core
|
https://api.github.com/repos/Aam-Digital/ndb-core
|
closed
|
user-defined, sortable groups for roll call
|
Priority: High Status: Ready to Work On Type: Feature
|
*As a teacher/socialworker I want to define my own groups of children in order to efficiently take attendance of a certain group that attends my class or center without having to skip over a lot of other students.*
**Objective**
While our app has a filter for the daily attendance / roll call feature (you can display children for a certain center or for a specific school) this is not flexible enough. Often class is attended by a group of students from several schools but not by all students of the center.
Users should be able to create custom groups of students and select from these when starting to take a day's attendance.
**Proposed Solutions & Alternatives**
* children group management view to select certain children, create and save a new group or edit an existing group
* daily attendance to allow selecting from the children groups (in addition to centers and schools as possible filters?)
> As this is a larger, critical feature the design (technical and UX) should be discussed within the team early for feedback.
**Further ideas for extension in the future**
* maybe an option to save shortcuts to "favourite groups" linked to the user account so that a user can very quickly jump to the daily attendance for his class of students.
|
1.0
|
user-defined, sortable groups for roll call - *As a teacher/socialworker I want to define my own groups of children in order to efficiently take attendance of a certain group that attends my class or center without having to skip over a lot of other students.*
**Objective**
While our app has a filter for the daily attendance / roll call feature (you can display children for a certain center or for a specific school) this is not flexible enough. Often class is attended by a group of students from several schools but not by all students of the center.
Users should be able to create custom groups of students and select from these when starting to take a day's attendance.
**Proposed Solutions & Alternatives**
* children group management view to select certain children, create and save a new group or edit an existing group
* daily attendance to allow selecting from the children groups (in addition to centers and schools as possible filters?)
> As this is a larger, critical feature the design (technical and UX) should be discussed within the team early for feedback.
**Further ideas for extension in the future**
* maybe an option to save shortcuts to "favourite groups" linked to the user account so that a user can very quickly jump to the daily attendance for his class of students.
|
non_process
|
user defined sortable groups for roll call as a teacher socialworker i want to define my own groups of children in order to efficiently take attendance of a certain group that attends my class or center without having to skip over a lot of other students objective while our app has a filter for the daily attendance roll call feature you can display children for a certain center or for a specific school this is not flexible enough often class is attended by a group of students from several schools but not by all students of the center users should be able to create custom groups of students and select from these when starting to take a day s attendance proposed solutions alternatives children group management view to select certain children create and save a new group or edit an existing group daily attendance to allow selecting from the children groups in addition to centers and schools as possible filters as this is a larger critical feature the design technical and ux should be discussed within the team early for feedback further ideas for extension in the future maybe an option to save shortcuts to favourite groups linked to the user account so that a user can very quickly jump to the daily attendance for his class of students
| 0
|
22,423
| 3,645,311,314
|
IssuesEvent
|
2016-02-15 14:07:57
|
MBB-team/VBA-toolbox
|
https://api.github.com/repos/MBB-team/VBA-toolbox
|
closed
|
VBA_ReDisplay_sessions : non-existent field 'i_abs'
|
auto-migrated Priority-Medium Type-Defect
|
_From @GoogleCodeExporter on October 7, 2015 9:29_
```
What steps will reproduce the problem?
1. Fit Extended RL model (6 sessions)
2. VBA_ReDisplay_sessions(posterior, out, {[1 2]}
3. The graphical output window appears with the Summary tab looking fine, no
error
4. When clicking on any of the other tabs ('VB inv, se...', ..., 'priors,
ses...'), the tab is blank and an Error is thrown :
??? Reference to non-existent field 'i_abs'.
Error in ==> VBA_ReDisplay_sessions>myPriors at 170
i_tab = ud.i_abs;
Error in ==> spm_uitab>doChoose at 195
feval(ud.callback);
??? Error while evaluating uicontrol Callback
Toolbox updated on 2013-10-11
Matlab 7.10.0 (R2010a) 32 bits
Debian 7.1 (wheezy) 32 bits
```
Original issue reported on code.google.com by `mail.ale...@gmail.com` on 13 Oct 2013 at 6:02
_Copied from original issue: lionel-rigoux/mbb-vb-toolbox#3_
|
1.0
|
VBA_ReDisplay_sessions : non-existent field 'i_abs' - _From @GoogleCodeExporter on October 7, 2015 9:29_
```
What steps will reproduce the problem?
1. Fit Extended RL model (6 sessions)
2. VBA_ReDisplay_sessions(posterior, out, {[1 2]}
3. The graphical output window appears with the Summary tab looking fine, no
error
4. When clicking on any of the other tabs ('VB inv, se...', ..., 'priors,
ses...'), the tab is blank and an Error is thrown :
??? Reference to non-existent field 'i_abs'.
Error in ==> VBA_ReDisplay_sessions>myPriors at 170
i_tab = ud.i_abs;
Error in ==> spm_uitab>doChoose at 195
feval(ud.callback);
??? Error while evaluating uicontrol Callback
Toolbox updated on 2013-10-11
Matlab 7.10.0 (R2010a) 32 bits
Debian 7.1 (wheezy) 32 bits
```
Original issue reported on code.google.com by `mail.ale...@gmail.com` on 13 Oct 2013 at 6:02
_Copied from original issue: lionel-rigoux/mbb-vb-toolbox#3_
|
non_process
|
vba redisplay sessions non existent field i abs from googlecodeexporter on october what steps will reproduce the problem fit extended rl model sessions vba redisplay sessions posterior out the graphical output window appears with the summary tab looking fine no error when clicking on any of the other tabs vb inv se priors ses the tab is blank and an error is thrown reference to non existent field i abs error in vba redisplay sessions mypriors at i tab ud i abs error in spm uitab dochoose at feval ud callback error while evaluating uicontrol callback toolbox updated on matlab bits debian wheezy bits original issue reported on code google com by mail ale gmail com on oct at copied from original issue lionel rigoux mbb vb toolbox
| 0
|
52,831
| 13,225,113,679
|
IssuesEvent
|
2020-08-17 20:31:05
|
icecube-trac/tix4
|
https://api.github.com/repos/icecube-trac/tix4
|
closed
|
sim-services/PropagatorServiceUtils::Propagate broken/non-functional (Trac #423)
|
Migrated from Trac combo reconstruction defect
|
I'd like to commit the following attached patch with the following commit message:
PropagatorServiceUtils::Propagate replaces its input pointer and should thus get it passed by reference. Also the wrong MCTree was modified (the original instead of the output copy).
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/423">https://code.icecube.wisc.edu/projects/icecube/ticket/423</a>, reported by claudio.kopperand owned by olivas</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2012-10-31T17:33:36",
"_ts": "1351704816000000",
"description": "I'd like to commit the following attached patch with the following commit message:\n\nPropagatorServiceUtils::Propagate replaces its input pointer and should thus get it passed by reference. Also the wrong MCTree was modified (the original instead of the output copy).\n",
"reporter": "claudio.kopper",
"cc": "",
"resolution": "fixed",
"time": "2012-06-25T00:59:43",
"component": "combo reconstruction",
"summary": "sim-services/PropagatorServiceUtils::Propagate broken/non-functional",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
sim-services/PropagatorServiceUtils::Propagate broken/non-functional (Trac #423) - I'd like to commit the following attached patch with the following commit message:
PropagatorServiceUtils::Propagate replaces its input pointer and should thus get it passed by reference. Also the wrong MCTree was modified (the original instead of the output copy).
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/423">https://code.icecube.wisc.edu/projects/icecube/ticket/423</a>, reported by claudio.kopperand owned by olivas</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2012-10-31T17:33:36",
"_ts": "1351704816000000",
"description": "I'd like to commit the following attached patch with the following commit message:\n\nPropagatorServiceUtils::Propagate replaces its input pointer and should thus get it passed by reference. Also the wrong MCTree was modified (the original instead of the output copy).\n",
"reporter": "claudio.kopper",
"cc": "",
"resolution": "fixed",
"time": "2012-06-25T00:59:43",
"component": "combo reconstruction",
"summary": "sim-services/PropagatorServiceUtils::Propagate broken/non-functional",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
</p>
</details>
|
non_process
|
sim services propagatorserviceutils propagate broken non functional trac i d like to commit the following attached patch with the following commit message propagatorserviceutils propagate replaces its input pointer and should thus get it passed by reference also the wrong mctree was modified the original instead of the output copy migrated from json status closed changetime ts description i d like to commit the following attached patch with the following commit message n npropagatorserviceutils propagate replaces its input pointer and should thus get it passed by reference also the wrong mctree was modified the original instead of the output copy n reporter claudio kopper cc resolution fixed time component combo reconstruction summary sim services propagatorserviceutils propagate broken non functional priority normal keywords milestone owner olivas type defect
| 0
|
22,277
| 3,619,793,015
|
IssuesEvent
|
2016-02-08 17:19:02
|
miracle091/transmission-remote-dotnet
|
https://api.github.com/repos/miracle091/transmission-remote-dotnet
|
closed
|
Crash
|
Priority-Medium Type-Defect
|
```
What steps will reproduce the problem?
No idea - i just got crash report.
What is the expected output? What do you see instead?
NA
What version of the products are you using?
OS:
Transmission: transmission - 2.42-1
Remote: 3.24(build 3)
Please provide any additional information below. Feel free to
attach screenshots or sample code which demonstrates the issue being
described.
```
Original issue reported on code.google.com by `chetan...@gmail.com` on 12 Mar 2012 at 7:05
Attachments:
* [trdcrash_20120312_062411.log](https://storage.googleapis.com/google-code-attachments/transmission-remote-dotnet/issue-438/comment-0/trdcrash_20120312_062411.log)
|
1.0
|
Crash - ```
What steps will reproduce the problem?
No idea - i just got crash report.
What is the expected output? What do you see instead?
NA
What version of the products are you using?
OS:
Transmission: transmission - 2.42-1
Remote: 3.24(build 3)
Please provide any additional information below. Feel free to
attach screenshots or sample code which demonstrates the issue being
described.
```
Original issue reported on code.google.com by `chetan...@gmail.com` on 12 Mar 2012 at 7:05
Attachments:
* [trdcrash_20120312_062411.log](https://storage.googleapis.com/google-code-attachments/transmission-remote-dotnet/issue-438/comment-0/trdcrash_20120312_062411.log)
|
non_process
|
crash what steps will reproduce the problem no idea i just got crash report what is the expected output what do you see instead na what version of the products are you using os transmission transmission remote build please provide any additional information below feel free to attach screenshots or sample code which demonstrates the issue being described original issue reported on code google com by chetan gmail com on mar at attachments
| 0
|
17,625
| 23,443,531,643
|
IssuesEvent
|
2022-08-15 17:12:05
|
crim-ca/weaver
|
https://api.github.com/repos/crim-ca/weaver
|
closed
|
[Feature] Handle Accept header for JSON/XML response in WPS 2.0
|
process/wps1 process/wps2 process/wps3
|
## Context
~Given that WPS 2.0 supports JSON response, we should handle a provided `Accept=application/json` header on the WPS route (eg: `/ows/wps` by default). By default this route returns XML.~
~Some code to do the detection of WPS version/format was introduced in https://github.com/crim-ca/weaver/commit/f0cd31e0c91c4a3a82a37d36fcdca8c0a7156c86 but is still unused.~
This is now functional
```http request
GET /ows/wps?service=wps&request=describeprocess&identifier=jsonarray2netcdf
Accept: application/json
```
<details>
<summary>See returned response (proper <i>OGCAPI - Processes</i> content)</summary>
<pre><code>{
"id": "jsonarray2netcdf",
"title": "JSON array to NetCDF",
"version": "1.1",
"description": "Extracts and fetches NetCDF files from a JSON file containing an URL string array, and provides them on the output directory.",
"keywords": [
"builtin"
],
"metadata": [],
"inputs": {
"input": {
"title": "input",
"minOccurs": 1,
"maxOccurs": 1,
"formats": [
{
"default": true,
"mediaType": "application/json"
}
]
}
},
"outputs": {
"output": {
"title": "output",
"formats": [
{
"default": true,
"mediaType": "application/x-netcdf"
}
]
}
},
"visibility": "public",
"jobControlOptions": [
"async-execute"
],
"outputTransmission": [
"reference"
],
"processDescriptionURL": "http://localhost:4002/processes/jsonarray2netcdf",
"processEndpointWPS1": "http://localhost:4002/ows/wps",
"executeEndpoint": "http://localhost:4002/processes/jsonarray2netcdf/jobs",
"links": [
{
"type": "application/json",
"title": "Current process description.",
"hreflang": "en-CA",
"href": "http://localhost:4002/processes/jsonarray2netcdf",
"rel": "self"
},
{
"type": "application/json",
"title": "Process definition.",
"hreflang": "en-CA",
"href": "http://localhost:4002/processes/jsonarray2netcdf",
"rel": "process-meta"
},
{
"type": "application/json",
"title": "Process execution endpoint for job submission.",
"hreflang": "en-CA",
"href": "http://localhost:4002/processes/jsonarray2netcdf/execution",
"rel": "http://www.opengis.net/def/rel/ogc/1.0/execute"
},
{
"type": "application/json",
"title": "List of registered processes.",
"hreflang": "en-CA",
"href": "http://localhost:4002/processes",
"rel": "http://www.opengis.net/def/rel/ogc/1.0/processes"
},
{
"type": "application/json",
"title": "List of job executions corresponding to this process.",
"hreflang": "en-CA",
"href": "http://localhost:4002/processes/jsonarray2netcdf/jobs",
"rel": "http://www.opengis.net/def/rel/ogc/1.0/job-list"
},
{
"type": "application/json",
"title": "List of processes registered under the service.",
"hreflang": "en-CA",
"href": "http://localhost:4002/processes",
"rel": "up"
}
]
}</code></pre>
</details>
~If version is WPS 1.x, the only supported format is XML. This should also be the default return on `/ows/wps` route if no `version=...` parameter was provided in the URL.~
Default behaviour is to provide the corresponding XML into JSON format, regardless of version:
```http request
GET /ows/wps?service=wps&request=describeprocess&identifier=jsonarray2netcdf&f=json
```
<details>
<summary>See produced response (JSON PyWPS content)</summary>
<pre><code>{
"pywps_version": "4.5.0",
"processes": [
{
"class": "weaver.processes.wps_package:WpsPackage",
"uuid": "None",
"workdir": null,
"version": "1.1",
"identifier": "jsonarray2netcdf",
"title": "JSON array to NetCDF",
"abstract": "Extracts and fetches NetCDF files from a JSON file containing an URL string array, and provides them on the output directory.",
"keywords": [
"builtin"
],
"metadata": [],
"inputs": [
{
"identifier": "input",
"title": "input",
"abstract": "",
"keywords": [],
"metadata": [],
"type": "complex",
"data_format": {
"mime_type": "application/json",
"encoding": "",
"schema": "",
"extension": ".json"
},
"asreference": false,
"supported_formats": [
{
"mime_type": "application/json",
"encoding": "",
"schema": "",
"extension": ".json"
}
],
"workdir": null,
"mode": 1,
"min_occurs": 1,
"max_occurs": 1,
"translations": null,
"mimetype": "application/json"
}
],
"outputs": [
{
"identifier": "output",
"title": "output",
"abstract": "",
"keywords": [],
"type": "complex",
"supported_formats": [
{
"mime_type": "application/x-netcdf",
"encoding": "",
"schema": "",
"extension": ".nc"
}
],
"asreference": false,
"data_format": {
"mime_type": "application/x-netcdf",
"encoding": "",
"schema": "",
"extension": ".nc"
},
"file": null,
"workdir": null,
"mode": 1,
"min_occurs": 1,
"max_occurs": 1,
"translations": null,
"mimetype": "application/x-netcdf"
}
],
"store_supported": "true",
"status_supported": "true",
"profile": [],
"translations": null
}
],
"language": "fr-CA"
}</code></pre>
</details>
## TODO
- [x] redirect a WPS-REST (3.0) call (eg: `/processes`, `/processes/{id}`, etc.) to their corresponding WPS 2.0 `/ows/wps` request if any from (`Accept: application/xml`, `Accept: text/xml`, `f=xml`, `format=xml`) is provided (inverse of what is currently functional).
Use [dict2xml](https://pypi.org/project/dict2xml/) (or similar) to provide an XML body corresponding to the original JSON.
- [x] handle other variation `format=` as alias of `f=` on OWS endpoint.
- [x] convert the PyWPS JSON to an actual (valid) *OGC API - Processes* representation rather than plain XML->JSON conversion.
- [x] Employ this tween definition to detect Accept header / format query and raise HTTP 422 if not acceptable :
https://github.com/Ouranosinc/Magpie/blob/f67a134d52ec8f3951cb1a08a07c1c5beb518882/magpie/api/generic.py#L106-L202
|
3.0
|
[Feature] Handle Accept header for JSON/XML response in WPS 2.0 - ## Context
~Given that WPS 2.0 supports JSON response, we should handle a provided `Accept=application/json` header on the WPS route (eg: `/ows/wps` by default). By default this route returns XML.~
~Some code to do the detection of WPS version/format was introduced in https://github.com/crim-ca/weaver/commit/f0cd31e0c91c4a3a82a37d36fcdca8c0a7156c86 but is still unused.~
This is now functional
```http request
GET /ows/wps?service=wps&request=describeprocess&identifier=jsonarray2netcdf
Accept: application/json
```
<details>
<summary>See returned response (proper <i>OGCAPI - Processes</i> content)</summary>
<pre><code>{
"id": "jsonarray2netcdf",
"title": "JSON array to NetCDF",
"version": "1.1",
"description": "Extracts and fetches NetCDF files from a JSON file containing an URL string array, and provides them on the output directory.",
"keywords": [
"builtin"
],
"metadata": [],
"inputs": {
"input": {
"title": "input",
"minOccurs": 1,
"maxOccurs": 1,
"formats": [
{
"default": true,
"mediaType": "application/json"
}
]
}
},
"outputs": {
"output": {
"title": "output",
"formats": [
{
"default": true,
"mediaType": "application/x-netcdf"
}
]
}
},
"visibility": "public",
"jobControlOptions": [
"async-execute"
],
"outputTransmission": [
"reference"
],
"processDescriptionURL": "http://localhost:4002/processes/jsonarray2netcdf",
"processEndpointWPS1": "http://localhost:4002/ows/wps",
"executeEndpoint": "http://localhost:4002/processes/jsonarray2netcdf/jobs",
"links": [
{
"type": "application/json",
"title": "Current process description.",
"hreflang": "en-CA",
"href": "http://localhost:4002/processes/jsonarray2netcdf",
"rel": "self"
},
{
"type": "application/json",
"title": "Process definition.",
"hreflang": "en-CA",
"href": "http://localhost:4002/processes/jsonarray2netcdf",
"rel": "process-meta"
},
{
"type": "application/json",
"title": "Process execution endpoint for job submission.",
"hreflang": "en-CA",
"href": "http://localhost:4002/processes/jsonarray2netcdf/execution",
"rel": "http://www.opengis.net/def/rel/ogc/1.0/execute"
},
{
"type": "application/json",
"title": "List of registered processes.",
"hreflang": "en-CA",
"href": "http://localhost:4002/processes",
"rel": "http://www.opengis.net/def/rel/ogc/1.0/processes"
},
{
"type": "application/json",
"title": "List of job executions corresponding to this process.",
"hreflang": "en-CA",
"href": "http://localhost:4002/processes/jsonarray2netcdf/jobs",
"rel": "http://www.opengis.net/def/rel/ogc/1.0/job-list"
},
{
"type": "application/json",
"title": "List of processes registered under the service.",
"hreflang": "en-CA",
"href": "http://localhost:4002/processes",
"rel": "up"
}
]
}</code></pre>
</details>
~If version is WPS 1.x, the only supported format is XML. This should also be the default return on `/ows/wps` route if no `version=...` parameter was provided in the URL.~
Default behaviour is to provide the corresponding XML into JSON format, regardless of version:
```http request
GET /ows/wps?service=wps&request=describeprocess&identifier=jsonarray2netcdf&f=json
```
<details>
<summary>See produced response (JSON PyWPS content)</summary>
<pre><code>{
"pywps_version": "4.5.0",
"processes": [
{
"class": "weaver.processes.wps_package:WpsPackage",
"uuid": "None",
"workdir": null,
"version": "1.1",
"identifier": "jsonarray2netcdf",
"title": "JSON array to NetCDF",
"abstract": "Extracts and fetches NetCDF files from a JSON file containing an URL string array, and provides them on the output directory.",
"keywords": [
"builtin"
],
"metadata": [],
"inputs": [
{
"identifier": "input",
"title": "input",
"abstract": "",
"keywords": [],
"metadata": [],
"type": "complex",
"data_format": {
"mime_type": "application/json",
"encoding": "",
"schema": "",
"extension": ".json"
},
"asreference": false,
"supported_formats": [
{
"mime_type": "application/json",
"encoding": "",
"schema": "",
"extension": ".json"
}
],
"workdir": null,
"mode": 1,
"min_occurs": 1,
"max_occurs": 1,
"translations": null,
"mimetype": "application/json"
}
],
"outputs": [
{
"identifier": "output",
"title": "output",
"abstract": "",
"keywords": [],
"type": "complex",
"supported_formats": [
{
"mime_type": "application/x-netcdf",
"encoding": "",
"schema": "",
"extension": ".nc"
}
],
"asreference": false,
"data_format": {
"mime_type": "application/x-netcdf",
"encoding": "",
"schema": "",
"extension": ".nc"
},
"file": null,
"workdir": null,
"mode": 1,
"min_occurs": 1,
"max_occurs": 1,
"translations": null,
"mimetype": "application/x-netcdf"
}
],
"store_supported": "true",
"status_supported": "true",
"profile": [],
"translations": null
}
],
"language": "fr-CA"
}</code></pre>
</details>
## TODO
- [x] redirect a WPS-REST (3.0) call (eg: `/processes`, `/processes/{id}`, etc.) to their corresponding WPS 2.0 `/ows/wps` request if any from (`Accept: application/xml`, `Accept: text/xml`, `f=xml`, `format=xml`) is provided (inverse of what is currently functional).
Use [dict2xml](https://pypi.org/project/dict2xml/) (or similar) to provide an XML body corresponding to the original JSON.
- [x] handle other variation `format=` as alias of `f=` on OWS endpoint.
- [x] convert the PyWPS JSON to an actual (valid) *OGC API - Processes* representation rather than plain XML->JSON conversion.
- [x] Employ this tween definition to detect Accept header / format query and raise HTTP 422 if not acceptable :
https://github.com/Ouranosinc/Magpie/blob/f67a134d52ec8f3951cb1a08a07c1c5beb518882/magpie/api/generic.py#L106-L202
|
process
|
handle accept header for json xml response in wps context given that wps supports json response we should handle a provided accept application json header on the wps route eg ows wps by default by default this route returns xml some code to do the detection of wps version format was introduced in but is still unused this is now functional http request get ows wps service wps request describeprocess identifier accept application json see returned response proper ogcapi processes content id title json array to netcdf version description extracts and fetches netcdf files from a json file containing an url string array and provides them on the output directory keywords builtin metadata inputs input title input minoccurs maxoccurs formats default true mediatype application json outputs output title output formats default true mediatype application x netcdf visibility public jobcontroloptions async execute outputtransmission reference processdescriptionurl executeendpoint links type application json title current process description hreflang en ca href rel self type application json title process definition hreflang en ca href rel process meta type application json title process execution endpoint for job submission hreflang en ca href rel type application json title list of registered processes hreflang en ca href rel type application json title list of job executions corresponding to this process hreflang en ca href rel type application json title list of processes registered under the service hreflang en ca href rel up if version is wps x the only supported format is xml this should also be the default return on ows wps route if no version parameter was provided in the url default behaviour is to provide the corresponding xml into json format regardless of version http request get ows wps service wps request describeprocess identifier f json see produced response json pywps content pywps version processes class weaver processes wps package wpspackage uuid none workdir null version identifier title json array to netcdf abstract extracts and fetches netcdf files from a json file containing an url string array and provides them on the output directory keywords builtin metadata inputs identifier input title input abstract keywords metadata type complex data format mime type application json encoding schema extension json asreference false supported formats mime type application json encoding schema extension json workdir null mode min occurs max occurs translations null mimetype application json outputs identifier output title output abstract keywords type complex supported formats mime type application x netcdf encoding schema extension nc asreference false data format mime type application x netcdf encoding schema extension nc file null workdir null mode min occurs max occurs translations null mimetype application x netcdf store supported true status supported true profile translations null language fr ca todo redirect a wps rest call eg processes processes id etc to their corresponding wps ows wps request if any from accept application xml accept text xml f xml format xml is provided inverse of what is currently functional use or similar to provide an xml body corresponding to the original json handle other variation format as alias of f on ows endpoint convert the pywps json to an actual valid ogc api processes representation rather than plain xml json conversion employ this tween definition to detect accept header format query and raise http if not acceptable
| 1
|
6,053
| 8,873,102,521
|
IssuesEvent
|
2019-01-11 17:06:24
|
18F/agile
|
https://api.github.com/repos/18F/agile
|
reopened
|
New Agile Guild lead
|
process
|
What we're after:
- So we can continue our guild leadership, we want to find a replacement for @porta-antiporta :pour-one-out:
Implementation steps:
- [x] Set up Google Form
- [x] Solicit nominations
- [x] Notify nominees of nomination, check interest in interviewing
- [x] Schedule/conduct interviews
- [x] Confer and select lead (Andrew/Porta)
- [x] Notify all nominees
- [x] Announce to TTS!
|
1.0
|
New Agile Guild lead - What we're after:
- So we can continue our guild leadership, we want to find a replacement for @porta-antiporta :pour-one-out:
Implementation steps:
- [x] Set up Google Form
- [x] Solicit nominations
- [x] Notify nominees of nomination, check interest in interviewing
- [x] Schedule/conduct interviews
- [x] Confer and select lead (Andrew/Porta)
- [x] Notify all nominees
- [x] Announce to TTS!
|
process
|
new agile guild lead what we re after so we can continue our guild leadership we want to find a replacement for porta antiporta pour one out implementation steps set up google form solicit nominations notify nominees of nomination check interest in interviewing schedule conduct interviews confer and select lead andrew porta notify all nominees announce to tts
| 1
|
224,700
| 17,770,019,925
|
IssuesEvent
|
2021-08-30 12:37:18
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
sql/catalog/lease: TestLeaseTxnDeadlineExtension failed
|
C-test-failure O-robot branch-master T-sql-schema
|
sql/catalog/lease.TestLeaseTxnDeadlineExtension [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=3374302&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=3374302&tab=artifacts#/) on master @ [392bd80db56a041099afe674c4436e0c8e3c02e6](https://github.com/cockroachdb/cockroach/commits/392bd80db56a041099afe674c4436e0c8e3c02e6):
```
github.com/cockroachdb/cockroach/pkg/sql/colflow.(*BatchFlowCoordinator).init()
/go/src/github.com/cockroachdb/cockroach/pkg/sql/colflow/flow_coordinator.go:233 +0x73
github.com/cockroachdb/cockroach/pkg/sql/colflow.(*BatchFlowCoordinator).Run()
/go/src/github.com/cockroachdb/cockroach/pkg/sql/colflow/flow_coordinator.go:274 +0x10b
github.com/cockroachdb/cockroach/pkg/sql/colflow.(*vectorizedFlow).Run()
/go/src/github.com/cockroachdb/cockroach/pkg/sql/colflow/vectorized_flow.go:257 +0x3d3
github.com/cockroachdb/cockroach/pkg/sql.(*DistSQLPlanner).Run()
/go/src/github.com/cockroachdb/cockroach/pkg/sql/distsql_running.go:535 +0x93d
github.com/cockroachdb/cockroach/pkg/sql.(*DistSQLPlanner).PlanAndRun()
/go/src/github.com/cockroachdb/cockroach/pkg/sql/distsql_running.go:1276 +0x236
github.com/cockroachdb/cockroach/pkg/sql.(*connExecutor).execWithDistSQLEngine()
/go/src/github.com/cockroachdb/cockroach/pkg/sql/conn_executor_exec.go:1326 +0xa2f
github.com/cockroachdb/cockroach/pkg/sql.(*connExecutor).dispatchToExecutionEngine()
/go/src/github.com/cockroachdb/cockroach/pkg/sql/conn_executor_exec.go:1062 +0xe25
github.com/cockroachdb/cockroach/pkg/sql.(*connExecutor).execStmtInOpenState()
/go/src/github.com/cockroachdb/cockroach/pkg/sql/conn_executor_exec.go:708 +0x16f5
github.com/cockroachdb/cockroach/pkg/sql.(*connExecutor).execStmt()
/go/src/github.com/cockroachdb/cockroach/pkg/sql/conn_executor_exec.go:129 +0x12be
github.com/cockroachdb/cockroach/pkg/sql.(*connExecutor).execCmd.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/sql/conn_executor.go:1687 +0x404
github.com/cockroachdb/cockroach/pkg/sql.(*connExecutor).execCmd()
/go/src/github.com/cockroachdb/cockroach/pkg/sql/conn_executor.go:1689 +0x1d14
github.com/cockroachdb/cockroach/pkg/sql.(*connExecutor).run()
/go/src/github.com/cockroachdb/cockroach/pkg/sql/conn_executor.go:1611 +0x565
github.com/cockroachdb/cockroach/pkg/sql.(*Server).ServeConn()
/go/src/github.com/cockroachdb/cockroach/pkg/sql/conn_executor.go:663 +0x134
github.com/cockroachdb/cockroach/pkg/sql/pgwire.(*conn).processCommandsAsync.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/sql/pgwire/conn.go:647 +0x3f2
==================
I210828 06:55:33.570200 65610 sql/table.go:159 [n1,client=127.0.0.1:34730,hostssl,user=root] 2965 queued new schema-change job 688335531442667521 for table 52, mutation 0
I210828 06:55:33.664387 65610 5@util/log/event_log.go:32 [n1,client=127.0.0.1:34730,hostssl,user=root] 2966 ={"Timestamp":1630133733562906239,"EventType":"alter_table","Statement":"ALTER TABLE defaultdb.public.t1 ALTER COLUMN val SET DEFAULT 5","Tag":"ALTER TABLE","User":"root","DescriptorID":52,"TableName":"defaultdb.public.t1"}
I210828 06:55:33.667417 65610 jobs/registry.go:345 [n1,client=127.0.0.1:34730,hostssl,user=root] 2967 scheduled jobs [688335531442667521]
I210828 06:55:33.680925 65633 sql/catalog/lease/descriptor_state.go:139 [n1] 2968 new lease: tabledesc.immutable: {ID: 52, Version: 2, ModificationTime: "1630133733.562906239,0", ParentID: 50, ParentSchemaID: 29, State: PUBLIC, NextColumnID: 3, Columns: [{ID: 1, TypeID: 20, Null: true, HasDefault: true}, {ID: 2, TypeID: 20, Null: false, Hidden: true, HasDefault: true}], NextFamilyID: 1, Families: [{ID: 0, Columns: [1, 2]}], PrimaryIndex: 1, NextIndexID: 2, Indexes: [{ID: 1, Unique: true, KeyColumns: [{ID: 2, Dir: ASC}], StoreColumns: [1]}]}
W210828 06:55:33.682774 64060 sql/catalog/lease/lease.go:921 [n1] 2969 error purging leases for descriptor 52(t1): renew lease on id
I210828 06:55:33.739890 65767 jobs/adopt.go:247 [-] 2970 job 688335531442667521: resuming execution
I210828 06:55:33.821111 65743 jobs/registry.go:1149 [n1] 2971 SCHEMA CHANGE job 688335531442667521: stepping through state running with error: <nil>
I210828 06:55:33.887983 65743 sql/schema_changer.go:586 [n1,job=688335531442667521,scExec,id=52] 2972 schema change on "t1" (v2) starting execution...
I210828 06:55:33.888412 65743 sql/schema_changer.go:1025 [n1,job=688335531442667521,scExec,id=52] 2973 waiting for a single version...
I210828 06:55:33.910069 65743 sql/schema_changer.go:1031 [n1,job=688335531442667521,scExec,id=52] 2974 waiting for a single version... done (at v 2)
I210828 06:55:33.910397 65743 jobs/registry.go:1149 [n1] 2975 SCHEMA CHANGE job 688335531442667521: stepping through state succeeded with error: <nil>
testing.go:1092: race detected during execution of test
I210828 06:55:34.086245 63639 testutils/testcluster/testcluster.go:101 [-] 2976 TestCluster quiescing nodes
I210828 06:55:34.087191 63735 kv/kvclient/kvcoord/transport_race.go:106 [-] 2977 transport race promotion: ran 27 iterations on up to 1999 requests
W210828 06:55:34.087462 65572 kv/kvserver/intentresolver/intent_resolver.go:790 [-] 2978 failed to gc transaction record: could not GC completed transaction anchored at /Local/Range/Table/47/RangeDescriptor: node unavailable; try another peer
W210828 06:55:34.089130 65869 sql/catalog/lease/storage.go:180 [-] 2979 error releasing lease "ID = 52 ver=2 expiration={2021-08-28 06:55:34.043825 +0000 UTC}": lease-release: failed to send RPC: sending to all replicas failed; last error: node unavailable; try another peer
W210828 06:55:34.089423 64062 jobs/registry.go:821 [-] 2981 canceling all adopted jobs due to stopper quiescing
E210828 06:55:34.089187 65868 kv/kvclient/kvcoord/txn_interceptor_committer.go:479 [-] 2980 making txn commit explicit failed for "unnamed" meta={id=60093a92 pri=100.00000000 epo=0 ts=1630133734.064565971,0 min=1630133734.064565971,0 seq=2} lock=true stat=STAGING rts=1630133734.064565971,0 wto=false gul=1630133734.564565971,0 ifw=1: result is ambiguous (server shutdown)
W210828 06:55:34.092745 63900 sql/sqlliveness/slinstance/slinstance.go:217 [n1] 2982 exiting heartbeat loop
W210828 06:55:34.093355 65841 sql/catalog/lease/storage.go:180 [-] 2983 error releasing lease "ID = 52 ver=1 expiration={2021-08-28 06:55:33.522339 +0000 UTC}": lease-release: failed to send RPC: sending to all replicas failed; last error: node unavailable; try another peer
--- FAIL: TestLeaseTxnDeadlineExtension/validate-lease-txn-deadline-ext-blocked (0.60s)
```
<details><summary>Reproduce</summary>
<p>
To reproduce, try:
```bash
make stressrace TESTS=TestLeaseTxnDeadlineExtension PKG=./pkg/sql/catalog/lease TESTTIMEOUT=5m STRESSFLAGS='-timeout 5m' 2>&1
```
Parameters in this failure:
- GOFLAGS=-race -parallel=4
</p>
</details>
/cc @cockroachdb/sql-schema fqazi
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestLeaseTxnDeadlineExtension.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
1.0
|
sql/catalog/lease: TestLeaseTxnDeadlineExtension failed - sql/catalog/lease.TestLeaseTxnDeadlineExtension [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=3374302&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=3374302&tab=artifacts#/) on master @ [392bd80db56a041099afe674c4436e0c8e3c02e6](https://github.com/cockroachdb/cockroach/commits/392bd80db56a041099afe674c4436e0c8e3c02e6):
```
github.com/cockroachdb/cockroach/pkg/sql/colflow.(*BatchFlowCoordinator).init()
/go/src/github.com/cockroachdb/cockroach/pkg/sql/colflow/flow_coordinator.go:233 +0x73
github.com/cockroachdb/cockroach/pkg/sql/colflow.(*BatchFlowCoordinator).Run()
/go/src/github.com/cockroachdb/cockroach/pkg/sql/colflow/flow_coordinator.go:274 +0x10b
github.com/cockroachdb/cockroach/pkg/sql/colflow.(*vectorizedFlow).Run()
/go/src/github.com/cockroachdb/cockroach/pkg/sql/colflow/vectorized_flow.go:257 +0x3d3
github.com/cockroachdb/cockroach/pkg/sql.(*DistSQLPlanner).Run()
/go/src/github.com/cockroachdb/cockroach/pkg/sql/distsql_running.go:535 +0x93d
github.com/cockroachdb/cockroach/pkg/sql.(*DistSQLPlanner).PlanAndRun()
/go/src/github.com/cockroachdb/cockroach/pkg/sql/distsql_running.go:1276 +0x236
github.com/cockroachdb/cockroach/pkg/sql.(*connExecutor).execWithDistSQLEngine()
/go/src/github.com/cockroachdb/cockroach/pkg/sql/conn_executor_exec.go:1326 +0xa2f
github.com/cockroachdb/cockroach/pkg/sql.(*connExecutor).dispatchToExecutionEngine()
/go/src/github.com/cockroachdb/cockroach/pkg/sql/conn_executor_exec.go:1062 +0xe25
github.com/cockroachdb/cockroach/pkg/sql.(*connExecutor).execStmtInOpenState()
/go/src/github.com/cockroachdb/cockroach/pkg/sql/conn_executor_exec.go:708 +0x16f5
github.com/cockroachdb/cockroach/pkg/sql.(*connExecutor).execStmt()
/go/src/github.com/cockroachdb/cockroach/pkg/sql/conn_executor_exec.go:129 +0x12be
github.com/cockroachdb/cockroach/pkg/sql.(*connExecutor).execCmd.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/sql/conn_executor.go:1687 +0x404
github.com/cockroachdb/cockroach/pkg/sql.(*connExecutor).execCmd()
/go/src/github.com/cockroachdb/cockroach/pkg/sql/conn_executor.go:1689 +0x1d14
github.com/cockroachdb/cockroach/pkg/sql.(*connExecutor).run()
/go/src/github.com/cockroachdb/cockroach/pkg/sql/conn_executor.go:1611 +0x565
github.com/cockroachdb/cockroach/pkg/sql.(*Server).ServeConn()
/go/src/github.com/cockroachdb/cockroach/pkg/sql/conn_executor.go:663 +0x134
github.com/cockroachdb/cockroach/pkg/sql/pgwire.(*conn).processCommandsAsync.func1()
/go/src/github.com/cockroachdb/cockroach/pkg/sql/pgwire/conn.go:647 +0x3f2
==================
I210828 06:55:33.570200 65610 sql/table.go:159 [n1,client=127.0.0.1:34730,hostssl,user=root] 2965 queued new schema-change job 688335531442667521 for table 52, mutation 0
I210828 06:55:33.664387 65610 5@util/log/event_log.go:32 [n1,client=127.0.0.1:34730,hostssl,user=root] 2966 ={"Timestamp":1630133733562906239,"EventType":"alter_table","Statement":"ALTER TABLE defaultdb.public.t1 ALTER COLUMN val SET DEFAULT 5","Tag":"ALTER TABLE","User":"root","DescriptorID":52,"TableName":"defaultdb.public.t1"}
I210828 06:55:33.667417 65610 jobs/registry.go:345 [n1,client=127.0.0.1:34730,hostssl,user=root] 2967 scheduled jobs [688335531442667521]
I210828 06:55:33.680925 65633 sql/catalog/lease/descriptor_state.go:139 [n1] 2968 new lease: tabledesc.immutable: {ID: 52, Version: 2, ModificationTime: "1630133733.562906239,0", ParentID: 50, ParentSchemaID: 29, State: PUBLIC, NextColumnID: 3, Columns: [{ID: 1, TypeID: 20, Null: true, HasDefault: true}, {ID: 2, TypeID: 20, Null: false, Hidden: true, HasDefault: true}], NextFamilyID: 1, Families: [{ID: 0, Columns: [1, 2]}], PrimaryIndex: 1, NextIndexID: 2, Indexes: [{ID: 1, Unique: true, KeyColumns: [{ID: 2, Dir: ASC}], StoreColumns: [1]}]}
W210828 06:55:33.682774 64060 sql/catalog/lease/lease.go:921 [n1] 2969 error purging leases for descriptor 52(t1): renew lease on id
I210828 06:55:33.739890 65767 jobs/adopt.go:247 [-] 2970 job 688335531442667521: resuming execution
I210828 06:55:33.821111 65743 jobs/registry.go:1149 [n1] 2971 SCHEMA CHANGE job 688335531442667521: stepping through state running with error: <nil>
I210828 06:55:33.887983 65743 sql/schema_changer.go:586 [n1,job=688335531442667521,scExec,id=52] 2972 schema change on "t1" (v2) starting execution...
I210828 06:55:33.888412 65743 sql/schema_changer.go:1025 [n1,job=688335531442667521,scExec,id=52] 2973 waiting for a single version...
I210828 06:55:33.910069 65743 sql/schema_changer.go:1031 [n1,job=688335531442667521,scExec,id=52] 2974 waiting for a single version... done (at v 2)
I210828 06:55:33.910397 65743 jobs/registry.go:1149 [n1] 2975 SCHEMA CHANGE job 688335531442667521: stepping through state succeeded with error: <nil>
testing.go:1092: race detected during execution of test
I210828 06:55:34.086245 63639 testutils/testcluster/testcluster.go:101 [-] 2976 TestCluster quiescing nodes
I210828 06:55:34.087191 63735 kv/kvclient/kvcoord/transport_race.go:106 [-] 2977 transport race promotion: ran 27 iterations on up to 1999 requests
W210828 06:55:34.087462 65572 kv/kvserver/intentresolver/intent_resolver.go:790 [-] 2978 failed to gc transaction record: could not GC completed transaction anchored at /Local/Range/Table/47/RangeDescriptor: node unavailable; try another peer
W210828 06:55:34.089130 65869 sql/catalog/lease/storage.go:180 [-] 2979 error releasing lease "ID = 52 ver=2 expiration={2021-08-28 06:55:34.043825 +0000 UTC}": lease-release: failed to send RPC: sending to all replicas failed; last error: node unavailable; try another peer
W210828 06:55:34.089423 64062 jobs/registry.go:821 [-] 2981 canceling all adopted jobs due to stopper quiescing
E210828 06:55:34.089187 65868 kv/kvclient/kvcoord/txn_interceptor_committer.go:479 [-] 2980 making txn commit explicit failed for "unnamed" meta={id=60093a92 pri=100.00000000 epo=0 ts=1630133734.064565971,0 min=1630133734.064565971,0 seq=2} lock=true stat=STAGING rts=1630133734.064565971,0 wto=false gul=1630133734.564565971,0 ifw=1: result is ambiguous (server shutdown)
W210828 06:55:34.092745 63900 sql/sqlliveness/slinstance/slinstance.go:217 [n1] 2982 exiting heartbeat loop
W210828 06:55:34.093355 65841 sql/catalog/lease/storage.go:180 [-] 2983 error releasing lease "ID = 52 ver=1 expiration={2021-08-28 06:55:33.522339 +0000 UTC}": lease-release: failed to send RPC: sending to all replicas failed; last error: node unavailable; try another peer
--- FAIL: TestLeaseTxnDeadlineExtension/validate-lease-txn-deadline-ext-blocked (0.60s)
```
<details><summary>Reproduce</summary>
<p>
To reproduce, try:
```bash
make stressrace TESTS=TestLeaseTxnDeadlineExtension PKG=./pkg/sql/catalog/lease TESTTIMEOUT=5m STRESSFLAGS='-timeout 5m' 2>&1
```
Parameters in this failure:
- GOFLAGS=-race -parallel=4
</p>
</details>
/cc @cockroachdb/sql-schema fqazi
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestLeaseTxnDeadlineExtension.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
non_process
|
sql catalog lease testleasetxndeadlineextension failed sql catalog lease testleasetxndeadlineextension with on master github com cockroachdb cockroach pkg sql colflow batchflowcoordinator init go src github com cockroachdb cockroach pkg sql colflow flow coordinator go github com cockroachdb cockroach pkg sql colflow batchflowcoordinator run go src github com cockroachdb cockroach pkg sql colflow flow coordinator go github com cockroachdb cockroach pkg sql colflow vectorizedflow run go src github com cockroachdb cockroach pkg sql colflow vectorized flow go github com cockroachdb cockroach pkg sql distsqlplanner run go src github com cockroachdb cockroach pkg sql distsql running go github com cockroachdb cockroach pkg sql distsqlplanner planandrun go src github com cockroachdb cockroach pkg sql distsql running go github com cockroachdb cockroach pkg sql connexecutor execwithdistsqlengine go src github com cockroachdb cockroach pkg sql conn executor exec go github com cockroachdb cockroach pkg sql connexecutor dispatchtoexecutionengine go src github com cockroachdb cockroach pkg sql conn executor exec go github com cockroachdb cockroach pkg sql connexecutor execstmtinopenstate go src github com cockroachdb cockroach pkg sql conn executor exec go github com cockroachdb cockroach pkg sql connexecutor execstmt go src github com cockroachdb cockroach pkg sql conn executor exec go github com cockroachdb cockroach pkg sql connexecutor execcmd go src github com cockroachdb cockroach pkg sql conn executor go github com cockroachdb cockroach pkg sql connexecutor execcmd go src github com cockroachdb cockroach pkg sql conn executor go github com cockroachdb cockroach pkg sql connexecutor run go src github com cockroachdb cockroach pkg sql conn executor go github com cockroachdb cockroach pkg sql server serveconn go src github com cockroachdb cockroach pkg sql conn executor go github com cockroachdb cockroach pkg sql pgwire conn processcommandsasync go src github com cockroachdb cockroach pkg sql pgwire conn go sql table go queued new schema change job for table mutation util log event log go timestamp eventtype alter table statement alter table defaultdb public alter column val set default tag alter table user root descriptorid tablename defaultdb public jobs registry go scheduled jobs sql catalog lease descriptor state go new lease tabledesc immutable id version modificationtime parentid parentschemaid state public nextcolumnid columns nextfamilyid families primaryindex nextindexid indexes storecolumns sql catalog lease lease go error purging leases for descriptor renew lease on id jobs adopt go job resuming execution jobs registry go schema change job stepping through state running with error sql schema changer go schema change on starting execution sql schema changer go waiting for a single version sql schema changer go waiting for a single version done at v jobs registry go schema change job stepping through state succeeded with error testing go race detected during execution of test testutils testcluster testcluster go testcluster quiescing nodes kv kvclient kvcoord transport race go transport race promotion ran iterations on up to requests kv kvserver intentresolver intent resolver go failed to gc transaction record could not gc completed transaction anchored at local range table rangedescriptor node unavailable try another peer sql catalog lease storage go error releasing lease id ver expiration utc lease release failed to send rpc sending to all replicas failed last error node unavailable try another peer jobs registry go canceling all adopted jobs due to stopper quiescing kv kvclient kvcoord txn interceptor committer go making txn commit explicit failed for unnamed meta id pri epo ts min seq lock true stat staging rts wto false gul ifw result is ambiguous server shutdown sql sqlliveness slinstance slinstance go exiting heartbeat loop sql catalog lease storage go error releasing lease id ver expiration utc lease release failed to send rpc sending to all replicas failed last error node unavailable try another peer fail testleasetxndeadlineextension validate lease txn deadline ext blocked reproduce to reproduce try bash make stressrace tests testleasetxndeadlineextension pkg pkg sql catalog lease testtimeout stressflags timeout parameters in this failure goflags race parallel cc cockroachdb sql schema fqazi
| 0
|
7,166
| 10,311,320,382
|
IssuesEvent
|
2019-08-29 17:04:36
|
material-components/material-components-ios
|
https://api.github.com/repos/material-components/material-components-ios
|
closed
|
[Banner] Make Banner be a Ready component
|
[Banner] type:Process
|
This was filed as an internal issue. If you are a Googler, please visit [b/118201071](http://b/118201071) for more details.
<!-- Auto-generated content below, do not modify -->
---
#### Internal data
- Associated internal bug: [b/118201071](http://b/118201071)
- Blocked by: https://github.com/material-components/material-components-ios/issues/7150
- Blocked by: https://github.com/material-components/material-components-ios/issues/5488
|
1.0
|
[Banner] Make Banner be a Ready component - This was filed as an internal issue. If you are a Googler, please visit [b/118201071](http://b/118201071) for more details.
<!-- Auto-generated content below, do not modify -->
---
#### Internal data
- Associated internal bug: [b/118201071](http://b/118201071)
- Blocked by: https://github.com/material-components/material-components-ios/issues/7150
- Blocked by: https://github.com/material-components/material-components-ios/issues/5488
|
process
|
make banner be a ready component this was filed as an internal issue if you are a googler please visit for more details internal data associated internal bug blocked by blocked by
| 1
|
121,808
| 26,035,798,383
|
IssuesEvent
|
2022-12-22 04:46:01
|
dotnet/runtime
|
https://api.github.com/repos/dotnet/runtime
|
closed
|
LSRA Reg Optional: Folding of read-modify-write mem operations
|
enhancement tenet-performance area-CodeGen-coreclr optimization
|
E.g. a = a + b/cns
If it is profitable to not to allocate a reg to 'a' - i.e. both to its Use and Def position, then we can generate
add [mem], reg/cns
category:cq
theme:register-allocator
skill-level:intermediate
cost:medium
|
1.0
|
LSRA Reg Optional: Folding of read-modify-write mem operations - E.g. a = a + b/cns
If it is profitable to not to allocate a reg to 'a' - i.e. both to its Use and Def position, then we can generate
add [mem], reg/cns
category:cq
theme:register-allocator
skill-level:intermediate
cost:medium
|
non_process
|
lsra reg optional folding of read modify write mem operations e g a a b cns if it is profitable to not to allocate a reg to a i e both to its use and def position then we can generate add reg cns category cq theme register allocator skill level intermediate cost medium
| 0
|
263,093
| 8,273,299,647
|
IssuesEvent
|
2018-09-17 04:51:01
|
minishift/minishift
|
https://api.github.com/repos/minishift/minishift
|
closed
|
Remove CI for b2d and minikube as we are not maintaining
|
kind/question priority/major
|
I think its time to completely remove the CI parts related to b2d and minikube ISO as it is creating more confusion than proving useful.
The one which we care about i.e CentOS iso nightly is working fine https://ci.centos.org/job/minishift-nightly-centos/.
|
1.0
|
Remove CI for b2d and minikube as we are not maintaining - I think its time to completely remove the CI parts related to b2d and minikube ISO as it is creating more confusion than proving useful.
The one which we care about i.e CentOS iso nightly is working fine https://ci.centos.org/job/minishift-nightly-centos/.
|
non_process
|
remove ci for and minikube as we are not maintaining i think its time to completely remove the ci parts related to and minikube iso as it is creating more confusion than proving useful the one which we care about i e centos iso nightly is working fine
| 0
|
12,170
| 14,741,827,576
|
IssuesEvent
|
2021-01-07 11:14:17
|
kdjstudios/SABillingGitlab
|
https://api.github.com/repos/kdjstudios/SABillingGitlab
|
closed
|
Revert Billing Cycle Improvements
|
anc-process anp-not prioritized ant-enhancement
|
In GitLab by @kdjstudios on Feb 13, 2019, 10:44
**Submitted by:** Kyle
**Helpdesk:** NA
**Server:** All
**Client/Site:** ALL
**Account:** All
**Issue:**
Hello, as pointed out in #1373 that when a site needs to revert a billing cycle not always to they want to delete the manual invoices that are apart of that billing cycle. I think we need to discuss a way to address this.
Possible solution:
- Provide a prompt to the user after clicking the revert button "Would you like to remove just the invoices created during the billing cycle process; or would you like to delete all invoices including any manual invoices?"
- If select "Billing Cycle Invoices"; then we simply undo the invoices created during the apply step.
- If select "All Invoices; then we revert as it currently stands and every invoices is removed.
- Also we would now need to have the "revert" button visable on the 'upload' step too. This way a user can revert all manual invoices too.
|
1.0
|
Revert Billing Cycle Improvements - In GitLab by @kdjstudios on Feb 13, 2019, 10:44
**Submitted by:** Kyle
**Helpdesk:** NA
**Server:** All
**Client/Site:** ALL
**Account:** All
**Issue:**
Hello, as pointed out in #1373 that when a site needs to revert a billing cycle not always to they want to delete the manual invoices that are apart of that billing cycle. I think we need to discuss a way to address this.
Possible solution:
- Provide a prompt to the user after clicking the revert button "Would you like to remove just the invoices created during the billing cycle process; or would you like to delete all invoices including any manual invoices?"
- If select "Billing Cycle Invoices"; then we simply undo the invoices created during the apply step.
- If select "All Invoices; then we revert as it currently stands and every invoices is removed.
- Also we would now need to have the "revert" button visable on the 'upload' step too. This way a user can revert all manual invoices too.
|
process
|
revert billing cycle improvements in gitlab by kdjstudios on feb submitted by kyle helpdesk na server all client site all account all issue hello as pointed out in that when a site needs to revert a billing cycle not always to they want to delete the manual invoices that are apart of that billing cycle i think we need to discuss a way to address this possible solution provide a prompt to the user after clicking the revert button would you like to remove just the invoices created during the billing cycle process or would you like to delete all invoices including any manual invoices if select billing cycle invoices then we simply undo the invoices created during the apply step if select all invoices then we revert as it currently stands and every invoices is removed also we would now need to have the revert button visable on the upload step too this way a user can revert all manual invoices too
| 1
|
186,098
| 14,394,638,097
|
IssuesEvent
|
2020-12-03 01:46:12
|
github-vet/rangeclosure-findings
|
https://api.github.com/repos/github-vet/rangeclosure-findings
|
closed
|
evergreen-ci/poplar: vendor/google.golang.org/grpc/credentials/alts/internal/conn/aes128gcm_test.go; 74 LoC
|
fresh medium test vendored
|
Found a possible issue in [evergreen-ci/poplar](https://www.github.com/evergreen-ci/poplar) at [vendor/google.golang.org/grpc/credentials/alts/internal/conn/aes128gcm_test.go](https://github.com/evergreen-ci/poplar/blob/5e03bb464d891b4578bd0c41deb4394147f7b8ce/vendor/google.golang.org/grpc/credentials/alts/internal/conn/aes128gcm_test.go#L85-L158)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> function call which takes a reference to test at line 154 may start a goroutine
[Click here to see the code in its original context.](https://github.com/evergreen-ci/poplar/blob/5e03bb464d891b4578bd0c41deb4394147f7b8ce/vendor/google.golang.org/grpc/credentials/alts/internal/conn/aes128gcm_test.go#L85-L158)
<details>
<summary>Click here to show the 74 line(s) of Go which triggered the analyzer.</summary>
```go
for _, test := range []cryptoTestVector{
{
key: dehex("11754cd72aec309bf52f7687212e8957"),
counter: dehex("3c819d9a9bed087615030b65"),
plaintext: nil,
ciphertext: nil,
tag: dehex("250327c674aaf477aef2675748cf6971"),
allocateDst: false,
},
{
key: dehex("ca47248ac0b6f8372a97ac43508308ed"),
counter: dehex("ffd2b598feabc9019262d2be"),
plaintext: nil,
ciphertext: nil,
tag: dehex("60d20404af527d248d893ae495707d1a"),
allocateDst: false,
},
{
key: dehex("7fddb57453c241d03efbed3ac44e371c"),
counter: dehex("ee283a3fc75575e33efd4887"),
plaintext: dehex("d5de42b461646c255c87bd2962d3b9a2"),
ciphertext: dehex("2ccda4a5415cb91e135c2a0f78c9b2fd"),
tag: dehex("b36d1df9b9d5e596f83e8b7f52971cb3"),
allocateDst: false,
},
{
key: dehex("ab72c77b97cb5fe9a382d9fe81ffdbed"),
counter: dehex("54cc7dc2c37ec006bcc6d1da"),
plaintext: dehex("007c5e5b3e59df24a7c355584fc1518d"),
ciphertext: dehex("0e1bde206a07a9c2c1b65300f8c64997"),
tag: dehex("2b4401346697138c7a4891ee59867d0c"),
allocateDst: false,
},
{
key: dehex("11754cd72aec309bf52f7687212e8957"),
counter: dehex("3c819d9a9bed087615030b65"),
plaintext: nil,
ciphertext: nil,
tag: dehex("250327c674aaf477aef2675748cf6971"),
allocateDst: true,
},
{
key: dehex("ca47248ac0b6f8372a97ac43508308ed"),
counter: dehex("ffd2b598feabc9019262d2be"),
plaintext: nil,
ciphertext: nil,
tag: dehex("60d20404af527d248d893ae495707d1a"),
allocateDst: true,
},
{
key: dehex("7fddb57453c241d03efbed3ac44e371c"),
counter: dehex("ee283a3fc75575e33efd4887"),
plaintext: dehex("d5de42b461646c255c87bd2962d3b9a2"),
ciphertext: dehex("2ccda4a5415cb91e135c2a0f78c9b2fd"),
tag: dehex("b36d1df9b9d5e596f83e8b7f52971cb3"),
allocateDst: true,
},
{
key: dehex("ab72c77b97cb5fe9a382d9fe81ffdbed"),
counter: dehex("54cc7dc2c37ec006bcc6d1da"),
plaintext: dehex("007c5e5b3e59df24a7c355584fc1518d"),
ciphertext: dehex("0e1bde206a07a9c2c1b65300f8c64997"),
tag: dehex("2b4401346697138c7a4891ee59867d0c"),
allocateDst: true,
},
} {
// Test encryption and decryption for aes128gcm.
client, server := getGCMCryptoPair(test.key, test.counter, t)
if CounterSide(test.counter) == core.ClientSide {
testGCMEncryptionDecryption(client, server, &test, false, t)
} else {
testGCMEncryptionDecryption(server, client, &test, false, t)
}
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 5e03bb464d891b4578bd0c41deb4394147f7b8ce
|
1.0
|
evergreen-ci/poplar: vendor/google.golang.org/grpc/credentials/alts/internal/conn/aes128gcm_test.go; 74 LoC -
Found a possible issue in [evergreen-ci/poplar](https://www.github.com/evergreen-ci/poplar) at [vendor/google.golang.org/grpc/credentials/alts/internal/conn/aes128gcm_test.go](https://github.com/evergreen-ci/poplar/blob/5e03bb464d891b4578bd0c41deb4394147f7b8ce/vendor/google.golang.org/grpc/credentials/alts/internal/conn/aes128gcm_test.go#L85-L158)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> function call which takes a reference to test at line 154 may start a goroutine
[Click here to see the code in its original context.](https://github.com/evergreen-ci/poplar/blob/5e03bb464d891b4578bd0c41deb4394147f7b8ce/vendor/google.golang.org/grpc/credentials/alts/internal/conn/aes128gcm_test.go#L85-L158)
<details>
<summary>Click here to show the 74 line(s) of Go which triggered the analyzer.</summary>
```go
for _, test := range []cryptoTestVector{
{
key: dehex("11754cd72aec309bf52f7687212e8957"),
counter: dehex("3c819d9a9bed087615030b65"),
plaintext: nil,
ciphertext: nil,
tag: dehex("250327c674aaf477aef2675748cf6971"),
allocateDst: false,
},
{
key: dehex("ca47248ac0b6f8372a97ac43508308ed"),
counter: dehex("ffd2b598feabc9019262d2be"),
plaintext: nil,
ciphertext: nil,
tag: dehex("60d20404af527d248d893ae495707d1a"),
allocateDst: false,
},
{
key: dehex("7fddb57453c241d03efbed3ac44e371c"),
counter: dehex("ee283a3fc75575e33efd4887"),
plaintext: dehex("d5de42b461646c255c87bd2962d3b9a2"),
ciphertext: dehex("2ccda4a5415cb91e135c2a0f78c9b2fd"),
tag: dehex("b36d1df9b9d5e596f83e8b7f52971cb3"),
allocateDst: false,
},
{
key: dehex("ab72c77b97cb5fe9a382d9fe81ffdbed"),
counter: dehex("54cc7dc2c37ec006bcc6d1da"),
plaintext: dehex("007c5e5b3e59df24a7c355584fc1518d"),
ciphertext: dehex("0e1bde206a07a9c2c1b65300f8c64997"),
tag: dehex("2b4401346697138c7a4891ee59867d0c"),
allocateDst: false,
},
{
key: dehex("11754cd72aec309bf52f7687212e8957"),
counter: dehex("3c819d9a9bed087615030b65"),
plaintext: nil,
ciphertext: nil,
tag: dehex("250327c674aaf477aef2675748cf6971"),
allocateDst: true,
},
{
key: dehex("ca47248ac0b6f8372a97ac43508308ed"),
counter: dehex("ffd2b598feabc9019262d2be"),
plaintext: nil,
ciphertext: nil,
tag: dehex("60d20404af527d248d893ae495707d1a"),
allocateDst: true,
},
{
key: dehex("7fddb57453c241d03efbed3ac44e371c"),
counter: dehex("ee283a3fc75575e33efd4887"),
plaintext: dehex("d5de42b461646c255c87bd2962d3b9a2"),
ciphertext: dehex("2ccda4a5415cb91e135c2a0f78c9b2fd"),
tag: dehex("b36d1df9b9d5e596f83e8b7f52971cb3"),
allocateDst: true,
},
{
key: dehex("ab72c77b97cb5fe9a382d9fe81ffdbed"),
counter: dehex("54cc7dc2c37ec006bcc6d1da"),
plaintext: dehex("007c5e5b3e59df24a7c355584fc1518d"),
ciphertext: dehex("0e1bde206a07a9c2c1b65300f8c64997"),
tag: dehex("2b4401346697138c7a4891ee59867d0c"),
allocateDst: true,
},
} {
// Test encryption and decryption for aes128gcm.
client, server := getGCMCryptoPair(test.key, test.counter, t)
if CounterSide(test.counter) == core.ClientSide {
testGCMEncryptionDecryption(client, server, &test, false, t)
} else {
testGCMEncryptionDecryption(server, client, &test, false, t)
}
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 5e03bb464d891b4578bd0c41deb4394147f7b8ce
|
non_process
|
evergreen ci poplar vendor google golang org grpc credentials alts internal conn test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message function call which takes a reference to test at line may start a goroutine click here to show the line s of go which triggered the analyzer go for test range cryptotestvector key dehex counter dehex plaintext nil ciphertext nil tag dehex allocatedst false key dehex counter dehex plaintext nil ciphertext nil tag dehex allocatedst false key dehex counter dehex plaintext dehex ciphertext dehex tag dehex allocatedst false key dehex counter dehex plaintext dehex ciphertext dehex tag dehex allocatedst false key dehex counter dehex plaintext nil ciphertext nil tag dehex allocatedst true key dehex counter dehex plaintext nil ciphertext nil tag dehex allocatedst true key dehex counter dehex plaintext dehex ciphertext dehex tag dehex allocatedst true key dehex counter dehex plaintext dehex ciphertext dehex tag dehex allocatedst true test encryption and decryption for client server getgcmcryptopair test key test counter t if counterside test counter core clientside testgcmencryptiondecryption client server test false t else testgcmencryptiondecryption server client test false t leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
| 0
|
16,168
| 20,605,072,275
|
IssuesEvent
|
2022-03-06 21:13:27
|
hoprnet/hoprnet
|
https://api.github.com/repos/hoprnet/hoprnet
|
opened
|
Update development processes
|
new issue processes
|
<!--- Please DO NOT remove the automatically added 'new issue' label -->
<!--- Provide a general summary of the issue in the Title above -->
After our call on 07/03/22 the new development processes were announced, update the development processes.
Updates: TBA
|
1.0
|
Update development processes - <!--- Please DO NOT remove the automatically added 'new issue' label -->
<!--- Provide a general summary of the issue in the Title above -->
After our call on 07/03/22 the new development processes were announced, update the development processes.
Updates: TBA
|
process
|
update development processes after our call on the new development processes were announced update the development processes updates tba
| 1
|
13,373
| 15,835,698,987
|
IssuesEvent
|
2021-04-06 18:21:06
|
EKGF/ekg-mm
|
https://api.github.com/repos/EKGF/ekg-mm
|
closed
|
Copy the generated PDF to Google Drive
|
ekg-mm-process
|
For easy access, it would be good to have the latest version of whichever PDF is generated, from any branch to the appropriate folders in Google Drive.
PS: This is NOT meant to be a solution for publishing it on the https://ekgf.org website. Ideally we can use the WiX APIs to do that.
@DennisWisnosky @avaddict @MikeAtkin @rivettp
|
1.0
|
Copy the generated PDF to Google Drive - For easy access, it would be good to have the latest version of whichever PDF is generated, from any branch to the appropriate folders in Google Drive.
PS: This is NOT meant to be a solution for publishing it on the https://ekgf.org website. Ideally we can use the WiX APIs to do that.
@DennisWisnosky @avaddict @MikeAtkin @rivettp
|
process
|
copy the generated pdf to google drive for easy access it would be good to have the latest version of whichever pdf is generated from any branch to the appropriate folders in google drive ps this is not meant to be a solution for publishing it on the website ideally we can use the wix apis to do that denniswisnosky avaddict mikeatkin rivettp
| 1
|
13,997
| 16,770,856,762
|
IssuesEvent
|
2021-06-14 14:38:29
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[iOS] Custom consent's visual steps are not shown in iOS app while 'Consent Review' process.
|
Bug P1 Process: Fixed Process: Tested dev iOS
|
**Steps:**
1. Mobile app participant is enrolled into the study
2. Configure a custom consent in Study builder with 'Visual Steps' enabled
3. Publish updates
4. Referesh the screen in iOS mobile
5. Click on 'Review' button in 'Consent Updated' pop-up
6. Observe the screens
**Actual:** Directly navigating to Review consent screen
**Expected:** Visual steps should be shown if enabled
**Note:**
- Issue not observed while enrolling into the study
- Issue not observed for RK/RS default consent sections i.e if only custom section is enabled, visual steps are not shown.
- Issue not observed in Android
|
2.0
|
[iOS] Custom consent's visual steps are not shown in iOS app while 'Consent Review' process. - **Steps:**
1. Mobile app participant is enrolled into the study
2. Configure a custom consent in Study builder with 'Visual Steps' enabled
3. Publish updates
4. Referesh the screen in iOS mobile
5. Click on 'Review' button in 'Consent Updated' pop-up
6. Observe the screens
**Actual:** Directly navigating to Review consent screen
**Expected:** Visual steps should be shown if enabled
**Note:**
- Issue not observed while enrolling into the study
- Issue not observed for RK/RS default consent sections i.e if only custom section is enabled, visual steps are not shown.
- Issue not observed in Android
|
process
|
custom consent s visual steps are not shown in ios app while consent review process steps mobile app participant is enrolled into the study configure a custom consent in study builder with visual steps enabled publish updates referesh the screen in ios mobile click on review button in consent updated pop up observe the screens actual directly navigating to review consent screen expected visual steps should be shown if enabled note issue not observed while enrolling into the study issue not observed for rk rs default consent sections i e if only custom section is enabled visual steps are not shown issue not observed in android
| 1
|
16,300
| 20,955,051,818
|
IssuesEvent
|
2022-03-27 01:48:56
|
fmnas/fmnas-site
|
https://api.github.com/repos/fmnas/fmnas-site
|
opened
|
Spambots/fuzzers can break application and create "failed applications"
|
backend form processor medium (3-8h)
|
Currently the only way this happened is with an invalid email address
|
1.0
|
Spambots/fuzzers can break application and create "failed applications" - Currently the only way this happened is with an invalid email address
|
process
|
spambots fuzzers can break application and create failed applications currently the only way this happened is with an invalid email address
| 1
|
10,002
| 13,042,438,786
|
IssuesEvent
|
2020-07-28 22:30:36
|
hashicorp/packer
|
https://api.github.com/repos/hashicorp/packer
|
closed
|
abort error on vmware-iso upload in packer 0.12
|
bug post-processor/vsphere
|
packer 0.12.0
After updating to packer 0.12.0 when trying to upload to vcenter using the vsphere postprocessor I now get the following error:
```
2016/11/17 13:33:54 ui: ==> vmware-iso: Running post-processor: vsphere
2016/11/17 13:33:54 ui: vmware-iso (vsphere): Uploading centos68-v1479406182/centos68-v1479406182/centos68-v1479406182.ovf to vSphere
vmware-iso (vsphere): Uploading centos68-v1479406182/centos68-v1479406182/centos68-v1479406182.ovf to vSphere
2016/11/17 13:33:54 packer: 2016/11/17 13:33:54 Starting ovftool with parameters: --acceptAllEulas --name=centos68-v1479406182 --datastore=NFS-Lab-Templates --noSSLVerify=true --diskMode=thin --vmFolder=/PPP --network=MGMT1 --overwrite --extraConfig:disk.enableUUID=true centos68-v1479406182/centos68-v1479406182/centos68-v1479406182.ovf vi://user:xxxxxx@vc01/Lab/host/LAB_A
==> vmware-iso: Running post-processor: vsphere
2016/11/17 13:59:25 ui: ==> vmware-iso: Running post-processor: vsphere
2016/11/17 13:59:25 ui: vmware-iso (vsphere): Uploading centos68-v1479406182/centos68-v1479406182/centos68-v1479406182.ovf to vSphere
vmware-iso (vsphere): Uploading centos68-v1479406182/centos68-v1479406182/centos68-v1479406182.ovf to vSphere
2016/11/17 13:59:25 packer: 2016/11/17 13:59:25 Starting ovftool with parameters: --acceptAllEulas --name=centos68-v1479406182 --datastore=NFS-Lab-Templates --noSSLVerify=true --diskMode=thin --vmFolder=/PPP --network=MGMT1 --overwrite --extraConfig:disk.enableUUID=true centos68-v1479406182/centos68-v1479406182/centos68-v1479406182.ovf vi://user:xxxxx@vc01/Lab/host/LAB_A
2016/11/17 13:59:54 Deleting original artifact for build 'vmware-iso'
Build 'vmware-iso' errored: 1 error(s) occurred:
* Post-processor failed: Failed: signal: aborted
2016/11/17 13:59:54 ui error: Build 'vmware-iso' errored: 1 error(s) occurred:
* Post-processor failed: Failed: signal: aborted
2016/11/17 13:59:54 Builds completed. Waiting on interrupt barrier...
2016/11/17 13:59:54 machine readable: error-count []string{"1"}
2016/11/17 13:59:54 ui error:
==> Some builds didn't complete successfully and had errors:
2016/11/17 13:59:54 machine readable: vmware-iso,error []string{"1 error(s) occurred:\n\n* Post-processor failed: Failed: signal: aborted\n"}
2016/11/17 13:59:54 ui error: --> vmware-iso: 1 error(s) occurred:
* Post-processor failed: Failed: signal: aborted
2016/11/17 13:59:54 ui:
==> Builds finished but no artifacts were created.
2016/11/17 13:59:54 waiting for all plugin processes to complete...
==> Some builds didn't complete successfully and had errors:
--> vmware-iso: 1 error(s) occurred:
* Post-processor failed: Failed: signal: aborted
```
|
1.0
|
abort error on vmware-iso upload in packer 0.12 - packer 0.12.0
After updating to packer 0.12.0 when trying to upload to vcenter using the vsphere postprocessor I now get the following error:
```
2016/11/17 13:33:54 ui: ==> vmware-iso: Running post-processor: vsphere
2016/11/17 13:33:54 ui: vmware-iso (vsphere): Uploading centos68-v1479406182/centos68-v1479406182/centos68-v1479406182.ovf to vSphere
vmware-iso (vsphere): Uploading centos68-v1479406182/centos68-v1479406182/centos68-v1479406182.ovf to vSphere
2016/11/17 13:33:54 packer: 2016/11/17 13:33:54 Starting ovftool with parameters: --acceptAllEulas --name=centos68-v1479406182 --datastore=NFS-Lab-Templates --noSSLVerify=true --diskMode=thin --vmFolder=/PPP --network=MGMT1 --overwrite --extraConfig:disk.enableUUID=true centos68-v1479406182/centos68-v1479406182/centos68-v1479406182.ovf vi://user:xxxxxx@vc01/Lab/host/LAB_A
==> vmware-iso: Running post-processor: vsphere
2016/11/17 13:59:25 ui: ==> vmware-iso: Running post-processor: vsphere
2016/11/17 13:59:25 ui: vmware-iso (vsphere): Uploading centos68-v1479406182/centos68-v1479406182/centos68-v1479406182.ovf to vSphere
vmware-iso (vsphere): Uploading centos68-v1479406182/centos68-v1479406182/centos68-v1479406182.ovf to vSphere
2016/11/17 13:59:25 packer: 2016/11/17 13:59:25 Starting ovftool with parameters: --acceptAllEulas --name=centos68-v1479406182 --datastore=NFS-Lab-Templates --noSSLVerify=true --diskMode=thin --vmFolder=/PPP --network=MGMT1 --overwrite --extraConfig:disk.enableUUID=true centos68-v1479406182/centos68-v1479406182/centos68-v1479406182.ovf vi://user:xxxxx@vc01/Lab/host/LAB_A
2016/11/17 13:59:54 Deleting original artifact for build 'vmware-iso'
Build 'vmware-iso' errored: 1 error(s) occurred:
* Post-processor failed: Failed: signal: aborted
2016/11/17 13:59:54 ui error: Build 'vmware-iso' errored: 1 error(s) occurred:
* Post-processor failed: Failed: signal: aborted
2016/11/17 13:59:54 Builds completed. Waiting on interrupt barrier...
2016/11/17 13:59:54 machine readable: error-count []string{"1"}
2016/11/17 13:59:54 ui error:
==> Some builds didn't complete successfully and had errors:
2016/11/17 13:59:54 machine readable: vmware-iso,error []string{"1 error(s) occurred:\n\n* Post-processor failed: Failed: signal: aborted\n"}
2016/11/17 13:59:54 ui error: --> vmware-iso: 1 error(s) occurred:
* Post-processor failed: Failed: signal: aborted
2016/11/17 13:59:54 ui:
==> Builds finished but no artifacts were created.
2016/11/17 13:59:54 waiting for all plugin processes to complete...
==> Some builds didn't complete successfully and had errors:
--> vmware-iso: 1 error(s) occurred:
* Post-processor failed: Failed: signal: aborted
```
|
process
|
abort error on vmware iso upload in packer packer after updating to packer when trying to upload to vcenter using the vsphere postprocessor i now get the following error ui vmware iso running post processor vsphere ui vmware iso vsphere uploading ovf to vsphere vmware iso vsphere uploading ovf to vsphere packer starting ovftool with parameters acceptalleulas name datastore nfs lab templates nosslverify true diskmode thin vmfolder ppp network overwrite extraconfig disk enableuuid true ovf vi user xxxxxx lab host lab a vmware iso running post processor vsphere ui vmware iso running post processor vsphere ui vmware iso vsphere uploading ovf to vsphere vmware iso vsphere uploading ovf to vsphere packer starting ovftool with parameters acceptalleulas name datastore nfs lab templates nosslverify true diskmode thin vmfolder ppp network overwrite extraconfig disk enableuuid true ovf vi user xxxxx lab host lab a deleting original artifact for build vmware iso build vmware iso errored error s occurred post processor failed failed signal aborted ui error build vmware iso errored error s occurred post processor failed failed signal aborted builds completed waiting on interrupt barrier machine readable error count string ui error some builds didn t complete successfully and had errors machine readable vmware iso error string error s occurred n n post processor failed failed signal aborted n ui error vmware iso error s occurred post processor failed failed signal aborted ui builds finished but no artifacts were created waiting for all plugin processes to complete some builds didn t complete successfully and had errors vmware iso error s occurred post processor failed failed signal aborted
| 1
|
20,504
| 27,167,253,523
|
IssuesEvent
|
2023-02-17 16:16:08
|
MicrosoftDocs/azure-devops-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
|
closed
|
´YAML is not yet supported in TFS.´ seems incorrect
|
devops/prod doc-bug Pri2 devops-cicd-process/tech
|
The following shows when filter is Azure DevOps Server 2019 (or 2020) and seems incorrect/not up-tp-date:
> "YAML is not yet supported in TFS."
Thanks,
--Neno
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 21e5cee4-eaae-3a96-db91-540ac759e83a
* Version Independent ID: 9bdc837c-ffe0-d999-f922-f3a5debc7f92
* Content: [Conditions - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/conditions?tabs=yaml&view=azure-devops-2020&WT.mc_id=DOP-MVP-21138)
* Content Source: [docs/pipelines/process/conditions.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/main/docs/pipelines/process/conditions.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
1.0
|
´YAML is not yet supported in TFS.´ seems incorrect - The following shows when filter is Azure DevOps Server 2019 (or 2020) and seems incorrect/not up-tp-date:
> "YAML is not yet supported in TFS."
Thanks,
--Neno
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 21e5cee4-eaae-3a96-db91-540ac759e83a
* Version Independent ID: 9bdc837c-ffe0-d999-f922-f3a5debc7f92
* Content: [Conditions - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/conditions?tabs=yaml&view=azure-devops-2020&WT.mc_id=DOP-MVP-21138)
* Content Source: [docs/pipelines/process/conditions.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/main/docs/pipelines/process/conditions.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
process
|
´yaml is not yet supported in tfs ´ seems incorrect the following shows when filter is azure devops server or and seems incorrect not up tp date yaml is not yet supported in tfs thanks neno document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id eaae version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
| 1
|
12,503
| 20,107,743,991
|
IssuesEvent
|
2022-02-07 12:15:15
|
RCDD-202110-TUR-BEW/backend-capstone-turkey-bursapediary
|
https://api.github.com/repos/RCDD-202110-TUR-BEW/backend-capstone-turkey-bursapediary
|
closed
|
testing for middlewares
|
enhancement requirement
|
- testing the isLogged middleware functionality that check the existence of the encrypted token that is generated and saved in cookies after the user logged in
- make sure that the token is valid and contains the required credentials for the user
- make sure to not let the user perform action if the token is not valid or expired
- test the functionality of the isAdmin middleware where its checking the already logged in user to perform admin level tasks
|
1.0
|
testing for middlewares - - testing the isLogged middleware functionality that check the existence of the encrypted token that is generated and saved in cookies after the user logged in
- make sure that the token is valid and contains the required credentials for the user
- make sure to not let the user perform action if the token is not valid or expired
- test the functionality of the isAdmin middleware where its checking the already logged in user to perform admin level tasks
|
non_process
|
testing for middlewares testing the islogged middleware functionality that check the existence of the encrypted token that is generated and saved in cookies after the user logged in make sure that the token is valid and contains the required credentials for the user make sure to not let the user perform action if the token is not valid or expired test the functionality of the isadmin middleware where its checking the already logged in user to perform admin level tasks
| 0
|
77,095
| 9,541,187,222
|
IssuesEvent
|
2019-04-30 21:34:50
|
well-typed/optics
|
https://api.github.com/repos/well-typed/optics
|
closed
|
Provide `from` or don't.
|
design
|
> Perhaps we should expose from in Optics.Iso as a specialisation of re, just for consistency with lens?
https://github.com/well-typed/optics/pull/43#discussion_r247121380
Options:
```
-- 1. alias
from = re
-- 2. restricted alias
from = re :: Iso s t a b -> Iso b a t s -- (working with Equality too)
-- 3. no alias
```
My preference is 3, then 1, then 2.
|
1.0
|
Provide `from` or don't. - > Perhaps we should expose from in Optics.Iso as a specialisation of re, just for consistency with lens?
https://github.com/well-typed/optics/pull/43#discussion_r247121380
Options:
```
-- 1. alias
from = re
-- 2. restricted alias
from = re :: Iso s t a b -> Iso b a t s -- (working with Equality too)
-- 3. no alias
```
My preference is 3, then 1, then 2.
|
non_process
|
provide from or don t perhaps we should expose from in optics iso as a specialisation of re just for consistency with lens options alias from re restricted alias from re iso s t a b iso b a t s working with equality too no alias my preference is then then
| 0
|
1,795
| 4,537,759,054
|
IssuesEvent
|
2016-09-09 02:16:28
|
MModel/MetaModel
|
https://api.github.com/repos/MModel/MetaModel
|
closed
|
Use var count instead of function count
|
[Difficulty] Easy [Propose] Enhancement [Status] In Process
|
Currently, when we need to fetch models count from database, we need to call
```swift
User.count()
```
But `User.count` is much simpler.
Need to both modify MetaModel-template Person.swift file and corresponding file in swift erb template file
|
1.0
|
Use var count instead of function count - Currently, when we need to fetch models count from database, we need to call
```swift
User.count()
```
But `User.count` is much simpler.
Need to both modify MetaModel-template Person.swift file and corresponding file in swift erb template file
|
process
|
use var count instead of function count currently when we need to fetch models count from database we need to call swift user count but user count is much simpler need to both modify metamodel template person swift file and corresponding file in swift erb template file
| 1
|
261,251
| 19,703,877,531
|
IssuesEvent
|
2022-01-12 19:33:19
|
rust-escalations/configuration
|
https://api.github.com/repos/rust-escalations/configuration
|
closed
|
Discord link werkt niet
|
documentation
|
nieuwe discord link die blijvend werk toevoegen aan de server
nieuwe link: https://discord.gg/bcn2ja4BMy
|
1.0
|
Discord link werkt niet - nieuwe discord link die blijvend werk toevoegen aan de server
nieuwe link: https://discord.gg/bcn2ja4BMy
|
non_process
|
discord link werkt niet nieuwe discord link die blijvend werk toevoegen aan de server nieuwe link
| 0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.