Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 4 112 | repo_url stringlengths 33 141 | action stringclasses 3 values | title stringlengths 1 1.02k | labels stringlengths 4 1.54k | body stringlengths 1 262k | index stringclasses 17 values | text_combine stringlengths 95 262k | label stringclasses 2 values | text stringlengths 96 252k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
51,845 | 3,014,617,585 | IssuesEvent | 2015-07-29 15:36:00 | jpchanson/BeSeenium | https://api.github.com/repos/jpchanson/BeSeenium | closed | do jetty proof of concept | Core functionality High Priority | get jetty to my controller test,
get jetty to display some stuff on the screen. | 1.0 | do jetty proof of concept - get jetty to my controller test,
get jetty to display some stuff on the screen. | non_test | do jetty proof of concept get jetty to my controller test get jetty to display some stuff on the screen | 0 |
1,868 | 4,697,449,435 | IssuesEvent | 2016-10-12 09:24:13 | nodejs/node | https://api.github.com/repos/nodejs/node | closed | degrading performance after using child_process | child_process confirmed-bug lts-watch-v4.x os x performance | After running a child process ( using exec or spawn ) I have found that the performance of my node.js application decreases by a factor of 10. Below is a contrived example and output.
```
var exec = require('child_process').exec;
function runExpensiveOperation(times) {
while(times > 0) {
console.time('expensiveOperation');
var str = 'lorem';
for ( var i=0;i< 10000000; i++) {
// string concatenation
str = str.length < 1000 ? str + str : '';
// math operation
i * i * i;
}
console.timeEnd('expensiveOperation');
times--;
}
}
console.log('PRE EXEC');
runExpensiveOperation(10);
exec('echo "hello"');
console.log('POST EXEC');
runExpensiveOperation(10);
```
Output:
```
PRE EXEC
expensiveOperation: 66.458ms
expensiveOperation: 65.735ms
expensiveOperation: 69.237ms
expensiveOperation: 65.269ms
expensiveOperation: 69.133ms
expensiveOperation: 65.639ms
expensiveOperation: 67.944ms
expensiveOperation: 63.595ms
expensiveOperation: 64.153ms
expensiveOperation: 65.093ms
POST EXEC
expensiveOperation: 715.861ms
expensiveOperation: 739.671ms
expensiveOperation: 714.546ms
expensiveOperation: 714.845ms
expensiveOperation: 745.719ms
expensiveOperation: 743.240ms
expensiveOperation: 716.481ms
expensiveOperation: 732.916ms
expensiveOperation: 736.576ms
expensiveOperation: 742.416ms
```
In addition, this problem only occurs if the string concatenation AND math operation are run in the expensiveOperation - if either are commented out then there is no issue.
* **Version**: 5.8.0
* **Platform**: Darwin Kernel Version 15.3.0: Thu Dec 10 18:40:58 PST 2015; root:xnu-3248.30.4~1/RELEASE_X86_64 x86_64 ( Macbook Air OS X El Capitan )
* **Subsystem**: child_process
| 1.0 | degrading performance after using child_process - After running a child process ( using exec or spawn ) I have found that the performance of my node.js application decreases by a factor of 10. Below is a contrived example and output.
```
var exec = require('child_process').exec;
function runExpensiveOperation(times) {
while(times > 0) {
console.time('expensiveOperation');
var str = 'lorem';
for ( var i=0;i< 10000000; i++) {
// string concatenation
str = str.length < 1000 ? str + str : '';
// math operation
i * i * i;
}
console.timeEnd('expensiveOperation');
times--;
}
}
console.log('PRE EXEC');
runExpensiveOperation(10);
exec('echo "hello"');
console.log('POST EXEC');
runExpensiveOperation(10);
```
Output:
```
PRE EXEC
expensiveOperation: 66.458ms
expensiveOperation: 65.735ms
expensiveOperation: 69.237ms
expensiveOperation: 65.269ms
expensiveOperation: 69.133ms
expensiveOperation: 65.639ms
expensiveOperation: 67.944ms
expensiveOperation: 63.595ms
expensiveOperation: 64.153ms
expensiveOperation: 65.093ms
POST EXEC
expensiveOperation: 715.861ms
expensiveOperation: 739.671ms
expensiveOperation: 714.546ms
expensiveOperation: 714.845ms
expensiveOperation: 745.719ms
expensiveOperation: 743.240ms
expensiveOperation: 716.481ms
expensiveOperation: 732.916ms
expensiveOperation: 736.576ms
expensiveOperation: 742.416ms
```
In addition, this problem only occurs if the string concatenation AND math operation are run in the expensiveOperation - if either are commented out then there is no issue.
* **Version**: 5.8.0
* **Platform**: Darwin Kernel Version 15.3.0: Thu Dec 10 18:40:58 PST 2015; root:xnu-3248.30.4~1/RELEASE_X86_64 x86_64 ( Macbook Air OS X El Capitan )
* **Subsystem**: child_process
| non_test | degrading performance after using child process after running a child process using exec or spawn i have found that the performance of my node js application decreases by a factor of below is a contrived example and output var exec require child process exec function runexpensiveoperation times while times console time expensiveoperation var str lorem for var i i i string concatenation str str length str str math operation i i i console timeend expensiveoperation times console log pre exec runexpensiveoperation exec echo hello console log post exec runexpensiveoperation output pre exec expensiveoperation expensiveoperation expensiveoperation expensiveoperation expensiveoperation expensiveoperation expensiveoperation expensiveoperation expensiveoperation expensiveoperation post exec expensiveoperation expensiveoperation expensiveoperation expensiveoperation expensiveoperation expensiveoperation expensiveoperation expensiveoperation expensiveoperation expensiveoperation in addition this problem only occurs if the string concatenation and math operation are run in the expensiveoperation if either are commented out then there is no issue version platform darwin kernel version thu dec pst root xnu release macbook air os x el capitan subsystem child process | 0 |
270,507 | 28,962,277,549 | IssuesEvent | 2023-05-10 04:19:23 | nidhi7598/external_curl_AOSP10_r33 | https://api.github.com/repos/nidhi7598/external_curl_AOSP10_r33 | opened | CVE-2022-32205 (Medium) detected in multiple libraries | Mend: dependency security vulnerability | ## CVE-2022-32205 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>curlcurl-7_64_1</b>, <b>curlcurl-7_64_1</b>, <b>curlcurl-7_64_1</b>, <b>curlcurl-7_64_1</b>, <b>curlcurl-7_64_1</b>, <b>curlcurl-7_64_1</b>, <b>curlcurl-7_64_1</b>, <b>curlcurl-7_64_1</b>, <b>curlcurl-7_64_1</b></p></summary>
<p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
A malicious server can serve excessive amounts of `Set-Cookie:` headers in a HTTP response to curl and curl < 7.84.0 stores all of them. A sufficiently large amount of (big) cookies make subsequent HTTP requests to this, or other servers to which the cookies match, create requests that become larger than the threshold that curl uses internally to avoid sending crazy large requests (1048576 bytes) and instead returns an error.This denial state might remain for as long as the same cookies are kept, match and haven't expired. Due to cookie matching rules, a server on `foo.example.com` can set cookies that also would match for `bar.example.com`, making it it possible for a "sister server" to effectively cause a denial of service for a sibling site on the same second level domain using this method.
<p>Publish Date: 2022-07-07
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-32205>CVE-2022-32205</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-07-07</p>
<p>Fix Resolution: curl-7_71_0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2022-32205 (Medium) detected in multiple libraries - ## CVE-2022-32205 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>curlcurl-7_64_1</b>, <b>curlcurl-7_64_1</b>, <b>curlcurl-7_64_1</b>, <b>curlcurl-7_64_1</b>, <b>curlcurl-7_64_1</b>, <b>curlcurl-7_64_1</b>, <b>curlcurl-7_64_1</b>, <b>curlcurl-7_64_1</b>, <b>curlcurl-7_64_1</b></p></summary>
<p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
A malicious server can serve excessive amounts of `Set-Cookie:` headers in a HTTP response to curl and curl < 7.84.0 stores all of them. A sufficiently large amount of (big) cookies make subsequent HTTP requests to this, or other servers to which the cookies match, create requests that become larger than the threshold that curl uses internally to avoid sending crazy large requests (1048576 bytes) and instead returns an error.This denial state might remain for as long as the same cookies are kept, match and haven't expired. Due to cookie matching rules, a server on `foo.example.com` can set cookies that also would match for `bar.example.com`, making it it possible for a "sister server" to effectively cause a denial of service for a sibling site on the same second level domain using this method.
<p>Publish Date: 2022-07-07
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-32205>CVE-2022-32205</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-07-07</p>
<p>Fix Resolution: curl-7_71_0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_test | cve medium detected in multiple libraries cve medium severity vulnerability vulnerable libraries curlcurl curlcurl curlcurl curlcurl curlcurl curlcurl curlcurl curlcurl curlcurl vulnerability details a malicious server can serve excessive amounts of set cookie headers in a http response to curl and curl stores all of them a sufficiently large amount of big cookies make subsequent http requests to this or other servers to which the cookies match create requests that become larger than the threshold that curl uses internally to avoid sending crazy large requests bytes and instead returns an error this denial state might remain for as long as the same cookies are kept match and haven t expired due to cookie matching rules a server on foo example com can set cookies that also would match for bar example com making it it possible for a sister server to effectively cause a denial of service for a sibling site on the same second level domain using this method publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version release date fix resolution curl step up your open source security game with mend | 0 |
686,881 | 23,507,525,505 | IssuesEvent | 2022-08-18 13:49:55 | twisted/twisted | https://api.github.com/repos/twisted/twisted | closed | TypeError: 'DelayedCall' object is not iterable | core bug priority-normal new | |<img alt="allenap's avatar" src="https://avatars.githubusercontent.com/u/0?s=50" width="50" height="50">| allenap reported|
|-|-|
|Trac ID|trac#8307|
|Type|defect|
|Created|2016-04-26 16:06:50Z|
```
Python 3.5.1+ (default, Mar 30 2016, 22:46:26)
[GCC 5.3.1 20160330] on linux
Type "help", "copyright", "credits" or "license" for more information.
>>> from twisted.internet.base import DelayedCall
>>> dc = DelayedCall(1, lambda: None, (), {}, lambda dc: None, lambda dc: None)
>>> dc.debug = True
>>> dc.cancel()
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "/home/gavin/GitHub/twisted/twisted/internet/base.py", line 94, in cancel
self._str = bytes(self)
TypeError: 'DelayedCall' object is not iterable
```
```
$ tail -n1 twisted/_version.py
version = versions.Version('twisted', 16, 1, 1)
```
<details><summary>Searchable metadata</summary>
```
trac-id__8307 8307
type__defect defect
reporter__allenap allenap
priority__normal normal
milestone__None None
branch__
branch_author__
status__new new
resolution__None None
component__core core
keywords__None None
time__1461686810312194 1461686810312194
changetime__1462290490013595 1462290490013595
version__None None
owner__None None
```
</details>
| 1.0 | TypeError: 'DelayedCall' object is not iterable - |<img alt="allenap's avatar" src="https://avatars.githubusercontent.com/u/0?s=50" width="50" height="50">| allenap reported|
|-|-|
|Trac ID|trac#8307|
|Type|defect|
|Created|2016-04-26 16:06:50Z|
```
Python 3.5.1+ (default, Mar 30 2016, 22:46:26)
[GCC 5.3.1 20160330] on linux
Type "help", "copyright", "credits" or "license" for more information.
>>> from twisted.internet.base import DelayedCall
>>> dc = DelayedCall(1, lambda: None, (), {}, lambda dc: None, lambda dc: None)
>>> dc.debug = True
>>> dc.cancel()
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
File "/home/gavin/GitHub/twisted/twisted/internet/base.py", line 94, in cancel
self._str = bytes(self)
TypeError: 'DelayedCall' object is not iterable
```
```
$ tail -n1 twisted/_version.py
version = versions.Version('twisted', 16, 1, 1)
```
<details><summary>Searchable metadata</summary>
```
trac-id__8307 8307
type__defect defect
reporter__allenap allenap
priority__normal normal
milestone__None None
branch__
branch_author__
status__new new
resolution__None None
component__core core
keywords__None None
time__1461686810312194 1461686810312194
changetime__1462290490013595 1462290490013595
version__None None
owner__None None
```
</details>
| non_test | typeerror delayedcall object is not iterable allenap reported trac id trac type defect created python default mar on linux type help copyright credits or license for more information from twisted internet base import delayedcall dc delayedcall lambda none lambda dc none lambda dc none dc debug true dc cancel traceback most recent call last file line in file home gavin github twisted twisted internet base py line in cancel self str bytes self typeerror delayedcall object is not iterable tail twisted version py version versions version twisted searchable metadata trac id type defect defect reporter allenap allenap priority normal normal milestone none none branch branch author status new new resolution none none component core core keywords none none time changetime version none none owner none none | 0 |
15,767 | 3,974,562,812 | IssuesEvent | 2016-05-04 22:46:02 | LigaData/Kamanja | https://api.github.com/repos/LigaData/Kamanja | closed | The documentation for creating I/O adapters on the website is not complete. | Bug (Documentation) P3 Verify | We have a document that can be put there. Some examples are necessary, as well. | 1.0 | The documentation for creating I/O adapters on the website is not complete. - We have a document that can be put there. Some examples are necessary, as well. | non_test | the documentation for creating i o adapters on the website is not complete we have a document that can be put there some examples are necessary as well | 0 |
261,929 | 22,781,153,799 | IssuesEvent | 2022-07-08 19:55:55 | NuGet/Home | https://api.github.com/repos/NuGet/Home | closed | Remove temporary Newtonsoft.Json workaround in dotnet integration tests | Type:Engineering Type:Test Pipeline:Backlog | Our dotnet integration tests copy a working .NET SDK folder, then patch the NuGet assemblies, and use this copy/patched .NET SDK to run tests.
We're upgrading to a version of newtonsoft.json higher than what the .NET SDK currently has (we're both upgrading at the same time), but until our CI pipeline start using builds of .NET SDK with a high enough version, we need to copy it ourselves.
Here's what needs to be removed:
https://github.com/NuGet/NuGet.Client/pull/4167/files/0e2f6a70900839b1f8670f8fd471d36bf1974df0#diff-8a81e906729d9dc1d11f635564b6eed54b859ab8aaacb3713fc29b19189f0f3b
| 1.0 | Remove temporary Newtonsoft.Json workaround in dotnet integration tests - Our dotnet integration tests copy a working .NET SDK folder, then patch the NuGet assemblies, and use this copy/patched .NET SDK to run tests.
We're upgrading to a version of newtonsoft.json higher than what the .NET SDK currently has (we're both upgrading at the same time), but until our CI pipeline start using builds of .NET SDK with a high enough version, we need to copy it ourselves.
Here's what needs to be removed:
https://github.com/NuGet/NuGet.Client/pull/4167/files/0e2f6a70900839b1f8670f8fd471d36bf1974df0#diff-8a81e906729d9dc1d11f635564b6eed54b859ab8aaacb3713fc29b19189f0f3b
| test | remove temporary newtonsoft json workaround in dotnet integration tests our dotnet integration tests copy a working net sdk folder then patch the nuget assemblies and use this copy patched net sdk to run tests we re upgrading to a version of newtonsoft json higher than what the net sdk currently has we re both upgrading at the same time but until our ci pipeline start using builds of net sdk with a high enough version we need to copy it ourselves here s what needs to be removed | 1 |
708,683 | 24,350,088,695 | IssuesEvent | 2022-10-02 20:47:38 | IAmTamal/Milan | https://api.github.com/repos/IAmTamal/Milan | closed | [DOCS] Readme + License changes 🛠 | 📄 aspect: text ✨ goal: improvement 🟨 priority: medium 🛠 status : under development hacktoberfest | ### Description
Hello!
I would like to help with the ReadMe documentation for the Milan project by making changes in the ReadMe for grammar and sentence formation as well as would like to brainstorm for anything that can be added that can make the project better along with linking the License in the file as well!
Would enjoy if you can assign this issue to me for hacktoberfest!
Thank you!
### Screenshots
_No response_
### Additional information
_No response_ | 1.0 | [DOCS] Readme + License changes 🛠 - ### Description
Hello!
I would like to help with the ReadMe documentation for the Milan project by making changes in the ReadMe for grammar and sentence formation as well as would like to brainstorm for anything that can be added that can make the project better along with linking the License in the file as well!
Would enjoy if you can assign this issue to me for hacktoberfest!
Thank you!
### Screenshots
_No response_
### Additional information
_No response_ | non_test | readme license changes 🛠 description hello i would like to help with the readme documentation for the milan project by making changes in the readme for grammar and sentence formation as well as would like to brainstorm for anything that can be added that can make the project better along with linking the license in the file as well would enjoy if you can assign this issue to me for hacktoberfest thank you screenshots no response additional information no response | 0 |
22,559 | 11,746,204,837 | IssuesEvent | 2020-03-12 11:14:58 | MicrosoftDocs/azure-docs | https://api.github.com/repos/MicrosoftDocs/azure-docs | closed | No mention of the Docker Compose YAML 4000 character limit | Pri2 app-service/svc cxp doc-enhancement triaged | As discussed here https://social.msdn.microsoft.com/Forums/azure/en-US/ff353717-fcb0-42d8-8237-5891e998c1d2/error-on-creating-web-app-with-docker-compose?forum=windowsazurewebsitespreview .
This limits quite basic docker compose setups.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: ad0cae44-bfc2-c879-d406-23ab0b756ab9
* Version Independent ID: 5d812d22-559c-2c9f-7676-c62d7f5c980a
* Content: [Quickstart: Create a multi-container app - Azure App Service](https://docs.microsoft.com/en-us/azure/app-service/containers/quickstart-multi-container#feedback)
* Content Source: [articles/app-service/containers/quickstart-multi-container.md](https://github.com/Microsoft/azure-docs/blob/master/articles/app-service/containers/quickstart-multi-container.md)
* Service: **app-service**
* GitHub Login: @msangapu-msft
* Microsoft Alias: **msangapu** | 1.0 | No mention of the Docker Compose YAML 4000 character limit - As discussed here https://social.msdn.microsoft.com/Forums/azure/en-US/ff353717-fcb0-42d8-8237-5891e998c1d2/error-on-creating-web-app-with-docker-compose?forum=windowsazurewebsitespreview .
This limits quite basic docker compose setups.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: ad0cae44-bfc2-c879-d406-23ab0b756ab9
* Version Independent ID: 5d812d22-559c-2c9f-7676-c62d7f5c980a
* Content: [Quickstart: Create a multi-container app - Azure App Service](https://docs.microsoft.com/en-us/azure/app-service/containers/quickstart-multi-container#feedback)
* Content Source: [articles/app-service/containers/quickstart-multi-container.md](https://github.com/Microsoft/azure-docs/blob/master/articles/app-service/containers/quickstart-multi-container.md)
* Service: **app-service**
* GitHub Login: @msangapu-msft
* Microsoft Alias: **msangapu** | non_test | no mention of the docker compose yaml character limit as discussed here this limits quite basic docker compose setups document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service app service github login msangapu msft microsoft alias msangapu | 0 |
61,328 | 6,733,262,256 | IssuesEvent | 2017-10-18 14:20:15 | ThePenguin1140/OpenCVObstructionTracking | https://api.github.com/repos/ThePenguin1140/OpenCVObstructionTracking | opened | Split video and upload pieces | help wanted testing | I created a test video as part of #4 but it's too big to put on github so I think the best thing might be split it into a few different 'test scenarios' and then upload the pieces?
The video file can be found here:
https://drive.google.com/open?id=0B55XddbN7M0zRG5PTE5VdzhIRzA
| 1.0 | Split video and upload pieces - I created a test video as part of #4 but it's too big to put on github so I think the best thing might be split it into a few different 'test scenarios' and then upload the pieces?
The video file can be found here:
https://drive.google.com/open?id=0B55XddbN7M0zRG5PTE5VdzhIRzA
| test | split video and upload pieces i created a test video as part of but it s too big to put on github so i think the best thing might be split it into a few different test scenarios and then upload the pieces the video file can be found here | 1 |
301,882 | 26,107,190,512 | IssuesEvent | 2022-12-27 14:40:52 | QubesOS/updates-status | https://api.github.com/repos/QubesOS/updates-status | closed | core-agent-linux v4.2.3 (r4.2) | r4.2-vm-bullseye-cur-test r4.2-vm-bookworm-cur-test r4.2-vm-fc37-cur-test r4.2-vm-fc36-cur-test r4.2-vm-centos-stream8-cur-test | Update of core-agent-linux to v4.2.3 for Qubes r4.2, see comments below for details and build status.
From commit: https://github.com/QubesOS/qubes-core-agent-linux/commit/e248fae8e4b7cc89dc6f1b26e68b114fabbdf78a
[Changes since previous version](https://github.com/QubesOS/qubes-core-agent-linux/compare/v4.2.2...v4.2.3):
QubesOS/qubes-core-agent-linux@e248fae8 version 4.2.3
QubesOS/qubes-core-agent-linux@f2bd5c5e firewall: remove debug print
QubesOS/qubes-core-agent-linux@f7f7a026 Merge remote-tracking branch 'origin/pr/399'
QubesOS/qubes-core-agent-linux@765661af ci: fix uploading coverage to codecov
QubesOS/qubes-core-agent-linux@f2db11ae archlinux: update example repo to r4.2 too
QubesOS/qubes-core-agent-linux@90478b0b Revert "temporarily pretend to be 4.1"
QubesOS/qubes-core-agent-linux@292a8ac1 Add purging of no longer allowed connections from conntrack
QubesOS/qubes-core-agent-linux@119eb3ac qubes-rpc/nautilus: Execute external commands asynchronously
QubesOS/qubes-core-agent-linux@0f7f0d6f qubes-rpc/nautilus: Add support for Nautilus API 4.0 The get_file_items method of Nautilus.MenuProvider no longer take the window argument.
Referenced issues:
QubesOS/qubes-issues#7916
QubesOS/qubes-issues#4141
If you're release manager, you can issue GPG-inline signed command:
* `Upload-component r4.2 core-agent-linux e248fae8e4b7cc89dc6f1b26e68b114fabbdf78a current all` (available 5 days from now)
* `Upload-component r4.2 core-agent-linux e248fae8e4b7cc89dc6f1b26e68b114fabbdf78a security-testing`
You can choose subset of distributions like:
* `Upload-component r4.2 core-agent-linux e248fae8e4b7cc89dc6f1b26e68b114fabbdf78a current vm-bookworm,vm-fc37` (available 5 days from now)
Above commands will work only if packages in current-testing repository were built from given commit (i.e. no new version superseded it).
For more information on how to test this update, please take a look at https://www.qubes-os.org/doc/testing/#updates.
| 5.0 | core-agent-linux v4.2.3 (r4.2) - Update of core-agent-linux to v4.2.3 for Qubes r4.2, see comments below for details and build status.
From commit: https://github.com/QubesOS/qubes-core-agent-linux/commit/e248fae8e4b7cc89dc6f1b26e68b114fabbdf78a
[Changes since previous version](https://github.com/QubesOS/qubes-core-agent-linux/compare/v4.2.2...v4.2.3):
QubesOS/qubes-core-agent-linux@e248fae8 version 4.2.3
QubesOS/qubes-core-agent-linux@f2bd5c5e firewall: remove debug print
QubesOS/qubes-core-agent-linux@f7f7a026 Merge remote-tracking branch 'origin/pr/399'
QubesOS/qubes-core-agent-linux@765661af ci: fix uploading coverage to codecov
QubesOS/qubes-core-agent-linux@f2db11ae archlinux: update example repo to r4.2 too
QubesOS/qubes-core-agent-linux@90478b0b Revert "temporarily pretend to be 4.1"
QubesOS/qubes-core-agent-linux@292a8ac1 Add purging of no longer allowed connections from conntrack
QubesOS/qubes-core-agent-linux@119eb3ac qubes-rpc/nautilus: Execute external commands asynchronously
QubesOS/qubes-core-agent-linux@0f7f0d6f qubes-rpc/nautilus: Add support for Nautilus API 4.0 The get_file_items method of Nautilus.MenuProvider no longer take the window argument.
Referenced issues:
QubesOS/qubes-issues#7916
QubesOS/qubes-issues#4141
If you're release manager, you can issue GPG-inline signed command:
* `Upload-component r4.2 core-agent-linux e248fae8e4b7cc89dc6f1b26e68b114fabbdf78a current all` (available 5 days from now)
* `Upload-component r4.2 core-agent-linux e248fae8e4b7cc89dc6f1b26e68b114fabbdf78a security-testing`
You can choose subset of distributions like:
* `Upload-component r4.2 core-agent-linux e248fae8e4b7cc89dc6f1b26e68b114fabbdf78a current vm-bookworm,vm-fc37` (available 5 days from now)
Above commands will work only if packages in current-testing repository were built from given commit (i.e. no new version superseded it).
For more information on how to test this update, please take a look at https://www.qubes-os.org/doc/testing/#updates.
| test | core agent linux update of core agent linux to for qubes see comments below for details and build status from commit qubesos qubes core agent linux version qubesos qubes core agent linux firewall remove debug print qubesos qubes core agent linux merge remote tracking branch origin pr qubesos qubes core agent linux ci fix uploading coverage to codecov qubesos qubes core agent linux archlinux update example repo to too qubesos qubes core agent linux revert temporarily pretend to be qubesos qubes core agent linux add purging of no longer allowed connections from conntrack qubesos qubes core agent linux qubes rpc nautilus execute external commands asynchronously qubesos qubes core agent linux qubes rpc nautilus add support for nautilus api the get file items method of nautilus menuprovider no longer take the window argument referenced issues qubesos qubes issues qubesos qubes issues if you re release manager you can issue gpg inline signed command upload component core agent linux current all available days from now upload component core agent linux security testing you can choose subset of distributions like upload component core agent linux current vm bookworm vm available days from now above commands will work only if packages in current testing repository were built from given commit i e no new version superseded it for more information on how to test this update please take a look at | 1 |
325,575 | 27,944,834,556 | IssuesEvent | 2023-03-24 01:34:24 | WiIIiam278/HuskHomes2 | https://api.github.com/repos/WiIIiam278/HuskHomes2 | closed | Teleport strange behaviour with tabcomplete name event | status: needs testing | Version: 3.0.4 from [spigot](https://www.spigotmc.org/resources/%E2%AD%90-huskhomes-1-16-1-19-%E2%AD%90-simple-intuitive-teleportation-suite-with-cross-server-support.83767/updates)
Using REDIS as messanger_type.
With this messenger_type we found a strange behaviour in tabcompleting names.
1. /tp and /tpa are ok if you are in two separate servers
2. /tp and /tpa player list is taken from the global player list instead of the player list from the servers with REDIS cache (I think it's normal because of this
` public CompletableFuture<List<String>> updatePlayerListCache(@NotNull HuskHomes plugin, @NotNull OnlineUser requester) {
if (plugin.getSettings().crossServer) {
return plugin.getNetworkMessenger().getOnlinePlayerNames(requester).thenApply(returnedPlayerList -> {
players.clear();
players.addAll(List.of(returnedPlayerList));
return players;
});
} else {
players.clear();
players.addAll(plugin.getOnlinePlayers()
.stream()
.filter(player -> !player.isVanished())
.map(onlineUser -> onlineUser.username)
.toList());
System.out.println(String.join(", ", players));
return CompletableFuture.completedFuture(players);
}
}`
3. The biggest problem comes when you are both in the same server: /tp and /tpa doesn't tab complete the player names in the same server. Only if you first write his name and teleport to him, the second time both commands work as defined.
| 1.0 | Teleport strange behaviour with tabcomplete name event - Version: 3.0.4 from [spigot](https://www.spigotmc.org/resources/%E2%AD%90-huskhomes-1-16-1-19-%E2%AD%90-simple-intuitive-teleportation-suite-with-cross-server-support.83767/updates)
Using REDIS as messanger_type.
With this messenger_type we found a strange behaviour in tabcompleting names.
1. /tp and /tpa are ok if you are in two separate servers
2. /tp and /tpa player list is taken from the global player list instead of the player list from the servers with REDIS cache (I think it's normal because of this
` public CompletableFuture<List<String>> updatePlayerListCache(@NotNull HuskHomes plugin, @NotNull OnlineUser requester) {
if (plugin.getSettings().crossServer) {
return plugin.getNetworkMessenger().getOnlinePlayerNames(requester).thenApply(returnedPlayerList -> {
players.clear();
players.addAll(List.of(returnedPlayerList));
return players;
});
} else {
players.clear();
players.addAll(plugin.getOnlinePlayers()
.stream()
.filter(player -> !player.isVanished())
.map(onlineUser -> onlineUser.username)
.toList());
System.out.println(String.join(", ", players));
return CompletableFuture.completedFuture(players);
}
}`
3. The biggest problem comes when you are both in the same server: /tp and /tpa doesn't tab complete the player names in the same server. Only if you first write his name and teleport to him, the second time both commands work as defined.
| test | teleport strange behaviour with tabcomplete name event version from using redis as messanger type with this messenger type we found a strange behaviour in tabcompleting names tp and tpa are ok if you are in two separate servers tp and tpa player list is taken from the global player list instead of the player list from the servers with redis cache i think it s normal because of this public completablefuture updateplayerlistcache notnull huskhomes plugin notnull onlineuser requester if plugin getsettings crossserver return plugin getnetworkmessenger getonlineplayernames requester thenapply returnedplayerlist players clear players addall list of returnedplayerlist return players else players clear players addall plugin getonlineplayers stream filter player player isvanished map onlineuser onlineuser username tolist system out println string join players return completablefuture completedfuture players the biggest problem comes when you are both in the same server tp and tpa doesn t tab complete the player names in the same server only if you first write his name and teleport to him the second time both commands work as defined | 1 |
228,382 | 18,173,514,805 | IssuesEvent | 2021-09-27 22:55:09 | microsoft/vscode | https://api.github.com/repos/microsoft/vscode | opened | Test Autocomplete Preselection Powered By Inline Suggestions | testplan-item | Refs: https://github.com/microsoft/vscode/issues/131940
- [ ] anyOS
- [ ] anyOS
Complexity: 3
[Create Issue](https://github.com/microsoft/vscode/issues/new?body=Testing+%23133906%0A%0A&assignees=hediet)
---
# Preparation
* Enable suggest preview: `"editor.suggest.preview": true,`.
* Install Copilot
# Tasks
* Create a TypeScript file and write `// Write an error to the console` followed by a new line and `console` to prompt copilot.
* Type `.` to trigger autocomplete.
* Verify that the inline suggestion stays and becomes non-italic
* Verify that an autocomplete item is preselected that is a prefix of the inline suggestion
| 1.0 | Test Autocomplete Preselection Powered By Inline Suggestions - Refs: https://github.com/microsoft/vscode/issues/131940
- [ ] anyOS
- [ ] anyOS
Complexity: 3
[Create Issue](https://github.com/microsoft/vscode/issues/new?body=Testing+%23133906%0A%0A&assignees=hediet)
---
# Preparation
* Enable suggest preview: `"editor.suggest.preview": true,`.
* Install Copilot
# Tasks
* Create a TypeScript file and write `// Write an error to the console` followed by a new line and `console` to prompt copilot.
* Type `.` to trigger autocomplete.
* Verify that the inline suggestion stays and becomes non-italic
* Verify that an autocomplete item is preselected that is a prefix of the inline suggestion
| test | test autocomplete preselection powered by inline suggestions refs anyos anyos complexity preparation enable suggest preview editor suggest preview true install copilot tasks create a typescript file and write write an error to the console followed by a new line and console to prompt copilot type to trigger autocomplete verify that the inline suggestion stays and becomes non italic verify that an autocomplete item is preselected that is a prefix of the inline suggestion | 1 |
13,624 | 3,350,354,496 | IssuesEvent | 2015-11-17 14:24:25 | servo/servo | https://api.github.com/repos/servo/servo | closed | Look through old layout bugs and write reftests | A-testing E-less easy | The reftest suite was neglected for a while. We should have more regression tests for bugs we fixed during that time. | 1.0 | Look through old layout bugs and write reftests - The reftest suite was neglected for a while. We should have more regression tests for bugs we fixed during that time. | test | look through old layout bugs and write reftests the reftest suite was neglected for a while we should have more regression tests for bugs we fixed during that time | 1 |
339,802 | 30,476,929,857 | IssuesEvent | 2023-07-17 17:12:59 | rancher/highlander | https://api.github.com/repos/rancher/highlander | opened | Migration of clusters between different Rancher instances | area/testing area/eks kind/qa area/hosted-providers | The Rancher migration procedure (migrating clusters to a 'replacement' Rancher) needs better testing, esp. for regressions.
Apparently migration for EKS clusters behaves differently compared to other downstream clusters.
| 1.0 | Migration of clusters between different Rancher instances - The Rancher migration procedure (migrating clusters to a 'replacement' Rancher) needs better testing, esp. for regressions.
Apparently migration for EKS clusters behaves differently compared to other downstream clusters.
| test | migration of clusters between different rancher instances the rancher migration procedure migrating clusters to a replacement rancher needs better testing esp for regressions apparently migration for eks clusters behaves differently compared to other downstream clusters | 1 |
101,241 | 8,782,994,040 | IssuesEvent | 2018-12-20 03:10:35 | EasyRPG/Player | https://api.github.com/repos/EasyRPG/Player | closed | Pictures won't load when loading savefile in a project with title screen disabled. | Patch available Testcase available Window/Scenes | When you have a RM2k3 >=1.10 project without Ttitle screen enabled in database, when you try to load a savefile with pictures in game, these won't are being loaded (and didn't shown in screen). However, if you enable Title Screen in Database editor, the images are shown without issue.
#### Player platform:
Tested in 0.5.4 and Master (Dec 15th).
#### Test case
Here's a test case:
[imgBug.zip](https://github.com/EasyRPG/Player/files/2683010/imgBug.zip)
Web: https://easyrpg.org/play/master/?game=issue-1571&engine=rpg2k3e (ignore the missing FaceSets)
A RM2k3 1.12a project with title screen disabled. It goes directly to a almost empty map with a automatic event that shows a dummy image, then it deletes itself, and a helper event for save/load calling.
- Save the current progress (with debug features or with the helper event).
- Then load the savefile.
#### Expected behaviour (RPG_RT 1.12a):
The dummy image is being shown when the map loads.
#### Current behaviour (Player 0.5.4 , Master):
No image is shown.
Talking with @fdelapena , maybe this bug can be related with https://github.com/EasyRPG/Player/issues/1524 | 1.0 | Pictures won't load when loading savefile in a project with title screen disabled. - When you have a RM2k3 >=1.10 project without Ttitle screen enabled in database, when you try to load a savefile with pictures in game, these won't are being loaded (and didn't shown in screen). However, if you enable Title Screen in Database editor, the images are shown without issue.
#### Player platform:
Tested in 0.5.4 and Master (Dec 15th).
#### Test case
Here's a test case:
[imgBug.zip](https://github.com/EasyRPG/Player/files/2683010/imgBug.zip)
Web: https://easyrpg.org/play/master/?game=issue-1571&engine=rpg2k3e (ignore the missing FaceSets)
A RM2k3 1.12a project with title screen disabled. It goes directly to a almost empty map with a automatic event that shows a dummy image, then it deletes itself, and a helper event for save/load calling.
- Save the current progress (with debug features or with the helper event).
- Then load the savefile.
#### Expected behaviour (RPG_RT 1.12a):
The dummy image is being shown when the map loads.
#### Current behaviour (Player 0.5.4 , Master):
No image is shown.
Talking with @fdelapena , maybe this bug can be related with https://github.com/EasyRPG/Player/issues/1524 | test | pictures won t load when loading savefile in a project with title screen disabled when you have a project without ttitle screen enabled in database when you try to load a savefile with pictures in game these won t are being loaded and didn t shown in screen however if you enable title screen in database editor the images are shown without issue player platform tested in and master dec test case here s a test case web ignore the missing facesets a project with title screen disabled it goes directly to a almost empty map with a automatic event that shows a dummy image then it deletes itself and a helper event for save load calling save the current progress with debug features or with the helper event then load the savefile expected behaviour rpg rt the dummy image is being shown when the map loads current behaviour player master no image is shown talking with fdelapena maybe this bug can be related with | 1 |
67,391 | 12,952,869,085 | IssuesEvent | 2020-07-19 22:15:20 | eucalypto/eucalyptapp | https://api.github.com/repos/eucalypto/eucalyptapp | opened | Replace DataBindingUtil in GameWonFragment | code enhancement | The use of DataBindingUtil is deprecated and should be replaced by the generated Class for the specific Fragment. | 1.0 | Replace DataBindingUtil in GameWonFragment - The use of DataBindingUtil is deprecated and should be replaced by the generated Class for the specific Fragment. | non_test | replace databindingutil in gamewonfragment the use of databindingutil is deprecated and should be replaced by the generated class for the specific fragment | 0 |
62,241 | 3,179,650,107 | IssuesEvent | 2015-09-25 03:21:07 | cjfields/redmine-test | https://api.github.com/repos/cjfields/redmine-test | opened | Bio::Seq object loses sequence data when blessed as Bio::Seq::Meta::Array | Category: Core Components Priority: Normal Status: New Tracker: Bug | ---
Author Name: **Roy Chaudhuri** (Roy Chaudhuri)
Original Redmine Issue: 2262, https://redmine.open-bio.org/issues/2262
Original Date: 2007-04-04
Original Assignee: Bioperl Guts
---
When I bless a Bio::Seq object as a Bio::Seq::Meta::Array (as instructed by the POD) it loses the primary sequence information (and other PrimarySeqI information such as length and accession). Features and annotation are not affected.
Bio::Seq::Meta suffers from a worse bug- Bio::SeqIO warns that the Bio::Seq::Meta object is not SeqI compliant. This seems to be due to the omission of Bio::Seq from the use base line in Bio::Seq::Meta (but not Bio::Seq::Meta::Array). When I add Bio::Seq into the use base line the behaviour is the same as for Bio::Seq::Meta::Array.
The following code demonstrates the problem:
\#!/usr/bin/perl
use warnings;
use strict;
use Bio::SeqIO;
use Bio::Seq::Meta::Array;
my $seq=Bio::SeqIO-\>new(-fh=\>\\\*ARGV, ~~format=\>’genbank’)~~\>next\_seq;
bless $seq, ‘Bio::Seq::Meta::Array’;
Bio::SeqIO-\>new(~~format=\>’genbank’)~~\>write\_seq($seq);
| 1.0 | Bio::Seq object loses sequence data when blessed as Bio::Seq::Meta::Array - ---
Author Name: **Roy Chaudhuri** (Roy Chaudhuri)
Original Redmine Issue: 2262, https://redmine.open-bio.org/issues/2262
Original Date: 2007-04-04
Original Assignee: Bioperl Guts
---
When I bless a Bio::Seq object as a Bio::Seq::Meta::Array (as instructed by the POD) it loses the primary sequence information (and other PrimarySeqI information such as length and accession). Features and annotation are not affected.
Bio::Seq::Meta suffers from a worse bug- Bio::SeqIO warns that the Bio::Seq::Meta object is not SeqI compliant. This seems to be due to the omission of Bio::Seq from the use base line in Bio::Seq::Meta (but not Bio::Seq::Meta::Array). When I add Bio::Seq into the use base line the behaviour is the same as for Bio::Seq::Meta::Array.
The following code demonstrates the problem:
\#!/usr/bin/perl
use warnings;
use strict;
use Bio::SeqIO;
use Bio::Seq::Meta::Array;
my $seq=Bio::SeqIO-\>new(-fh=\>\\\*ARGV, ~~format=\>’genbank’)~~\>next\_seq;
bless $seq, ‘Bio::Seq::Meta::Array’;
Bio::SeqIO-\>new(~~format=\>’genbank’)~~\>write\_seq($seq);
| non_test | bio seq object loses sequence data when blessed as bio seq meta array author name roy chaudhuri roy chaudhuri original redmine issue original date original assignee bioperl guts when i bless a bio seq object as a bio seq meta array as instructed by the pod it loses the primary sequence information and other primaryseqi information such as length and accession features and annotation are not affected bio seq meta suffers from a worse bug bio seqio warns that the bio seq meta object is not seqi compliant this seems to be due to the omission of bio seq from the use base line in bio seq meta but not bio seq meta array when i add bio seq into the use base line the behaviour is the same as for bio seq meta array the following code demonstrates the problem usr bin perl use warnings use strict use bio seqio use bio seq meta array my seq bio seqio new fh argv format ’genbank’ next seq bless seq ‘bio seq meta array’ bio seqio new format ’genbank’ write seq seq | 0 |
390,436 | 11,543,724,741 | IssuesEvent | 2020-02-18 10:07:15 | wso2/product-apim | https://api.github.com/repos/wso2/product-apim | opened | Issue in invoking SOAP APIs | Priority/Normal Type/Bug | ### Description:
Following issue occurred when invoking Pass through API created using SOAP back-end.
1. NPE when invoking:
```
ERROR - ServerWorker Error processing POST request for : /through/1.
java.lang.NullPointerException: null
at org.wso2.carbon.apimgt.gateway.utils.OpenAPIUtils.getPathItemSecurityScopes_aroundBody10(OpenAPIUtils.java:205) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?]
at org.wso2.carbon.apimgt.gateway.utils.OpenAPIUtils.getPathItemSecurityScopes(OpenAPIUtils.java:191) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?]
at org.wso2.carbon.apimgt.gateway.utils.OpenAPIUtils.getScopesOfResource_aroundBody2(OpenAPIUtils.java:77) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?]
at org.wso2.carbon.apimgt.gateway.utils.OpenAPIUtils.getScopesOfResource(OpenAPIUtils.java:71) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?]
at org.wso2.carbon.apimgt.gateway.handlers.security.jwt.JWTValidator.validateScopes_aroundBody8(JWTValidator.java:542) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?]
at org.wso2.carbon.apimgt.gateway.handlers.security.jwt.JWTValidator.validateScopes(JWTValidator.java:530) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?]
at org.wso2.carbon.apimgt.gateway.handlers.security.jwt.JWTValidator.authenticate_aroundBody0(JWTValidator.java:230) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?]
at org.wso2.carbon.apimgt.gateway.handlers.security.jwt.JWTValidator.authenticate(JWTValidator.java:91) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?]
at org.wso2.carbon.apimgt.gateway.handlers.security.oauth.OAuthAuthenticator.authenticate_aroundBody4(OAuthAuthenticator.java:333) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?]
at org.wso2.carbon.apimgt.gateway.handlers.security.oauth.OAuthAuthenticator.authenticate(OAuthAuthenticator.java:109) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?]
at org.wso2.carbon.apimgt.gateway.handlers.security.APIAuthenticationHandler.isAuthenticate_aroundBody42(APIAuthenticationHandler.java:419) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?]
at org.wso2.carbon.apimgt.gateway.handlers.security.APIAuthenticationHandler.isAuthenticate(APIAuthenticationHandler.java:413) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?]
at org.wso2.carbon.apimgt.gateway.handlers.security.APIAuthenticationHandler.handleRequest_aroundBody36(APIAuthenticationHandler.java:349) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?]
at org.wso2.carbon.apimgt.gateway.handlers.security.APIAuthenticationHandler.handleRequest(APIAuthenticationHandler.java:320) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?]
at org.apache.synapse.rest.API.process(API.java:367) ~[synapse-core_2.1.7.wso2v143.jar:2.1.7-wso2v143]
at org.apache.synapse.rest.RESTRequestHandler.apiProcessNonDefaultStrategy(RESTRequestHandler.java:149) ~[synapse-core_2.1.7.wso2v143.jar:2.1.7-wso2v143]
at org.apache.synapse.rest.RESTRequestHandler.dispatchToAPI(RESTRequestHandler.java:95) ~[synapse-core_2.1.7.wso2v143.jar:2.1.7-wso2v143]
at org.apache.synapse.rest.RESTRequestHandler.process(RESTRequestHandler.java:71) ~[synapse-core_2.1.7.wso2v143.jar:2.1.7-wso2v143]
at org.apache.synapse.core.axis2.Axis2SynapseEnvironment.injectMessage(Axis2SynapseEnvironment.java:327) ~[synapse-core_2.1.7.wso2v143.jar:2.1.7-wso2v143]
at org.apache.synapse.core.axis2.SynapseMessageReceiver.receive(SynapseMessageReceiver.java:98) ~[synapse-core_2.1.7.wso2v143.jar:2.1.7-wso2v143]
at org.apache.axis2.engine.AxisEngine.receive(AxisEngine.java:180) ~[axis2_1.6.1.wso2v40.jar:?]
at org.apache.synapse.transport.passthru.ServerWorker.processNonEntityEnclosingRESTHandler(ServerWorker.java:368) [synapse-nhttp-transport_2.1.7.wso2v143.jar:?]
at org.apache.synapse.transport.passthru.ServerWorker.processEntityEnclosingRequest(ServerWorker.java:427) [synapse-nhttp-transport_2.1.7.wso2v143.jar:?]
at org.apache.synapse.transport.passthru.ServerWorker.run(ServerWorker.java:182) [synapse-nhttp-transport_2.1.7.wso2v143.jar:?]
at org.apache.axis2.transport.base.threads.NativeWorkerPool$1.run(NativeWorkerPool.java:172) [axis2_1.6.1.wso2v40.jar:?]
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) [?:1.8.0_201]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) [?:1.8.0_201]
at java.lang.Thread.run(Thread.java:748) [?:1.8.0_201]
```
2. Generated swagger does not have the content-type of input (The default content type should be **Text/xml**). Hence, it is not possible to select the content-type in Swagger UI in Dev Portal. Refer the following screenshot:

| 1.0 | Issue in invoking SOAP APIs - ### Description:
Following issue occurred when invoking Pass through API created using SOAP back-end.
1. NPE when invoking:
```
ERROR - ServerWorker Error processing POST request for : /through/1.
java.lang.NullPointerException: null
at org.wso2.carbon.apimgt.gateway.utils.OpenAPIUtils.getPathItemSecurityScopes_aroundBody10(OpenAPIUtils.java:205) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?]
at org.wso2.carbon.apimgt.gateway.utils.OpenAPIUtils.getPathItemSecurityScopes(OpenAPIUtils.java:191) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?]
at org.wso2.carbon.apimgt.gateway.utils.OpenAPIUtils.getScopesOfResource_aroundBody2(OpenAPIUtils.java:77) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?]
at org.wso2.carbon.apimgt.gateway.utils.OpenAPIUtils.getScopesOfResource(OpenAPIUtils.java:71) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?]
at org.wso2.carbon.apimgt.gateway.handlers.security.jwt.JWTValidator.validateScopes_aroundBody8(JWTValidator.java:542) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?]
at org.wso2.carbon.apimgt.gateway.handlers.security.jwt.JWTValidator.validateScopes(JWTValidator.java:530) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?]
at org.wso2.carbon.apimgt.gateway.handlers.security.jwt.JWTValidator.authenticate_aroundBody0(JWTValidator.java:230) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?]
at org.wso2.carbon.apimgt.gateway.handlers.security.jwt.JWTValidator.authenticate(JWTValidator.java:91) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?]
at org.wso2.carbon.apimgt.gateway.handlers.security.oauth.OAuthAuthenticator.authenticate_aroundBody4(OAuthAuthenticator.java:333) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?]
at org.wso2.carbon.apimgt.gateway.handlers.security.oauth.OAuthAuthenticator.authenticate(OAuthAuthenticator.java:109) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?]
at org.wso2.carbon.apimgt.gateway.handlers.security.APIAuthenticationHandler.isAuthenticate_aroundBody42(APIAuthenticationHandler.java:419) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?]
at org.wso2.carbon.apimgt.gateway.handlers.security.APIAuthenticationHandler.isAuthenticate(APIAuthenticationHandler.java:413) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?]
at org.wso2.carbon.apimgt.gateway.handlers.security.APIAuthenticationHandler.handleRequest_aroundBody36(APIAuthenticationHandler.java:349) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?]
at org.wso2.carbon.apimgt.gateway.handlers.security.APIAuthenticationHandler.handleRequest(APIAuthenticationHandler.java:320) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?]
at org.apache.synapse.rest.API.process(API.java:367) ~[synapse-core_2.1.7.wso2v143.jar:2.1.7-wso2v143]
at org.apache.synapse.rest.RESTRequestHandler.apiProcessNonDefaultStrategy(RESTRequestHandler.java:149) ~[synapse-core_2.1.7.wso2v143.jar:2.1.7-wso2v143]
at org.apache.synapse.rest.RESTRequestHandler.dispatchToAPI(RESTRequestHandler.java:95) ~[synapse-core_2.1.7.wso2v143.jar:2.1.7-wso2v143]
at org.apache.synapse.rest.RESTRequestHandler.process(RESTRequestHandler.java:71) ~[synapse-core_2.1.7.wso2v143.jar:2.1.7-wso2v143]
at org.apache.synapse.core.axis2.Axis2SynapseEnvironment.injectMessage(Axis2SynapseEnvironment.java:327) ~[synapse-core_2.1.7.wso2v143.jar:2.1.7-wso2v143]
at org.apache.synapse.core.axis2.SynapseMessageReceiver.receive(SynapseMessageReceiver.java:98) ~[synapse-core_2.1.7.wso2v143.jar:2.1.7-wso2v143]
at org.apache.axis2.engine.AxisEngine.receive(AxisEngine.java:180) ~[axis2_1.6.1.wso2v40.jar:?]
at org.apache.synapse.transport.passthru.ServerWorker.processNonEntityEnclosingRESTHandler(ServerWorker.java:368) [synapse-nhttp-transport_2.1.7.wso2v143.jar:?]
at org.apache.synapse.transport.passthru.ServerWorker.processEntityEnclosingRequest(ServerWorker.java:427) [synapse-nhttp-transport_2.1.7.wso2v143.jar:?]
at org.apache.synapse.transport.passthru.ServerWorker.run(ServerWorker.java:182) [synapse-nhttp-transport_2.1.7.wso2v143.jar:?]
at org.apache.axis2.transport.base.threads.NativeWorkerPool$1.run(NativeWorkerPool.java:172) [axis2_1.6.1.wso2v40.jar:?]
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) [?:1.8.0_201]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) [?:1.8.0_201]
at java.lang.Thread.run(Thread.java:748) [?:1.8.0_201]
```
2. Generated swagger does not have the content-type of input (The default content type should be **Text/xml**). Hence, it is not possible to select the content-type in Swagger UI in Dev Portal. Refer the following screenshot:

| non_test | issue in invoking soap apis description following issue occurred when invoking pass through api created using soap back end npe when invoking error serverworker error processing post request for through java lang nullpointerexception null at org carbon apimgt gateway utils openapiutils getpathitemsecurityscopes openapiutils java at org carbon apimgt gateway utils openapiutils getpathitemsecurityscopes openapiutils java at org carbon apimgt gateway utils openapiutils getscopesofresource openapiutils java at org carbon apimgt gateway utils openapiutils getscopesofresource openapiutils java at org carbon apimgt gateway handlers security jwt jwtvalidator validatescopes jwtvalidator java at org carbon apimgt gateway handlers security jwt jwtvalidator validatescopes jwtvalidator java at org carbon apimgt gateway handlers security jwt jwtvalidator authenticate jwtvalidator java at org carbon apimgt gateway handlers security jwt jwtvalidator authenticate jwtvalidator java at org carbon apimgt gateway handlers security oauth oauthauthenticator authenticate oauthauthenticator java at org carbon apimgt gateway handlers security oauth oauthauthenticator authenticate oauthauthenticator java at org carbon apimgt gateway handlers security apiauthenticationhandler isauthenticate apiauthenticationhandler java at org carbon apimgt gateway handlers security apiauthenticationhandler isauthenticate apiauthenticationhandler java at org carbon apimgt gateway handlers security apiauthenticationhandler handlerequest apiauthenticationhandler java at org carbon apimgt gateway handlers security apiauthenticationhandler handlerequest apiauthenticationhandler java at org apache synapse rest api process api java at org apache synapse rest restrequesthandler apiprocessnondefaultstrategy restrequesthandler java at org apache synapse rest restrequesthandler dispatchtoapi restrequesthandler java at org apache synapse rest restrequesthandler process restrequesthandler java at org apache synapse core injectmessage java at org apache synapse core synapsemessagereceiver receive synapsemessagereceiver java at org apache engine axisengine receive axisengine java at org apache synapse transport passthru serverworker processnonentityenclosingresthandler serverworker java at org apache synapse transport passthru serverworker processentityenclosingrequest serverworker java at org apache synapse transport passthru serverworker run serverworker java at org apache transport base threads nativeworkerpool run nativeworkerpool java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java generated swagger does not have the content type of input the default content type should be text xml hence it is not possible to select the content type in swagger ui in dev portal refer the following screenshot | 0 |
258,038 | 22,272,201,756 | IssuesEvent | 2022-06-10 13:25:23 | mozilla-mobile/fenix | https://api.github.com/repos/mozilla-mobile/fenix | opened | Intermittent UI test failure - TabbedBrowsingTest.closeTabTest | eng:ui-test | ### Firebase Test Run:
https://console.firebase.google.com/u/0/project/moz-fenix/testlab/histories/bh.66b7091e15d53d45/matrices/6696243506005561227/executions/bs.6f49200709fbc39c
### Stacktrace:
`java.lang.RuntimeException: Error while connecting UiAutomation@3740472[id=-1, flags=0]
at android.app.UiAutomation.connect(UiAutomation.java:259)
at android.app.Instrumentation.getUiAutomation(Instrumentation.java:2176)
at androidx.test.uiautomator.UiDevice.getUiAutomation(UiDevice.java:1129)
at androidx.test.uiautomator.QueryController.<init>(QueryController.java:95)
at androidx.test.uiautomator.UiDevice.<init>(UiDevice.java:109)
at androidx.test.uiautomator.UiDevice.getInstance(UiDevice.java:261)
at org.mozilla.fenix.ui.TabbedBrowsingTest.<init>(TabbedBrowsingTest.kt:43)`
### Build:
on main 6/9
Known issue: https://github.com/mozilla-mobile/fenix/issues/25132 | 1.0 | Intermittent UI test failure - TabbedBrowsingTest.closeTabTest - ### Firebase Test Run:
https://console.firebase.google.com/u/0/project/moz-fenix/testlab/histories/bh.66b7091e15d53d45/matrices/6696243506005561227/executions/bs.6f49200709fbc39c
### Stacktrace:
`java.lang.RuntimeException: Error while connecting UiAutomation@3740472[id=-1, flags=0]
at android.app.UiAutomation.connect(UiAutomation.java:259)
at android.app.Instrumentation.getUiAutomation(Instrumentation.java:2176)
at androidx.test.uiautomator.UiDevice.getUiAutomation(UiDevice.java:1129)
at androidx.test.uiautomator.QueryController.<init>(QueryController.java:95)
at androidx.test.uiautomator.UiDevice.<init>(UiDevice.java:109)
at androidx.test.uiautomator.UiDevice.getInstance(UiDevice.java:261)
at org.mozilla.fenix.ui.TabbedBrowsingTest.<init>(TabbedBrowsingTest.kt:43)`
### Build:
on main 6/9
Known issue: https://github.com/mozilla-mobile/fenix/issues/25132 | test | intermittent ui test failure tabbedbrowsingtest closetabtest firebase test run stacktrace java lang runtimeexception error while connecting uiautomation at android app uiautomation connect uiautomation java at android app instrumentation getuiautomation instrumentation java at androidx test uiautomator uidevice getuiautomation uidevice java at androidx test uiautomator querycontroller querycontroller java at androidx test uiautomator uidevice uidevice java at androidx test uiautomator uidevice getinstance uidevice java at org mozilla fenix ui tabbedbrowsingtest tabbedbrowsingtest kt build on main known issue | 1 |
174,691 | 13,505,085,402 | IssuesEvent | 2020-09-13 21:04:22 | thexerteproject/xerteonlinetoolkits | https://api.github.com/repos/thexerteproject/xerteonlinetoolkits | closed | Refactor Xenith engine code | Needs testing New feature XOT template enhancement | Xenith engine has got rather huge and spaghetti like and it's really hard for me to follow (having not been around the code for a while) so it must be impossible for anyone new coming in and it's only going to get more difficult for us to support unless we try and modernise it, group related code and refactor. It is getting on 8 years old now!
Anyway, i've run this past @FayCross a while ago but didn't have the time to implement any of it but with this self-isolation thing i've got all the time in the world!!
To start i'm going to refactor:
- Variables code (that's a good 600 lines alone)
- Glossary
- Menu
- Dialog
I'll leave them in xenith.js for now (xenith will get bigger just now but will eventually have the code blocks separated out to separate files) but I also have a loader which can load on the fly each of the blocks, when needed, but need to work on that to make sure it works offline and in scorm.
I know there are probably other bugs and developments that need done but i'm hope this refactor will reduce the complication of tracing and fixing those issues.
I've done VARIABLES and it's working fine as far as I can see. So i'll start committing them as separate commits and tag this issue so we can keep track of them.
Hope you guys are all safe! | 1.0 | Refactor Xenith engine code - Xenith engine has got rather huge and spaghetti like and it's really hard for me to follow (having not been around the code for a while) so it must be impossible for anyone new coming in and it's only going to get more difficult for us to support unless we try and modernise it, group related code and refactor. It is getting on 8 years old now!
Anyway, i've run this past @FayCross a while ago but didn't have the time to implement any of it but with this self-isolation thing i've got all the time in the world!!
To start i'm going to refactor:
- Variables code (that's a good 600 lines alone)
- Glossary
- Menu
- Dialog
I'll leave them in xenith.js for now (xenith will get bigger just now but will eventually have the code blocks separated out to separate files) but I also have a loader which can load on the fly each of the blocks, when needed, but need to work on that to make sure it works offline and in scorm.
I know there are probably other bugs and developments that need done but i'm hope this refactor will reduce the complication of tracing and fixing those issues.
I've done VARIABLES and it's working fine as far as I can see. So i'll start committing them as separate commits and tag this issue so we can keep track of them.
Hope you guys are all safe! | test | refactor xenith engine code xenith engine has got rather huge and spaghetti like and it s really hard for me to follow having not been around the code for a while so it must be impossible for anyone new coming in and it s only going to get more difficult for us to support unless we try and modernise it group related code and refactor it is getting on years old now anyway i ve run this past faycross a while ago but didn t have the time to implement any of it but with this self isolation thing i ve got all the time in the world to start i m going to refactor variables code that s a good lines alone glossary menu dialog i ll leave them in xenith js for now xenith will get bigger just now but will eventually have the code blocks separated out to separate files but i also have a loader which can load on the fly each of the blocks when needed but need to work on that to make sure it works offline and in scorm i know there are probably other bugs and developments that need done but i m hope this refactor will reduce the complication of tracing and fixing those issues i ve done variables and it s working fine as far as i can see so i ll start committing them as separate commits and tag this issue so we can keep track of them hope you guys are all safe | 1 |
202,484 | 15,286,694,431 | IssuesEvent | 2021-02-23 14:58:55 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | roachtest: kv95/enc=false/nodes=3/cpu=32/seq failed | C-test-failure O-roachtest O-robot branch-release-20.1 release-blocker | [(roachtest).kv95/enc=false/nodes=3/cpu=32/seq failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2661561&tab=buildLog) on [release-20.1@90f78268f3b5b08ba838ac3ad164821d2f5a5362](https://github.com/cockroachdb/cockroach/commits/90f78268f3b5b08ba838ac3ad164821d2f5a5362):
```
| github.com/cockroachdb/cockroach/vendor/golang.org/x/sync/errgroup.(*Group).Go.func1
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/golang.org/x/sync/errgroup/errgroup.go:57
| runtime.goexit
| /usr/local/go/src/runtime/asm_amd64.s:1374
Wraps: (2) 2 safe details enclosed
Wraps: (3) output in run_074156.709_n4_workload_run_kv
Wraps: (4) /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-2661561-1612941349-20-n4cpu32:4 -- ./workload run kv --init --histograms=perf/stats.json --concurrency=192 --duration=10m0s --read-percent=95 --sequential {pgurl:1-3} returned
| stderr:
| ./workload: error while loading shared libraries: libncurses.so.6: cannot open shared object file: No such file or directory
| Error: COMMAND_PROBLEM: exit status 127
| (1) COMMAND_PROBLEM
| Wraps: (2) Node 4. Command with error:
| | ```
| | ./workload run kv --init --histograms=perf/stats.json --concurrency=192 --duration=10m0s --read-percent=95 --sequential {pgurl:1-3}
| | ```
| Wraps: (3) exit status 127
| Error types: (1) errors.Cmd (2) *hintdetail.withDetail (3) *exec.ExitError
|
| stdout:
Wraps: (5) exit status 20
Error types: (1) *withstack.withStack (2) *safedetails.withSafeDetails (3) *errutil.withMessage (4) *main.withCommandDetails (5) *exec.ExitError
cluster.go:2628,kv.go:96,kv.go:183,test_runner.go:749: monitor failure: monitor task failed: t.Fatal() was called
(1) attached stack trace
| main.(*monitor).WaitE
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2616
| main.(*monitor).Wait
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2624
| main.registerKV.func2
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/kv.go:96
| main.registerKV.func3
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/kv.go:183
| main.(*testRunner).runTest.func2
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/test_runner.go:749
Wraps: (2) monitor failure
Wraps: (3) attached stack trace
| main.(*monitor).wait.func2
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2672
Wraps: (4) monitor task failed
Wraps: (5) attached stack trace
| main.init
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2586
| runtime.doInit
| /usr/local/go/src/runtime/proc.go:5652
| runtime.main
| /usr/local/go/src/runtime/proc.go:191
| runtime.goexit
| /usr/local/go/src/runtime/asm_amd64.s:1374
Wraps: (6) t.Fatal() was called
Error types: (1) *withstack.withStack (2) *errutil.withMessage (3) *withstack.withStack (4) *errutil.withMessage (5) *withstack.withStack (6) *errors.errorString
```
<details><summary>More</summary><p>
Artifacts: [/kv95/enc=false/nodes=3/cpu=32/seq](https://teamcity.cockroachdb.com/viewLog.html?buildId=2661561&tab=artifacts#/kv95/enc=false/nodes=3/cpu=32/seq)
Related:
- #60224 roachtest: kv95/enc=false/nodes=3/cpu=32/seq failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-60149](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-60149) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #60077 roachtest: kv95/enc=false/nodes=3/cpu=32/seq failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-20.2](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-20.2) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #59924 roachtest: kv95/enc=false/nodes=3/cpu=32/seq failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-master](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-master) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Akv95%2Fenc%3Dfalse%2Fnodes%3D3%2Fcpu%3D32%2Fseq.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
| 2.0 | roachtest: kv95/enc=false/nodes=3/cpu=32/seq failed - [(roachtest).kv95/enc=false/nodes=3/cpu=32/seq failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2661561&tab=buildLog) on [release-20.1@90f78268f3b5b08ba838ac3ad164821d2f5a5362](https://github.com/cockroachdb/cockroach/commits/90f78268f3b5b08ba838ac3ad164821d2f5a5362):
```
| github.com/cockroachdb/cockroach/vendor/golang.org/x/sync/errgroup.(*Group).Go.func1
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/golang.org/x/sync/errgroup/errgroup.go:57
| runtime.goexit
| /usr/local/go/src/runtime/asm_amd64.s:1374
Wraps: (2) 2 safe details enclosed
Wraps: (3) output in run_074156.709_n4_workload_run_kv
Wraps: (4) /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-2661561-1612941349-20-n4cpu32:4 -- ./workload run kv --init --histograms=perf/stats.json --concurrency=192 --duration=10m0s --read-percent=95 --sequential {pgurl:1-3} returned
| stderr:
| ./workload: error while loading shared libraries: libncurses.so.6: cannot open shared object file: No such file or directory
| Error: COMMAND_PROBLEM: exit status 127
| (1) COMMAND_PROBLEM
| Wraps: (2) Node 4. Command with error:
| | ```
| | ./workload run kv --init --histograms=perf/stats.json --concurrency=192 --duration=10m0s --read-percent=95 --sequential {pgurl:1-3}
| | ```
| Wraps: (3) exit status 127
| Error types: (1) errors.Cmd (2) *hintdetail.withDetail (3) *exec.ExitError
|
| stdout:
Wraps: (5) exit status 20
Error types: (1) *withstack.withStack (2) *safedetails.withSafeDetails (3) *errutil.withMessage (4) *main.withCommandDetails (5) *exec.ExitError
cluster.go:2628,kv.go:96,kv.go:183,test_runner.go:749: monitor failure: monitor task failed: t.Fatal() was called
(1) attached stack trace
| main.(*monitor).WaitE
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2616
| main.(*monitor).Wait
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2624
| main.registerKV.func2
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/kv.go:96
| main.registerKV.func3
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/kv.go:183
| main.(*testRunner).runTest.func2
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/test_runner.go:749
Wraps: (2) monitor failure
Wraps: (3) attached stack trace
| main.(*monitor).wait.func2
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2672
Wraps: (4) monitor task failed
Wraps: (5) attached stack trace
| main.init
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2586
| runtime.doInit
| /usr/local/go/src/runtime/proc.go:5652
| runtime.main
| /usr/local/go/src/runtime/proc.go:191
| runtime.goexit
| /usr/local/go/src/runtime/asm_amd64.s:1374
Wraps: (6) t.Fatal() was called
Error types: (1) *withstack.withStack (2) *errutil.withMessage (3) *withstack.withStack (4) *errutil.withMessage (5) *withstack.withStack (6) *errors.errorString
```
<details><summary>More</summary><p>
Artifacts: [/kv95/enc=false/nodes=3/cpu=32/seq](https://teamcity.cockroachdb.com/viewLog.html?buildId=2661561&tab=artifacts#/kv95/enc=false/nodes=3/cpu=32/seq)
Related:
- #60224 roachtest: kv95/enc=false/nodes=3/cpu=32/seq failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-60149](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-60149) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #60077 roachtest: kv95/enc=false/nodes=3/cpu=32/seq failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-20.2](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-20.2) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #59924 roachtest: kv95/enc=false/nodes=3/cpu=32/seq failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-master](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-master) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Akv95%2Fenc%3Dfalse%2Fnodes%3D3%2Fcpu%3D32%2Fseq.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
| test | roachtest enc false nodes cpu seq failed on github com cockroachdb cockroach vendor golang org x sync errgroup group go home agent work go src github com cockroachdb cockroach vendor golang org x sync errgroup errgroup go runtime goexit usr local go src runtime asm s wraps safe details enclosed wraps output in run workload run kv wraps home agent work go src github com cockroachdb cockroach bin roachprod run teamcity workload run kv init histograms perf stats json concurrency duration read percent sequential pgurl returned stderr workload error while loading shared libraries libncurses so cannot open shared object file no such file or directory error command problem exit status command problem wraps node command with error workload run kv init histograms perf stats json concurrency duration read percent sequential pgurl wraps exit status error types errors cmd hintdetail withdetail exec exiterror stdout wraps exit status error types withstack withstack safedetails withsafedetails errutil withmessage main withcommanddetails exec exiterror cluster go kv go kv go test runner go monitor failure monitor task failed t fatal was called attached stack trace main monitor waite home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go main monitor wait home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go main registerkv home agent work go src github com cockroachdb cockroach pkg cmd roachtest kv go main registerkv home agent work go src github com cockroachdb cockroach pkg cmd roachtest kv go main testrunner runtest home agent work go src github com cockroachdb cockroach pkg cmd roachtest test runner go wraps monitor failure wraps attached stack trace main monitor wait home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go wraps monitor task failed wraps attached stack trace main init home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go runtime doinit usr local go src runtime proc go runtime main usr local go src runtime proc go runtime goexit usr local go src runtime asm s wraps t fatal was called error types withstack withstack errutil withmessage withstack withstack errutil withmessage withstack withstack errors errorstring more artifacts related roachtest enc false nodes cpu seq failed roachtest enc false nodes cpu seq failed roachtest enc false nodes cpu seq failed powered by | 1 |
422,804 | 12,287,484,931 | IssuesEvent | 2020-05-09 12:25:33 | googleapis/elixir-google-api | https://api.github.com/repos/googleapis/elixir-google-api | opened | Synthesis failed for IAMCredentials | autosynth failure priority: p1 type: bug | Hello! Autosynth couldn't regenerate IAMCredentials. :broken_heart:
Here's the output from running `synth.py`:
```
2020-05-09 05:14:26 [INFO] logs will be written to: /tmpfs/src/github/synthtool/logs/googleapis/elixir-google-api
2020-05-09 05:14:26,326 autosynth > logs will be written to: /tmpfs/src/github/synthtool/logs/googleapis/elixir-google-api
Switched to branch 'autosynth-iamcredentials'
2020-05-09 05:14:27 [INFO] Running synthtool
2020-05-09 05:14:27,964 autosynth > Running synthtool
2020-05-09 05:14:27 [INFO] ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/iam_credentials/synth.metadata', 'synth.py', '--']
2020-05-09 05:14:27,964 autosynth > ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/iam_credentials/synth.metadata', 'synth.py', '--']
2020-05-09 05:14:28,174 synthtool > Executing /home/kbuilder/.cache/synthtool/elixir-google-api/synth.py.
On branch autosynth-iamcredentials
nothing to commit, working tree clean
2020-05-09 05:14:28,464 synthtool > Cloning https://github.com/googleapis/elixir-google-api.git.
2020-05-09 05:14:28,945 synthtool > Running: docker run --rm -v/home/kbuilder/.cache/synthtool/elixir-google-api:/workspace -v/var/run/docker.sock:/var/run/docker.sock -e USER_GROUP=1000:1000 -w /workspace gcr.io/cloud-devrel-public-resources/elixir19 scripts/generate_client.sh IAMCredentials
2020-05-09 05:14:32,556 synthtool > No files in sources /home/kbuilder/.cache/synthtool/elixir-google-api/clients were copied. Does the source contain files?
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module>
main()
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 829, in __call__
return self.main(*args, **kwargs)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 782, in main
rv = self.invoke(ctx)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1066, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 610, in invoke
return callback(*args, **kwargs)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main
spec.loader.exec_module(synth_module) # type: ignore
File "/tmpfs/src/github/synthtool/synthtool/metadata.py", line 180, in __exit__
write(self.metadata_file_path)
File "/tmpfs/src/github/synthtool/synthtool/metadata.py", line 112, in write
with open(outfile, "w") as fh:
FileNotFoundError: [Errno 2] No such file or directory: 'clients/iam_credentials/synth.metadata'
2020-05-09 05:14:32 [ERROR] Synthesis failed
2020-05-09 05:14:32,586 autosynth > Synthesis failed
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 599, in <module>
main()
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 471, in main
return _inner_main(temp_dir)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 549, in _inner_main
).synthesize(base_synth_log_path)
File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 118, in synthesize
synth_proc.check_returncode() # Raise an exception.
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode
self.stderr)
subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/iam_credentials/synth.metadata', 'synth.py', '--', 'IAMCredentials']' returned non-zero exit status 1.
```
Google internal developers can see the full log [here](https://sponge/11ff3741-9158-4831-8681-fff828f77e1a).
| 1.0 | Synthesis failed for IAMCredentials - Hello! Autosynth couldn't regenerate IAMCredentials. :broken_heart:
Here's the output from running `synth.py`:
```
2020-05-09 05:14:26 [INFO] logs will be written to: /tmpfs/src/github/synthtool/logs/googleapis/elixir-google-api
2020-05-09 05:14:26,326 autosynth > logs will be written to: /tmpfs/src/github/synthtool/logs/googleapis/elixir-google-api
Switched to branch 'autosynth-iamcredentials'
2020-05-09 05:14:27 [INFO] Running synthtool
2020-05-09 05:14:27,964 autosynth > Running synthtool
2020-05-09 05:14:27 [INFO] ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/iam_credentials/synth.metadata', 'synth.py', '--']
2020-05-09 05:14:27,964 autosynth > ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/iam_credentials/synth.metadata', 'synth.py', '--']
2020-05-09 05:14:28,174 synthtool > Executing /home/kbuilder/.cache/synthtool/elixir-google-api/synth.py.
On branch autosynth-iamcredentials
nothing to commit, working tree clean
2020-05-09 05:14:28,464 synthtool > Cloning https://github.com/googleapis/elixir-google-api.git.
2020-05-09 05:14:28,945 synthtool > Running: docker run --rm -v/home/kbuilder/.cache/synthtool/elixir-google-api:/workspace -v/var/run/docker.sock:/var/run/docker.sock -e USER_GROUP=1000:1000 -w /workspace gcr.io/cloud-devrel-public-resources/elixir19 scripts/generate_client.sh IAMCredentials
2020-05-09 05:14:32,556 synthtool > No files in sources /home/kbuilder/.cache/synthtool/elixir-google-api/clients were copied. Does the source contain files?
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module>
main()
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 829, in __call__
return self.main(*args, **kwargs)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 782, in main
rv = self.invoke(ctx)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1066, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 610, in invoke
return callback(*args, **kwargs)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main
spec.loader.exec_module(synth_module) # type: ignore
File "/tmpfs/src/github/synthtool/synthtool/metadata.py", line 180, in __exit__
write(self.metadata_file_path)
File "/tmpfs/src/github/synthtool/synthtool/metadata.py", line 112, in write
with open(outfile, "w") as fh:
FileNotFoundError: [Errno 2] No such file or directory: 'clients/iam_credentials/synth.metadata'
2020-05-09 05:14:32 [ERROR] Synthesis failed
2020-05-09 05:14:32,586 autosynth > Synthesis failed
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 599, in <module>
main()
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 471, in main
return _inner_main(temp_dir)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 549, in _inner_main
).synthesize(base_synth_log_path)
File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 118, in synthesize
synth_proc.check_returncode() # Raise an exception.
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode
self.stderr)
subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/iam_credentials/synth.metadata', 'synth.py', '--', 'IAMCredentials']' returned non-zero exit status 1.
```
Google internal developers can see the full log [here](https://sponge/11ff3741-9158-4831-8681-fff828f77e1a).
| non_test | synthesis failed for iamcredentials hello autosynth couldn t regenerate iamcredentials broken heart here s the output from running synth py logs will be written to tmpfs src github synthtool logs googleapis elixir google api autosynth logs will be written to tmpfs src github synthtool logs googleapis elixir google api switched to branch autosynth iamcredentials running synthtool autosynth running synthtool autosynth synthtool executing home kbuilder cache synthtool elixir google api synth py on branch autosynth iamcredentials nothing to commit working tree clean synthtool cloning synthtool running docker run rm v home kbuilder cache synthtool elixir google api workspace v var run docker sock var run docker sock e user group w workspace gcr io cloud devrel public resources scripts generate client sh iamcredentials synthtool no files in sources home kbuilder cache synthtool elixir google api clients were copied does the source contain files traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool synthtool main py line in main file tmpfs src github synthtool env lib site packages click core py line in call return self main args kwargs file tmpfs src github synthtool env lib site packages click core py line in main rv self invoke ctx file tmpfs src github synthtool env lib site packages click core py line in invoke return ctx invoke self callback ctx params file tmpfs src github synthtool env lib site packages click core py line in invoke return callback args kwargs file tmpfs src github synthtool synthtool main py line in main spec loader exec module synth module type ignore file tmpfs src github synthtool synthtool metadata py line in exit write self metadata file path file tmpfs src github synthtool synthtool metadata py line in write with open outfile w as fh filenotfounderror no such file or directory clients iam credentials synth metadata synthesis failed autosynth synthesis failed traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool autosynth synth py line in main file tmpfs src github synthtool autosynth synth py line in main return inner main temp dir file tmpfs src github synthtool autosynth synth py line in inner main synthesize base synth log path file tmpfs src github synthtool autosynth synthesizer py line in synthesize synth proc check returncode raise an exception file home kbuilder pyenv versions lib subprocess py line in check returncode self stderr subprocess calledprocesserror command returned non zero exit status google internal developers can see the full log | 0 |
173,362 | 13,398,307,305 | IssuesEvent | 2020-09-03 12:59:02 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | roachtest: jobs/mixed-versions failed | C-test-failure O-roachtest O-robot branch-provisional_202007220233_v20.2.0-alpha.2 release-blocker | [(roachtest).jobs/mixed-versions failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2107811&tab=buildLog) on [provisional_202007220233_v20.2.0-alpha.2@d3119926d33d808c6384cf3e99a7f7435f395489](https://github.com/cockroachdb/cockroach/commits/d3119926d33d808c6384cf3e99a7f7435f395489):
```
The test failed on branch=provisional_202007220233_v20.2.0-alpha.2, cloud=gce:
test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/jobs/mixed-versions/run_1
mixed_version_jobs.go:198,versionupgrade.go:167,mixed_version_jobs.go:296,mixed_version_jobs.go:320,test_runner.go:757: Cluster info
Node 1: 20.1
Node 2: 20.1-9
Node 3: 20.1-9
Node 4: 20.1
Unsuccessful job 574495805866573825 of type IMPORT, description IMPORT TABLE tpcc.public.customer (c_id INT8 NOT NULL, c_d_id INT8 NOT NULL, c_w_id INT8 NOT NULL, c_first VARCHAR(16), c_middle CHAR(2), c_last VARCHAR(16), c_street_1 VARCHAR(20), c_street_2 VARCHAR(20), c_city VARCHAR(20), c_state CHAR(2), c_zip CHAR(9), c_phone CHAR(16), c_since TIMESTAMP, c_credit CHAR(2), c_credit_lim DECIMAL(12,2), c_discount DECIMAL(4,4), c_balance DECIMAL(12,2), c_ytd_payment DECIMAL(12,2), c_payment_cnt INT8, c_delivery_cnt INT8, c_data VARCHAR(500), PRIMARY KEY (c_w_id, c_d_id, c_id), INDEX customer_idx (c_w_id, c_d_id, c_last, c_first)) CSV DATA ('workload:///csv/tpcc/customer?fks=true&interleaved=false&row-end=1500000&row-start=0&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/customer?fks=true&interleaved=false&row-end=3000000&row-start=1500000&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/customer?fks=true&interleaved=false&row-end=4500000&row-start=3000000&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/customer?fks=true&interleaved=false&row-end=6000000&row-start=4500000&seed=1&version=2.1.0&warehouses=200') WITH "nullif" = 'NULL', status failed, error version mismatch in flow request: 30; this node accepts 27 through 28, coordinator 3
Unsuccessful job 574495805866868737 of type IMPORT, description IMPORT TABLE tpcc.public.stock (s_i_id INT8 NOT NULL, s_w_id INT8 NOT NULL, s_quantity INT8, s_dist_01 CHAR(24), s_dist_02 CHAR(24), s_dist_03 CHAR(24), s_dist_04 CHAR(24), s_dist_05 CHAR(24), s_dist_06 CHAR(24), s_dist_07 CHAR(24), s_dist_08 CHAR(24), s_dist_09 CHAR(24), s_dist_10 CHAR(24), s_ytd INT8, s_order_cnt INT8, s_remote_cnt INT8, s_data VARCHAR(50), PRIMARY KEY (s_w_id, s_i_id), INDEX stock_item_fk_idx (s_i_id)) CSV DATA ('workload:///csv/tpcc/stock?fks=true&interleaved=false&row-end=5000000&row-start=0&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/stock?fks=true&interleaved=false&row-end=10000000&row-start=5000000&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/stock?fks=true&interleaved=false&row-end=15000000&row-start=10000000&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/stock?fks=true&interleaved=false&row-end=20000000&row-start=15000000&seed=1&version=2.1.0&warehouses=200') WITH "nullif" = 'NULL', status failed, error version mismatch in flow request: 30; this node accepts 27 through 28, coordinator 3
Unsuccessful job 574495805867360257 of type IMPORT, description IMPORT TABLE tpcc.public.history (rowid UUID NOT NULL DEFAULT gen_random_uuid(), h_c_id INT8 NOT NULL, h_c_d_id INT8 NOT NULL, h_c_w_id INT8 NOT NULL, h_d_id INT8 NOT NULL, h_w_id INT8 NOT NULL, h_date TIMESTAMP, h_amount DECIMAL(6,2), h_data VARCHAR(24), PRIMARY KEY (h_w_id, rowid), INDEX history_customer_fk_idx (h_c_w_id, h_c_d_id, h_c_id), INDEX history_district_fk_idx (h_w_id, h_d_id)) CSV DATA ('workload:///csv/tpcc/history?fks=true&interleaved=false&row-end=1500000&row-start=0&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/history?fks=true&interleaved=false&row-end=3000000&row-start=1500000&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/history?fks=true&interleaved=false&row-end=4500000&row-start=3000000&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/history?fks=true&interleaved=false&row-end=6000000&row-start=4500000&seed=1&version=2.1.0&warehouses=200') WITH "nullif" = 'NULL', status failed, error version mismatch in flow request: 30; this node accepts 27 through 28, coordinator 3
Unsuccessful job 574495805868441601 of type IMPORT, description IMPORT TABLE tpcc.public.order_line (ol_o_id INT8 NOT NULL, ol_d_id INT8 NOT NULL, ol_w_id INT8 NOT NULL, ol_number INT8 NOT NULL, ol_i_id INT8 NOT NULL, ol_supply_w_id INT8, ol_delivery_d TIMESTAMP, ol_quantity INT8, ol_amount DECIMAL(6,2), ol_dist_info CHAR(24), PRIMARY KEY (ol_w_id, ol_d_id, ol_o_id DESC, ol_number), INDEX order_line_stock_fk_idx (ol_supply_w_id, ol_i_id)) CSV DATA ('workload:///csv/tpcc/order_line?fks=true&interleaved=false&row-end=1500000&row-start=0&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/order_line?fks=true&interleaved=false&row-end=3000000&row-start=1500000&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/order_line?fks=true&interleaved=false&row-end=4500000&row-start=3000000&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/order_line?fks=true&interleaved=false&row-end=6000000&row-start=4500000&seed=1&version=2.1.0&warehouses=200') WITH "nullif" = 'NULL', status failed, error version mismatch in flow request: 30; this node accepts 27 through 28, coordinator 3
Unsuccessful job 574495805872570369 of type IMPORT, description IMPORT TABLE tpcc.public.item (i_id INT8 NOT NULL, i_im_id INT8, i_name VARCHAR(24), i_price DECIMAL(5,2), i_data VARCHAR(50), PRIMARY KEY (i_id)) CSV DATA ('workload:///csv/tpcc/item?fks=true&interleaved=false&row-end=25000&row-start=0&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/item?fks=true&interleaved=false&row-end=50000&row-start=25000&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/item?fks=true&interleaved=false&row-end=75000&row-start=50000&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/item?fks=true&interleaved=false&row-end=100000&row-start=75000&seed=1&version=2.1.0&warehouses=200') WITH "nullif" = 'NULL', status failed, error job 574495805872570369: cannot be reverted, manual cleanup may be required: rolling back tables: unexpected value: raw_bytes:"\026\334\374M\003\n\232\003\n\004item\030\250\001 \242\001(\002:\n\010\243\275\234\344\366\243\376\221\026B\032\n\004i_id\020\001\032\014\010\001\020@\030\0000\000P\024`\000 \0000\000B\035\n\007i_im_id\020\002\032\014\010\001\020@\030\0000\000P\024`\000 \0010\000B\035\n\006i_name\020\003\032\r\010\007\020\030\030\0000\007P\223\010`\000 \0010\000B\036\n\007i_price\020\004\032\r\010\003\020\002\030\0050\000P\244\r`\000 \0010\000B\035\n\006i_data\020\005\032\r\010\007\0202\030\0000\007P\223\010`\000 \0010\000H\006RG\n\007primary\020\001\030\001\"\004i_id0\001@\000J\020\010\000\020\000\032\000 \000(\0000\0008\000@\000Z\000z\002\010\000\200\001\000\210\001\000\220\001\001\230\001\000\242\001\006\010\000\022\000\030\000\250\001\000`\002j\025\n\t\n\005admin\020\002\n\010\n\004root\020\002\200\001\001\210\001\003\230\001\000\262\001?\n\007primary\020\000\032\004i_id\032\007i_im_id\032\006i_name\032\007i_price\032\006i_data \001 \002 \003 \004 \005(\000\270\001\001\302\001\000\350\001\000\362\001\004\010\000\022\000\370\001\000\200\002\000\222\002\000\232\002\n\010\243\275\234\344\366\243\376\221\026\262\002\timporting\270\002\000\300\002\035" timestamp:<wall_time:1595392626971668696 logical:1 > , coordinator 3
Unsuccessful job 574495805872603137 of type IMPORT, description IMPORT TABLE tpcc.public."order" (o_id INT8 NOT NULL, o_d_id INT8 NOT NULL, o_w_id INT8 NOT NULL, o_c_id INT8, o_entry_d TIMESTAMP, o_carrier_id INT8, o_ol_cnt INT8, o_all_local INT8, PRIMARY KEY (o_w_id, o_d_id, o_id DESC), CONSTRAINT order_idx UNIQUE (o_w_id, o_d_id, o_c_id, o_id DESC) STORING (o_entry_d, o_carrier_id)) CSV DATA ('workload:///csv/tpcc/order?fks=true&interleaved=false&row-end=1500000&row-start=0&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/order?fks=true&interleaved=false&row-end=3000000&row-start=1500000&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/order?fks=true&interleaved=false&row-end=4500000&row-start=3000000&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/order?fks=true&interleaved=false&row-end=6000000&row-start=4500000&seed=1&version=2.1.0&warehouses=200') WITH "nullif" = 'NULL', status failed, error version mismatch in flow request: 30; this node accepts 27 through 28, coordinator 3
```
<details><summary>More</summary><p>
Artifacts: [/jobs/mixed-versions](https://teamcity.cockroachdb.com/viewLog.html?buildId=2107811&tab=artifacts#/jobs/mixed-versions)
Related:
- #51186 roachtest: jobs/mixed-versions failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202007081918_v20.2.0-alpha.2](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202007081918_v20.2.0-alpha.2) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #51100 roachtest: jobs/mixed-versions failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202007071743_v20.2.0-alpha.2](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202007071743_v20.2.0-alpha.2) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #50026 roachtest: jobs/mixed-versions failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202006032224_v20.2.0-alpha.1](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202006032224_v20.2.0-alpha.1) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #49281 roachtest: jobs/mixed-versions failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202005191400_v20.1.1](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202005191400_v20.1.1) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #49233 roachtest: jobs/mixed-versions failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202005182011_v20.1.1](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202005182011_v20.1.1) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #48407 roachtest: jobs/mixed-versions failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202005041945_v19.1.9](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202005041945_v19.1.9) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #48315 roachtest: jobs/mixed-versions failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-master](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-master) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #48194 roachtest: jobs/mixed-versions failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-20.1](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-20.1) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #48193 roachtest: jobs/mixed-versions failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-19.1](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-19.1) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Ajobs%2Fmixed-versions.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
| 2.0 | roachtest: jobs/mixed-versions failed - [(roachtest).jobs/mixed-versions failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2107811&tab=buildLog) on [provisional_202007220233_v20.2.0-alpha.2@d3119926d33d808c6384cf3e99a7f7435f395489](https://github.com/cockroachdb/cockroach/commits/d3119926d33d808c6384cf3e99a7f7435f395489):
```
The test failed on branch=provisional_202007220233_v20.2.0-alpha.2, cloud=gce:
test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/jobs/mixed-versions/run_1
mixed_version_jobs.go:198,versionupgrade.go:167,mixed_version_jobs.go:296,mixed_version_jobs.go:320,test_runner.go:757: Cluster info
Node 1: 20.1
Node 2: 20.1-9
Node 3: 20.1-9
Node 4: 20.1
Unsuccessful job 574495805866573825 of type IMPORT, description IMPORT TABLE tpcc.public.customer (c_id INT8 NOT NULL, c_d_id INT8 NOT NULL, c_w_id INT8 NOT NULL, c_first VARCHAR(16), c_middle CHAR(2), c_last VARCHAR(16), c_street_1 VARCHAR(20), c_street_2 VARCHAR(20), c_city VARCHAR(20), c_state CHAR(2), c_zip CHAR(9), c_phone CHAR(16), c_since TIMESTAMP, c_credit CHAR(2), c_credit_lim DECIMAL(12,2), c_discount DECIMAL(4,4), c_balance DECIMAL(12,2), c_ytd_payment DECIMAL(12,2), c_payment_cnt INT8, c_delivery_cnt INT8, c_data VARCHAR(500), PRIMARY KEY (c_w_id, c_d_id, c_id), INDEX customer_idx (c_w_id, c_d_id, c_last, c_first)) CSV DATA ('workload:///csv/tpcc/customer?fks=true&interleaved=false&row-end=1500000&row-start=0&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/customer?fks=true&interleaved=false&row-end=3000000&row-start=1500000&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/customer?fks=true&interleaved=false&row-end=4500000&row-start=3000000&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/customer?fks=true&interleaved=false&row-end=6000000&row-start=4500000&seed=1&version=2.1.0&warehouses=200') WITH "nullif" = 'NULL', status failed, error version mismatch in flow request: 30; this node accepts 27 through 28, coordinator 3
Unsuccessful job 574495805866868737 of type IMPORT, description IMPORT TABLE tpcc.public.stock (s_i_id INT8 NOT NULL, s_w_id INT8 NOT NULL, s_quantity INT8, s_dist_01 CHAR(24), s_dist_02 CHAR(24), s_dist_03 CHAR(24), s_dist_04 CHAR(24), s_dist_05 CHAR(24), s_dist_06 CHAR(24), s_dist_07 CHAR(24), s_dist_08 CHAR(24), s_dist_09 CHAR(24), s_dist_10 CHAR(24), s_ytd INT8, s_order_cnt INT8, s_remote_cnt INT8, s_data VARCHAR(50), PRIMARY KEY (s_w_id, s_i_id), INDEX stock_item_fk_idx (s_i_id)) CSV DATA ('workload:///csv/tpcc/stock?fks=true&interleaved=false&row-end=5000000&row-start=0&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/stock?fks=true&interleaved=false&row-end=10000000&row-start=5000000&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/stock?fks=true&interleaved=false&row-end=15000000&row-start=10000000&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/stock?fks=true&interleaved=false&row-end=20000000&row-start=15000000&seed=1&version=2.1.0&warehouses=200') WITH "nullif" = 'NULL', status failed, error version mismatch in flow request: 30; this node accepts 27 through 28, coordinator 3
Unsuccessful job 574495805867360257 of type IMPORT, description IMPORT TABLE tpcc.public.history (rowid UUID NOT NULL DEFAULT gen_random_uuid(), h_c_id INT8 NOT NULL, h_c_d_id INT8 NOT NULL, h_c_w_id INT8 NOT NULL, h_d_id INT8 NOT NULL, h_w_id INT8 NOT NULL, h_date TIMESTAMP, h_amount DECIMAL(6,2), h_data VARCHAR(24), PRIMARY KEY (h_w_id, rowid), INDEX history_customer_fk_idx (h_c_w_id, h_c_d_id, h_c_id), INDEX history_district_fk_idx (h_w_id, h_d_id)) CSV DATA ('workload:///csv/tpcc/history?fks=true&interleaved=false&row-end=1500000&row-start=0&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/history?fks=true&interleaved=false&row-end=3000000&row-start=1500000&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/history?fks=true&interleaved=false&row-end=4500000&row-start=3000000&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/history?fks=true&interleaved=false&row-end=6000000&row-start=4500000&seed=1&version=2.1.0&warehouses=200') WITH "nullif" = 'NULL', status failed, error version mismatch in flow request: 30; this node accepts 27 through 28, coordinator 3
Unsuccessful job 574495805868441601 of type IMPORT, description IMPORT TABLE tpcc.public.order_line (ol_o_id INT8 NOT NULL, ol_d_id INT8 NOT NULL, ol_w_id INT8 NOT NULL, ol_number INT8 NOT NULL, ol_i_id INT8 NOT NULL, ol_supply_w_id INT8, ol_delivery_d TIMESTAMP, ol_quantity INT8, ol_amount DECIMAL(6,2), ol_dist_info CHAR(24), PRIMARY KEY (ol_w_id, ol_d_id, ol_o_id DESC, ol_number), INDEX order_line_stock_fk_idx (ol_supply_w_id, ol_i_id)) CSV DATA ('workload:///csv/tpcc/order_line?fks=true&interleaved=false&row-end=1500000&row-start=0&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/order_line?fks=true&interleaved=false&row-end=3000000&row-start=1500000&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/order_line?fks=true&interleaved=false&row-end=4500000&row-start=3000000&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/order_line?fks=true&interleaved=false&row-end=6000000&row-start=4500000&seed=1&version=2.1.0&warehouses=200') WITH "nullif" = 'NULL', status failed, error version mismatch in flow request: 30; this node accepts 27 through 28, coordinator 3
Unsuccessful job 574495805872570369 of type IMPORT, description IMPORT TABLE tpcc.public.item (i_id INT8 NOT NULL, i_im_id INT8, i_name VARCHAR(24), i_price DECIMAL(5,2), i_data VARCHAR(50), PRIMARY KEY (i_id)) CSV DATA ('workload:///csv/tpcc/item?fks=true&interleaved=false&row-end=25000&row-start=0&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/item?fks=true&interleaved=false&row-end=50000&row-start=25000&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/item?fks=true&interleaved=false&row-end=75000&row-start=50000&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/item?fks=true&interleaved=false&row-end=100000&row-start=75000&seed=1&version=2.1.0&warehouses=200') WITH "nullif" = 'NULL', status failed, error job 574495805872570369: cannot be reverted, manual cleanup may be required: rolling back tables: unexpected value: raw_bytes:"\026\334\374M\003\n\232\003\n\004item\030\250\001 \242\001(\002:\n\010\243\275\234\344\366\243\376\221\026B\032\n\004i_id\020\001\032\014\010\001\020@\030\0000\000P\024`\000 \0000\000B\035\n\007i_im_id\020\002\032\014\010\001\020@\030\0000\000P\024`\000 \0010\000B\035\n\006i_name\020\003\032\r\010\007\020\030\030\0000\007P\223\010`\000 \0010\000B\036\n\007i_price\020\004\032\r\010\003\020\002\030\0050\000P\244\r`\000 \0010\000B\035\n\006i_data\020\005\032\r\010\007\0202\030\0000\007P\223\010`\000 \0010\000H\006RG\n\007primary\020\001\030\001\"\004i_id0\001@\000J\020\010\000\020\000\032\000 \000(\0000\0008\000@\000Z\000z\002\010\000\200\001\000\210\001\000\220\001\001\230\001\000\242\001\006\010\000\022\000\030\000\250\001\000`\002j\025\n\t\n\005admin\020\002\n\010\n\004root\020\002\200\001\001\210\001\003\230\001\000\262\001?\n\007primary\020\000\032\004i_id\032\007i_im_id\032\006i_name\032\007i_price\032\006i_data \001 \002 \003 \004 \005(\000\270\001\001\302\001\000\350\001\000\362\001\004\010\000\022\000\370\001\000\200\002\000\222\002\000\232\002\n\010\243\275\234\344\366\243\376\221\026\262\002\timporting\270\002\000\300\002\035" timestamp:<wall_time:1595392626971668696 logical:1 > , coordinator 3
Unsuccessful job 574495805872603137 of type IMPORT, description IMPORT TABLE tpcc.public."order" (o_id INT8 NOT NULL, o_d_id INT8 NOT NULL, o_w_id INT8 NOT NULL, o_c_id INT8, o_entry_d TIMESTAMP, o_carrier_id INT8, o_ol_cnt INT8, o_all_local INT8, PRIMARY KEY (o_w_id, o_d_id, o_id DESC), CONSTRAINT order_idx UNIQUE (o_w_id, o_d_id, o_c_id, o_id DESC) STORING (o_entry_d, o_carrier_id)) CSV DATA ('workload:///csv/tpcc/order?fks=true&interleaved=false&row-end=1500000&row-start=0&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/order?fks=true&interleaved=false&row-end=3000000&row-start=1500000&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/order?fks=true&interleaved=false&row-end=4500000&row-start=3000000&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/order?fks=true&interleaved=false&row-end=6000000&row-start=4500000&seed=1&version=2.1.0&warehouses=200') WITH "nullif" = 'NULL', status failed, error version mismatch in flow request: 30; this node accepts 27 through 28, coordinator 3
```
<details><summary>More</summary><p>
Artifacts: [/jobs/mixed-versions](https://teamcity.cockroachdb.com/viewLog.html?buildId=2107811&tab=artifacts#/jobs/mixed-versions)
Related:
- #51186 roachtest: jobs/mixed-versions failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202007081918_v20.2.0-alpha.2](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202007081918_v20.2.0-alpha.2) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #51100 roachtest: jobs/mixed-versions failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202007071743_v20.2.0-alpha.2](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202007071743_v20.2.0-alpha.2) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #50026 roachtest: jobs/mixed-versions failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202006032224_v20.2.0-alpha.1](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202006032224_v20.2.0-alpha.1) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #49281 roachtest: jobs/mixed-versions failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202005191400_v20.1.1](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202005191400_v20.1.1) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #49233 roachtest: jobs/mixed-versions failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202005182011_v20.1.1](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202005182011_v20.1.1) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #48407 roachtest: jobs/mixed-versions failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202005041945_v19.1.9](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202005041945_v19.1.9) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #48315 roachtest: jobs/mixed-versions failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-master](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-master) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #48194 roachtest: jobs/mixed-versions failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-20.1](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-20.1) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #48193 roachtest: jobs/mixed-versions failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-19.1](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-19.1) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Ajobs%2Fmixed-versions.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
| test | roachtest jobs mixed versions failed on the test failed on branch provisional alpha cloud gce test artifacts and logs in home agent work go src github com cockroachdb cockroach artifacts jobs mixed versions run mixed version jobs go versionupgrade go mixed version jobs go mixed version jobs go test runner go cluster info node node node node unsuccessful job of type import description import table tpcc public customer c id not null c d id not null c w id not null c first varchar c middle char c last varchar c street varchar c street varchar c city varchar c state char c zip char c phone char c since timestamp c credit char c credit lim decimal c discount decimal c balance decimal c ytd payment decimal c payment cnt c delivery cnt c data varchar primary key c w id c d id c id index customer idx c w id c d id c last c first csv data workload csv tpcc customer fks true interleaved false row end row start seed version warehouses workload csv tpcc customer fks true interleaved false row end row start seed version warehouses workload csv tpcc customer fks true interleaved false row end row start seed version warehouses workload csv tpcc customer fks true interleaved false row end row start seed version warehouses with nullif null status failed error version mismatch in flow request this node accepts through coordinator unsuccessful job of type import description import table tpcc public stock s i id not null s w id not null s quantity s dist char s dist char s dist char s dist char s dist char s dist char s dist char s dist char s dist char s dist char s ytd s order cnt s remote cnt s data varchar primary key s w id s i id index stock item fk idx s i id csv data workload csv tpcc stock fks true interleaved false row end row start seed version warehouses workload csv tpcc stock fks true interleaved false row end row start seed version warehouses workload csv tpcc stock fks true interleaved false row end row start seed version warehouses workload csv tpcc stock fks true interleaved false row end row start seed version warehouses with nullif null status failed error version mismatch in flow request this node accepts through coordinator unsuccessful job of type import description import table tpcc public history rowid uuid not null default gen random uuid h c id not null h c d id not null h c w id not null h d id not null h w id not null h date timestamp h amount decimal h data varchar primary key h w id rowid index history customer fk idx h c w id h c d id h c id index history district fk idx h w id h d id csv data workload csv tpcc history fks true interleaved false row end row start seed version warehouses workload csv tpcc history fks true interleaved false row end row start seed version warehouses workload csv tpcc history fks true interleaved false row end row start seed version warehouses workload csv tpcc history fks true interleaved false row end row start seed version warehouses with nullif null status failed error version mismatch in flow request this node accepts through coordinator unsuccessful job of type import description import table tpcc public order line ol o id not null ol d id not null ol w id not null ol number not null ol i id not null ol supply w id ol delivery d timestamp ol quantity ol amount decimal ol dist info char primary key ol w id ol d id ol o id desc ol number index order line stock fk idx ol supply w id ol i id csv data workload csv tpcc order line fks true interleaved false row end row start seed version warehouses workload csv tpcc order line fks true interleaved false row end row start seed version warehouses workload csv tpcc order line fks true interleaved false row end row start seed version warehouses workload csv tpcc order line fks true interleaved false row end row start seed version warehouses with nullif null status failed error version mismatch in flow request this node accepts through coordinator unsuccessful job of type import description import table tpcc public item i id not null i im id i name varchar i price decimal i data varchar primary key i id csv data workload csv tpcc item fks true interleaved false row end row start seed version warehouses workload csv tpcc item fks true interleaved false row end row start seed version warehouses workload csv tpcc item fks true interleaved false row end row start seed version warehouses workload csv tpcc item fks true interleaved false row end row start seed version warehouses with nullif null status failed error job cannot be reverted manual cleanup may be required rolling back tables unexpected value raw bytes n n n n id n im id n name r n price r r n data r n n t n n n n id im id name price data n timporting timestamp coordinator unsuccessful job of type import description import table tpcc public order o id not null o d id not null o w id not null o c id o entry d timestamp o carrier id o ol cnt o all local primary key o w id o d id o id desc constraint order idx unique o w id o d id o c id o id desc storing o entry d o carrier id csv data workload csv tpcc order fks true interleaved false row end row start seed version warehouses workload csv tpcc order fks true interleaved false row end row start seed version warehouses workload csv tpcc order fks true interleaved false row end row start seed version warehouses workload csv tpcc order fks true interleaved false row end row start seed version warehouses with nullif null status failed error version mismatch in flow request this node accepts through coordinator more artifacts related roachtest jobs mixed versions failed roachtest jobs mixed versions failed roachtest jobs mixed versions failed roachtest jobs mixed versions failed roachtest jobs mixed versions failed roachtest jobs mixed versions failed roachtest jobs mixed versions failed roachtest jobs mixed versions failed roachtest jobs mixed versions failed powered by | 1 |
86,305 | 8,031,396,503 | IssuesEvent | 2018-07-28 01:02:58 | rancher/rancher | https://api.github.com/repos/rancher/rancher | closed | Increasing the project resource quota does not adjust the quota on the namespace | kind/bug status/resolved status/to-test version/2.0 | **Rancher versions:**
rancher/server or rancher/rancher: master
**Steps to Reproduce:**
1. Create a Project with resource quota as below:
```
{"limit":{"pods": "10",
"services": "10",
"replicationControllers": "10",
"secrets": "10",
"configMaps": "10",
"persistentVolumeClaims": "10",
"servicesNodePorts": "100",
"servicesLoadBalancers": "10",
"requestsCpu": "2000m",
"requestsMemory": "100Gi",
"limitsCpu": "2000m",
"limitsMemory": "100Gi"
}}
```
2. Create a Resourcequota template as below:
```
{"pods": "10",
"services": "10",
"replicationControllers": "10",
"secrets": "10",
"configMaps": "10",
"persistentVolumeClaims": "10",
"servicesNodePorts": "100",
"servicesLoadBalancers": "10",
"requestsCpu": "1000m",
"requestsMemory": "10Gi",
"limitsCpu": "1000m",
"limitsMemory": "50Gi"
}
```
3. Create namespace/workloads exceeding the cpu resource quota lets say 2000m.
Workload creation will fail.
4. Edit and Increase the project cpu resource quota on the project to 3000m as below:
```
{"limit":{"pods": "10",
"services": "10",
"replicationControllers": "10",
"secrets": "10",
"configMaps": "10",
"persistentVolumeClaims": "10",
"servicesNodePorts": "100",
"servicesLoadBalancers": "10",
"requestsCpu": "3000m",
"requestsMemory": "100Gi",
"limitsCpu": "3000m",
"limitsMemory": "100Gi"
}}
```
Increasing the project quota does not adjust the quota on the namespace. The pod creation does not continue within the allowed new updated quota
<img width="1187" alt="screen shot 2018-07-25 at 4 18 25 pm" src="https://user-images.githubusercontent.com/18536626/43232574-85fd6fbc-9026-11e8-85df-4119c8bdc870.png">
| 1.0 | Increasing the project resource quota does not adjust the quota on the namespace - **Rancher versions:**
rancher/server or rancher/rancher: master
**Steps to Reproduce:**
1. Create a Project with resource quota as below:
```
{"limit":{"pods": "10",
"services": "10",
"replicationControllers": "10",
"secrets": "10",
"configMaps": "10",
"persistentVolumeClaims": "10",
"servicesNodePorts": "100",
"servicesLoadBalancers": "10",
"requestsCpu": "2000m",
"requestsMemory": "100Gi",
"limitsCpu": "2000m",
"limitsMemory": "100Gi"
}}
```
2. Create a Resourcequota template as below:
```
{"pods": "10",
"services": "10",
"replicationControllers": "10",
"secrets": "10",
"configMaps": "10",
"persistentVolumeClaims": "10",
"servicesNodePorts": "100",
"servicesLoadBalancers": "10",
"requestsCpu": "1000m",
"requestsMemory": "10Gi",
"limitsCpu": "1000m",
"limitsMemory": "50Gi"
}
```
3. Create namespace/workloads exceeding the cpu resource quota lets say 2000m.
Workload creation will fail.
4. Edit and Increase the project cpu resource quota on the project to 3000m as below:
```
{"limit":{"pods": "10",
"services": "10",
"replicationControllers": "10",
"secrets": "10",
"configMaps": "10",
"persistentVolumeClaims": "10",
"servicesNodePorts": "100",
"servicesLoadBalancers": "10",
"requestsCpu": "3000m",
"requestsMemory": "100Gi",
"limitsCpu": "3000m",
"limitsMemory": "100Gi"
}}
```
Increasing the project quota does not adjust the quota on the namespace. The pod creation does not continue within the allowed new updated quota
<img width="1187" alt="screen shot 2018-07-25 at 4 18 25 pm" src="https://user-images.githubusercontent.com/18536626/43232574-85fd6fbc-9026-11e8-85df-4119c8bdc870.png">
| test | increasing the project resource quota does not adjust the quota on the namespace rancher versions rancher server or rancher rancher master steps to reproduce create a project with resource quota as below limit pods services replicationcontrollers secrets configmaps persistentvolumeclaims servicesnodeports servicesloadbalancers requestscpu requestsmemory limitscpu limitsmemory create a resourcequota template as below pods services replicationcontrollers secrets configmaps persistentvolumeclaims servicesnodeports servicesloadbalancers requestscpu requestsmemory limitscpu limitsmemory create namespace workloads exceeding the cpu resource quota lets say workload creation will fail edit and increase the project cpu resource quota on the project to as below limit pods services replicationcontrollers secrets configmaps persistentvolumeclaims servicesnodeports servicesloadbalancers requestscpu requestsmemory limitscpu limitsmemory increasing the project quota does not adjust the quota on the namespace the pod creation does not continue within the allowed new updated quota img width alt screen shot at pm src | 1 |
32,673 | 4,781,803,019 | IssuesEvent | 2016-10-28 10:44:49 | itheof/42sh-posix | https://api.github.com/repos/itheof/42sh-posix | closed | Test de retour NULL de malloc: idées | help wanted question test | Je pense qu'il est possible de tester les retours NULL de malloc de manière propre.
On pourrait charger notre propre librarie dynamique malloc (contenant malloc, realloc, free...) et 'wrapper' les vrais fonctions du système (conflicts de noms ?), en les glissants dans ce genre de scope:
```c
#ifnedef
# define MALLOC_CRASH_AT 0 // each compilation increment this
#endif
static g_malloc_count = 0;
void *malloc(size_t size)
{
if (g_malloc_count == MALLOC_CRASH)
{
++g_malloc_count;
return (NULL);
}
++g_malloc_count;
return (malloc(size)); // this is the real malloc but name conflicts...
}
```
`MALLOC_CRASH_AT` est incrémenté à chaque compilation et à chaque fois le programme passe une série de tests | 1.0 | Test de retour NULL de malloc: idées - Je pense qu'il est possible de tester les retours NULL de malloc de manière propre.
On pourrait charger notre propre librarie dynamique malloc (contenant malloc, realloc, free...) et 'wrapper' les vrais fonctions du système (conflicts de noms ?), en les glissants dans ce genre de scope:
```c
#ifnedef
# define MALLOC_CRASH_AT 0 // each compilation increment this
#endif
static g_malloc_count = 0;
void *malloc(size_t size)
{
if (g_malloc_count == MALLOC_CRASH)
{
++g_malloc_count;
return (NULL);
}
++g_malloc_count;
return (malloc(size)); // this is the real malloc but name conflicts...
}
```
`MALLOC_CRASH_AT` est incrémenté à chaque compilation et à chaque fois le programme passe une série de tests | test | test de retour null de malloc idées je pense qu il est possible de tester les retours null de malloc de manière propre on pourrait charger notre propre librarie dynamique malloc contenant malloc realloc free et wrapper les vrais fonctions du système conflicts de noms en les glissants dans ce genre de scope c ifnedef define malloc crash at each compilation increment this endif static g malloc count void malloc size t size if g malloc count malloc crash g malloc count return null g malloc count return malloc size this is the real malloc but name conflicts malloc crash at est incrémenté à chaque compilation et à chaque fois le programme passe une série de tests | 1 |
4,586 | 7,428,876,262 | IssuesEvent | 2018-03-24 07:33:06 | kookmin-sw/2018-cap1-2 | https://api.github.com/repos/kookmin-sw/2018-cap1-2 | opened | 암묵적 의미가 동일한 기호처리 | ImageProcessing | for 구문에서 동일한 의미를 가지는 -> ~ 기호가 들어왔을 때,
<현재>
- 인터프리터 대상에서 제외되는 -> 기호를 그대로 반환해서 로지컬 에러 발생
이를 영상처리 단계에서 -> 기호를 ~ 기호로 대체 변환해줘
인터프리팅 성공 가능성을 높이는 것에 대해 고려.
+ >=, <=, != 기호를 하나의 기호로 합치는 것이 기호 표기 상 가능하다면
해당 기호들도 고려 대상에 포함. | 1.0 | 암묵적 의미가 동일한 기호처리 - for 구문에서 동일한 의미를 가지는 -> ~ 기호가 들어왔을 때,
<현재>
- 인터프리터 대상에서 제외되는 -> 기호를 그대로 반환해서 로지컬 에러 발생
이를 영상처리 단계에서 -> 기호를 ~ 기호로 대체 변환해줘
인터프리팅 성공 가능성을 높이는 것에 대해 고려.
+ >=, <=, != 기호를 하나의 기호로 합치는 것이 기호 표기 상 가능하다면
해당 기호들도 고려 대상에 포함. | non_test | 암묵적 의미가 동일한 기호처리 for 구문에서 동일한 의미를 가지는 기호가 들어왔을 때 인터프리터 대상에서 제외되는 기호를 그대로 반환해서 로지컬 에러 발생 이를 영상처리 단계에서 기호를 기호로 대체 변환해줘 인터프리팅 성공 가능성을 높이는 것에 대해 고려 기호를 하나의 기호로 합치는 것이 기호 표기 상 가능하다면 해당 기호들도 고려 대상에 포함 | 0 |
233,067 | 18,945,405,528 | IssuesEvent | 2021-11-18 09:38:33 | open-telemetry/opentelemetry-collector-contrib | https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib | opened | Flaky test: ??? on tailsamplingprocessor | bug flaky test | Seen on https://github.com/open-telemetry/opentelemetry-collector-contrib/runs/4243284022?check_suite_focus=true
Full logs: [logs_18069.zip](https://github.com/open-telemetry/opentelemetry-collector-contrib/files/7561721/logs_18069.zip)
```
=== RUN TestStatusCodeSampling/filter_on_OK_and_UNSET_-_matches
--- PASS: TestStatusCodeSampling (0.00s)
--- PASS: TestStatusCodeSampling/filter_on_ERROR_-_none_match (0.00s)
--- PASS: TestStatusCodeSampling/filter_on_OK_and_ERROR_-_none_match (0.00s)
make[2]: *** [../../Makefile.Common:45: test] Error 1
make[1]: *** [Makefile:149: for-all-target-./processor/tailsamplingprocessor] Error 2
make: *** [Makefile:69: gotest] Error 2
--- PASS: TestStatusCodeSampling/filter_on_UNSET_-_matches (0.00s)
--- PASS: TestStatusCodeSampling/filter_on_OK_and_UNSET_-_matches (0.00s)
=== RUN TestOnLateArrivingSpans_StatusCodeSampling
--- PASS: TestOnLateArrivingSpans_StatusCodeSampling (0.00s)
=== RUN TestStringTagFilter
=== RUN TestStringTagFilter/nonmatching_node_attribute_key
=== RUN TestStringTagFilter/nonmatching_node_attribute_value
=== RUN TestStringTagFilter/matching_node_attribute
=== RUN TestStringTagFilter/nonmatching_span_attribute_key
=== RUN TestStringTagFilter/nonmatching_span_attribute_value
=== RUN TestStringTagFilter/matching_span_attribute
=== RUN TestStringTagFilter/matching_span_attribute_with_regex
=== RUN TestStringTagFilter/nonmatching_span_attribute_with_regex
=== RUN TestStringTagFilter/matching_span_attribute_with_regex_without_CacheSize_provided_in_config
=== RUN TestStringTagFilter/matching_plain_text_node_attribute_in_regex
=== RUN TestStringTagFilter/nonmatching_span_attribute_on_empty_filter_list
=== RUN TestStringTagFilter/invert_nonmatching_node_attribute_key
=== RUN TestStringTagFilter/invert_nonmatching_node_attribute_value
=== RUN TestStringTagFilter/invert_nonmatching_node_attribute_list
=== RUN TestStringTagFilter/invert_matching_node_attribute
=== RUN TestStringTagFilter/invert_matching_node_attribute_list
=== RUN TestStringTagFilter/invert_nonmatching_span_attribute_key
=== RUN TestStringTagFilter/invert_nonmatching_span_attribute_value
=== RUN TestStringTagFilter/invert_nonmatching_span_attribute_list
=== RUN TestStringTagFilter/invert_matching_span_attribute
=== RUN TestStringTagFilter/invert_matching_span_attribute_list
=== RUN TestStringTagFilter/invert_matching_span_attribute_with_regex
=== RUN TestStringTagFilter/invert_matching_span_attribute_with_regex_list
=== RUN TestStringTagFilter/invert_nonmatching_span_attribute_with_regex
=== RUN TestStringTagFilter/invert_nonmatching_span_attribute_with_regex_list
=== RUN TestStringTagFilter/invert_matching_plain_text_node_attribute_in_regex
=== RUN TestStringTagFilter/invert_matching_plain_text_node_attribute_in_regex_list
=== RUN TestStringTagFilter/invert_nonmatching_span_attribute_on_empty_filter_list
--- PASS: TestStringTagFilter (0.01s)
--- PASS: TestStringTagFilter/nonmatching_node_attribute_key (0.00s)
--- PASS: TestStringTagFilter/nonmatching_node_attribute_value (0.00s)
--- PASS: TestStringTagFilter/matching_node_attribute (0.00s)
--- PASS: TestStringTagFilter/nonmatching_span_attribute_key (0.00s)
--- PASS: TestStringTagFilter/nonmatching_span_attribute_value (0.00s)
--- PASS: TestStringTagFilter/matching_span_attribute (0.00s)
--- PASS: TestStringTagFilter/matching_span_attribute_with_regex (0.00s)
--- PASS: TestStringTagFilter/nonmatching_span_attribute_with_regex (0.00s)
--- PASS: TestStringTagFilter/matching_span_attribute_with_regex_without_CacheSize_provided_in_config (0.00s)
--- PASS: TestStringTagFilter/matching_plain_text_node_attribute_in_regex (0.00s)
--- PASS: TestStringTagFilter/nonmatching_span_attribute_on_empty_filter_list (0.00s)
--- PASS: TestStringTagFilter/invert_nonmatching_node_attribute_key (0.00s)
--- PASS: TestStringTagFilter/invert_nonmatching_node_attribute_value (0.00s)
--- PASS: TestStringTagFilter/invert_nonmatching_node_attribute_list (0.00s)
--- PASS: TestStringTagFilter/invert_matching_node_attribute (0.00s)
--- PASS: TestStringTagFilter/invert_matching_node_attribute_list (0.00s)
--- PASS: TestStringTagFilter/invert_nonmatching_span_attribute_key (0.00s)
--- PASS: TestStringTagFilter/invert_nonmatching_span_attribute_value (0.00s)
--- PASS: TestStringTagFilter/invert_nonmatching_span_attribute_list (0.00s)
--- PASS: TestStringTagFilter/invert_matching_span_attribute (0.00s)
--- PASS: TestStringTagFilter/invert_matching_span_attribute_list (0.00s)
--- PASS: TestStringTagFilter/invert_matching_span_attribute_with_regex (0.00s)
--- PASS: TestStringTagFilter/invert_matching_span_attribute_with_regex_list (0.00s)
--- PASS: TestStringTagFilter/invert_nonmatching_span_attribute_with_regex (0.00s)
--- PASS: TestStringTagFilter/invert_nonmatching_span_attribute_with_regex_list (0.00s)
--- PASS: TestStringTagFilter/invert_matching_plain_text_node_attribute_in_regex (0.00s)
--- PASS: TestStringTagFilter/invert_matching_plain_text_node_attribute_in_regex_list (0.00s)
--- PASS: TestStringTagFilter/invert_nonmatching_span_attribute_on_empty_filter_list (0.00s)
=== RUN TestOnLateArrivingSpans_StringAttribute
--- PASS: TestOnLateArrivingSpans_StringAttribute (0.00s)
=== RUN TestTimeProvider
--- PASS: TestTimeProvider (0.00s)
PASS
ok github.com/open-telemetry/opentelemetry-collector-contrib/processor/tailsamplingprocessor/internal/sampling 4.407s
FAIL
make[2]: Leaving directory '/home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/tailsamplingprocessor'
make[1]: Leaving directory '/home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib'
``` | 1.0 | Flaky test: ??? on tailsamplingprocessor - Seen on https://github.com/open-telemetry/opentelemetry-collector-contrib/runs/4243284022?check_suite_focus=true
Full logs: [logs_18069.zip](https://github.com/open-telemetry/opentelemetry-collector-contrib/files/7561721/logs_18069.zip)
```
=== RUN TestStatusCodeSampling/filter_on_OK_and_UNSET_-_matches
--- PASS: TestStatusCodeSampling (0.00s)
--- PASS: TestStatusCodeSampling/filter_on_ERROR_-_none_match (0.00s)
--- PASS: TestStatusCodeSampling/filter_on_OK_and_ERROR_-_none_match (0.00s)
make[2]: *** [../../Makefile.Common:45: test] Error 1
make[1]: *** [Makefile:149: for-all-target-./processor/tailsamplingprocessor] Error 2
make: *** [Makefile:69: gotest] Error 2
--- PASS: TestStatusCodeSampling/filter_on_UNSET_-_matches (0.00s)
--- PASS: TestStatusCodeSampling/filter_on_OK_and_UNSET_-_matches (0.00s)
=== RUN TestOnLateArrivingSpans_StatusCodeSampling
--- PASS: TestOnLateArrivingSpans_StatusCodeSampling (0.00s)
=== RUN TestStringTagFilter
=== RUN TestStringTagFilter/nonmatching_node_attribute_key
=== RUN TestStringTagFilter/nonmatching_node_attribute_value
=== RUN TestStringTagFilter/matching_node_attribute
=== RUN TestStringTagFilter/nonmatching_span_attribute_key
=== RUN TestStringTagFilter/nonmatching_span_attribute_value
=== RUN TestStringTagFilter/matching_span_attribute
=== RUN TestStringTagFilter/matching_span_attribute_with_regex
=== RUN TestStringTagFilter/nonmatching_span_attribute_with_regex
=== RUN TestStringTagFilter/matching_span_attribute_with_regex_without_CacheSize_provided_in_config
=== RUN TestStringTagFilter/matching_plain_text_node_attribute_in_regex
=== RUN TestStringTagFilter/nonmatching_span_attribute_on_empty_filter_list
=== RUN TestStringTagFilter/invert_nonmatching_node_attribute_key
=== RUN TestStringTagFilter/invert_nonmatching_node_attribute_value
=== RUN TestStringTagFilter/invert_nonmatching_node_attribute_list
=== RUN TestStringTagFilter/invert_matching_node_attribute
=== RUN TestStringTagFilter/invert_matching_node_attribute_list
=== RUN TestStringTagFilter/invert_nonmatching_span_attribute_key
=== RUN TestStringTagFilter/invert_nonmatching_span_attribute_value
=== RUN TestStringTagFilter/invert_nonmatching_span_attribute_list
=== RUN TestStringTagFilter/invert_matching_span_attribute
=== RUN TestStringTagFilter/invert_matching_span_attribute_list
=== RUN TestStringTagFilter/invert_matching_span_attribute_with_regex
=== RUN TestStringTagFilter/invert_matching_span_attribute_with_regex_list
=== RUN TestStringTagFilter/invert_nonmatching_span_attribute_with_regex
=== RUN TestStringTagFilter/invert_nonmatching_span_attribute_with_regex_list
=== RUN TestStringTagFilter/invert_matching_plain_text_node_attribute_in_regex
=== RUN TestStringTagFilter/invert_matching_plain_text_node_attribute_in_regex_list
=== RUN TestStringTagFilter/invert_nonmatching_span_attribute_on_empty_filter_list
--- PASS: TestStringTagFilter (0.01s)
--- PASS: TestStringTagFilter/nonmatching_node_attribute_key (0.00s)
--- PASS: TestStringTagFilter/nonmatching_node_attribute_value (0.00s)
--- PASS: TestStringTagFilter/matching_node_attribute (0.00s)
--- PASS: TestStringTagFilter/nonmatching_span_attribute_key (0.00s)
--- PASS: TestStringTagFilter/nonmatching_span_attribute_value (0.00s)
--- PASS: TestStringTagFilter/matching_span_attribute (0.00s)
--- PASS: TestStringTagFilter/matching_span_attribute_with_regex (0.00s)
--- PASS: TestStringTagFilter/nonmatching_span_attribute_with_regex (0.00s)
--- PASS: TestStringTagFilter/matching_span_attribute_with_regex_without_CacheSize_provided_in_config (0.00s)
--- PASS: TestStringTagFilter/matching_plain_text_node_attribute_in_regex (0.00s)
--- PASS: TestStringTagFilter/nonmatching_span_attribute_on_empty_filter_list (0.00s)
--- PASS: TestStringTagFilter/invert_nonmatching_node_attribute_key (0.00s)
--- PASS: TestStringTagFilter/invert_nonmatching_node_attribute_value (0.00s)
--- PASS: TestStringTagFilter/invert_nonmatching_node_attribute_list (0.00s)
--- PASS: TestStringTagFilter/invert_matching_node_attribute (0.00s)
--- PASS: TestStringTagFilter/invert_matching_node_attribute_list (0.00s)
--- PASS: TestStringTagFilter/invert_nonmatching_span_attribute_key (0.00s)
--- PASS: TestStringTagFilter/invert_nonmatching_span_attribute_value (0.00s)
--- PASS: TestStringTagFilter/invert_nonmatching_span_attribute_list (0.00s)
--- PASS: TestStringTagFilter/invert_matching_span_attribute (0.00s)
--- PASS: TestStringTagFilter/invert_matching_span_attribute_list (0.00s)
--- PASS: TestStringTagFilter/invert_matching_span_attribute_with_regex (0.00s)
--- PASS: TestStringTagFilter/invert_matching_span_attribute_with_regex_list (0.00s)
--- PASS: TestStringTagFilter/invert_nonmatching_span_attribute_with_regex (0.00s)
--- PASS: TestStringTagFilter/invert_nonmatching_span_attribute_with_regex_list (0.00s)
--- PASS: TestStringTagFilter/invert_matching_plain_text_node_attribute_in_regex (0.00s)
--- PASS: TestStringTagFilter/invert_matching_plain_text_node_attribute_in_regex_list (0.00s)
--- PASS: TestStringTagFilter/invert_nonmatching_span_attribute_on_empty_filter_list (0.00s)
=== RUN TestOnLateArrivingSpans_StringAttribute
--- PASS: TestOnLateArrivingSpans_StringAttribute (0.00s)
=== RUN TestTimeProvider
--- PASS: TestTimeProvider (0.00s)
PASS
ok github.com/open-telemetry/opentelemetry-collector-contrib/processor/tailsamplingprocessor/internal/sampling 4.407s
FAIL
make[2]: Leaving directory '/home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/tailsamplingprocessor'
make[1]: Leaving directory '/home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib'
``` | test | flaky test on tailsamplingprocessor seen on full logs run teststatuscodesampling filter on ok and unset matches pass teststatuscodesampling pass teststatuscodesampling filter on error none match pass teststatuscodesampling filter on ok and error none match make error make error make error pass teststatuscodesampling filter on unset matches pass teststatuscodesampling filter on ok and unset matches run testonlatearrivingspans statuscodesampling pass testonlatearrivingspans statuscodesampling run teststringtagfilter run teststringtagfilter nonmatching node attribute key run teststringtagfilter nonmatching node attribute value run teststringtagfilter matching node attribute run teststringtagfilter nonmatching span attribute key run teststringtagfilter nonmatching span attribute value run teststringtagfilter matching span attribute run teststringtagfilter matching span attribute with regex run teststringtagfilter nonmatching span attribute with regex run teststringtagfilter matching span attribute with regex without cachesize provided in config run teststringtagfilter matching plain text node attribute in regex run teststringtagfilter nonmatching span attribute on empty filter list run teststringtagfilter invert nonmatching node attribute key run teststringtagfilter invert nonmatching node attribute value run teststringtagfilter invert nonmatching node attribute list run teststringtagfilter invert matching node attribute run teststringtagfilter invert matching node attribute list run teststringtagfilter invert nonmatching span attribute key run teststringtagfilter invert nonmatching span attribute value run teststringtagfilter invert nonmatching span attribute list run teststringtagfilter invert matching span attribute run teststringtagfilter invert matching span attribute list run teststringtagfilter invert matching span attribute with regex run teststringtagfilter invert matching span attribute with regex list run teststringtagfilter invert nonmatching span attribute with regex run teststringtagfilter invert nonmatching span attribute with regex list run teststringtagfilter invert matching plain text node attribute in regex run teststringtagfilter invert matching plain text node attribute in regex list run teststringtagfilter invert nonmatching span attribute on empty filter list pass teststringtagfilter pass teststringtagfilter nonmatching node attribute key pass teststringtagfilter nonmatching node attribute value pass teststringtagfilter matching node attribute pass teststringtagfilter nonmatching span attribute key pass teststringtagfilter nonmatching span attribute value pass teststringtagfilter matching span attribute pass teststringtagfilter matching span attribute with regex pass teststringtagfilter nonmatching span attribute with regex pass teststringtagfilter matching span attribute with regex without cachesize provided in config pass teststringtagfilter matching plain text node attribute in regex pass teststringtagfilter nonmatching span attribute on empty filter list pass teststringtagfilter invert nonmatching node attribute key pass teststringtagfilter invert nonmatching node attribute value pass teststringtagfilter invert nonmatching node attribute list pass teststringtagfilter invert matching node attribute pass teststringtagfilter invert matching node attribute list pass teststringtagfilter invert nonmatching span attribute key pass teststringtagfilter invert nonmatching span attribute value pass teststringtagfilter invert nonmatching span attribute list pass teststringtagfilter invert matching span attribute pass teststringtagfilter invert matching span attribute list pass teststringtagfilter invert matching span attribute with regex pass teststringtagfilter invert matching span attribute with regex list pass teststringtagfilter invert nonmatching span attribute with regex pass teststringtagfilter invert nonmatching span attribute with regex list pass teststringtagfilter invert matching plain text node attribute in regex pass teststringtagfilter invert matching plain text node attribute in regex list pass teststringtagfilter invert nonmatching span attribute on empty filter list run testonlatearrivingspans stringattribute pass testonlatearrivingspans stringattribute run testtimeprovider pass testtimeprovider pass ok github com open telemetry opentelemetry collector contrib processor tailsamplingprocessor internal sampling fail make leaving directory home runner work opentelemetry collector contrib opentelemetry collector contrib processor tailsamplingprocessor make leaving directory home runner work opentelemetry collector contrib opentelemetry collector contrib | 1 |
56,620 | 14,078,460,808 | IssuesEvent | 2020-11-04 13:36:20 | themagicalmammal/android_kernel_samsung_a3xelte | https://api.github.com/repos/themagicalmammal/android_kernel_samsung_a3xelte | opened | CVE-2018-20669 (High) detected in multiple libraries | security vulnerability | ## CVE-2018-20669 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linuxv3.10</b>, <b>linuxlinux-3.10.11</b>, <b>linuxlinux-3.5</b></p></summary>
<p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue where a provided address with access_ok() is not checked was discovered in i915_gem_execbuffer2_ioctl in drivers/gpu/drm/i915/i915_gem_execbuffer.c in the Linux kernel through 4.19.13. A local attacker can craft a malicious IOCTL function call to overwrite arbitrary kernel memory, resulting in a Denial of Service or privilege escalation.
<p>Publish Date: 2019-03-21
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20669>CVE-2018-20669</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20669">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20669</a></p>
<p>Release Date: 2019-03-21</p>
<p>Fix Resolution: v5.0-rc1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2018-20669 (High) detected in multiple libraries - ## CVE-2018-20669 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linuxv3.10</b>, <b>linuxlinux-3.10.11</b>, <b>linuxlinux-3.5</b></p></summary>
<p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue where a provided address with access_ok() is not checked was discovered in i915_gem_execbuffer2_ioctl in drivers/gpu/drm/i915/i915_gem_execbuffer.c in the Linux kernel through 4.19.13. A local attacker can craft a malicious IOCTL function call to overwrite arbitrary kernel memory, resulting in a Denial of Service or privilege escalation.
<p>Publish Date: 2019-03-21
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20669>CVE-2018-20669</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20669">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20669</a></p>
<p>Release Date: 2019-03-21</p>
<p>Fix Resolution: v5.0-rc1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_test | cve high detected in multiple libraries cve high severity vulnerability vulnerable libraries linuxlinux linuxlinux vulnerability details an issue where a provided address with access ok is not checked was discovered in gem ioctl in drivers gpu drm gem execbuffer c in the linux kernel through a local attacker can craft a malicious ioctl function call to overwrite arbitrary kernel memory resulting in a denial of service or privilege escalation publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
330,729 | 28,484,922,517 | IssuesEvent | 2023-04-18 07:09:34 | unifyai/ivy | https://api.github.com/repos/unifyai/ivy | closed | Fix exponents_and_logarithms.test_numpy_exp2 | NumPy Frontend Sub Task Failing Test | | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4722833970/jobs/8377976075" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/4722833970/jobs/8377965187" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4722833970/jobs/8377966037" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/4645654528/jobs/8221610959" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
<details>
<summary>Not found</summary>
Not found
</details>
<details>
<summary>FAILED ivy_tests/test_ivy/test_frontends/test_numpy/test_mathematical_functions/test_exponents_and_logarithms.py::test_numpy_exp2[cpu-ivy.functional.backends.jax-False-False]</summary>
2023-04-08T14:48:36.1956634Z E IndexError: list index out of range2023-04-08T14:48:36.1956957Z E Falsifying example: test_numpy_exp2(2023-04-08T14:48:36.1957633Z E dtypes_values_casting=(['float16'],2023-04-08T14:48:36.1958008Z E [array(-1., dtype=float16)],2023-04-08T14:48:36.1958298Z E 'no',2023-04-08T14:48:36.1958524Z E None),2023-04-08T14:48:36.1958779Z E where=[array(False)],2023-04-08T14:48:36.1959097Z E test_flags=FrontendFunctionTestFlags(2023-04-08T14:48:36.1959672Z E num_positional_args=1,2023-04-08T14:48:36.1959941Z E with_out=False,2023-04-08T14:48:36.1960185Z E inplace=False,2023-04-08T14:48:36.1960452Z E as_variable=[False],2023-04-08T14:48:36.1960727Z E native_arrays=[False],2023-04-08T14:48:36.1961021Z E generate_frontend_arrays=True,2023-04-08T14:48:36.1961282Z E ),2023-04-08T14:48:36.1961668Z E fn_tree='ivy.functional.frontends.numpy.exp2',2023-04-08T14:48:36.1962040Z E frontend='numpy',2023-04-08T14:48:36.1962335Z E on_device='cpu',2023-04-08T14:48:36.1962568Z E )2023-04-08T14:48:36.1962772Z E 2023-04-08T14:48:36.1963383Z E You can reproduce this example by temporarily adding @reproduce_failure('6.71.0', b'AXicY2AAAUYGBiSaEQAALAAE') as a decorator on your test case
</details>
| 1.0 | Fix exponents_and_logarithms.test_numpy_exp2 - | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4722833970/jobs/8377976075" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/4722833970/jobs/8377965187" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4722833970/jobs/8377966037" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/4645654528/jobs/8221610959" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
<details>
<summary>Not found</summary>
Not found
</details>
<details>
<summary>FAILED ivy_tests/test_ivy/test_frontends/test_numpy/test_mathematical_functions/test_exponents_and_logarithms.py::test_numpy_exp2[cpu-ivy.functional.backends.jax-False-False]</summary>
2023-04-08T14:48:36.1956634Z E IndexError: list index out of range2023-04-08T14:48:36.1956957Z E Falsifying example: test_numpy_exp2(2023-04-08T14:48:36.1957633Z E dtypes_values_casting=(['float16'],2023-04-08T14:48:36.1958008Z E [array(-1., dtype=float16)],2023-04-08T14:48:36.1958298Z E 'no',2023-04-08T14:48:36.1958524Z E None),2023-04-08T14:48:36.1958779Z E where=[array(False)],2023-04-08T14:48:36.1959097Z E test_flags=FrontendFunctionTestFlags(2023-04-08T14:48:36.1959672Z E num_positional_args=1,2023-04-08T14:48:36.1959941Z E with_out=False,2023-04-08T14:48:36.1960185Z E inplace=False,2023-04-08T14:48:36.1960452Z E as_variable=[False],2023-04-08T14:48:36.1960727Z E native_arrays=[False],2023-04-08T14:48:36.1961021Z E generate_frontend_arrays=True,2023-04-08T14:48:36.1961282Z E ),2023-04-08T14:48:36.1961668Z E fn_tree='ivy.functional.frontends.numpy.exp2',2023-04-08T14:48:36.1962040Z E frontend='numpy',2023-04-08T14:48:36.1962335Z E on_device='cpu',2023-04-08T14:48:36.1962568Z E )2023-04-08T14:48:36.1962772Z E 2023-04-08T14:48:36.1963383Z E You can reproduce this example by temporarily adding @reproduce_failure('6.71.0', b'AXicY2AAAUYGBiSaEQAALAAE') as a decorator on your test case
</details>
| test | fix exponents and logarithms test numpy tensorflow img src torch img src numpy img src jax img src not found not found failed ivy tests test ivy test frontends test numpy test mathematical functions test exponents and logarithms py test numpy e indexerror list index out of e falsifying example test numpy e dtypes values casting e e no e none e where e test flags frontendfunctiontestflags e num positional args e with out false e inplace false e as variable e native arrays e generate frontend arrays true e e fn tree ivy functional frontends numpy e frontend numpy e on device cpu e e e you can reproduce this example by temporarily adding reproduce failure b as a decorator on your test case | 1 |
176,122 | 13,627,748,350 | IssuesEvent | 2020-09-24 13:02:29 | Scholar-6/brillder | https://api.github.com/repos/Scholar-6/brillder | closed | Ticket #67 Synthesis has gone missing | Betatester Request Critical Blocker | https://brillder.scholar6.org/build/brick/257/investigation/synthesis
A lot of my content has seemingly disappeared from...
Requester: Sophie
Assignee: Support
A lot of my content has seemingly disappeared from the editor. It's still present in the mobile-screen-preview, but it seems to be missing in the editor itself.
------------------
Submitted from: https://brillder.scholar6.org/build/brick/257/investigation/question-component/30403



- [x] seems Wiris related as pure text components still showing and editable
- [x] could it also be because of new fixed question text component? | 1.0 | Ticket #67 Synthesis has gone missing - https://brillder.scholar6.org/build/brick/257/investigation/synthesis
A lot of my content has seemingly disappeared from...
Requester: Sophie
Assignee: Support
A lot of my content has seemingly disappeared from the editor. It's still present in the mobile-screen-preview, but it seems to be missing in the editor itself.
------------------
Submitted from: https://brillder.scholar6.org/build/brick/257/investigation/question-component/30403



- [x] seems Wiris related as pure text components still showing and editable
- [x] could it also be because of new fixed question text component? | test | ticket synthesis has gone missing a lot of my content has seemingly disappeared from requester sophie assignee support a lot of my content has seemingly disappeared from the editor it s still present in the mobile screen preview but it seems to be missing in the editor itself submitted from seems wiris related as pure text components still showing and editable could it also be because of new fixed question text component | 1 |
19,762 | 3,779,582,586 | IssuesEvent | 2016-03-18 09:07:10 | koying/SPMC | https://api.github.com/repos/koying/SPMC | closed | Generic issue with Android < 4.4 | bug To be tested | When I open SPMC 16.1 , It crashed. But SPMC 15.0 and KODI 16.1 is OK.
System info from SPMC:
CPU : ARMV7 Processor rev1(V7I)
GPU : Mail-450MP
System : Android 4.3.0 API Level18 (kernel:linux 3.1.10)
OpenGL vendor : ARM
OpenGl version :OpenGL ES2.0 | 1.0 | Generic issue with Android < 4.4 - When I open SPMC 16.1 , It crashed. But SPMC 15.0 and KODI 16.1 is OK.
System info from SPMC:
CPU : ARMV7 Processor rev1(V7I)
GPU : Mail-450MP
System : Android 4.3.0 API Level18 (kernel:linux 3.1.10)
OpenGL vendor : ARM
OpenGl version :OpenGL ES2.0 | test | generic issue with android when i open spmc it crashed but spmc and kodi is ok system info from spmc cpu processor gpu mail system android api kernel linux opengl vendor arm opengl version opengl | 1 |
328,167 | 24,172,873,737 | IssuesEvent | 2022-09-22 21:01:12 | fleetdm/fleet | https://api.github.com/repos/fleetdm/fleet | opened | Document using a password manager during set up is the best practice | :improve documentation | ### Goals
Test that using a password manager (ex. 1password) works when setting up Fleet and logging into Fleet.
In the "Running the Fleet server" docs: Document that using a password manager is the best practice when setting up Fleet: https://fleetdm.com/docs/deploying/server-installation#running-the-fleet-server
This is because, if you forget your password, you will have to wipe the database or redeploy Fleet.
| 1.0 | Document using a password manager during set up is the best practice - ### Goals
Test that using a password manager (ex. 1password) works when setting up Fleet and logging into Fleet.
In the "Running the Fleet server" docs: Document that using a password manager is the best practice when setting up Fleet: https://fleetdm.com/docs/deploying/server-installation#running-the-fleet-server
This is because, if you forget your password, you will have to wipe the database or redeploy Fleet.
| non_test | document using a password manager during set up is the best practice goals test that using a password manager ex works when setting up fleet and logging into fleet in the running the fleet server docs document that using a password manager is the best practice when setting up fleet this is because if you forget your password you will have to wipe the database or redeploy fleet | 0 |
284,791 | 21,470,291,549 | IssuesEvent | 2022-04-26 08:54:17 | OpenRefine/OpenRefine | https://api.github.com/repos/OpenRefine/OpenRefine | closed | Update user manual and in-app documentation for splitByLength() | bug documentation grel | The documentation, specifically for splitByLength() does not inform the user that they have the possibility to only return values within a given range of indices, Iv since updated the document but wanted to make a bug report on it for submitting the PR.
ex:
`"123456789012345678901234567890".splitByLengths(3,4,5)[0,2]`
`returns [ "123", "4567" ]`
More about this issue can be seen in this thread: https://github.com/OpenRefine/OpenRefine/issues/3929 | 1.0 | Update user manual and in-app documentation for splitByLength() - The documentation, specifically for splitByLength() does not inform the user that they have the possibility to only return values within a given range of indices, Iv since updated the document but wanted to make a bug report on it for submitting the PR.
ex:
`"123456789012345678901234567890".splitByLengths(3,4,5)[0,2]`
`returns [ "123", "4567" ]`
More about this issue can be seen in this thread: https://github.com/OpenRefine/OpenRefine/issues/3929 | non_test | update user manual and in app documentation for splitbylength the documentation specifically for splitbylength does not inform the user that they have the possibility to only return values within a given range of indices iv since updated the document but wanted to make a bug report on it for submitting the pr ex splitbylengths returns more about this issue can be seen in this thread | 0 |
41,745 | 5,395,699,608 | IssuesEvent | 2017-02-27 09:29:18 | mautic/mautic | https://api.github.com/repos/mautic/mautic | closed | Unable To See focus data | Pending Feedback Ready To Test | What type of report is this:
| Q | A
| ---| ---
| Bug report? | Y
| Feature request? | N
| Enhancement? |
## Description:
While saving (SAVE and CLOSE). it stuck. and then you can go back and try to click on a link. you will get only publish detail in rough data.
You will not get the script.
## If a bug:
| Q | A
| --- | ---
| Mautic version|2.6.0
| PHP version| 5.6
### Steps to reproduce:
1. Create focus.
2. While saving (SAVE and CLOSE). it stuck.
I can't enter in it to get script link.
### Log errors:
I have created the focus but unable to see it. Mean I can't enter in it to get script link.
... (Version 2.6.0)
@escopecz
[2017-02-23 10:37:50] mautic.WARNING: PHP Warning - fwrite() expects parameter 1 to be resource, boolean given - in file /home/mona/mautic-master/app/bundles/CoreBundle/IpLookup/AbstractLocalDataLookup.php - at line 98 [] []
[2017-02-23 10:37:50] mautic.WARNING: PHP Warning - fwrite() expects parameter 1 to be resource, boolean given - in file /home/mona/mautic-master/app/bundles/CoreBundle/IpLookup/AbstractLocalDataLookup.php - at line 98 [] []
[2017-02-23 10:37:50] mautic.WARNING: PHP Warning - fwrite() expects parameter 1 to be resource, boolean given - in file /home/mona/mautic-master/app/bundles/CoreBundle/IpLookup/AbstractLocalDataLookup.php - at line 98 [] []
[2017-02-23 10:37:50] mautic.WARNING: PHP Warning - fwrite() expects parameter 1 to be resource, boolean given - in file /home/mona/mautic-master/app/bundles/CoreBundle/IpLookup/AbstractLocalDataLookup.php - at line 98 [] []
[2017-02-23 10:37:50] mautic.WARNING: PHP Warning - fclose() expects parameter 1 to be resource, boolean given - in file /home/mona/mautic-master/app/bundles/CoreBundle/IpLookup/AbstractLocalDataLookup.php - at line 100 [] []
[2017-02-23 10:37:59] mautic.NOTICE: PHP Notice - Undefined variable: dateRangeForm - in file /home/mona/mautic-master/plugins/MauticFocusBundle/Views/Focus/details.html.php - at line 103 [] []
[2017-02-23 10:37:59] mautic.ERROR: PHP Error - Argument 1 passed to Symfony\Bundle\FrameworkBundle\Templating\Helper\FormHelper::start() must be an instance of Symfony\Component\Form\FormView, null given, called in /home/mona/mautic-master/app/bundles/CoreBundle/Views/Helper/graph_dateselect.html.php on line 16 and defined - in file /home/mona/mautic-master/vendor/symfony/framework-bundle/Templating/Helper/FormHelper.php - at line 101 [] []
[2017-02-23 10:38:03] mautic.CRITICAL: Uncaught PHP Exception InvalidArgumentException: "Method getTimezone not exists" at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php line 39 {"exception":"[object] (InvalidArgumentException(code: 0): Method getTimezone not exists at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php:39)"} []
[2017-02-23 10:38:03] mautic.NOTICE: PHP Notice - Undefined variable: dateRangeForm - in file /home/mona/mautic-master/plugins/MauticFocusBundle/Views/Focus/details.html.php - at line 103 [] []
[2017-02-23 10:38:03] mautic.ERROR: PHP Error - Argument 1 passed to Symfony\Bundle\FrameworkBundle\Templating\Helper\FormHelper::start() must be an instance of Symfony\Component\Form\FormView, null given, called in /home/mona/mautic-master/app/bundles/CoreBundle/Views/Helper/graph_dateselect.html.php on line 16 and defined - in file /home/mona/mautic-master/vendor/symfony/framework-bundle/Templating/Helper/FormHelper.php - at line 101 [] []
[2017-02-23 10:38:15] mautic.NOTICE: PHP Notice - Undefined variable: dateRangeForm - in file /home/mona/mautic-master/plugins/MauticFocusBundle/Views/Focus/details.html.php - at line 103 [] []
[2017-02-23 10:38:15] mautic.ERROR: PHP Error - Argument 1 passed to Symfony\Bundle\FrameworkBundle\Templating\Helper\FormHelper::start() must be an instance of Symfony\Component\Form\FormView, null given, called in /home/mona/mautic-master/app/bundles/CoreBundle/Views/Helper/graph_dateselect.html.php on line 16 and defined - in file /home/mona/mautic-master/vendor/symfony/framework-bundle/Templating/Helper/FormHelper.php - at line 101 [] []
[2017-02-23 10:38:30] mautic.CRITICAL: Uncaught PHP Exception InvalidArgumentException: "Method getTimezone not exists" at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php line 39 {"exception":"[object] (InvalidArgumentException(code: 0): Method getTimezone not exists at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php:39)"} []
[2017-02-23 10:38:52] mautic.CRITICAL: Uncaught PHP Exception InvalidArgumentException: "Method getTimezone not exists" at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php line 39 {"exception":"[object] (InvalidArgumentException(code: 0): Method getTimezone not exists at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php:39)"} []
[2017-02-23 10:38:56] mautic.CRITICAL: Uncaught PHP Exception InvalidArgumentException: "Method getTimezone not exists" at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php line 39 {"exception":"[object] (InvalidArgumentException(code: 0): Method getTimezone not exists at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php:39)"} []
[2017-02-23 10:38:57] mautic.NOTICE: PHP Notice - Undefined variable: dateRangeForm - in file /home/mona/mautic-master/plugins/MauticFocusBundle/Views/Focus/details.html.php - at line 103 [] []
[2017-02-23 10:38:57] mautic.ERROR: PHP Error - Argument 1 passed to Symfony\Bundle\FrameworkBundle\Templating\Helper\FormHelper::start() must be an instance of Symfony\Component\Form\FormView, null given, called in /home/mona/mautic-master/app/bundles/CoreBundle/Views/Helper/graph_dateselect.html.php on line 16 and defined - in file /home/mona/mautic-master/vendor/symfony/framework-bundle/Templating/Helper/FormHelper.php - at line 101 [] []
[2017-02-23 10:39:05] mautic.CRITICAL: Uncaught PHP Exception InvalidArgumentException: "Method getTimezone not exists" at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php line 39 {"exception":"[object] (InvalidArgumentException(code: 0): Method getTimezone not exists at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php:39)"} []
[2017-02-23 10:39:12] mautic.CRITICAL: Uncaught PHP Exception InvalidArgumentException: "Method getTimezone not exists" at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php line 39 {"exception":"[object] (InvalidArgumentException(code: 0): Method getTimezone not exists at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php:39)"} []
[2017-02-23 10:39:18] mautic.CRITICAL: Uncaught PHP Exception InvalidArgumentException: "Method getTimezone not exists" at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php line 39 {"exception":"[object] (InvalidArgumentException(code: 0): Method getTimezone not exists at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php:39)"} []
[2017-02-23 10:39:41] mautic.CRITICAL: Uncaught PHP Exception InvalidArgumentException: "Method getTimezone not exists" at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php line 39 {"exception":"[object] (InvalidArgumentException(code: 0): Method getTimezone not exists at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php:39)"} []
[2017-02-23 10:39:49] mautic.CRITICAL: Uncaught PHP Exception InvalidArgumentException: "Method getTimezone not exists" at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php line 39 {"exception":"[object] (InvalidArgumentException(code: 0): Method getTimezone not exists at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php:39)"} []
[2017-02-23 10:39:50] mautic.CRITICAL: Uncaught PHP Exception InvalidArgumentException: "Method getTimezone not exists" at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php line 39 {"exception":"[object] (InvalidArgumentException(code: 0): Method getTimezone not exists at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php:39)"} []
[2017-02-23 10:39:50] mautic.NOTICE: PHP Notice - Undefined variable: dateRangeForm - in file /home/mona/mautic-master/plugins/MauticFocusBundle/Views/Focus/details.html.php - at line 103 [] []
[2017-02-23 10:39:50] mautic.ERROR: PHP Error - Argument 1 passed to Symfony\Bundle\FrameworkBundle\Templating\Helper\FormHelper::start() must be an instance of Symfony\Component\Form\FormView, null given, called in /home/mona/mautic-master/app/bundles/CoreBundle/Views/Helper/graph_dateselect.html.php on line 16 and defined - in file /home/mona/mautic-master/vendor/symfony/framework-bundle/Templating/Helper/FormHelper.php - at line 101 [] []
[2017-02-23 10:40:06] mautic.CRITICAL: Uncaught PHP Exception InvalidArgumentException: "Method getTimezone not exists" at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php line 39 {"exception":"[object] (InvalidArgumentException(code: 0): Method getTimezone not exists at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php:39)"} []
[2017-02-23 10:40:23] mautic.NOTICE: PHP Notice - Undefined variable: dateRangeForm - in file /home/mona/mautic-master/plugins/MauticFocusBundle/Views/Focus/details.html.php - at line 103 [] []
[2017-02-23 10:40:23] mautic.ERROR: PHP Error - Argument 1 passed to Symfony\Bundle\FrameworkBundle\Templating\Helper\FormHelper::start() must be an instance of Symfony\Component\Form\FormView, null given, called in /home/mona/mautic-master/app/bundles/CoreBundle/Views/Helper/graph_dateselect.html.php on line 16 and defined - in file /home/mona/mautic-master/vendor/symfony/framework-bundle/Templating/Helper/FormHelper.php - at line 101 [] []
| 1.0 | Unable To See focus data - What type of report is this:
| Q | A
| ---| ---
| Bug report? | Y
| Feature request? | N
| Enhancement? |
## Description:
While saving (SAVE and CLOSE). it stuck. and then you can go back and try to click on a link. you will get only publish detail in rough data.
You will not get the script.
## If a bug:
| Q | A
| --- | ---
| Mautic version|2.6.0
| PHP version| 5.6
### Steps to reproduce:
1. Create focus.
2. While saving (SAVE and CLOSE). it stuck.
I can't enter in it to get script link.
### Log errors:
I have created the focus but unable to see it. Mean I can't enter in it to get script link.
... (Version 2.6.0)
@escopecz
[2017-02-23 10:37:50] mautic.WARNING: PHP Warning - fwrite() expects parameter 1 to be resource, boolean given - in file /home/mona/mautic-master/app/bundles/CoreBundle/IpLookup/AbstractLocalDataLookup.php - at line 98 [] []
[2017-02-23 10:37:50] mautic.WARNING: PHP Warning - fwrite() expects parameter 1 to be resource, boolean given - in file /home/mona/mautic-master/app/bundles/CoreBundle/IpLookup/AbstractLocalDataLookup.php - at line 98 [] []
[2017-02-23 10:37:50] mautic.WARNING: PHP Warning - fwrite() expects parameter 1 to be resource, boolean given - in file /home/mona/mautic-master/app/bundles/CoreBundle/IpLookup/AbstractLocalDataLookup.php - at line 98 [] []
[2017-02-23 10:37:50] mautic.WARNING: PHP Warning - fwrite() expects parameter 1 to be resource, boolean given - in file /home/mona/mautic-master/app/bundles/CoreBundle/IpLookup/AbstractLocalDataLookup.php - at line 98 [] []
[2017-02-23 10:37:50] mautic.WARNING: PHP Warning - fclose() expects parameter 1 to be resource, boolean given - in file /home/mona/mautic-master/app/bundles/CoreBundle/IpLookup/AbstractLocalDataLookup.php - at line 100 [] []
[2017-02-23 10:37:59] mautic.NOTICE: PHP Notice - Undefined variable: dateRangeForm - in file /home/mona/mautic-master/plugins/MauticFocusBundle/Views/Focus/details.html.php - at line 103 [] []
[2017-02-23 10:37:59] mautic.ERROR: PHP Error - Argument 1 passed to Symfony\Bundle\FrameworkBundle\Templating\Helper\FormHelper::start() must be an instance of Symfony\Component\Form\FormView, null given, called in /home/mona/mautic-master/app/bundles/CoreBundle/Views/Helper/graph_dateselect.html.php on line 16 and defined - in file /home/mona/mautic-master/vendor/symfony/framework-bundle/Templating/Helper/FormHelper.php - at line 101 [] []
[2017-02-23 10:38:03] mautic.CRITICAL: Uncaught PHP Exception InvalidArgumentException: "Method getTimezone not exists" at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php line 39 {"exception":"[object] (InvalidArgumentException(code: 0): Method getTimezone not exists at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php:39)"} []
[2017-02-23 10:38:03] mautic.NOTICE: PHP Notice - Undefined variable: dateRangeForm - in file /home/mona/mautic-master/plugins/MauticFocusBundle/Views/Focus/details.html.php - at line 103 [] []
[2017-02-23 10:38:03] mautic.ERROR: PHP Error - Argument 1 passed to Symfony\Bundle\FrameworkBundle\Templating\Helper\FormHelper::start() must be an instance of Symfony\Component\Form\FormView, null given, called in /home/mona/mautic-master/app/bundles/CoreBundle/Views/Helper/graph_dateselect.html.php on line 16 and defined - in file /home/mona/mautic-master/vendor/symfony/framework-bundle/Templating/Helper/FormHelper.php - at line 101 [] []
[2017-02-23 10:38:15] mautic.NOTICE: PHP Notice - Undefined variable: dateRangeForm - in file /home/mona/mautic-master/plugins/MauticFocusBundle/Views/Focus/details.html.php - at line 103 [] []
[2017-02-23 10:38:15] mautic.ERROR: PHP Error - Argument 1 passed to Symfony\Bundle\FrameworkBundle\Templating\Helper\FormHelper::start() must be an instance of Symfony\Component\Form\FormView, null given, called in /home/mona/mautic-master/app/bundles/CoreBundle/Views/Helper/graph_dateselect.html.php on line 16 and defined - in file /home/mona/mautic-master/vendor/symfony/framework-bundle/Templating/Helper/FormHelper.php - at line 101 [] []
[2017-02-23 10:38:30] mautic.CRITICAL: Uncaught PHP Exception InvalidArgumentException: "Method getTimezone not exists" at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php line 39 {"exception":"[object] (InvalidArgumentException(code: 0): Method getTimezone not exists at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php:39)"} []
[2017-02-23 10:38:52] mautic.CRITICAL: Uncaught PHP Exception InvalidArgumentException: "Method getTimezone not exists" at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php line 39 {"exception":"[object] (InvalidArgumentException(code: 0): Method getTimezone not exists at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php:39)"} []
[2017-02-23 10:38:56] mautic.CRITICAL: Uncaught PHP Exception InvalidArgumentException: "Method getTimezone not exists" at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php line 39 {"exception":"[object] (InvalidArgumentException(code: 0): Method getTimezone not exists at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php:39)"} []
[2017-02-23 10:38:57] mautic.NOTICE: PHP Notice - Undefined variable: dateRangeForm - in file /home/mona/mautic-master/plugins/MauticFocusBundle/Views/Focus/details.html.php - at line 103 [] []
[2017-02-23 10:38:57] mautic.ERROR: PHP Error - Argument 1 passed to Symfony\Bundle\FrameworkBundle\Templating\Helper\FormHelper::start() must be an instance of Symfony\Component\Form\FormView, null given, called in /home/mona/mautic-master/app/bundles/CoreBundle/Views/Helper/graph_dateselect.html.php on line 16 and defined - in file /home/mona/mautic-master/vendor/symfony/framework-bundle/Templating/Helper/FormHelper.php - at line 101 [] []
[2017-02-23 10:39:05] mautic.CRITICAL: Uncaught PHP Exception InvalidArgumentException: "Method getTimezone not exists" at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php line 39 {"exception":"[object] (InvalidArgumentException(code: 0): Method getTimezone not exists at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php:39)"} []
[2017-02-23 10:39:12] mautic.CRITICAL: Uncaught PHP Exception InvalidArgumentException: "Method getTimezone not exists" at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php line 39 {"exception":"[object] (InvalidArgumentException(code: 0): Method getTimezone not exists at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php:39)"} []
[2017-02-23 10:39:18] mautic.CRITICAL: Uncaught PHP Exception InvalidArgumentException: "Method getTimezone not exists" at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php line 39 {"exception":"[object] (InvalidArgumentException(code: 0): Method getTimezone not exists at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php:39)"} []
[2017-02-23 10:39:41] mautic.CRITICAL: Uncaught PHP Exception InvalidArgumentException: "Method getTimezone not exists" at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php line 39 {"exception":"[object] (InvalidArgumentException(code: 0): Method getTimezone not exists at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php:39)"} []
[2017-02-23 10:39:49] mautic.CRITICAL: Uncaught PHP Exception InvalidArgumentException: "Method getTimezone not exists" at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php line 39 {"exception":"[object] (InvalidArgumentException(code: 0): Method getTimezone not exists at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php:39)"} []
[2017-02-23 10:39:50] mautic.CRITICAL: Uncaught PHP Exception InvalidArgumentException: "Method getTimezone not exists" at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php line 39 {"exception":"[object] (InvalidArgumentException(code: 0): Method getTimezone not exists at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php:39)"} []
[2017-02-23 10:39:50] mautic.NOTICE: PHP Notice - Undefined variable: dateRangeForm - in file /home/mona/mautic-master/plugins/MauticFocusBundle/Views/Focus/details.html.php - at line 103 [] []
[2017-02-23 10:39:50] mautic.ERROR: PHP Error - Argument 1 passed to Symfony\Bundle\FrameworkBundle\Templating\Helper\FormHelper::start() must be an instance of Symfony\Component\Form\FormView, null given, called in /home/mona/mautic-master/app/bundles/CoreBundle/Views/Helper/graph_dateselect.html.php on line 16 and defined - in file /home/mona/mautic-master/vendor/symfony/framework-bundle/Templating/Helper/FormHelper.php - at line 101 [] []
[2017-02-23 10:40:06] mautic.CRITICAL: Uncaught PHP Exception InvalidArgumentException: "Method getTimezone not exists" at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php line 39 {"exception":"[object] (InvalidArgumentException(code: 0): Method getTimezone not exists at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php:39)"} []
[2017-02-23 10:40:23] mautic.NOTICE: PHP Notice - Undefined variable: dateRangeForm - in file /home/mona/mautic-master/plugins/MauticFocusBundle/Views/Focus/details.html.php - at line 103 [] []
[2017-02-23 10:40:23] mautic.ERROR: PHP Error - Argument 1 passed to Symfony\Bundle\FrameworkBundle\Templating\Helper\FormHelper::start() must be an instance of Symfony\Component\Form\FormView, null given, called in /home/mona/mautic-master/app/bundles/CoreBundle/Views/Helper/graph_dateselect.html.php on line 16 and defined - in file /home/mona/mautic-master/vendor/symfony/framework-bundle/Templating/Helper/FormHelper.php - at line 101 [] []
| test | unable to see focus data what type of report is this q a bug report y feature request n enhancement description while saving save and close it stuck and then you can go back and try to click on a link you will get only publish detail in rough data you will not get the script if a bug q a mautic version php version steps to reproduce create focus while saving save and close it stuck i can t enter in it to get script link log errors i have created the focus but unable to see it mean i can t enter in it to get script link version escopecz mautic warning php warning fwrite expects parameter to be resource boolean given in file home mona mautic master app bundles corebundle iplookup abstractlocaldatalookup php at line mautic warning php warning fwrite expects parameter to be resource boolean given in file home mona mautic master app bundles corebundle iplookup abstractlocaldatalookup php at line mautic warning php warning fwrite expects parameter to be resource boolean given in file home mona mautic master app bundles corebundle iplookup abstractlocaldatalookup php at line mautic warning php warning fwrite expects parameter to be resource boolean given in file home mona mautic master app bundles corebundle iplookup abstractlocaldatalookup php at line mautic warning php warning fclose expects parameter to be resource boolean given in file home mona mautic master app bundles corebundle iplookup abstractlocaldatalookup php at line mautic notice php notice undefined variable daterangeform in file home mona mautic master plugins mauticfocusbundle views focus details html php at line mautic error php error argument passed to symfony bundle frameworkbundle templating helper formhelper start must be an instance of symfony component form formview null given called in home mona mautic master app bundles corebundle views helper graph dateselect html php on line and defined in file home mona mautic master vendor symfony framework bundle templating helper formhelper php at line mautic critical uncaught php exception invalidargumentexception method gettimezone not exists at home mona mautic master app bundles corebundle entity commonentity php line exception invalidargumentexception code method gettimezone not exists at home mona mautic master app bundles corebundle entity commonentity php mautic notice php notice undefined variable daterangeform in file home mona mautic master plugins mauticfocusbundle views focus details html php at line mautic error php error argument passed to symfony bundle frameworkbundle templating helper formhelper start must be an instance of symfony component form formview null given called in home mona mautic master app bundles corebundle views helper graph dateselect html php on line and defined in file home mona mautic master vendor symfony framework bundle templating helper formhelper php at line mautic notice php notice undefined variable daterangeform in file home mona mautic master plugins mauticfocusbundle views focus details html php at line mautic error php error argument passed to symfony bundle frameworkbundle templating helper formhelper start must be an instance of symfony component form formview null given called in home mona mautic master app bundles corebundle views helper graph dateselect html php on line and defined in file home mona mautic master vendor symfony framework bundle templating helper formhelper php at line mautic critical uncaught php exception invalidargumentexception method gettimezone not exists at home mona mautic master app bundles corebundle entity commonentity php line exception invalidargumentexception code method gettimezone not exists at home mona mautic master app bundles corebundle entity commonentity php mautic critical uncaught php exception invalidargumentexception method gettimezone not exists at home mona mautic master app bundles corebundle entity commonentity php line exception invalidargumentexception code method gettimezone not exists at home mona mautic master app bundles corebundle entity commonentity php mautic critical uncaught php exception invalidargumentexception method gettimezone not exists at home mona mautic master app bundles corebundle entity commonentity php line exception invalidargumentexception code method gettimezone not exists at home mona mautic master app bundles corebundle entity commonentity php mautic notice php notice undefined variable daterangeform in file home mona mautic master plugins mauticfocusbundle views focus details html php at line mautic error php error argument passed to symfony bundle frameworkbundle templating helper formhelper start must be an instance of symfony component form formview null given called in home mona mautic master app bundles corebundle views helper graph dateselect html php on line and defined in file home mona mautic master vendor symfony framework bundle templating helper formhelper php at line mautic critical uncaught php exception invalidargumentexception method gettimezone not exists at home mona mautic master app bundles corebundle entity commonentity php line exception invalidargumentexception code method gettimezone not exists at home mona mautic master app bundles corebundle entity commonentity php mautic critical uncaught php exception invalidargumentexception method gettimezone not exists at home mona mautic master app bundles corebundle entity commonentity php line exception invalidargumentexception code method gettimezone not exists at home mona mautic master app bundles corebundle entity commonentity php mautic critical uncaught php exception invalidargumentexception method gettimezone not exists at home mona mautic master app bundles corebundle entity commonentity php line exception invalidargumentexception code method gettimezone not exists at home mona mautic master app bundles corebundle entity commonentity php mautic critical uncaught php exception invalidargumentexception method gettimezone not exists at home mona mautic master app bundles corebundle entity commonentity php line exception invalidargumentexception code method gettimezone not exists at home mona mautic master app bundles corebundle entity commonentity php mautic critical uncaught php exception invalidargumentexception method gettimezone not exists at home mona mautic master app bundles corebundle entity commonentity php line exception invalidargumentexception code method gettimezone not exists at home mona mautic master app bundles corebundle entity commonentity php mautic critical uncaught php exception invalidargumentexception method gettimezone not exists at home mona mautic master app bundles corebundle entity commonentity php line exception invalidargumentexception code method gettimezone not exists at home mona mautic master app bundles corebundle entity commonentity php mautic notice php notice undefined variable daterangeform in file home mona mautic master plugins mauticfocusbundle views focus details html php at line mautic error php error argument passed to symfony bundle frameworkbundle templating helper formhelper start must be an instance of symfony component form formview null given called in home mona mautic master app bundles corebundle views helper graph dateselect html php on line and defined in file home mona mautic master vendor symfony framework bundle templating helper formhelper php at line mautic critical uncaught php exception invalidargumentexception method gettimezone not exists at home mona mautic master app bundles corebundle entity commonentity php line exception invalidargumentexception code method gettimezone not exists at home mona mautic master app bundles corebundle entity commonentity php mautic notice php notice undefined variable daterangeform in file home mona mautic master plugins mauticfocusbundle views focus details html php at line mautic error php error argument passed to symfony bundle frameworkbundle templating helper formhelper start must be an instance of symfony component form formview null given called in home mona mautic master app bundles corebundle views helper graph dateselect html php on line and defined in file home mona mautic master vendor symfony framework bundle templating helper formhelper php at line | 1 |
161,609 | 12,554,323,268 | IssuesEvent | 2020-06-07 01:34:19 | kotest/kotest | https://api.github.com/repos/kotest/kotest | closed | Property tests should report exception of running shrunk input | bug property-testing | During property testing, if an input fails, the `AssertionError` is being rendered as part of the test report. But if the input has successfully shrunk, the exception that failed the shrunk input is ignored.
Ideally the exception that the user sees in the final test report should be the one that the shrunk input failed with. It might also be ok to render both exceptions since also both inputs (original and shrunk) are rendered. | 1.0 | Property tests should report exception of running shrunk input - During property testing, if an input fails, the `AssertionError` is being rendered as part of the test report. But if the input has successfully shrunk, the exception that failed the shrunk input is ignored.
Ideally the exception that the user sees in the final test report should be the one that the shrunk input failed with. It might also be ok to render both exceptions since also both inputs (original and shrunk) are rendered. | test | property tests should report exception of running shrunk input during property testing if an input fails the assertionerror is being rendered as part of the test report but if the input has successfully shrunk the exception that failed the shrunk input is ignored ideally the exception that the user sees in the final test report should be the one that the shrunk input failed with it might also be ok to render both exceptions since also both inputs original and shrunk are rendered | 1 |
48,202 | 5,949,071,483 | IssuesEvent | 2017-05-26 13:22:05 | MohammadYounes/AlertifyJS | https://api.github.com/repos/MohammadYounes/AlertifyJS | closed | alert dialog can be hidden by some elements | needs test case troubleshooting | below is an example image

the vertical bar is an element of [ace editor](https://ace.c9.io)
styles of the vertical bar

| 1.0 | alert dialog can be hidden by some elements - below is an example image

the vertical bar is an element of [ace editor](https://ace.c9.io)
styles of the vertical bar

| test | alert dialog can be hidden by some elements below is an example image the vertical bar is an element of styles of the vertical bar | 1 |
150,954 | 11,995,036,021 | IssuesEvent | 2020-04-08 14:36:59 | prysmaticlabs/prysm | https://api.github.com/repos/prysmaticlabs/prysm | closed | Allow Multiple Proposal Slots in an Epoch | API Testnet Tracking | Currently is is assumed that a validator will only have a single proposal slot in an epoch, however due to the fact that proposal slots are computed in on a per slot basis, there is the possibility that a validator could have multiple proposal slots in an epoch instead of once. This issue is to track the implementation of the required changes for this. We will need to modify our ethereum-api and the relevant code in prysm | 1.0 | Allow Multiple Proposal Slots in an Epoch - Currently is is assumed that a validator will only have a single proposal slot in an epoch, however due to the fact that proposal slots are computed in on a per slot basis, there is the possibility that a validator could have multiple proposal slots in an epoch instead of once. This issue is to track the implementation of the required changes for this. We will need to modify our ethereum-api and the relevant code in prysm | test | allow multiple proposal slots in an epoch currently is is assumed that a validator will only have a single proposal slot in an epoch however due to the fact that proposal slots are computed in on a per slot basis there is the possibility that a validator could have multiple proposal slots in an epoch instead of once this issue is to track the implementation of the required changes for this we will need to modify our ethereum api and the relevant code in prysm | 1 |
143,864 | 11,581,475,554 | IssuesEvent | 2020-02-21 22:45:25 | pvlib/pvlib-python | https://api.github.com/repos/pvlib/pvlib-python | closed | the test suite is too slow | testing | The test suite is getting to be quite slow. Here are the 20 slowest tests on my mac:
```
pytest --durations=20
18.96s call pvlib/tests/iotools/test_pvgis.py::test_get_pvgis_tmy_kwargs
16.32s call pvlib/tests/iotools/test_psm3.py::test_get_psm3_singleyear
14.82s setup pvlib/tests/test_forecast.py::test_process_data[HRRR]
8.12s call pvlib/tests/iotools/test_psm3.py::test_get_psm3_tmy
6.82s setup pvlib/tests/test_forecast.py::test_process_data[NDFD]
5.47s call pvlib/tests/iotools/test_pvgis.py::test_get_pvgis_tmy_epw
4.78s call pvlib/tests/iotools/test_pvgis.py::test_get_pvgis_tmy
3.99s setup pvlib/tests/test_forecast.py::test_process_data[RAP]
3.45s call pvlib/tests/iotools/test_srml.py::test_read_srml_month_from_solardat
3.20s call pvlib/tests/test_forecast.py::test_bad_kwarg_get_data
3.16s call pvlib/tests/test_forecast.py::test_bad_kwarg_get_processed_data
3.13s call pvlib/tests/test_forecast.py::test_how_kwarg_get_processed_data
3.13s setup pvlib/tests/test_forecast.py::test_process_data[NAM]
3.12s call pvlib/tests/test_forecast.py::test_vert_level
3.08s call pvlib/tests/test_forecast.py::test_datetime
2.98s call pvlib/tests/test_irradiance.py::test_get_extra_radiation_nrel_numba
2.85s call pvlib/tests/test_solarposition.py::test_spa_python_numba_physical
2.82s call pvlib/tests/test_solarposition.py::test_get_solarposition_deltat[None-nrel_numba]
2.73s setup pvlib/tests/test_spa.py::NumbaSpaTest::test_aberration_correction
2.69s call pvlib/tests/iotools/test_pvgis.py::test_get_pvgis_tmy_basic
```
I will work on the `forecast` tests. Any ideas for the `iotools` tests? | 1.0 | the test suite is too slow - The test suite is getting to be quite slow. Here are the 20 slowest tests on my mac:
```
pytest --durations=20
18.96s call pvlib/tests/iotools/test_pvgis.py::test_get_pvgis_tmy_kwargs
16.32s call pvlib/tests/iotools/test_psm3.py::test_get_psm3_singleyear
14.82s setup pvlib/tests/test_forecast.py::test_process_data[HRRR]
8.12s call pvlib/tests/iotools/test_psm3.py::test_get_psm3_tmy
6.82s setup pvlib/tests/test_forecast.py::test_process_data[NDFD]
5.47s call pvlib/tests/iotools/test_pvgis.py::test_get_pvgis_tmy_epw
4.78s call pvlib/tests/iotools/test_pvgis.py::test_get_pvgis_tmy
3.99s setup pvlib/tests/test_forecast.py::test_process_data[RAP]
3.45s call pvlib/tests/iotools/test_srml.py::test_read_srml_month_from_solardat
3.20s call pvlib/tests/test_forecast.py::test_bad_kwarg_get_data
3.16s call pvlib/tests/test_forecast.py::test_bad_kwarg_get_processed_data
3.13s call pvlib/tests/test_forecast.py::test_how_kwarg_get_processed_data
3.13s setup pvlib/tests/test_forecast.py::test_process_data[NAM]
3.12s call pvlib/tests/test_forecast.py::test_vert_level
3.08s call pvlib/tests/test_forecast.py::test_datetime
2.98s call pvlib/tests/test_irradiance.py::test_get_extra_radiation_nrel_numba
2.85s call pvlib/tests/test_solarposition.py::test_spa_python_numba_physical
2.82s call pvlib/tests/test_solarposition.py::test_get_solarposition_deltat[None-nrel_numba]
2.73s setup pvlib/tests/test_spa.py::NumbaSpaTest::test_aberration_correction
2.69s call pvlib/tests/iotools/test_pvgis.py::test_get_pvgis_tmy_basic
```
I will work on the `forecast` tests. Any ideas for the `iotools` tests? | test | the test suite is too slow the test suite is getting to be quite slow here are the slowest tests on my mac pytest durations call pvlib tests iotools test pvgis py test get pvgis tmy kwargs call pvlib tests iotools test py test get singleyear setup pvlib tests test forecast py test process data call pvlib tests iotools test py test get tmy setup pvlib tests test forecast py test process data call pvlib tests iotools test pvgis py test get pvgis tmy epw call pvlib tests iotools test pvgis py test get pvgis tmy setup pvlib tests test forecast py test process data call pvlib tests iotools test srml py test read srml month from solardat call pvlib tests test forecast py test bad kwarg get data call pvlib tests test forecast py test bad kwarg get processed data call pvlib tests test forecast py test how kwarg get processed data setup pvlib tests test forecast py test process data call pvlib tests test forecast py test vert level call pvlib tests test forecast py test datetime call pvlib tests test irradiance py test get extra radiation nrel numba call pvlib tests test solarposition py test spa python numba physical call pvlib tests test solarposition py test get solarposition deltat setup pvlib tests test spa py numbaspatest test aberration correction call pvlib tests iotools test pvgis py test get pvgis tmy basic i will work on the forecast tests any ideas for the iotools tests | 1 |
261,900 | 22,779,236,994 | IssuesEvent | 2022-07-08 17:39:48 | nasa/openmct | https://api.github.com/repos/nasa/openmct | closed | [Edit] Remove available from arrow next to title | type:bug unverified severity:medium bug:retest | Steps to reproduce:
1. Create a new Display Layout. This should take you immediately into Edit mode.
2. Click Save and choose some appropriate location (e.g. My Items) then click OK. You should no longer be in Edit mode, but should still be navigated to the same layout.
3. Click the Edit button. You should again be taken into Edit mode.
4. Hover to the right of the title of the Display Layout ("Unnamed Display Layout" at top of screen). An arrow pointing down should appear.
5. Click on the arrow. A context menu should appear.
6. **Observe:** The context menu contains the Remove action. This is unexpected, as Remove effects the _parent_ of the edited object, which is outside of the scope for editing. (Additionally, the object is unsaved.)
7. Click on Remove.
8. **Observe:** Navigation takes you back to the parent container, but the object has not really been removed.
Believe correct behavior here is for Remove to be unavailable at Step 6.
| 1.0 | [Edit] Remove available from arrow next to title - Steps to reproduce:
1. Create a new Display Layout. This should take you immediately into Edit mode.
2. Click Save and choose some appropriate location (e.g. My Items) then click OK. You should no longer be in Edit mode, but should still be navigated to the same layout.
3. Click the Edit button. You should again be taken into Edit mode.
4. Hover to the right of the title of the Display Layout ("Unnamed Display Layout" at top of screen). An arrow pointing down should appear.
5. Click on the arrow. A context menu should appear.
6. **Observe:** The context menu contains the Remove action. This is unexpected, as Remove effects the _parent_ of the edited object, which is outside of the scope for editing. (Additionally, the object is unsaved.)
7. Click on Remove.
8. **Observe:** Navigation takes you back to the parent container, but the object has not really been removed.
Believe correct behavior here is for Remove to be unavailable at Step 6.
| test | remove available from arrow next to title steps to reproduce create a new display layout this should take you immediately into edit mode click save and choose some appropriate location e g my items then click ok you should no longer be in edit mode but should still be navigated to the same layout click the edit button you should again be taken into edit mode hover to the right of the title of the display layout unnamed display layout at top of screen an arrow pointing down should appear click on the arrow a context menu should appear observe the context menu contains the remove action this is unexpected as remove effects the parent of the edited object which is outside of the scope for editing additionally the object is unsaved click on remove observe navigation takes you back to the parent container but the object has not really been removed believe correct behavior here is for remove to be unavailable at step | 1 |
128,195 | 10,519,545,769 | IssuesEvent | 2019-09-29 18:44:54 | commercialhaskell/stackage | https://api.github.com/repos/commercialhaskell/stackage | closed | ghc-exactprint test failure | failure: test-suite | ```
Test suite failure for package ghc-exactprint-0.6.1
test: executable not found
```
CC @alanz | 1.0 | ghc-exactprint test failure - ```
Test suite failure for package ghc-exactprint-0.6.1
test: executable not found
```
CC @alanz | test | ghc exactprint test failure test suite failure for package ghc exactprint test executable not found cc alanz | 1 |
516,096 | 14,975,282,581 | IssuesEvent | 2021-01-28 05:46:46 | nlpsandbox/nlpsandbox-controller | https://api.github.com/repos/nlpsandbox/nlpsandbox-controller | closed | Should remove content of a submission annotation store if it already exists? | Priority: High | This is the scenario when re-running a submission... | 1.0 | Should remove content of a submission annotation store if it already exists? - This is the scenario when re-running a submission... | non_test | should remove content of a submission annotation store if it already exists this is the scenario when re running a submission | 0 |
18,616 | 4,288,534,585 | IssuesEvent | 2016-07-17 14:34:23 | swisnl/jQuery-contextMenu | https://api.github.com/repos/swisnl/jQuery-contextMenu | closed | documentation does not state how to destroy a contextmenu | Documentation | i found a solution in #110 to use:
```javascript
$.contextMenu("destroy", ".selector");
``` | 1.0 | documentation does not state how to destroy a contextmenu - i found a solution in #110 to use:
```javascript
$.contextMenu("destroy", ".selector");
``` | non_test | documentation does not state how to destroy a contextmenu i found a solution in to use javascript contextmenu destroy selector | 0 |
1,665 | 2,800,476,578 | IssuesEvent | 2015-05-13 10:15:07 | akka/akka | https://api.github.com/repos/akka/akka | opened | Add feature to run a test until it fails k times, dumping the failure output somewhere | 1 - triaged t:build | Should allow specifying a number ("I want 4 failures") and then run a test until it collects the required amount of failures (or some upper iteration/time limit is reached), then it dumps the outputs to different files | 1.0 | Add feature to run a test until it fails k times, dumping the failure output somewhere - Should allow specifying a number ("I want 4 failures") and then run a test until it collects the required amount of failures (or some upper iteration/time limit is reached), then it dumps the outputs to different files | non_test | add feature to run a test until it fails k times dumping the failure output somewhere should allow specifying a number i want failures and then run a test until it collects the required amount of failures or some upper iteration time limit is reached then it dumps the outputs to different files | 0 |
90,637 | 8,251,418,035 | IssuesEvent | 2018-09-12 07:45:09 | Kademi/kademi-dev | https://api.github.com/repos/Kademi/kademi-dev | closed | KCRM: Create a task from add a note window on dashboard component cause error | Ready to Test - Dev bug | 
When create save it gives you method not implemented

http://vladtest53wweb-vladtest53w.kademi-ci.co/dashboard
user: adminqa or kcrmadmin | 1.0 | KCRM: Create a task from add a note window on dashboard component cause error - 
When create save it gives you method not implemented

http://vladtest53wweb-vladtest53w.kademi-ci.co/dashboard
user: adminqa or kcrmadmin | test | kcrm create a task from add a note window on dashboard component cause error when create save it gives you method not implemented user adminqa or kcrmadmin | 1 |
94,963 | 3,933,558,680 | IssuesEvent | 2016-04-25 19:33:18 | ghutchis/avogadro | https://api.github.com/repos/ghutchis/avogadro | closed | Construction Tool doesn't Work | auto-migrated Construction Tool Doesn't Work low priority v_0.1.0 | When I'm going to make a new molecula nothing happens. I try to use the construction tool, but it doesn't work. I have used the program before an all is OK. But the last times doesn't work.
Reported by: *anonymous | 1.0 | Construction Tool doesn't Work - When I'm going to make a new molecula nothing happens. I try to use the construction tool, but it doesn't work. I have used the program before an all is OK. But the last times doesn't work.
Reported by: *anonymous | non_test | construction tool doesn t work when i m going to make a new molecula nothing happens i try to use the construction tool but it doesn t work i have used the program before an all is ok but the last times doesn t work reported by anonymous | 0 |
173,327 | 6,523,581,184 | IssuesEvent | 2017-08-29 09:14:16 | BuckleScript/bucklescript | https://api.github.com/repos/BuckleScript/bucklescript | closed | get rid of no export special handling | PRIORITY:HIGH | `no_export` is only used for demo purpose,
currently we already provide `[@@bs.config { no_export } ]` for such purpose, we can make all these behind the scene by using
```ocaml
[@@@bs.config{no_export}]
# 1 "toplevel" (* first two lines are hidden from user when clicking remove unused code *)
let b = 3
```
I am convinced to remove this ad hoc support to reduce maintenance overhead, the main thing is with such ad-hoc support - the invariant that the length of exported identifiers and export lambdas are the same is not held any more, we need add a special handling in the pipeline | 1.0 | get rid of no export special handling - `no_export` is only used for demo purpose,
currently we already provide `[@@bs.config { no_export } ]` for such purpose, we can make all these behind the scene by using
```ocaml
[@@@bs.config{no_export}]
# 1 "toplevel" (* first two lines are hidden from user when clicking remove unused code *)
let b = 3
```
I am convinced to remove this ad hoc support to reduce maintenance overhead, the main thing is with such ad-hoc support - the invariant that the length of exported identifiers and export lambdas are the same is not held any more, we need add a special handling in the pipeline | non_test | get rid of no export special handling no export is only used for demo purpose currently we already provide for such purpose we can make all these behind the scene by using ocaml toplevel first two lines are hidden from user when clicking remove unused code let b i am convinced to remove this ad hoc support to reduce maintenance overhead the main thing is with such ad hoc support the invariant that the length of exported identifiers and export lambdas are the same is not held any more we need add a special handling in the pipeline | 0 |
72,609 | 3,388,490,203 | IssuesEvent | 2015-11-29 10:29:38 | rathena/rathena | https://api.github.com/repos/rathena/rathena | closed | This NPC is not belong to npc/re/instances/WolfchevLaboratory.txt | bug:script mode:renewal priority:low server:map | I just tried to finding rA script while playing idRO. Found the reward NPC technically is not instance dependent.
```
// Weird Old Man (Rewards)
//============================================================
lighthalzen,342,291,4 script Weird old man#Bio4Reward 4_M_EINOLD,{
}
```
Why, because the required items can be dropped from lhz_dun03 and lhz_dun04, not only from `1@lhz` instance.
and also with this enchant NPC
```
// Sorcerer (Rewards)
//============================================================
lhz_cube,233,24,4 script Sorcerer#Bio4Reward 4_M_UMDANCEKID,{
}
```
So, if user want disable instance but just enable the quest, when instance file is commented, the quest is still.
| 1.0 | This NPC is not belong to npc/re/instances/WolfchevLaboratory.txt - I just tried to finding rA script while playing idRO. Found the reward NPC technically is not instance dependent.
```
// Weird Old Man (Rewards)
//============================================================
lighthalzen,342,291,4 script Weird old man#Bio4Reward 4_M_EINOLD,{
}
```
Why, because the required items can be dropped from lhz_dun03 and lhz_dun04, not only from `1@lhz` instance.
and also with this enchant NPC
```
// Sorcerer (Rewards)
//============================================================
lhz_cube,233,24,4 script Sorcerer#Bio4Reward 4_M_UMDANCEKID,{
}
```
So, if user want disable instance but just enable the quest, when instance file is commented, the quest is still.
| non_test | this npc is not belong to npc re instances wolfchevlaboratory txt i just tried to finding ra script while playing idro found the reward npc technically is not instance dependent weird old man rewards lighthalzen script weird old man m einold why because the required items can be dropped from lhz and lhz not only from lhz instance and also with this enchant npc sorcerer rewards lhz cube script sorcerer m umdancekid so if user want disable instance but just enable the quest when instance file is commented the quest is still | 0 |
372,320 | 25,995,571,496 | IssuesEvent | 2022-12-20 11:15:38 | aws/aws-cli | https://api.github.com/repos/aws/aws-cli | closed | Update aws cloudformation package docs to include missing resources | documentation feature-request cloudformation package-deploy customization community p2 | Confirm by changing [ ] to [x] below:
- [X] I've gone though the [User Guide](https://docs.aws.amazon.com/cli/latest/userguide/cli-chap-welcome.html) and the [API reference](https://docs.aws.amazon.com/cli/latest/reference/)
- [X] I've searched for [previous similar issues](https://github.com/aws/aws-cli/issues) and didn't find any solution
Issue is about usage on:
- [ ] Service API : I want to do X using Y service, what should I do?
- [ ] CLI : passing arguments or cli configurations.
- [X] Other/Not sure.
It appears that the `aws cloudformation package` command supports more resources than listed in the documentation.
Specifically, I can see these three resources supported, but not mentioned in the docs.
* `Location` property of the `ServerlessApplicationResource` (`AWS::Serverless::Application`),
* `ContentUri` of the `ServerlessLayerVersionResource` (`AWS::Serverless::LayerVersion`)
* `Content` property of the `LambdaLayerVersionResource` (`AWS::Lambda::LayerVersion`)
Resources are taken from https://github.com/aws/aws-cli/blob/daeaf7d46b7d14140a6c96ce931575ccf8dbf2d8/awscli/customizations/cloudformation/artifact_exporter.py#L500
Docs are https://awscli.amazonaws.com/v2/documentation/api/latest/reference/cloudformation/package.html
I think this file needs to be updated
https://github.com/aws/aws-cli/blob/develop/awscli/examples/cloudformation/_package_description.rst
I'm happy to make the PR, but I'll wait for a maintainer to respond first. | 1.0 | Update aws cloudformation package docs to include missing resources - Confirm by changing [ ] to [x] below:
- [X] I've gone though the [User Guide](https://docs.aws.amazon.com/cli/latest/userguide/cli-chap-welcome.html) and the [API reference](https://docs.aws.amazon.com/cli/latest/reference/)
- [X] I've searched for [previous similar issues](https://github.com/aws/aws-cli/issues) and didn't find any solution
Issue is about usage on:
- [ ] Service API : I want to do X using Y service, what should I do?
- [ ] CLI : passing arguments or cli configurations.
- [X] Other/Not sure.
It appears that the `aws cloudformation package` command supports more resources than listed in the documentation.
Specifically, I can see these three resources supported, but not mentioned in the docs.
* `Location` property of the `ServerlessApplicationResource` (`AWS::Serverless::Application`),
* `ContentUri` of the `ServerlessLayerVersionResource` (`AWS::Serverless::LayerVersion`)
* `Content` property of the `LambdaLayerVersionResource` (`AWS::Lambda::LayerVersion`)
Resources are taken from https://github.com/aws/aws-cli/blob/daeaf7d46b7d14140a6c96ce931575ccf8dbf2d8/awscli/customizations/cloudformation/artifact_exporter.py#L500
Docs are https://awscli.amazonaws.com/v2/documentation/api/latest/reference/cloudformation/package.html
I think this file needs to be updated
https://github.com/aws/aws-cli/blob/develop/awscli/examples/cloudformation/_package_description.rst
I'm happy to make the PR, but I'll wait for a maintainer to respond first. | non_test | update aws cloudformation package docs to include missing resources confirm by changing to below i ve gone though the and the i ve searched for and didn t find any solution issue is about usage on service api i want to do x using y service what should i do cli passing arguments or cli configurations other not sure it appears that the aws cloudformation package command supports more resources than listed in the documentation specifically i can see these three resources supported but not mentioned in the docs location property of the serverlessapplicationresource aws serverless application contenturi of the serverlesslayerversionresource aws serverless layerversion content property of the lambdalayerversionresource aws lambda layerversion resources are taken from docs are i think this file needs to be updated i m happy to make the pr but i ll wait for a maintainer to respond first | 0 |
23,994 | 4,055,456,822 | IssuesEvent | 2016-05-24 15:31:48 | ngageoint/hootenanny | https://api.github.com/repos/ngageoint/hootenanny | closed | Modify PrintNamesCmdTest.sh to sort the output | Category: Test Status: In Progress Type: Bug | Sometimes the output isn't in the same order and the test fails. Sorting should help alleviate this.
Be sure and use the "C" LANG when sorting or you may get different results on Ubuntu vs. CentOS.
See this for an example failure: http://selma:8765/view/critical/job/hoot-rpms-centos67/254/console | 1.0 | Modify PrintNamesCmdTest.sh to sort the output - Sometimes the output isn't in the same order and the test fails. Sorting should help alleviate this.
Be sure and use the "C" LANG when sorting or you may get different results on Ubuntu vs. CentOS.
See this for an example failure: http://selma:8765/view/critical/job/hoot-rpms-centos67/254/console | test | modify printnamescmdtest sh to sort the output sometimes the output isn t in the same order and the test fails sorting should help alleviate this be sure and use the c lang when sorting or you may get different results on ubuntu vs centos see this for an example failure | 1 |
14,343 | 17,370,285,464 | IssuesEvent | 2021-07-30 13:08:57 | 2i2c-org/team-compass | https://api.github.com/repos/2i2c-org/team-compass | opened | Use a password manager to share infrastructure passwords between the team | :label: team-process type: enhancement | # Summary
There are a few places where we share accounts to access the same services or webpages. An example of this is the grafana of a hub, where we don't necessarily want to create a new admin username for every single hub engineer.
For these cases, we currently follow a practice of "ask a team member what the password is". This makes it hard to know who has access to which passwords, and is an extra step team members must follow to get access. It's also something that has to be done _each time_ a new password is needed.
# Proposal
We use a team password application like [1Password](https://1password.com/). This would allow us to store the passwords in an encrypted service, and we could purchase a team account that would provide each of us access to them (I believe their base team account is $20/mo for 10 people, which isn't bad).
We could then use this to store any team passwords that don't require 2FA in order to log-in, and then the only step for providing access to a new team member is to get them a 1Password account.
# Actions
- [ ] Answer questions below, and if we wish to proceed...
- [ ] Write up a proposed process for how we share passwords
- [ ] Set up password manager and accounts for team members
- [ ] Write it up in team compass
# Questions
- Is anybody opposed to this idea? Would it lead us to any obvious anti-patterns?
- Any strong preferences for a particular password manager? 1Password vs. LastPass, for example.
| 1.0 | Use a password manager to share infrastructure passwords between the team - # Summary
There are a few places where we share accounts to access the same services or webpages. An example of this is the grafana of a hub, where we don't necessarily want to create a new admin username for every single hub engineer.
For these cases, we currently follow a practice of "ask a team member what the password is". This makes it hard to know who has access to which passwords, and is an extra step team members must follow to get access. It's also something that has to be done _each time_ a new password is needed.
# Proposal
We use a team password application like [1Password](https://1password.com/). This would allow us to store the passwords in an encrypted service, and we could purchase a team account that would provide each of us access to them (I believe their base team account is $20/mo for 10 people, which isn't bad).
We could then use this to store any team passwords that don't require 2FA in order to log-in, and then the only step for providing access to a new team member is to get them a 1Password account.
# Actions
- [ ] Answer questions below, and if we wish to proceed...
- [ ] Write up a proposed process for how we share passwords
- [ ] Set up password manager and accounts for team members
- [ ] Write it up in team compass
# Questions
- Is anybody opposed to this idea? Would it lead us to any obvious anti-patterns?
- Any strong preferences for a particular password manager? 1Password vs. LastPass, for example.
| non_test | use a password manager to share infrastructure passwords between the team summary there are a few places where we share accounts to access the same services or webpages an example of this is the grafana of a hub where we don t necessarily want to create a new admin username for every single hub engineer for these cases we currently follow a practice of ask a team member what the password is this makes it hard to know who has access to which passwords and is an extra step team members must follow to get access it s also something that has to be done each time a new password is needed proposal we use a team password application like this would allow us to store the passwords in an encrypted service and we could purchase a team account that would provide each of us access to them i believe their base team account is mo for people which isn t bad we could then use this to store any team passwords that don t require in order to log in and then the only step for providing access to a new team member is to get them a account actions answer questions below and if we wish to proceed write up a proposed process for how we share passwords set up password manager and accounts for team members write it up in team compass questions is anybody opposed to this idea would it lead us to any obvious anti patterns any strong preferences for a particular password manager vs lastpass for example | 0 |
154,899 | 19,765,546,834 | IssuesEvent | 2022-01-17 01:26:24 | tuanducdesign/web-netflix | https://api.github.com/repos/tuanducdesign/web-netflix | opened | CVE-2021-3918 (High) detected in json-schema-0.2.3.tgz | security vulnerability | ## CVE-2021-3918 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>json-schema-0.2.3.tgz</b></p></summary>
<p>JSON Schema validation and specifications</p>
<p>Library home page: <a href="https://registry.npmjs.org/json-schema/-/json-schema-0.2.3.tgz">https://registry.npmjs.org/json-schema/-/json-schema-0.2.3.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/json-schema/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-4.0.3.tgz (Root Library)
- jest-circus-26.6.0.tgz
- jest-runner-26.6.3.tgz
- jest-config-26.6.3.tgz
- jest-environment-jsdom-26.6.2.tgz
- jsdom-16.4.0.tgz
- request-2.88.2.tgz
- http-signature-1.2.0.tgz
- jsprim-1.4.1.tgz
- :x: **json-schema-0.2.3.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
json-schema is vulnerable to Improperly Controlled Modification of Object Prototype Attributes ('Prototype Pollution')
<p>Publish Date: 2021-11-13
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3918>CVE-2021-3918</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2021-3918">https://nvd.nist.gov/vuln/detail/CVE-2021-3918</a></p>
<p>Release Date: 2021-11-13</p>
<p>Fix Resolution: json-schema - 0.4.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-3918 (High) detected in json-schema-0.2.3.tgz - ## CVE-2021-3918 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>json-schema-0.2.3.tgz</b></p></summary>
<p>JSON Schema validation and specifications</p>
<p>Library home page: <a href="https://registry.npmjs.org/json-schema/-/json-schema-0.2.3.tgz">https://registry.npmjs.org/json-schema/-/json-schema-0.2.3.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/json-schema/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-4.0.3.tgz (Root Library)
- jest-circus-26.6.0.tgz
- jest-runner-26.6.3.tgz
- jest-config-26.6.3.tgz
- jest-environment-jsdom-26.6.2.tgz
- jsdom-16.4.0.tgz
- request-2.88.2.tgz
- http-signature-1.2.0.tgz
- jsprim-1.4.1.tgz
- :x: **json-schema-0.2.3.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
json-schema is vulnerable to Improperly Controlled Modification of Object Prototype Attributes ('Prototype Pollution')
<p>Publish Date: 2021-11-13
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3918>CVE-2021-3918</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2021-3918">https://nvd.nist.gov/vuln/detail/CVE-2021-3918</a></p>
<p>Release Date: 2021-11-13</p>
<p>Fix Resolution: json-schema - 0.4.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_test | cve high detected in json schema tgz cve high severity vulnerability vulnerable library json schema tgz json schema validation and specifications library home page a href path to dependency file package json path to vulnerable library node modules json schema package json dependency hierarchy react scripts tgz root library jest circus tgz jest runner tgz jest config tgz jest environment jsdom tgz jsdom tgz request tgz http signature tgz jsprim tgz x json schema tgz vulnerable library found in base branch master vulnerability details json schema is vulnerable to improperly controlled modification of object prototype attributes prototype pollution publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution json schema step up your open source security game with whitesource | 0 |
86,457 | 10,499,746,716 | IssuesEvent | 2019-09-26 09:02:44 | spring-projects/spring-framework | https://api.github.com/repos/spring-projects/spring-framework | closed | Improve docs for AnnotatedBeanDefinitionReader, @Configuration, and @ContextConfiguration regarding "annotated classes" | in: core in: test type: documentation | ## Original Description
I find this test interesting:
```java
@ExtendWith(SpringExtension.class)
@ContextConfiguration(classes = RestTemplate.class)
class MyTest {
@Autowired
RestTemplate restTemplate;
@Test
void check() {
assertThat(restTemplate).isNotNull();
}
}
```
This works because while loading a context, `AnnotationConfigContextLoader` registers `@ContextConfiguration` provided classes(`RestTemplate` in this case) to bean definitions, so that they are available for injections.
So, this is also possible:
```java
@ExtendWith(SpringExtension.class)
@ContextConfiguration(classes = { MyTest2.Foo.class, MyTest2.Bar.class })
class MyTest2 {
static class Foo {
public Foo(Bar bar) { // dependency to Bar
}
}
static class Bar {
public Bar() {
}
}
@Autowired
Foo foo;
@Autowired
Bar bar;
@Test
void check() {
assertThat(this.foo).isNotNull();
assertThat(this.bar).isNotNull();
}
}
```
I think this is not an intended usage of `@ContextConfiguration#classes`.
Probably, by default, filter-out or validate those classes to be `@Configuration` classes.
For the case of allowing non `@Configuration` classes (for example, lite-mode), probably provide an explicit option(new attribute) on `@ContextConfiguration`. e.g.: `@ContextConfiguration(classes=MyBean.class, liteMode=true)`
----
This is what I found in real world code base:
```java
@ExtendWith({MockitoExtension.class, SpringExtension.class})
@ContextConfiguration(classes = {ResourceBundleMessageSource.class, MyService.class,
MyExceptionHandler.class, MyPropertyConfiguration.class, RestTemplate.class})
public class MyExceptionHandlerTest {
// ...
}
```
So, would be nice not seeing such test class :)
---
## Deliverables
Improve documentation for the following regarding "annotated classes".
- [x] `AnnotationConfigRegistry`
- [x] `AnnotationConfigApplicationContext`
- [x] `AnnotationConfigWebApplicationContext`
- [x] `AnnotatedBeanDefinitionReader`
- [x] `@Configuration`
- [x] `@Import`
- [x] `AnnotationConfigContextLoader`
- [x] `AnnotationConfigContextLoaderUtils`
- [x] `@ContextConfiguration`
- [x] Reference Manual: `testing.adoc`
| 1.0 | Improve docs for AnnotatedBeanDefinitionReader, @Configuration, and @ContextConfiguration regarding "annotated classes" - ## Original Description
I find this test interesting:
```java
@ExtendWith(SpringExtension.class)
@ContextConfiguration(classes = RestTemplate.class)
class MyTest {
@Autowired
RestTemplate restTemplate;
@Test
void check() {
assertThat(restTemplate).isNotNull();
}
}
```
This works because while loading a context, `AnnotationConfigContextLoader` registers `@ContextConfiguration` provided classes(`RestTemplate` in this case) to bean definitions, so that they are available for injections.
So, this is also possible:
```java
@ExtendWith(SpringExtension.class)
@ContextConfiguration(classes = { MyTest2.Foo.class, MyTest2.Bar.class })
class MyTest2 {
static class Foo {
public Foo(Bar bar) { // dependency to Bar
}
}
static class Bar {
public Bar() {
}
}
@Autowired
Foo foo;
@Autowired
Bar bar;
@Test
void check() {
assertThat(this.foo).isNotNull();
assertThat(this.bar).isNotNull();
}
}
```
I think this is not an intended usage of `@ContextConfiguration#classes`.
Probably, by default, filter-out or validate those classes to be `@Configuration` classes.
For the case of allowing non `@Configuration` classes (for example, lite-mode), probably provide an explicit option(new attribute) on `@ContextConfiguration`. e.g.: `@ContextConfiguration(classes=MyBean.class, liteMode=true)`
----
This is what I found in real world code base:
```java
@ExtendWith({MockitoExtension.class, SpringExtension.class})
@ContextConfiguration(classes = {ResourceBundleMessageSource.class, MyService.class,
MyExceptionHandler.class, MyPropertyConfiguration.class, RestTemplate.class})
public class MyExceptionHandlerTest {
// ...
}
```
So, would be nice not seeing such test class :)
---
## Deliverables
Improve documentation for the following regarding "annotated classes".
- [x] `AnnotationConfigRegistry`
- [x] `AnnotationConfigApplicationContext`
- [x] `AnnotationConfigWebApplicationContext`
- [x] `AnnotatedBeanDefinitionReader`
- [x] `@Configuration`
- [x] `@Import`
- [x] `AnnotationConfigContextLoader`
- [x] `AnnotationConfigContextLoaderUtils`
- [x] `@ContextConfiguration`
- [x] Reference Manual: `testing.adoc`
| non_test | improve docs for annotatedbeandefinitionreader configuration and contextconfiguration regarding annotated classes original description i find this test interesting java extendwith springextension class contextconfiguration classes resttemplate class class mytest autowired resttemplate resttemplate test void check assertthat resttemplate isnotnull this works because while loading a context annotationconfigcontextloader registers contextconfiguration provided classes resttemplate in this case to bean definitions so that they are available for injections so this is also possible java extendwith springextension class contextconfiguration classes foo class bar class class static class foo public foo bar bar dependency to bar static class bar public bar autowired foo foo autowired bar bar test void check assertthat this foo isnotnull assertthat this bar isnotnull i think this is not an intended usage of contextconfiguration classes probably by default filter out or validate those classes to be configuration classes for the case of allowing non configuration classes for example lite mode probably provide an explicit option new attribute on contextconfiguration e g contextconfiguration classes mybean class litemode true this is what i found in real world code base java extendwith mockitoextension class springextension class contextconfiguration classes resourcebundlemessagesource class myservice class myexceptionhandler class mypropertyconfiguration class resttemplate class public class myexceptionhandlertest so would be nice not seeing such test class deliverables improve documentation for the following regarding annotated classes annotationconfigregistry annotationconfigapplicationcontext annotationconfigwebapplicationcontext annotatedbeandefinitionreader configuration import annotationconfigcontextloader annotationconfigcontextloaderutils contextconfiguration reference manual testing adoc | 0 |
8,352 | 4,231,022,521 | IssuesEvent | 2016-07-04 14:21:39 | ParabolInc/action | https://api.github.com/repos/ParabolInc/action | closed | npm build:server depends on theme.json too | building | ## Issue - Bug
Consider adding `WebpackShellPlugin` to server webpack build configuration.
- **Estimated effort:** 1 points ([see CONTRIBUTING.md](https://github.com/ParabolInc/action/blob/master/CONTRIBUTING.md#points-and-sizes))
| 1.0 | npm build:server depends on theme.json too - ## Issue - Bug
Consider adding `WebpackShellPlugin` to server webpack build configuration.
- **Estimated effort:** 1 points ([see CONTRIBUTING.md](https://github.com/ParabolInc/action/blob/master/CONTRIBUTING.md#points-and-sizes))
| non_test | npm build server depends on theme json too issue bug consider adding webpackshellplugin to server webpack build configuration estimated effort points | 0 |
491,540 | 14,165,963,005 | IssuesEvent | 2020-11-12 08:11:09 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | opened | [0.9.2 develop-111] Glitch/exploit with Steam tractor Plow | Category: Gameplay Priority: Medium | Step to reproduce:
- spawn tractor with plow, start tilling:

- press Q to turn off module and imideatly press Q to turn on module. You will have module turned on with basic tractor speed. Also plow is not in right position.
video:
https://drive.google.com/file/d/1SKVNKdx8nGK9MU3XrdFw3fchyfmQp6DX/view?usp=sharing | 1.0 | [0.9.2 develop-111] Glitch/exploit with Steam tractor Plow - Step to reproduce:
- spawn tractor with plow, start tilling:

- press Q to turn off module and imideatly press Q to turn on module. You will have module turned on with basic tractor speed. Also plow is not in right position.
video:
https://drive.google.com/file/d/1SKVNKdx8nGK9MU3XrdFw3fchyfmQp6DX/view?usp=sharing | non_test | glitch exploit with steam tractor plow step to reproduce spawn tractor with plow start tilling press q to turn off module and imideatly press q to turn on module you will have module turned on with basic tractor speed also plow is not in right position video | 0 |
103,271 | 8,893,615,621 | IssuesEvent | 2019-01-16 00:09:41 | astropy/astropy | https://api.github.com/repos/astropy/astropy | closed | Test failure with latest developer version of Matplotlib | testing visualization | Here's an example log:
https://circleci.com/gh/astrofrog/astropy/1075?utm_campaign=vcs-integration-link&utm_medium=referral&utm_source=github-build-link
This seems related to the axis unit functionality. We should do a bisect to figure out which change in Matplotlib caused this. | 1.0 | Test failure with latest developer version of Matplotlib - Here's an example log:
https://circleci.com/gh/astrofrog/astropy/1075?utm_campaign=vcs-integration-link&utm_medium=referral&utm_source=github-build-link
This seems related to the axis unit functionality. We should do a bisect to figure out which change in Matplotlib caused this. | test | test failure with latest developer version of matplotlib here s an example log this seems related to the axis unit functionality we should do a bisect to figure out which change in matplotlib caused this | 1 |
690,894 | 23,676,455,519 | IssuesEvent | 2022-08-28 06:33:06 | Gilded-Games/The-Aether | https://api.github.com/repos/Gilded-Games/The-Aether | closed | Feature: Boss Advancements | priority/medium status/pending-review type/feature version/1.19 | - Advancements for bosses.
- Likely don't need a custom trigger, just need to manually call to the existing killed entity trigger for everyone in the boss room when the boss is killed. | 1.0 | Feature: Boss Advancements - - Advancements for bosses.
- Likely don't need a custom trigger, just need to manually call to the existing killed entity trigger for everyone in the boss room when the boss is killed. | non_test | feature boss advancements advancements for bosses likely don t need a custom trigger just need to manually call to the existing killed entity trigger for everyone in the boss room when the boss is killed | 0 |
196,514 | 14,876,634,287 | IssuesEvent | 2021-01-20 01:18:09 | jasonfleming/asgs | https://api.github.com/repos/jasonfleming/asgs | closed | remove ALL conditional env set ups in platforms.sh | in integration testing | This may require creating or moving it to exiting mechanisms (like config/operator_defaults.sh) to provide for additional customizations in the asgs shell, but this should be discouraged.
Once we clean up `platforms.sh` we will be in a better position to evaluate the utility of `config/operator_defaults.sh` even being in the repository or move the role of it to the use of a `~/.asgsh_profile` type approach (that should be discouraged). | 1.0 | remove ALL conditional env set ups in platforms.sh - This may require creating or moving it to exiting mechanisms (like config/operator_defaults.sh) to provide for additional customizations in the asgs shell, but this should be discouraged.
Once we clean up `platforms.sh` we will be in a better position to evaluate the utility of `config/operator_defaults.sh` even being in the repository or move the role of it to the use of a `~/.asgsh_profile` type approach (that should be discouraged). | test | remove all conditional env set ups in platforms sh this may require creating or moving it to exiting mechanisms like config operator defaults sh to provide for additional customizations in the asgs shell but this should be discouraged once we clean up platforms sh we will be in a better position to evaluate the utility of config operator defaults sh even being in the repository or move the role of it to the use of a asgsh profile type approach that should be discouraged | 1 |
803,077 | 29,116,576,610 | IssuesEvent | 2023-05-17 01:53:22 | ML4GW/BBHNet | https://api.github.com/repos/ML4GW/BBHNet | closed | "hopeless" injections | high priority data | To get the most value out of VT calculations from injection runs, the CBC pipelines typically will not inject `hopeless` injections. These are defined as injections with SNR below some value such that they are essentially impossible to recover.
This would require calculating the SNR of signals in the `Sampler` using `ml4gw`, and rejecting those below some SNR threshold until we've generated `num_signals` events above the threshold.
The changes required for this might fit in well with #267
| 1.0 | "hopeless" injections - To get the most value out of VT calculations from injection runs, the CBC pipelines typically will not inject `hopeless` injections. These are defined as injections with SNR below some value such that they are essentially impossible to recover.
This would require calculating the SNR of signals in the `Sampler` using `ml4gw`, and rejecting those below some SNR threshold until we've generated `num_signals` events above the threshold.
The changes required for this might fit in well with #267
| non_test | hopeless injections to get the most value out of vt calculations from injection runs the cbc pipelines typically will not inject hopeless injections these are defined as injections with snr below some value such that they are essentially impossible to recover this would require calculating the snr of signals in the sampler using and rejecting those below some snr threshold until we ve generated num signals events above the threshold the changes required for this might fit in well with | 0 |
94,412 | 27,186,451,551 | IssuesEvent | 2023-02-19 09:15:57 | curl/curl | https://api.github.com/repos/curl/curl | closed | Unable detect OpenSSL while set host to arm-linux-gnueabi | build TLS | While I configure the cURL 7.88.0 source with:
> ./configure --with-openssl
It works correctly.
But if I set host to arm-linux-gnueabi, configure like these:
> ./configure --with-openssl --host=arm-linux-gnueabi
> ./configure --with-openssl --host=arm-linux-gnueabi --target=arm-linux-gnueabi
> ./configure --with-openssl --host=arm-linux-gnueabi --target=arm-linux-gnueabi CC=arm-linux-gnueabi-gcc
I got an error:
> configure: error: --with-openssl was given but OpenSSL could not be detected
Ubuntu version is 14.04.6, arm-linux-gnueabi-gcc version is 4.7.3, OpenSSL version is 1.0.1f.
What's the reason and how can I solve it? Please help me, Thanks. | 1.0 | Unable detect OpenSSL while set host to arm-linux-gnueabi - While I configure the cURL 7.88.0 source with:
> ./configure --with-openssl
It works correctly.
But if I set host to arm-linux-gnueabi, configure like these:
> ./configure --with-openssl --host=arm-linux-gnueabi
> ./configure --with-openssl --host=arm-linux-gnueabi --target=arm-linux-gnueabi
> ./configure --with-openssl --host=arm-linux-gnueabi --target=arm-linux-gnueabi CC=arm-linux-gnueabi-gcc
I got an error:
> configure: error: --with-openssl was given but OpenSSL could not be detected
Ubuntu version is 14.04.6, arm-linux-gnueabi-gcc version is 4.7.3, OpenSSL version is 1.0.1f.
What's the reason and how can I solve it? Please help me, Thanks. | non_test | unable detect openssl while set host to arm linux gnueabi while i configure the curl source with configure with openssl it works correctly but if i set host to arm linux gnueabi configure like these configure with openssl host arm linux gnueabi configure with openssl host arm linux gnueabi target arm linux gnueabi configure with openssl host arm linux gnueabi target arm linux gnueabi cc arm linux gnueabi gcc i got an error configure error with openssl was given but openssl could not be detected ubuntu version is arm linux gnueabi gcc version is openssl version is what s the reason and how can i solve it please help me thanks | 0 |
121,494 | 4,817,285,569 | IssuesEvent | 2016-11-04 13:13:31 | handsontable/handsontable | https://api.github.com/repos/handsontable/handsontable | closed | Allow beforeAutofill to be blocked via returning false | Core: callbacks Priority: normal Type: Feature | Currently we are not able to block [`beforeAutofill`](https://docs.handsontable.com/pro/1.7.1/Hooks.html#event:beforeAutofill) by returning `false`.
| 1.0 | Allow beforeAutofill to be blocked via returning false - Currently we are not able to block [`beforeAutofill`](https://docs.handsontable.com/pro/1.7.1/Hooks.html#event:beforeAutofill) by returning `false`.
| non_test | allow beforeautofill to be blocked via returning false currently we are not able to block by returning false | 0 |
282,979 | 24,509,703,143 | IssuesEvent | 2022-10-10 20:02:15 | commercialhaskell/stackage | https://api.github.com/repos/commercialhaskell/stackage | opened | hspec-contrib 0.5.1.1 | failure: test-suite | ```
Preprocessing test suite 'spec' for hspec-contrib-0.5.1.1..
Building test suite 'spec' for hspec-contrib-0.5.1.1..
[3 of 5] Compiling Test.Hspec.Contrib.HUnitSpec
/var/stackage/work/unpack-dir/unpacked/hspec-contrib-0.5.1.1-e6ff010b10d5f3707f4cfbd0af0c39ce36efa3a71a789e5b68d2c87aaa8bdbab/test/Test/Hspec/Contrib/HUnitSpec.hs:11:70: error:
• Couldn't match type: [SpecTree ()]
with: (a0, [Tree a1 (Item a2)])
Expected: IO (a0, [Tree a1 (Item a2)])
Actual: IO [SpecTree ()]
• In the second argument of ‘(<$>)’, namely
‘runSpecM (fromHUnitTest a)’
In the first argument of ‘shouldReturn’, namely
‘bimapForest (const ()) itemRequirement . snd
<$> runSpecM (fromHUnitTest a)’
In the expression:
bimapForest (const ()) itemRequirement . snd
<$> runSpecM (fromHUnitTest a)
`shouldReturn` b
|
11 | a `shouldYield` b = bimapForest (const ()) itemRequirement . snd <$> runSpecM (fromHUnitTest a) `shouldReturn` b
| ^^^^^^^^^^^^^^^^^^^^^^^^^^
curator: Received ExitFailure 1 when running
Raw command: /usr/bin/stack --terminal --system-ghc build --test --bench --test-suite-timeout=600 --no-rerun-tests --no-run-benchmarks --haddock --no-interleaved-output --jobs=3
Run from: unpack-dir
``` | 1.0 | hspec-contrib 0.5.1.1 - ```
Preprocessing test suite 'spec' for hspec-contrib-0.5.1.1..
Building test suite 'spec' for hspec-contrib-0.5.1.1..
[3 of 5] Compiling Test.Hspec.Contrib.HUnitSpec
/var/stackage/work/unpack-dir/unpacked/hspec-contrib-0.5.1.1-e6ff010b10d5f3707f4cfbd0af0c39ce36efa3a71a789e5b68d2c87aaa8bdbab/test/Test/Hspec/Contrib/HUnitSpec.hs:11:70: error:
• Couldn't match type: [SpecTree ()]
with: (a0, [Tree a1 (Item a2)])
Expected: IO (a0, [Tree a1 (Item a2)])
Actual: IO [SpecTree ()]
• In the second argument of ‘(<$>)’, namely
‘runSpecM (fromHUnitTest a)’
In the first argument of ‘shouldReturn’, namely
‘bimapForest (const ()) itemRequirement . snd
<$> runSpecM (fromHUnitTest a)’
In the expression:
bimapForest (const ()) itemRequirement . snd
<$> runSpecM (fromHUnitTest a)
`shouldReturn` b
|
11 | a `shouldYield` b = bimapForest (const ()) itemRequirement . snd <$> runSpecM (fromHUnitTest a) `shouldReturn` b
| ^^^^^^^^^^^^^^^^^^^^^^^^^^
curator: Received ExitFailure 1 when running
Raw command: /usr/bin/stack --terminal --system-ghc build --test --bench --test-suite-timeout=600 --no-rerun-tests --no-run-benchmarks --haddock --no-interleaved-output --jobs=3
Run from: unpack-dir
``` | test | hspec contrib preprocessing test suite spec for hspec contrib building test suite spec for hspec contrib compiling test hspec contrib hunitspec var stackage work unpack dir unpacked hspec contrib test test hspec contrib hunitspec hs error • couldn t match type with expected io actual io • in the second argument of ‘ ’ namely ‘runspecm fromhunittest a ’ in the first argument of ‘shouldreturn’ namely ‘bimapforest const itemrequirement snd runspecm fromhunittest a ’ in the expression bimapforest const itemrequirement snd runspecm fromhunittest a shouldreturn b a shouldyield b bimapforest const itemrequirement snd runspecm fromhunittest a shouldreturn b curator received exitfailure when running raw command usr bin stack terminal system ghc build test bench test suite timeout no rerun tests no run benchmarks haddock no interleaved output jobs run from unpack dir | 1 |
419,942 | 12,230,534,177 | IssuesEvent | 2020-05-04 05:11:17 | roed314/seminars | https://api.github.com/repos/roed314/seminars | opened | Logos on acknowledgments page | low priority | It would be nice to add some logos to the acknowledgments page (e.g., the Simons Foundation logo). | 1.0 | Logos on acknowledgments page - It would be nice to add some logos to the acknowledgments page (e.g., the Simons Foundation logo). | non_test | logos on acknowledgments page it would be nice to add some logos to the acknowledgments page e g the simons foundation logo | 0 |
92,517 | 8,366,483,286 | IssuesEvent | 2018-10-04 09:14:52 | stratisproject/StratisBitcoinFullNode | https://api.github.com/repos/stratisproject/StratisBitcoinFullNode | opened | MiningAndPropagatingPOS [Mining and Staking tests impl] | Size XS: 0.5-2 days Test Coverage | **MiningAndPropagatingPOS [[Mining and Staking tests impl]**
- MineBlockCheckPeerHasNewBlock | Mine a block and propagate to peers (success)
- MineBlockStakeAtInsufficientHeightError | Mine a block insufficient height to stake (error)
| 1.0 | MiningAndPropagatingPOS [Mining and Staking tests impl] - **MiningAndPropagatingPOS [[Mining and Staking tests impl]**
- MineBlockCheckPeerHasNewBlock | Mine a block and propagate to peers (success)
- MineBlockStakeAtInsufficientHeightError | Mine a block insufficient height to stake (error)
| test | miningandpropagatingpos miningandpropagatingpos mineblockcheckpeerhasnewblock mine a block and propagate to peers success mineblockstakeatinsufficientheighterror mine a block insufficient height to stake error | 1 |
6,485 | 2,848,367,610 | IssuesEvent | 2015-05-29 22:31:19 | gosha1128/NYUAMS | https://api.github.com/repos/gosha1128/NYUAMS | closed | Test Case: Admin>emails - table stops populating | 0.0 DEV 1.4 VERIFIED IN DEV 2.3 high fixed load testing | Table stops populating after 21 lines have been populated
https://drive.google.com/open?id=0B_9JEK6SVk1bamNTMEJqUExNQW8&authuser=0 | 1.0 | Test Case: Admin>emails - table stops populating - Table stops populating after 21 lines have been populated
https://drive.google.com/open?id=0B_9JEK6SVk1bamNTMEJqUExNQW8&authuser=0 | test | test case admin emails table stops populating table stops populating after lines have been populated | 1 |
12,898 | 15,169,784,361 | IssuesEvent | 2021-02-12 21:49:10 | Foundry376/Mailspring | https://api.github.com/repos/Foundry376/Mailspring | closed | Can't configure archive folder for Gmail account? | audit bug mailserver-incompatibility needs-reproduction | <!--
Thanks for taking the time to file an issue! The Mailspring community uses GitHub issues to coordinate development.
Before filling out this form, try searching for both open and closed issues here: https://github.com/Foundry376/Mailspring/issues. Upvoting an existing sync issue is MUCH better than submitting a new one - it often takes a while for them to be merged and we use issue upvotes to prioritize development.
-->
##### What is the behavior you're seeing?
New mailspring user here. I setup my gmail account and I was surprised to note that I can't configure the "Archive" mailbox for gmail:

I am able to configure this mailbox for the other accounts I've setup. This is particularly important to me as I have all of my unattended mailing lists messages in the default archive mailbox "[Gmail]/All Mail" and I want to use a different archive mailbox, just like my other mail clients (K9).
##### What email provider (Gmail, Exchange, etc.) are you trying to use?
Gmail
##### What operating system are you using?
NixOS
##### What version of Mailspring are you using?
1.7.8
##### Is the issue reproducible with a particular attachment, message, signature, etc?
Irrelevant.
##### Other comments
Damn this application is beautiful! :heart_eyes: | True | Can't configure archive folder for Gmail account? - <!--
Thanks for taking the time to file an issue! The Mailspring community uses GitHub issues to coordinate development.
Before filling out this form, try searching for both open and closed issues here: https://github.com/Foundry376/Mailspring/issues. Upvoting an existing sync issue is MUCH better than submitting a new one - it often takes a while for them to be merged and we use issue upvotes to prioritize development.
-->
##### What is the behavior you're seeing?
New mailspring user here. I setup my gmail account and I was surprised to note that I can't configure the "Archive" mailbox for gmail:

I am able to configure this mailbox for the other accounts I've setup. This is particularly important to me as I have all of my unattended mailing lists messages in the default archive mailbox "[Gmail]/All Mail" and I want to use a different archive mailbox, just like my other mail clients (K9).
##### What email provider (Gmail, Exchange, etc.) are you trying to use?
Gmail
##### What operating system are you using?
NixOS
##### What version of Mailspring are you using?
1.7.8
##### Is the issue reproducible with a particular attachment, message, signature, etc?
Irrelevant.
##### Other comments
Damn this application is beautiful! :heart_eyes: | non_test | can t configure archive folder for gmail account thanks for taking the time to file an issue the mailspring community uses github issues to coordinate development before filling out this form try searching for both open and closed issues here upvoting an existing sync issue is much better than submitting a new one it often takes a while for them to be merged and we use issue upvotes to prioritize development what is the behavior you re seeing new mailspring user here i setup my gmail account and i was surprised to note that i can t configure the archive mailbox for gmail i am able to configure this mailbox for the other accounts i ve setup this is particularly important to me as i have all of my unattended mailing lists messages in the default archive mailbox all mail and i want to use a different archive mailbox just like my other mail clients what email provider gmail exchange etc are you trying to use gmail what operating system are you using nixos what version of mailspring are you using is the issue reproducible with a particular attachment message signature etc irrelevant other comments damn this application is beautiful heart eyes | 0 |
268,111 | 23,345,643,343 | IssuesEvent | 2022-08-09 17:41:16 | wso2/product-is | https://api.github.com/repos/wso2/product-is | closed | Console app does not display device_code grant type | ui bug frontend 6.0.0-rc-testing | **Describe the issue:**
he console can be used to view service providers and update the necessary configurations. The protocol section of the service provider only shows limited grant types for service provider applications. device_code grant type is not under allowed grant types (where as management console does list that)

**How to reproduce:**
Log into console app
Go to applications and select Console app.
View protocol section of the service provider app.
**Expected behavior:**
device_code should be listed under allowed grant_types
| 1.0 | Console app does not display device_code grant type - **Describe the issue:**
he console can be used to view service providers and update the necessary configurations. The protocol section of the service provider only shows limited grant types for service provider applications. device_code grant type is not under allowed grant types (where as management console does list that)

**How to reproduce:**
Log into console app
Go to applications and select Console app.
View protocol section of the service provider app.
**Expected behavior:**
device_code should be listed under allowed grant_types
| test | console app does not display device code grant type describe the issue he console can be used to view service providers and update the necessary configurations the protocol section of the service provider only shows limited grant types for service provider applications device code grant type is not under allowed grant types where as management console does list that how to reproduce log into console app go to applications and select console app view protocol section of the service provider app expected behavior device code should be listed under allowed grant types | 1 |
371,759 | 10,981,078,813 | IssuesEvent | 2019-11-30 18:59:43 | BendroCorp/bendrocorp-app | https://api.github.com/repos/BendroCorp/bendrocorp-app | opened | System Map Search View | api supported effort: high feature priority:high | - [ ] Offer a map view or a searchable list view - switch via toggle and persist preference
- [ ] Search View (will be the default mobile view when added)
- [ ] Searching for something adds it to the recently viewed list at the top of the system map
- [ ] Has a list of all items contained within system map and their essential data
- [ ] Searchable by title
- [ ] Opens to an object details page | 1.0 | System Map Search View - - [ ] Offer a map view or a searchable list view - switch via toggle and persist preference
- [ ] Search View (will be the default mobile view when added)
- [ ] Searching for something adds it to the recently viewed list at the top of the system map
- [ ] Has a list of all items contained within system map and their essential data
- [ ] Searchable by title
- [ ] Opens to an object details page | non_test | system map search view offer a map view or a searchable list view switch via toggle and persist preference search view will be the default mobile view when added searching for something adds it to the recently viewed list at the top of the system map has a list of all items contained within system map and their essential data searchable by title opens to an object details page | 0 |
193,712 | 14,660,464,836 | IssuesEvent | 2020-12-28 23:47:39 | github-vet/rangeloop-pointer-findings | https://api.github.com/repos/github-vet/rangeloop-pointer-findings | closed | gravitational/teleport: integration/integration_test.go; 281 LoC | fresh huge test |
Found a possible issue in [gravitational/teleport](https://www.github.com/gravitational/teleport) at [integration/integration_test.go](https://github.com/gravitational/teleport/blob/38c9804738112eceedd209d9cb5bda7ffe3ab6f5/integration/integration_test.go#L299-L579)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message.
> range-loop variable tt used in defer or goroutine at line 378
[Click here to see the code in its original context.](https://github.com/gravitational/teleport/blob/38c9804738112eceedd209d9cb5bda7ffe3ab6f5/integration/integration_test.go#L299-L579)
<details>
<summary>Click here to show the 281 line(s) of Go which triggered the analyzer.</summary>
```go
for _, tt := range tests {
comment := check.Commentf(tt.comment)
makeConfig := func() (*check.C, []string, []*InstanceSecrets, *service.Config) {
clusterConfig, err := services.NewClusterConfig(services.ClusterConfigSpecV3{
SessionRecording: tt.inRecordLocation,
Audit: services.AuditConfig{AuditSessionsURI: tt.auditSessionsURI},
LocalAuth: services.NewBool(true),
})
c.Assert(err, check.IsNil, comment)
tconf := s.defaultServiceConfig()
tconf.Auth.Enabled = true
tconf.Auth.ClusterConfig = clusterConfig
tconf.Proxy.Enabled = true
tconf.Proxy.DisableWebService = true
tconf.Proxy.DisableWebInterface = true
tconf.SSH.Enabled = true
return c, nil, nil, tconf
}
t := s.newTeleportWithConfig(makeConfig())
defer t.StopAll()
// Start a node.
nodeSSHPort := s.getPorts(1)[0]
nodeConfig := func() *service.Config {
tconf := s.defaultServiceConfig()
tconf.HostUUID = "node"
tconf.Hostname = "node"
tconf.SSH.Enabled = true
tconf.SSH.Addr.Addr = net.JoinHostPort(t.Hostname, fmt.Sprintf("%v", nodeSSHPort))
return tconf
}
nodeProcess, err := t.StartNode(nodeConfig())
c.Assert(err, check.IsNil, comment)
// get access to a authClient for the cluster
site := t.GetSiteAPI(Site)
c.Assert(site, check.NotNil, comment)
// wait 10 seconds for both nodes to show up, otherwise
// we'll have trouble connecting to the node below.
waitForNodes := func(site auth.ClientI, count int) error {
tickCh := time.Tick(500 * time.Millisecond)
stopCh := time.After(10 * time.Second)
for {
select {
case <-tickCh:
nodesInSite, err := site.GetNodes(defaults.Namespace, services.SkipValidation())
if err != nil && !trace.IsNotFound(err) {
return trace.Wrap(err)
}
if got, want := len(nodesInSite), count; got == want {
return nil
}
case <-stopCh:
return trace.BadParameter("waited 10s, did find %v nodes", count)
}
}
}
err = waitForNodes(site, 2)
c.Assert(err, check.IsNil, comment)
// should have no sessions:
sessions, err := site.GetSessions(defaults.Namespace)
c.Assert(err, check.IsNil, comment)
c.Assert(len(sessions), check.Equals, 0, comment)
// create interactive session (this goroutine is this user's terminal time)
endC := make(chan error)
myTerm := NewTerminal(250)
go func() {
cl, err := t.NewClient(ClientConfig{
Login: s.me.Username,
Cluster: Site,
Host: Host,
Port: nodeSSHPort,
ForwardAgent: tt.inForwardAgent,
})
c.Assert(err, check.IsNil, comment)
cl.Stdout = myTerm
cl.Stdin = myTerm
err = cl.SSH(context.TODO(), []string{}, false)
endC <- err
}()
// wait until we've found the session in the audit log
getSession := func(site auth.ClientI) (*session.Session, error) {
tickCh := time.Tick(500 * time.Millisecond)
stopCh := time.After(10 * time.Second)
for {
select {
case <-tickCh:
sessions, err = site.GetSessions(defaults.Namespace)
if err != nil {
return nil, trace.Wrap(err)
}
if len(sessions) != 1 {
continue
}
return &sessions[0], nil
case <-stopCh:
return nil, trace.BadParameter("unable to find sessions after 10s (mode=%v)", tt.inRecordLocation)
}
}
}
session, err := getSession(site)
c.Assert(err, check.IsNil, comment)
// wait for the user to join this session:
for len(session.Parties) == 0 {
time.Sleep(time.Millisecond * 5)
session, err = site.GetSession(defaults.Namespace, sessions[0].ID)
c.Assert(err, check.IsNil, comment)
}
// make sure it's us who joined! :)
c.Assert(session.Parties[0].User, check.Equals, s.me.Username, comment)
// lets type "echo hi" followed by "enter" and then "exit" + "enter":
myTerm.Type("\aecho hi\n\r\aexit\n\r\a")
// wait for session to end:
select {
case <-endC:
case <-time.After(10 * time.Second):
c.Fatalf("%s: Timeout waiting for session to finish.", tt.comment)
}
// wait for the upload of the right session to complete
timeoutC := time.After(10 * time.Second)
loop:
for {
select {
case event := <-t.UploadEventsC:
if event.SessionID != string(session.ID) {
c.Logf("Skipping mismatching session %v, expecting upload of %v.", event.SessionID, session.ID)
continue
}
break loop
case <-timeoutC:
dumpGoroutineProfile()
c.Fatalf("%s: Timeout waiting for upload of session %v to complete to %v",
tt.comment, session.ID, tt.auditSessionsURI)
}
}
// read back the entire session (we have to try several times until we get back
// everything because the session is closing)
var sessionStream []byte
for i := 0; i < 6; i++ {
sessionStream, err = site.GetSessionChunk(defaults.Namespace, session.ID, 0, events.MaxChunkBytes)
c.Assert(err, check.IsNil, comment)
if strings.Contains(string(sessionStream), "exit") {
break
}
time.Sleep(time.Millisecond * 250)
if i >= 5 {
// session stream keeps coming back short
c.Fatalf("%s: Stream is not getting data: %q.", tt.comment, string(sessionStream))
}
}
// see what we got. It looks different based on bash settings, but here it is
// on Ev's machine (hostname is 'edsger'):
//
// edsger ~: echo hi
// hi
// edsger ~: exit
// logout
//
sessionComment := check.Commentf("%q", string(sessionStream))
c.Assert(strings.Contains(string(sessionStream), "echo hi"), check.Equals, true, sessionComment)
c.Assert(strings.Contains(string(sessionStream), "exit"), check.Equals, true, sessionComment)
// Wait until session.start, session.leave, and session.end events have arrived.
getSessions := func(site auth.ClientI) ([]events.EventFields, error) {
tickCh := time.Tick(500 * time.Millisecond)
stopCh := time.After(10 * time.Second)
for {
select {
case <-tickCh:
// Get all session events from the backend.
sessionEvents, err := site.GetSessionEvents(defaults.Namespace, session.ID, 0, false)
if err != nil {
return nil, trace.Wrap(err)
}
// Look through all session events for the three wanted.
var hasStart bool
var hasEnd bool
var hasLeave bool
for _, se := range sessionEvents {
if se.GetType() == events.SessionStartEvent {
hasStart = true
}
if se.GetType() == events.SessionEndEvent {
hasEnd = true
}
if se.GetType() == events.SessionLeaveEvent {
hasLeave = true
}
}
// Make sure all three events were found.
if hasStart && hasEnd && hasLeave {
return sessionEvents, nil
}
case <-stopCh:
return nil, trace.BadParameter("unable to find all session events after 10s (mode=%v)", tt.inRecordLocation)
}
}
}
history, err := getSessions(site)
c.Assert(err, check.IsNil, comment)
getChunk := func(e events.EventFields, maxlen int) string {
offset := e.GetInt("offset")
length := e.GetInt("bytes")
if length == 0 {
return ""
}
if length > maxlen {
length = maxlen
}
return string(sessionStream[offset : offset+length])
}
findByType := func(et string) events.EventFields {
for _, e := range history {
if e.GetType() == et {
return e
}
}
return nil
}
// there should alwys be 'session.start' event (and it must be first)
first := history[0]
start := findByType(events.SessionStartEvent)
c.Assert(start, check.DeepEquals, first, comment)
c.Assert(start.GetInt("bytes"), check.Equals, 0, comment)
c.Assert(start.GetString(events.SessionEventID) != "", check.Equals, true, comment)
c.Assert(start.GetString(events.TerminalSize) != "", check.Equals, true, comment)
// If session are being recorded at nodes, the SessionServerID should contain
// the ID of the node. If sessions are being recorded at the proxy, then
// SessionServerID should be that of the proxy.
expectedServerID := nodeProcess.Config.HostUUID
if services.IsRecordAtProxy(tt.inRecordLocation) {
expectedServerID = t.Process.Config.HostUUID
}
c.Assert(start.GetString(events.SessionServerID), check.Equals, expectedServerID, comment)
// make sure data is recorded properly
out := &bytes.Buffer{}
for _, e := range history {
out.WriteString(getChunk(e, 1000))
}
recorded := replaceNewlines(out.String())
c.Assert(recorded, check.Matches, ".*exit.*", comment)
c.Assert(recorded, check.Matches, ".*echo hi.*", comment)
// there should alwys be 'session.end' event
end := findByType(events.SessionEndEvent)
c.Assert(end, check.NotNil)
c.Assert(end.GetInt("bytes"), check.Equals, 0)
c.Assert(end.GetString(events.SessionEventID) != "", check.Equals, true)
// there should alwys be 'session.leave' event
leave := findByType(events.SessionLeaveEvent)
c.Assert(leave, check.NotNil, comment)
c.Assert(leave.GetInt("bytes"), check.Equals, 0, comment)
c.Assert(leave.GetString(events.SessionEventID) != "", check.Equals, true, comment)
// all of them should have a proper time:
for _, e := range history {
c.Assert(e.GetTime("time").IsZero(), check.Equals, false, comment)
}
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 38c9804738112eceedd209d9cb5bda7ffe3ab6f5
| 1.0 | gravitational/teleport: integration/integration_test.go; 281 LoC -
Found a possible issue in [gravitational/teleport](https://www.github.com/gravitational/teleport) at [integration/integration_test.go](https://github.com/gravitational/teleport/blob/38c9804738112eceedd209d9cb5bda7ffe3ab6f5/integration/integration_test.go#L299-L579)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message.
> range-loop variable tt used in defer or goroutine at line 378
[Click here to see the code in its original context.](https://github.com/gravitational/teleport/blob/38c9804738112eceedd209d9cb5bda7ffe3ab6f5/integration/integration_test.go#L299-L579)
<details>
<summary>Click here to show the 281 line(s) of Go which triggered the analyzer.</summary>
```go
for _, tt := range tests {
comment := check.Commentf(tt.comment)
makeConfig := func() (*check.C, []string, []*InstanceSecrets, *service.Config) {
clusterConfig, err := services.NewClusterConfig(services.ClusterConfigSpecV3{
SessionRecording: tt.inRecordLocation,
Audit: services.AuditConfig{AuditSessionsURI: tt.auditSessionsURI},
LocalAuth: services.NewBool(true),
})
c.Assert(err, check.IsNil, comment)
tconf := s.defaultServiceConfig()
tconf.Auth.Enabled = true
tconf.Auth.ClusterConfig = clusterConfig
tconf.Proxy.Enabled = true
tconf.Proxy.DisableWebService = true
tconf.Proxy.DisableWebInterface = true
tconf.SSH.Enabled = true
return c, nil, nil, tconf
}
t := s.newTeleportWithConfig(makeConfig())
defer t.StopAll()
// Start a node.
nodeSSHPort := s.getPorts(1)[0]
nodeConfig := func() *service.Config {
tconf := s.defaultServiceConfig()
tconf.HostUUID = "node"
tconf.Hostname = "node"
tconf.SSH.Enabled = true
tconf.SSH.Addr.Addr = net.JoinHostPort(t.Hostname, fmt.Sprintf("%v", nodeSSHPort))
return tconf
}
nodeProcess, err := t.StartNode(nodeConfig())
c.Assert(err, check.IsNil, comment)
// get access to a authClient for the cluster
site := t.GetSiteAPI(Site)
c.Assert(site, check.NotNil, comment)
// wait 10 seconds for both nodes to show up, otherwise
// we'll have trouble connecting to the node below.
waitForNodes := func(site auth.ClientI, count int) error {
tickCh := time.Tick(500 * time.Millisecond)
stopCh := time.After(10 * time.Second)
for {
select {
case <-tickCh:
nodesInSite, err := site.GetNodes(defaults.Namespace, services.SkipValidation())
if err != nil && !trace.IsNotFound(err) {
return trace.Wrap(err)
}
if got, want := len(nodesInSite), count; got == want {
return nil
}
case <-stopCh:
return trace.BadParameter("waited 10s, did find %v nodes", count)
}
}
}
err = waitForNodes(site, 2)
c.Assert(err, check.IsNil, comment)
// should have no sessions:
sessions, err := site.GetSessions(defaults.Namespace)
c.Assert(err, check.IsNil, comment)
c.Assert(len(sessions), check.Equals, 0, comment)
// create interactive session (this goroutine is this user's terminal time)
endC := make(chan error)
myTerm := NewTerminal(250)
go func() {
cl, err := t.NewClient(ClientConfig{
Login: s.me.Username,
Cluster: Site,
Host: Host,
Port: nodeSSHPort,
ForwardAgent: tt.inForwardAgent,
})
c.Assert(err, check.IsNil, comment)
cl.Stdout = myTerm
cl.Stdin = myTerm
err = cl.SSH(context.TODO(), []string{}, false)
endC <- err
}()
// wait until we've found the session in the audit log
getSession := func(site auth.ClientI) (*session.Session, error) {
tickCh := time.Tick(500 * time.Millisecond)
stopCh := time.After(10 * time.Second)
for {
select {
case <-tickCh:
sessions, err = site.GetSessions(defaults.Namespace)
if err != nil {
return nil, trace.Wrap(err)
}
if len(sessions) != 1 {
continue
}
return &sessions[0], nil
case <-stopCh:
return nil, trace.BadParameter("unable to find sessions after 10s (mode=%v)", tt.inRecordLocation)
}
}
}
session, err := getSession(site)
c.Assert(err, check.IsNil, comment)
// wait for the user to join this session:
for len(session.Parties) == 0 {
time.Sleep(time.Millisecond * 5)
session, err = site.GetSession(defaults.Namespace, sessions[0].ID)
c.Assert(err, check.IsNil, comment)
}
// make sure it's us who joined! :)
c.Assert(session.Parties[0].User, check.Equals, s.me.Username, comment)
// lets type "echo hi" followed by "enter" and then "exit" + "enter":
myTerm.Type("\aecho hi\n\r\aexit\n\r\a")
// wait for session to end:
select {
case <-endC:
case <-time.After(10 * time.Second):
c.Fatalf("%s: Timeout waiting for session to finish.", tt.comment)
}
// wait for the upload of the right session to complete
timeoutC := time.After(10 * time.Second)
loop:
for {
select {
case event := <-t.UploadEventsC:
if event.SessionID != string(session.ID) {
c.Logf("Skipping mismatching session %v, expecting upload of %v.", event.SessionID, session.ID)
continue
}
break loop
case <-timeoutC:
dumpGoroutineProfile()
c.Fatalf("%s: Timeout waiting for upload of session %v to complete to %v",
tt.comment, session.ID, tt.auditSessionsURI)
}
}
// read back the entire session (we have to try several times until we get back
// everything because the session is closing)
var sessionStream []byte
for i := 0; i < 6; i++ {
sessionStream, err = site.GetSessionChunk(defaults.Namespace, session.ID, 0, events.MaxChunkBytes)
c.Assert(err, check.IsNil, comment)
if strings.Contains(string(sessionStream), "exit") {
break
}
time.Sleep(time.Millisecond * 250)
if i >= 5 {
// session stream keeps coming back short
c.Fatalf("%s: Stream is not getting data: %q.", tt.comment, string(sessionStream))
}
}
// see what we got. It looks different based on bash settings, but here it is
// on Ev's machine (hostname is 'edsger'):
//
// edsger ~: echo hi
// hi
// edsger ~: exit
// logout
//
sessionComment := check.Commentf("%q", string(sessionStream))
c.Assert(strings.Contains(string(sessionStream), "echo hi"), check.Equals, true, sessionComment)
c.Assert(strings.Contains(string(sessionStream), "exit"), check.Equals, true, sessionComment)
// Wait until session.start, session.leave, and session.end events have arrived.
getSessions := func(site auth.ClientI) ([]events.EventFields, error) {
tickCh := time.Tick(500 * time.Millisecond)
stopCh := time.After(10 * time.Second)
for {
select {
case <-tickCh:
// Get all session events from the backend.
sessionEvents, err := site.GetSessionEvents(defaults.Namespace, session.ID, 0, false)
if err != nil {
return nil, trace.Wrap(err)
}
// Look through all session events for the three wanted.
var hasStart bool
var hasEnd bool
var hasLeave bool
for _, se := range sessionEvents {
if se.GetType() == events.SessionStartEvent {
hasStart = true
}
if se.GetType() == events.SessionEndEvent {
hasEnd = true
}
if se.GetType() == events.SessionLeaveEvent {
hasLeave = true
}
}
// Make sure all three events were found.
if hasStart && hasEnd && hasLeave {
return sessionEvents, nil
}
case <-stopCh:
return nil, trace.BadParameter("unable to find all session events after 10s (mode=%v)", tt.inRecordLocation)
}
}
}
history, err := getSessions(site)
c.Assert(err, check.IsNil, comment)
getChunk := func(e events.EventFields, maxlen int) string {
offset := e.GetInt("offset")
length := e.GetInt("bytes")
if length == 0 {
return ""
}
if length > maxlen {
length = maxlen
}
return string(sessionStream[offset : offset+length])
}
findByType := func(et string) events.EventFields {
for _, e := range history {
if e.GetType() == et {
return e
}
}
return nil
}
// there should alwys be 'session.start' event (and it must be first)
first := history[0]
start := findByType(events.SessionStartEvent)
c.Assert(start, check.DeepEquals, first, comment)
c.Assert(start.GetInt("bytes"), check.Equals, 0, comment)
c.Assert(start.GetString(events.SessionEventID) != "", check.Equals, true, comment)
c.Assert(start.GetString(events.TerminalSize) != "", check.Equals, true, comment)
// If session are being recorded at nodes, the SessionServerID should contain
// the ID of the node. If sessions are being recorded at the proxy, then
// SessionServerID should be that of the proxy.
expectedServerID := nodeProcess.Config.HostUUID
if services.IsRecordAtProxy(tt.inRecordLocation) {
expectedServerID = t.Process.Config.HostUUID
}
c.Assert(start.GetString(events.SessionServerID), check.Equals, expectedServerID, comment)
// make sure data is recorded properly
out := &bytes.Buffer{}
for _, e := range history {
out.WriteString(getChunk(e, 1000))
}
recorded := replaceNewlines(out.String())
c.Assert(recorded, check.Matches, ".*exit.*", comment)
c.Assert(recorded, check.Matches, ".*echo hi.*", comment)
// there should alwys be 'session.end' event
end := findByType(events.SessionEndEvent)
c.Assert(end, check.NotNil)
c.Assert(end.GetInt("bytes"), check.Equals, 0)
c.Assert(end.GetString(events.SessionEventID) != "", check.Equals, true)
// there should alwys be 'session.leave' event
leave := findByType(events.SessionLeaveEvent)
c.Assert(leave, check.NotNil, comment)
c.Assert(leave.GetInt("bytes"), check.Equals, 0, comment)
c.Assert(leave.GetString(events.SessionEventID) != "", check.Equals, true, comment)
// all of them should have a proper time:
for _, e := range history {
c.Assert(e.GetTime("time").IsZero(), check.Equals, false, comment)
}
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 38c9804738112eceedd209d9cb5bda7ffe3ab6f5
| test | gravitational teleport integration integration test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message range loop variable tt used in defer or goroutine at line click here to show the line s of go which triggered the analyzer go for tt range tests comment check commentf tt comment makeconfig func check c string instancesecrets service config clusterconfig err services newclusterconfig services sessionrecording tt inrecordlocation audit services auditconfig auditsessionsuri tt auditsessionsuri localauth services newbool true c assert err check isnil comment tconf s defaultserviceconfig tconf auth enabled true tconf auth clusterconfig clusterconfig tconf proxy enabled true tconf proxy disablewebservice true tconf proxy disablewebinterface true tconf ssh enabled true return c nil nil tconf t s newteleportwithconfig makeconfig defer t stopall start a node nodesshport s getports nodeconfig func service config tconf s defaultserviceconfig tconf hostuuid node tconf hostname node tconf ssh enabled true tconf ssh addr addr net joinhostport t hostname fmt sprintf v nodesshport return tconf nodeprocess err t startnode nodeconfig c assert err check isnil comment get access to a authclient for the cluster site t getsiteapi site c assert site check notnil comment wait seconds for both nodes to show up otherwise we ll have trouble connecting to the node below waitfornodes func site auth clienti count int error tickch time tick time millisecond stopch time after time second for select case tickch nodesinsite err site getnodes defaults namespace services skipvalidation if err nil trace isnotfound err return trace wrap err if got want len nodesinsite count got want return nil case stopch return trace badparameter waited did find v nodes count err waitfornodes site c assert err check isnil comment should have no sessions sessions err site getsessions defaults namespace c assert err check isnil comment c assert len sessions check equals comment create interactive session this goroutine is this user s terminal time endc make chan error myterm newterminal go func cl err t newclient clientconfig login s me username cluster site host host port nodesshport forwardagent tt inforwardagent c assert err check isnil comment cl stdout myterm cl stdin myterm err cl ssh context todo string false endc err wait until we ve found the session in the audit log getsession func site auth clienti session session error tickch time tick time millisecond stopch time after time second for select case tickch sessions err site getsessions defaults namespace if err nil return nil trace wrap err if len sessions continue return sessions nil case stopch return nil trace badparameter unable to find sessions after mode v tt inrecordlocation session err getsession site c assert err check isnil comment wait for the user to join this session for len session parties time sleep time millisecond session err site getsession defaults namespace sessions id c assert err check isnil comment make sure it s us who joined c assert session parties user check equals s me username comment lets type echo hi followed by enter and then exit enter myterm type aecho hi n r aexit n r a wait for session to end select case endc case time after time second c fatalf s timeout waiting for session to finish tt comment wait for the upload of the right session to complete timeoutc time after time second loop for select case event t uploadeventsc if event sessionid string session id c logf skipping mismatching session v expecting upload of v event sessionid session id continue break loop case timeoutc dumpgoroutineprofile c fatalf s timeout waiting for upload of session v to complete to v tt comment session id tt auditsessionsuri read back the entire session we have to try several times until we get back everything because the session is closing var sessionstream byte for i i i sessionstream err site getsessionchunk defaults namespace session id events maxchunkbytes c assert err check isnil comment if strings contains string sessionstream exit break time sleep time millisecond if i session stream keeps coming back short c fatalf s stream is not getting data q tt comment string sessionstream see what we got it looks different based on bash settings but here it is on ev s machine hostname is edsger edsger echo hi hi edsger exit logout sessioncomment check commentf q string sessionstream c assert strings contains string sessionstream echo hi check equals true sessioncomment c assert strings contains string sessionstream exit check equals true sessioncomment wait until session start session leave and session end events have arrived getsessions func site auth clienti events eventfields error tickch time tick time millisecond stopch time after time second for select case tickch get all session events from the backend sessionevents err site getsessionevents defaults namespace session id false if err nil return nil trace wrap err look through all session events for the three wanted var hasstart bool var hasend bool var hasleave bool for se range sessionevents if se gettype events sessionstartevent hasstart true if se gettype events sessionendevent hasend true if se gettype events sessionleaveevent hasleave true make sure all three events were found if hasstart hasend hasleave return sessionevents nil case stopch return nil trace badparameter unable to find all session events after mode v tt inrecordlocation history err getsessions site c assert err check isnil comment getchunk func e events eventfields maxlen int string offset e getint offset length e getint bytes if length return if length maxlen length maxlen return string sessionstream findbytype func et string events eventfields for e range history if e gettype et return e return nil there should alwys be session start event and it must be first first history start findbytype events sessionstartevent c assert start check deepequals first comment c assert start getint bytes check equals comment c assert start getstring events sessioneventid check equals true comment c assert start getstring events terminalsize check equals true comment if session are being recorded at nodes the sessionserverid should contain the id of the node if sessions are being recorded at the proxy then sessionserverid should be that of the proxy expectedserverid nodeprocess config hostuuid if services isrecordatproxy tt inrecordlocation expectedserverid t process config hostuuid c assert start getstring events sessionserverid check equals expectedserverid comment make sure data is recorded properly out bytes buffer for e range history out writestring getchunk e recorded replacenewlines out string c assert recorded check matches exit comment c assert recorded check matches echo hi comment there should alwys be session end event end findbytype events sessionendevent c assert end check notnil c assert end getint bytes check equals c assert end getstring events sessioneventid check equals true there should alwys be session leave event leave findbytype events sessionleaveevent c assert leave check notnil comment c assert leave getint bytes check equals comment c assert leave getstring events sessioneventid check equals true comment all of them should have a proper time for e range history c assert e gettime time iszero check equals false comment leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id | 1 |
29,280 | 8,316,494,766 | IssuesEvent | 2018-09-25 09:12:35 | openshiftio/openshift.io | https://api.github.com/repos/openshiftio/openshift.io | closed | Update osio-pipeline to process resources of same kind from different template file | area/pipelines status/in-progress team/build-cd type/task | Update osio-pipeline to process multiple resources of the same kind from multiple template file and can be passed to build and deploy API.
https://openshift.io/openshiftio/Openshift_io/plan/detail/673
| 1.0 | Update osio-pipeline to process resources of same kind from different template file - Update osio-pipeline to process multiple resources of the same kind from multiple template file and can be passed to build and deploy API.
https://openshift.io/openshiftio/Openshift_io/plan/detail/673
| non_test | update osio pipeline to process resources of same kind from different template file update osio pipeline to process multiple resources of the same kind from multiple template file and can be passed to build and deploy api | 0 |
339,411 | 30,446,197,789 | IssuesEvent | 2023-07-15 17:47:42 | natiatabatadzebtu/mid-term-versioning | https://api.github.com/repos/natiatabatadzebtu/mid-term-versioning | opened | 640c851 failed unit and formatting tests. | ci-pytest ci-black | Automatically generated message
640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests.
Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689442899/pytest.html
Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689442899/black.html
Automatically generated message
640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests.
Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689442927/pytest.html
Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689442927/black.html
Automatically generated message
640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests.
Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689442954/pytest.html
Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689442954/black.html
Automatically generated message
640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests.
Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689442981/pytest.html
Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689442981/black.html
Automatically generated message
640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests.
Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443009/pytest.html
Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443009/black.html
Automatically generated message
640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests.
Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443037/pytest.html
Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443037/black.html
Automatically generated message
640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests.
Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443064/pytest.html
Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443064/black.html
Automatically generated message
640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests.
Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443091/pytest.html
Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443091/black.html
Automatically generated message
640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests.
Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443119/pytest.html
Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443119/black.html
Automatically generated message
640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests.
Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443147/pytest.html
Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443147/black.html
Automatically generated message
640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests.
Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443175/pytest.html
Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443175/black.html
Automatically generated message
640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests.
Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443203/pytest.html
Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443203/black.html
Automatically generated message
640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests.
Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443230/pytest.html
Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443230/black.html
Automatically generated message
640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests.
Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443258/pytest.html
Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443258/black.html
| 1.0 | 640c851 failed unit and formatting tests. - Automatically generated message
640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests.
Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689442899/pytest.html
Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689442899/black.html
Automatically generated message
640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests.
Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689442927/pytest.html
Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689442927/black.html
Automatically generated message
640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests.
Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689442954/pytest.html
Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689442954/black.html
Automatically generated message
640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests.
Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689442981/pytest.html
Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689442981/black.html
Automatically generated message
640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests.
Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443009/pytest.html
Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443009/black.html
Automatically generated message
640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests.
Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443037/pytest.html
Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443037/black.html
Automatically generated message
640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests.
Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443064/pytest.html
Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443064/black.html
Automatically generated message
640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests.
Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443091/pytest.html
Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443091/black.html
Automatically generated message
640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests.
Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443119/pytest.html
Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443119/black.html
Automatically generated message
640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests.
Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443147/pytest.html
Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443147/black.html
Automatically generated message
640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests.
Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443175/pytest.html
Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443175/black.html
Automatically generated message
640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests.
Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443203/pytest.html
Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443203/black.html
Automatically generated message
640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests.
Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443230/pytest.html
Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443230/black.html
Automatically generated message
640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests.
Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443258/pytest.html
Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443258/black.html
| test | failed unit and formatting tests automatically generated message failed unit and formatting tests pytest report black report automatically generated message failed unit and formatting tests pytest report black report automatically generated message failed unit and formatting tests pytest report black report automatically generated message failed unit and formatting tests pytest report black report automatically generated message failed unit and formatting tests pytest report black report automatically generated message failed unit and formatting tests pytest report black report automatically generated message failed unit and formatting tests pytest report black report automatically generated message failed unit and formatting tests pytest report black report automatically generated message failed unit and formatting tests pytest report black report automatically generated message failed unit and formatting tests pytest report black report automatically generated message failed unit and formatting tests pytest report black report automatically generated message failed unit and formatting tests pytest report black report automatically generated message failed unit and formatting tests pytest report black report automatically generated message failed unit and formatting tests pytest report black report | 1 |
50,391 | 6,086,442,927 | IssuesEvent | 2017-06-18 01:05:02 | Semantic-Org/Semantic-UI | https://api.github.com/repos/Semantic-Org/Semantic-UI | closed | Popup keeps appear in 'bottom center' even if i set it's position as 'top center' | Needs Test Case Usage Question | So, what is this about? I set properly `position: 'top center'`, but popup appears in bottom center position without any errors and warnings thrown. | 1.0 | Popup keeps appear in 'bottom center' even if i set it's position as 'top center' - So, what is this about? I set properly `position: 'top center'`, but popup appears in bottom center position without any errors and warnings thrown. | test | popup keeps appear in bottom center even if i set it s position as top center so what is this about i set properly position top center but popup appears in bottom center position without any errors and warnings thrown | 1 |
492,596 | 14,216,432,889 | IssuesEvent | 2020-11-17 08:59:14 | redhat-developer/vscode-openshift-tools | https://api.github.com/repos/redhat-developer/vscode-openshift-tools | closed | Create Service command support for creating services backed by operators | kind/feature priority/major resolution/duplicate upstream/odo | The command should support creating services based from service catalogs and based on operators using odo options explained in https://github.com/openshift/odo/issues/2613.
https://odo.dev/docs/operator-hub/
Component to service should work when Service Binding Operator (SBO) installed. | 1.0 | Create Service command support for creating services backed by operators - The command should support creating services based from service catalogs and based on operators using odo options explained in https://github.com/openshift/odo/issues/2613.
https://odo.dev/docs/operator-hub/
Component to service should work when Service Binding Operator (SBO) installed. | non_test | create service command support for creating services backed by operators the command should support creating services based from service catalogs and based on operators using odo options explained in component to service should work when service binding operator sbo installed | 0 |
290,544 | 21,884,484,542 | IssuesEvent | 2022-05-19 17:10:48 | a-little-org-called-mario/a-little-game-called-mario | https://api.github.com/repos/a-little-org-called-mario/a-little-game-called-mario | closed | Collision Layer Diligence and Proposal | documentation stale | A lot of the content that is being added, and that contains collision boxes, are all set to layer 1 mask 1.
This will start to become a major issue moving forward. I already ran into a problem trying with commit #252
Here is what I propose:
* We do a major PR that changes all current assets to no longer use layer 1 or mask 1.
* Create a wiki page that explains the reasoning behind this choice and what they need to do to fix it.
* (If this is possible) We add a check on commits that checks for `collision_mask` and `collision_layer` in `*.tscn` files to see if it is an odd number.
* These numbers can only be odd if layer 1 or mask 1 is set. All the other options are even numbers.
* If odd numbers are found, tell the contributor and link to the document explaining what needs to be done. | 1.0 | Collision Layer Diligence and Proposal - A lot of the content that is being added, and that contains collision boxes, are all set to layer 1 mask 1.
This will start to become a major issue moving forward. I already ran into a problem trying with commit #252
Here is what I propose:
* We do a major PR that changes all current assets to no longer use layer 1 or mask 1.
* Create a wiki page that explains the reasoning behind this choice and what they need to do to fix it.
* (If this is possible) We add a check on commits that checks for `collision_mask` and `collision_layer` in `*.tscn` files to see if it is an odd number.
* These numbers can only be odd if layer 1 or mask 1 is set. All the other options are even numbers.
* If odd numbers are found, tell the contributor and link to the document explaining what needs to be done. | non_test | collision layer diligence and proposal a lot of the content that is being added and that contains collision boxes are all set to layer mask this will start to become a major issue moving forward i already ran into a problem trying with commit here is what i propose we do a major pr that changes all current assets to no longer use layer or mask create a wiki page that explains the reasoning behind this choice and what they need to do to fix it if this is possible we add a check on commits that checks for collision mask and collision layer in tscn files to see if it is an odd number these numbers can only be odd if layer or mask is set all the other options are even numbers if odd numbers are found tell the contributor and link to the document explaining what needs to be done | 0 |
159,127 | 12,464,095,497 | IssuesEvent | 2020-05-28 11:51:37 | dasch-swiss/knora-app | https://api.github.com/repos/dasch-swiss/knora-app | closed | Homepage : external links are not easy to spot | styling user-testing | **Describe the bug**
I find the link to external websites are not very easy to spot, unless you mouse over and look at the bottom left corner of the screen where the URL where the image or text redirects is displayed.
**To Reproduce Steps to reproduce the behavior:**
1. Go to '...'
2. Click on '....'
3. Scroll down to '....'
4. See error
**OPTIONAL: Expected behavior**
Maybe underline when a text that is an hyperlink?
**Screenshots**
If applicable, add screenshots to help explain your problem (drag-and-drop the image directly here).
**Desktop (please complete the following information):**
- OS: macOS 10.14.6 (18G103), French
- Browser Firefox
- Version 70.01.1
**Additional context**
Add any other context about the problem here.
| 1.0 | Homepage : external links are not easy to spot - **Describe the bug**
I find the link to external websites are not very easy to spot, unless you mouse over and look at the bottom left corner of the screen where the URL where the image or text redirects is displayed.
**To Reproduce Steps to reproduce the behavior:**
1. Go to '...'
2. Click on '....'
3. Scroll down to '....'
4. See error
**OPTIONAL: Expected behavior**
Maybe underline when a text that is an hyperlink?
**Screenshots**
If applicable, add screenshots to help explain your problem (drag-and-drop the image directly here).
**Desktop (please complete the following information):**
- OS: macOS 10.14.6 (18G103), French
- Browser Firefox
- Version 70.01.1
**Additional context**
Add any other context about the problem here.
| test | homepage external links are not easy to spot describe the bug i find the link to external websites are not very easy to spot unless you mouse over and look at the bottom left corner of the screen where the url where the image or text redirects is displayed to reproduce steps to reproduce the behavior go to click on scroll down to see error optional expected behavior maybe underline when a text that is an hyperlink screenshots if applicable add screenshots to help explain your problem drag and drop the image directly here desktop please complete the following information os macos french browser firefox version additional context add any other context about the problem here | 1 |
161,222 | 12,534,179,299 | IssuesEvent | 2020-06-04 18:57:05 | astropy/astropy | https://api.github.com/repos/astropy/astropy | closed | tox -e codestyle fails on astropy/_erfa/core.py | Bug Docs testing | <!-- This comments are hidden when you submit the issue,
so you do not need to remove them! -->
<!-- Please be sure to check out our contributing guidelines,
https://github.com/astropy/astropy/blob/master/CONTRIBUTING.md .
Please be sure to check out our code of conduct,
https://github.com/astropy/astropy/blob/master/CODE_OF_CONDUCT.md . -->
<!-- Please have a search on our GitHub repository to see if a similar
issue has already been posted.
If a similar issue is closed, have a quick look to see if you are satisfied
by the resolution.
If not please go ahead and open an issue! -->
<!-- Please check that the development version still produces the same bug.
You can install development version with
pip install git+https://github.com/astropy/astropy
command. -->
### Description
<!-- Provide a general description of the bug. -->
`tox -e codestyle` always fails with a huge stream of code style violations in `astropy/_erfa/core.py`. This renders it largely useless for checking my own code for style violations.
I assume the file is auto-generated, so simply telling the job to ignore code style violations in that file might solve the problem. If not, a once-over with `black` or `autopep8` would probably resolve the issue.
### Expected behavior
<!-- What did you expect to happen. -->
`tox -e codestyle` succeeds on a clean master.
### Actual behavior
<!-- What actually happened. -->
<!-- Was the output confusing or poorly described? -->
```
...
astropy/_erfa/core.py:20058:37: W291 trailing whitespace
astropy/_erfa/core.py:20101:37: W291 trailing whitespace
astropy/_erfa/core.py:20146:37: W291 trailing whitespace
astropy/_erfa/core.py:20187:37: W291 trailing whitespace
astropy/_erfa/core.py:20231:37: W291 trailing whitespace
astropy/_erfa/core.py:20272:37: W291 trailing whitespace
astropy/_erfa/core.py:20342:22: W292 no newline at end of file
244
ERROR: InvocationError for command /home/archibald/projects/astropy/astropy/.tox/codestyle/bin/flake8 astropy --count --select=E101,W191,W291,W292,W293,W391,E111,E112,E113,E30,E502,E722,E901,E902,E999,F822,F823 (exited with code 1)
_______________________________________________________ summary _______________________________________________________
ERROR: codestyle: commands failed
```
### Steps to Reproduce
<!-- Ideally a code example could be provided so we can run it ourselves. -->
<!-- If you are pasting code, use triple backticks (```) around
your code snippet. -->
<!-- If necessary, sanitize your screen output to be pasted so you do not
reveal secrets like tokens and passwords. -->
```sh
tox -e codestyle
```
### System Details
<!-- Even if you do not think this is necessary, it is useful information for the maintainers.
Please run the following snippet and paste the output below:
import platform; print(platform.platform())
import sys; print("Python", sys.version)
import numpy; print("Numpy", numpy.__version__)
import astropy; print("astropy", astropy.__version__)
import scipy; print("Scipy", scipy.__version__)
import matplotlib; print("Matplotlib", matplotlib.__version__)
-->
```
>>> import platform; print(platform.platform())
Linux-5.4.0-33-generic-x86_64-with-glibc2.29
>>> import sys; print("Python", sys.version)
Python 3.8.2 (default, Apr 27 2020, 15:53:34)
[GCC 9.3.0]
>>> import numpy; print("Numpy", numpy.__version__)
Numpy 1.18.4
>>> import astropy; print("astropy", astropy.__version__)
astropy 4.2.dev162+gd0a1e469f
>>> import scipy; print("Scipy", scipy.__version__)
Scipy 1.4.1
>>> import matplotlib; print("Matplotlib", matplotlib.__version__)
Matplotlib 3.2.1
``` | 1.0 | tox -e codestyle fails on astropy/_erfa/core.py - <!-- This comments are hidden when you submit the issue,
so you do not need to remove them! -->
<!-- Please be sure to check out our contributing guidelines,
https://github.com/astropy/astropy/blob/master/CONTRIBUTING.md .
Please be sure to check out our code of conduct,
https://github.com/astropy/astropy/blob/master/CODE_OF_CONDUCT.md . -->
<!-- Please have a search on our GitHub repository to see if a similar
issue has already been posted.
If a similar issue is closed, have a quick look to see if you are satisfied
by the resolution.
If not please go ahead and open an issue! -->
<!-- Please check that the development version still produces the same bug.
You can install development version with
pip install git+https://github.com/astropy/astropy
command. -->
### Description
<!-- Provide a general description of the bug. -->
`tox -e codestyle` always fails with a huge stream of code style violations in `astropy/_erfa/core.py`. This renders it largely useless for checking my own code for style violations.
I assume the file is auto-generated, so simply telling the job to ignore code style violations in that file might solve the problem. If not, a once-over with `black` or `autopep8` would probably resolve the issue.
### Expected behavior
<!-- What did you expect to happen. -->
`tox -e codestyle` succeeds on a clean master.
### Actual behavior
<!-- What actually happened. -->
<!-- Was the output confusing or poorly described? -->
```
...
astropy/_erfa/core.py:20058:37: W291 trailing whitespace
astropy/_erfa/core.py:20101:37: W291 trailing whitespace
astropy/_erfa/core.py:20146:37: W291 trailing whitespace
astropy/_erfa/core.py:20187:37: W291 trailing whitespace
astropy/_erfa/core.py:20231:37: W291 trailing whitespace
astropy/_erfa/core.py:20272:37: W291 trailing whitespace
astropy/_erfa/core.py:20342:22: W292 no newline at end of file
244
ERROR: InvocationError for command /home/archibald/projects/astropy/astropy/.tox/codestyle/bin/flake8 astropy --count --select=E101,W191,W291,W292,W293,W391,E111,E112,E113,E30,E502,E722,E901,E902,E999,F822,F823 (exited with code 1)
_______________________________________________________ summary _______________________________________________________
ERROR: codestyle: commands failed
```
### Steps to Reproduce
<!-- Ideally a code example could be provided so we can run it ourselves. -->
<!-- If you are pasting code, use triple backticks (```) around
your code snippet. -->
<!-- If necessary, sanitize your screen output to be pasted so you do not
reveal secrets like tokens and passwords. -->
```sh
tox -e codestyle
```
### System Details
<!-- Even if you do not think this is necessary, it is useful information for the maintainers.
Please run the following snippet and paste the output below:
import platform; print(platform.platform())
import sys; print("Python", sys.version)
import numpy; print("Numpy", numpy.__version__)
import astropy; print("astropy", astropy.__version__)
import scipy; print("Scipy", scipy.__version__)
import matplotlib; print("Matplotlib", matplotlib.__version__)
-->
```
>>> import platform; print(platform.platform())
Linux-5.4.0-33-generic-x86_64-with-glibc2.29
>>> import sys; print("Python", sys.version)
Python 3.8.2 (default, Apr 27 2020, 15:53:34)
[GCC 9.3.0]
>>> import numpy; print("Numpy", numpy.__version__)
Numpy 1.18.4
>>> import astropy; print("astropy", astropy.__version__)
astropy 4.2.dev162+gd0a1e469f
>>> import scipy; print("Scipy", scipy.__version__)
Scipy 1.4.1
>>> import matplotlib; print("Matplotlib", matplotlib.__version__)
Matplotlib 3.2.1
``` | test | tox e codestyle fails on astropy erfa core py this comments are hidden when you submit the issue so you do not need to remove them please be sure to check out our contributing guidelines please be sure to check out our code of conduct please have a search on our github repository to see if a similar issue has already been posted if a similar issue is closed have a quick look to see if you are satisfied by the resolution if not please go ahead and open an issue please check that the development version still produces the same bug you can install development version with pip install git command description tox e codestyle always fails with a huge stream of code style violations in astropy erfa core py this renders it largely useless for checking my own code for style violations i assume the file is auto generated so simply telling the job to ignore code style violations in that file might solve the problem if not a once over with black or would probably resolve the issue expected behavior tox e codestyle succeeds on a clean master actual behavior astropy erfa core py trailing whitespace astropy erfa core py trailing whitespace astropy erfa core py trailing whitespace astropy erfa core py trailing whitespace astropy erfa core py trailing whitespace astropy erfa core py trailing whitespace astropy erfa core py no newline at end of file error invocationerror for command home archibald projects astropy astropy tox codestyle bin astropy count select exited with code summary error codestyle commands failed steps to reproduce if you are pasting code use triple backticks around your code snippet if necessary sanitize your screen output to be pasted so you do not reveal secrets like tokens and passwords sh tox e codestyle system details even if you do not think this is necessary it is useful information for the maintainers please run the following snippet and paste the output below import platform print platform platform import sys print python sys version import numpy print numpy numpy version import astropy print astropy astropy version import scipy print scipy scipy version import matplotlib print matplotlib matplotlib version import platform print platform platform linux generic with import sys print python sys version python default apr import numpy print numpy numpy version numpy import astropy print astropy astropy version astropy import scipy print scipy scipy version scipy import matplotlib print matplotlib matplotlib version matplotlib | 1 |
183,425 | 14,230,121,603 | IssuesEvent | 2020-11-18 07:33:15 | remirror/remirror | https://api.github.com/repos/remirror/remirror | opened | Reactivate e2e tests for SSR | status: experimental :test_tube: test: e2e :heavy_check_mark: type: feature :sparkles: | ### Description
Some issues with server-side rendering (**SSR**) have been raised in third-party codebases.
This issue should track progress on reactivating end-to-end tests for SSR to ensure that all extensions contained within this repository work as expected. | 2.0 | Reactivate e2e tests for SSR - ### Description
Some issues with server-side rendering (**SSR**) have been raised in third-party codebases.
This issue should track progress on reactivating end-to-end tests for SSR to ensure that all extensions contained within this repository work as expected. | test | reactivate tests for ssr description some issues with server side rendering ssr have been raised in third party codebases this issue should track progress on reactivating end to end tests for ssr to ensure that all extensions contained within this repository work as expected | 1 |
725,804 | 24,976,359,539 | IssuesEvent | 2022-11-02 08:11:04 | Lightning-AI/lightning | https://api.github.com/repos/Lightning-AI/lightning | closed | Testing custom work dependencies example fails due to timeout in CI | ci priority: 1 app | ### Bug description
One of our app example tests fails due to timeout.
- CI file: https://github.com/Lightning-AI/lightning/blob/c39c8eb2e4f49e35ef38a2ae6de2765ec71623db/.azure/app-cloud-e2e.yml#L66
- CI run: https://dev.azure.com/Lightning-AI/lightning/_build/results?buildId=111050&view=logs&j=4e1b52d3-134d-51bf-b10a-6dde7facebae&t=60ace6bb-5e20-5573-9db7-5550ec53d684&l=1463
### How to reproduce the bug
_No response_
### Error messages and logs
```
...
flow 10/30/2022 17:05:43 DEBUG: The url http://10.1.113.254:47157/ is not ready.
flow 10/30/2022 17:05:43 DEBUG: The url http://10.1.113.254:47157/ is not ready.
flow 10/30/2022 17:05:43 DEBUG: The url http://10.1.113.254:47157/ is not ready.
(node:446) MaxListenersExceededWarning: Possible EventEmitter memory leak detected. 11 end listeners added to [ReadStream]. Use emitter.setMaxListeners() to increase limit
(Use `node --trace-warnings ...` to show where the warning was created)
(node:446) MaxListenersExceededWarning: Possible EventEmitter memory leak detected. 11 error listeners added to [ReadStream]. Use emitter.setMaxListeners() to increase limit
/bin/sh: 1: ***: not found
FAILED
=================================== FAILURES ===================================
_________________ test_custom_work_dependencies_example_cloud __________________
@pytest.mark.cloud
def test_custom_work_dependencies_example_cloud() -> None:
# if requirements not installed, the app will fail
with run_app_in_cloud(
os.path.join(_PROJECT_ROOT, "tests/tests_app_examples/custom_work_dependencies/"),
app_name="app.py",
) as (_, _, fetch_logs, _):
has_logs = False
while not has_logs:
> for log in fetch_logs(["flow"]):
tests/tests_app_examples/test_custom_work_dependencies.py:19:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
src/***/app/testing/testing.py:440: in fetch_logs
for log_event in gen:
src/***/app/utilities/app_logs.py:96: in _app_logs_reader
log_event: _LogEvent = read_queue.get(timeout=None if follow else 1.0)
/usr/lib/python3.8/queue.py:179: in get
self.not_empty.wait(remaining)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <Condition(<unlocked _thread.lock object at 0x7fe4bb27f300>, 0)>
timeout = 0.9999995000000581
def wait(self, timeout=None):
"""Wait until notified or until a timeout occurs.
If the calling thread has not acquired the lock when this method is
called, a RuntimeError is raised.
This method releases the underlying lock, and then blocks until it is
awakened by a notify() or notify_all() call for the same condition
variable in another thread, or until the optional timeout occurs. Once
awakened or timed out, it re-acquires the lock and returns.
When the timeout argument is present and not None, it should be a
floating point number specifying a timeout for the operation in seconds
(or fractions thereof).
When the underlying lock is an RLock, it is not released using its
release() method, since this may not actually unlock the lock when it
was acquired multiple times recursively. Instead, an internal interface
of the RLock class is used, which really unlocks it even when it has
been recursively acquired several times. Another internal interface is
then used to restore the recursion level when the lock is reacquired.
"""
if not self._is_owned():
raise RuntimeError("cannot wait on un-acquired lock")
waiter = _allocate_lock()
waiter.acquire()
self._waiters.append(waiter)
saved_state = self._release_save()
gotit = False
try: # restore state no matter what (e.g., KeyboardInterrupt)
if timeout is None:
waiter.acquire()
gotit = True
else:
if timeout > 0:
> gotit = waiter.acquire(True, timeout)
E Failed: Timeout >1200.0s
/usr/lib/python3.8/threading.py:306: Failed
```
```
=========================== short test summary info ============================
FAILED tests/tests_app_examples/test_custom_work_dependencies.py::test_custom_work_dependencies_example_cloud
======================== 1 failed in 1373.06s (0:22:53) ========================
+++++++++++++++++++++++++++++++++++ Timeout ++++++++++++++++++++++++++++++++++++
~~~~~~~~~~~~~~~~~~~~ Stack of Thread-241 (140620377880320) ~~~~~~~~~~~~~~~~~~~~~
File "/usr/lib/python3.8/threading.py", line 890, in _bootstrap
self._bootstrap_inner()
File "/usr/lib/python3.8/threading.py", line 932, in _bootstrap_inner
self.run()
File "/usr/lib/python3.8/threading.py", line 870, in run
self._target(*self._args, **self._kwargs)
File "/home/AzDevOps_azpcontainer/.local/lib/python3.8/site-packages/websocket/_app.py", line 449, in run_forever
setSock()
File "/home/AzDevOps_azpcontainer/.local/lib/python3.8/site-packages/websocket/_app.py", line 378, in setSock
dispatcher.read(self.sock.sock, read, check)
File "/home/AzDevOps_azpcontainer/.local/lib/python3.8/site-packages/websocket/_app.py", line 72, in read
r = sel.select(self.ping_timeout)
File "/usr/lib/python3.8/selectors.py", line 468, in select
fd_event_list = self._selector.poll(timeout, max_ev)
~~~~~~~~~~~~~~~~~~~~~ Stack of waitpid-0 (140621050640128) ~~~~~~~~~~~~~~~~~~~~~
File "/usr/lib/python3.8/threading.py", line 890, in _bootstrap
self._bootstrap_inner()
File "/usr/lib/python3.8/threading.py", line 932, in _bootstrap_inner
self.run()
File "/usr/lib/python3.8/threading.py", line 870, in run
self._target(*self._args, **self._kwargs)
File "/usr/lib/python3.8/asyncio/unix_events.py", line 1318, in _do_waitpid
pid, status = os.waitpid(expected_pid, 0)
+++++++++++++++++++++++++++++++++++ Timeout ++++++++++++++++++++++++++++++++++++
##[error]Bash exited with code '1'.
```
### Environment
_No response_
### More info
_No response_
cc @tchaton @rohitgr7 @carmocca @akihironitta @borda | 1.0 | Testing custom work dependencies example fails due to timeout in CI - ### Bug description
One of our app example tests fails due to timeout.
- CI file: https://github.com/Lightning-AI/lightning/blob/c39c8eb2e4f49e35ef38a2ae6de2765ec71623db/.azure/app-cloud-e2e.yml#L66
- CI run: https://dev.azure.com/Lightning-AI/lightning/_build/results?buildId=111050&view=logs&j=4e1b52d3-134d-51bf-b10a-6dde7facebae&t=60ace6bb-5e20-5573-9db7-5550ec53d684&l=1463
### How to reproduce the bug
_No response_
### Error messages and logs
```
...
flow 10/30/2022 17:05:43 DEBUG: The url http://10.1.113.254:47157/ is not ready.
flow 10/30/2022 17:05:43 DEBUG: The url http://10.1.113.254:47157/ is not ready.
flow 10/30/2022 17:05:43 DEBUG: The url http://10.1.113.254:47157/ is not ready.
(node:446) MaxListenersExceededWarning: Possible EventEmitter memory leak detected. 11 end listeners added to [ReadStream]. Use emitter.setMaxListeners() to increase limit
(Use `node --trace-warnings ...` to show where the warning was created)
(node:446) MaxListenersExceededWarning: Possible EventEmitter memory leak detected. 11 error listeners added to [ReadStream]. Use emitter.setMaxListeners() to increase limit
/bin/sh: 1: ***: not found
FAILED
=================================== FAILURES ===================================
_________________ test_custom_work_dependencies_example_cloud __________________
@pytest.mark.cloud
def test_custom_work_dependencies_example_cloud() -> None:
# if requirements not installed, the app will fail
with run_app_in_cloud(
os.path.join(_PROJECT_ROOT, "tests/tests_app_examples/custom_work_dependencies/"),
app_name="app.py",
) as (_, _, fetch_logs, _):
has_logs = False
while not has_logs:
> for log in fetch_logs(["flow"]):
tests/tests_app_examples/test_custom_work_dependencies.py:19:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
src/***/app/testing/testing.py:440: in fetch_logs
for log_event in gen:
src/***/app/utilities/app_logs.py:96: in _app_logs_reader
log_event: _LogEvent = read_queue.get(timeout=None if follow else 1.0)
/usr/lib/python3.8/queue.py:179: in get
self.not_empty.wait(remaining)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <Condition(<unlocked _thread.lock object at 0x7fe4bb27f300>, 0)>
timeout = 0.9999995000000581
def wait(self, timeout=None):
"""Wait until notified or until a timeout occurs.
If the calling thread has not acquired the lock when this method is
called, a RuntimeError is raised.
This method releases the underlying lock, and then blocks until it is
awakened by a notify() or notify_all() call for the same condition
variable in another thread, or until the optional timeout occurs. Once
awakened or timed out, it re-acquires the lock and returns.
When the timeout argument is present and not None, it should be a
floating point number specifying a timeout for the operation in seconds
(or fractions thereof).
When the underlying lock is an RLock, it is not released using its
release() method, since this may not actually unlock the lock when it
was acquired multiple times recursively. Instead, an internal interface
of the RLock class is used, which really unlocks it even when it has
been recursively acquired several times. Another internal interface is
then used to restore the recursion level when the lock is reacquired.
"""
if not self._is_owned():
raise RuntimeError("cannot wait on un-acquired lock")
waiter = _allocate_lock()
waiter.acquire()
self._waiters.append(waiter)
saved_state = self._release_save()
gotit = False
try: # restore state no matter what (e.g., KeyboardInterrupt)
if timeout is None:
waiter.acquire()
gotit = True
else:
if timeout > 0:
> gotit = waiter.acquire(True, timeout)
E Failed: Timeout >1200.0s
/usr/lib/python3.8/threading.py:306: Failed
```
```
=========================== short test summary info ============================
FAILED tests/tests_app_examples/test_custom_work_dependencies.py::test_custom_work_dependencies_example_cloud
======================== 1 failed in 1373.06s (0:22:53) ========================
+++++++++++++++++++++++++++++++++++ Timeout ++++++++++++++++++++++++++++++++++++
~~~~~~~~~~~~~~~~~~~~ Stack of Thread-241 (140620377880320) ~~~~~~~~~~~~~~~~~~~~~
File "/usr/lib/python3.8/threading.py", line 890, in _bootstrap
self._bootstrap_inner()
File "/usr/lib/python3.8/threading.py", line 932, in _bootstrap_inner
self.run()
File "/usr/lib/python3.8/threading.py", line 870, in run
self._target(*self._args, **self._kwargs)
File "/home/AzDevOps_azpcontainer/.local/lib/python3.8/site-packages/websocket/_app.py", line 449, in run_forever
setSock()
File "/home/AzDevOps_azpcontainer/.local/lib/python3.8/site-packages/websocket/_app.py", line 378, in setSock
dispatcher.read(self.sock.sock, read, check)
File "/home/AzDevOps_azpcontainer/.local/lib/python3.8/site-packages/websocket/_app.py", line 72, in read
r = sel.select(self.ping_timeout)
File "/usr/lib/python3.8/selectors.py", line 468, in select
fd_event_list = self._selector.poll(timeout, max_ev)
~~~~~~~~~~~~~~~~~~~~~ Stack of waitpid-0 (140621050640128) ~~~~~~~~~~~~~~~~~~~~~
File "/usr/lib/python3.8/threading.py", line 890, in _bootstrap
self._bootstrap_inner()
File "/usr/lib/python3.8/threading.py", line 932, in _bootstrap_inner
self.run()
File "/usr/lib/python3.8/threading.py", line 870, in run
self._target(*self._args, **self._kwargs)
File "/usr/lib/python3.8/asyncio/unix_events.py", line 1318, in _do_waitpid
pid, status = os.waitpid(expected_pid, 0)
+++++++++++++++++++++++++++++++++++ Timeout ++++++++++++++++++++++++++++++++++++
##[error]Bash exited with code '1'.
```
### Environment
_No response_
### More info
_No response_
cc @tchaton @rohitgr7 @carmocca @akihironitta @borda | non_test | testing custom work dependencies example fails due to timeout in ci bug description one of our app example tests fails due to timeout ci file ci run how to reproduce the bug no response error messages and logs flow debug the url is not ready flow debug the url is not ready flow debug the url is not ready node maxlistenersexceededwarning possible eventemitter memory leak detected end listeners added to use emitter setmaxlisteners to increase limit use node trace warnings to show where the warning was created node maxlistenersexceededwarning possible eventemitter memory leak detected error listeners added to use emitter setmaxlisteners to increase limit bin sh not found failed failures test custom work dependencies example cloud pytest mark cloud def test custom work dependencies example cloud none if requirements not installed the app will fail with run app in cloud os path join project root tests tests app examples custom work dependencies app name app py as fetch logs has logs false while not has logs for log in fetch logs tests tests app examples test custom work dependencies py src app testing testing py in fetch logs for log event in gen src app utilities app logs py in app logs reader log event logevent read queue get timeout none if follow else usr lib queue py in get self not empty wait remaining self timeout def wait self timeout none wait until notified or until a timeout occurs if the calling thread has not acquired the lock when this method is called a runtimeerror is raised this method releases the underlying lock and then blocks until it is awakened by a notify or notify all call for the same condition variable in another thread or until the optional timeout occurs once awakened or timed out it re acquires the lock and returns when the timeout argument is present and not none it should be a floating point number specifying a timeout for the operation in seconds or fractions thereof when the underlying lock is an rlock it is not released using its release method since this may not actually unlock the lock when it was acquired multiple times recursively instead an internal interface of the rlock class is used which really unlocks it even when it has been recursively acquired several times another internal interface is then used to restore the recursion level when the lock is reacquired if not self is owned raise runtimeerror cannot wait on un acquired lock waiter allocate lock waiter acquire self waiters append waiter saved state self release save gotit false try restore state no matter what e g keyboardinterrupt if timeout is none waiter acquire gotit true else if timeout gotit waiter acquire true timeout e failed timeout usr lib threading py failed short test summary info failed tests tests app examples test custom work dependencies py test custom work dependencies example cloud failed in timeout stack of thread file usr lib threading py line in bootstrap self bootstrap inner file usr lib threading py line in bootstrap inner self run file usr lib threading py line in run self target self args self kwargs file home azdevops azpcontainer local lib site packages websocket app py line in run forever setsock file home azdevops azpcontainer local lib site packages websocket app py line in setsock dispatcher read self sock sock read check file home azdevops azpcontainer local lib site packages websocket app py line in read r sel select self ping timeout file usr lib selectors py line in select fd event list self selector poll timeout max ev stack of waitpid file usr lib threading py line in bootstrap self bootstrap inner file usr lib threading py line in bootstrap inner self run file usr lib threading py line in run self target self args self kwargs file usr lib asyncio unix events py line in do waitpid pid status os waitpid expected pid timeout bash exited with code environment no response more info no response cc tchaton carmocca akihironitta borda | 0 |
102,875 | 8,868,648,076 | IssuesEvent | 2019-01-11 00:59:19 | rancher/rancher | https://api.github.com/repos/rancher/rancher | closed | AD - Group changes not reflected in user attributes even after auth-user-info-max-age-seconds passed or forcing rancher login using cli | area/authentication status/reopened status/resolved status/to-test team/az version/2.0 | Original issue: https://github.com/rancherlabs/rancher-security/issues/40
Original text:
Rancher server version - build from master
Steps to reproduce the problem:
AD auth enabled setup
Create a cluster with 3 projects:
p1 - member access given to "group1"
p2 - member access given to "group2"
p3 - member access given to "group3"
Log in to rancher UI as "user1" who is part of "group1" and "group2"
Create API keys.
Set "auth-user-info-max-age-seconds" to 120 seconds.
Use the above created keys to log in using rancher-cli.
User is provided access to both projects p1 and p2 as expected (./rancher project)
From AD console, remove p1 from being part of "group2".
Wait for more than > "auth-user-info-max-age-seconds" to pass.
scenario 1:
In the current logged in rancher-cli session , execute ```./rancher project``` .
Both projects p1 and p2 are listed.
scenario 2:
Log into rancher-cli using the same token , execute ```./rancher project``` .
Both projects p1 and p2 are listed.
In this case , following error logs seen in rancher-server logs:
```
2018/12/14 01:22:48 [DEBUG] REST UPDATE apis/management.cattle.io/v3//userattributes/u-66zjeognlk
2018/12/14 01:22:48 [DEBUG] UserAttributeController calling handler mgmt-auth-userattributes-controller u-66zjeognlk
2018/12/14 01:22:48 [DEBUG] REST GET apis/management.cattle.io/v3//authconfigs/activedirectory
2018/12/14 01:22:48 [DEBUG] REST GET api//v1/mgmt-secrets/secrets/activedirectoryconfig-serviceaccountpassword
2018/12/14 01:22:48 [DEBUG] Now creating Ldap connection
2018/12/14 01:22:48 [DEBUG] Binding service account username password
2018/12/14 01:22:48 [DEBUG] LDAP Search query: {(sAMAccountName=activedirectory_user://CN=testguser1,CN=Users,DC=testad,DC=rancher,DC=io)}
2018/12/14 01:22:48 [ERROR] Error refreshing token principals, skipping: Cannot locate user information for (sAMAccountName=activedirectory_user://CN=testguser1,CN=Users,DC=testad,DC=rancher,DC=io)
2018/12/14 01:22:48 [DEBUG] REST UPDATE apis/management.cattle.io/v3//userattributes/u-66zjeognlk
2018/12/14 01:22:48 [DEBUG] UserAttributeController calling handler mgmt-auth-userattributes-controller u-66zjeognlk
```
In both scenario1 and scenario2 , we expect user to have access to only project p1 since they are not part of "group2" anymore.
Also the user-attributes does not reflect the group change (```kubectl get user attribute <user-id> -o json```)
Note:
When "user1" logs in using the UI , I see the user-attributes reflect the group changes as expected.
Further log ins done from rancher-cli using the older token shows only project p1.
| 1.0 | AD - Group changes not reflected in user attributes even after auth-user-info-max-age-seconds passed or forcing rancher login using cli - Original issue: https://github.com/rancherlabs/rancher-security/issues/40
Original text:
Rancher server version - build from master
Steps to reproduce the problem:
AD auth enabled setup
Create a cluster with 3 projects:
p1 - member access given to "group1"
p2 - member access given to "group2"
p3 - member access given to "group3"
Log in to rancher UI as "user1" who is part of "group1" and "group2"
Create API keys.
Set "auth-user-info-max-age-seconds" to 120 seconds.
Use the above created keys to log in using rancher-cli.
User is provided access to both projects p1 and p2 as expected (./rancher project)
From AD console, remove p1 from being part of "group2".
Wait for more than > "auth-user-info-max-age-seconds" to pass.
scenario 1:
In the current logged in rancher-cli session , execute ```./rancher project``` .
Both projects p1 and p2 are listed.
scenario 2:
Log into rancher-cli using the same token , execute ```./rancher project``` .
Both projects p1 and p2 are listed.
In this case , following error logs seen in rancher-server logs:
```
2018/12/14 01:22:48 [DEBUG] REST UPDATE apis/management.cattle.io/v3//userattributes/u-66zjeognlk
2018/12/14 01:22:48 [DEBUG] UserAttributeController calling handler mgmt-auth-userattributes-controller u-66zjeognlk
2018/12/14 01:22:48 [DEBUG] REST GET apis/management.cattle.io/v3//authconfigs/activedirectory
2018/12/14 01:22:48 [DEBUG] REST GET api//v1/mgmt-secrets/secrets/activedirectoryconfig-serviceaccountpassword
2018/12/14 01:22:48 [DEBUG] Now creating Ldap connection
2018/12/14 01:22:48 [DEBUG] Binding service account username password
2018/12/14 01:22:48 [DEBUG] LDAP Search query: {(sAMAccountName=activedirectory_user://CN=testguser1,CN=Users,DC=testad,DC=rancher,DC=io)}
2018/12/14 01:22:48 [ERROR] Error refreshing token principals, skipping: Cannot locate user information for (sAMAccountName=activedirectory_user://CN=testguser1,CN=Users,DC=testad,DC=rancher,DC=io)
2018/12/14 01:22:48 [DEBUG] REST UPDATE apis/management.cattle.io/v3//userattributes/u-66zjeognlk
2018/12/14 01:22:48 [DEBUG] UserAttributeController calling handler mgmt-auth-userattributes-controller u-66zjeognlk
```
In both scenario1 and scenario2 , we expect user to have access to only project p1 since they are not part of "group2" anymore.
Also the user-attributes does not reflect the group change (```kubectl get user attribute <user-id> -o json```)
Note:
When "user1" logs in using the UI , I see the user-attributes reflect the group changes as expected.
Further log ins done from rancher-cli using the older token shows only project p1.
| test | ad group changes not reflected in user attributes even after auth user info max age seconds passed or forcing rancher login using cli original issue original text rancher server version build from master steps to reproduce the problem ad auth enabled setup create a cluster with projects member access given to member access given to member access given to log in to rancher ui as who is part of and create api keys set auth user info max age seconds to seconds use the above created keys to log in using rancher cli user is provided access to both projects and as expected rancher project from ad console remove from being part of wait for more than auth user info max age seconds to pass scenario in the current logged in rancher cli session execute rancher project both projects and are listed scenario log into rancher cli using the same token execute rancher project both projects and are listed in this case following error logs seen in rancher server logs rest update apis management cattle io userattributes u userattributecontroller calling handler mgmt auth userattributes controller u rest get apis management cattle io authconfigs activedirectory rest get api mgmt secrets secrets activedirectoryconfig serviceaccountpassword now creating ldap connection binding service account username password ldap search query samaccountname activedirectory user cn cn users dc testad dc rancher dc io error refreshing token principals skipping cannot locate user information for samaccountname activedirectory user cn cn users dc testad dc rancher dc io rest update apis management cattle io userattributes u userattributecontroller calling handler mgmt auth userattributes controller u in both and we expect user to have access to only project since they are not part of anymore also the user attributes does not reflect the group change kubectl get user attribute o json note when logs in using the ui i see the user attributes reflect the group changes as expected further log ins done from rancher cli using the older token shows only project | 1 |
345,688 | 30,833,304,188 | IssuesEvent | 2023-08-02 04:49:35 | teamhanko/hanko | https://api.github.com/repos/teamhanko/hanko | closed | Make e2e test check for component ids instead of button texts etc. | javascript tests stale | Changing e.g. a button text in `hanko-auth` currently breaks e2e tests because the tests are based on strings ("see if button with text 'Sign in with a passkey" appears").
Wouldn't it be better to set ids for all relevant components in the element and make the tests check for the ids? | 1.0 | Make e2e test check for component ids instead of button texts etc. - Changing e.g. a button text in `hanko-auth` currently breaks e2e tests because the tests are based on strings ("see if button with text 'Sign in with a passkey" appears").
Wouldn't it be better to set ids for all relevant components in the element and make the tests check for the ids? | test | make test check for component ids instead of button texts etc changing e g a button text in hanko auth currently breaks tests because the tests are based on strings see if button with text sign in with a passkey appears wouldn t it be better to set ids for all relevant components in the element and make the tests check for the ids | 1 |
222,736 | 17,470,661,078 | IssuesEvent | 2021-08-07 04:17:27 | dotnet/aspnetcore | https://api.github.com/repos/dotnet/aspnetcore | opened | Microsoft.AspNetCore.Hosting.Tests.HostingApplicationTests.IHttpActivityFeatureIsPopulated | test-failure | ## Failing Test(s)
- Microsoft.AspNetCore.Hosting.Tests.HostingApplicationTests.IHttpActivityFeatureIsPopulated
## Error Message
<!--
Provide the error message associated with the test failure, if applicable.
-->
```text
Assert.NotEqual() Failure
Expected: Not Activity { ActivityTraceFlags = None, Baggage = [], Context = System.Diagnostics.ActivityContext, DisplayName = "Microsoft.AspNetCore.Hosting.HttpRequestIn", Duration = 00:00:00, ... }
Actual: Activity { ActivityTraceFlags = None, Baggage = [], Context = System.Diagnostics.ActivityContext, DisplayName = "Microsoft.AspNetCore.Hosting.HttpRequestIn", Duration = 00:00:00, ... }
```
## Stacktrace
<details>
<!--
Provide the stack trace associated with the test failure, if applicable.
-->
```text
at Microsoft.AspNetCore.Hosting.Tests.HostingApplicationTests.IHttpActivityFeatureIsPopulated() in /_/src/Hosting/Hosting/test/HostingApplicationTests.cs:line 122
```
</details>
## Logs
<details>
<!--
Provide the (helix) logs associated with the test failure, if applicable.
-->
```text
[xUnit.net 00:00:01.21] Microsoft.AspNetCore.Hosting.Tests.HostingApplicationTests.IHttpActivityFeatureIsPopulated [FAIL]
[xUnit.net 00:00:01.21] Assert.NotEqual() Failure
[xUnit.net 00:00:01.21] Expected: Not Activity { ActivityTraceFlags = None, Baggage = [], Context = System.Diagnostics.ActivityContext, DisplayName = "Microsoft.AspNetCore.Hosting.HttpRequestIn", Duration = 00:00:00, ... }
[xUnit.net 00:00:01.21] Actual: Activity { ActivityTraceFlags = None, Baggage = [], Context = System.Diagnostics.ActivityContext, DisplayName = "Microsoft.AspNetCore.Hosting.HttpRequestIn", Duration = 00:00:00, ... }
[xUnit.net 00:00:01.21] Stack Trace:
[xUnit.net 00:00:01.21] /_/src/Hosting/Hosting/test/HostingApplicationTests.cs(122,0): at Microsoft.AspNetCore.Hosting.Tests.HostingApplicationTests.IHttpActivityFeatureIsPopulated()
```
</details>
## Build
<!--
Provide a link to the build where the test failure occurred.
-->
https://dev.azure.com/dnceng/public/_build/results?buildId=1280595&view=ms.vss-test-web.build-test-results-tab&runId=37844528&resultId=103322&paneView=debug | 1.0 | Microsoft.AspNetCore.Hosting.Tests.HostingApplicationTests.IHttpActivityFeatureIsPopulated - ## Failing Test(s)
- Microsoft.AspNetCore.Hosting.Tests.HostingApplicationTests.IHttpActivityFeatureIsPopulated
## Error Message
<!--
Provide the error message associated with the test failure, if applicable.
-->
```text
Assert.NotEqual() Failure
Expected: Not Activity { ActivityTraceFlags = None, Baggage = [], Context = System.Diagnostics.ActivityContext, DisplayName = "Microsoft.AspNetCore.Hosting.HttpRequestIn", Duration = 00:00:00, ... }
Actual: Activity { ActivityTraceFlags = None, Baggage = [], Context = System.Diagnostics.ActivityContext, DisplayName = "Microsoft.AspNetCore.Hosting.HttpRequestIn", Duration = 00:00:00, ... }
```
## Stacktrace
<details>
<!--
Provide the stack trace associated with the test failure, if applicable.
-->
```text
at Microsoft.AspNetCore.Hosting.Tests.HostingApplicationTests.IHttpActivityFeatureIsPopulated() in /_/src/Hosting/Hosting/test/HostingApplicationTests.cs:line 122
```
</details>
## Logs
<details>
<!--
Provide the (helix) logs associated with the test failure, if applicable.
-->
```text
[xUnit.net 00:00:01.21] Microsoft.AspNetCore.Hosting.Tests.HostingApplicationTests.IHttpActivityFeatureIsPopulated [FAIL]
[xUnit.net 00:00:01.21] Assert.NotEqual() Failure
[xUnit.net 00:00:01.21] Expected: Not Activity { ActivityTraceFlags = None, Baggage = [], Context = System.Diagnostics.ActivityContext, DisplayName = "Microsoft.AspNetCore.Hosting.HttpRequestIn", Duration = 00:00:00, ... }
[xUnit.net 00:00:01.21] Actual: Activity { ActivityTraceFlags = None, Baggage = [], Context = System.Diagnostics.ActivityContext, DisplayName = "Microsoft.AspNetCore.Hosting.HttpRequestIn", Duration = 00:00:00, ... }
[xUnit.net 00:00:01.21] Stack Trace:
[xUnit.net 00:00:01.21] /_/src/Hosting/Hosting/test/HostingApplicationTests.cs(122,0): at Microsoft.AspNetCore.Hosting.Tests.HostingApplicationTests.IHttpActivityFeatureIsPopulated()
```
</details>
## Build
<!--
Provide a link to the build where the test failure occurred.
-->
https://dev.azure.com/dnceng/public/_build/results?buildId=1280595&view=ms.vss-test-web.build-test-results-tab&runId=37844528&resultId=103322&paneView=debug | test | microsoft aspnetcore hosting tests hostingapplicationtests ihttpactivityfeatureispopulated failing test s microsoft aspnetcore hosting tests hostingapplicationtests ihttpactivityfeatureispopulated error message provide the error message associated with the test failure if applicable text assert notequal failure expected not activity activitytraceflags none baggage context system diagnostics activitycontext displayname microsoft aspnetcore hosting httprequestin duration actual activity activitytraceflags none baggage context system diagnostics activitycontext displayname microsoft aspnetcore hosting httprequestin duration stacktrace provide the stack trace associated with the test failure if applicable text at microsoft aspnetcore hosting tests hostingapplicationtests ihttpactivityfeatureispopulated in src hosting hosting test hostingapplicationtests cs line logs provide the helix logs associated with the test failure if applicable text microsoft aspnetcore hosting tests hostingapplicationtests ihttpactivityfeatureispopulated assert notequal failure expected not activity activitytraceflags none baggage context system diagnostics activitycontext displayname microsoft aspnetcore hosting httprequestin duration actual activity activitytraceflags none baggage context system diagnostics activitycontext displayname microsoft aspnetcore hosting httprequestin duration stack trace src hosting hosting test hostingapplicationtests cs at microsoft aspnetcore hosting tests hostingapplicationtests ihttpactivityfeatureispopulated build provide a link to the build where the test failure occurred | 1 |
81,089 | 23,386,144,926 | IssuesEvent | 2022-08-11 13:54:54 | xamarin/xamarin-android | https://api.github.com/repos/xamarin/xamarin-android | opened | BroadcastReceiver with targetSdkVersion="31" causes java.exe exited with code 1 | Area: App+Library Build needs-triage | ### Android application type
Classic Xamarin.Android (MonoAndroid12.0, etc.)
### Affected platform version
VS 2022
### Description
Placing this class in a Xamarin.Android app that targets sdk 31 (Android 12) and compiles with Android 12 causes a build error
java.exe exited with code 1
` [BroadcastReceiver]
[IntentFilter(new[] { Android.Content.Intent.ActionBatteryChanged })]
public class MyReceiver : BroadcastReceiver
{
public override void OnReceive(Context context, Intent intent)
{
}
}
`
Including example project
[XamApp1.zip](https://github.com/xamarin/xamarin-android/files/9309213/XamApp1.zip)
### Steps to Reproduce
1. Create standard Xamarin.Android app using VS2022
2. Add the Broadcast reciever
3. Make sure your android:targetSdkVersion="31" in AndroidManifest.xml
4. Make sure the Compile Using Android Version: (Target Framework) setting is set to Android 12 (S)
5. Compile Debug->AnyCPU
### Did you find any workaround?
Target Android 11 or below
### Relevant log output
_No response_ | 1.0 | BroadcastReceiver with targetSdkVersion="31" causes java.exe exited with code 1 - ### Android application type
Classic Xamarin.Android (MonoAndroid12.0, etc.)
### Affected platform version
VS 2022
### Description
Placing this class in a Xamarin.Android app that targets sdk 31 (Android 12) and compiles with Android 12 causes a build error
java.exe exited with code 1
` [BroadcastReceiver]
[IntentFilter(new[] { Android.Content.Intent.ActionBatteryChanged })]
public class MyReceiver : BroadcastReceiver
{
public override void OnReceive(Context context, Intent intent)
{
}
}
`
Including example project
[XamApp1.zip](https://github.com/xamarin/xamarin-android/files/9309213/XamApp1.zip)
### Steps to Reproduce
1. Create standard Xamarin.Android app using VS2022
2. Add the Broadcast reciever
3. Make sure your android:targetSdkVersion="31" in AndroidManifest.xml
4. Make sure the Compile Using Android Version: (Target Framework) setting is set to Android 12 (S)
5. Compile Debug->AnyCPU
### Did you find any workaround?
Target Android 11 or below
### Relevant log output
_No response_ | non_test | broadcastreceiver with targetsdkversion causes java exe exited with code android application type classic xamarin android etc affected platform version vs description placing this class in a xamarin android app that targets sdk android and compiles with android causes a build error java exe exited with code android content intent actionbatterychanged public class myreceiver broadcastreceiver public override void onreceive context context intent intent including example project steps to reproduce create standard xamarin android app using add the broadcast reciever make sure your android targetsdkversion in androidmanifest xml make sure the compile using android version target framework setting is set to android s compile debug anycpu did you find any workaround target android or below relevant log output no response | 0 |
177,042 | 13,675,858,965 | IssuesEvent | 2020-09-29 13:13:24 | brave/brave-browser | https://api.github.com/repos/brave/brave-browser | closed | [Desktop] Sync v2 - Apps are not automatically installed | OS/Desktop QA/Test-Plan-Specified QA/Yes feature/sync priority/P4 | App is synced but the App is not installed
According to @darkdh this is an upstream bug
## Steps to Reproduce
<!--Please add a series of steps to reproduce the issue-->
1. Clean install on `DeviceA` and `DeviceB`
1. Establish sync chain between `DeviceA` and `DeviceB`
1. Enable Apps sync on `DeviceA` and `DeviceB`
1. `DeviceA`: Open interia.pl
1. `DeviceA`: Click `Install Interia` in the URL field 
1. Wait for sync update
1. `DeviceB`: Open interia.pl
## Actual result:
<!--Please add screenshots if needed-->
`DeviceB`: App is synced

`DeviceB` : App is not installed

## Expected result:
`DeviceB` : App is installed

## Reproduces how often:
<!--[Easily reproduced/Intermittent issue/No steps to reproduce]-->
100% repro rate
## Brave version (brave://version info)
<!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details-->
Brave | 1.12.108 Chromium: 84.0.4147.105 (Official Build) (64-bit)
-- | --
Revision | a6b12dfad6663f13a7e16e9a42a6a4975374096b-refs/branch-heads/4147@{#943}
OS | Ubuntu 18.04 LTS
cc @brave/legacy_qa @darkdh @jsecretan @AlexeyBarabash @rebron | 1.0 | [Desktop] Sync v2 - Apps are not automatically installed - App is synced but the App is not installed
According to @darkdh this is an upstream bug
## Steps to Reproduce
<!--Please add a series of steps to reproduce the issue-->
1. Clean install on `DeviceA` and `DeviceB`
1. Establish sync chain between `DeviceA` and `DeviceB`
1. Enable Apps sync on `DeviceA` and `DeviceB`
1. `DeviceA`: Open interia.pl
1. `DeviceA`: Click `Install Interia` in the URL field 
1. Wait for sync update
1. `DeviceB`: Open interia.pl
## Actual result:
<!--Please add screenshots if needed-->
`DeviceB`: App is synced

`DeviceB` : App is not installed

## Expected result:
`DeviceB` : App is installed

## Reproduces how often:
<!--[Easily reproduced/Intermittent issue/No steps to reproduce]-->
100% repro rate
## Brave version (brave://version info)
<!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details-->
Brave | 1.12.108 Chromium: 84.0.4147.105 (Official Build) (64-bit)
-- | --
Revision | a6b12dfad6663f13a7e16e9a42a6a4975374096b-refs/branch-heads/4147@{#943}
OS | Ubuntu 18.04 LTS
cc @brave/legacy_qa @darkdh @jsecretan @AlexeyBarabash @rebron | test | sync apps are not automatically installed app is synced but the app is not installed according to darkdh this is an upstream bug steps to reproduce clean install on devicea and deviceb establish sync chain between devicea and deviceb enable apps sync on devicea and deviceb devicea open interia pl devicea click install interia in the url field wait for sync update deviceb open interia pl actual result deviceb app is synced deviceb app is not installed expected result deviceb app is installed reproduces how often repro rate brave version brave version info brave chromium official build bit revision refs branch heads os ubuntu lts cc brave legacy qa darkdh jsecretan alexeybarabash rebron | 1 |
147,274 | 23,192,583,755 | IssuesEvent | 2022-08-01 13:50:32 | team-ribbon/ribborn-frontend | https://api.github.com/repos/team-ribbon/ribborn-frontend | closed | [공유] 공유하기 위치 조정이 필요합니다 | design | <img src="https://user-images.githubusercontent.com/105181604/181536396-92705f5c-5f60-40c5-bd3c-505351fd9854.jpg" />
PC > 공유하기 클릭 시 좌우 스크롤 생기는 문제
모바일 > 공유하기 클릭하면 플로팅 버튼 위치 이동되는 문제 => 모바일에서는 공유하기 누르면 좌우로가 아니라, 위아래로 뜨면 좋을 것 같습니다. | 1.0 | [공유] 공유하기 위치 조정이 필요합니다 - <img src="https://user-images.githubusercontent.com/105181604/181536396-92705f5c-5f60-40c5-bd3c-505351fd9854.jpg" />
PC > 공유하기 클릭 시 좌우 스크롤 생기는 문제
모바일 > 공유하기 클릭하면 플로팅 버튼 위치 이동되는 문제 => 모바일에서는 공유하기 누르면 좌우로가 아니라, 위아래로 뜨면 좋을 것 같습니다. | non_test | 공유하기 위치 조정이 필요합니다 pc 공유하기 클릭 시 좌우 스크롤 생기는 문제 모바일 공유하기 클릭하면 플로팅 버튼 위치 이동되는 문제 모바일에서는 공유하기 누르면 좌우로가 아니라 위아래로 뜨면 좋을 것 같습니다 | 0 |
221,986 | 17,380,614,979 | IssuesEvent | 2021-07-31 16:26:59 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | opened | roachtest: clearrange/zfs/checks=true failed | C-test-failure O-roachtest O-robot branch-master release-blocker | roachtest.clearrange/zfs/checks=true [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=3246869&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=3246869&tab=artifacts#/clearrange/zfs/checks=true) on master @ [701b177d8f4b81d8654dfb4090a2cd3cf82e63a7](https://github.com/cockroachdb/cockroach/commits/701b177d8f4b81d8654dfb4090a2cd3cf82e63a7):
```
The test failed on branch=master, cloud=gce:
test timed out (see artifacts for details)
```
<details><summary>Reproduce</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/roachtest)
</p>
</details>
/cc @cockroachdb/storage
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*clearrange/zfs/checks=true.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
| 2.0 | roachtest: clearrange/zfs/checks=true failed - roachtest.clearrange/zfs/checks=true [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=3246869&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=3246869&tab=artifacts#/clearrange/zfs/checks=true) on master @ [701b177d8f4b81d8654dfb4090a2cd3cf82e63a7](https://github.com/cockroachdb/cockroach/commits/701b177d8f4b81d8654dfb4090a2cd3cf82e63a7):
```
The test failed on branch=master, cloud=gce:
test timed out (see artifacts for details)
```
<details><summary>Reproduce</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/roachtest)
</p>
</details>
/cc @cockroachdb/storage
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*clearrange/zfs/checks=true.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
| test | roachtest clearrange zfs checks true failed roachtest clearrange zfs checks true with on master the test failed on branch master cloud gce test timed out see artifacts for details reproduce see cc cockroachdb storage | 1 |
18,372 | 10,948,170,205 | IssuesEvent | 2019-11-26 08:17:08 | microsoft/botbuilder-dotnet | https://api.github.com/repos/microsoft/botbuilder-dotnet | closed | TurnContext.SendActivitiesAsync changes the activities' ID | Bot Services R7 bug customer-replied-to customer-reported | ## Version
master
## Describe the bug
On [this line](https://github.com/microsoft/botbuilder-dotnet/blob/master/libraries/Microsoft.Bot.Builder/TurnContext.cs#L274), we can see where the turn context modifies the ID of each outgoing activity:
```c#
activity.Id = responses[index].Id;
```
While outgoing activities are necessarily modified when a conversation reference is applied, modifying the ID based on the resource response creates strange behavior when the activity is saved and used again, like when it's a retry prompt. The Emulator for example will not display an activity if the ID is the same as an activity already in the conversation history. Several customers are running into issues because of this:
https://github.com/microsoft/BotFramework-Emulator/issues/1684
https://stackoverflow.com/questions/58082537/why-does-the-retry-prompt-only-show-once
The turn context in the Node SDK does not modify an outgoing activity's ID.
## Expected behavior
There is reason to think that simply sending an activity shouldn't modify it at all, even by applying the conversation reference. That is to say, the necessary modifications should be done on a *copy* of the activity that gets passed to `SendActivityAsync` for example. However, changing this now could be breaking behavior if some bots expect the activity to be modified. Also, an easy workaround is to send an activity through the adapter or the connector, so bot developers have a choice about whether they want outgoing activities to be modified automatically.
The ID in particular may be a special case for a few reasons. Not only is this modification causing the issues we see in Emulator, but the behavior is inconsistent between repos. Even if we don't want to change this ID-modifying behavior in the .NET SDK, we should still make sure the different SDK's work the same way.
[bug]
| 1.0 | TurnContext.SendActivitiesAsync changes the activities' ID - ## Version
master
## Describe the bug
On [this line](https://github.com/microsoft/botbuilder-dotnet/blob/master/libraries/Microsoft.Bot.Builder/TurnContext.cs#L274), we can see where the turn context modifies the ID of each outgoing activity:
```c#
activity.Id = responses[index].Id;
```
While outgoing activities are necessarily modified when a conversation reference is applied, modifying the ID based on the resource response creates strange behavior when the activity is saved and used again, like when it's a retry prompt. The Emulator for example will not display an activity if the ID is the same as an activity already in the conversation history. Several customers are running into issues because of this:
https://github.com/microsoft/BotFramework-Emulator/issues/1684
https://stackoverflow.com/questions/58082537/why-does-the-retry-prompt-only-show-once
The turn context in the Node SDK does not modify an outgoing activity's ID.
## Expected behavior
There is reason to think that simply sending an activity shouldn't modify it at all, even by applying the conversation reference. That is to say, the necessary modifications should be done on a *copy* of the activity that gets passed to `SendActivityAsync` for example. However, changing this now could be breaking behavior if some bots expect the activity to be modified. Also, an easy workaround is to send an activity through the adapter or the connector, so bot developers have a choice about whether they want outgoing activities to be modified automatically.
The ID in particular may be a special case for a few reasons. Not only is this modification causing the issues we see in Emulator, but the behavior is inconsistent between repos. Even if we don't want to change this ID-modifying behavior in the .NET SDK, we should still make sure the different SDK's work the same way.
[bug]
| non_test | turncontext sendactivitiesasync changes the activities id version master describe the bug on we can see where the turn context modifies the id of each outgoing activity c activity id responses id while outgoing activities are necessarily modified when a conversation reference is applied modifying the id based on the resource response creates strange behavior when the activity is saved and used again like when it s a retry prompt the emulator for example will not display an activity if the id is the same as an activity already in the conversation history several customers are running into issues because of this the turn context in the node sdk does not modify an outgoing activity s id expected behavior there is reason to think that simply sending an activity shouldn t modify it at all even by applying the conversation reference that is to say the necessary modifications should be done on a copy of the activity that gets passed to sendactivityasync for example however changing this now could be breaking behavior if some bots expect the activity to be modified also an easy workaround is to send an activity through the adapter or the connector so bot developers have a choice about whether they want outgoing activities to be modified automatically the id in particular may be a special case for a few reasons not only is this modification causing the issues we see in emulator but the behavior is inconsistent between repos even if we don t want to change this id modifying behavior in the net sdk we should still make sure the different sdk s work the same way | 0 |
68,538 | 21,699,680,193 | IssuesEvent | 2022-05-10 01:39:59 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | closed | "your settings" in the search initialisation fail message must be styled as a link | T-Defect S-Tolerable Help Wanted A-Timeline-Search A-Indexing O-Occasional good first issue | 
"your settings" must be styled as a link. Right now it's highlighten only on mouse over.
That link should open the user setting, jump to the "Security & Privacy" tab and scroll down to the "Message search" section
element-desktop 1.7.33 debian sid | 1.0 | "your settings" in the search initialisation fail message must be styled as a link - 
"your settings" must be styled as a link. Right now it's highlighten only on mouse over.
That link should open the user setting, jump to the "Security & Privacy" tab and scroll down to the "Message search" section
element-desktop 1.7.33 debian sid | non_test | your settings in the search initialisation fail message must be styled as a link your settings must be styled as a link right now it s highlighten only on mouse over that link should open the user setting jump to the security privacy tab and scroll down to the message search section element desktop debian sid | 0 |
229,382 | 18,294,436,303 | IssuesEvent | 2021-10-05 18:53:31 | elastic/kibana | https://api.github.com/repos/elastic/kibana | opened | Failing test: Jest Tests.x-pack/plugins/fleet/storybook - Storyshots Sections/EPM/Package List Grid List | failed-test | A test failed on a tracked branch
```
Error: expect(received).not.toContain(expected) // indexOf
Expected substring: not "euiErrorBoundary"
Received string: "<div class=\"kbnRedirectCrossAppLinks\"><div class=\"euiErrorBoundary\"><div class=\"euiErrorBoundary__text\"><div class=\"euiText euiText--extraSmall\"><h1>Error</h1><pre class=\"euiErrorBoundary__stack\"><p>Error: Cannot read property 'startsWith' of undefined
at PackageCard (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/x-pack/plugins/fleet/public/applications/integrations/sections/epm/components/package_card.tsx:55:19)
at renderWithHooks (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:16260:18)
at mountIndeterminateComponent (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:18794:13)
at beginWork$1 (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:20162:16)
at HTMLUnknownElement.callCallback (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:336:14)
at HTMLUnknownElement.callTheUserObjectsOperation (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-environment-jsdom/node_modules/jsdom/lib/jsdom/living/generated/EventListener.js:26:30)
at innerInvokeEventListeners (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-environment-jsdom/node_modules/jsdom/lib/jsdom/living/events/EventTarget-impl.js:318:25)
at invokeEventListeners (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-environment-jsdom/node_modules/jsdom/lib/jsdom/living/events/EventTarget-impl.js:274:3)
at HTMLUnknownElementImpl._dispatch (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-environment-jsdom/node_modules/jsdom/lib/jsdom/living/events/EventTarget-impl.js:221:9)
at HTMLUnknownElementImpl.dispatchEvent (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-environment-jsdom/node_modules/jsdom/lib/jsdom/living/events/EventTarget-impl.js:94:17)
at HTMLUnknownElement.dispatchEvent (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-environment-jsdom/node_modules/jsdom/lib/jsdom/living/generated/EventTarget.js:231:34)
at Object.invokeGuardedCallbackDev (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:385:16)
at invokeGuardedCallback (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:440:31)
at beginWork$$1 (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:25780:7)
at performUnitOfWork (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:24698:12)
at workLoopSync (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:24671:22)
at performSyncWorkOnRoot (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:24270:11)
at /opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:12199:24
at unstable_runWithPriority (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/scheduler/cjs/scheduler.development.js:697:12)
at runWithPriority$2 (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:12149:10)
at flushSyncCallbackQueueImpl (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:12194:7)
at flushSyncCallbackQueue (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:12182:3)
at batchedUpdates$1 (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:24392:7)
at Object.act (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom-test-utils.development.js:1092:14)
at wrapAct (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/enzyme-adapter-react-16/src/ReactSixteenAdapter.js:405:13)
at Object.render (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/enzyme-adapter-react-16/src/ReactSixteenAdapter.js:474:16)
at new ReactWrapper (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/enzyme/src/ReactWrapper.js:115:16)
at mount (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/enzyme/src/mount.js:10:10)
at test (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/x-pack/plugins/fleet/storybook/smoke.test.tsx:17:22)
at Object.<anonymous> (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/@storybook/addon-storyshots/dist/api/snapshotsTestsTemplate.js:42:20)
at Promise.then.completed (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/utils.js:276:28)
at new Promise (<anonymous>)
at callAsyncCircusFn (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/utils.js:216:10)
at _callCircusTest (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:212:40)
at _runTest (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:149:3)
at _runTestsForDescribeBlock (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:63:9)
at _runTestsForDescribeBlock (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:57:9)
at _runTestsForDescribeBlock (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:57:9)
at run (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:25:3)
at runAndTransformResultsToJestFormat (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/legacy-code-todo-rewrite/jestAdapterInit.js:176:21)
at jestAdapter (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/legacy-code-todo-rewrite/jestAdapter.js:109:19)
at runTestInternal (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-runner/build/runTest.js:380:16)
at runTest (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-runner/build/runTest.js:472:34)
at Object.worker (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-runner/build/testWorker.js:133:12)</p></pre></div></div></div></div>"
at test (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/x-pack/plugins/fleet/storybook/smoke.test.tsx:20:33)
at _callCircusTest (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:212:5)
at _runTest (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:149:3)
at _runTestsForDescribeBlock (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:63:9)
at _runTestsForDescribeBlock (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:57:9)
at _runTestsForDescribeBlock (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:57:9)
at run (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:25:3)
at runAndTransformResultsToJestFormat (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/legacy-code-todo-rewrite/jestAdapterInit.js:176:21)
at jestAdapter (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/legacy-code-todo-rewrite/jestAdapter.js:109:19)
at runTestInternal (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-runner/build/runTest.js:380:16)
at runTest (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-runner/build/runTest.js:472:34)
at Object.worker (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-runner/build/testWorker.js:133:12)
```
First failure: [CI Build - master](https://buildkite.com/elastic/kibana-hourly/builds/946#ff42e543-5004-441b-887b-0e66b9018892)
<!-- kibanaCiData = {"failed-test":{"test.class":"Jest Tests.x-pack/plugins/fleet/storybook","test.name":"Storyshots Sections/EPM/Package List Grid List","test.failCount":1}} --> | 1.0 | Failing test: Jest Tests.x-pack/plugins/fleet/storybook - Storyshots Sections/EPM/Package List Grid List - A test failed on a tracked branch
```
Error: expect(received).not.toContain(expected) // indexOf
Expected substring: not "euiErrorBoundary"
Received string: "<div class=\"kbnRedirectCrossAppLinks\"><div class=\"euiErrorBoundary\"><div class=\"euiErrorBoundary__text\"><div class=\"euiText euiText--extraSmall\"><h1>Error</h1><pre class=\"euiErrorBoundary__stack\"><p>Error: Cannot read property 'startsWith' of undefined
at PackageCard (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/x-pack/plugins/fleet/public/applications/integrations/sections/epm/components/package_card.tsx:55:19)
at renderWithHooks (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:16260:18)
at mountIndeterminateComponent (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:18794:13)
at beginWork$1 (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:20162:16)
at HTMLUnknownElement.callCallback (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:336:14)
at HTMLUnknownElement.callTheUserObjectsOperation (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-environment-jsdom/node_modules/jsdom/lib/jsdom/living/generated/EventListener.js:26:30)
at innerInvokeEventListeners (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-environment-jsdom/node_modules/jsdom/lib/jsdom/living/events/EventTarget-impl.js:318:25)
at invokeEventListeners (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-environment-jsdom/node_modules/jsdom/lib/jsdom/living/events/EventTarget-impl.js:274:3)
at HTMLUnknownElementImpl._dispatch (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-environment-jsdom/node_modules/jsdom/lib/jsdom/living/events/EventTarget-impl.js:221:9)
at HTMLUnknownElementImpl.dispatchEvent (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-environment-jsdom/node_modules/jsdom/lib/jsdom/living/events/EventTarget-impl.js:94:17)
at HTMLUnknownElement.dispatchEvent (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-environment-jsdom/node_modules/jsdom/lib/jsdom/living/generated/EventTarget.js:231:34)
at Object.invokeGuardedCallbackDev (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:385:16)
at invokeGuardedCallback (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:440:31)
at beginWork$$1 (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:25780:7)
at performUnitOfWork (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:24698:12)
at workLoopSync (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:24671:22)
at performSyncWorkOnRoot (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:24270:11)
at /opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:12199:24
at unstable_runWithPriority (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/scheduler/cjs/scheduler.development.js:697:12)
at runWithPriority$2 (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:12149:10)
at flushSyncCallbackQueueImpl (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:12194:7)
at flushSyncCallbackQueue (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:12182:3)
at batchedUpdates$1 (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:24392:7)
at Object.act (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom-test-utils.development.js:1092:14)
at wrapAct (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/enzyme-adapter-react-16/src/ReactSixteenAdapter.js:405:13)
at Object.render (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/enzyme-adapter-react-16/src/ReactSixteenAdapter.js:474:16)
at new ReactWrapper (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/enzyme/src/ReactWrapper.js:115:16)
at mount (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/enzyme/src/mount.js:10:10)
at test (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/x-pack/plugins/fleet/storybook/smoke.test.tsx:17:22)
at Object.<anonymous> (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/@storybook/addon-storyshots/dist/api/snapshotsTestsTemplate.js:42:20)
at Promise.then.completed (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/utils.js:276:28)
at new Promise (<anonymous>)
at callAsyncCircusFn (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/utils.js:216:10)
at _callCircusTest (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:212:40)
at _runTest (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:149:3)
at _runTestsForDescribeBlock (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:63:9)
at _runTestsForDescribeBlock (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:57:9)
at _runTestsForDescribeBlock (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:57:9)
at run (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:25:3)
at runAndTransformResultsToJestFormat (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/legacy-code-todo-rewrite/jestAdapterInit.js:176:21)
at jestAdapter (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/legacy-code-todo-rewrite/jestAdapter.js:109:19)
at runTestInternal (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-runner/build/runTest.js:380:16)
at runTest (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-runner/build/runTest.js:472:34)
at Object.worker (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-runner/build/testWorker.js:133:12)</p></pre></div></div></div></div>"
at test (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/x-pack/plugins/fleet/storybook/smoke.test.tsx:20:33)
at _callCircusTest (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:212:5)
at _runTest (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:149:3)
at _runTestsForDescribeBlock (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:63:9)
at _runTestsForDescribeBlock (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:57:9)
at _runTestsForDescribeBlock (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:57:9)
at run (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:25:3)
at runAndTransformResultsToJestFormat (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/legacy-code-todo-rewrite/jestAdapterInit.js:176:21)
at jestAdapter (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/legacy-code-todo-rewrite/jestAdapter.js:109:19)
at runTestInternal (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-runner/build/runTest.js:380:16)
at runTest (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-runner/build/runTest.js:472:34)
at Object.worker (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-runner/build/testWorker.js:133:12)
```
First failure: [CI Build - master](https://buildkite.com/elastic/kibana-hourly/builds/946#ff42e543-5004-441b-887b-0e66b9018892)
<!-- kibanaCiData = {"failed-test":{"test.class":"Jest Tests.x-pack/plugins/fleet/storybook","test.name":"Storyshots Sections/EPM/Package List Grid List","test.failCount":1}} --> | test | failing test jest tests x pack plugins fleet storybook storyshots sections epm package list grid list a test failed on a tracked branch error expect received not tocontain expected indexof expected substring not euierrorboundary received string error error cannot read property startswith of undefined at packagecard opt local ssd buildkite builds kb elastic kibana hourly kibana x pack plugins fleet public applications integrations sections epm components package card tsx at renderwithhooks opt local ssd buildkite builds kb elastic kibana hourly kibana node modules react dom cjs react dom development js at mountindeterminatecomponent opt local ssd buildkite builds kb elastic kibana hourly kibana node modules react dom cjs react dom development js at beginwork opt local ssd buildkite builds kb elastic kibana hourly kibana node modules react dom cjs react dom development js at htmlunknownelement callcallback opt local ssd buildkite builds kb elastic kibana hourly kibana node modules react dom cjs react dom development js at htmlunknownelement calltheuserobjectsoperation opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest environment jsdom node modules jsdom lib jsdom living generated eventlistener js at innerinvokeeventlisteners opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest environment jsdom node modules jsdom lib jsdom living events eventtarget impl js at invokeeventlisteners opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest environment jsdom node modules jsdom lib jsdom living events eventtarget impl js at htmlunknownelementimpl dispatch opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest environment jsdom node modules jsdom lib jsdom living events eventtarget impl js at htmlunknownelementimpl dispatchevent opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest environment jsdom node modules jsdom lib jsdom living events eventtarget impl js at htmlunknownelement dispatchevent opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest environment jsdom node modules jsdom lib jsdom living generated eventtarget js at object invokeguardedcallbackdev opt local ssd buildkite builds kb elastic kibana hourly kibana node modules react dom cjs react dom development js at invokeguardedcallback opt local ssd buildkite builds kb elastic kibana hourly kibana node modules react dom cjs react dom development js at beginwork opt local ssd buildkite builds kb elastic kibana hourly kibana node modules react dom cjs react dom development js at performunitofwork opt local ssd buildkite builds kb elastic kibana hourly kibana node modules react dom cjs react dom development js at workloopsync opt local ssd buildkite builds kb elastic kibana hourly kibana node modules react dom cjs react dom development js at performsyncworkonroot opt local ssd buildkite builds kb elastic kibana hourly kibana node modules react dom cjs react dom development js at opt local ssd buildkite builds kb elastic kibana hourly kibana node modules react dom cjs react dom development js at unstable runwithpriority opt local ssd buildkite builds kb elastic kibana hourly kibana node modules scheduler cjs scheduler development js at runwithpriority opt local ssd buildkite builds kb elastic kibana hourly kibana node modules react dom cjs react dom development js at flushsynccallbackqueueimpl opt local ssd buildkite builds kb elastic kibana hourly kibana node modules react dom cjs react dom development js at flushsynccallbackqueue opt local ssd buildkite builds kb elastic kibana hourly kibana node modules react dom cjs react dom development js at batchedupdates opt local ssd buildkite builds kb elastic kibana hourly kibana node modules react dom cjs react dom development js at object act opt local ssd buildkite builds kb elastic kibana hourly kibana node modules react dom cjs react dom test utils development js at wrapact opt local ssd buildkite builds kb elastic kibana hourly kibana node modules enzyme adapter react src reactsixteenadapter js at object render opt local ssd buildkite builds kb elastic kibana hourly kibana node modules enzyme adapter react src reactsixteenadapter js at new reactwrapper opt local ssd buildkite builds kb elastic kibana hourly kibana node modules enzyme src reactwrapper js at mount opt local ssd buildkite builds kb elastic kibana hourly kibana node modules enzyme src mount js at test opt local ssd buildkite builds kb elastic kibana hourly kibana x pack plugins fleet storybook smoke test tsx at object lt anonymous gt opt local ssd buildkite builds kb elastic kibana hourly kibana node modules storybook addon storyshots dist api snapshotsteststemplate js at promise then completed opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build utils js at new promise lt anonymous gt at callasynccircusfn opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build utils js at callcircustest opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build run js at runtest opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build run js at runtestsfordescribeblock opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build run js at runtestsfordescribeblock opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build run js at runtestsfordescribeblock opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build run js at run opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build run js at runandtransformresultstojestformat opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build legacy code todo rewrite jestadapterinit js at jestadapter opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build legacy code todo rewrite jestadapter js at runtestinternal opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest runner build runtest js at runtest opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest runner build runtest js at object worker opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest runner build testworker js at test opt local ssd buildkite builds kb elastic kibana hourly kibana x pack plugins fleet storybook smoke test tsx at callcircustest opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build run js at runtest opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build run js at runtestsfordescribeblock opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build run js at runtestsfordescribeblock opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build run js at runtestsfordescribeblock opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build run js at run opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build run js at runandtransformresultstojestformat opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build legacy code todo rewrite jestadapterinit js at jestadapter opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build legacy code todo rewrite jestadapter js at runtestinternal opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest runner build runtest js at runtest opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest runner build runtest js at object worker opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest runner build testworker js first failure | 1 |
36,916 | 2,813,567,513 | IssuesEvent | 2015-05-18 15:20:07 | CruxFramework/crux | https://api.github.com/repos/CruxFramework/crux | closed | Error DateBox mask format. | bug imported invalid Module-CruxWidgets Priority-Medium | _From [guilhe...@triggolabs.com](https://code.google.com/u/113083826916919097787/) on August 27, 2014 10:03:43_
Was Identified que DateBox you assume the unexpected behavior in cases for the Following:
1 - If the component is declared without any "pattern" attribute, the mask will be unconfigured.
2 - When is passed the "pattern" attribute, the mask will be unconfigured Following in the cases:
2.1- DateBox in to the "storyboard"
2.2 - DateBox off the panel
obs: DateBox Within the \<gwt: HTMLPanel> or \<crux: styledPanel> Correctly only works WHEN reporting the pattern.
Version: 5.2.0-SNAPSHOT WIDGETS.
Attached a project using the DateBox in several cases.
**Attachment:** [test-components.rar](http://code.google.com/p/crux-framework/issues/detail?id=492)
_Original issue: http://code.google.com/p/crux-framework/issues/detail?id=492_ | 1.0 | Error DateBox mask format. - _From [guilhe...@triggolabs.com](https://code.google.com/u/113083826916919097787/) on August 27, 2014 10:03:43_
Was Identified que DateBox you assume the unexpected behavior in cases for the Following:
1 - If the component is declared without any "pattern" attribute, the mask will be unconfigured.
2 - When is passed the "pattern" attribute, the mask will be unconfigured Following in the cases:
2.1- DateBox in to the "storyboard"
2.2 - DateBox off the panel
obs: DateBox Within the \<gwt: HTMLPanel> or \<crux: styledPanel> Correctly only works WHEN reporting the pattern.
Version: 5.2.0-SNAPSHOT WIDGETS.
Attached a project using the DateBox in several cases.
**Attachment:** [test-components.rar](http://code.google.com/p/crux-framework/issues/detail?id=492)
_Original issue: http://code.google.com/p/crux-framework/issues/detail?id=492_ | non_test | error datebox mask format from on august was identified que datebox you assume the unexpected behavior in cases for the following if the component is declared without any pattern attribute the mask will be unconfigured when is passed the pattern attribute the mask will be unconfigured following in the cases datebox in to the storyboard datebox off the panel obs datebox within the or correctly only works when reporting the pattern version snapshot widgets attached a project using the datebox in several cases attachment original issue | 0 |
190,127 | 15,218,236,534 | IssuesEvent | 2021-02-17 17:33:59 | Carisk/Carisk-backend | https://api.github.com/repos/Carisk/Carisk-backend | closed | Novo problema a se resolver com aplicativo | Arquitetura documentation | **Descrição**
O escopo do aplicativo atual é demasiadamente simple e pode ser exapandido de forma a se tornar mais útil ao usuário final. Isso é possível de ser feito com a detecção de novos problemas a serem resolvidos.
**Critérios de aceitação**
- [ ] Novo problema a ser resolvido, com sua possível solução e utilidade, bem como comunicação com o time e documentação. | 1.0 | Novo problema a se resolver com aplicativo - **Descrição**
O escopo do aplicativo atual é demasiadamente simple e pode ser exapandido de forma a se tornar mais útil ao usuário final. Isso é possível de ser feito com a detecção de novos problemas a serem resolvidos.
**Critérios de aceitação**
- [ ] Novo problema a ser resolvido, com sua possível solução e utilidade, bem como comunicação com o time e documentação. | non_test | novo problema a se resolver com aplicativo descrição o escopo do aplicativo atual é demasiadamente simple e pode ser exapandido de forma a se tornar mais útil ao usuário final isso é possível de ser feito com a detecção de novos problemas a serem resolvidos critérios de aceitação novo problema a ser resolvido com sua possível solução e utilidade bem como comunicação com o time e documentação | 0 |
25,383 | 18,670,745,421 | IssuesEvent | 2021-10-30 17:02:10 | battlecode/battlecode21 | https://api.github.com/repos/battlecode/battlecode21 | closed | In tournament result publishers, make tournament url a command line arg | infrastructure | tour url is a var right now. this makes it really hard to run concurrently :((
also avoids funny pushes and things | 1.0 | In tournament result publishers, make tournament url a command line arg - tour url is a var right now. this makes it really hard to run concurrently :((
also avoids funny pushes and things | non_test | in tournament result publishers make tournament url a command line arg tour url is a var right now this makes it really hard to run concurrently also avoids funny pushes and things | 0 |
73,657 | 3,419,254,383 | IssuesEvent | 2015-12-08 08:47:15 | movabletype/smartphone-app | https://api.github.com/repos/movabletype/smartphone-app | opened | Would be great if user can configure image width freely. | enhancement Priority: MID | ## Specification
* The user must choose the width from a list, or enter a value.
* The system must be provided default settings.
| 1.0 | Would be great if user can configure image width freely. - ## Specification
* The user must choose the width from a list, or enter a value.
* The system must be provided default settings.
| non_test | would be great if user can configure image width freely specification the user must choose the width from a list or enter a value the system must be provided default settings | 0 |
577,808 | 17,135,390,342 | IssuesEvent | 2021-07-13 00:54:29 | internetarchive/openlibrary | https://api.github.com/repos/internetarchive/openlibrary | opened | Upgrade to HTTP/2 | Lead: @jamesachamp Needs: Investigation Priority: 3 Type: Feature Request | <!-- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] -->
### Describe the problem that you'd like solved
<!-- A clear and concise description of what you want to happen. -->
While page load times have improved over the past year, they could still be faster.
### Proposal & Constraints
<!-- What is the proposed solution / implementation? Is there a precedent of this approach succeeding elsewhere? -->
Upgrade our web and covers servers to use HTTP/2. Doing so will greatly limit the amount of TCP connections needed to serve a page to a patron. Server push may also be utilized to further improve performance.
<!-- Which suggestions or requirements should be considered for how feature needs to appear or be implemented? -->
### Additional context
<!-- Add any other context or screenshots about the feature request here. -->
### Stakeholders
<!-- @ tag stakeholders of this bug -->
| 1.0 | Upgrade to HTTP/2 - <!-- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] -->
### Describe the problem that you'd like solved
<!-- A clear and concise description of what you want to happen. -->
While page load times have improved over the past year, they could still be faster.
### Proposal & Constraints
<!-- What is the proposed solution / implementation? Is there a precedent of this approach succeeding elsewhere? -->
Upgrade our web and covers servers to use HTTP/2. Doing so will greatly limit the amount of TCP connections needed to serve a page to a patron. Server push may also be utilized to further improve performance.
<!-- Which suggestions or requirements should be considered for how feature needs to appear or be implemented? -->
### Additional context
<!-- Add any other context or screenshots about the feature request here. -->
### Stakeholders
<!-- @ tag stakeholders of this bug -->
| non_test | upgrade to http describe the problem that you d like solved while page load times have improved over the past year they could still be faster proposal constraints upgrade our web and covers servers to use http doing so will greatly limit the amount of tcp connections needed to serve a page to a patron server push may also be utilized to further improve performance additional context stakeholders | 0 |
44,918 | 11,545,257,740 | IssuesEvent | 2020-02-18 13:07:15 | imageio/imageio | https://api.github.com/repos/imageio/imageio | closed | Add proxy support to ```fetching.py``` | building/installing plugin: ffmpeg | Installing ffmpeg automatically causes some problems behind a firewall using the ```imageio.plugins.ffmpeg.download()``` method.
Proxy support can be enhanced by adding/changing the following within [```_fetch_file```](https://github.com/imageio/imageio/blob/368c724957cc67652f9c24b32628e4fdc19ecd3c/imageio/core/fetching.py#L159).
``` python
authinfo = urllib.request.HTTPBasicAuthHandler()
proxy_support = urllib.request.ProxyHandler({'https': 'http://yourproxy.corporate.zombie.net:2011'})
opener = urllib.request.build_opener(proxy_support, authinfo,urllib.request.CacheFTPHandler)
urllib.request.install_opener(opener)
remote_file = urllib.request.urlopen(url)
```
Perhaps it would be good to pass a proxy location through ```imageio.plugins.ffmpeg.download()``` to make the process suitable when needing a proxy? | 1.0 | Add proxy support to ```fetching.py``` - Installing ffmpeg automatically causes some problems behind a firewall using the ```imageio.plugins.ffmpeg.download()``` method.
Proxy support can be enhanced by adding/changing the following within [```_fetch_file```](https://github.com/imageio/imageio/blob/368c724957cc67652f9c24b32628e4fdc19ecd3c/imageio/core/fetching.py#L159).
``` python
authinfo = urllib.request.HTTPBasicAuthHandler()
proxy_support = urllib.request.ProxyHandler({'https': 'http://yourproxy.corporate.zombie.net:2011'})
opener = urllib.request.build_opener(proxy_support, authinfo,urllib.request.CacheFTPHandler)
urllib.request.install_opener(opener)
remote_file = urllib.request.urlopen(url)
```
Perhaps it would be good to pass a proxy location through ```imageio.plugins.ffmpeg.download()``` to make the process suitable when needing a proxy? | non_test | add proxy support to fetching py installing ffmpeg automatically causes some problems behind a firewall using the imageio plugins ffmpeg download method proxy support can be enhanced by adding changing the following within python authinfo urllib request httpbasicauthhandler proxy support urllib request proxyhandler https opener urllib request build opener proxy support authinfo urllib request cacheftphandler urllib request install opener opener remote file urllib request urlopen url perhaps it would be good to pass a proxy location through imageio plugins ffmpeg download to make the process suitable when needing a proxy | 0 |
142,519 | 11,483,933,125 | IssuesEvent | 2020-02-11 01:37:56 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | opened | roachtest: pgjdbc failed | C-test-failure O-roachtest O-robot branch-provisional_202002062136_v19.1.8 release-blocker | [(roachtest).pgjdbc failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=1742234&tab=buildLog) on [provisional_202002062136_v19.1.8@407017cad14dfa63f19578055082dc10f3283cc4](https://github.com/cockroachdb/cockroach/commits/407017cad14dfa63f19578055082dc10f3283cc4):
```
The test failed on branch=provisional_202002062136_v19.1.8, cloud=gce:
test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/pgjdbc/run_1
java_helpers.go:121,java_helpers.go:213,pgjdbc.go:172,pgjdbc.go:184,test_runner.go:741: XML syntax error on line 15002: illegal character code U+0001
```
<details><summary>More</summary><p>
Artifacts: [/pgjdbc](https://teamcity.cockroachdb.com/viewLog.html?buildId=1742234&tab=artifacts#/pgjdbc)
Related:
- #44932 roachtest: pgjdbc failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202002062122_v19.2.4](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202002062122_v19.2.4) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #44844 roachtest: pgjdbc failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-19.2](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-19.2)
- #44284 roachtest: pgjdbc failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-19.1](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-19.1)
- #44283 roachtest: pgjdbc failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-2.1](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-2.1)
- #44254 roachtest: pgjdbc failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202001221958_v2.1.11](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202001221958_v2.1.11)
- #42905 roachtest: pgjdbc failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-master](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-master)
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Apgjdbc.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
| 2.0 | roachtest: pgjdbc failed - [(roachtest).pgjdbc failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=1742234&tab=buildLog) on [provisional_202002062136_v19.1.8@407017cad14dfa63f19578055082dc10f3283cc4](https://github.com/cockroachdb/cockroach/commits/407017cad14dfa63f19578055082dc10f3283cc4):
```
The test failed on branch=provisional_202002062136_v19.1.8, cloud=gce:
test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/pgjdbc/run_1
java_helpers.go:121,java_helpers.go:213,pgjdbc.go:172,pgjdbc.go:184,test_runner.go:741: XML syntax error on line 15002: illegal character code U+0001
```
<details><summary>More</summary><p>
Artifacts: [/pgjdbc](https://teamcity.cockroachdb.com/viewLog.html?buildId=1742234&tab=artifacts#/pgjdbc)
Related:
- #44932 roachtest: pgjdbc failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202002062122_v19.2.4](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202002062122_v19.2.4) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #44844 roachtest: pgjdbc failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-19.2](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-19.2)
- #44284 roachtest: pgjdbc failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-19.1](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-19.1)
- #44283 roachtest: pgjdbc failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-2.1](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-2.1)
- #44254 roachtest: pgjdbc failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202001221958_v2.1.11](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202001221958_v2.1.11)
- #42905 roachtest: pgjdbc failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-master](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-master)
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Apgjdbc.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
| test | roachtest pgjdbc failed on the test failed on branch provisional cloud gce test artifacts and logs in home agent work go src github com cockroachdb cockroach artifacts pgjdbc run java helpers go java helpers go pgjdbc go pgjdbc go test runner go xml syntax error on line illegal character code u more artifacts related roachtest pgjdbc failed roachtest pgjdbc failed roachtest pgjdbc failed roachtest pgjdbc failed roachtest pgjdbc failed roachtest pgjdbc failed powered by | 1 |
336,707 | 30,217,337,629 | IssuesEvent | 2023-07-05 16:30:20 | BoBAdministration/QA-Bug-Reports | https://api.github.com/repos/BoBAdministration/QA-Bug-Reports | closed | Acro doesnt have its idle animation in its emote wheel | Fixed-PendingTesting | **Describe the Bug**
Acro doesnt have its idle animation in the emote wheel
**To Reproduce**
1. log into tester server
2. spawn in as acro
3. open emote menu
**Expected behavior**
emote menu has idle anim
**Actual behavior**
emote menu does not have idle anim
**Screenshots & Video**

**Branch Version**
Tester
**Character Information**
1.2 Acro
**Additional Information**
Tested on titania | 1.0 | Acro doesnt have its idle animation in its emote wheel - **Describe the Bug**
Acro doesnt have its idle animation in the emote wheel
**To Reproduce**
1. log into tester server
2. spawn in as acro
3. open emote menu
**Expected behavior**
emote menu has idle anim
**Actual behavior**
emote menu does not have idle anim
**Screenshots & Video**

**Branch Version**
Tester
**Character Information**
1.2 Acro
**Additional Information**
Tested on titania | test | acro doesnt have its idle animation in its emote wheel describe the bug acro doesnt have its idle animation in the emote wheel to reproduce log into tester server spawn in as acro open emote menu expected behavior emote menu has idle anim actual behavior emote menu does not have idle anim screenshots video branch version tester character information acro additional information tested on titania | 1 |
230,489 | 17,618,833,591 | IssuesEvent | 2021-08-18 13:09:41 | UnBArqDsw2021-1/2021.1_G02_TaNaMesa_docs | https://api.github.com/repos/UnBArqDsw2021-1/2021.1_G02_TaNaMesa_docs | opened | Protótipo de Alta Fidelidade | documentation | ## Descrição da issue
Issue destinada a criação e documentação do protótipo de alta fidelidade, baseado no protótipo de baixa fidelidade e nos requisitos elicitados para o projeto.
## Contexto adicional
https://github.com/UnBArqDsw2021-1/2021.1_G02_TaNaMesa_docs/issues/69
## Tópicos:
- [ ] Criar documento
- [ ] Criar protótipo
- [ ] Documentar telas
- [ ] Revisar documento
| 1.0 | Protótipo de Alta Fidelidade - ## Descrição da issue
Issue destinada a criação e documentação do protótipo de alta fidelidade, baseado no protótipo de baixa fidelidade e nos requisitos elicitados para o projeto.
## Contexto adicional
https://github.com/UnBArqDsw2021-1/2021.1_G02_TaNaMesa_docs/issues/69
## Tópicos:
- [ ] Criar documento
- [ ] Criar protótipo
- [ ] Documentar telas
- [ ] Revisar documento
| non_test | protótipo de alta fidelidade descrição da issue issue destinada a criação e documentação do protótipo de alta fidelidade baseado no protótipo de baixa fidelidade e nos requisitos elicitados para o projeto contexto adicional tópicos criar documento criar protótipo documentar telas revisar documento | 0 |
91,999 | 18,758,275,952 | IssuesEvent | 2021-11-05 13:40:07 | Onelinerhub/onelinerhub | https://api.github.com/repos/Onelinerhub/onelinerhub | closed | Write shortest possible code: How to import module from another directory (python) | help wanted good first issue code python | Please write shortest code example for this question:
**How to import module from another directory** in python
### How to do it:
1. Go to [python codes](https://github.com/Onelinerhub/onelinerhub/tree/main/python)
2. Create new file (named in underscore case, should contain key words from title) with `md` extension (markdown file).
3. Propose new file with following content (please use all three blocks if possible - title, code itself and explanations list):
~~~
# How to import module from another directory
```python
code part1 part2 part3 ...
```
- part1 - explain code part 1
- part2 - explain code part 2
- ...
~~~
More [advanced template](https://github.com/Onelinerhub/onelinerhub/blob/main/template.md) for examples and linked solutions. More [docs here](https://github.com/Onelinerhub/onelinerhub#onelinerhub). | 1.0 | Write shortest possible code: How to import module from another directory (python) - Please write shortest code example for this question:
**How to import module from another directory** in python
### How to do it:
1. Go to [python codes](https://github.com/Onelinerhub/onelinerhub/tree/main/python)
2. Create new file (named in underscore case, should contain key words from title) with `md` extension (markdown file).
3. Propose new file with following content (please use all three blocks if possible - title, code itself and explanations list):
~~~
# How to import module from another directory
```python
code part1 part2 part3 ...
```
- part1 - explain code part 1
- part2 - explain code part 2
- ...
~~~
More [advanced template](https://github.com/Onelinerhub/onelinerhub/blob/main/template.md) for examples and linked solutions. More [docs here](https://github.com/Onelinerhub/onelinerhub#onelinerhub). | non_test | write shortest possible code how to import module from another directory python please write shortest code example for this question how to import module from another directory in python how to do it go to create new file named in underscore case should contain key words from title with md extension markdown file propose new file with following content please use all three blocks if possible title code itself and explanations list how to import module from another directory python code explain code part explain code part more for examples and linked solutions more | 0 |
328,972 | 28,143,241,456 | IssuesEvent | 2023-04-02 07:07:37 | unifyai/ivy | https://api.github.com/repos/unifyai/ivy | opened | Fix backend_handler.test_variables | Sub Task Failing Test | | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4586867454/jobs/8100010916" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/4586867454/jobs/8100010808" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4586867454/jobs/8100007678" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/4586867454/jobs/8100009708" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
<details>
<summary>Not found</summary>
Not found
</details>
<details>
<summary>Not found</summary>
Not found
</details>
<details>
<summary>Not found</summary>
Not found
</details>
<details>
<summary>Not found</summary>
Not found
</details>
| 1.0 | Fix backend_handler.test_variables - | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4586867454/jobs/8100010916" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/4586867454/jobs/8100010808" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4586867454/jobs/8100007678" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/4586867454/jobs/8100009708" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
<details>
<summary>Not found</summary>
Not found
</details>
<details>
<summary>Not found</summary>
Not found
</details>
<details>
<summary>Not found</summary>
Not found
</details>
<details>
<summary>Not found</summary>
Not found
</details>
| test | fix backend handler test variables tensorflow img src torch img src numpy img src jax img src not found not found not found not found not found not found not found not found | 1 |
149,941 | 11,940,788,156 | IssuesEvent | 2020-04-02 17:18:12 | rancher/rancher | https://api.github.com/repos/rancher/rancher | opened | test_catalog_refresh flaky test | kind/flaky-test | https://drone-publish.rancher.io/rancher/rancher/2676/1/2
https://drone-pr.rancher.io/rancher/rancher/7191/2/2
```
=================================== FAILURES ===================================
--
904 | _____________________________ test_catalog_refresh _____________________________
905 | [gw5] linux -- Python 3.7.5 /go/src/github.com/rancher/rancher/tests/integration/.tox/py37/bin/python3.7
906 |
907 | admin_mc = <suite.conftest.ManagementContext object at 0x7f813b3d1ed0>
908 |
909 | def test_catalog_refresh(admin_mc):
910 | """Test that on refresh the response includes the names of the catalogs
911 | that are being refreshed"""
912 | client = admin_mc.client
913 | catalog = client.by_id_catalog("library")
914 | out = client.action(obj=catalog, action_name="refresh")
915 | assert out['catalogs'][0] == "library"
916 |
917 | catalogs = client.list_catalog()
918 | > out = client.action(obj=catalogs, action_name="refresh")
919 |
920 | test_catalog.py:282:
921 | _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
922 | ../.tox/py37/lib/python3.7/site-packages/rancher.py:471: in action
923 | return self._post_and_retry(url, *args, **kw)
924 | ../.tox/py37/lib/python3.7/site-packages/rancher.py:430: in _post_and_retry
925 | raise e
926 | ../.tox/py37/lib/python3.7/site-packages/rancher.py:423: in _post_and_retry
927 | return self._post(url, data=self._to_dict(*args, **kw))
928 | ../.tox/py37/lib/python3.7/site-packages/rancher.py:68: in wrapped
929 | return fn(*args, **kw)
930 | ../.tox/py37/lib/python3.7/site-packages/rancher.py:319: in _post
931 | self._error(r.text)
932 | _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
933 |
934 | self = <rancher.Client object at 0x7f813b745e90>
935 | text = '{"baseType":"error","code":"ServerError","message":"catalogs.management.cattle.io \\"random-284755-1585846570\\" not found","status":500,"type":"error"}\n'
936 |
937 | def _error(self, text):
938 | > raise ApiError(self._unmarshall(text))
939 | E rancher.ApiError: (ApiError(...), 'ServerError : catalogs.management.cattle.io "random-284755-1585846570" not found\n\t{\'baseType\': \'error\', \'code\': \'ServerError\', \'message\': \'catalogs.management.cattle.io "random-284755-1585846570" not found\', \'status\': 500, \'type\': \'error\'}')
940 |
941 | ../.tox/py37/lib/python3.7/site-packages/rancher.py:299: ApiError
``` | 1.0 | test_catalog_refresh flaky test - https://drone-publish.rancher.io/rancher/rancher/2676/1/2
https://drone-pr.rancher.io/rancher/rancher/7191/2/2
```
=================================== FAILURES ===================================
--
904 | _____________________________ test_catalog_refresh _____________________________
905 | [gw5] linux -- Python 3.7.5 /go/src/github.com/rancher/rancher/tests/integration/.tox/py37/bin/python3.7
906 |
907 | admin_mc = <suite.conftest.ManagementContext object at 0x7f813b3d1ed0>
908 |
909 | def test_catalog_refresh(admin_mc):
910 | """Test that on refresh the response includes the names of the catalogs
911 | that are being refreshed"""
912 | client = admin_mc.client
913 | catalog = client.by_id_catalog("library")
914 | out = client.action(obj=catalog, action_name="refresh")
915 | assert out['catalogs'][0] == "library"
916 |
917 | catalogs = client.list_catalog()
918 | > out = client.action(obj=catalogs, action_name="refresh")
919 |
920 | test_catalog.py:282:
921 | _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
922 | ../.tox/py37/lib/python3.7/site-packages/rancher.py:471: in action
923 | return self._post_and_retry(url, *args, **kw)
924 | ../.tox/py37/lib/python3.7/site-packages/rancher.py:430: in _post_and_retry
925 | raise e
926 | ../.tox/py37/lib/python3.7/site-packages/rancher.py:423: in _post_and_retry
927 | return self._post(url, data=self._to_dict(*args, **kw))
928 | ../.tox/py37/lib/python3.7/site-packages/rancher.py:68: in wrapped
929 | return fn(*args, **kw)
930 | ../.tox/py37/lib/python3.7/site-packages/rancher.py:319: in _post
931 | self._error(r.text)
932 | _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
933 |
934 | self = <rancher.Client object at 0x7f813b745e90>
935 | text = '{"baseType":"error","code":"ServerError","message":"catalogs.management.cattle.io \\"random-284755-1585846570\\" not found","status":500,"type":"error"}\n'
936 |
937 | def _error(self, text):
938 | > raise ApiError(self._unmarshall(text))
939 | E rancher.ApiError: (ApiError(...), 'ServerError : catalogs.management.cattle.io "random-284755-1585846570" not found\n\t{\'baseType\': \'error\', \'code\': \'ServerError\', \'message\': \'catalogs.management.cattle.io "random-284755-1585846570" not found\', \'status\': 500, \'type\': \'error\'}')
940 |
941 | ../.tox/py37/lib/python3.7/site-packages/rancher.py:299: ApiError
``` | test | test catalog refresh flaky test failures test catalog refresh linux python go src github com rancher rancher tests integration tox bin admin mc def test catalog refresh admin mc test that on refresh the response includes the names of the catalogs that are being refreshed client admin mc client catalog client by id catalog library out client action obj catalog action name refresh assert out library catalogs client list catalog out client action obj catalogs action name refresh test catalog py tox lib site packages rancher py in action return self post and retry url args kw tox lib site packages rancher py in post and retry raise e tox lib site packages rancher py in post and retry return self post url data self to dict args kw tox lib site packages rancher py in wrapped return fn args kw tox lib site packages rancher py in post self error r text self text basetype error code servererror message catalogs management cattle io random not found status type error n def error self text raise apierror self unmarshall text e rancher apierror apierror servererror catalogs management cattle io random not found n t basetype error code servererror message catalogs management cattle io random not found status type error tox lib site packages rancher py apierror | 1 |
70,212 | 23,051,623,561 | IssuesEvent | 2022-07-24 18:11:11 | vector-im/element-android | https://api.github.com/repos/vector-im/element-android | closed | Voice message lock mechanism is incompatible with gesture navigation | T-Defect A-Composer A-Voice-Messages S-Major O-Occasional Z-Platform Specific | #### Describe the bug
If your phone uses gesture navigation, it is impossible to use the voice message slide-to-lock feature, since sliding upwards from the voice message button is interpreted as trying to close the app.
#### To Reproduce
1. Enable gesture navigation on your phone
2. Press and hold the voice message button
3. Attempt to slide upwards to lock the recording
#### Expected behavior
An upwards swipe should lock the recording instead of trying to close the app
#### Screenshots
https://user-images.githubusercontent.com/48614497/130335718-ac906310-2be0-4a5e-ae9a-64f3babebd85.mp4
#### Smartphone (please complete the following information):
- Device: Google Pixel 3a
- OS: LineageOS 15.1
#### Additional context
- App version and store: 1.2.0 - F-Droid | 1.0 | Voice message lock mechanism is incompatible with gesture navigation - #### Describe the bug
If your phone uses gesture navigation, it is impossible to use the voice message slide-to-lock feature, since sliding upwards from the voice message button is interpreted as trying to close the app.
#### To Reproduce
1. Enable gesture navigation on your phone
2. Press and hold the voice message button
3. Attempt to slide upwards to lock the recording
#### Expected behavior
An upwards swipe should lock the recording instead of trying to close the app
#### Screenshots
https://user-images.githubusercontent.com/48614497/130335718-ac906310-2be0-4a5e-ae9a-64f3babebd85.mp4
#### Smartphone (please complete the following information):
- Device: Google Pixel 3a
- OS: LineageOS 15.1
#### Additional context
- App version and store: 1.2.0 - F-Droid | non_test | voice message lock mechanism is incompatible with gesture navigation describe the bug if your phone uses gesture navigation it is impossible to use the voice message slide to lock feature since sliding upwards from the voice message button is interpreted as trying to close the app to reproduce enable gesture navigation on your phone press and hold the voice message button attempt to slide upwards to lock the recording expected behavior an upwards swipe should lock the recording instead of trying to close the app screenshots smartphone please complete the following information device google pixel os lineageos additional context app version and store f droid | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.