Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
4
112
repo_url
stringlengths
33
141
action
stringclasses
3 values
title
stringlengths
1
1.02k
labels
stringlengths
4
1.54k
body
stringlengths
1
262k
index
stringclasses
17 values
text_combine
stringlengths
95
262k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
51,845
3,014,617,585
IssuesEvent
2015-07-29 15:36:00
jpchanson/BeSeenium
https://api.github.com/repos/jpchanson/BeSeenium
closed
do jetty proof of concept
Core functionality High Priority
get jetty to my controller test, get jetty to display some stuff on the screen.
1.0
do jetty proof of concept - get jetty to my controller test, get jetty to display some stuff on the screen.
non_test
do jetty proof of concept get jetty to my controller test get jetty to display some stuff on the screen
0
1,868
4,697,449,435
IssuesEvent
2016-10-12 09:24:13
nodejs/node
https://api.github.com/repos/nodejs/node
closed
degrading performance after using child_process
child_process confirmed-bug lts-watch-v4.x os x performance
After running a child process ( using exec or spawn ) I have found that the performance of my node.js application decreases by a factor of 10. Below is a contrived example and output. ``` var exec = require('child_process').exec; function runExpensiveOperation(times) { while(times > 0) { console.time('expensiveOperation'); var str = 'lorem'; for ( var i=0;i< 10000000; i++) { // string concatenation str = str.length < 1000 ? str + str : ''; // math operation i * i * i; } console.timeEnd('expensiveOperation'); times--; } } console.log('PRE EXEC'); runExpensiveOperation(10); exec('echo "hello"'); console.log('POST EXEC'); runExpensiveOperation(10); ``` Output: ``` PRE EXEC expensiveOperation: 66.458ms expensiveOperation: 65.735ms expensiveOperation: 69.237ms expensiveOperation: 65.269ms expensiveOperation: 69.133ms expensiveOperation: 65.639ms expensiveOperation: 67.944ms expensiveOperation: 63.595ms expensiveOperation: 64.153ms expensiveOperation: 65.093ms POST EXEC expensiveOperation: 715.861ms expensiveOperation: 739.671ms expensiveOperation: 714.546ms expensiveOperation: 714.845ms expensiveOperation: 745.719ms expensiveOperation: 743.240ms expensiveOperation: 716.481ms expensiveOperation: 732.916ms expensiveOperation: 736.576ms expensiveOperation: 742.416ms ``` In addition, this problem only occurs if the string concatenation AND math operation are run in the expensiveOperation - if either are commented out then there is no issue. * **Version**: 5.8.0 * **Platform**: Darwin Kernel Version 15.3.0: Thu Dec 10 18:40:58 PST 2015; root:xnu-3248.30.4~1/RELEASE_X86_64 x86_64 ( Macbook Air OS X El Capitan ) * **Subsystem**: child_process
1.0
degrading performance after using child_process - After running a child process ( using exec or spawn ) I have found that the performance of my node.js application decreases by a factor of 10. Below is a contrived example and output. ``` var exec = require('child_process').exec; function runExpensiveOperation(times) { while(times > 0) { console.time('expensiveOperation'); var str = 'lorem'; for ( var i=0;i< 10000000; i++) { // string concatenation str = str.length < 1000 ? str + str : ''; // math operation i * i * i; } console.timeEnd('expensiveOperation'); times--; } } console.log('PRE EXEC'); runExpensiveOperation(10); exec('echo "hello"'); console.log('POST EXEC'); runExpensiveOperation(10); ``` Output: ``` PRE EXEC expensiveOperation: 66.458ms expensiveOperation: 65.735ms expensiveOperation: 69.237ms expensiveOperation: 65.269ms expensiveOperation: 69.133ms expensiveOperation: 65.639ms expensiveOperation: 67.944ms expensiveOperation: 63.595ms expensiveOperation: 64.153ms expensiveOperation: 65.093ms POST EXEC expensiveOperation: 715.861ms expensiveOperation: 739.671ms expensiveOperation: 714.546ms expensiveOperation: 714.845ms expensiveOperation: 745.719ms expensiveOperation: 743.240ms expensiveOperation: 716.481ms expensiveOperation: 732.916ms expensiveOperation: 736.576ms expensiveOperation: 742.416ms ``` In addition, this problem only occurs if the string concatenation AND math operation are run in the expensiveOperation - if either are commented out then there is no issue. * **Version**: 5.8.0 * **Platform**: Darwin Kernel Version 15.3.0: Thu Dec 10 18:40:58 PST 2015; root:xnu-3248.30.4~1/RELEASE_X86_64 x86_64 ( Macbook Air OS X El Capitan ) * **Subsystem**: child_process
non_test
degrading performance after using child process after running a child process using exec or spawn i have found that the performance of my node js application decreases by a factor of below is a contrived example and output var exec require child process exec function runexpensiveoperation times while times console time expensiveoperation var str lorem for var i i i string concatenation str str length str str math operation i i i console timeend expensiveoperation times console log pre exec runexpensiveoperation exec echo hello console log post exec runexpensiveoperation output pre exec expensiveoperation expensiveoperation expensiveoperation expensiveoperation expensiveoperation expensiveoperation expensiveoperation expensiveoperation expensiveoperation expensiveoperation post exec expensiveoperation expensiveoperation expensiveoperation expensiveoperation expensiveoperation expensiveoperation expensiveoperation expensiveoperation expensiveoperation expensiveoperation in addition this problem only occurs if the string concatenation and math operation are run in the expensiveoperation if either are commented out then there is no issue version platform darwin kernel version thu dec pst root xnu release macbook air os x el capitan subsystem child process
0
270,507
28,962,277,549
IssuesEvent
2023-05-10 04:19:23
nidhi7598/external_curl_AOSP10_r33
https://api.github.com/repos/nidhi7598/external_curl_AOSP10_r33
opened
CVE-2022-32205 (Medium) detected in multiple libraries
Mend: dependency security vulnerability
## CVE-2022-32205 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>curlcurl-7_64_1</b>, <b>curlcurl-7_64_1</b>, <b>curlcurl-7_64_1</b>, <b>curlcurl-7_64_1</b>, <b>curlcurl-7_64_1</b>, <b>curlcurl-7_64_1</b>, <b>curlcurl-7_64_1</b>, <b>curlcurl-7_64_1</b>, <b>curlcurl-7_64_1</b></p></summary> <p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> A malicious server can serve excessive amounts of `Set-Cookie:` headers in a HTTP response to curl and curl < 7.84.0 stores all of them. A sufficiently large amount of (big) cookies make subsequent HTTP requests to this, or other servers to which the cookies match, create requests that become larger than the threshold that curl uses internally to avoid sending crazy large requests (1048576 bytes) and instead returns an error.This denial state might remain for as long as the same cookies are kept, match and haven't expired. Due to cookie matching rules, a server on `foo.example.com` can set cookies that also would match for `bar.example.com`, making it it possible for a "sister server" to effectively cause a denial of service for a sibling site on the same second level domain using this method. <p>Publish Date: 2022-07-07 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-32205>CVE-2022-32205</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-07-07</p> <p>Fix Resolution: curl-7_71_0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-32205 (Medium) detected in multiple libraries - ## CVE-2022-32205 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>curlcurl-7_64_1</b>, <b>curlcurl-7_64_1</b>, <b>curlcurl-7_64_1</b>, <b>curlcurl-7_64_1</b>, <b>curlcurl-7_64_1</b>, <b>curlcurl-7_64_1</b>, <b>curlcurl-7_64_1</b>, <b>curlcurl-7_64_1</b>, <b>curlcurl-7_64_1</b></p></summary> <p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> A malicious server can serve excessive amounts of `Set-Cookie:` headers in a HTTP response to curl and curl < 7.84.0 stores all of them. A sufficiently large amount of (big) cookies make subsequent HTTP requests to this, or other servers to which the cookies match, create requests that become larger than the threshold that curl uses internally to avoid sending crazy large requests (1048576 bytes) and instead returns an error.This denial state might remain for as long as the same cookies are kept, match and haven't expired. Due to cookie matching rules, a server on `foo.example.com` can set cookies that also would match for `bar.example.com`, making it it possible for a "sister server" to effectively cause a denial of service for a sibling site on the same second level domain using this method. <p>Publish Date: 2022-07-07 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-32205>CVE-2022-32205</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-07-07</p> <p>Fix Resolution: curl-7_71_0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve medium detected in multiple libraries cve medium severity vulnerability vulnerable libraries curlcurl curlcurl curlcurl curlcurl curlcurl curlcurl curlcurl curlcurl curlcurl vulnerability details a malicious server can serve excessive amounts of set cookie headers in a http response to curl and curl stores all of them a sufficiently large amount of big cookies make subsequent http requests to this or other servers to which the cookies match create requests that become larger than the threshold that curl uses internally to avoid sending crazy large requests bytes and instead returns an error this denial state might remain for as long as the same cookies are kept match and haven t expired due to cookie matching rules a server on foo example com can set cookies that also would match for bar example com making it it possible for a sister server to effectively cause a denial of service for a sibling site on the same second level domain using this method publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version release date fix resolution curl step up your open source security game with mend
0
686,881
23,507,525,505
IssuesEvent
2022-08-18 13:49:55
twisted/twisted
https://api.github.com/repos/twisted/twisted
closed
TypeError: 'DelayedCall' object is not iterable
core bug priority-normal new
|<img alt="allenap's avatar" src="https://avatars.githubusercontent.com/u/0?s=50" width="50" height="50">| allenap reported| |-|-| |Trac ID|trac#8307| |Type|defect| |Created|2016-04-26 16:06:50Z| ``` Python 3.5.1+ (default, Mar 30 2016, 22:46:26) [GCC 5.3.1 20160330] on linux Type "help", "copyright", "credits" or "license" for more information. >>> from twisted.internet.base import DelayedCall >>> dc = DelayedCall(1, lambda: None, (), {}, lambda dc: None, lambda dc: None) >>> dc.debug = True >>> dc.cancel() Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/home/gavin/GitHub/twisted/twisted/internet/base.py", line 94, in cancel self._str = bytes(self) TypeError: 'DelayedCall' object is not iterable ``` ``` $ tail -n1 twisted/_version.py version = versions.Version('twisted', 16, 1, 1) ``` <details><summary>Searchable metadata</summary> ``` trac-id__8307 8307 type__defect defect reporter__allenap allenap priority__normal normal milestone__None None branch__ branch_author__ status__new new resolution__None None component__core core keywords__None None time__1461686810312194 1461686810312194 changetime__1462290490013595 1462290490013595 version__None None owner__None None ``` </details>
1.0
TypeError: 'DelayedCall' object is not iterable - |<img alt="allenap's avatar" src="https://avatars.githubusercontent.com/u/0?s=50" width="50" height="50">| allenap reported| |-|-| |Trac ID|trac#8307| |Type|defect| |Created|2016-04-26 16:06:50Z| ``` Python 3.5.1+ (default, Mar 30 2016, 22:46:26) [GCC 5.3.1 20160330] on linux Type "help", "copyright", "credits" or "license" for more information. >>> from twisted.internet.base import DelayedCall >>> dc = DelayedCall(1, lambda: None, (), {}, lambda dc: None, lambda dc: None) >>> dc.debug = True >>> dc.cancel() Traceback (most recent call last): File "<stdin>", line 1, in <module> File "/home/gavin/GitHub/twisted/twisted/internet/base.py", line 94, in cancel self._str = bytes(self) TypeError: 'DelayedCall' object is not iterable ``` ``` $ tail -n1 twisted/_version.py version = versions.Version('twisted', 16, 1, 1) ``` <details><summary>Searchable metadata</summary> ``` trac-id__8307 8307 type__defect defect reporter__allenap allenap priority__normal normal milestone__None None branch__ branch_author__ status__new new resolution__None None component__core core keywords__None None time__1461686810312194 1461686810312194 changetime__1462290490013595 1462290490013595 version__None None owner__None None ``` </details>
non_test
typeerror delayedcall object is not iterable allenap reported trac id trac type defect created python default mar on linux type help copyright credits or license for more information from twisted internet base import delayedcall dc delayedcall lambda none lambda dc none lambda dc none dc debug true dc cancel traceback most recent call last file line in file home gavin github twisted twisted internet base py line in cancel self str bytes self typeerror delayedcall object is not iterable tail twisted version py version versions version twisted searchable metadata trac id type defect defect reporter allenap allenap priority normal normal milestone none none branch branch author status new new resolution none none component core core keywords none none time changetime version none none owner none none
0
15,767
3,974,562,812
IssuesEvent
2016-05-04 22:46:02
LigaData/Kamanja
https://api.github.com/repos/LigaData/Kamanja
closed
The documentation for creating I/O adapters on the website is not complete.
Bug (Documentation) P3 Verify
We have a document that can be put there. Some examples are necessary, as well.
1.0
The documentation for creating I/O adapters on the website is not complete. - We have a document that can be put there. Some examples are necessary, as well.
non_test
the documentation for creating i o adapters on the website is not complete we have a document that can be put there some examples are necessary as well
0
261,929
22,781,153,799
IssuesEvent
2022-07-08 19:55:55
NuGet/Home
https://api.github.com/repos/NuGet/Home
closed
Remove temporary Newtonsoft.Json workaround in dotnet integration tests
Type:Engineering Type:Test Pipeline:Backlog
Our dotnet integration tests copy a working .NET SDK folder, then patch the NuGet assemblies, and use this copy/patched .NET SDK to run tests. We're upgrading to a version of newtonsoft.json higher than what the .NET SDK currently has (we're both upgrading at the same time), but until our CI pipeline start using builds of .NET SDK with a high enough version, we need to copy it ourselves. Here's what needs to be removed: https://github.com/NuGet/NuGet.Client/pull/4167/files/0e2f6a70900839b1f8670f8fd471d36bf1974df0#diff-8a81e906729d9dc1d11f635564b6eed54b859ab8aaacb3713fc29b19189f0f3b
1.0
Remove temporary Newtonsoft.Json workaround in dotnet integration tests - Our dotnet integration tests copy a working .NET SDK folder, then patch the NuGet assemblies, and use this copy/patched .NET SDK to run tests. We're upgrading to a version of newtonsoft.json higher than what the .NET SDK currently has (we're both upgrading at the same time), but until our CI pipeline start using builds of .NET SDK with a high enough version, we need to copy it ourselves. Here's what needs to be removed: https://github.com/NuGet/NuGet.Client/pull/4167/files/0e2f6a70900839b1f8670f8fd471d36bf1974df0#diff-8a81e906729d9dc1d11f635564b6eed54b859ab8aaacb3713fc29b19189f0f3b
test
remove temporary newtonsoft json workaround in dotnet integration tests our dotnet integration tests copy a working net sdk folder then patch the nuget assemblies and use this copy patched net sdk to run tests we re upgrading to a version of newtonsoft json higher than what the net sdk currently has we re both upgrading at the same time but until our ci pipeline start using builds of net sdk with a high enough version we need to copy it ourselves here s what needs to be removed
1
708,683
24,350,088,695
IssuesEvent
2022-10-02 20:47:38
IAmTamal/Milan
https://api.github.com/repos/IAmTamal/Milan
closed
[DOCS] Readme + License changes 🛠
📄 aspect: text ✨ goal: improvement 🟨 priority: medium 🛠 status : under development hacktoberfest
### Description Hello! I would like to help with the ReadMe documentation for the Milan project by making changes in the ReadMe for grammar and sentence formation as well as would like to brainstorm for anything that can be added that can make the project better along with linking the License in the file as well! Would enjoy if you can assign this issue to me for hacktoberfest! Thank you! ### Screenshots _No response_ ### Additional information _No response_
1.0
[DOCS] Readme + License changes 🛠 - ### Description Hello! I would like to help with the ReadMe documentation for the Milan project by making changes in the ReadMe for grammar and sentence formation as well as would like to brainstorm for anything that can be added that can make the project better along with linking the License in the file as well! Would enjoy if you can assign this issue to me for hacktoberfest! Thank you! ### Screenshots _No response_ ### Additional information _No response_
non_test
readme license changes 🛠 description hello i would like to help with the readme documentation for the milan project by making changes in the readme for grammar and sentence formation as well as would like to brainstorm for anything that can be added that can make the project better along with linking the license in the file as well would enjoy if you can assign this issue to me for hacktoberfest thank you screenshots no response additional information no response
0
22,559
11,746,204,837
IssuesEvent
2020-03-12 11:14:58
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
No mention of the Docker Compose YAML 4000 character limit
Pri2 app-service/svc cxp doc-enhancement triaged
As discussed here https://social.msdn.microsoft.com/Forums/azure/en-US/ff353717-fcb0-42d8-8237-5891e998c1d2/error-on-creating-web-app-with-docker-compose?forum=windowsazurewebsitespreview . This limits quite basic docker compose setups. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: ad0cae44-bfc2-c879-d406-23ab0b756ab9 * Version Independent ID: 5d812d22-559c-2c9f-7676-c62d7f5c980a * Content: [Quickstart: Create a multi-container app - Azure App Service](https://docs.microsoft.com/en-us/azure/app-service/containers/quickstart-multi-container#feedback) * Content Source: [articles/app-service/containers/quickstart-multi-container.md](https://github.com/Microsoft/azure-docs/blob/master/articles/app-service/containers/quickstart-multi-container.md) * Service: **app-service** * GitHub Login: @msangapu-msft * Microsoft Alias: **msangapu**
1.0
No mention of the Docker Compose YAML 4000 character limit - As discussed here https://social.msdn.microsoft.com/Forums/azure/en-US/ff353717-fcb0-42d8-8237-5891e998c1d2/error-on-creating-web-app-with-docker-compose?forum=windowsazurewebsitespreview . This limits quite basic docker compose setups. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: ad0cae44-bfc2-c879-d406-23ab0b756ab9 * Version Independent ID: 5d812d22-559c-2c9f-7676-c62d7f5c980a * Content: [Quickstart: Create a multi-container app - Azure App Service](https://docs.microsoft.com/en-us/azure/app-service/containers/quickstart-multi-container#feedback) * Content Source: [articles/app-service/containers/quickstart-multi-container.md](https://github.com/Microsoft/azure-docs/blob/master/articles/app-service/containers/quickstart-multi-container.md) * Service: **app-service** * GitHub Login: @msangapu-msft * Microsoft Alias: **msangapu**
non_test
no mention of the docker compose yaml character limit as discussed here this limits quite basic docker compose setups document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service app service github login msangapu msft microsoft alias msangapu
0
61,328
6,733,262,256
IssuesEvent
2017-10-18 14:20:15
ThePenguin1140/OpenCVObstructionTracking
https://api.github.com/repos/ThePenguin1140/OpenCVObstructionTracking
opened
Split video and upload pieces
help wanted testing
I created a test video as part of #4 but it's too big to put on github so I think the best thing might be split it into a few different 'test scenarios' and then upload the pieces? The video file can be found here: https://drive.google.com/open?id=0B55XddbN7M0zRG5PTE5VdzhIRzA
1.0
Split video and upload pieces - I created a test video as part of #4 but it's too big to put on github so I think the best thing might be split it into a few different 'test scenarios' and then upload the pieces? The video file can be found here: https://drive.google.com/open?id=0B55XddbN7M0zRG5PTE5VdzhIRzA
test
split video and upload pieces i created a test video as part of but it s too big to put on github so i think the best thing might be split it into a few different test scenarios and then upload the pieces the video file can be found here
1
301,882
26,107,190,512
IssuesEvent
2022-12-27 14:40:52
QubesOS/updates-status
https://api.github.com/repos/QubesOS/updates-status
closed
core-agent-linux v4.2.3 (r4.2)
r4.2-vm-bullseye-cur-test r4.2-vm-bookworm-cur-test r4.2-vm-fc37-cur-test r4.2-vm-fc36-cur-test r4.2-vm-centos-stream8-cur-test
Update of core-agent-linux to v4.2.3 for Qubes r4.2, see comments below for details and build status. From commit: https://github.com/QubesOS/qubes-core-agent-linux/commit/e248fae8e4b7cc89dc6f1b26e68b114fabbdf78a [Changes since previous version](https://github.com/QubesOS/qubes-core-agent-linux/compare/v4.2.2...v4.2.3): QubesOS/qubes-core-agent-linux@e248fae8 version 4.2.3 QubesOS/qubes-core-agent-linux@f2bd5c5e firewall: remove debug print QubesOS/qubes-core-agent-linux@f7f7a026 Merge remote-tracking branch 'origin/pr/399' QubesOS/qubes-core-agent-linux@765661af ci: fix uploading coverage to codecov QubesOS/qubes-core-agent-linux@f2db11ae archlinux: update example repo to r4.2 too QubesOS/qubes-core-agent-linux@90478b0b Revert "temporarily pretend to be 4.1" QubesOS/qubes-core-agent-linux@292a8ac1 Add purging of no longer allowed connections from conntrack QubesOS/qubes-core-agent-linux@119eb3ac qubes-rpc/nautilus: Execute external commands asynchronously QubesOS/qubes-core-agent-linux@0f7f0d6f qubes-rpc/nautilus: Add support for Nautilus API 4.0 The get_file_items method of Nautilus.MenuProvider no longer take the window argument. Referenced issues: QubesOS/qubes-issues#7916 QubesOS/qubes-issues#4141 If you're release manager, you can issue GPG-inline signed command: * `Upload-component r4.2 core-agent-linux e248fae8e4b7cc89dc6f1b26e68b114fabbdf78a current all` (available 5 days from now) * `Upload-component r4.2 core-agent-linux e248fae8e4b7cc89dc6f1b26e68b114fabbdf78a security-testing` You can choose subset of distributions like: * `Upload-component r4.2 core-agent-linux e248fae8e4b7cc89dc6f1b26e68b114fabbdf78a current vm-bookworm,vm-fc37` (available 5 days from now) Above commands will work only if packages in current-testing repository were built from given commit (i.e. no new version superseded it). For more information on how to test this update, please take a look at https://www.qubes-os.org/doc/testing/#updates.
5.0
core-agent-linux v4.2.3 (r4.2) - Update of core-agent-linux to v4.2.3 for Qubes r4.2, see comments below for details and build status. From commit: https://github.com/QubesOS/qubes-core-agent-linux/commit/e248fae8e4b7cc89dc6f1b26e68b114fabbdf78a [Changes since previous version](https://github.com/QubesOS/qubes-core-agent-linux/compare/v4.2.2...v4.2.3): QubesOS/qubes-core-agent-linux@e248fae8 version 4.2.3 QubesOS/qubes-core-agent-linux@f2bd5c5e firewall: remove debug print QubesOS/qubes-core-agent-linux@f7f7a026 Merge remote-tracking branch 'origin/pr/399' QubesOS/qubes-core-agent-linux@765661af ci: fix uploading coverage to codecov QubesOS/qubes-core-agent-linux@f2db11ae archlinux: update example repo to r4.2 too QubesOS/qubes-core-agent-linux@90478b0b Revert "temporarily pretend to be 4.1" QubesOS/qubes-core-agent-linux@292a8ac1 Add purging of no longer allowed connections from conntrack QubesOS/qubes-core-agent-linux@119eb3ac qubes-rpc/nautilus: Execute external commands asynchronously QubesOS/qubes-core-agent-linux@0f7f0d6f qubes-rpc/nautilus: Add support for Nautilus API 4.0 The get_file_items method of Nautilus.MenuProvider no longer take the window argument. Referenced issues: QubesOS/qubes-issues#7916 QubesOS/qubes-issues#4141 If you're release manager, you can issue GPG-inline signed command: * `Upload-component r4.2 core-agent-linux e248fae8e4b7cc89dc6f1b26e68b114fabbdf78a current all` (available 5 days from now) * `Upload-component r4.2 core-agent-linux e248fae8e4b7cc89dc6f1b26e68b114fabbdf78a security-testing` You can choose subset of distributions like: * `Upload-component r4.2 core-agent-linux e248fae8e4b7cc89dc6f1b26e68b114fabbdf78a current vm-bookworm,vm-fc37` (available 5 days from now) Above commands will work only if packages in current-testing repository were built from given commit (i.e. no new version superseded it). For more information on how to test this update, please take a look at https://www.qubes-os.org/doc/testing/#updates.
test
core agent linux update of core agent linux to for qubes see comments below for details and build status from commit qubesos qubes core agent linux version qubesos qubes core agent linux firewall remove debug print qubesos qubes core agent linux merge remote tracking branch origin pr qubesos qubes core agent linux ci fix uploading coverage to codecov qubesos qubes core agent linux archlinux update example repo to too qubesos qubes core agent linux revert temporarily pretend to be qubesos qubes core agent linux add purging of no longer allowed connections from conntrack qubesos qubes core agent linux qubes rpc nautilus execute external commands asynchronously qubesos qubes core agent linux qubes rpc nautilus add support for nautilus api the get file items method of nautilus menuprovider no longer take the window argument referenced issues qubesos qubes issues qubesos qubes issues if you re release manager you can issue gpg inline signed command upload component core agent linux current all available days from now upload component core agent linux security testing you can choose subset of distributions like upload component core agent linux current vm bookworm vm available days from now above commands will work only if packages in current testing repository were built from given commit i e no new version superseded it for more information on how to test this update please take a look at
1
325,575
27,944,834,556
IssuesEvent
2023-03-24 01:34:24
WiIIiam278/HuskHomes2
https://api.github.com/repos/WiIIiam278/HuskHomes2
closed
Teleport strange behaviour with tabcomplete name event
status: needs testing
Version: 3.0.4 from [spigot](https://www.spigotmc.org/resources/%E2%AD%90-huskhomes-1-16-1-19-%E2%AD%90-simple-intuitive-teleportation-suite-with-cross-server-support.83767/updates) Using REDIS as messanger_type. With this messenger_type we found a strange behaviour in tabcompleting names. 1. /tp and /tpa are ok if you are in two separate servers 2. /tp and /tpa player list is taken from the global player list instead of the player list from the servers with REDIS cache (I think it's normal because of this ` public CompletableFuture<List<String>> updatePlayerListCache(@NotNull HuskHomes plugin, @NotNull OnlineUser requester) { if (plugin.getSettings().crossServer) { return plugin.getNetworkMessenger().getOnlinePlayerNames(requester).thenApply(returnedPlayerList -> { players.clear(); players.addAll(List.of(returnedPlayerList)); return players; }); } else { players.clear(); players.addAll(plugin.getOnlinePlayers() .stream() .filter(player -> !player.isVanished()) .map(onlineUser -> onlineUser.username) .toList()); System.out.println(String.join(", ", players)); return CompletableFuture.completedFuture(players); } }` 3. The biggest problem comes when you are both in the same server: /tp and /tpa doesn't tab complete the player names in the same server. Only if you first write his name and teleport to him, the second time both commands work as defined.
1.0
Teleport strange behaviour with tabcomplete name event - Version: 3.0.4 from [spigot](https://www.spigotmc.org/resources/%E2%AD%90-huskhomes-1-16-1-19-%E2%AD%90-simple-intuitive-teleportation-suite-with-cross-server-support.83767/updates) Using REDIS as messanger_type. With this messenger_type we found a strange behaviour in tabcompleting names. 1. /tp and /tpa are ok if you are in two separate servers 2. /tp and /tpa player list is taken from the global player list instead of the player list from the servers with REDIS cache (I think it's normal because of this ` public CompletableFuture<List<String>> updatePlayerListCache(@NotNull HuskHomes plugin, @NotNull OnlineUser requester) { if (plugin.getSettings().crossServer) { return plugin.getNetworkMessenger().getOnlinePlayerNames(requester).thenApply(returnedPlayerList -> { players.clear(); players.addAll(List.of(returnedPlayerList)); return players; }); } else { players.clear(); players.addAll(plugin.getOnlinePlayers() .stream() .filter(player -> !player.isVanished()) .map(onlineUser -> onlineUser.username) .toList()); System.out.println(String.join(", ", players)); return CompletableFuture.completedFuture(players); } }` 3. The biggest problem comes when you are both in the same server: /tp and /tpa doesn't tab complete the player names in the same server. Only if you first write his name and teleport to him, the second time both commands work as defined.
test
teleport strange behaviour with tabcomplete name event version from using redis as messanger type with this messenger type we found a strange behaviour in tabcompleting names tp and tpa are ok if you are in two separate servers tp and tpa player list is taken from the global player list instead of the player list from the servers with redis cache i think it s normal because of this public completablefuture updateplayerlistcache notnull huskhomes plugin notnull onlineuser requester if plugin getsettings crossserver return plugin getnetworkmessenger getonlineplayernames requester thenapply returnedplayerlist players clear players addall list of returnedplayerlist return players else players clear players addall plugin getonlineplayers stream filter player player isvanished map onlineuser onlineuser username tolist system out println string join players return completablefuture completedfuture players the biggest problem comes when you are both in the same server tp and tpa doesn t tab complete the player names in the same server only if you first write his name and teleport to him the second time both commands work as defined
1
228,382
18,173,514,805
IssuesEvent
2021-09-27 22:55:09
microsoft/vscode
https://api.github.com/repos/microsoft/vscode
opened
Test Autocomplete Preselection Powered By Inline Suggestions
testplan-item
Refs: https://github.com/microsoft/vscode/issues/131940 - [ ] anyOS - [ ] anyOS Complexity: 3 [Create Issue](https://github.com/microsoft/vscode/issues/new?body=Testing+%23133906%0A%0A&assignees=hediet) --- # Preparation * Enable suggest preview: `"editor.suggest.preview": true,`. * Install Copilot # Tasks * Create a TypeScript file and write `// Write an error to the console` followed by a new line and `console` to prompt copilot. * Type `.` to trigger autocomplete. * Verify that the inline suggestion stays and becomes non-italic * Verify that an autocomplete item is preselected that is a prefix of the inline suggestion
1.0
Test Autocomplete Preselection Powered By Inline Suggestions - Refs: https://github.com/microsoft/vscode/issues/131940 - [ ] anyOS - [ ] anyOS Complexity: 3 [Create Issue](https://github.com/microsoft/vscode/issues/new?body=Testing+%23133906%0A%0A&assignees=hediet) --- # Preparation * Enable suggest preview: `"editor.suggest.preview": true,`. * Install Copilot # Tasks * Create a TypeScript file and write `// Write an error to the console` followed by a new line and `console` to prompt copilot. * Type `.` to trigger autocomplete. * Verify that the inline suggestion stays and becomes non-italic * Verify that an autocomplete item is preselected that is a prefix of the inline suggestion
test
test autocomplete preselection powered by inline suggestions refs anyos anyos complexity preparation enable suggest preview editor suggest preview true install copilot tasks create a typescript file and write write an error to the console followed by a new line and console to prompt copilot type to trigger autocomplete verify that the inline suggestion stays and becomes non italic verify that an autocomplete item is preselected that is a prefix of the inline suggestion
1
13,624
3,350,354,496
IssuesEvent
2015-11-17 14:24:25
servo/servo
https://api.github.com/repos/servo/servo
closed
Look through old layout bugs and write reftests
A-testing E-less easy
The reftest suite was neglected for a while. We should have more regression tests for bugs we fixed during that time.
1.0
Look through old layout bugs and write reftests - The reftest suite was neglected for a while. We should have more regression tests for bugs we fixed during that time.
test
look through old layout bugs and write reftests the reftest suite was neglected for a while we should have more regression tests for bugs we fixed during that time
1
339,802
30,476,929,857
IssuesEvent
2023-07-17 17:12:59
rancher/highlander
https://api.github.com/repos/rancher/highlander
opened
Migration of clusters between different Rancher instances
area/testing area/eks kind/qa area/hosted-providers
The Rancher migration procedure (migrating clusters to a 'replacement' Rancher) needs better testing, esp. for regressions. Apparently migration for EKS clusters behaves differently compared to other downstream clusters.
1.0
Migration of clusters between different Rancher instances - The Rancher migration procedure (migrating clusters to a 'replacement' Rancher) needs better testing, esp. for regressions. Apparently migration for EKS clusters behaves differently compared to other downstream clusters.
test
migration of clusters between different rancher instances the rancher migration procedure migrating clusters to a replacement rancher needs better testing esp for regressions apparently migration for eks clusters behaves differently compared to other downstream clusters
1
101,241
8,782,994,040
IssuesEvent
2018-12-20 03:10:35
EasyRPG/Player
https://api.github.com/repos/EasyRPG/Player
closed
Pictures won't load when loading savefile in a project with title screen disabled.
Patch available Testcase available Window/Scenes
When you have a RM2k3 >=1.10 project without Ttitle screen enabled in database, when you try to load a savefile with pictures in game, these won't are being loaded (and didn't shown in screen). However, if you enable Title Screen in Database editor, the images are shown without issue. #### Player platform: Tested in 0.5.4 and Master (Dec 15th). #### Test case Here's a test case: [imgBug.zip](https://github.com/EasyRPG/Player/files/2683010/imgBug.zip) Web: https://easyrpg.org/play/master/?game=issue-1571&engine=rpg2k3e (ignore the missing FaceSets) A RM2k3 1.12a project with title screen disabled. It goes directly to a almost empty map with a automatic event that shows a dummy image, then it deletes itself, and a helper event for save/load calling. - Save the current progress (with debug features or with the helper event). - Then load the savefile. #### Expected behaviour (RPG_RT 1.12a): The dummy image is being shown when the map loads. #### Current behaviour (Player 0.5.4 , Master): No image is shown. Talking with @fdelapena , maybe this bug can be related with https://github.com/EasyRPG/Player/issues/1524
1.0
Pictures won't load when loading savefile in a project with title screen disabled. - When you have a RM2k3 >=1.10 project without Ttitle screen enabled in database, when you try to load a savefile with pictures in game, these won't are being loaded (and didn't shown in screen). However, if you enable Title Screen in Database editor, the images are shown without issue. #### Player platform: Tested in 0.5.4 and Master (Dec 15th). #### Test case Here's a test case: [imgBug.zip](https://github.com/EasyRPG/Player/files/2683010/imgBug.zip) Web: https://easyrpg.org/play/master/?game=issue-1571&engine=rpg2k3e (ignore the missing FaceSets) A RM2k3 1.12a project with title screen disabled. It goes directly to a almost empty map with a automatic event that shows a dummy image, then it deletes itself, and a helper event for save/load calling. - Save the current progress (with debug features or with the helper event). - Then load the savefile. #### Expected behaviour (RPG_RT 1.12a): The dummy image is being shown when the map loads. #### Current behaviour (Player 0.5.4 , Master): No image is shown. Talking with @fdelapena , maybe this bug can be related with https://github.com/EasyRPG/Player/issues/1524
test
pictures won t load when loading savefile in a project with title screen disabled when you have a project without ttitle screen enabled in database when you try to load a savefile with pictures in game these won t are being loaded and didn t shown in screen however if you enable title screen in database editor the images are shown without issue player platform tested in and master dec test case here s a test case web ignore the missing facesets a project with title screen disabled it goes directly to a almost empty map with a automatic event that shows a dummy image then it deletes itself and a helper event for save load calling save the current progress with debug features or with the helper event then load the savefile expected behaviour rpg rt the dummy image is being shown when the map loads current behaviour player master no image is shown talking with fdelapena maybe this bug can be related with
1
67,391
12,952,869,085
IssuesEvent
2020-07-19 22:15:20
eucalypto/eucalyptapp
https://api.github.com/repos/eucalypto/eucalyptapp
opened
Replace DataBindingUtil in GameWonFragment
code enhancement
The use of DataBindingUtil is deprecated and should be replaced by the generated Class for the specific Fragment.
1.0
Replace DataBindingUtil in GameWonFragment - The use of DataBindingUtil is deprecated and should be replaced by the generated Class for the specific Fragment.
non_test
replace databindingutil in gamewonfragment the use of databindingutil is deprecated and should be replaced by the generated class for the specific fragment
0
62,241
3,179,650,107
IssuesEvent
2015-09-25 03:21:07
cjfields/redmine-test
https://api.github.com/repos/cjfields/redmine-test
opened
Bio::Seq object loses sequence data when blessed as Bio::Seq::Meta::Array
Category: Core Components Priority: Normal Status: New Tracker: Bug
--- Author Name: **Roy Chaudhuri** (Roy Chaudhuri) Original Redmine Issue: 2262, https://redmine.open-bio.org/issues/2262 Original Date: 2007-04-04 Original Assignee: Bioperl Guts --- When I bless a Bio::Seq object as a Bio::Seq::Meta::Array (as instructed by the POD) it loses the primary sequence information (and other PrimarySeqI information such as length and accession). Features and annotation are not affected. Bio::Seq::Meta suffers from a worse bug- Bio::SeqIO warns that the Bio::Seq::Meta object is not SeqI compliant. This seems to be due to the omission of Bio::Seq from the use base line in Bio::Seq::Meta (but not Bio::Seq::Meta::Array). When I add Bio::Seq into the use base line the behaviour is the same as for Bio::Seq::Meta::Array. The following code demonstrates the problem: \#!/usr/bin/perl use warnings; use strict; use Bio::SeqIO; use Bio::Seq::Meta::Array; my $seq=Bio::SeqIO-\>new(-fh=\>\\\*ARGV, ~~format=\>’genbank’)~~\>next\_seq; bless $seq, ‘Bio::Seq::Meta::Array’; Bio::SeqIO-\>new(~~format=\>’genbank’)~~\>write\_seq($seq);
1.0
Bio::Seq object loses sequence data when blessed as Bio::Seq::Meta::Array - --- Author Name: **Roy Chaudhuri** (Roy Chaudhuri) Original Redmine Issue: 2262, https://redmine.open-bio.org/issues/2262 Original Date: 2007-04-04 Original Assignee: Bioperl Guts --- When I bless a Bio::Seq object as a Bio::Seq::Meta::Array (as instructed by the POD) it loses the primary sequence information (and other PrimarySeqI information such as length and accession). Features and annotation are not affected. Bio::Seq::Meta suffers from a worse bug- Bio::SeqIO warns that the Bio::Seq::Meta object is not SeqI compliant. This seems to be due to the omission of Bio::Seq from the use base line in Bio::Seq::Meta (but not Bio::Seq::Meta::Array). When I add Bio::Seq into the use base line the behaviour is the same as for Bio::Seq::Meta::Array. The following code demonstrates the problem: \#!/usr/bin/perl use warnings; use strict; use Bio::SeqIO; use Bio::Seq::Meta::Array; my $seq=Bio::SeqIO-\>new(-fh=\>\\\*ARGV, ~~format=\>’genbank’)~~\>next\_seq; bless $seq, ‘Bio::Seq::Meta::Array’; Bio::SeqIO-\>new(~~format=\>’genbank’)~~\>write\_seq($seq);
non_test
bio seq object loses sequence data when blessed as bio seq meta array author name roy chaudhuri roy chaudhuri original redmine issue original date original assignee bioperl guts when i bless a bio seq object as a bio seq meta array as instructed by the pod it loses the primary sequence information and other primaryseqi information such as length and accession features and annotation are not affected bio seq meta suffers from a worse bug bio seqio warns that the bio seq meta object is not seqi compliant this seems to be due to the omission of bio seq from the use base line in bio seq meta but not bio seq meta array when i add bio seq into the use base line the behaviour is the same as for bio seq meta array the following code demonstrates the problem usr bin perl use warnings use strict use bio seqio use bio seq meta array my seq bio seqio new fh argv format ’genbank’ next seq bless seq ‘bio seq meta array’ bio seqio new format ’genbank’ write seq seq
0
390,436
11,543,724,741
IssuesEvent
2020-02-18 10:07:15
wso2/product-apim
https://api.github.com/repos/wso2/product-apim
opened
Issue in invoking SOAP APIs
Priority/Normal Type/Bug
### Description: Following issue occurred when invoking Pass through API created using SOAP back-end. 1. NPE when invoking: ``` ERROR - ServerWorker Error processing POST request for : /through/1. java.lang.NullPointerException: null at org.wso2.carbon.apimgt.gateway.utils.OpenAPIUtils.getPathItemSecurityScopes_aroundBody10(OpenAPIUtils.java:205) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?] at org.wso2.carbon.apimgt.gateway.utils.OpenAPIUtils.getPathItemSecurityScopes(OpenAPIUtils.java:191) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?] at org.wso2.carbon.apimgt.gateway.utils.OpenAPIUtils.getScopesOfResource_aroundBody2(OpenAPIUtils.java:77) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?] at org.wso2.carbon.apimgt.gateway.utils.OpenAPIUtils.getScopesOfResource(OpenAPIUtils.java:71) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?] at org.wso2.carbon.apimgt.gateway.handlers.security.jwt.JWTValidator.validateScopes_aroundBody8(JWTValidator.java:542) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?] at org.wso2.carbon.apimgt.gateway.handlers.security.jwt.JWTValidator.validateScopes(JWTValidator.java:530) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?] at org.wso2.carbon.apimgt.gateway.handlers.security.jwt.JWTValidator.authenticate_aroundBody0(JWTValidator.java:230) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?] at org.wso2.carbon.apimgt.gateway.handlers.security.jwt.JWTValidator.authenticate(JWTValidator.java:91) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?] at org.wso2.carbon.apimgt.gateway.handlers.security.oauth.OAuthAuthenticator.authenticate_aroundBody4(OAuthAuthenticator.java:333) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?] at org.wso2.carbon.apimgt.gateway.handlers.security.oauth.OAuthAuthenticator.authenticate(OAuthAuthenticator.java:109) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?] at org.wso2.carbon.apimgt.gateway.handlers.security.APIAuthenticationHandler.isAuthenticate_aroundBody42(APIAuthenticationHandler.java:419) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?] at org.wso2.carbon.apimgt.gateway.handlers.security.APIAuthenticationHandler.isAuthenticate(APIAuthenticationHandler.java:413) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?] at org.wso2.carbon.apimgt.gateway.handlers.security.APIAuthenticationHandler.handleRequest_aroundBody36(APIAuthenticationHandler.java:349) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?] at org.wso2.carbon.apimgt.gateway.handlers.security.APIAuthenticationHandler.handleRequest(APIAuthenticationHandler.java:320) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?] at org.apache.synapse.rest.API.process(API.java:367) ~[synapse-core_2.1.7.wso2v143.jar:2.1.7-wso2v143] at org.apache.synapse.rest.RESTRequestHandler.apiProcessNonDefaultStrategy(RESTRequestHandler.java:149) ~[synapse-core_2.1.7.wso2v143.jar:2.1.7-wso2v143] at org.apache.synapse.rest.RESTRequestHandler.dispatchToAPI(RESTRequestHandler.java:95) ~[synapse-core_2.1.7.wso2v143.jar:2.1.7-wso2v143] at org.apache.synapse.rest.RESTRequestHandler.process(RESTRequestHandler.java:71) ~[synapse-core_2.1.7.wso2v143.jar:2.1.7-wso2v143] at org.apache.synapse.core.axis2.Axis2SynapseEnvironment.injectMessage(Axis2SynapseEnvironment.java:327) ~[synapse-core_2.1.7.wso2v143.jar:2.1.7-wso2v143] at org.apache.synapse.core.axis2.SynapseMessageReceiver.receive(SynapseMessageReceiver.java:98) ~[synapse-core_2.1.7.wso2v143.jar:2.1.7-wso2v143] at org.apache.axis2.engine.AxisEngine.receive(AxisEngine.java:180) ~[axis2_1.6.1.wso2v40.jar:?] at org.apache.synapse.transport.passthru.ServerWorker.processNonEntityEnclosingRESTHandler(ServerWorker.java:368) [synapse-nhttp-transport_2.1.7.wso2v143.jar:?] at org.apache.synapse.transport.passthru.ServerWorker.processEntityEnclosingRequest(ServerWorker.java:427) [synapse-nhttp-transport_2.1.7.wso2v143.jar:?] at org.apache.synapse.transport.passthru.ServerWorker.run(ServerWorker.java:182) [synapse-nhttp-transport_2.1.7.wso2v143.jar:?] at org.apache.axis2.transport.base.threads.NativeWorkerPool$1.run(NativeWorkerPool.java:172) [axis2_1.6.1.wso2v40.jar:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) [?:1.8.0_201] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) [?:1.8.0_201] at java.lang.Thread.run(Thread.java:748) [?:1.8.0_201] ``` 2. Generated swagger does not have the content-type of input (The default content type should be **Text/xml**). Hence, it is not possible to select the content-type in Swagger UI in Dev Portal. Refer the following screenshot: ![soap](https://user-images.githubusercontent.com/19728269/74725705-453c2c00-5264-11ea-9207-608d22754ed2.png)
1.0
Issue in invoking SOAP APIs - ### Description: Following issue occurred when invoking Pass through API created using SOAP back-end. 1. NPE when invoking: ``` ERROR - ServerWorker Error processing POST request for : /through/1. java.lang.NullPointerException: null at org.wso2.carbon.apimgt.gateway.utils.OpenAPIUtils.getPathItemSecurityScopes_aroundBody10(OpenAPIUtils.java:205) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?] at org.wso2.carbon.apimgt.gateway.utils.OpenAPIUtils.getPathItemSecurityScopes(OpenAPIUtils.java:191) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?] at org.wso2.carbon.apimgt.gateway.utils.OpenAPIUtils.getScopesOfResource_aroundBody2(OpenAPIUtils.java:77) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?] at org.wso2.carbon.apimgt.gateway.utils.OpenAPIUtils.getScopesOfResource(OpenAPIUtils.java:71) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?] at org.wso2.carbon.apimgt.gateway.handlers.security.jwt.JWTValidator.validateScopes_aroundBody8(JWTValidator.java:542) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?] at org.wso2.carbon.apimgt.gateway.handlers.security.jwt.JWTValidator.validateScopes(JWTValidator.java:530) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?] at org.wso2.carbon.apimgt.gateway.handlers.security.jwt.JWTValidator.authenticate_aroundBody0(JWTValidator.java:230) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?] at org.wso2.carbon.apimgt.gateway.handlers.security.jwt.JWTValidator.authenticate(JWTValidator.java:91) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?] at org.wso2.carbon.apimgt.gateway.handlers.security.oauth.OAuthAuthenticator.authenticate_aroundBody4(OAuthAuthenticator.java:333) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?] at org.wso2.carbon.apimgt.gateway.handlers.security.oauth.OAuthAuthenticator.authenticate(OAuthAuthenticator.java:109) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?] at org.wso2.carbon.apimgt.gateway.handlers.security.APIAuthenticationHandler.isAuthenticate_aroundBody42(APIAuthenticationHandler.java:419) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?] at org.wso2.carbon.apimgt.gateway.handlers.security.APIAuthenticationHandler.isAuthenticate(APIAuthenticationHandler.java:413) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?] at org.wso2.carbon.apimgt.gateway.handlers.security.APIAuthenticationHandler.handleRequest_aroundBody36(APIAuthenticationHandler.java:349) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?] at org.wso2.carbon.apimgt.gateway.handlers.security.APIAuthenticationHandler.handleRequest(APIAuthenticationHandler.java:320) ~[org.wso2.carbon.apimgt.gateway_6.6.64.jar:?] at org.apache.synapse.rest.API.process(API.java:367) ~[synapse-core_2.1.7.wso2v143.jar:2.1.7-wso2v143] at org.apache.synapse.rest.RESTRequestHandler.apiProcessNonDefaultStrategy(RESTRequestHandler.java:149) ~[synapse-core_2.1.7.wso2v143.jar:2.1.7-wso2v143] at org.apache.synapse.rest.RESTRequestHandler.dispatchToAPI(RESTRequestHandler.java:95) ~[synapse-core_2.1.7.wso2v143.jar:2.1.7-wso2v143] at org.apache.synapse.rest.RESTRequestHandler.process(RESTRequestHandler.java:71) ~[synapse-core_2.1.7.wso2v143.jar:2.1.7-wso2v143] at org.apache.synapse.core.axis2.Axis2SynapseEnvironment.injectMessage(Axis2SynapseEnvironment.java:327) ~[synapse-core_2.1.7.wso2v143.jar:2.1.7-wso2v143] at org.apache.synapse.core.axis2.SynapseMessageReceiver.receive(SynapseMessageReceiver.java:98) ~[synapse-core_2.1.7.wso2v143.jar:2.1.7-wso2v143] at org.apache.axis2.engine.AxisEngine.receive(AxisEngine.java:180) ~[axis2_1.6.1.wso2v40.jar:?] at org.apache.synapse.transport.passthru.ServerWorker.processNonEntityEnclosingRESTHandler(ServerWorker.java:368) [synapse-nhttp-transport_2.1.7.wso2v143.jar:?] at org.apache.synapse.transport.passthru.ServerWorker.processEntityEnclosingRequest(ServerWorker.java:427) [synapse-nhttp-transport_2.1.7.wso2v143.jar:?] at org.apache.synapse.transport.passthru.ServerWorker.run(ServerWorker.java:182) [synapse-nhttp-transport_2.1.7.wso2v143.jar:?] at org.apache.axis2.transport.base.threads.NativeWorkerPool$1.run(NativeWorkerPool.java:172) [axis2_1.6.1.wso2v40.jar:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) [?:1.8.0_201] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) [?:1.8.0_201] at java.lang.Thread.run(Thread.java:748) [?:1.8.0_201] ``` 2. Generated swagger does not have the content-type of input (The default content type should be **Text/xml**). Hence, it is not possible to select the content-type in Swagger UI in Dev Portal. Refer the following screenshot: ![soap](https://user-images.githubusercontent.com/19728269/74725705-453c2c00-5264-11ea-9207-608d22754ed2.png)
non_test
issue in invoking soap apis description following issue occurred when invoking pass through api created using soap back end npe when invoking error serverworker error processing post request for through java lang nullpointerexception null at org carbon apimgt gateway utils openapiutils getpathitemsecurityscopes openapiutils java at org carbon apimgt gateway utils openapiutils getpathitemsecurityscopes openapiutils java at org carbon apimgt gateway utils openapiutils getscopesofresource openapiutils java at org carbon apimgt gateway utils openapiutils getscopesofresource openapiutils java at org carbon apimgt gateway handlers security jwt jwtvalidator validatescopes jwtvalidator java at org carbon apimgt gateway handlers security jwt jwtvalidator validatescopes jwtvalidator java at org carbon apimgt gateway handlers security jwt jwtvalidator authenticate jwtvalidator java at org carbon apimgt gateway handlers security jwt jwtvalidator authenticate jwtvalidator java at org carbon apimgt gateway handlers security oauth oauthauthenticator authenticate oauthauthenticator java at org carbon apimgt gateway handlers security oauth oauthauthenticator authenticate oauthauthenticator java at org carbon apimgt gateway handlers security apiauthenticationhandler isauthenticate apiauthenticationhandler java at org carbon apimgt gateway handlers security apiauthenticationhandler isauthenticate apiauthenticationhandler java at org carbon apimgt gateway handlers security apiauthenticationhandler handlerequest apiauthenticationhandler java at org carbon apimgt gateway handlers security apiauthenticationhandler handlerequest apiauthenticationhandler java at org apache synapse rest api process api java at org apache synapse rest restrequesthandler apiprocessnondefaultstrategy restrequesthandler java at org apache synapse rest restrequesthandler dispatchtoapi restrequesthandler java at org apache synapse rest restrequesthandler process restrequesthandler java at org apache synapse core injectmessage java at org apache synapse core synapsemessagereceiver receive synapsemessagereceiver java at org apache engine axisengine receive axisengine java at org apache synapse transport passthru serverworker processnonentityenclosingresthandler serverworker java at org apache synapse transport passthru serverworker processentityenclosingrequest serverworker java at org apache synapse transport passthru serverworker run serverworker java at org apache transport base threads nativeworkerpool run nativeworkerpool java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java generated swagger does not have the content type of input the default content type should be text xml hence it is not possible to select the content type in swagger ui in dev portal refer the following screenshot
0
258,038
22,272,201,756
IssuesEvent
2022-06-10 13:25:23
mozilla-mobile/fenix
https://api.github.com/repos/mozilla-mobile/fenix
opened
Intermittent UI test failure - TabbedBrowsingTest.closeTabTest
eng:ui-test
### Firebase Test Run: https://console.firebase.google.com/u/0/project/moz-fenix/testlab/histories/bh.66b7091e15d53d45/matrices/6696243506005561227/executions/bs.6f49200709fbc39c ### Stacktrace: `java.lang.RuntimeException: Error while connecting UiAutomation@3740472[id=-1, flags=0] at android.app.UiAutomation.connect(UiAutomation.java:259) at android.app.Instrumentation.getUiAutomation(Instrumentation.java:2176) at androidx.test.uiautomator.UiDevice.getUiAutomation(UiDevice.java:1129) at androidx.test.uiautomator.QueryController.<init>(QueryController.java:95) at androidx.test.uiautomator.UiDevice.<init>(UiDevice.java:109) at androidx.test.uiautomator.UiDevice.getInstance(UiDevice.java:261) at org.mozilla.fenix.ui.TabbedBrowsingTest.<init>(TabbedBrowsingTest.kt:43)` ### Build: on main 6/9 Known issue: https://github.com/mozilla-mobile/fenix/issues/25132
1.0
Intermittent UI test failure - TabbedBrowsingTest.closeTabTest - ### Firebase Test Run: https://console.firebase.google.com/u/0/project/moz-fenix/testlab/histories/bh.66b7091e15d53d45/matrices/6696243506005561227/executions/bs.6f49200709fbc39c ### Stacktrace: `java.lang.RuntimeException: Error while connecting UiAutomation@3740472[id=-1, flags=0] at android.app.UiAutomation.connect(UiAutomation.java:259) at android.app.Instrumentation.getUiAutomation(Instrumentation.java:2176) at androidx.test.uiautomator.UiDevice.getUiAutomation(UiDevice.java:1129) at androidx.test.uiautomator.QueryController.<init>(QueryController.java:95) at androidx.test.uiautomator.UiDevice.<init>(UiDevice.java:109) at androidx.test.uiautomator.UiDevice.getInstance(UiDevice.java:261) at org.mozilla.fenix.ui.TabbedBrowsingTest.<init>(TabbedBrowsingTest.kt:43)` ### Build: on main 6/9 Known issue: https://github.com/mozilla-mobile/fenix/issues/25132
test
intermittent ui test failure tabbedbrowsingtest closetabtest firebase test run stacktrace java lang runtimeexception error while connecting uiautomation at android app uiautomation connect uiautomation java at android app instrumentation getuiautomation instrumentation java at androidx test uiautomator uidevice getuiautomation uidevice java at androidx test uiautomator querycontroller querycontroller java at androidx test uiautomator uidevice uidevice java at androidx test uiautomator uidevice getinstance uidevice java at org mozilla fenix ui tabbedbrowsingtest tabbedbrowsingtest kt build on main known issue
1
174,691
13,505,085,402
IssuesEvent
2020-09-13 21:04:22
thexerteproject/xerteonlinetoolkits
https://api.github.com/repos/thexerteproject/xerteonlinetoolkits
closed
Refactor Xenith engine code
Needs testing New feature XOT template enhancement
Xenith engine has got rather huge and spaghetti like and it's really hard for me to follow (having not been around the code for a while) so it must be impossible for anyone new coming in and it's only going to get more difficult for us to support unless we try and modernise it, group related code and refactor. It is getting on 8 years old now! Anyway, i've run this past @FayCross a while ago but didn't have the time to implement any of it but with this self-isolation thing i've got all the time in the world!! To start i'm going to refactor: - Variables code (that's a good 600 lines alone) - Glossary - Menu - Dialog I'll leave them in xenith.js for now (xenith will get bigger just now but will eventually have the code blocks separated out to separate files) but I also have a loader which can load on the fly each of the blocks, when needed, but need to work on that to make sure it works offline and in scorm. I know there are probably other bugs and developments that need done but i'm hope this refactor will reduce the complication of tracing and fixing those issues. I've done VARIABLES and it's working fine as far as I can see. So i'll start committing them as separate commits and tag this issue so we can keep track of them. Hope you guys are all safe!
1.0
Refactor Xenith engine code - Xenith engine has got rather huge and spaghetti like and it's really hard for me to follow (having not been around the code for a while) so it must be impossible for anyone new coming in and it's only going to get more difficult for us to support unless we try and modernise it, group related code and refactor. It is getting on 8 years old now! Anyway, i've run this past @FayCross a while ago but didn't have the time to implement any of it but with this self-isolation thing i've got all the time in the world!! To start i'm going to refactor: - Variables code (that's a good 600 lines alone) - Glossary - Menu - Dialog I'll leave them in xenith.js for now (xenith will get bigger just now but will eventually have the code blocks separated out to separate files) but I also have a loader which can load on the fly each of the blocks, when needed, but need to work on that to make sure it works offline and in scorm. I know there are probably other bugs and developments that need done but i'm hope this refactor will reduce the complication of tracing and fixing those issues. I've done VARIABLES and it's working fine as far as I can see. So i'll start committing them as separate commits and tag this issue so we can keep track of them. Hope you guys are all safe!
test
refactor xenith engine code xenith engine has got rather huge and spaghetti like and it s really hard for me to follow having not been around the code for a while so it must be impossible for anyone new coming in and it s only going to get more difficult for us to support unless we try and modernise it group related code and refactor it is getting on years old now anyway i ve run this past faycross a while ago but didn t have the time to implement any of it but with this self isolation thing i ve got all the time in the world to start i m going to refactor variables code that s a good lines alone glossary menu dialog i ll leave them in xenith js for now xenith will get bigger just now but will eventually have the code blocks separated out to separate files but i also have a loader which can load on the fly each of the blocks when needed but need to work on that to make sure it works offline and in scorm i know there are probably other bugs and developments that need done but i m hope this refactor will reduce the complication of tracing and fixing those issues i ve done variables and it s working fine as far as i can see so i ll start committing them as separate commits and tag this issue so we can keep track of them hope you guys are all safe
1
202,484
15,286,694,431
IssuesEvent
2021-02-23 14:58:55
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachtest: kv95/enc=false/nodes=3/cpu=32/seq failed
C-test-failure O-roachtest O-robot branch-release-20.1 release-blocker
[(roachtest).kv95/enc=false/nodes=3/cpu=32/seq failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2661561&tab=buildLog) on [release-20.1@90f78268f3b5b08ba838ac3ad164821d2f5a5362](https://github.com/cockroachdb/cockroach/commits/90f78268f3b5b08ba838ac3ad164821d2f5a5362): ``` | github.com/cockroachdb/cockroach/vendor/golang.org/x/sync/errgroup.(*Group).Go.func1 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/golang.org/x/sync/errgroup/errgroup.go:57 | runtime.goexit | /usr/local/go/src/runtime/asm_amd64.s:1374 Wraps: (2) 2 safe details enclosed Wraps: (3) output in run_074156.709_n4_workload_run_kv Wraps: (4) /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-2661561-1612941349-20-n4cpu32:4 -- ./workload run kv --init --histograms=perf/stats.json --concurrency=192 --duration=10m0s --read-percent=95 --sequential {pgurl:1-3} returned | stderr: | ./workload: error while loading shared libraries: libncurses.so.6: cannot open shared object file: No such file or directory | Error: COMMAND_PROBLEM: exit status 127 | (1) COMMAND_PROBLEM | Wraps: (2) Node 4. Command with error: | | ``` | | ./workload run kv --init --histograms=perf/stats.json --concurrency=192 --duration=10m0s --read-percent=95 --sequential {pgurl:1-3} | | ``` | Wraps: (3) exit status 127 | Error types: (1) errors.Cmd (2) *hintdetail.withDetail (3) *exec.ExitError | | stdout: Wraps: (5) exit status 20 Error types: (1) *withstack.withStack (2) *safedetails.withSafeDetails (3) *errutil.withMessage (4) *main.withCommandDetails (5) *exec.ExitError cluster.go:2628,kv.go:96,kv.go:183,test_runner.go:749: monitor failure: monitor task failed: t.Fatal() was called (1) attached stack trace | main.(*monitor).WaitE | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2616 | main.(*monitor).Wait | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2624 | main.registerKV.func2 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/kv.go:96 | main.registerKV.func3 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/kv.go:183 | main.(*testRunner).runTest.func2 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/test_runner.go:749 Wraps: (2) monitor failure Wraps: (3) attached stack trace | main.(*monitor).wait.func2 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2672 Wraps: (4) monitor task failed Wraps: (5) attached stack trace | main.init | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2586 | runtime.doInit | /usr/local/go/src/runtime/proc.go:5652 | runtime.main | /usr/local/go/src/runtime/proc.go:191 | runtime.goexit | /usr/local/go/src/runtime/asm_amd64.s:1374 Wraps: (6) t.Fatal() was called Error types: (1) *withstack.withStack (2) *errutil.withMessage (3) *withstack.withStack (4) *errutil.withMessage (5) *withstack.withStack (6) *errors.errorString ``` <details><summary>More</summary><p> Artifacts: [/kv95/enc=false/nodes=3/cpu=32/seq](https://teamcity.cockroachdb.com/viewLog.html?buildId=2661561&tab=artifacts#/kv95/enc=false/nodes=3/cpu=32/seq) Related: - #60224 roachtest: kv95/enc=false/nodes=3/cpu=32/seq failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-60149](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-60149) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) - #60077 roachtest: kv95/enc=false/nodes=3/cpu=32/seq failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-20.2](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-20.2) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) - #59924 roachtest: kv95/enc=false/nodes=3/cpu=32/seq failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-master](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-master) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) [See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Akv95%2Fenc%3Dfalse%2Fnodes%3D3%2Fcpu%3D32%2Fseq.%2A&sort=title&restgroup=false&display=lastcommented+project) <sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
2.0
roachtest: kv95/enc=false/nodes=3/cpu=32/seq failed - [(roachtest).kv95/enc=false/nodes=3/cpu=32/seq failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2661561&tab=buildLog) on [release-20.1@90f78268f3b5b08ba838ac3ad164821d2f5a5362](https://github.com/cockroachdb/cockroach/commits/90f78268f3b5b08ba838ac3ad164821d2f5a5362): ``` | github.com/cockroachdb/cockroach/vendor/golang.org/x/sync/errgroup.(*Group).Go.func1 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/golang.org/x/sync/errgroup/errgroup.go:57 | runtime.goexit | /usr/local/go/src/runtime/asm_amd64.s:1374 Wraps: (2) 2 safe details enclosed Wraps: (3) output in run_074156.709_n4_workload_run_kv Wraps: (4) /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-2661561-1612941349-20-n4cpu32:4 -- ./workload run kv --init --histograms=perf/stats.json --concurrency=192 --duration=10m0s --read-percent=95 --sequential {pgurl:1-3} returned | stderr: | ./workload: error while loading shared libraries: libncurses.so.6: cannot open shared object file: No such file or directory | Error: COMMAND_PROBLEM: exit status 127 | (1) COMMAND_PROBLEM | Wraps: (2) Node 4. Command with error: | | ``` | | ./workload run kv --init --histograms=perf/stats.json --concurrency=192 --duration=10m0s --read-percent=95 --sequential {pgurl:1-3} | | ``` | Wraps: (3) exit status 127 | Error types: (1) errors.Cmd (2) *hintdetail.withDetail (3) *exec.ExitError | | stdout: Wraps: (5) exit status 20 Error types: (1) *withstack.withStack (2) *safedetails.withSafeDetails (3) *errutil.withMessage (4) *main.withCommandDetails (5) *exec.ExitError cluster.go:2628,kv.go:96,kv.go:183,test_runner.go:749: monitor failure: monitor task failed: t.Fatal() was called (1) attached stack trace | main.(*monitor).WaitE | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2616 | main.(*monitor).Wait | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2624 | main.registerKV.func2 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/kv.go:96 | main.registerKV.func3 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/kv.go:183 | main.(*testRunner).runTest.func2 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/test_runner.go:749 Wraps: (2) monitor failure Wraps: (3) attached stack trace | main.(*monitor).wait.func2 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2672 Wraps: (4) monitor task failed Wraps: (5) attached stack trace | main.init | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2586 | runtime.doInit | /usr/local/go/src/runtime/proc.go:5652 | runtime.main | /usr/local/go/src/runtime/proc.go:191 | runtime.goexit | /usr/local/go/src/runtime/asm_amd64.s:1374 Wraps: (6) t.Fatal() was called Error types: (1) *withstack.withStack (2) *errutil.withMessage (3) *withstack.withStack (4) *errutil.withMessage (5) *withstack.withStack (6) *errors.errorString ``` <details><summary>More</summary><p> Artifacts: [/kv95/enc=false/nodes=3/cpu=32/seq](https://teamcity.cockroachdb.com/viewLog.html?buildId=2661561&tab=artifacts#/kv95/enc=false/nodes=3/cpu=32/seq) Related: - #60224 roachtest: kv95/enc=false/nodes=3/cpu=32/seq failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-60149](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-60149) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) - #60077 roachtest: kv95/enc=false/nodes=3/cpu=32/seq failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-20.2](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-20.2) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) - #59924 roachtest: kv95/enc=false/nodes=3/cpu=32/seq failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-master](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-master) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) [See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Akv95%2Fenc%3Dfalse%2Fnodes%3D3%2Fcpu%3D32%2Fseq.%2A&sort=title&restgroup=false&display=lastcommented+project) <sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
test
roachtest enc false nodes cpu seq failed on github com cockroachdb cockroach vendor golang org x sync errgroup group go home agent work go src github com cockroachdb cockroach vendor golang org x sync errgroup errgroup go runtime goexit usr local go src runtime asm s wraps safe details enclosed wraps output in run workload run kv wraps home agent work go src github com cockroachdb cockroach bin roachprod run teamcity workload run kv init histograms perf stats json concurrency duration read percent sequential pgurl returned stderr workload error while loading shared libraries libncurses so cannot open shared object file no such file or directory error command problem exit status command problem wraps node command with error workload run kv init histograms perf stats json concurrency duration read percent sequential pgurl wraps exit status error types errors cmd hintdetail withdetail exec exiterror stdout wraps exit status error types withstack withstack safedetails withsafedetails errutil withmessage main withcommanddetails exec exiterror cluster go kv go kv go test runner go monitor failure monitor task failed t fatal was called attached stack trace main monitor waite home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go main monitor wait home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go main registerkv home agent work go src github com cockroachdb cockroach pkg cmd roachtest kv go main registerkv home agent work go src github com cockroachdb cockroach pkg cmd roachtest kv go main testrunner runtest home agent work go src github com cockroachdb cockroach pkg cmd roachtest test runner go wraps monitor failure wraps attached stack trace main monitor wait home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go wraps monitor task failed wraps attached stack trace main init home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go runtime doinit usr local go src runtime proc go runtime main usr local go src runtime proc go runtime goexit usr local go src runtime asm s wraps t fatal was called error types withstack withstack errutil withmessage withstack withstack errutil withmessage withstack withstack errors errorstring more artifacts related roachtest enc false nodes cpu seq failed roachtest enc false nodes cpu seq failed roachtest enc false nodes cpu seq failed powered by
1
422,804
12,287,484,931
IssuesEvent
2020-05-09 12:25:33
googleapis/elixir-google-api
https://api.github.com/repos/googleapis/elixir-google-api
opened
Synthesis failed for IAMCredentials
autosynth failure priority: p1 type: bug
Hello! Autosynth couldn't regenerate IAMCredentials. :broken_heart: Here's the output from running `synth.py`: ``` 2020-05-09 05:14:26 [INFO] logs will be written to: /tmpfs/src/github/synthtool/logs/googleapis/elixir-google-api 2020-05-09 05:14:26,326 autosynth > logs will be written to: /tmpfs/src/github/synthtool/logs/googleapis/elixir-google-api Switched to branch 'autosynth-iamcredentials' 2020-05-09 05:14:27 [INFO] Running synthtool 2020-05-09 05:14:27,964 autosynth > Running synthtool 2020-05-09 05:14:27 [INFO] ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/iam_credentials/synth.metadata', 'synth.py', '--'] 2020-05-09 05:14:27,964 autosynth > ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/iam_credentials/synth.metadata', 'synth.py', '--'] 2020-05-09 05:14:28,174 synthtool > Executing /home/kbuilder/.cache/synthtool/elixir-google-api/synth.py. On branch autosynth-iamcredentials nothing to commit, working tree clean 2020-05-09 05:14:28,464 synthtool > Cloning https://github.com/googleapis/elixir-google-api.git. 2020-05-09 05:14:28,945 synthtool > Running: docker run --rm -v/home/kbuilder/.cache/synthtool/elixir-google-api:/workspace -v/var/run/docker.sock:/var/run/docker.sock -e USER_GROUP=1000:1000 -w /workspace gcr.io/cloud-devrel-public-resources/elixir19 scripts/generate_client.sh IAMCredentials 2020-05-09 05:14:32,556 synthtool > No files in sources /home/kbuilder/.cache/synthtool/elixir-google-api/clients were copied. Does the source contain files? Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module> main() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 829, in __call__ return self.main(*args, **kwargs) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 782, in main rv = self.invoke(ctx) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1066, in invoke return ctx.invoke(self.callback, **ctx.params) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 610, in invoke return callback(*args, **kwargs) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main spec.loader.exec_module(synth_module) # type: ignore File "/tmpfs/src/github/synthtool/synthtool/metadata.py", line 180, in __exit__ write(self.metadata_file_path) File "/tmpfs/src/github/synthtool/synthtool/metadata.py", line 112, in write with open(outfile, "w") as fh: FileNotFoundError: [Errno 2] No such file or directory: 'clients/iam_credentials/synth.metadata' 2020-05-09 05:14:32 [ERROR] Synthesis failed 2020-05-09 05:14:32,586 autosynth > Synthesis failed Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 599, in <module> main() File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 471, in main return _inner_main(temp_dir) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 549, in _inner_main ).synthesize(base_synth_log_path) File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 118, in synthesize synth_proc.check_returncode() # Raise an exception. File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode self.stderr) subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/iam_credentials/synth.metadata', 'synth.py', '--', 'IAMCredentials']' returned non-zero exit status 1. ``` Google internal developers can see the full log [here](https://sponge/11ff3741-9158-4831-8681-fff828f77e1a).
1.0
Synthesis failed for IAMCredentials - Hello! Autosynth couldn't regenerate IAMCredentials. :broken_heart: Here's the output from running `synth.py`: ``` 2020-05-09 05:14:26 [INFO] logs will be written to: /tmpfs/src/github/synthtool/logs/googleapis/elixir-google-api 2020-05-09 05:14:26,326 autosynth > logs will be written to: /tmpfs/src/github/synthtool/logs/googleapis/elixir-google-api Switched to branch 'autosynth-iamcredentials' 2020-05-09 05:14:27 [INFO] Running synthtool 2020-05-09 05:14:27,964 autosynth > Running synthtool 2020-05-09 05:14:27 [INFO] ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/iam_credentials/synth.metadata', 'synth.py', '--'] 2020-05-09 05:14:27,964 autosynth > ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/iam_credentials/synth.metadata', 'synth.py', '--'] 2020-05-09 05:14:28,174 synthtool > Executing /home/kbuilder/.cache/synthtool/elixir-google-api/synth.py. On branch autosynth-iamcredentials nothing to commit, working tree clean 2020-05-09 05:14:28,464 synthtool > Cloning https://github.com/googleapis/elixir-google-api.git. 2020-05-09 05:14:28,945 synthtool > Running: docker run --rm -v/home/kbuilder/.cache/synthtool/elixir-google-api:/workspace -v/var/run/docker.sock:/var/run/docker.sock -e USER_GROUP=1000:1000 -w /workspace gcr.io/cloud-devrel-public-resources/elixir19 scripts/generate_client.sh IAMCredentials 2020-05-09 05:14:32,556 synthtool > No files in sources /home/kbuilder/.cache/synthtool/elixir-google-api/clients were copied. Does the source contain files? Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module> main() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 829, in __call__ return self.main(*args, **kwargs) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 782, in main rv = self.invoke(ctx) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1066, in invoke return ctx.invoke(self.callback, **ctx.params) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 610, in invoke return callback(*args, **kwargs) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main spec.loader.exec_module(synth_module) # type: ignore File "/tmpfs/src/github/synthtool/synthtool/metadata.py", line 180, in __exit__ write(self.metadata_file_path) File "/tmpfs/src/github/synthtool/synthtool/metadata.py", line 112, in write with open(outfile, "w") as fh: FileNotFoundError: [Errno 2] No such file or directory: 'clients/iam_credentials/synth.metadata' 2020-05-09 05:14:32 [ERROR] Synthesis failed 2020-05-09 05:14:32,586 autosynth > Synthesis failed Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 599, in <module> main() File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 471, in main return _inner_main(temp_dir) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 549, in _inner_main ).synthesize(base_synth_log_path) File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 118, in synthesize synth_proc.check_returncode() # Raise an exception. File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode self.stderr) subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/iam_credentials/synth.metadata', 'synth.py', '--', 'IAMCredentials']' returned non-zero exit status 1. ``` Google internal developers can see the full log [here](https://sponge/11ff3741-9158-4831-8681-fff828f77e1a).
non_test
synthesis failed for iamcredentials hello autosynth couldn t regenerate iamcredentials broken heart here s the output from running synth py logs will be written to tmpfs src github synthtool logs googleapis elixir google api autosynth logs will be written to tmpfs src github synthtool logs googleapis elixir google api switched to branch autosynth iamcredentials running synthtool autosynth running synthtool autosynth synthtool executing home kbuilder cache synthtool elixir google api synth py on branch autosynth iamcredentials nothing to commit working tree clean synthtool cloning synthtool running docker run rm v home kbuilder cache synthtool elixir google api workspace v var run docker sock var run docker sock e user group w workspace gcr io cloud devrel public resources scripts generate client sh iamcredentials synthtool no files in sources home kbuilder cache synthtool elixir google api clients were copied does the source contain files traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool synthtool main py line in main file tmpfs src github synthtool env lib site packages click core py line in call return self main args kwargs file tmpfs src github synthtool env lib site packages click core py line in main rv self invoke ctx file tmpfs src github synthtool env lib site packages click core py line in invoke return ctx invoke self callback ctx params file tmpfs src github synthtool env lib site packages click core py line in invoke return callback args kwargs file tmpfs src github synthtool synthtool main py line in main spec loader exec module synth module type ignore file tmpfs src github synthtool synthtool metadata py line in exit write self metadata file path file tmpfs src github synthtool synthtool metadata py line in write with open outfile w as fh filenotfounderror no such file or directory clients iam credentials synth metadata synthesis failed autosynth synthesis failed traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool autosynth synth py line in main file tmpfs src github synthtool autosynth synth py line in main return inner main temp dir file tmpfs src github synthtool autosynth synth py line in inner main synthesize base synth log path file tmpfs src github synthtool autosynth synthesizer py line in synthesize synth proc check returncode raise an exception file home kbuilder pyenv versions lib subprocess py line in check returncode self stderr subprocess calledprocesserror command returned non zero exit status google internal developers can see the full log
0
173,362
13,398,307,305
IssuesEvent
2020-09-03 12:59:02
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachtest: jobs/mixed-versions failed
C-test-failure O-roachtest O-robot branch-provisional_202007220233_v20.2.0-alpha.2 release-blocker
[(roachtest).jobs/mixed-versions failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2107811&tab=buildLog) on [provisional_202007220233_v20.2.0-alpha.2@d3119926d33d808c6384cf3e99a7f7435f395489](https://github.com/cockroachdb/cockroach/commits/d3119926d33d808c6384cf3e99a7f7435f395489): ``` The test failed on branch=provisional_202007220233_v20.2.0-alpha.2, cloud=gce: test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/jobs/mixed-versions/run_1 mixed_version_jobs.go:198,versionupgrade.go:167,mixed_version_jobs.go:296,mixed_version_jobs.go:320,test_runner.go:757: Cluster info Node 1: 20.1 Node 2: 20.1-9 Node 3: 20.1-9 Node 4: 20.1 Unsuccessful job 574495805866573825 of type IMPORT, description IMPORT TABLE tpcc.public.customer (c_id INT8 NOT NULL, c_d_id INT8 NOT NULL, c_w_id INT8 NOT NULL, c_first VARCHAR(16), c_middle CHAR(2), c_last VARCHAR(16), c_street_1 VARCHAR(20), c_street_2 VARCHAR(20), c_city VARCHAR(20), c_state CHAR(2), c_zip CHAR(9), c_phone CHAR(16), c_since TIMESTAMP, c_credit CHAR(2), c_credit_lim DECIMAL(12,2), c_discount DECIMAL(4,4), c_balance DECIMAL(12,2), c_ytd_payment DECIMAL(12,2), c_payment_cnt INT8, c_delivery_cnt INT8, c_data VARCHAR(500), PRIMARY KEY (c_w_id, c_d_id, c_id), INDEX customer_idx (c_w_id, c_d_id, c_last, c_first)) CSV DATA ('workload:///csv/tpcc/customer?fks=true&interleaved=false&row-end=1500000&row-start=0&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/customer?fks=true&interleaved=false&row-end=3000000&row-start=1500000&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/customer?fks=true&interleaved=false&row-end=4500000&row-start=3000000&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/customer?fks=true&interleaved=false&row-end=6000000&row-start=4500000&seed=1&version=2.1.0&warehouses=200') WITH "nullif" = 'NULL', status failed, error version mismatch in flow request: 30; this node accepts 27 through 28, coordinator 3 Unsuccessful job 574495805866868737 of type IMPORT, description IMPORT TABLE tpcc.public.stock (s_i_id INT8 NOT NULL, s_w_id INT8 NOT NULL, s_quantity INT8, s_dist_01 CHAR(24), s_dist_02 CHAR(24), s_dist_03 CHAR(24), s_dist_04 CHAR(24), s_dist_05 CHAR(24), s_dist_06 CHAR(24), s_dist_07 CHAR(24), s_dist_08 CHAR(24), s_dist_09 CHAR(24), s_dist_10 CHAR(24), s_ytd INT8, s_order_cnt INT8, s_remote_cnt INT8, s_data VARCHAR(50), PRIMARY KEY (s_w_id, s_i_id), INDEX stock_item_fk_idx (s_i_id)) CSV DATA ('workload:///csv/tpcc/stock?fks=true&interleaved=false&row-end=5000000&row-start=0&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/stock?fks=true&interleaved=false&row-end=10000000&row-start=5000000&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/stock?fks=true&interleaved=false&row-end=15000000&row-start=10000000&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/stock?fks=true&interleaved=false&row-end=20000000&row-start=15000000&seed=1&version=2.1.0&warehouses=200') WITH "nullif" = 'NULL', status failed, error version mismatch in flow request: 30; this node accepts 27 through 28, coordinator 3 Unsuccessful job 574495805867360257 of type IMPORT, description IMPORT TABLE tpcc.public.history (rowid UUID NOT NULL DEFAULT gen_random_uuid(), h_c_id INT8 NOT NULL, h_c_d_id INT8 NOT NULL, h_c_w_id INT8 NOT NULL, h_d_id INT8 NOT NULL, h_w_id INT8 NOT NULL, h_date TIMESTAMP, h_amount DECIMAL(6,2), h_data VARCHAR(24), PRIMARY KEY (h_w_id, rowid), INDEX history_customer_fk_idx (h_c_w_id, h_c_d_id, h_c_id), INDEX history_district_fk_idx (h_w_id, h_d_id)) CSV DATA ('workload:///csv/tpcc/history?fks=true&interleaved=false&row-end=1500000&row-start=0&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/history?fks=true&interleaved=false&row-end=3000000&row-start=1500000&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/history?fks=true&interleaved=false&row-end=4500000&row-start=3000000&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/history?fks=true&interleaved=false&row-end=6000000&row-start=4500000&seed=1&version=2.1.0&warehouses=200') WITH "nullif" = 'NULL', status failed, error version mismatch in flow request: 30; this node accepts 27 through 28, coordinator 3 Unsuccessful job 574495805868441601 of type IMPORT, description IMPORT TABLE tpcc.public.order_line (ol_o_id INT8 NOT NULL, ol_d_id INT8 NOT NULL, ol_w_id INT8 NOT NULL, ol_number INT8 NOT NULL, ol_i_id INT8 NOT NULL, ol_supply_w_id INT8, ol_delivery_d TIMESTAMP, ol_quantity INT8, ol_amount DECIMAL(6,2), ol_dist_info CHAR(24), PRIMARY KEY (ol_w_id, ol_d_id, ol_o_id DESC, ol_number), INDEX order_line_stock_fk_idx (ol_supply_w_id, ol_i_id)) CSV DATA ('workload:///csv/tpcc/order_line?fks=true&interleaved=false&row-end=1500000&row-start=0&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/order_line?fks=true&interleaved=false&row-end=3000000&row-start=1500000&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/order_line?fks=true&interleaved=false&row-end=4500000&row-start=3000000&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/order_line?fks=true&interleaved=false&row-end=6000000&row-start=4500000&seed=1&version=2.1.0&warehouses=200') WITH "nullif" = 'NULL', status failed, error version mismatch in flow request: 30; this node accepts 27 through 28, coordinator 3 Unsuccessful job 574495805872570369 of type IMPORT, description IMPORT TABLE tpcc.public.item (i_id INT8 NOT NULL, i_im_id INT8, i_name VARCHAR(24), i_price DECIMAL(5,2), i_data VARCHAR(50), PRIMARY KEY (i_id)) CSV DATA ('workload:///csv/tpcc/item?fks=true&interleaved=false&row-end=25000&row-start=0&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/item?fks=true&interleaved=false&row-end=50000&row-start=25000&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/item?fks=true&interleaved=false&row-end=75000&row-start=50000&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/item?fks=true&interleaved=false&row-end=100000&row-start=75000&seed=1&version=2.1.0&warehouses=200') WITH "nullif" = 'NULL', status failed, error job 574495805872570369: cannot be reverted, manual cleanup may be required: rolling back tables: unexpected value: raw_bytes:"\026\334\374M\003\n\232\003\n\004item\030\250\001 \242\001(\002:\n\010\243\275\234\344\366\243\376\221\026B\032\n\004i_id\020\001\032\014\010\001\020@\030\0000\000P\024`\000 \0000\000B\035\n\007i_im_id\020\002\032\014\010\001\020@\030\0000\000P\024`\000 \0010\000B\035\n\006i_name\020\003\032\r\010\007\020\030\030\0000\007P\223\010`\000 \0010\000B\036\n\007i_price\020\004\032\r\010\003\020\002\030\0050\000P\244\r`\000 \0010\000B\035\n\006i_data\020\005\032\r\010\007\0202\030\0000\007P\223\010`\000 \0010\000H\006RG\n\007primary\020\001\030\001\"\004i_id0\001@\000J\020\010\000\020\000\032\000 \000(\0000\0008\000@\000Z\000z\002\010\000\200\001\000\210\001\000\220\001\001\230\001\000\242\001\006\010\000\022\000\030\000\250\001\000`\002j\025\n\t\n\005admin\020\002\n\010\n\004root\020\002\200\001\001\210\001\003\230\001\000\262\001?\n\007primary\020\000\032\004i_id\032\007i_im_id\032\006i_name\032\007i_price\032\006i_data \001 \002 \003 \004 \005(\000\270\001\001\302\001\000\350\001\000\362\001\004\010\000\022\000\370\001\000\200\002\000\222\002\000\232\002\n\010\243\275\234\344\366\243\376\221\026\262\002\timporting\270\002\000\300\002\035" timestamp:<wall_time:1595392626971668696 logical:1 > , coordinator 3 Unsuccessful job 574495805872603137 of type IMPORT, description IMPORT TABLE tpcc.public."order" (o_id INT8 NOT NULL, o_d_id INT8 NOT NULL, o_w_id INT8 NOT NULL, o_c_id INT8, o_entry_d TIMESTAMP, o_carrier_id INT8, o_ol_cnt INT8, o_all_local INT8, PRIMARY KEY (o_w_id, o_d_id, o_id DESC), CONSTRAINT order_idx UNIQUE (o_w_id, o_d_id, o_c_id, o_id DESC) STORING (o_entry_d, o_carrier_id)) CSV DATA ('workload:///csv/tpcc/order?fks=true&interleaved=false&row-end=1500000&row-start=0&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/order?fks=true&interleaved=false&row-end=3000000&row-start=1500000&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/order?fks=true&interleaved=false&row-end=4500000&row-start=3000000&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/order?fks=true&interleaved=false&row-end=6000000&row-start=4500000&seed=1&version=2.1.0&warehouses=200') WITH "nullif" = 'NULL', status failed, error version mismatch in flow request: 30; this node accepts 27 through 28, coordinator 3 ``` <details><summary>More</summary><p> Artifacts: [/jobs/mixed-versions](https://teamcity.cockroachdb.com/viewLog.html?buildId=2107811&tab=artifacts#/jobs/mixed-versions) Related: - #51186 roachtest: jobs/mixed-versions failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202007081918_v20.2.0-alpha.2](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202007081918_v20.2.0-alpha.2) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) - #51100 roachtest: jobs/mixed-versions failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202007071743_v20.2.0-alpha.2](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202007071743_v20.2.0-alpha.2) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) - #50026 roachtest: jobs/mixed-versions failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202006032224_v20.2.0-alpha.1](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202006032224_v20.2.0-alpha.1) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) - #49281 roachtest: jobs/mixed-versions failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202005191400_v20.1.1](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202005191400_v20.1.1) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) - #49233 roachtest: jobs/mixed-versions failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202005182011_v20.1.1](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202005182011_v20.1.1) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) - #48407 roachtest: jobs/mixed-versions failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202005041945_v19.1.9](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202005041945_v19.1.9) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) - #48315 roachtest: jobs/mixed-versions failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-master](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-master) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) - #48194 roachtest: jobs/mixed-versions failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-20.1](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-20.1) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) - #48193 roachtest: jobs/mixed-versions failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-19.1](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-19.1) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) [See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Ajobs%2Fmixed-versions.%2A&sort=title&restgroup=false&display=lastcommented+project) <sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
2.0
roachtest: jobs/mixed-versions failed - [(roachtest).jobs/mixed-versions failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2107811&tab=buildLog) on [provisional_202007220233_v20.2.0-alpha.2@d3119926d33d808c6384cf3e99a7f7435f395489](https://github.com/cockroachdb/cockroach/commits/d3119926d33d808c6384cf3e99a7f7435f395489): ``` The test failed on branch=provisional_202007220233_v20.2.0-alpha.2, cloud=gce: test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/jobs/mixed-versions/run_1 mixed_version_jobs.go:198,versionupgrade.go:167,mixed_version_jobs.go:296,mixed_version_jobs.go:320,test_runner.go:757: Cluster info Node 1: 20.1 Node 2: 20.1-9 Node 3: 20.1-9 Node 4: 20.1 Unsuccessful job 574495805866573825 of type IMPORT, description IMPORT TABLE tpcc.public.customer (c_id INT8 NOT NULL, c_d_id INT8 NOT NULL, c_w_id INT8 NOT NULL, c_first VARCHAR(16), c_middle CHAR(2), c_last VARCHAR(16), c_street_1 VARCHAR(20), c_street_2 VARCHAR(20), c_city VARCHAR(20), c_state CHAR(2), c_zip CHAR(9), c_phone CHAR(16), c_since TIMESTAMP, c_credit CHAR(2), c_credit_lim DECIMAL(12,2), c_discount DECIMAL(4,4), c_balance DECIMAL(12,2), c_ytd_payment DECIMAL(12,2), c_payment_cnt INT8, c_delivery_cnt INT8, c_data VARCHAR(500), PRIMARY KEY (c_w_id, c_d_id, c_id), INDEX customer_idx (c_w_id, c_d_id, c_last, c_first)) CSV DATA ('workload:///csv/tpcc/customer?fks=true&interleaved=false&row-end=1500000&row-start=0&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/customer?fks=true&interleaved=false&row-end=3000000&row-start=1500000&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/customer?fks=true&interleaved=false&row-end=4500000&row-start=3000000&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/customer?fks=true&interleaved=false&row-end=6000000&row-start=4500000&seed=1&version=2.1.0&warehouses=200') WITH "nullif" = 'NULL', status failed, error version mismatch in flow request: 30; this node accepts 27 through 28, coordinator 3 Unsuccessful job 574495805866868737 of type IMPORT, description IMPORT TABLE tpcc.public.stock (s_i_id INT8 NOT NULL, s_w_id INT8 NOT NULL, s_quantity INT8, s_dist_01 CHAR(24), s_dist_02 CHAR(24), s_dist_03 CHAR(24), s_dist_04 CHAR(24), s_dist_05 CHAR(24), s_dist_06 CHAR(24), s_dist_07 CHAR(24), s_dist_08 CHAR(24), s_dist_09 CHAR(24), s_dist_10 CHAR(24), s_ytd INT8, s_order_cnt INT8, s_remote_cnt INT8, s_data VARCHAR(50), PRIMARY KEY (s_w_id, s_i_id), INDEX stock_item_fk_idx (s_i_id)) CSV DATA ('workload:///csv/tpcc/stock?fks=true&interleaved=false&row-end=5000000&row-start=0&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/stock?fks=true&interleaved=false&row-end=10000000&row-start=5000000&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/stock?fks=true&interleaved=false&row-end=15000000&row-start=10000000&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/stock?fks=true&interleaved=false&row-end=20000000&row-start=15000000&seed=1&version=2.1.0&warehouses=200') WITH "nullif" = 'NULL', status failed, error version mismatch in flow request: 30; this node accepts 27 through 28, coordinator 3 Unsuccessful job 574495805867360257 of type IMPORT, description IMPORT TABLE tpcc.public.history (rowid UUID NOT NULL DEFAULT gen_random_uuid(), h_c_id INT8 NOT NULL, h_c_d_id INT8 NOT NULL, h_c_w_id INT8 NOT NULL, h_d_id INT8 NOT NULL, h_w_id INT8 NOT NULL, h_date TIMESTAMP, h_amount DECIMAL(6,2), h_data VARCHAR(24), PRIMARY KEY (h_w_id, rowid), INDEX history_customer_fk_idx (h_c_w_id, h_c_d_id, h_c_id), INDEX history_district_fk_idx (h_w_id, h_d_id)) CSV DATA ('workload:///csv/tpcc/history?fks=true&interleaved=false&row-end=1500000&row-start=0&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/history?fks=true&interleaved=false&row-end=3000000&row-start=1500000&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/history?fks=true&interleaved=false&row-end=4500000&row-start=3000000&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/history?fks=true&interleaved=false&row-end=6000000&row-start=4500000&seed=1&version=2.1.0&warehouses=200') WITH "nullif" = 'NULL', status failed, error version mismatch in flow request: 30; this node accepts 27 through 28, coordinator 3 Unsuccessful job 574495805868441601 of type IMPORT, description IMPORT TABLE tpcc.public.order_line (ol_o_id INT8 NOT NULL, ol_d_id INT8 NOT NULL, ol_w_id INT8 NOT NULL, ol_number INT8 NOT NULL, ol_i_id INT8 NOT NULL, ol_supply_w_id INT8, ol_delivery_d TIMESTAMP, ol_quantity INT8, ol_amount DECIMAL(6,2), ol_dist_info CHAR(24), PRIMARY KEY (ol_w_id, ol_d_id, ol_o_id DESC, ol_number), INDEX order_line_stock_fk_idx (ol_supply_w_id, ol_i_id)) CSV DATA ('workload:///csv/tpcc/order_line?fks=true&interleaved=false&row-end=1500000&row-start=0&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/order_line?fks=true&interleaved=false&row-end=3000000&row-start=1500000&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/order_line?fks=true&interleaved=false&row-end=4500000&row-start=3000000&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/order_line?fks=true&interleaved=false&row-end=6000000&row-start=4500000&seed=1&version=2.1.0&warehouses=200') WITH "nullif" = 'NULL', status failed, error version mismatch in flow request: 30; this node accepts 27 through 28, coordinator 3 Unsuccessful job 574495805872570369 of type IMPORT, description IMPORT TABLE tpcc.public.item (i_id INT8 NOT NULL, i_im_id INT8, i_name VARCHAR(24), i_price DECIMAL(5,2), i_data VARCHAR(50), PRIMARY KEY (i_id)) CSV DATA ('workload:///csv/tpcc/item?fks=true&interleaved=false&row-end=25000&row-start=0&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/item?fks=true&interleaved=false&row-end=50000&row-start=25000&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/item?fks=true&interleaved=false&row-end=75000&row-start=50000&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/item?fks=true&interleaved=false&row-end=100000&row-start=75000&seed=1&version=2.1.0&warehouses=200') WITH "nullif" = 'NULL', status failed, error job 574495805872570369: cannot be reverted, manual cleanup may be required: rolling back tables: unexpected value: raw_bytes:"\026\334\374M\003\n\232\003\n\004item\030\250\001 \242\001(\002:\n\010\243\275\234\344\366\243\376\221\026B\032\n\004i_id\020\001\032\014\010\001\020@\030\0000\000P\024`\000 \0000\000B\035\n\007i_im_id\020\002\032\014\010\001\020@\030\0000\000P\024`\000 \0010\000B\035\n\006i_name\020\003\032\r\010\007\020\030\030\0000\007P\223\010`\000 \0010\000B\036\n\007i_price\020\004\032\r\010\003\020\002\030\0050\000P\244\r`\000 \0010\000B\035\n\006i_data\020\005\032\r\010\007\0202\030\0000\007P\223\010`\000 \0010\000H\006RG\n\007primary\020\001\030\001\"\004i_id0\001@\000J\020\010\000\020\000\032\000 \000(\0000\0008\000@\000Z\000z\002\010\000\200\001\000\210\001\000\220\001\001\230\001\000\242\001\006\010\000\022\000\030\000\250\001\000`\002j\025\n\t\n\005admin\020\002\n\010\n\004root\020\002\200\001\001\210\001\003\230\001\000\262\001?\n\007primary\020\000\032\004i_id\032\007i_im_id\032\006i_name\032\007i_price\032\006i_data \001 \002 \003 \004 \005(\000\270\001\001\302\001\000\350\001\000\362\001\004\010\000\022\000\370\001\000\200\002\000\222\002\000\232\002\n\010\243\275\234\344\366\243\376\221\026\262\002\timporting\270\002\000\300\002\035" timestamp:<wall_time:1595392626971668696 logical:1 > , coordinator 3 Unsuccessful job 574495805872603137 of type IMPORT, description IMPORT TABLE tpcc.public."order" (o_id INT8 NOT NULL, o_d_id INT8 NOT NULL, o_w_id INT8 NOT NULL, o_c_id INT8, o_entry_d TIMESTAMP, o_carrier_id INT8, o_ol_cnt INT8, o_all_local INT8, PRIMARY KEY (o_w_id, o_d_id, o_id DESC), CONSTRAINT order_idx UNIQUE (o_w_id, o_d_id, o_c_id, o_id DESC) STORING (o_entry_d, o_carrier_id)) CSV DATA ('workload:///csv/tpcc/order?fks=true&interleaved=false&row-end=1500000&row-start=0&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/order?fks=true&interleaved=false&row-end=3000000&row-start=1500000&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/order?fks=true&interleaved=false&row-end=4500000&row-start=3000000&seed=1&version=2.1.0&warehouses=200', 'workload:///csv/tpcc/order?fks=true&interleaved=false&row-end=6000000&row-start=4500000&seed=1&version=2.1.0&warehouses=200') WITH "nullif" = 'NULL', status failed, error version mismatch in flow request: 30; this node accepts 27 through 28, coordinator 3 ``` <details><summary>More</summary><p> Artifacts: [/jobs/mixed-versions](https://teamcity.cockroachdb.com/viewLog.html?buildId=2107811&tab=artifacts#/jobs/mixed-versions) Related: - #51186 roachtest: jobs/mixed-versions failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202007081918_v20.2.0-alpha.2](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202007081918_v20.2.0-alpha.2) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) - #51100 roachtest: jobs/mixed-versions failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202007071743_v20.2.0-alpha.2](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202007071743_v20.2.0-alpha.2) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) - #50026 roachtest: jobs/mixed-versions failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202006032224_v20.2.0-alpha.1](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202006032224_v20.2.0-alpha.1) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) - #49281 roachtest: jobs/mixed-versions failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202005191400_v20.1.1](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202005191400_v20.1.1) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) - #49233 roachtest: jobs/mixed-versions failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202005182011_v20.1.1](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202005182011_v20.1.1) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) - #48407 roachtest: jobs/mixed-versions failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202005041945_v19.1.9](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202005041945_v19.1.9) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) - #48315 roachtest: jobs/mixed-versions failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-master](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-master) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) - #48194 roachtest: jobs/mixed-versions failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-20.1](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-20.1) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) - #48193 roachtest: jobs/mixed-versions failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-19.1](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-19.1) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) [See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Ajobs%2Fmixed-versions.%2A&sort=title&restgroup=false&display=lastcommented+project) <sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
test
roachtest jobs mixed versions failed on the test failed on branch provisional alpha cloud gce test artifacts and logs in home agent work go src github com cockroachdb cockroach artifacts jobs mixed versions run mixed version jobs go versionupgrade go mixed version jobs go mixed version jobs go test runner go cluster info node node node node unsuccessful job of type import description import table tpcc public customer c id not null c d id not null c w id not null c first varchar c middle char c last varchar c street varchar c street varchar c city varchar c state char c zip char c phone char c since timestamp c credit char c credit lim decimal c discount decimal c balance decimal c ytd payment decimal c payment cnt c delivery cnt c data varchar primary key c w id c d id c id index customer idx c w id c d id c last c first csv data workload csv tpcc customer fks true interleaved false row end row start seed version warehouses workload csv tpcc customer fks true interleaved false row end row start seed version warehouses workload csv tpcc customer fks true interleaved false row end row start seed version warehouses workload csv tpcc customer fks true interleaved false row end row start seed version warehouses with nullif null status failed error version mismatch in flow request this node accepts through coordinator unsuccessful job of type import description import table tpcc public stock s i id not null s w id not null s quantity s dist char s dist char s dist char s dist char s dist char s dist char s dist char s dist char s dist char s dist char s ytd s order cnt s remote cnt s data varchar primary key s w id s i id index stock item fk idx s i id csv data workload csv tpcc stock fks true interleaved false row end row start seed version warehouses workload csv tpcc stock fks true interleaved false row end row start seed version warehouses workload csv tpcc stock fks true interleaved false row end row start seed version warehouses workload csv tpcc stock fks true interleaved false row end row start seed version warehouses with nullif null status failed error version mismatch in flow request this node accepts through coordinator unsuccessful job of type import description import table tpcc public history rowid uuid not null default gen random uuid h c id not null h c d id not null h c w id not null h d id not null h w id not null h date timestamp h amount decimal h data varchar primary key h w id rowid index history customer fk idx h c w id h c d id h c id index history district fk idx h w id h d id csv data workload csv tpcc history fks true interleaved false row end row start seed version warehouses workload csv tpcc history fks true interleaved false row end row start seed version warehouses workload csv tpcc history fks true interleaved false row end row start seed version warehouses workload csv tpcc history fks true interleaved false row end row start seed version warehouses with nullif null status failed error version mismatch in flow request this node accepts through coordinator unsuccessful job of type import description import table tpcc public order line ol o id not null ol d id not null ol w id not null ol number not null ol i id not null ol supply w id ol delivery d timestamp ol quantity ol amount decimal ol dist info char primary key ol w id ol d id ol o id desc ol number index order line stock fk idx ol supply w id ol i id csv data workload csv tpcc order line fks true interleaved false row end row start seed version warehouses workload csv tpcc order line fks true interleaved false row end row start seed version warehouses workload csv tpcc order line fks true interleaved false row end row start seed version warehouses workload csv tpcc order line fks true interleaved false row end row start seed version warehouses with nullif null status failed error version mismatch in flow request this node accepts through coordinator unsuccessful job of type import description import table tpcc public item i id not null i im id i name varchar i price decimal i data varchar primary key i id csv data workload csv tpcc item fks true interleaved false row end row start seed version warehouses workload csv tpcc item fks true interleaved false row end row start seed version warehouses workload csv tpcc item fks true interleaved false row end row start seed version warehouses workload csv tpcc item fks true interleaved false row end row start seed version warehouses with nullif null status failed error job cannot be reverted manual cleanup may be required rolling back tables unexpected value raw bytes n n n n id n im id n name r n price r r n data r n n t n n n n id im id name price data n timporting timestamp coordinator unsuccessful job of type import description import table tpcc public order o id not null o d id not null o w id not null o c id o entry d timestamp o carrier id o ol cnt o all local primary key o w id o d id o id desc constraint order idx unique o w id o d id o c id o id desc storing o entry d o carrier id csv data workload csv tpcc order fks true interleaved false row end row start seed version warehouses workload csv tpcc order fks true interleaved false row end row start seed version warehouses workload csv tpcc order fks true interleaved false row end row start seed version warehouses workload csv tpcc order fks true interleaved false row end row start seed version warehouses with nullif null status failed error version mismatch in flow request this node accepts through coordinator more artifacts related roachtest jobs mixed versions failed roachtest jobs mixed versions failed roachtest jobs mixed versions failed roachtest jobs mixed versions failed roachtest jobs mixed versions failed roachtest jobs mixed versions failed roachtest jobs mixed versions failed roachtest jobs mixed versions failed roachtest jobs mixed versions failed powered by
1
86,305
8,031,396,503
IssuesEvent
2018-07-28 01:02:58
rancher/rancher
https://api.github.com/repos/rancher/rancher
closed
Increasing the project resource quota does not adjust the quota on the namespace
kind/bug status/resolved status/to-test version/2.0
**Rancher versions:** rancher/server or rancher/rancher: master **Steps to Reproduce:** 1. Create a Project with resource quota as below: ``` {"limit":{"pods": "10", "services": "10", "replicationControllers": "10", "secrets": "10", "configMaps": "10", "persistentVolumeClaims": "10", "servicesNodePorts": "100", "servicesLoadBalancers": "10", "requestsCpu": "2000m", "requestsMemory": "100Gi", "limitsCpu": "2000m", "limitsMemory": "100Gi" }} ``` 2. Create a Resourcequota template as below: ``` {"pods": "10", "services": "10", "replicationControllers": "10", "secrets": "10", "configMaps": "10", "persistentVolumeClaims": "10", "servicesNodePorts": "100", "servicesLoadBalancers": "10", "requestsCpu": "1000m", "requestsMemory": "10Gi", "limitsCpu": "1000m", "limitsMemory": "50Gi" } ``` 3. Create namespace/workloads exceeding the cpu resource quota lets say 2000m. Workload creation will fail. 4. Edit and Increase the project cpu resource quota on the project to 3000m as below: ``` {"limit":{"pods": "10", "services": "10", "replicationControllers": "10", "secrets": "10", "configMaps": "10", "persistentVolumeClaims": "10", "servicesNodePorts": "100", "servicesLoadBalancers": "10", "requestsCpu": "3000m", "requestsMemory": "100Gi", "limitsCpu": "3000m", "limitsMemory": "100Gi" }} ``` Increasing the project quota does not adjust the quota on the namespace. The pod creation does not continue within the allowed new updated quota <img width="1187" alt="screen shot 2018-07-25 at 4 18 25 pm" src="https://user-images.githubusercontent.com/18536626/43232574-85fd6fbc-9026-11e8-85df-4119c8bdc870.png">
1.0
Increasing the project resource quota does not adjust the quota on the namespace - **Rancher versions:** rancher/server or rancher/rancher: master **Steps to Reproduce:** 1. Create a Project with resource quota as below: ``` {"limit":{"pods": "10", "services": "10", "replicationControllers": "10", "secrets": "10", "configMaps": "10", "persistentVolumeClaims": "10", "servicesNodePorts": "100", "servicesLoadBalancers": "10", "requestsCpu": "2000m", "requestsMemory": "100Gi", "limitsCpu": "2000m", "limitsMemory": "100Gi" }} ``` 2. Create a Resourcequota template as below: ``` {"pods": "10", "services": "10", "replicationControllers": "10", "secrets": "10", "configMaps": "10", "persistentVolumeClaims": "10", "servicesNodePorts": "100", "servicesLoadBalancers": "10", "requestsCpu": "1000m", "requestsMemory": "10Gi", "limitsCpu": "1000m", "limitsMemory": "50Gi" } ``` 3. Create namespace/workloads exceeding the cpu resource quota lets say 2000m. Workload creation will fail. 4. Edit and Increase the project cpu resource quota on the project to 3000m as below: ``` {"limit":{"pods": "10", "services": "10", "replicationControllers": "10", "secrets": "10", "configMaps": "10", "persistentVolumeClaims": "10", "servicesNodePorts": "100", "servicesLoadBalancers": "10", "requestsCpu": "3000m", "requestsMemory": "100Gi", "limitsCpu": "3000m", "limitsMemory": "100Gi" }} ``` Increasing the project quota does not adjust the quota on the namespace. The pod creation does not continue within the allowed new updated quota <img width="1187" alt="screen shot 2018-07-25 at 4 18 25 pm" src="https://user-images.githubusercontent.com/18536626/43232574-85fd6fbc-9026-11e8-85df-4119c8bdc870.png">
test
increasing the project resource quota does not adjust the quota on the namespace rancher versions rancher server or rancher rancher master steps to reproduce create a project with resource quota as below limit pods services replicationcontrollers secrets configmaps persistentvolumeclaims servicesnodeports servicesloadbalancers requestscpu requestsmemory limitscpu limitsmemory create a resourcequota template as below pods services replicationcontrollers secrets configmaps persistentvolumeclaims servicesnodeports servicesloadbalancers requestscpu requestsmemory limitscpu limitsmemory create namespace workloads exceeding the cpu resource quota lets say workload creation will fail edit and increase the project cpu resource quota on the project to as below limit pods services replicationcontrollers secrets configmaps persistentvolumeclaims servicesnodeports servicesloadbalancers requestscpu requestsmemory limitscpu limitsmemory increasing the project quota does not adjust the quota on the namespace the pod creation does not continue within the allowed new updated quota img width alt screen shot at pm src
1
32,673
4,781,803,019
IssuesEvent
2016-10-28 10:44:49
itheof/42sh-posix
https://api.github.com/repos/itheof/42sh-posix
closed
Test de retour NULL de malloc: idées
help wanted question test
Je pense qu'il est possible de tester les retours NULL de malloc de manière propre. On pourrait charger notre propre librarie dynamique malloc (contenant malloc, realloc, free...) et 'wrapper' les vrais fonctions du système (conflicts de noms ?), en les glissants dans ce genre de scope: ```c #ifnedef # define MALLOC_CRASH_AT 0 // each compilation increment this #endif static g_malloc_count = 0; void *malloc(size_t size) { if (g_malloc_count == MALLOC_CRASH) { ++g_malloc_count; return (NULL); } ++g_malloc_count; return (malloc(size)); // this is the real malloc but name conflicts... } ``` `MALLOC_CRASH_AT` est incrémenté à chaque compilation et à chaque fois le programme passe une série de tests
1.0
Test de retour NULL de malloc: idées - Je pense qu'il est possible de tester les retours NULL de malloc de manière propre. On pourrait charger notre propre librarie dynamique malloc (contenant malloc, realloc, free...) et 'wrapper' les vrais fonctions du système (conflicts de noms ?), en les glissants dans ce genre de scope: ```c #ifnedef # define MALLOC_CRASH_AT 0 // each compilation increment this #endif static g_malloc_count = 0; void *malloc(size_t size) { if (g_malloc_count == MALLOC_CRASH) { ++g_malloc_count; return (NULL); } ++g_malloc_count; return (malloc(size)); // this is the real malloc but name conflicts... } ``` `MALLOC_CRASH_AT` est incrémenté à chaque compilation et à chaque fois le programme passe une série de tests
test
test de retour null de malloc idées je pense qu il est possible de tester les retours null de malloc de manière propre on pourrait charger notre propre librarie dynamique malloc contenant malloc realloc free et wrapper les vrais fonctions du système conflicts de noms en les glissants dans ce genre de scope c ifnedef define malloc crash at each compilation increment this endif static g malloc count void malloc size t size if g malloc count malloc crash g malloc count return null g malloc count return malloc size this is the real malloc but name conflicts malloc crash at est incrémenté à chaque compilation et à chaque fois le programme passe une série de tests
1
4,586
7,428,876,262
IssuesEvent
2018-03-24 07:33:06
kookmin-sw/2018-cap1-2
https://api.github.com/repos/kookmin-sw/2018-cap1-2
opened
암묵적 의미가 동일한 기호처리
ImageProcessing
for 구문에서 동일한 의미를 가지는 -> ~ 기호가 들어왔을 때, <현재> - 인터프리터 대상에서 제외되는 -> 기호를 그대로 반환해서 로지컬 에러 발생 이를 영상처리 단계에서 -> 기호를 ~ 기호로 대체 변환해줘 인터프리팅 성공 가능성을 높이는 것에 대해 고려. + >=, <=, != 기호를 하나의 기호로 합치는 것이 기호 표기 상 가능하다면 해당 기호들도 고려 대상에 포함.
1.0
암묵적 의미가 동일한 기호처리 - for 구문에서 동일한 의미를 가지는 -> ~ 기호가 들어왔을 때, <현재> - 인터프리터 대상에서 제외되는 -> 기호를 그대로 반환해서 로지컬 에러 발생 이를 영상처리 단계에서 -> 기호를 ~ 기호로 대체 변환해줘 인터프리팅 성공 가능성을 높이는 것에 대해 고려. + >=, <=, != 기호를 하나의 기호로 합치는 것이 기호 표기 상 가능하다면 해당 기호들도 고려 대상에 포함.
non_test
암묵적 의미가 동일한 기호처리 for 구문에서 동일한 의미를 가지는 기호가 들어왔을 때 인터프리터 대상에서 제외되는 기호를 그대로 반환해서 로지컬 에러 발생 이를 영상처리 단계에서 기호를 기호로 대체 변환해줘 인터프리팅 성공 가능성을 높이는 것에 대해 고려 기호를 하나의 기호로 합치는 것이 기호 표기 상 가능하다면 해당 기호들도 고려 대상에 포함
0
233,067
18,945,405,528
IssuesEvent
2021-11-18 09:38:33
open-telemetry/opentelemetry-collector-contrib
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
opened
Flaky test: ??? on tailsamplingprocessor
bug flaky test
Seen on https://github.com/open-telemetry/opentelemetry-collector-contrib/runs/4243284022?check_suite_focus=true Full logs: [logs_18069.zip](https://github.com/open-telemetry/opentelemetry-collector-contrib/files/7561721/logs_18069.zip) ``` === RUN TestStatusCodeSampling/filter_on_OK_and_UNSET_-_matches --- PASS: TestStatusCodeSampling (0.00s) --- PASS: TestStatusCodeSampling/filter_on_ERROR_-_none_match (0.00s) --- PASS: TestStatusCodeSampling/filter_on_OK_and_ERROR_-_none_match (0.00s) make[2]: *** [../../Makefile.Common:45: test] Error 1 make[1]: *** [Makefile:149: for-all-target-./processor/tailsamplingprocessor] Error 2 make: *** [Makefile:69: gotest] Error 2 --- PASS: TestStatusCodeSampling/filter_on_UNSET_-_matches (0.00s) --- PASS: TestStatusCodeSampling/filter_on_OK_and_UNSET_-_matches (0.00s) === RUN TestOnLateArrivingSpans_StatusCodeSampling --- PASS: TestOnLateArrivingSpans_StatusCodeSampling (0.00s) === RUN TestStringTagFilter === RUN TestStringTagFilter/nonmatching_node_attribute_key === RUN TestStringTagFilter/nonmatching_node_attribute_value === RUN TestStringTagFilter/matching_node_attribute === RUN TestStringTagFilter/nonmatching_span_attribute_key === RUN TestStringTagFilter/nonmatching_span_attribute_value === RUN TestStringTagFilter/matching_span_attribute === RUN TestStringTagFilter/matching_span_attribute_with_regex === RUN TestStringTagFilter/nonmatching_span_attribute_with_regex === RUN TestStringTagFilter/matching_span_attribute_with_regex_without_CacheSize_provided_in_config === RUN TestStringTagFilter/matching_plain_text_node_attribute_in_regex === RUN TestStringTagFilter/nonmatching_span_attribute_on_empty_filter_list === RUN TestStringTagFilter/invert_nonmatching_node_attribute_key === RUN TestStringTagFilter/invert_nonmatching_node_attribute_value === RUN TestStringTagFilter/invert_nonmatching_node_attribute_list === RUN TestStringTagFilter/invert_matching_node_attribute === RUN TestStringTagFilter/invert_matching_node_attribute_list === RUN TestStringTagFilter/invert_nonmatching_span_attribute_key === RUN TestStringTagFilter/invert_nonmatching_span_attribute_value === RUN TestStringTagFilter/invert_nonmatching_span_attribute_list === RUN TestStringTagFilter/invert_matching_span_attribute === RUN TestStringTagFilter/invert_matching_span_attribute_list === RUN TestStringTagFilter/invert_matching_span_attribute_with_regex === RUN TestStringTagFilter/invert_matching_span_attribute_with_regex_list === RUN TestStringTagFilter/invert_nonmatching_span_attribute_with_regex === RUN TestStringTagFilter/invert_nonmatching_span_attribute_with_regex_list === RUN TestStringTagFilter/invert_matching_plain_text_node_attribute_in_regex === RUN TestStringTagFilter/invert_matching_plain_text_node_attribute_in_regex_list === RUN TestStringTagFilter/invert_nonmatching_span_attribute_on_empty_filter_list --- PASS: TestStringTagFilter (0.01s) --- PASS: TestStringTagFilter/nonmatching_node_attribute_key (0.00s) --- PASS: TestStringTagFilter/nonmatching_node_attribute_value (0.00s) --- PASS: TestStringTagFilter/matching_node_attribute (0.00s) --- PASS: TestStringTagFilter/nonmatching_span_attribute_key (0.00s) --- PASS: TestStringTagFilter/nonmatching_span_attribute_value (0.00s) --- PASS: TestStringTagFilter/matching_span_attribute (0.00s) --- PASS: TestStringTagFilter/matching_span_attribute_with_regex (0.00s) --- PASS: TestStringTagFilter/nonmatching_span_attribute_with_regex (0.00s) --- PASS: TestStringTagFilter/matching_span_attribute_with_regex_without_CacheSize_provided_in_config (0.00s) --- PASS: TestStringTagFilter/matching_plain_text_node_attribute_in_regex (0.00s) --- PASS: TestStringTagFilter/nonmatching_span_attribute_on_empty_filter_list (0.00s) --- PASS: TestStringTagFilter/invert_nonmatching_node_attribute_key (0.00s) --- PASS: TestStringTagFilter/invert_nonmatching_node_attribute_value (0.00s) --- PASS: TestStringTagFilter/invert_nonmatching_node_attribute_list (0.00s) --- PASS: TestStringTagFilter/invert_matching_node_attribute (0.00s) --- PASS: TestStringTagFilter/invert_matching_node_attribute_list (0.00s) --- PASS: TestStringTagFilter/invert_nonmatching_span_attribute_key (0.00s) --- PASS: TestStringTagFilter/invert_nonmatching_span_attribute_value (0.00s) --- PASS: TestStringTagFilter/invert_nonmatching_span_attribute_list (0.00s) --- PASS: TestStringTagFilter/invert_matching_span_attribute (0.00s) --- PASS: TestStringTagFilter/invert_matching_span_attribute_list (0.00s) --- PASS: TestStringTagFilter/invert_matching_span_attribute_with_regex (0.00s) --- PASS: TestStringTagFilter/invert_matching_span_attribute_with_regex_list (0.00s) --- PASS: TestStringTagFilter/invert_nonmatching_span_attribute_with_regex (0.00s) --- PASS: TestStringTagFilter/invert_nonmatching_span_attribute_with_regex_list (0.00s) --- PASS: TestStringTagFilter/invert_matching_plain_text_node_attribute_in_regex (0.00s) --- PASS: TestStringTagFilter/invert_matching_plain_text_node_attribute_in_regex_list (0.00s) --- PASS: TestStringTagFilter/invert_nonmatching_span_attribute_on_empty_filter_list (0.00s) === RUN TestOnLateArrivingSpans_StringAttribute --- PASS: TestOnLateArrivingSpans_StringAttribute (0.00s) === RUN TestTimeProvider --- PASS: TestTimeProvider (0.00s) PASS ok github.com/open-telemetry/opentelemetry-collector-contrib/processor/tailsamplingprocessor/internal/sampling 4.407s FAIL make[2]: Leaving directory '/home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/tailsamplingprocessor' make[1]: Leaving directory '/home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib' ```
1.0
Flaky test: ??? on tailsamplingprocessor - Seen on https://github.com/open-telemetry/opentelemetry-collector-contrib/runs/4243284022?check_suite_focus=true Full logs: [logs_18069.zip](https://github.com/open-telemetry/opentelemetry-collector-contrib/files/7561721/logs_18069.zip) ``` === RUN TestStatusCodeSampling/filter_on_OK_and_UNSET_-_matches --- PASS: TestStatusCodeSampling (0.00s) --- PASS: TestStatusCodeSampling/filter_on_ERROR_-_none_match (0.00s) --- PASS: TestStatusCodeSampling/filter_on_OK_and_ERROR_-_none_match (0.00s) make[2]: *** [../../Makefile.Common:45: test] Error 1 make[1]: *** [Makefile:149: for-all-target-./processor/tailsamplingprocessor] Error 2 make: *** [Makefile:69: gotest] Error 2 --- PASS: TestStatusCodeSampling/filter_on_UNSET_-_matches (0.00s) --- PASS: TestStatusCodeSampling/filter_on_OK_and_UNSET_-_matches (0.00s) === RUN TestOnLateArrivingSpans_StatusCodeSampling --- PASS: TestOnLateArrivingSpans_StatusCodeSampling (0.00s) === RUN TestStringTagFilter === RUN TestStringTagFilter/nonmatching_node_attribute_key === RUN TestStringTagFilter/nonmatching_node_attribute_value === RUN TestStringTagFilter/matching_node_attribute === RUN TestStringTagFilter/nonmatching_span_attribute_key === RUN TestStringTagFilter/nonmatching_span_attribute_value === RUN TestStringTagFilter/matching_span_attribute === RUN TestStringTagFilter/matching_span_attribute_with_regex === RUN TestStringTagFilter/nonmatching_span_attribute_with_regex === RUN TestStringTagFilter/matching_span_attribute_with_regex_without_CacheSize_provided_in_config === RUN TestStringTagFilter/matching_plain_text_node_attribute_in_regex === RUN TestStringTagFilter/nonmatching_span_attribute_on_empty_filter_list === RUN TestStringTagFilter/invert_nonmatching_node_attribute_key === RUN TestStringTagFilter/invert_nonmatching_node_attribute_value === RUN TestStringTagFilter/invert_nonmatching_node_attribute_list === RUN TestStringTagFilter/invert_matching_node_attribute === RUN TestStringTagFilter/invert_matching_node_attribute_list === RUN TestStringTagFilter/invert_nonmatching_span_attribute_key === RUN TestStringTagFilter/invert_nonmatching_span_attribute_value === RUN TestStringTagFilter/invert_nonmatching_span_attribute_list === RUN TestStringTagFilter/invert_matching_span_attribute === RUN TestStringTagFilter/invert_matching_span_attribute_list === RUN TestStringTagFilter/invert_matching_span_attribute_with_regex === RUN TestStringTagFilter/invert_matching_span_attribute_with_regex_list === RUN TestStringTagFilter/invert_nonmatching_span_attribute_with_regex === RUN TestStringTagFilter/invert_nonmatching_span_attribute_with_regex_list === RUN TestStringTagFilter/invert_matching_plain_text_node_attribute_in_regex === RUN TestStringTagFilter/invert_matching_plain_text_node_attribute_in_regex_list === RUN TestStringTagFilter/invert_nonmatching_span_attribute_on_empty_filter_list --- PASS: TestStringTagFilter (0.01s) --- PASS: TestStringTagFilter/nonmatching_node_attribute_key (0.00s) --- PASS: TestStringTagFilter/nonmatching_node_attribute_value (0.00s) --- PASS: TestStringTagFilter/matching_node_attribute (0.00s) --- PASS: TestStringTagFilter/nonmatching_span_attribute_key (0.00s) --- PASS: TestStringTagFilter/nonmatching_span_attribute_value (0.00s) --- PASS: TestStringTagFilter/matching_span_attribute (0.00s) --- PASS: TestStringTagFilter/matching_span_attribute_with_regex (0.00s) --- PASS: TestStringTagFilter/nonmatching_span_attribute_with_regex (0.00s) --- PASS: TestStringTagFilter/matching_span_attribute_with_regex_without_CacheSize_provided_in_config (0.00s) --- PASS: TestStringTagFilter/matching_plain_text_node_attribute_in_regex (0.00s) --- PASS: TestStringTagFilter/nonmatching_span_attribute_on_empty_filter_list (0.00s) --- PASS: TestStringTagFilter/invert_nonmatching_node_attribute_key (0.00s) --- PASS: TestStringTagFilter/invert_nonmatching_node_attribute_value (0.00s) --- PASS: TestStringTagFilter/invert_nonmatching_node_attribute_list (0.00s) --- PASS: TestStringTagFilter/invert_matching_node_attribute (0.00s) --- PASS: TestStringTagFilter/invert_matching_node_attribute_list (0.00s) --- PASS: TestStringTagFilter/invert_nonmatching_span_attribute_key (0.00s) --- PASS: TestStringTagFilter/invert_nonmatching_span_attribute_value (0.00s) --- PASS: TestStringTagFilter/invert_nonmatching_span_attribute_list (0.00s) --- PASS: TestStringTagFilter/invert_matching_span_attribute (0.00s) --- PASS: TestStringTagFilter/invert_matching_span_attribute_list (0.00s) --- PASS: TestStringTagFilter/invert_matching_span_attribute_with_regex (0.00s) --- PASS: TestStringTagFilter/invert_matching_span_attribute_with_regex_list (0.00s) --- PASS: TestStringTagFilter/invert_nonmatching_span_attribute_with_regex (0.00s) --- PASS: TestStringTagFilter/invert_nonmatching_span_attribute_with_regex_list (0.00s) --- PASS: TestStringTagFilter/invert_matching_plain_text_node_attribute_in_regex (0.00s) --- PASS: TestStringTagFilter/invert_matching_plain_text_node_attribute_in_regex_list (0.00s) --- PASS: TestStringTagFilter/invert_nonmatching_span_attribute_on_empty_filter_list (0.00s) === RUN TestOnLateArrivingSpans_StringAttribute --- PASS: TestOnLateArrivingSpans_StringAttribute (0.00s) === RUN TestTimeProvider --- PASS: TestTimeProvider (0.00s) PASS ok github.com/open-telemetry/opentelemetry-collector-contrib/processor/tailsamplingprocessor/internal/sampling 4.407s FAIL make[2]: Leaving directory '/home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/tailsamplingprocessor' make[1]: Leaving directory '/home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib' ```
test
flaky test on tailsamplingprocessor seen on full logs run teststatuscodesampling filter on ok and unset matches pass teststatuscodesampling pass teststatuscodesampling filter on error none match pass teststatuscodesampling filter on ok and error none match make error make error make error pass teststatuscodesampling filter on unset matches pass teststatuscodesampling filter on ok and unset matches run testonlatearrivingspans statuscodesampling pass testonlatearrivingspans statuscodesampling run teststringtagfilter run teststringtagfilter nonmatching node attribute key run teststringtagfilter nonmatching node attribute value run teststringtagfilter matching node attribute run teststringtagfilter nonmatching span attribute key run teststringtagfilter nonmatching span attribute value run teststringtagfilter matching span attribute run teststringtagfilter matching span attribute with regex run teststringtagfilter nonmatching span attribute with regex run teststringtagfilter matching span attribute with regex without cachesize provided in config run teststringtagfilter matching plain text node attribute in regex run teststringtagfilter nonmatching span attribute on empty filter list run teststringtagfilter invert nonmatching node attribute key run teststringtagfilter invert nonmatching node attribute value run teststringtagfilter invert nonmatching node attribute list run teststringtagfilter invert matching node attribute run teststringtagfilter invert matching node attribute list run teststringtagfilter invert nonmatching span attribute key run teststringtagfilter invert nonmatching span attribute value run teststringtagfilter invert nonmatching span attribute list run teststringtagfilter invert matching span attribute run teststringtagfilter invert matching span attribute list run teststringtagfilter invert matching span attribute with regex run teststringtagfilter invert matching span attribute with regex list run teststringtagfilter invert nonmatching span attribute with regex run teststringtagfilter invert nonmatching span attribute with regex list run teststringtagfilter invert matching plain text node attribute in regex run teststringtagfilter invert matching plain text node attribute in regex list run teststringtagfilter invert nonmatching span attribute on empty filter list pass teststringtagfilter pass teststringtagfilter nonmatching node attribute key pass teststringtagfilter nonmatching node attribute value pass teststringtagfilter matching node attribute pass teststringtagfilter nonmatching span attribute key pass teststringtagfilter nonmatching span attribute value pass teststringtagfilter matching span attribute pass teststringtagfilter matching span attribute with regex pass teststringtagfilter nonmatching span attribute with regex pass teststringtagfilter matching span attribute with regex without cachesize provided in config pass teststringtagfilter matching plain text node attribute in regex pass teststringtagfilter nonmatching span attribute on empty filter list pass teststringtagfilter invert nonmatching node attribute key pass teststringtagfilter invert nonmatching node attribute value pass teststringtagfilter invert nonmatching node attribute list pass teststringtagfilter invert matching node attribute pass teststringtagfilter invert matching node attribute list pass teststringtagfilter invert nonmatching span attribute key pass teststringtagfilter invert nonmatching span attribute value pass teststringtagfilter invert nonmatching span attribute list pass teststringtagfilter invert matching span attribute pass teststringtagfilter invert matching span attribute list pass teststringtagfilter invert matching span attribute with regex pass teststringtagfilter invert matching span attribute with regex list pass teststringtagfilter invert nonmatching span attribute with regex pass teststringtagfilter invert nonmatching span attribute with regex list pass teststringtagfilter invert matching plain text node attribute in regex pass teststringtagfilter invert matching plain text node attribute in regex list pass teststringtagfilter invert nonmatching span attribute on empty filter list run testonlatearrivingspans stringattribute pass testonlatearrivingspans stringattribute run testtimeprovider pass testtimeprovider pass ok github com open telemetry opentelemetry collector contrib processor tailsamplingprocessor internal sampling fail make leaving directory home runner work opentelemetry collector contrib opentelemetry collector contrib processor tailsamplingprocessor make leaving directory home runner work opentelemetry collector contrib opentelemetry collector contrib
1
56,620
14,078,460,808
IssuesEvent
2020-11-04 13:36:20
themagicalmammal/android_kernel_samsung_a3xelte
https://api.github.com/repos/themagicalmammal/android_kernel_samsung_a3xelte
opened
CVE-2018-20669 (High) detected in multiple libraries
security vulnerability
## CVE-2018-20669 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linuxv3.10</b>, <b>linuxlinux-3.10.11</b>, <b>linuxlinux-3.5</b></p></summary> <p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue where a provided address with access_ok() is not checked was discovered in i915_gem_execbuffer2_ioctl in drivers/gpu/drm/i915/i915_gem_execbuffer.c in the Linux kernel through 4.19.13. A local attacker can craft a malicious IOCTL function call to overwrite arbitrary kernel memory, resulting in a Denial of Service or privilege escalation. <p>Publish Date: 2019-03-21 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20669>CVE-2018-20669</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20669">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20669</a></p> <p>Release Date: 2019-03-21</p> <p>Fix Resolution: v5.0-rc1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-20669 (High) detected in multiple libraries - ## CVE-2018-20669 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linuxv3.10</b>, <b>linuxlinux-3.10.11</b>, <b>linuxlinux-3.5</b></p></summary> <p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue where a provided address with access_ok() is not checked was discovered in i915_gem_execbuffer2_ioctl in drivers/gpu/drm/i915/i915_gem_execbuffer.c in the Linux kernel through 4.19.13. A local attacker can craft a malicious IOCTL function call to overwrite arbitrary kernel memory, resulting in a Denial of Service or privilege escalation. <p>Publish Date: 2019-03-21 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20669>CVE-2018-20669</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20669">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20669</a></p> <p>Release Date: 2019-03-21</p> <p>Fix Resolution: v5.0-rc1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in multiple libraries cve high severity vulnerability vulnerable libraries linuxlinux linuxlinux vulnerability details an issue where a provided address with access ok is not checked was discovered in gem ioctl in drivers gpu drm gem execbuffer c in the linux kernel through a local attacker can craft a malicious ioctl function call to overwrite arbitrary kernel memory resulting in a denial of service or privilege escalation publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
330,729
28,484,922,517
IssuesEvent
2023-04-18 07:09:34
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
closed
Fix exponents_and_logarithms.test_numpy_exp2
NumPy Frontend Sub Task Failing Test
| | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4722833970/jobs/8377976075" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/4722833970/jobs/8377965187" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4722833970/jobs/8377966037" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/4645654528/jobs/8221610959" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> <details> <summary>Not found</summary> Not found </details> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_numpy/test_mathematical_functions/test_exponents_and_logarithms.py::test_numpy_exp2[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-04-08T14:48:36.1956634Z E IndexError: list index out of range2023-04-08T14:48:36.1956957Z E Falsifying example: test_numpy_exp2(2023-04-08T14:48:36.1957633Z E dtypes_values_casting=(['float16'],2023-04-08T14:48:36.1958008Z E [array(-1., dtype=float16)],2023-04-08T14:48:36.1958298Z E 'no',2023-04-08T14:48:36.1958524Z E None),2023-04-08T14:48:36.1958779Z E where=[array(False)],2023-04-08T14:48:36.1959097Z E test_flags=FrontendFunctionTestFlags(2023-04-08T14:48:36.1959672Z E num_positional_args=1,2023-04-08T14:48:36.1959941Z E with_out=False,2023-04-08T14:48:36.1960185Z E inplace=False,2023-04-08T14:48:36.1960452Z E as_variable=[False],2023-04-08T14:48:36.1960727Z E native_arrays=[False],2023-04-08T14:48:36.1961021Z E generate_frontend_arrays=True,2023-04-08T14:48:36.1961282Z E ),2023-04-08T14:48:36.1961668Z E fn_tree='ivy.functional.frontends.numpy.exp2',2023-04-08T14:48:36.1962040Z E frontend='numpy',2023-04-08T14:48:36.1962335Z E on_device='cpu',2023-04-08T14:48:36.1962568Z E )2023-04-08T14:48:36.1962772Z E 2023-04-08T14:48:36.1963383Z E You can reproduce this example by temporarily adding @reproduce_failure('6.71.0', b'AXicY2AAAUYGBiSaEQAALAAE') as a decorator on your test case </details>
1.0
Fix exponents_and_logarithms.test_numpy_exp2 - | | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4722833970/jobs/8377976075" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/4722833970/jobs/8377965187" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4722833970/jobs/8377966037" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/4645654528/jobs/8221610959" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> <details> <summary>Not found</summary> Not found </details> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_numpy/test_mathematical_functions/test_exponents_and_logarithms.py::test_numpy_exp2[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-04-08T14:48:36.1956634Z E IndexError: list index out of range2023-04-08T14:48:36.1956957Z E Falsifying example: test_numpy_exp2(2023-04-08T14:48:36.1957633Z E dtypes_values_casting=(['float16'],2023-04-08T14:48:36.1958008Z E [array(-1., dtype=float16)],2023-04-08T14:48:36.1958298Z E 'no',2023-04-08T14:48:36.1958524Z E None),2023-04-08T14:48:36.1958779Z E where=[array(False)],2023-04-08T14:48:36.1959097Z E test_flags=FrontendFunctionTestFlags(2023-04-08T14:48:36.1959672Z E num_positional_args=1,2023-04-08T14:48:36.1959941Z E with_out=False,2023-04-08T14:48:36.1960185Z E inplace=False,2023-04-08T14:48:36.1960452Z E as_variable=[False],2023-04-08T14:48:36.1960727Z E native_arrays=[False],2023-04-08T14:48:36.1961021Z E generate_frontend_arrays=True,2023-04-08T14:48:36.1961282Z E ),2023-04-08T14:48:36.1961668Z E fn_tree='ivy.functional.frontends.numpy.exp2',2023-04-08T14:48:36.1962040Z E frontend='numpy',2023-04-08T14:48:36.1962335Z E on_device='cpu',2023-04-08T14:48:36.1962568Z E )2023-04-08T14:48:36.1962772Z E 2023-04-08T14:48:36.1963383Z E You can reproduce this example by temporarily adding @reproduce_failure('6.71.0', b'AXicY2AAAUYGBiSaEQAALAAE') as a decorator on your test case </details>
test
fix exponents and logarithms test numpy tensorflow img src torch img src numpy img src jax img src not found not found failed ivy tests test ivy test frontends test numpy test mathematical functions test exponents and logarithms py test numpy e indexerror list index out of e falsifying example test numpy e dtypes values casting e e no e none e where e test flags frontendfunctiontestflags e num positional args e with out false e inplace false e as variable e native arrays e generate frontend arrays true e e fn tree ivy functional frontends numpy e frontend numpy e on device cpu e e e you can reproduce this example by temporarily adding reproduce failure b as a decorator on your test case
1
176,122
13,627,748,350
IssuesEvent
2020-09-24 13:02:29
Scholar-6/brillder
https://api.github.com/repos/Scholar-6/brillder
closed
Ticket #67 Synthesis has gone missing
Betatester Request Critical Blocker
https://brillder.scholar6.org/build/brick/257/investigation/synthesis A lot of my content has seemingly disappeared from... Requester: Sophie Assignee: Support A lot of my content has seemingly disappeared from the editor. It's still present in the mobile-screen-preview, but it seems to be missing in the editor itself. ------------------ Submitted from: https://brillder.scholar6.org/build/brick/257/investigation/question-component/30403 ![](https://scholar6.zendesk.com/attachments/token/fpcUMJhoYkSxt2LjMCFe54urg/?name=Screenshot+2020-09-24+at+09.17.20.png) ![](https://scholar6.zendesk.com/attachments/token/3ZQ5HLw9KIifyrWssZnNDMxl7/?name=Screenshot+2020-09-24+at+09.17.07.png) ![](https://scholar6.zendesk.com/attachments/token/HE4itSLEj1vFtrpSEjheBlF8H/?name=Screenshot+2020-09-24+at+09.17.42.png) - [x] seems Wiris related as pure text components still showing and editable - [x] could it also be because of new fixed question text component?
1.0
Ticket #67 Synthesis has gone missing - https://brillder.scholar6.org/build/brick/257/investigation/synthesis A lot of my content has seemingly disappeared from... Requester: Sophie Assignee: Support A lot of my content has seemingly disappeared from the editor. It's still present in the mobile-screen-preview, but it seems to be missing in the editor itself. ------------------ Submitted from: https://brillder.scholar6.org/build/brick/257/investigation/question-component/30403 ![](https://scholar6.zendesk.com/attachments/token/fpcUMJhoYkSxt2LjMCFe54urg/?name=Screenshot+2020-09-24+at+09.17.20.png) ![](https://scholar6.zendesk.com/attachments/token/3ZQ5HLw9KIifyrWssZnNDMxl7/?name=Screenshot+2020-09-24+at+09.17.07.png) ![](https://scholar6.zendesk.com/attachments/token/HE4itSLEj1vFtrpSEjheBlF8H/?name=Screenshot+2020-09-24+at+09.17.42.png) - [x] seems Wiris related as pure text components still showing and editable - [x] could it also be because of new fixed question text component?
test
ticket synthesis has gone missing a lot of my content has seemingly disappeared from requester sophie assignee support a lot of my content has seemingly disappeared from the editor it s still present in the mobile screen preview but it seems to be missing in the editor itself submitted from seems wiris related as pure text components still showing and editable could it also be because of new fixed question text component
1
19,762
3,779,582,586
IssuesEvent
2016-03-18 09:07:10
koying/SPMC
https://api.github.com/repos/koying/SPMC
closed
Generic issue with Android < 4.4
bug To be tested
When I open SPMC 16.1 , It crashed. But SPMC 15.0 and KODI 16.1 is OK. System info from SPMC: CPU : ARMV7 Processor rev1(V7I) GPU : Mail-450MP System : Android 4.3.0 API Level18 (kernel:linux 3.1.10) OpenGL vendor : ARM OpenGl version :OpenGL ES2.0
1.0
Generic issue with Android < 4.4 - When I open SPMC 16.1 , It crashed. But SPMC 15.0 and KODI 16.1 is OK. System info from SPMC: CPU : ARMV7 Processor rev1(V7I) GPU : Mail-450MP System : Android 4.3.0 API Level18 (kernel:linux 3.1.10) OpenGL vendor : ARM OpenGl version :OpenGL ES2.0
test
generic issue with android when i open spmc it crashed but spmc and kodi is ok system info from spmc cpu processor gpu mail system android api kernel linux opengl vendor arm opengl version opengl
1
328,167
24,172,873,737
IssuesEvent
2022-09-22 21:01:12
fleetdm/fleet
https://api.github.com/repos/fleetdm/fleet
opened
Document using a password manager during set up is the best practice
:improve documentation
### Goals Test that using a password manager (ex. 1password) works when setting up Fleet and logging into Fleet. In the "Running the Fleet server" docs: Document that using a password manager is the best practice when setting up Fleet: https://fleetdm.com/docs/deploying/server-installation#running-the-fleet-server This is because, if you forget your password, you will have to wipe the database or redeploy Fleet.
1.0
Document using a password manager during set up is the best practice - ### Goals Test that using a password manager (ex. 1password) works when setting up Fleet and logging into Fleet. In the "Running the Fleet server" docs: Document that using a password manager is the best practice when setting up Fleet: https://fleetdm.com/docs/deploying/server-installation#running-the-fleet-server This is because, if you forget your password, you will have to wipe the database or redeploy Fleet.
non_test
document using a password manager during set up is the best practice goals test that using a password manager ex works when setting up fleet and logging into fleet in the running the fleet server docs document that using a password manager is the best practice when setting up fleet this is because if you forget your password you will have to wipe the database or redeploy fleet
0
284,791
21,470,291,549
IssuesEvent
2022-04-26 08:54:17
OpenRefine/OpenRefine
https://api.github.com/repos/OpenRefine/OpenRefine
closed
Update user manual and in-app documentation for splitByLength()
bug documentation grel
The documentation, specifically for splitByLength() does not inform the user that they have the possibility to only return values within a given range of indices, Iv since updated the document but wanted to make a bug report on it for submitting the PR. ex: `"123456789012345678901234567890".splitByLengths(3,4,5)[0,2]` `returns [ "123", "4567" ]` More about this issue can be seen in this thread: https://github.com/OpenRefine/OpenRefine/issues/3929
1.0
Update user manual and in-app documentation for splitByLength() - The documentation, specifically for splitByLength() does not inform the user that they have the possibility to only return values within a given range of indices, Iv since updated the document but wanted to make a bug report on it for submitting the PR. ex: `"123456789012345678901234567890".splitByLengths(3,4,5)[0,2]` `returns [ "123", "4567" ]` More about this issue can be seen in this thread: https://github.com/OpenRefine/OpenRefine/issues/3929
non_test
update user manual and in app documentation for splitbylength the documentation specifically for splitbylength does not inform the user that they have the possibility to only return values within a given range of indices iv since updated the document but wanted to make a bug report on it for submitting the pr ex splitbylengths returns more about this issue can be seen in this thread
0
41,745
5,395,699,608
IssuesEvent
2017-02-27 09:29:18
mautic/mautic
https://api.github.com/repos/mautic/mautic
closed
Unable To See focus data
Pending Feedback Ready To Test
What type of report is this: | Q | A | ---| --- | Bug report? | Y | Feature request? | N | Enhancement? | ## Description: While saving (SAVE and CLOSE). it stuck. and then you can go back and try to click on a link. you will get only publish detail in rough data. You will not get the script. ## If a bug: | Q | A | --- | --- | Mautic version|2.6.0 | PHP version| 5.6 ### Steps to reproduce: 1. Create focus. 2. While saving (SAVE and CLOSE). it stuck. I can't enter in it to get script link. ### Log errors: I have created the focus but unable to see it. Mean I can't enter in it to get script link. ... (Version 2.6.0) @escopecz [2017-02-23 10:37:50] mautic.WARNING: PHP Warning - fwrite() expects parameter 1 to be resource, boolean given - in file /home/mona/mautic-master/app/bundles/CoreBundle/IpLookup/AbstractLocalDataLookup.php - at line 98 [] [] [2017-02-23 10:37:50] mautic.WARNING: PHP Warning - fwrite() expects parameter 1 to be resource, boolean given - in file /home/mona/mautic-master/app/bundles/CoreBundle/IpLookup/AbstractLocalDataLookup.php - at line 98 [] [] [2017-02-23 10:37:50] mautic.WARNING: PHP Warning - fwrite() expects parameter 1 to be resource, boolean given - in file /home/mona/mautic-master/app/bundles/CoreBundle/IpLookup/AbstractLocalDataLookup.php - at line 98 [] [] [2017-02-23 10:37:50] mautic.WARNING: PHP Warning - fwrite() expects parameter 1 to be resource, boolean given - in file /home/mona/mautic-master/app/bundles/CoreBundle/IpLookup/AbstractLocalDataLookup.php - at line 98 [] [] [2017-02-23 10:37:50] mautic.WARNING: PHP Warning - fclose() expects parameter 1 to be resource, boolean given - in file /home/mona/mautic-master/app/bundles/CoreBundle/IpLookup/AbstractLocalDataLookup.php - at line 100 [] [] [2017-02-23 10:37:59] mautic.NOTICE: PHP Notice - Undefined variable: dateRangeForm - in file /home/mona/mautic-master/plugins/MauticFocusBundle/Views/Focus/details.html.php - at line 103 [] [] [2017-02-23 10:37:59] mautic.ERROR: PHP Error - Argument 1 passed to Symfony\Bundle\FrameworkBundle\Templating\Helper\FormHelper::start() must be an instance of Symfony\Component\Form\FormView, null given, called in /home/mona/mautic-master/app/bundles/CoreBundle/Views/Helper/graph_dateselect.html.php on line 16 and defined - in file /home/mona/mautic-master/vendor/symfony/framework-bundle/Templating/Helper/FormHelper.php - at line 101 [] [] [2017-02-23 10:38:03] mautic.CRITICAL: Uncaught PHP Exception InvalidArgumentException: "Method getTimezone not exists" at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php line 39 {"exception":"[object] (InvalidArgumentException(code: 0): Method getTimezone not exists at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php:39)"} [] [2017-02-23 10:38:03] mautic.NOTICE: PHP Notice - Undefined variable: dateRangeForm - in file /home/mona/mautic-master/plugins/MauticFocusBundle/Views/Focus/details.html.php - at line 103 [] [] [2017-02-23 10:38:03] mautic.ERROR: PHP Error - Argument 1 passed to Symfony\Bundle\FrameworkBundle\Templating\Helper\FormHelper::start() must be an instance of Symfony\Component\Form\FormView, null given, called in /home/mona/mautic-master/app/bundles/CoreBundle/Views/Helper/graph_dateselect.html.php on line 16 and defined - in file /home/mona/mautic-master/vendor/symfony/framework-bundle/Templating/Helper/FormHelper.php - at line 101 [] [] [2017-02-23 10:38:15] mautic.NOTICE: PHP Notice - Undefined variable: dateRangeForm - in file /home/mona/mautic-master/plugins/MauticFocusBundle/Views/Focus/details.html.php - at line 103 [] [] [2017-02-23 10:38:15] mautic.ERROR: PHP Error - Argument 1 passed to Symfony\Bundle\FrameworkBundle\Templating\Helper\FormHelper::start() must be an instance of Symfony\Component\Form\FormView, null given, called in /home/mona/mautic-master/app/bundles/CoreBundle/Views/Helper/graph_dateselect.html.php on line 16 and defined - in file /home/mona/mautic-master/vendor/symfony/framework-bundle/Templating/Helper/FormHelper.php - at line 101 [] [] [2017-02-23 10:38:30] mautic.CRITICAL: Uncaught PHP Exception InvalidArgumentException: "Method getTimezone not exists" at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php line 39 {"exception":"[object] (InvalidArgumentException(code: 0): Method getTimezone not exists at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php:39)"} [] [2017-02-23 10:38:52] mautic.CRITICAL: Uncaught PHP Exception InvalidArgumentException: "Method getTimezone not exists" at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php line 39 {"exception":"[object] (InvalidArgumentException(code: 0): Method getTimezone not exists at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php:39)"} [] [2017-02-23 10:38:56] mautic.CRITICAL: Uncaught PHP Exception InvalidArgumentException: "Method getTimezone not exists" at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php line 39 {"exception":"[object] (InvalidArgumentException(code: 0): Method getTimezone not exists at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php:39)"} [] [2017-02-23 10:38:57] mautic.NOTICE: PHP Notice - Undefined variable: dateRangeForm - in file /home/mona/mautic-master/plugins/MauticFocusBundle/Views/Focus/details.html.php - at line 103 [] [] [2017-02-23 10:38:57] mautic.ERROR: PHP Error - Argument 1 passed to Symfony\Bundle\FrameworkBundle\Templating\Helper\FormHelper::start() must be an instance of Symfony\Component\Form\FormView, null given, called in /home/mona/mautic-master/app/bundles/CoreBundle/Views/Helper/graph_dateselect.html.php on line 16 and defined - in file /home/mona/mautic-master/vendor/symfony/framework-bundle/Templating/Helper/FormHelper.php - at line 101 [] [] [2017-02-23 10:39:05] mautic.CRITICAL: Uncaught PHP Exception InvalidArgumentException: "Method getTimezone not exists" at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php line 39 {"exception":"[object] (InvalidArgumentException(code: 0): Method getTimezone not exists at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php:39)"} [] [2017-02-23 10:39:12] mautic.CRITICAL: Uncaught PHP Exception InvalidArgumentException: "Method getTimezone not exists" at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php line 39 {"exception":"[object] (InvalidArgumentException(code: 0): Method getTimezone not exists at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php:39)"} [] [2017-02-23 10:39:18] mautic.CRITICAL: Uncaught PHP Exception InvalidArgumentException: "Method getTimezone not exists" at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php line 39 {"exception":"[object] (InvalidArgumentException(code: 0): Method getTimezone not exists at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php:39)"} [] [2017-02-23 10:39:41] mautic.CRITICAL: Uncaught PHP Exception InvalidArgumentException: "Method getTimezone not exists" at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php line 39 {"exception":"[object] (InvalidArgumentException(code: 0): Method getTimezone not exists at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php:39)"} [] [2017-02-23 10:39:49] mautic.CRITICAL: Uncaught PHP Exception InvalidArgumentException: "Method getTimezone not exists" at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php line 39 {"exception":"[object] (InvalidArgumentException(code: 0): Method getTimezone not exists at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php:39)"} [] [2017-02-23 10:39:50] mautic.CRITICAL: Uncaught PHP Exception InvalidArgumentException: "Method getTimezone not exists" at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php line 39 {"exception":"[object] (InvalidArgumentException(code: 0): Method getTimezone not exists at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php:39)"} [] [2017-02-23 10:39:50] mautic.NOTICE: PHP Notice - Undefined variable: dateRangeForm - in file /home/mona/mautic-master/plugins/MauticFocusBundle/Views/Focus/details.html.php - at line 103 [] [] [2017-02-23 10:39:50] mautic.ERROR: PHP Error - Argument 1 passed to Symfony\Bundle\FrameworkBundle\Templating\Helper\FormHelper::start() must be an instance of Symfony\Component\Form\FormView, null given, called in /home/mona/mautic-master/app/bundles/CoreBundle/Views/Helper/graph_dateselect.html.php on line 16 and defined - in file /home/mona/mautic-master/vendor/symfony/framework-bundle/Templating/Helper/FormHelper.php - at line 101 [] [] [2017-02-23 10:40:06] mautic.CRITICAL: Uncaught PHP Exception InvalidArgumentException: "Method getTimezone not exists" at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php line 39 {"exception":"[object] (InvalidArgumentException(code: 0): Method getTimezone not exists at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php:39)"} [] [2017-02-23 10:40:23] mautic.NOTICE: PHP Notice - Undefined variable: dateRangeForm - in file /home/mona/mautic-master/plugins/MauticFocusBundle/Views/Focus/details.html.php - at line 103 [] [] [2017-02-23 10:40:23] mautic.ERROR: PHP Error - Argument 1 passed to Symfony\Bundle\FrameworkBundle\Templating\Helper\FormHelper::start() must be an instance of Symfony\Component\Form\FormView, null given, called in /home/mona/mautic-master/app/bundles/CoreBundle/Views/Helper/graph_dateselect.html.php on line 16 and defined - in file /home/mona/mautic-master/vendor/symfony/framework-bundle/Templating/Helper/FormHelper.php - at line 101 [] []
1.0
Unable To See focus data - What type of report is this: | Q | A | ---| --- | Bug report? | Y | Feature request? | N | Enhancement? | ## Description: While saving (SAVE and CLOSE). it stuck. and then you can go back and try to click on a link. you will get only publish detail in rough data. You will not get the script. ## If a bug: | Q | A | --- | --- | Mautic version|2.6.0 | PHP version| 5.6 ### Steps to reproduce: 1. Create focus. 2. While saving (SAVE and CLOSE). it stuck. I can't enter in it to get script link. ### Log errors: I have created the focus but unable to see it. Mean I can't enter in it to get script link. ... (Version 2.6.0) @escopecz [2017-02-23 10:37:50] mautic.WARNING: PHP Warning - fwrite() expects parameter 1 to be resource, boolean given - in file /home/mona/mautic-master/app/bundles/CoreBundle/IpLookup/AbstractLocalDataLookup.php - at line 98 [] [] [2017-02-23 10:37:50] mautic.WARNING: PHP Warning - fwrite() expects parameter 1 to be resource, boolean given - in file /home/mona/mautic-master/app/bundles/CoreBundle/IpLookup/AbstractLocalDataLookup.php - at line 98 [] [] [2017-02-23 10:37:50] mautic.WARNING: PHP Warning - fwrite() expects parameter 1 to be resource, boolean given - in file /home/mona/mautic-master/app/bundles/CoreBundle/IpLookup/AbstractLocalDataLookup.php - at line 98 [] [] [2017-02-23 10:37:50] mautic.WARNING: PHP Warning - fwrite() expects parameter 1 to be resource, boolean given - in file /home/mona/mautic-master/app/bundles/CoreBundle/IpLookup/AbstractLocalDataLookup.php - at line 98 [] [] [2017-02-23 10:37:50] mautic.WARNING: PHP Warning - fclose() expects parameter 1 to be resource, boolean given - in file /home/mona/mautic-master/app/bundles/CoreBundle/IpLookup/AbstractLocalDataLookup.php - at line 100 [] [] [2017-02-23 10:37:59] mautic.NOTICE: PHP Notice - Undefined variable: dateRangeForm - in file /home/mona/mautic-master/plugins/MauticFocusBundle/Views/Focus/details.html.php - at line 103 [] [] [2017-02-23 10:37:59] mautic.ERROR: PHP Error - Argument 1 passed to Symfony\Bundle\FrameworkBundle\Templating\Helper\FormHelper::start() must be an instance of Symfony\Component\Form\FormView, null given, called in /home/mona/mautic-master/app/bundles/CoreBundle/Views/Helper/graph_dateselect.html.php on line 16 and defined - in file /home/mona/mautic-master/vendor/symfony/framework-bundle/Templating/Helper/FormHelper.php - at line 101 [] [] [2017-02-23 10:38:03] mautic.CRITICAL: Uncaught PHP Exception InvalidArgumentException: "Method getTimezone not exists" at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php line 39 {"exception":"[object] (InvalidArgumentException(code: 0): Method getTimezone not exists at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php:39)"} [] [2017-02-23 10:38:03] mautic.NOTICE: PHP Notice - Undefined variable: dateRangeForm - in file /home/mona/mautic-master/plugins/MauticFocusBundle/Views/Focus/details.html.php - at line 103 [] [] [2017-02-23 10:38:03] mautic.ERROR: PHP Error - Argument 1 passed to Symfony\Bundle\FrameworkBundle\Templating\Helper\FormHelper::start() must be an instance of Symfony\Component\Form\FormView, null given, called in /home/mona/mautic-master/app/bundles/CoreBundle/Views/Helper/graph_dateselect.html.php on line 16 and defined - in file /home/mona/mautic-master/vendor/symfony/framework-bundle/Templating/Helper/FormHelper.php - at line 101 [] [] [2017-02-23 10:38:15] mautic.NOTICE: PHP Notice - Undefined variable: dateRangeForm - in file /home/mona/mautic-master/plugins/MauticFocusBundle/Views/Focus/details.html.php - at line 103 [] [] [2017-02-23 10:38:15] mautic.ERROR: PHP Error - Argument 1 passed to Symfony\Bundle\FrameworkBundle\Templating\Helper\FormHelper::start() must be an instance of Symfony\Component\Form\FormView, null given, called in /home/mona/mautic-master/app/bundles/CoreBundle/Views/Helper/graph_dateselect.html.php on line 16 and defined - in file /home/mona/mautic-master/vendor/symfony/framework-bundle/Templating/Helper/FormHelper.php - at line 101 [] [] [2017-02-23 10:38:30] mautic.CRITICAL: Uncaught PHP Exception InvalidArgumentException: "Method getTimezone not exists" at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php line 39 {"exception":"[object] (InvalidArgumentException(code: 0): Method getTimezone not exists at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php:39)"} [] [2017-02-23 10:38:52] mautic.CRITICAL: Uncaught PHP Exception InvalidArgumentException: "Method getTimezone not exists" at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php line 39 {"exception":"[object] (InvalidArgumentException(code: 0): Method getTimezone not exists at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php:39)"} [] [2017-02-23 10:38:56] mautic.CRITICAL: Uncaught PHP Exception InvalidArgumentException: "Method getTimezone not exists" at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php line 39 {"exception":"[object] (InvalidArgumentException(code: 0): Method getTimezone not exists at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php:39)"} [] [2017-02-23 10:38:57] mautic.NOTICE: PHP Notice - Undefined variable: dateRangeForm - in file /home/mona/mautic-master/plugins/MauticFocusBundle/Views/Focus/details.html.php - at line 103 [] [] [2017-02-23 10:38:57] mautic.ERROR: PHP Error - Argument 1 passed to Symfony\Bundle\FrameworkBundle\Templating\Helper\FormHelper::start() must be an instance of Symfony\Component\Form\FormView, null given, called in /home/mona/mautic-master/app/bundles/CoreBundle/Views/Helper/graph_dateselect.html.php on line 16 and defined - in file /home/mona/mautic-master/vendor/symfony/framework-bundle/Templating/Helper/FormHelper.php - at line 101 [] [] [2017-02-23 10:39:05] mautic.CRITICAL: Uncaught PHP Exception InvalidArgumentException: "Method getTimezone not exists" at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php line 39 {"exception":"[object] (InvalidArgumentException(code: 0): Method getTimezone not exists at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php:39)"} [] [2017-02-23 10:39:12] mautic.CRITICAL: Uncaught PHP Exception InvalidArgumentException: "Method getTimezone not exists" at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php line 39 {"exception":"[object] (InvalidArgumentException(code: 0): Method getTimezone not exists at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php:39)"} [] [2017-02-23 10:39:18] mautic.CRITICAL: Uncaught PHP Exception InvalidArgumentException: "Method getTimezone not exists" at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php line 39 {"exception":"[object] (InvalidArgumentException(code: 0): Method getTimezone not exists at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php:39)"} [] [2017-02-23 10:39:41] mautic.CRITICAL: Uncaught PHP Exception InvalidArgumentException: "Method getTimezone not exists" at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php line 39 {"exception":"[object] (InvalidArgumentException(code: 0): Method getTimezone not exists at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php:39)"} [] [2017-02-23 10:39:49] mautic.CRITICAL: Uncaught PHP Exception InvalidArgumentException: "Method getTimezone not exists" at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php line 39 {"exception":"[object] (InvalidArgumentException(code: 0): Method getTimezone not exists at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php:39)"} [] [2017-02-23 10:39:50] mautic.CRITICAL: Uncaught PHP Exception InvalidArgumentException: "Method getTimezone not exists" at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php line 39 {"exception":"[object] (InvalidArgumentException(code: 0): Method getTimezone not exists at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php:39)"} [] [2017-02-23 10:39:50] mautic.NOTICE: PHP Notice - Undefined variable: dateRangeForm - in file /home/mona/mautic-master/plugins/MauticFocusBundle/Views/Focus/details.html.php - at line 103 [] [] [2017-02-23 10:39:50] mautic.ERROR: PHP Error - Argument 1 passed to Symfony\Bundle\FrameworkBundle\Templating\Helper\FormHelper::start() must be an instance of Symfony\Component\Form\FormView, null given, called in /home/mona/mautic-master/app/bundles/CoreBundle/Views/Helper/graph_dateselect.html.php on line 16 and defined - in file /home/mona/mautic-master/vendor/symfony/framework-bundle/Templating/Helper/FormHelper.php - at line 101 [] [] [2017-02-23 10:40:06] mautic.CRITICAL: Uncaught PHP Exception InvalidArgumentException: "Method getTimezone not exists" at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php line 39 {"exception":"[object] (InvalidArgumentException(code: 0): Method getTimezone not exists at /home/mona/mautic-master/app/bundles/CoreBundle/Entity/CommonEntity.php:39)"} [] [2017-02-23 10:40:23] mautic.NOTICE: PHP Notice - Undefined variable: dateRangeForm - in file /home/mona/mautic-master/plugins/MauticFocusBundle/Views/Focus/details.html.php - at line 103 [] [] [2017-02-23 10:40:23] mautic.ERROR: PHP Error - Argument 1 passed to Symfony\Bundle\FrameworkBundle\Templating\Helper\FormHelper::start() must be an instance of Symfony\Component\Form\FormView, null given, called in /home/mona/mautic-master/app/bundles/CoreBundle/Views/Helper/graph_dateselect.html.php on line 16 and defined - in file /home/mona/mautic-master/vendor/symfony/framework-bundle/Templating/Helper/FormHelper.php - at line 101 [] []
test
unable to see focus data what type of report is this q a bug report y feature request n enhancement description while saving save and close it stuck and then you can go back and try to click on a link you will get only publish detail in rough data you will not get the script if a bug q a mautic version php version steps to reproduce create focus while saving save and close it stuck i can t enter in it to get script link log errors i have created the focus but unable to see it mean i can t enter in it to get script link version escopecz mautic warning php warning fwrite expects parameter to be resource boolean given in file home mona mautic master app bundles corebundle iplookup abstractlocaldatalookup php at line mautic warning php warning fwrite expects parameter to be resource boolean given in file home mona mautic master app bundles corebundle iplookup abstractlocaldatalookup php at line mautic warning php warning fwrite expects parameter to be resource boolean given in file home mona mautic master app bundles corebundle iplookup abstractlocaldatalookup php at line mautic warning php warning fwrite expects parameter to be resource boolean given in file home mona mautic master app bundles corebundle iplookup abstractlocaldatalookup php at line mautic warning php warning fclose expects parameter to be resource boolean given in file home mona mautic master app bundles corebundle iplookup abstractlocaldatalookup php at line mautic notice php notice undefined variable daterangeform in file home mona mautic master plugins mauticfocusbundle views focus details html php at line mautic error php error argument passed to symfony bundle frameworkbundle templating helper formhelper start must be an instance of symfony component form formview null given called in home mona mautic master app bundles corebundle views helper graph dateselect html php on line and defined in file home mona mautic master vendor symfony framework bundle templating helper formhelper php at line mautic critical uncaught php exception invalidargumentexception method gettimezone not exists at home mona mautic master app bundles corebundle entity commonentity php line exception invalidargumentexception code method gettimezone not exists at home mona mautic master app bundles corebundle entity commonentity php mautic notice php notice undefined variable daterangeform in file home mona mautic master plugins mauticfocusbundle views focus details html php at line mautic error php error argument passed to symfony bundle frameworkbundle templating helper formhelper start must be an instance of symfony component form formview null given called in home mona mautic master app bundles corebundle views helper graph dateselect html php on line and defined in file home mona mautic master vendor symfony framework bundle templating helper formhelper php at line mautic notice php notice undefined variable daterangeform in file home mona mautic master plugins mauticfocusbundle views focus details html php at line mautic error php error argument passed to symfony bundle frameworkbundle templating helper formhelper start must be an instance of symfony component form formview null given called in home mona mautic master app bundles corebundle views helper graph dateselect html php on line and defined in file home mona mautic master vendor symfony framework bundle templating helper formhelper php at line mautic critical uncaught php exception invalidargumentexception method gettimezone not exists at home mona mautic master app bundles corebundle entity commonentity php line exception invalidargumentexception code method gettimezone not exists at home mona mautic master app bundles corebundle entity commonentity php mautic critical uncaught php exception invalidargumentexception method gettimezone not exists at home mona mautic master app bundles corebundle entity commonentity php line exception invalidargumentexception code method gettimezone not exists at home mona mautic master app bundles corebundle entity commonentity php mautic critical uncaught php exception invalidargumentexception method gettimezone not exists at home mona mautic master app bundles corebundle entity commonentity php line exception invalidargumentexception code method gettimezone not exists at home mona mautic master app bundles corebundle entity commonentity php mautic notice php notice undefined variable daterangeform in file home mona mautic master plugins mauticfocusbundle views focus details html php at line mautic error php error argument passed to symfony bundle frameworkbundle templating helper formhelper start must be an instance of symfony component form formview null given called in home mona mautic master app bundles corebundle views helper graph dateselect html php on line and defined in file home mona mautic master vendor symfony framework bundle templating helper formhelper php at line mautic critical uncaught php exception invalidargumentexception method gettimezone not exists at home mona mautic master app bundles corebundle entity commonentity php line exception invalidargumentexception code method gettimezone not exists at home mona mautic master app bundles corebundle entity commonentity php mautic critical uncaught php exception invalidargumentexception method gettimezone not exists at home mona mautic master app bundles corebundle entity commonentity php line exception invalidargumentexception code method gettimezone not exists at home mona mautic master app bundles corebundle entity commonentity php mautic critical uncaught php exception invalidargumentexception method gettimezone not exists at home mona mautic master app bundles corebundle entity commonentity php line exception invalidargumentexception code method gettimezone not exists at home mona mautic master app bundles corebundle entity commonentity php mautic critical uncaught php exception invalidargumentexception method gettimezone not exists at home mona mautic master app bundles corebundle entity commonentity php line exception invalidargumentexception code method gettimezone not exists at home mona mautic master app bundles corebundle entity commonentity php mautic critical uncaught php exception invalidargumentexception method gettimezone not exists at home mona mautic master app bundles corebundle entity commonentity php line exception invalidargumentexception code method gettimezone not exists at home mona mautic master app bundles corebundle entity commonentity php mautic critical uncaught php exception invalidargumentexception method gettimezone not exists at home mona mautic master app bundles corebundle entity commonentity php line exception invalidargumentexception code method gettimezone not exists at home mona mautic master app bundles corebundle entity commonentity php mautic notice php notice undefined variable daterangeform in file home mona mautic master plugins mauticfocusbundle views focus details html php at line mautic error php error argument passed to symfony bundle frameworkbundle templating helper formhelper start must be an instance of symfony component form formview null given called in home mona mautic master app bundles corebundle views helper graph dateselect html php on line and defined in file home mona mautic master vendor symfony framework bundle templating helper formhelper php at line mautic critical uncaught php exception invalidargumentexception method gettimezone not exists at home mona mautic master app bundles corebundle entity commonentity php line exception invalidargumentexception code method gettimezone not exists at home mona mautic master app bundles corebundle entity commonentity php mautic notice php notice undefined variable daterangeform in file home mona mautic master plugins mauticfocusbundle views focus details html php at line mautic error php error argument passed to symfony bundle frameworkbundle templating helper formhelper start must be an instance of symfony component form formview null given called in home mona mautic master app bundles corebundle views helper graph dateselect html php on line and defined in file home mona mautic master vendor symfony framework bundle templating helper formhelper php at line
1
161,609
12,554,323,268
IssuesEvent
2020-06-07 01:34:19
kotest/kotest
https://api.github.com/repos/kotest/kotest
closed
Property tests should report exception of running shrunk input
bug property-testing
During property testing, if an input fails, the `AssertionError` is being rendered as part of the test report. But if the input has successfully shrunk, the exception that failed the shrunk input is ignored. Ideally the exception that the user sees in the final test report should be the one that the shrunk input failed with. It might also be ok to render both exceptions since also both inputs (original and shrunk) are rendered.
1.0
Property tests should report exception of running shrunk input - During property testing, if an input fails, the `AssertionError` is being rendered as part of the test report. But if the input has successfully shrunk, the exception that failed the shrunk input is ignored. Ideally the exception that the user sees in the final test report should be the one that the shrunk input failed with. It might also be ok to render both exceptions since also both inputs (original and shrunk) are rendered.
test
property tests should report exception of running shrunk input during property testing if an input fails the assertionerror is being rendered as part of the test report but if the input has successfully shrunk the exception that failed the shrunk input is ignored ideally the exception that the user sees in the final test report should be the one that the shrunk input failed with it might also be ok to render both exceptions since also both inputs original and shrunk are rendered
1
48,202
5,949,071,483
IssuesEvent
2017-05-26 13:22:05
MohammadYounes/AlertifyJS
https://api.github.com/repos/MohammadYounes/AlertifyJS
closed
alert dialog can be hidden by some elements
needs test case troubleshooting
below is an example image ![image](https://cloud.githubusercontent.com/assets/3401416/24072347/80a3cd86-0c29-11e7-8464-8a3f99de5246.png) the vertical bar is an element of [ace editor](https://ace.c9.io) styles of the vertical bar ![image](https://cloud.githubusercontent.com/assets/3401416/24072362/b8ade464-0c29-11e7-96a1-a0ce8243b866.png)
1.0
alert dialog can be hidden by some elements - below is an example image ![image](https://cloud.githubusercontent.com/assets/3401416/24072347/80a3cd86-0c29-11e7-8464-8a3f99de5246.png) the vertical bar is an element of [ace editor](https://ace.c9.io) styles of the vertical bar ![image](https://cloud.githubusercontent.com/assets/3401416/24072362/b8ade464-0c29-11e7-96a1-a0ce8243b866.png)
test
alert dialog can be hidden by some elements below is an example image the vertical bar is an element of styles of the vertical bar
1
150,954
11,995,036,021
IssuesEvent
2020-04-08 14:36:59
prysmaticlabs/prysm
https://api.github.com/repos/prysmaticlabs/prysm
closed
Allow Multiple Proposal Slots in an Epoch
API Testnet Tracking
Currently is is assumed that a validator will only have a single proposal slot in an epoch, however due to the fact that proposal slots are computed in on a per slot basis, there is the possibility that a validator could have multiple proposal slots in an epoch instead of once. This issue is to track the implementation of the required changes for this. We will need to modify our ethereum-api and the relevant code in prysm
1.0
Allow Multiple Proposal Slots in an Epoch - Currently is is assumed that a validator will only have a single proposal slot in an epoch, however due to the fact that proposal slots are computed in on a per slot basis, there is the possibility that a validator could have multiple proposal slots in an epoch instead of once. This issue is to track the implementation of the required changes for this. We will need to modify our ethereum-api and the relevant code in prysm
test
allow multiple proposal slots in an epoch currently is is assumed that a validator will only have a single proposal slot in an epoch however due to the fact that proposal slots are computed in on a per slot basis there is the possibility that a validator could have multiple proposal slots in an epoch instead of once this issue is to track the implementation of the required changes for this we will need to modify our ethereum api and the relevant code in prysm
1
143,864
11,581,475,554
IssuesEvent
2020-02-21 22:45:25
pvlib/pvlib-python
https://api.github.com/repos/pvlib/pvlib-python
closed
the test suite is too slow
testing
The test suite is getting to be quite slow. Here are the 20 slowest tests on my mac: ``` pytest --durations=20 18.96s call pvlib/tests/iotools/test_pvgis.py::test_get_pvgis_tmy_kwargs 16.32s call pvlib/tests/iotools/test_psm3.py::test_get_psm3_singleyear 14.82s setup pvlib/tests/test_forecast.py::test_process_data[HRRR] 8.12s call pvlib/tests/iotools/test_psm3.py::test_get_psm3_tmy 6.82s setup pvlib/tests/test_forecast.py::test_process_data[NDFD] 5.47s call pvlib/tests/iotools/test_pvgis.py::test_get_pvgis_tmy_epw 4.78s call pvlib/tests/iotools/test_pvgis.py::test_get_pvgis_tmy 3.99s setup pvlib/tests/test_forecast.py::test_process_data[RAP] 3.45s call pvlib/tests/iotools/test_srml.py::test_read_srml_month_from_solardat 3.20s call pvlib/tests/test_forecast.py::test_bad_kwarg_get_data 3.16s call pvlib/tests/test_forecast.py::test_bad_kwarg_get_processed_data 3.13s call pvlib/tests/test_forecast.py::test_how_kwarg_get_processed_data 3.13s setup pvlib/tests/test_forecast.py::test_process_data[NAM] 3.12s call pvlib/tests/test_forecast.py::test_vert_level 3.08s call pvlib/tests/test_forecast.py::test_datetime 2.98s call pvlib/tests/test_irradiance.py::test_get_extra_radiation_nrel_numba 2.85s call pvlib/tests/test_solarposition.py::test_spa_python_numba_physical 2.82s call pvlib/tests/test_solarposition.py::test_get_solarposition_deltat[None-nrel_numba] 2.73s setup pvlib/tests/test_spa.py::NumbaSpaTest::test_aberration_correction 2.69s call pvlib/tests/iotools/test_pvgis.py::test_get_pvgis_tmy_basic ``` I will work on the `forecast` tests. Any ideas for the `iotools` tests?
1.0
the test suite is too slow - The test suite is getting to be quite slow. Here are the 20 slowest tests on my mac: ``` pytest --durations=20 18.96s call pvlib/tests/iotools/test_pvgis.py::test_get_pvgis_tmy_kwargs 16.32s call pvlib/tests/iotools/test_psm3.py::test_get_psm3_singleyear 14.82s setup pvlib/tests/test_forecast.py::test_process_data[HRRR] 8.12s call pvlib/tests/iotools/test_psm3.py::test_get_psm3_tmy 6.82s setup pvlib/tests/test_forecast.py::test_process_data[NDFD] 5.47s call pvlib/tests/iotools/test_pvgis.py::test_get_pvgis_tmy_epw 4.78s call pvlib/tests/iotools/test_pvgis.py::test_get_pvgis_tmy 3.99s setup pvlib/tests/test_forecast.py::test_process_data[RAP] 3.45s call pvlib/tests/iotools/test_srml.py::test_read_srml_month_from_solardat 3.20s call pvlib/tests/test_forecast.py::test_bad_kwarg_get_data 3.16s call pvlib/tests/test_forecast.py::test_bad_kwarg_get_processed_data 3.13s call pvlib/tests/test_forecast.py::test_how_kwarg_get_processed_data 3.13s setup pvlib/tests/test_forecast.py::test_process_data[NAM] 3.12s call pvlib/tests/test_forecast.py::test_vert_level 3.08s call pvlib/tests/test_forecast.py::test_datetime 2.98s call pvlib/tests/test_irradiance.py::test_get_extra_radiation_nrel_numba 2.85s call pvlib/tests/test_solarposition.py::test_spa_python_numba_physical 2.82s call pvlib/tests/test_solarposition.py::test_get_solarposition_deltat[None-nrel_numba] 2.73s setup pvlib/tests/test_spa.py::NumbaSpaTest::test_aberration_correction 2.69s call pvlib/tests/iotools/test_pvgis.py::test_get_pvgis_tmy_basic ``` I will work on the `forecast` tests. Any ideas for the `iotools` tests?
test
the test suite is too slow the test suite is getting to be quite slow here are the slowest tests on my mac pytest durations call pvlib tests iotools test pvgis py test get pvgis tmy kwargs call pvlib tests iotools test py test get singleyear setup pvlib tests test forecast py test process data call pvlib tests iotools test py test get tmy setup pvlib tests test forecast py test process data call pvlib tests iotools test pvgis py test get pvgis tmy epw call pvlib tests iotools test pvgis py test get pvgis tmy setup pvlib tests test forecast py test process data call pvlib tests iotools test srml py test read srml month from solardat call pvlib tests test forecast py test bad kwarg get data call pvlib tests test forecast py test bad kwarg get processed data call pvlib tests test forecast py test how kwarg get processed data setup pvlib tests test forecast py test process data call pvlib tests test forecast py test vert level call pvlib tests test forecast py test datetime call pvlib tests test irradiance py test get extra radiation nrel numba call pvlib tests test solarposition py test spa python numba physical call pvlib tests test solarposition py test get solarposition deltat setup pvlib tests test spa py numbaspatest test aberration correction call pvlib tests iotools test pvgis py test get pvgis tmy basic i will work on the forecast tests any ideas for the iotools tests
1
261,900
22,779,236,994
IssuesEvent
2022-07-08 17:39:48
nasa/openmct
https://api.github.com/repos/nasa/openmct
closed
[Edit] Remove available from arrow next to title
type:bug unverified severity:medium bug:retest
Steps to reproduce: 1. Create a new Display Layout. This should take you immediately into Edit mode. 2. Click Save and choose some appropriate location (e.g. My Items) then click OK. You should no longer be in Edit mode, but should still be navigated to the same layout. 3. Click the Edit button. You should again be taken into Edit mode. 4. Hover to the right of the title of the Display Layout ("Unnamed Display Layout" at top of screen). An arrow pointing down should appear. 5. Click on the arrow. A context menu should appear. 6. **Observe:** The context menu contains the Remove action. This is unexpected, as Remove effects the _parent_ of the edited object, which is outside of the scope for editing. (Additionally, the object is unsaved.) 7. Click on Remove. 8. **Observe:** Navigation takes you back to the parent container, but the object has not really been removed. Believe correct behavior here is for Remove to be unavailable at Step 6.
1.0
[Edit] Remove available from arrow next to title - Steps to reproduce: 1. Create a new Display Layout. This should take you immediately into Edit mode. 2. Click Save and choose some appropriate location (e.g. My Items) then click OK. You should no longer be in Edit mode, but should still be navigated to the same layout. 3. Click the Edit button. You should again be taken into Edit mode. 4. Hover to the right of the title of the Display Layout ("Unnamed Display Layout" at top of screen). An arrow pointing down should appear. 5. Click on the arrow. A context menu should appear. 6. **Observe:** The context menu contains the Remove action. This is unexpected, as Remove effects the _parent_ of the edited object, which is outside of the scope for editing. (Additionally, the object is unsaved.) 7. Click on Remove. 8. **Observe:** Navigation takes you back to the parent container, but the object has not really been removed. Believe correct behavior here is for Remove to be unavailable at Step 6.
test
remove available from arrow next to title steps to reproduce create a new display layout this should take you immediately into edit mode click save and choose some appropriate location e g my items then click ok you should no longer be in edit mode but should still be navigated to the same layout click the edit button you should again be taken into edit mode hover to the right of the title of the display layout unnamed display layout at top of screen an arrow pointing down should appear click on the arrow a context menu should appear observe the context menu contains the remove action this is unexpected as remove effects the parent of the edited object which is outside of the scope for editing additionally the object is unsaved click on remove observe navigation takes you back to the parent container but the object has not really been removed believe correct behavior here is for remove to be unavailable at step
1
128,195
10,519,545,769
IssuesEvent
2019-09-29 18:44:54
commercialhaskell/stackage
https://api.github.com/repos/commercialhaskell/stackage
closed
ghc-exactprint test failure
failure: test-suite
``` Test suite failure for package ghc-exactprint-0.6.1 test: executable not found ``` CC @alanz
1.0
ghc-exactprint test failure - ``` Test suite failure for package ghc-exactprint-0.6.1 test: executable not found ``` CC @alanz
test
ghc exactprint test failure test suite failure for package ghc exactprint test executable not found cc alanz
1
516,096
14,975,282,581
IssuesEvent
2021-01-28 05:46:46
nlpsandbox/nlpsandbox-controller
https://api.github.com/repos/nlpsandbox/nlpsandbox-controller
closed
Should remove content of a submission annotation store if it already exists?
Priority: High
This is the scenario when re-running a submission...
1.0
Should remove content of a submission annotation store if it already exists? - This is the scenario when re-running a submission...
non_test
should remove content of a submission annotation store if it already exists this is the scenario when re running a submission
0
18,616
4,288,534,585
IssuesEvent
2016-07-17 14:34:23
swisnl/jQuery-contextMenu
https://api.github.com/repos/swisnl/jQuery-contextMenu
closed
documentation does not state how to destroy a contextmenu
Documentation
i found a solution in #110 to use: ```javascript $.contextMenu("destroy", ".selector"); ```
1.0
documentation does not state how to destroy a contextmenu - i found a solution in #110 to use: ```javascript $.contextMenu("destroy", ".selector"); ```
non_test
documentation does not state how to destroy a contextmenu i found a solution in to use javascript contextmenu destroy selector
0
1,665
2,800,476,578
IssuesEvent
2015-05-13 10:15:07
akka/akka
https://api.github.com/repos/akka/akka
opened
Add feature to run a test until it fails k times, dumping the failure output somewhere
1 - triaged t:build
Should allow specifying a number ("I want 4 failures") and then run a test until it collects the required amount of failures (or some upper iteration/time limit is reached), then it dumps the outputs to different files
1.0
Add feature to run a test until it fails k times, dumping the failure output somewhere - Should allow specifying a number ("I want 4 failures") and then run a test until it collects the required amount of failures (or some upper iteration/time limit is reached), then it dumps the outputs to different files
non_test
add feature to run a test until it fails k times dumping the failure output somewhere should allow specifying a number i want failures and then run a test until it collects the required amount of failures or some upper iteration time limit is reached then it dumps the outputs to different files
0
90,637
8,251,418,035
IssuesEvent
2018-09-12 07:45:09
Kademi/kademi-dev
https://api.github.com/repos/Kademi/kademi-dev
closed
KCRM: Create a task from add a note window on dashboard component cause error
Ready to Test - Dev bug
![image](https://user-images.githubusercontent.com/1945230/45361239-6a8e2a80-b5da-11e8-8831-0d03fbceccb1.png) When create save it gives you method not implemented ![image](https://user-images.githubusercontent.com/1945230/45361254-7679ec80-b5da-11e8-9898-545210f2c820.png) http://vladtest53wweb-vladtest53w.kademi-ci.co/dashboard user: adminqa or kcrmadmin
1.0
KCRM: Create a task from add a note window on dashboard component cause error - ![image](https://user-images.githubusercontent.com/1945230/45361239-6a8e2a80-b5da-11e8-8831-0d03fbceccb1.png) When create save it gives you method not implemented ![image](https://user-images.githubusercontent.com/1945230/45361254-7679ec80-b5da-11e8-9898-545210f2c820.png) http://vladtest53wweb-vladtest53w.kademi-ci.co/dashboard user: adminqa or kcrmadmin
test
kcrm create a task from add a note window on dashboard component cause error when create save it gives you method not implemented user adminqa or kcrmadmin
1
94,963
3,933,558,680
IssuesEvent
2016-04-25 19:33:18
ghutchis/avogadro
https://api.github.com/repos/ghutchis/avogadro
closed
Construction Tool doesn't Work
auto-migrated Construction Tool Doesn't Work low priority v_0.1.0
When I'm going to make a new molecula nothing happens. I try to use the construction tool, but it doesn't work. I have used the program before an all is OK. But the last times doesn't work. Reported by: *anonymous
1.0
Construction Tool doesn't Work - When I'm going to make a new molecula nothing happens. I try to use the construction tool, but it doesn't work. I have used the program before an all is OK. But the last times doesn't work. Reported by: *anonymous
non_test
construction tool doesn t work when i m going to make a new molecula nothing happens i try to use the construction tool but it doesn t work i have used the program before an all is ok but the last times doesn t work reported by anonymous
0
173,327
6,523,581,184
IssuesEvent
2017-08-29 09:14:16
BuckleScript/bucklescript
https://api.github.com/repos/BuckleScript/bucklescript
closed
get rid of no export special handling
PRIORITY:HIGH
`no_export` is only used for demo purpose, currently we already provide `[@@bs.config { no_export } ]` for such purpose, we can make all these behind the scene by using ```ocaml [@@@bs.config{no_export}] # 1 "toplevel" (* first two lines are hidden from user when clicking remove unused code *) let b = 3 ``` I am convinced to remove this ad hoc support to reduce maintenance overhead, the main thing is with such ad-hoc support - the invariant that the length of exported identifiers and export lambdas are the same is not held any more, we need add a special handling in the pipeline
1.0
get rid of no export special handling - `no_export` is only used for demo purpose, currently we already provide `[@@bs.config { no_export } ]` for such purpose, we can make all these behind the scene by using ```ocaml [@@@bs.config{no_export}] # 1 "toplevel" (* first two lines are hidden from user when clicking remove unused code *) let b = 3 ``` I am convinced to remove this ad hoc support to reduce maintenance overhead, the main thing is with such ad-hoc support - the invariant that the length of exported identifiers and export lambdas are the same is not held any more, we need add a special handling in the pipeline
non_test
get rid of no export special handling no export is only used for demo purpose currently we already provide for such purpose we can make all these behind the scene by using ocaml toplevel first two lines are hidden from user when clicking remove unused code let b i am convinced to remove this ad hoc support to reduce maintenance overhead the main thing is with such ad hoc support the invariant that the length of exported identifiers and export lambdas are the same is not held any more we need add a special handling in the pipeline
0
72,609
3,388,490,203
IssuesEvent
2015-11-29 10:29:38
rathena/rathena
https://api.github.com/repos/rathena/rathena
closed
This NPC is not belong to npc/re/instances/WolfchevLaboratory.txt
bug:script mode:renewal priority:low server:map
I just tried to finding rA script while playing idRO. Found the reward NPC technically is not instance dependent. ``` // Weird Old Man (Rewards) //============================================================ lighthalzen,342,291,4 script Weird old man#Bio4Reward 4_M_EINOLD,{ } ``` Why, because the required items can be dropped from lhz_dun03 and lhz_dun04, not only from `1@lhz` instance. and also with this enchant NPC ``` // Sorcerer (Rewards) //============================================================ lhz_cube,233,24,4 script Sorcerer#Bio4Reward 4_M_UMDANCEKID,{ } ``` So, if user want disable instance but just enable the quest, when instance file is commented, the quest is still.
1.0
This NPC is not belong to npc/re/instances/WolfchevLaboratory.txt - I just tried to finding rA script while playing idRO. Found the reward NPC technically is not instance dependent. ``` // Weird Old Man (Rewards) //============================================================ lighthalzen,342,291,4 script Weird old man#Bio4Reward 4_M_EINOLD,{ } ``` Why, because the required items can be dropped from lhz_dun03 and lhz_dun04, not only from `1@lhz` instance. and also with this enchant NPC ``` // Sorcerer (Rewards) //============================================================ lhz_cube,233,24,4 script Sorcerer#Bio4Reward 4_M_UMDANCEKID,{ } ``` So, if user want disable instance but just enable the quest, when instance file is commented, the quest is still.
non_test
this npc is not belong to npc re instances wolfchevlaboratory txt i just tried to finding ra script while playing idro found the reward npc technically is not instance dependent weird old man rewards lighthalzen script weird old man m einold why because the required items can be dropped from lhz and lhz not only from lhz instance and also with this enchant npc sorcerer rewards lhz cube script sorcerer m umdancekid so if user want disable instance but just enable the quest when instance file is commented the quest is still
0
372,320
25,995,571,496
IssuesEvent
2022-12-20 11:15:38
aws/aws-cli
https://api.github.com/repos/aws/aws-cli
closed
Update aws cloudformation package docs to include missing resources
documentation feature-request cloudformation package-deploy customization community p2
Confirm by changing [ ] to [x] below: - [X] I've gone though the [User Guide](https://docs.aws.amazon.com/cli/latest/userguide/cli-chap-welcome.html) and the [API reference](https://docs.aws.amazon.com/cli/latest/reference/) - [X] I've searched for [previous similar issues](https://github.com/aws/aws-cli/issues) and didn't find any solution Issue is about usage on: - [ ] Service API : I want to do X using Y service, what should I do? - [ ] CLI : passing arguments or cli configurations. - [X] Other/Not sure. It appears that the `aws cloudformation package` command supports more resources than listed in the documentation. Specifically, I can see these three resources supported, but not mentioned in the docs. * `Location` property of the `ServerlessApplicationResource` (`AWS::Serverless::Application`), * `ContentUri` of the `ServerlessLayerVersionResource` (`AWS::Serverless::LayerVersion`) * `Content` property of the `LambdaLayerVersionResource` (`AWS::Lambda::LayerVersion`) Resources are taken from https://github.com/aws/aws-cli/blob/daeaf7d46b7d14140a6c96ce931575ccf8dbf2d8/awscli/customizations/cloudformation/artifact_exporter.py#L500 Docs are https://awscli.amazonaws.com/v2/documentation/api/latest/reference/cloudformation/package.html I think this file needs to be updated https://github.com/aws/aws-cli/blob/develop/awscli/examples/cloudformation/_package_description.rst I'm happy to make the PR, but I'll wait for a maintainer to respond first.
1.0
Update aws cloudformation package docs to include missing resources - Confirm by changing [ ] to [x] below: - [X] I've gone though the [User Guide](https://docs.aws.amazon.com/cli/latest/userguide/cli-chap-welcome.html) and the [API reference](https://docs.aws.amazon.com/cli/latest/reference/) - [X] I've searched for [previous similar issues](https://github.com/aws/aws-cli/issues) and didn't find any solution Issue is about usage on: - [ ] Service API : I want to do X using Y service, what should I do? - [ ] CLI : passing arguments or cli configurations. - [X] Other/Not sure. It appears that the `aws cloudformation package` command supports more resources than listed in the documentation. Specifically, I can see these three resources supported, but not mentioned in the docs. * `Location` property of the `ServerlessApplicationResource` (`AWS::Serverless::Application`), * `ContentUri` of the `ServerlessLayerVersionResource` (`AWS::Serverless::LayerVersion`) * `Content` property of the `LambdaLayerVersionResource` (`AWS::Lambda::LayerVersion`) Resources are taken from https://github.com/aws/aws-cli/blob/daeaf7d46b7d14140a6c96ce931575ccf8dbf2d8/awscli/customizations/cloudformation/artifact_exporter.py#L500 Docs are https://awscli.amazonaws.com/v2/documentation/api/latest/reference/cloudformation/package.html I think this file needs to be updated https://github.com/aws/aws-cli/blob/develop/awscli/examples/cloudformation/_package_description.rst I'm happy to make the PR, but I'll wait for a maintainer to respond first.
non_test
update aws cloudformation package docs to include missing resources confirm by changing to below i ve gone though the and the i ve searched for and didn t find any solution issue is about usage on service api i want to do x using y service what should i do cli passing arguments or cli configurations other not sure it appears that the aws cloudformation package command supports more resources than listed in the documentation specifically i can see these three resources supported but not mentioned in the docs location property of the serverlessapplicationresource aws serverless application contenturi of the serverlesslayerversionresource aws serverless layerversion content property of the lambdalayerversionresource aws lambda layerversion resources are taken from docs are i think this file needs to be updated i m happy to make the pr but i ll wait for a maintainer to respond first
0
23,994
4,055,456,822
IssuesEvent
2016-05-24 15:31:48
ngageoint/hootenanny
https://api.github.com/repos/ngageoint/hootenanny
closed
Modify PrintNamesCmdTest.sh to sort the output
Category: Test Status: In Progress Type: Bug
Sometimes the output isn't in the same order and the test fails. Sorting should help alleviate this. Be sure and use the "C" LANG when sorting or you may get different results on Ubuntu vs. CentOS. See this for an example failure: http://selma:8765/view/critical/job/hoot-rpms-centos67/254/console
1.0
Modify PrintNamesCmdTest.sh to sort the output - Sometimes the output isn't in the same order and the test fails. Sorting should help alleviate this. Be sure and use the "C" LANG when sorting or you may get different results on Ubuntu vs. CentOS. See this for an example failure: http://selma:8765/view/critical/job/hoot-rpms-centos67/254/console
test
modify printnamescmdtest sh to sort the output sometimes the output isn t in the same order and the test fails sorting should help alleviate this be sure and use the c lang when sorting or you may get different results on ubuntu vs centos see this for an example failure
1
14,343
17,370,285,464
IssuesEvent
2021-07-30 13:08:57
2i2c-org/team-compass
https://api.github.com/repos/2i2c-org/team-compass
opened
Use a password manager to share infrastructure passwords between the team
:label: team-process type: enhancement
# Summary There are a few places where we share accounts to access the same services or webpages. An example of this is the grafana of a hub, where we don't necessarily want to create a new admin username for every single hub engineer. For these cases, we currently follow a practice of "ask a team member what the password is". This makes it hard to know who has access to which passwords, and is an extra step team members must follow to get access. It's also something that has to be done _each time_ a new password is needed. # Proposal We use a team password application like [1Password](https://1password.com/). This would allow us to store the passwords in an encrypted service, and we could purchase a team account that would provide each of us access to them (I believe their base team account is $20/mo for 10 people, which isn't bad). We could then use this to store any team passwords that don't require 2FA in order to log-in, and then the only step for providing access to a new team member is to get them a 1Password account. # Actions - [ ] Answer questions below, and if we wish to proceed... - [ ] Write up a proposed process for how we share passwords - [ ] Set up password manager and accounts for team members - [ ] Write it up in team compass # Questions - Is anybody opposed to this idea? Would it lead us to any obvious anti-patterns? - Any strong preferences for a particular password manager? 1Password vs. LastPass, for example.
1.0
Use a password manager to share infrastructure passwords between the team - # Summary There are a few places where we share accounts to access the same services or webpages. An example of this is the grafana of a hub, where we don't necessarily want to create a new admin username for every single hub engineer. For these cases, we currently follow a practice of "ask a team member what the password is". This makes it hard to know who has access to which passwords, and is an extra step team members must follow to get access. It's also something that has to be done _each time_ a new password is needed. # Proposal We use a team password application like [1Password](https://1password.com/). This would allow us to store the passwords in an encrypted service, and we could purchase a team account that would provide each of us access to them (I believe their base team account is $20/mo for 10 people, which isn't bad). We could then use this to store any team passwords that don't require 2FA in order to log-in, and then the only step for providing access to a new team member is to get them a 1Password account. # Actions - [ ] Answer questions below, and if we wish to proceed... - [ ] Write up a proposed process for how we share passwords - [ ] Set up password manager and accounts for team members - [ ] Write it up in team compass # Questions - Is anybody opposed to this idea? Would it lead us to any obvious anti-patterns? - Any strong preferences for a particular password manager? 1Password vs. LastPass, for example.
non_test
use a password manager to share infrastructure passwords between the team summary there are a few places where we share accounts to access the same services or webpages an example of this is the grafana of a hub where we don t necessarily want to create a new admin username for every single hub engineer for these cases we currently follow a practice of ask a team member what the password is this makes it hard to know who has access to which passwords and is an extra step team members must follow to get access it s also something that has to be done each time a new password is needed proposal we use a team password application like this would allow us to store the passwords in an encrypted service and we could purchase a team account that would provide each of us access to them i believe their base team account is mo for people which isn t bad we could then use this to store any team passwords that don t require in order to log in and then the only step for providing access to a new team member is to get them a account actions answer questions below and if we wish to proceed write up a proposed process for how we share passwords set up password manager and accounts for team members write it up in team compass questions is anybody opposed to this idea would it lead us to any obvious anti patterns any strong preferences for a particular password manager vs lastpass for example
0
154,899
19,765,546,834
IssuesEvent
2022-01-17 01:26:24
tuanducdesign/web-netflix
https://api.github.com/repos/tuanducdesign/web-netflix
opened
CVE-2021-3918 (High) detected in json-schema-0.2.3.tgz
security vulnerability
## CVE-2021-3918 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>json-schema-0.2.3.tgz</b></p></summary> <p>JSON Schema validation and specifications</p> <p>Library home page: <a href="https://registry.npmjs.org/json-schema/-/json-schema-0.2.3.tgz">https://registry.npmjs.org/json-schema/-/json-schema-0.2.3.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/json-schema/package.json</p> <p> Dependency Hierarchy: - react-scripts-4.0.3.tgz (Root Library) - jest-circus-26.6.0.tgz - jest-runner-26.6.3.tgz - jest-config-26.6.3.tgz - jest-environment-jsdom-26.6.2.tgz - jsdom-16.4.0.tgz - request-2.88.2.tgz - http-signature-1.2.0.tgz - jsprim-1.4.1.tgz - :x: **json-schema-0.2.3.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> json-schema is vulnerable to Improperly Controlled Modification of Object Prototype Attributes ('Prototype Pollution') <p>Publish Date: 2021-11-13 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3918>CVE-2021-3918</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2021-3918">https://nvd.nist.gov/vuln/detail/CVE-2021-3918</a></p> <p>Release Date: 2021-11-13</p> <p>Fix Resolution: json-schema - 0.4.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-3918 (High) detected in json-schema-0.2.3.tgz - ## CVE-2021-3918 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>json-schema-0.2.3.tgz</b></p></summary> <p>JSON Schema validation and specifications</p> <p>Library home page: <a href="https://registry.npmjs.org/json-schema/-/json-schema-0.2.3.tgz">https://registry.npmjs.org/json-schema/-/json-schema-0.2.3.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/json-schema/package.json</p> <p> Dependency Hierarchy: - react-scripts-4.0.3.tgz (Root Library) - jest-circus-26.6.0.tgz - jest-runner-26.6.3.tgz - jest-config-26.6.3.tgz - jest-environment-jsdom-26.6.2.tgz - jsdom-16.4.0.tgz - request-2.88.2.tgz - http-signature-1.2.0.tgz - jsprim-1.4.1.tgz - :x: **json-schema-0.2.3.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> json-schema is vulnerable to Improperly Controlled Modification of Object Prototype Attributes ('Prototype Pollution') <p>Publish Date: 2021-11-13 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3918>CVE-2021-3918</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2021-3918">https://nvd.nist.gov/vuln/detail/CVE-2021-3918</a></p> <p>Release Date: 2021-11-13</p> <p>Fix Resolution: json-schema - 0.4.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in json schema tgz cve high severity vulnerability vulnerable library json schema tgz json schema validation and specifications library home page a href path to dependency file package json path to vulnerable library node modules json schema package json dependency hierarchy react scripts tgz root library jest circus tgz jest runner tgz jest config tgz jest environment jsdom tgz jsdom tgz request tgz http signature tgz jsprim tgz x json schema tgz vulnerable library found in base branch master vulnerability details json schema is vulnerable to improperly controlled modification of object prototype attributes prototype pollution publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution json schema step up your open source security game with whitesource
0
86,457
10,499,746,716
IssuesEvent
2019-09-26 09:02:44
spring-projects/spring-framework
https://api.github.com/repos/spring-projects/spring-framework
closed
Improve docs for AnnotatedBeanDefinitionReader, @Configuration, and @ContextConfiguration regarding "annotated classes"
in: core in: test type: documentation
## Original Description I find this test interesting: ```java @ExtendWith(SpringExtension.class) @ContextConfiguration(classes = RestTemplate.class) class MyTest { @Autowired RestTemplate restTemplate; @Test void check() { assertThat(restTemplate).isNotNull(); } } ``` This works because while loading a context, `AnnotationConfigContextLoader` registers `@ContextConfiguration` provided classes(`RestTemplate` in this case) to bean definitions, so that they are available for injections. So, this is also possible: ```java @ExtendWith(SpringExtension.class) @ContextConfiguration(classes = { MyTest2.Foo.class, MyTest2.Bar.class }) class MyTest2 { static class Foo { public Foo(Bar bar) { // dependency to Bar } } static class Bar { public Bar() { } } @Autowired Foo foo; @Autowired Bar bar; @Test void check() { assertThat(this.foo).isNotNull(); assertThat(this.bar).isNotNull(); } } ``` I think this is not an intended usage of `@ContextConfiguration#classes`. Probably, by default, filter-out or validate those classes to be `@Configuration` classes. For the case of allowing non `@Configuration` classes (for example, lite-mode), probably provide an explicit option(new attribute) on `@ContextConfiguration`. e.g.: `@ContextConfiguration(classes=MyBean.class, liteMode=true)` ---- This is what I found in real world code base: ```java @ExtendWith({MockitoExtension.class, SpringExtension.class}) @ContextConfiguration(classes = {ResourceBundleMessageSource.class, MyService.class, MyExceptionHandler.class, MyPropertyConfiguration.class, RestTemplate.class}) public class MyExceptionHandlerTest { // ... } ``` So, would be nice not seeing such test class :) --- ## Deliverables Improve documentation for the following regarding "annotated classes". - [x] `AnnotationConfigRegistry` - [x] `AnnotationConfigApplicationContext` - [x] `AnnotationConfigWebApplicationContext` - [x] `AnnotatedBeanDefinitionReader` - [x] `@Configuration` - [x] `@Import` - [x] `AnnotationConfigContextLoader` - [x] `AnnotationConfigContextLoaderUtils` - [x] `@ContextConfiguration` - [x] Reference Manual: `testing.adoc`
1.0
Improve docs for AnnotatedBeanDefinitionReader, @Configuration, and @ContextConfiguration regarding "annotated classes" - ## Original Description I find this test interesting: ```java @ExtendWith(SpringExtension.class) @ContextConfiguration(classes = RestTemplate.class) class MyTest { @Autowired RestTemplate restTemplate; @Test void check() { assertThat(restTemplate).isNotNull(); } } ``` This works because while loading a context, `AnnotationConfigContextLoader` registers `@ContextConfiguration` provided classes(`RestTemplate` in this case) to bean definitions, so that they are available for injections. So, this is also possible: ```java @ExtendWith(SpringExtension.class) @ContextConfiguration(classes = { MyTest2.Foo.class, MyTest2.Bar.class }) class MyTest2 { static class Foo { public Foo(Bar bar) { // dependency to Bar } } static class Bar { public Bar() { } } @Autowired Foo foo; @Autowired Bar bar; @Test void check() { assertThat(this.foo).isNotNull(); assertThat(this.bar).isNotNull(); } } ``` I think this is not an intended usage of `@ContextConfiguration#classes`. Probably, by default, filter-out or validate those classes to be `@Configuration` classes. For the case of allowing non `@Configuration` classes (for example, lite-mode), probably provide an explicit option(new attribute) on `@ContextConfiguration`. e.g.: `@ContextConfiguration(classes=MyBean.class, liteMode=true)` ---- This is what I found in real world code base: ```java @ExtendWith({MockitoExtension.class, SpringExtension.class}) @ContextConfiguration(classes = {ResourceBundleMessageSource.class, MyService.class, MyExceptionHandler.class, MyPropertyConfiguration.class, RestTemplate.class}) public class MyExceptionHandlerTest { // ... } ``` So, would be nice not seeing such test class :) --- ## Deliverables Improve documentation for the following regarding "annotated classes". - [x] `AnnotationConfigRegistry` - [x] `AnnotationConfigApplicationContext` - [x] `AnnotationConfigWebApplicationContext` - [x] `AnnotatedBeanDefinitionReader` - [x] `@Configuration` - [x] `@Import` - [x] `AnnotationConfigContextLoader` - [x] `AnnotationConfigContextLoaderUtils` - [x] `@ContextConfiguration` - [x] Reference Manual: `testing.adoc`
non_test
improve docs for annotatedbeandefinitionreader configuration and contextconfiguration regarding annotated classes original description i find this test interesting java extendwith springextension class contextconfiguration classes resttemplate class class mytest autowired resttemplate resttemplate test void check assertthat resttemplate isnotnull this works because while loading a context annotationconfigcontextloader registers contextconfiguration provided classes resttemplate in this case to bean definitions so that they are available for injections so this is also possible java extendwith springextension class contextconfiguration classes foo class bar class class static class foo public foo bar bar dependency to bar static class bar public bar autowired foo foo autowired bar bar test void check assertthat this foo isnotnull assertthat this bar isnotnull i think this is not an intended usage of contextconfiguration classes probably by default filter out or validate those classes to be configuration classes for the case of allowing non configuration classes for example lite mode probably provide an explicit option new attribute on contextconfiguration e g contextconfiguration classes mybean class litemode true this is what i found in real world code base java extendwith mockitoextension class springextension class contextconfiguration classes resourcebundlemessagesource class myservice class myexceptionhandler class mypropertyconfiguration class resttemplate class public class myexceptionhandlertest so would be nice not seeing such test class deliverables improve documentation for the following regarding annotated classes annotationconfigregistry annotationconfigapplicationcontext annotationconfigwebapplicationcontext annotatedbeandefinitionreader configuration import annotationconfigcontextloader annotationconfigcontextloaderutils contextconfiguration reference manual testing adoc
0
8,352
4,231,022,521
IssuesEvent
2016-07-04 14:21:39
ParabolInc/action
https://api.github.com/repos/ParabolInc/action
closed
npm build:server depends on theme.json too
building
## Issue - Bug Consider adding `WebpackShellPlugin` to server webpack build configuration. - **Estimated effort:** 1 points ([see CONTRIBUTING.md](https://github.com/ParabolInc/action/blob/master/CONTRIBUTING.md#points-and-sizes))
1.0
npm build:server depends on theme.json too - ## Issue - Bug Consider adding `WebpackShellPlugin` to server webpack build configuration. - **Estimated effort:** 1 points ([see CONTRIBUTING.md](https://github.com/ParabolInc/action/blob/master/CONTRIBUTING.md#points-and-sizes))
non_test
npm build server depends on theme json too issue bug consider adding webpackshellplugin to server webpack build configuration estimated effort points
0
491,540
14,165,963,005
IssuesEvent
2020-11-12 08:11:09
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
opened
[0.9.2 develop-111] Glitch/exploit with Steam tractor Plow
Category: Gameplay Priority: Medium
Step to reproduce: - spawn tractor with plow, start tilling: ![image](https://user-images.githubusercontent.com/45708377/98912515-43438880-24d7-11eb-9633-e14953a46dc8.png) - press Q to turn off module and imideatly press Q to turn on module. You will have module turned on with basic tractor speed. Also plow is not in right position. video: https://drive.google.com/file/d/1SKVNKdx8nGK9MU3XrdFw3fchyfmQp6DX/view?usp=sharing
1.0
[0.9.2 develop-111] Glitch/exploit with Steam tractor Plow - Step to reproduce: - spawn tractor with plow, start tilling: ![image](https://user-images.githubusercontent.com/45708377/98912515-43438880-24d7-11eb-9633-e14953a46dc8.png) - press Q to turn off module and imideatly press Q to turn on module. You will have module turned on with basic tractor speed. Also plow is not in right position. video: https://drive.google.com/file/d/1SKVNKdx8nGK9MU3XrdFw3fchyfmQp6DX/view?usp=sharing
non_test
glitch exploit with steam tractor plow step to reproduce spawn tractor with plow start tilling press q to turn off module and imideatly press q to turn on module you will have module turned on with basic tractor speed also plow is not in right position video
0
103,271
8,893,615,621
IssuesEvent
2019-01-16 00:09:41
astropy/astropy
https://api.github.com/repos/astropy/astropy
closed
Test failure with latest developer version of Matplotlib
testing visualization
Here's an example log: https://circleci.com/gh/astrofrog/astropy/1075?utm_campaign=vcs-integration-link&utm_medium=referral&utm_source=github-build-link This seems related to the axis unit functionality. We should do a bisect to figure out which change in Matplotlib caused this.
1.0
Test failure with latest developer version of Matplotlib - Here's an example log: https://circleci.com/gh/astrofrog/astropy/1075?utm_campaign=vcs-integration-link&utm_medium=referral&utm_source=github-build-link This seems related to the axis unit functionality. We should do a bisect to figure out which change in Matplotlib caused this.
test
test failure with latest developer version of matplotlib here s an example log this seems related to the axis unit functionality we should do a bisect to figure out which change in matplotlib caused this
1
690,894
23,676,455,519
IssuesEvent
2022-08-28 06:33:06
Gilded-Games/The-Aether
https://api.github.com/repos/Gilded-Games/The-Aether
closed
Feature: Boss Advancements
priority/medium status/pending-review type/feature version/1.19
- Advancements for bosses. - Likely don't need a custom trigger, just need to manually call to the existing killed entity trigger for everyone in the boss room when the boss is killed.
1.0
Feature: Boss Advancements - - Advancements for bosses. - Likely don't need a custom trigger, just need to manually call to the existing killed entity trigger for everyone in the boss room when the boss is killed.
non_test
feature boss advancements advancements for bosses likely don t need a custom trigger just need to manually call to the existing killed entity trigger for everyone in the boss room when the boss is killed
0
196,514
14,876,634,287
IssuesEvent
2021-01-20 01:18:09
jasonfleming/asgs
https://api.github.com/repos/jasonfleming/asgs
closed
remove ALL conditional env set ups in platforms.sh
in integration testing
This may require creating or moving it to exiting mechanisms (like config/operator_defaults.sh) to provide for additional customizations in the asgs shell, but this should be discouraged. Once we clean up `platforms.sh` we will be in a better position to evaluate the utility of `config/operator_defaults.sh` even being in the repository or move the role of it to the use of a `~/.asgsh_profile` type approach (that should be discouraged).
1.0
remove ALL conditional env set ups in platforms.sh - This may require creating or moving it to exiting mechanisms (like config/operator_defaults.sh) to provide for additional customizations in the asgs shell, but this should be discouraged. Once we clean up `platforms.sh` we will be in a better position to evaluate the utility of `config/operator_defaults.sh` even being in the repository or move the role of it to the use of a `~/.asgsh_profile` type approach (that should be discouraged).
test
remove all conditional env set ups in platforms sh this may require creating or moving it to exiting mechanisms like config operator defaults sh to provide for additional customizations in the asgs shell but this should be discouraged once we clean up platforms sh we will be in a better position to evaluate the utility of config operator defaults sh even being in the repository or move the role of it to the use of a asgsh profile type approach that should be discouraged
1
803,077
29,116,576,610
IssuesEvent
2023-05-17 01:53:22
ML4GW/BBHNet
https://api.github.com/repos/ML4GW/BBHNet
closed
"hopeless" injections
high priority data
To get the most value out of VT calculations from injection runs, the CBC pipelines typically will not inject `hopeless` injections. These are defined as injections with SNR below some value such that they are essentially impossible to recover. This would require calculating the SNR of signals in the `Sampler` using `ml4gw`, and rejecting those below some SNR threshold until we've generated `num_signals` events above the threshold. The changes required for this might fit in well with #267
1.0
"hopeless" injections - To get the most value out of VT calculations from injection runs, the CBC pipelines typically will not inject `hopeless` injections. These are defined as injections with SNR below some value such that they are essentially impossible to recover. This would require calculating the SNR of signals in the `Sampler` using `ml4gw`, and rejecting those below some SNR threshold until we've generated `num_signals` events above the threshold. The changes required for this might fit in well with #267
non_test
hopeless injections to get the most value out of vt calculations from injection runs the cbc pipelines typically will not inject hopeless injections these are defined as injections with snr below some value such that they are essentially impossible to recover this would require calculating the snr of signals in the sampler using and rejecting those below some snr threshold until we ve generated num signals events above the threshold the changes required for this might fit in well with
0
94,412
27,186,451,551
IssuesEvent
2023-02-19 09:15:57
curl/curl
https://api.github.com/repos/curl/curl
closed
Unable detect OpenSSL while set host to arm-linux-gnueabi
build TLS
While I configure the cURL 7.88.0 source with: > ./configure --with-openssl It works correctly. But if I set host to arm-linux-gnueabi, configure like these: > ./configure --with-openssl --host=arm-linux-gnueabi > ./configure --with-openssl --host=arm-linux-gnueabi --target=arm-linux-gnueabi > ./configure --with-openssl --host=arm-linux-gnueabi --target=arm-linux-gnueabi CC=arm-linux-gnueabi-gcc I got an error: > configure: error: --with-openssl was given but OpenSSL could not be detected Ubuntu version is 14.04.6, arm-linux-gnueabi-gcc version is 4.7.3, OpenSSL version is 1.0.1f. What's the reason and how can I solve it? Please help me, Thanks.
1.0
Unable detect OpenSSL while set host to arm-linux-gnueabi - While I configure the cURL 7.88.0 source with: > ./configure --with-openssl It works correctly. But if I set host to arm-linux-gnueabi, configure like these: > ./configure --with-openssl --host=arm-linux-gnueabi > ./configure --with-openssl --host=arm-linux-gnueabi --target=arm-linux-gnueabi > ./configure --with-openssl --host=arm-linux-gnueabi --target=arm-linux-gnueabi CC=arm-linux-gnueabi-gcc I got an error: > configure: error: --with-openssl was given but OpenSSL could not be detected Ubuntu version is 14.04.6, arm-linux-gnueabi-gcc version is 4.7.3, OpenSSL version is 1.0.1f. What's the reason and how can I solve it? Please help me, Thanks.
non_test
unable detect openssl while set host to arm linux gnueabi while i configure the curl source with configure with openssl it works correctly but if i set host to arm linux gnueabi configure like these configure with openssl host arm linux gnueabi configure with openssl host arm linux gnueabi target arm linux gnueabi configure with openssl host arm linux gnueabi target arm linux gnueabi cc arm linux gnueabi gcc i got an error configure error with openssl was given but openssl could not be detected ubuntu version is arm linux gnueabi gcc version is openssl version is what s the reason and how can i solve it please help me thanks
0
121,494
4,817,285,569
IssuesEvent
2016-11-04 13:13:31
handsontable/handsontable
https://api.github.com/repos/handsontable/handsontable
closed
Allow beforeAutofill to be blocked via returning false
Core: callbacks Priority: normal Type: Feature
Currently we are not able to block [`beforeAutofill`](https://docs.handsontable.com/pro/1.7.1/Hooks.html#event:beforeAutofill) by returning `false`.
1.0
Allow beforeAutofill to be blocked via returning false - Currently we are not able to block [`beforeAutofill`](https://docs.handsontable.com/pro/1.7.1/Hooks.html#event:beforeAutofill) by returning `false`.
non_test
allow beforeautofill to be blocked via returning false currently we are not able to block by returning false
0
282,979
24,509,703,143
IssuesEvent
2022-10-10 20:02:15
commercialhaskell/stackage
https://api.github.com/repos/commercialhaskell/stackage
opened
hspec-contrib 0.5.1.1
failure: test-suite
``` Preprocessing test suite 'spec' for hspec-contrib-0.5.1.1.. Building test suite 'spec' for hspec-contrib-0.5.1.1.. [3 of 5] Compiling Test.Hspec.Contrib.HUnitSpec /var/stackage/work/unpack-dir/unpacked/hspec-contrib-0.5.1.1-e6ff010b10d5f3707f4cfbd0af0c39ce36efa3a71a789e5b68d2c87aaa8bdbab/test/Test/Hspec/Contrib/HUnitSpec.hs:11:70: error: • Couldn't match type: [SpecTree ()] with: (a0, [Tree a1 (Item a2)]) Expected: IO (a0, [Tree a1 (Item a2)]) Actual: IO [SpecTree ()] • In the second argument of ‘(<$>)’, namely ‘runSpecM (fromHUnitTest a)’ In the first argument of ‘shouldReturn’, namely ‘bimapForest (const ()) itemRequirement . snd <$> runSpecM (fromHUnitTest a)’ In the expression: bimapForest (const ()) itemRequirement . snd <$> runSpecM (fromHUnitTest a) `shouldReturn` b | 11 | a `shouldYield` b = bimapForest (const ()) itemRequirement . snd <$> runSpecM (fromHUnitTest a) `shouldReturn` b | ^^^^^^^^^^^^^^^^^^^^^^^^^^ curator: Received ExitFailure 1 when running Raw command: /usr/bin/stack --terminal --system-ghc build --test --bench --test-suite-timeout=600 --no-rerun-tests --no-run-benchmarks --haddock --no-interleaved-output --jobs=3 Run from: unpack-dir ```
1.0
hspec-contrib 0.5.1.1 - ``` Preprocessing test suite 'spec' for hspec-contrib-0.5.1.1.. Building test suite 'spec' for hspec-contrib-0.5.1.1.. [3 of 5] Compiling Test.Hspec.Contrib.HUnitSpec /var/stackage/work/unpack-dir/unpacked/hspec-contrib-0.5.1.1-e6ff010b10d5f3707f4cfbd0af0c39ce36efa3a71a789e5b68d2c87aaa8bdbab/test/Test/Hspec/Contrib/HUnitSpec.hs:11:70: error: • Couldn't match type: [SpecTree ()] with: (a0, [Tree a1 (Item a2)]) Expected: IO (a0, [Tree a1 (Item a2)]) Actual: IO [SpecTree ()] • In the second argument of ‘(<$>)’, namely ‘runSpecM (fromHUnitTest a)’ In the first argument of ‘shouldReturn’, namely ‘bimapForest (const ()) itemRequirement . snd <$> runSpecM (fromHUnitTest a)’ In the expression: bimapForest (const ()) itemRequirement . snd <$> runSpecM (fromHUnitTest a) `shouldReturn` b | 11 | a `shouldYield` b = bimapForest (const ()) itemRequirement . snd <$> runSpecM (fromHUnitTest a) `shouldReturn` b | ^^^^^^^^^^^^^^^^^^^^^^^^^^ curator: Received ExitFailure 1 when running Raw command: /usr/bin/stack --terminal --system-ghc build --test --bench --test-suite-timeout=600 --no-rerun-tests --no-run-benchmarks --haddock --no-interleaved-output --jobs=3 Run from: unpack-dir ```
test
hspec contrib preprocessing test suite spec for hspec contrib building test suite spec for hspec contrib compiling test hspec contrib hunitspec var stackage work unpack dir unpacked hspec contrib test test hspec contrib hunitspec hs error • couldn t match type with expected io actual io • in the second argument of ‘ ’ namely ‘runspecm fromhunittest a ’ in the first argument of ‘shouldreturn’ namely ‘bimapforest const itemrequirement snd runspecm fromhunittest a ’ in the expression bimapforest const itemrequirement snd runspecm fromhunittest a shouldreturn b a shouldyield b bimapforest const itemrequirement snd runspecm fromhunittest a shouldreturn b curator received exitfailure when running raw command usr bin stack terminal system ghc build test bench test suite timeout no rerun tests no run benchmarks haddock no interleaved output jobs run from unpack dir
1
419,942
12,230,534,177
IssuesEvent
2020-05-04 05:11:17
roed314/seminars
https://api.github.com/repos/roed314/seminars
opened
Logos on acknowledgments page
low priority
It would be nice to add some logos to the acknowledgments page (e.g., the Simons Foundation logo).
1.0
Logos on acknowledgments page - It would be nice to add some logos to the acknowledgments page (e.g., the Simons Foundation logo).
non_test
logos on acknowledgments page it would be nice to add some logos to the acknowledgments page e g the simons foundation logo
0
92,517
8,366,483,286
IssuesEvent
2018-10-04 09:14:52
stratisproject/StratisBitcoinFullNode
https://api.github.com/repos/stratisproject/StratisBitcoinFullNode
opened
MiningAndPropagatingPOS [Mining and Staking tests impl]
Size XS: 0.5-2 days Test Coverage
**MiningAndPropagatingPOS [[Mining and Staking tests impl]** - MineBlockCheckPeerHasNewBlock | Mine a block and propagate to peers (success) - MineBlockStakeAtInsufficientHeightError | Mine a block insufficient height to stake (error)
1.0
MiningAndPropagatingPOS [Mining and Staking tests impl] - **MiningAndPropagatingPOS [[Mining and Staking tests impl]** - MineBlockCheckPeerHasNewBlock | Mine a block and propagate to peers (success) - MineBlockStakeAtInsufficientHeightError | Mine a block insufficient height to stake (error)
test
miningandpropagatingpos miningandpropagatingpos mineblockcheckpeerhasnewblock mine a block and propagate to peers success mineblockstakeatinsufficientheighterror mine a block insufficient height to stake error
1
6,485
2,848,367,610
IssuesEvent
2015-05-29 22:31:19
gosha1128/NYUAMS
https://api.github.com/repos/gosha1128/NYUAMS
closed
Test Case: Admin>emails - table stops populating
0.0 DEV 1.4 VERIFIED IN DEV 2.3 high fixed load testing
Table stops populating after 21 lines have been populated https://drive.google.com/open?id=0B_9JEK6SVk1bamNTMEJqUExNQW8&authuser=0
1.0
Test Case: Admin>emails - table stops populating - Table stops populating after 21 lines have been populated https://drive.google.com/open?id=0B_9JEK6SVk1bamNTMEJqUExNQW8&authuser=0
test
test case admin emails table stops populating table stops populating after lines have been populated
1
12,898
15,169,784,361
IssuesEvent
2021-02-12 21:49:10
Foundry376/Mailspring
https://api.github.com/repos/Foundry376/Mailspring
closed
Can't configure archive folder for Gmail account?
audit bug mailserver-incompatibility needs-reproduction
<!-- Thanks for taking the time to file an issue! The Mailspring community uses GitHub issues to coordinate development. Before filling out this form, try searching for both open and closed issues here: https://github.com/Foundry376/Mailspring/issues. Upvoting an existing sync issue is MUCH better than submitting a new one - it often takes a while for them to be merged and we use issue upvotes to prioritize development. --> ##### What is the behavior you're seeing? New mailspring user here. I setup my gmail account and I was surprised to note that I can't configure the "Archive" mailbox for gmail: ![Screenshot from 2020-09-22 20-05-40](https://user-images.githubusercontent.com/10998835/93915330-65026980-fcf7-11ea-8e7c-5ca5f1ebba3e.png) I am able to configure this mailbox for the other accounts I've setup. This is particularly important to me as I have all of my unattended mailing lists messages in the default archive mailbox "[Gmail]/All Mail" and I want to use a different archive mailbox, just like my other mail clients (K9). ##### What email provider (Gmail, Exchange, etc.) are you trying to use? Gmail ##### What operating system are you using? NixOS ##### What version of Mailspring are you using? 1.7.8 ##### Is the issue reproducible with a particular attachment, message, signature, etc? Irrelevant. ##### Other comments Damn this application is beautiful! :heart_eyes:
True
Can't configure archive folder for Gmail account? - <!-- Thanks for taking the time to file an issue! The Mailspring community uses GitHub issues to coordinate development. Before filling out this form, try searching for both open and closed issues here: https://github.com/Foundry376/Mailspring/issues. Upvoting an existing sync issue is MUCH better than submitting a new one - it often takes a while for them to be merged and we use issue upvotes to prioritize development. --> ##### What is the behavior you're seeing? New mailspring user here. I setup my gmail account and I was surprised to note that I can't configure the "Archive" mailbox for gmail: ![Screenshot from 2020-09-22 20-05-40](https://user-images.githubusercontent.com/10998835/93915330-65026980-fcf7-11ea-8e7c-5ca5f1ebba3e.png) I am able to configure this mailbox for the other accounts I've setup. This is particularly important to me as I have all of my unattended mailing lists messages in the default archive mailbox "[Gmail]/All Mail" and I want to use a different archive mailbox, just like my other mail clients (K9). ##### What email provider (Gmail, Exchange, etc.) are you trying to use? Gmail ##### What operating system are you using? NixOS ##### What version of Mailspring are you using? 1.7.8 ##### Is the issue reproducible with a particular attachment, message, signature, etc? Irrelevant. ##### Other comments Damn this application is beautiful! :heart_eyes:
non_test
can t configure archive folder for gmail account thanks for taking the time to file an issue the mailspring community uses github issues to coordinate development before filling out this form try searching for both open and closed issues here upvoting an existing sync issue is much better than submitting a new one it often takes a while for them to be merged and we use issue upvotes to prioritize development what is the behavior you re seeing new mailspring user here i setup my gmail account and i was surprised to note that i can t configure the archive mailbox for gmail i am able to configure this mailbox for the other accounts i ve setup this is particularly important to me as i have all of my unattended mailing lists messages in the default archive mailbox all mail and i want to use a different archive mailbox just like my other mail clients what email provider gmail exchange etc are you trying to use gmail what operating system are you using nixos what version of mailspring are you using is the issue reproducible with a particular attachment message signature etc irrelevant other comments damn this application is beautiful heart eyes
0
268,111
23,345,643,343
IssuesEvent
2022-08-09 17:41:16
wso2/product-is
https://api.github.com/repos/wso2/product-is
closed
Console app does not display device_code grant type
ui bug frontend 6.0.0-rc-testing
**Describe the issue:** he console can be used to view service providers and update the necessary configurations. The protocol section of the service provider only shows limited grant types for service provider applications. device_code grant type is not under allowed grant types (where as management console does list that) ![Screenshot from 2022-08-09 19-31-00](https://user-images.githubusercontent.com/32198547/183668920-5faf6f25-e469-45e3-bbc0-369b85eea56b.png) **How to reproduce:** Log into console app Go to applications and select Console app. View protocol section of the service provider app. **Expected behavior:** device_code should be listed under allowed grant_types
1.0
Console app does not display device_code grant type - **Describe the issue:** he console can be used to view service providers and update the necessary configurations. The protocol section of the service provider only shows limited grant types for service provider applications. device_code grant type is not under allowed grant types (where as management console does list that) ![Screenshot from 2022-08-09 19-31-00](https://user-images.githubusercontent.com/32198547/183668920-5faf6f25-e469-45e3-bbc0-369b85eea56b.png) **How to reproduce:** Log into console app Go to applications and select Console app. View protocol section of the service provider app. **Expected behavior:** device_code should be listed under allowed grant_types
test
console app does not display device code grant type describe the issue he console can be used to view service providers and update the necessary configurations the protocol section of the service provider only shows limited grant types for service provider applications device code grant type is not under allowed grant types where as management console does list that how to reproduce log into console app go to applications and select console app view protocol section of the service provider app expected behavior device code should be listed under allowed grant types
1
371,759
10,981,078,813
IssuesEvent
2019-11-30 18:59:43
BendroCorp/bendrocorp-app
https://api.github.com/repos/BendroCorp/bendrocorp-app
opened
System Map Search View
api supported effort: high feature priority:high
- [ ] Offer a map view or a searchable list view - switch via toggle and persist preference - [ ] Search View (will be the default mobile view when added) - [ ] Searching for something adds it to the recently viewed list at the top of the system map - [ ] Has a list of all items contained within system map and their essential data - [ ] Searchable by title - [ ] Opens to an object details page
1.0
System Map Search View - - [ ] Offer a map view or a searchable list view - switch via toggle and persist preference - [ ] Search View (will be the default mobile view when added) - [ ] Searching for something adds it to the recently viewed list at the top of the system map - [ ] Has a list of all items contained within system map and their essential data - [ ] Searchable by title - [ ] Opens to an object details page
non_test
system map search view offer a map view or a searchable list view switch via toggle and persist preference search view will be the default mobile view when added searching for something adds it to the recently viewed list at the top of the system map has a list of all items contained within system map and their essential data searchable by title opens to an object details page
0
193,712
14,660,464,836
IssuesEvent
2020-12-28 23:47:39
github-vet/rangeloop-pointer-findings
https://api.github.com/repos/github-vet/rangeloop-pointer-findings
closed
gravitational/teleport: integration/integration_test.go; 281 LoC
fresh huge test
Found a possible issue in [gravitational/teleport](https://www.github.com/gravitational/teleport) at [integration/integration_test.go](https://github.com/gravitational/teleport/blob/38c9804738112eceedd209d9cb5bda7ffe3ab6f5/integration/integration_test.go#L299-L579) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > range-loop variable tt used in defer or goroutine at line 378 [Click here to see the code in its original context.](https://github.com/gravitational/teleport/blob/38c9804738112eceedd209d9cb5bda7ffe3ab6f5/integration/integration_test.go#L299-L579) <details> <summary>Click here to show the 281 line(s) of Go which triggered the analyzer.</summary> ```go for _, tt := range tests { comment := check.Commentf(tt.comment) makeConfig := func() (*check.C, []string, []*InstanceSecrets, *service.Config) { clusterConfig, err := services.NewClusterConfig(services.ClusterConfigSpecV3{ SessionRecording: tt.inRecordLocation, Audit: services.AuditConfig{AuditSessionsURI: tt.auditSessionsURI}, LocalAuth: services.NewBool(true), }) c.Assert(err, check.IsNil, comment) tconf := s.defaultServiceConfig() tconf.Auth.Enabled = true tconf.Auth.ClusterConfig = clusterConfig tconf.Proxy.Enabled = true tconf.Proxy.DisableWebService = true tconf.Proxy.DisableWebInterface = true tconf.SSH.Enabled = true return c, nil, nil, tconf } t := s.newTeleportWithConfig(makeConfig()) defer t.StopAll() // Start a node. nodeSSHPort := s.getPorts(1)[0] nodeConfig := func() *service.Config { tconf := s.defaultServiceConfig() tconf.HostUUID = "node" tconf.Hostname = "node" tconf.SSH.Enabled = true tconf.SSH.Addr.Addr = net.JoinHostPort(t.Hostname, fmt.Sprintf("%v", nodeSSHPort)) return tconf } nodeProcess, err := t.StartNode(nodeConfig()) c.Assert(err, check.IsNil, comment) // get access to a authClient for the cluster site := t.GetSiteAPI(Site) c.Assert(site, check.NotNil, comment) // wait 10 seconds for both nodes to show up, otherwise // we'll have trouble connecting to the node below. waitForNodes := func(site auth.ClientI, count int) error { tickCh := time.Tick(500 * time.Millisecond) stopCh := time.After(10 * time.Second) for { select { case <-tickCh: nodesInSite, err := site.GetNodes(defaults.Namespace, services.SkipValidation()) if err != nil && !trace.IsNotFound(err) { return trace.Wrap(err) } if got, want := len(nodesInSite), count; got == want { return nil } case <-stopCh: return trace.BadParameter("waited 10s, did find %v nodes", count) } } } err = waitForNodes(site, 2) c.Assert(err, check.IsNil, comment) // should have no sessions: sessions, err := site.GetSessions(defaults.Namespace) c.Assert(err, check.IsNil, comment) c.Assert(len(sessions), check.Equals, 0, comment) // create interactive session (this goroutine is this user's terminal time) endC := make(chan error) myTerm := NewTerminal(250) go func() { cl, err := t.NewClient(ClientConfig{ Login: s.me.Username, Cluster: Site, Host: Host, Port: nodeSSHPort, ForwardAgent: tt.inForwardAgent, }) c.Assert(err, check.IsNil, comment) cl.Stdout = myTerm cl.Stdin = myTerm err = cl.SSH(context.TODO(), []string{}, false) endC <- err }() // wait until we've found the session in the audit log getSession := func(site auth.ClientI) (*session.Session, error) { tickCh := time.Tick(500 * time.Millisecond) stopCh := time.After(10 * time.Second) for { select { case <-tickCh: sessions, err = site.GetSessions(defaults.Namespace) if err != nil { return nil, trace.Wrap(err) } if len(sessions) != 1 { continue } return &sessions[0], nil case <-stopCh: return nil, trace.BadParameter("unable to find sessions after 10s (mode=%v)", tt.inRecordLocation) } } } session, err := getSession(site) c.Assert(err, check.IsNil, comment) // wait for the user to join this session: for len(session.Parties) == 0 { time.Sleep(time.Millisecond * 5) session, err = site.GetSession(defaults.Namespace, sessions[0].ID) c.Assert(err, check.IsNil, comment) } // make sure it's us who joined! :) c.Assert(session.Parties[0].User, check.Equals, s.me.Username, comment) // lets type "echo hi" followed by "enter" and then "exit" + "enter": myTerm.Type("\aecho hi\n\r\aexit\n\r\a") // wait for session to end: select { case <-endC: case <-time.After(10 * time.Second): c.Fatalf("%s: Timeout waiting for session to finish.", tt.comment) } // wait for the upload of the right session to complete timeoutC := time.After(10 * time.Second) loop: for { select { case event := <-t.UploadEventsC: if event.SessionID != string(session.ID) { c.Logf("Skipping mismatching session %v, expecting upload of %v.", event.SessionID, session.ID) continue } break loop case <-timeoutC: dumpGoroutineProfile() c.Fatalf("%s: Timeout waiting for upload of session %v to complete to %v", tt.comment, session.ID, tt.auditSessionsURI) } } // read back the entire session (we have to try several times until we get back // everything because the session is closing) var sessionStream []byte for i := 0; i < 6; i++ { sessionStream, err = site.GetSessionChunk(defaults.Namespace, session.ID, 0, events.MaxChunkBytes) c.Assert(err, check.IsNil, comment) if strings.Contains(string(sessionStream), "exit") { break } time.Sleep(time.Millisecond * 250) if i >= 5 { // session stream keeps coming back short c.Fatalf("%s: Stream is not getting data: %q.", tt.comment, string(sessionStream)) } } // see what we got. It looks different based on bash settings, but here it is // on Ev's machine (hostname is 'edsger'): // // edsger ~: echo hi // hi // edsger ~: exit // logout // sessionComment := check.Commentf("%q", string(sessionStream)) c.Assert(strings.Contains(string(sessionStream), "echo hi"), check.Equals, true, sessionComment) c.Assert(strings.Contains(string(sessionStream), "exit"), check.Equals, true, sessionComment) // Wait until session.start, session.leave, and session.end events have arrived. getSessions := func(site auth.ClientI) ([]events.EventFields, error) { tickCh := time.Tick(500 * time.Millisecond) stopCh := time.After(10 * time.Second) for { select { case <-tickCh: // Get all session events from the backend. sessionEvents, err := site.GetSessionEvents(defaults.Namespace, session.ID, 0, false) if err != nil { return nil, trace.Wrap(err) } // Look through all session events for the three wanted. var hasStart bool var hasEnd bool var hasLeave bool for _, se := range sessionEvents { if se.GetType() == events.SessionStartEvent { hasStart = true } if se.GetType() == events.SessionEndEvent { hasEnd = true } if se.GetType() == events.SessionLeaveEvent { hasLeave = true } } // Make sure all three events were found. if hasStart && hasEnd && hasLeave { return sessionEvents, nil } case <-stopCh: return nil, trace.BadParameter("unable to find all session events after 10s (mode=%v)", tt.inRecordLocation) } } } history, err := getSessions(site) c.Assert(err, check.IsNil, comment) getChunk := func(e events.EventFields, maxlen int) string { offset := e.GetInt("offset") length := e.GetInt("bytes") if length == 0 { return "" } if length > maxlen { length = maxlen } return string(sessionStream[offset : offset+length]) } findByType := func(et string) events.EventFields { for _, e := range history { if e.GetType() == et { return e } } return nil } // there should alwys be 'session.start' event (and it must be first) first := history[0] start := findByType(events.SessionStartEvent) c.Assert(start, check.DeepEquals, first, comment) c.Assert(start.GetInt("bytes"), check.Equals, 0, comment) c.Assert(start.GetString(events.SessionEventID) != "", check.Equals, true, comment) c.Assert(start.GetString(events.TerminalSize) != "", check.Equals, true, comment) // If session are being recorded at nodes, the SessionServerID should contain // the ID of the node. If sessions are being recorded at the proxy, then // SessionServerID should be that of the proxy. expectedServerID := nodeProcess.Config.HostUUID if services.IsRecordAtProxy(tt.inRecordLocation) { expectedServerID = t.Process.Config.HostUUID } c.Assert(start.GetString(events.SessionServerID), check.Equals, expectedServerID, comment) // make sure data is recorded properly out := &bytes.Buffer{} for _, e := range history { out.WriteString(getChunk(e, 1000)) } recorded := replaceNewlines(out.String()) c.Assert(recorded, check.Matches, ".*exit.*", comment) c.Assert(recorded, check.Matches, ".*echo hi.*", comment) // there should alwys be 'session.end' event end := findByType(events.SessionEndEvent) c.Assert(end, check.NotNil) c.Assert(end.GetInt("bytes"), check.Equals, 0) c.Assert(end.GetString(events.SessionEventID) != "", check.Equals, true) // there should alwys be 'session.leave' event leave := findByType(events.SessionLeaveEvent) c.Assert(leave, check.NotNil, comment) c.Assert(leave.GetInt("bytes"), check.Equals, 0, comment) c.Assert(leave.GetString(events.SessionEventID) != "", check.Equals, true, comment) // all of them should have a proper time: for _, e := range history { c.Assert(e.GetTime("time").IsZero(), check.Equals, false, comment) } } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 38c9804738112eceedd209d9cb5bda7ffe3ab6f5
1.0
gravitational/teleport: integration/integration_test.go; 281 LoC - Found a possible issue in [gravitational/teleport](https://www.github.com/gravitational/teleport) at [integration/integration_test.go](https://github.com/gravitational/teleport/blob/38c9804738112eceedd209d9cb5bda7ffe3ab6f5/integration/integration_test.go#L299-L579) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > range-loop variable tt used in defer or goroutine at line 378 [Click here to see the code in its original context.](https://github.com/gravitational/teleport/blob/38c9804738112eceedd209d9cb5bda7ffe3ab6f5/integration/integration_test.go#L299-L579) <details> <summary>Click here to show the 281 line(s) of Go which triggered the analyzer.</summary> ```go for _, tt := range tests { comment := check.Commentf(tt.comment) makeConfig := func() (*check.C, []string, []*InstanceSecrets, *service.Config) { clusterConfig, err := services.NewClusterConfig(services.ClusterConfigSpecV3{ SessionRecording: tt.inRecordLocation, Audit: services.AuditConfig{AuditSessionsURI: tt.auditSessionsURI}, LocalAuth: services.NewBool(true), }) c.Assert(err, check.IsNil, comment) tconf := s.defaultServiceConfig() tconf.Auth.Enabled = true tconf.Auth.ClusterConfig = clusterConfig tconf.Proxy.Enabled = true tconf.Proxy.DisableWebService = true tconf.Proxy.DisableWebInterface = true tconf.SSH.Enabled = true return c, nil, nil, tconf } t := s.newTeleportWithConfig(makeConfig()) defer t.StopAll() // Start a node. nodeSSHPort := s.getPorts(1)[0] nodeConfig := func() *service.Config { tconf := s.defaultServiceConfig() tconf.HostUUID = "node" tconf.Hostname = "node" tconf.SSH.Enabled = true tconf.SSH.Addr.Addr = net.JoinHostPort(t.Hostname, fmt.Sprintf("%v", nodeSSHPort)) return tconf } nodeProcess, err := t.StartNode(nodeConfig()) c.Assert(err, check.IsNil, comment) // get access to a authClient for the cluster site := t.GetSiteAPI(Site) c.Assert(site, check.NotNil, comment) // wait 10 seconds for both nodes to show up, otherwise // we'll have trouble connecting to the node below. waitForNodes := func(site auth.ClientI, count int) error { tickCh := time.Tick(500 * time.Millisecond) stopCh := time.After(10 * time.Second) for { select { case <-tickCh: nodesInSite, err := site.GetNodes(defaults.Namespace, services.SkipValidation()) if err != nil && !trace.IsNotFound(err) { return trace.Wrap(err) } if got, want := len(nodesInSite), count; got == want { return nil } case <-stopCh: return trace.BadParameter("waited 10s, did find %v nodes", count) } } } err = waitForNodes(site, 2) c.Assert(err, check.IsNil, comment) // should have no sessions: sessions, err := site.GetSessions(defaults.Namespace) c.Assert(err, check.IsNil, comment) c.Assert(len(sessions), check.Equals, 0, comment) // create interactive session (this goroutine is this user's terminal time) endC := make(chan error) myTerm := NewTerminal(250) go func() { cl, err := t.NewClient(ClientConfig{ Login: s.me.Username, Cluster: Site, Host: Host, Port: nodeSSHPort, ForwardAgent: tt.inForwardAgent, }) c.Assert(err, check.IsNil, comment) cl.Stdout = myTerm cl.Stdin = myTerm err = cl.SSH(context.TODO(), []string{}, false) endC <- err }() // wait until we've found the session in the audit log getSession := func(site auth.ClientI) (*session.Session, error) { tickCh := time.Tick(500 * time.Millisecond) stopCh := time.After(10 * time.Second) for { select { case <-tickCh: sessions, err = site.GetSessions(defaults.Namespace) if err != nil { return nil, trace.Wrap(err) } if len(sessions) != 1 { continue } return &sessions[0], nil case <-stopCh: return nil, trace.BadParameter("unable to find sessions after 10s (mode=%v)", tt.inRecordLocation) } } } session, err := getSession(site) c.Assert(err, check.IsNil, comment) // wait for the user to join this session: for len(session.Parties) == 0 { time.Sleep(time.Millisecond * 5) session, err = site.GetSession(defaults.Namespace, sessions[0].ID) c.Assert(err, check.IsNil, comment) } // make sure it's us who joined! :) c.Assert(session.Parties[0].User, check.Equals, s.me.Username, comment) // lets type "echo hi" followed by "enter" and then "exit" + "enter": myTerm.Type("\aecho hi\n\r\aexit\n\r\a") // wait for session to end: select { case <-endC: case <-time.After(10 * time.Second): c.Fatalf("%s: Timeout waiting for session to finish.", tt.comment) } // wait for the upload of the right session to complete timeoutC := time.After(10 * time.Second) loop: for { select { case event := <-t.UploadEventsC: if event.SessionID != string(session.ID) { c.Logf("Skipping mismatching session %v, expecting upload of %v.", event.SessionID, session.ID) continue } break loop case <-timeoutC: dumpGoroutineProfile() c.Fatalf("%s: Timeout waiting for upload of session %v to complete to %v", tt.comment, session.ID, tt.auditSessionsURI) } } // read back the entire session (we have to try several times until we get back // everything because the session is closing) var sessionStream []byte for i := 0; i < 6; i++ { sessionStream, err = site.GetSessionChunk(defaults.Namespace, session.ID, 0, events.MaxChunkBytes) c.Assert(err, check.IsNil, comment) if strings.Contains(string(sessionStream), "exit") { break } time.Sleep(time.Millisecond * 250) if i >= 5 { // session stream keeps coming back short c.Fatalf("%s: Stream is not getting data: %q.", tt.comment, string(sessionStream)) } } // see what we got. It looks different based on bash settings, but here it is // on Ev's machine (hostname is 'edsger'): // // edsger ~: echo hi // hi // edsger ~: exit // logout // sessionComment := check.Commentf("%q", string(sessionStream)) c.Assert(strings.Contains(string(sessionStream), "echo hi"), check.Equals, true, sessionComment) c.Assert(strings.Contains(string(sessionStream), "exit"), check.Equals, true, sessionComment) // Wait until session.start, session.leave, and session.end events have arrived. getSessions := func(site auth.ClientI) ([]events.EventFields, error) { tickCh := time.Tick(500 * time.Millisecond) stopCh := time.After(10 * time.Second) for { select { case <-tickCh: // Get all session events from the backend. sessionEvents, err := site.GetSessionEvents(defaults.Namespace, session.ID, 0, false) if err != nil { return nil, trace.Wrap(err) } // Look through all session events for the three wanted. var hasStart bool var hasEnd bool var hasLeave bool for _, se := range sessionEvents { if se.GetType() == events.SessionStartEvent { hasStart = true } if se.GetType() == events.SessionEndEvent { hasEnd = true } if se.GetType() == events.SessionLeaveEvent { hasLeave = true } } // Make sure all three events were found. if hasStart && hasEnd && hasLeave { return sessionEvents, nil } case <-stopCh: return nil, trace.BadParameter("unable to find all session events after 10s (mode=%v)", tt.inRecordLocation) } } } history, err := getSessions(site) c.Assert(err, check.IsNil, comment) getChunk := func(e events.EventFields, maxlen int) string { offset := e.GetInt("offset") length := e.GetInt("bytes") if length == 0 { return "" } if length > maxlen { length = maxlen } return string(sessionStream[offset : offset+length]) } findByType := func(et string) events.EventFields { for _, e := range history { if e.GetType() == et { return e } } return nil } // there should alwys be 'session.start' event (and it must be first) first := history[0] start := findByType(events.SessionStartEvent) c.Assert(start, check.DeepEquals, first, comment) c.Assert(start.GetInt("bytes"), check.Equals, 0, comment) c.Assert(start.GetString(events.SessionEventID) != "", check.Equals, true, comment) c.Assert(start.GetString(events.TerminalSize) != "", check.Equals, true, comment) // If session are being recorded at nodes, the SessionServerID should contain // the ID of the node. If sessions are being recorded at the proxy, then // SessionServerID should be that of the proxy. expectedServerID := nodeProcess.Config.HostUUID if services.IsRecordAtProxy(tt.inRecordLocation) { expectedServerID = t.Process.Config.HostUUID } c.Assert(start.GetString(events.SessionServerID), check.Equals, expectedServerID, comment) // make sure data is recorded properly out := &bytes.Buffer{} for _, e := range history { out.WriteString(getChunk(e, 1000)) } recorded := replaceNewlines(out.String()) c.Assert(recorded, check.Matches, ".*exit.*", comment) c.Assert(recorded, check.Matches, ".*echo hi.*", comment) // there should alwys be 'session.end' event end := findByType(events.SessionEndEvent) c.Assert(end, check.NotNil) c.Assert(end.GetInt("bytes"), check.Equals, 0) c.Assert(end.GetString(events.SessionEventID) != "", check.Equals, true) // there should alwys be 'session.leave' event leave := findByType(events.SessionLeaveEvent) c.Assert(leave, check.NotNil, comment) c.Assert(leave.GetInt("bytes"), check.Equals, 0, comment) c.Assert(leave.GetString(events.SessionEventID) != "", check.Equals, true, comment) // all of them should have a proper time: for _, e := range history { c.Assert(e.GetTime("time").IsZero(), check.Equals, false, comment) } } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 38c9804738112eceedd209d9cb5bda7ffe3ab6f5
test
gravitational teleport integration integration test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message range loop variable tt used in defer or goroutine at line click here to show the line s of go which triggered the analyzer go for tt range tests comment check commentf tt comment makeconfig func check c string instancesecrets service config clusterconfig err services newclusterconfig services sessionrecording tt inrecordlocation audit services auditconfig auditsessionsuri tt auditsessionsuri localauth services newbool true c assert err check isnil comment tconf s defaultserviceconfig tconf auth enabled true tconf auth clusterconfig clusterconfig tconf proxy enabled true tconf proxy disablewebservice true tconf proxy disablewebinterface true tconf ssh enabled true return c nil nil tconf t s newteleportwithconfig makeconfig defer t stopall start a node nodesshport s getports nodeconfig func service config tconf s defaultserviceconfig tconf hostuuid node tconf hostname node tconf ssh enabled true tconf ssh addr addr net joinhostport t hostname fmt sprintf v nodesshport return tconf nodeprocess err t startnode nodeconfig c assert err check isnil comment get access to a authclient for the cluster site t getsiteapi site c assert site check notnil comment wait seconds for both nodes to show up otherwise we ll have trouble connecting to the node below waitfornodes func site auth clienti count int error tickch time tick time millisecond stopch time after time second for select case tickch nodesinsite err site getnodes defaults namespace services skipvalidation if err nil trace isnotfound err return trace wrap err if got want len nodesinsite count got want return nil case stopch return trace badparameter waited did find v nodes count err waitfornodes site c assert err check isnil comment should have no sessions sessions err site getsessions defaults namespace c assert err check isnil comment c assert len sessions check equals comment create interactive session this goroutine is this user s terminal time endc make chan error myterm newterminal go func cl err t newclient clientconfig login s me username cluster site host host port nodesshport forwardagent tt inforwardagent c assert err check isnil comment cl stdout myterm cl stdin myterm err cl ssh context todo string false endc err wait until we ve found the session in the audit log getsession func site auth clienti session session error tickch time tick time millisecond stopch time after time second for select case tickch sessions err site getsessions defaults namespace if err nil return nil trace wrap err if len sessions continue return sessions nil case stopch return nil trace badparameter unable to find sessions after mode v tt inrecordlocation session err getsession site c assert err check isnil comment wait for the user to join this session for len session parties time sleep time millisecond session err site getsession defaults namespace sessions id c assert err check isnil comment make sure it s us who joined c assert session parties user check equals s me username comment lets type echo hi followed by enter and then exit enter myterm type aecho hi n r aexit n r a wait for session to end select case endc case time after time second c fatalf s timeout waiting for session to finish tt comment wait for the upload of the right session to complete timeoutc time after time second loop for select case event t uploadeventsc if event sessionid string session id c logf skipping mismatching session v expecting upload of v event sessionid session id continue break loop case timeoutc dumpgoroutineprofile c fatalf s timeout waiting for upload of session v to complete to v tt comment session id tt auditsessionsuri read back the entire session we have to try several times until we get back everything because the session is closing var sessionstream byte for i i i sessionstream err site getsessionchunk defaults namespace session id events maxchunkbytes c assert err check isnil comment if strings contains string sessionstream exit break time sleep time millisecond if i session stream keeps coming back short c fatalf s stream is not getting data q tt comment string sessionstream see what we got it looks different based on bash settings but here it is on ev s machine hostname is edsger edsger echo hi hi edsger exit logout sessioncomment check commentf q string sessionstream c assert strings contains string sessionstream echo hi check equals true sessioncomment c assert strings contains string sessionstream exit check equals true sessioncomment wait until session start session leave and session end events have arrived getsessions func site auth clienti events eventfields error tickch time tick time millisecond stopch time after time second for select case tickch get all session events from the backend sessionevents err site getsessionevents defaults namespace session id false if err nil return nil trace wrap err look through all session events for the three wanted var hasstart bool var hasend bool var hasleave bool for se range sessionevents if se gettype events sessionstartevent hasstart true if se gettype events sessionendevent hasend true if se gettype events sessionleaveevent hasleave true make sure all three events were found if hasstart hasend hasleave return sessionevents nil case stopch return nil trace badparameter unable to find all session events after mode v tt inrecordlocation history err getsessions site c assert err check isnil comment getchunk func e events eventfields maxlen int string offset e getint offset length e getint bytes if length return if length maxlen length maxlen return string sessionstream findbytype func et string events eventfields for e range history if e gettype et return e return nil there should alwys be session start event and it must be first first history start findbytype events sessionstartevent c assert start check deepequals first comment c assert start getint bytes check equals comment c assert start getstring events sessioneventid check equals true comment c assert start getstring events terminalsize check equals true comment if session are being recorded at nodes the sessionserverid should contain the id of the node if sessions are being recorded at the proxy then sessionserverid should be that of the proxy expectedserverid nodeprocess config hostuuid if services isrecordatproxy tt inrecordlocation expectedserverid t process config hostuuid c assert start getstring events sessionserverid check equals expectedserverid comment make sure data is recorded properly out bytes buffer for e range history out writestring getchunk e recorded replacenewlines out string c assert recorded check matches exit comment c assert recorded check matches echo hi comment there should alwys be session end event end findbytype events sessionendevent c assert end check notnil c assert end getint bytes check equals c assert end getstring events sessioneventid check equals true there should alwys be session leave event leave findbytype events sessionleaveevent c assert leave check notnil comment c assert leave getint bytes check equals comment c assert leave getstring events sessioneventid check equals true comment all of them should have a proper time for e range history c assert e gettime time iszero check equals false comment leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
1
29,280
8,316,494,766
IssuesEvent
2018-09-25 09:12:35
openshiftio/openshift.io
https://api.github.com/repos/openshiftio/openshift.io
closed
Update osio-pipeline to process resources of same kind from different template file
area/pipelines status/in-progress team/build-cd type/task
Update osio-pipeline to process multiple resources of the same kind from multiple template file and can be passed to build and deploy API. https://openshift.io/openshiftio/Openshift_io/plan/detail/673
1.0
Update osio-pipeline to process resources of same kind from different template file - Update osio-pipeline to process multiple resources of the same kind from multiple template file and can be passed to build and deploy API. https://openshift.io/openshiftio/Openshift_io/plan/detail/673
non_test
update osio pipeline to process resources of same kind from different template file update osio pipeline to process multiple resources of the same kind from multiple template file and can be passed to build and deploy api
0
339,411
30,446,197,789
IssuesEvent
2023-07-15 17:47:42
natiatabatadzebtu/mid-term-versioning
https://api.github.com/repos/natiatabatadzebtu/mid-term-versioning
opened
640c851 failed unit and formatting tests.
ci-pytest ci-black
Automatically generated message 640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests. Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689442899/pytest.html Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689442899/black.html Automatically generated message 640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests. Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689442927/pytest.html Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689442927/black.html Automatically generated message 640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests. Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689442954/pytest.html Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689442954/black.html Automatically generated message 640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests. Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689442981/pytest.html Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689442981/black.html Automatically generated message 640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests. Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443009/pytest.html Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443009/black.html Automatically generated message 640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests. Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443037/pytest.html Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443037/black.html Automatically generated message 640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests. Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443064/pytest.html Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443064/black.html Automatically generated message 640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests. Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443091/pytest.html Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443091/black.html Automatically generated message 640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests. Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443119/pytest.html Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443119/black.html Automatically generated message 640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests. Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443147/pytest.html Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443147/black.html Automatically generated message 640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests. Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443175/pytest.html Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443175/black.html Automatically generated message 640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests. Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443203/pytest.html Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443203/black.html Automatically generated message 640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests. Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443230/pytest.html Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443230/black.html Automatically generated message 640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests. Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443258/pytest.html Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443258/black.html
1.0
640c851 failed unit and formatting tests. - Automatically generated message 640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests. Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689442899/pytest.html Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689442899/black.html Automatically generated message 640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests. Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689442927/pytest.html Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689442927/black.html Automatically generated message 640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests. Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689442954/pytest.html Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689442954/black.html Automatically generated message 640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests. Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689442981/pytest.html Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689442981/black.html Automatically generated message 640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests. Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443009/pytest.html Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443009/black.html Automatically generated message 640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests. Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443037/pytest.html Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443037/black.html Automatically generated message 640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests. Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443064/pytest.html Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443064/black.html Automatically generated message 640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests. Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443091/pytest.html Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443091/black.html Automatically generated message 640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests. Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443119/pytest.html Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443119/black.html Automatically generated message 640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests. Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443147/pytest.html Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443147/black.html Automatically generated message 640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests. Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443175/pytest.html Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443175/black.html Automatically generated message 640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests. Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443203/pytest.html Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443203/black.html Automatically generated message 640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests. Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443230/pytest.html Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443230/black.html Automatically generated message 640c8517801ba0baab4fcca90ce913ede6618f37 failed unit and formatting tests. Pytest report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443258/pytest.html Black report: https://natiatabatadzebtu.github.io/mid-term-versioning-ci/640c8517801ba0baab4fcca90ce913ede6618f37-1689443258/black.html
test
failed unit and formatting tests automatically generated message failed unit and formatting tests pytest report black report automatically generated message failed unit and formatting tests pytest report black report automatically generated message failed unit and formatting tests pytest report black report automatically generated message failed unit and formatting tests pytest report black report automatically generated message failed unit and formatting tests pytest report black report automatically generated message failed unit and formatting tests pytest report black report automatically generated message failed unit and formatting tests pytest report black report automatically generated message failed unit and formatting tests pytest report black report automatically generated message failed unit and formatting tests pytest report black report automatically generated message failed unit and formatting tests pytest report black report automatically generated message failed unit and formatting tests pytest report black report automatically generated message failed unit and formatting tests pytest report black report automatically generated message failed unit and formatting tests pytest report black report automatically generated message failed unit and formatting tests pytest report black report
1
50,391
6,086,442,927
IssuesEvent
2017-06-18 01:05:02
Semantic-Org/Semantic-UI
https://api.github.com/repos/Semantic-Org/Semantic-UI
closed
Popup keeps appear in 'bottom center' even if i set it's position as 'top center'
Needs Test Case Usage Question
So, what is this about? I set properly `position: 'top center'`, but popup appears in bottom center position without any errors and warnings thrown.
1.0
Popup keeps appear in 'bottom center' even if i set it's position as 'top center' - So, what is this about? I set properly `position: 'top center'`, but popup appears in bottom center position without any errors and warnings thrown.
test
popup keeps appear in bottom center even if i set it s position as top center so what is this about i set properly position top center but popup appears in bottom center position without any errors and warnings thrown
1
492,596
14,216,432,889
IssuesEvent
2020-11-17 08:59:14
redhat-developer/vscode-openshift-tools
https://api.github.com/repos/redhat-developer/vscode-openshift-tools
closed
Create Service command support for creating services backed by operators
kind/feature priority/major resolution/duplicate upstream/odo
The command should support creating services based from service catalogs and based on operators using odo options explained in https://github.com/openshift/odo/issues/2613. https://odo.dev/docs/operator-hub/ Component to service should work when Service Binding Operator (SBO) installed.
1.0
Create Service command support for creating services backed by operators - The command should support creating services based from service catalogs and based on operators using odo options explained in https://github.com/openshift/odo/issues/2613. https://odo.dev/docs/operator-hub/ Component to service should work when Service Binding Operator (SBO) installed.
non_test
create service command support for creating services backed by operators the command should support creating services based from service catalogs and based on operators using odo options explained in component to service should work when service binding operator sbo installed
0
290,544
21,884,484,542
IssuesEvent
2022-05-19 17:10:48
a-little-org-called-mario/a-little-game-called-mario
https://api.github.com/repos/a-little-org-called-mario/a-little-game-called-mario
closed
Collision Layer Diligence and Proposal
documentation stale
A lot of the content that is being added, and that contains collision boxes, are all set to layer 1 mask 1. This will start to become a major issue moving forward. I already ran into a problem trying with commit #252 Here is what I propose: * We do a major PR that changes all current assets to no longer use layer 1 or mask 1. * Create a wiki page that explains the reasoning behind this choice and what they need to do to fix it. * (If this is possible) We add a check on commits that checks for `collision_mask` and `collision_layer` in `*.tscn` files to see if it is an odd number. * These numbers can only be odd if layer 1 or mask 1 is set. All the other options are even numbers. * If odd numbers are found, tell the contributor and link to the document explaining what needs to be done.
1.0
Collision Layer Diligence and Proposal - A lot of the content that is being added, and that contains collision boxes, are all set to layer 1 mask 1. This will start to become a major issue moving forward. I already ran into a problem trying with commit #252 Here is what I propose: * We do a major PR that changes all current assets to no longer use layer 1 or mask 1. * Create a wiki page that explains the reasoning behind this choice and what they need to do to fix it. * (If this is possible) We add a check on commits that checks for `collision_mask` and `collision_layer` in `*.tscn` files to see if it is an odd number. * These numbers can only be odd if layer 1 or mask 1 is set. All the other options are even numbers. * If odd numbers are found, tell the contributor and link to the document explaining what needs to be done.
non_test
collision layer diligence and proposal a lot of the content that is being added and that contains collision boxes are all set to layer mask this will start to become a major issue moving forward i already ran into a problem trying with commit here is what i propose we do a major pr that changes all current assets to no longer use layer or mask create a wiki page that explains the reasoning behind this choice and what they need to do to fix it if this is possible we add a check on commits that checks for collision mask and collision layer in tscn files to see if it is an odd number these numbers can only be odd if layer or mask is set all the other options are even numbers if odd numbers are found tell the contributor and link to the document explaining what needs to be done
0
159,127
12,464,095,497
IssuesEvent
2020-05-28 11:51:37
dasch-swiss/knora-app
https://api.github.com/repos/dasch-swiss/knora-app
closed
Homepage : external links are not easy to spot
styling user-testing
**Describe the bug** I find the link to external websites are not very easy to spot, unless you mouse over and look at the bottom left corner of the screen where the URL where the image or text redirects is displayed. **To Reproduce Steps to reproduce the behavior:** 1. Go to '...' 2. Click on '....' 3. Scroll down to '....' 4. See error **OPTIONAL: Expected behavior** Maybe underline when a text that is an hyperlink? **Screenshots** If applicable, add screenshots to help explain your problem (drag-and-drop the image directly here). **Desktop (please complete the following information):** - OS: macOS 10.14.6 (18G103), French - Browser Firefox - Version 70.01.1 **Additional context** Add any other context about the problem here.
1.0
Homepage : external links are not easy to spot - **Describe the bug** I find the link to external websites are not very easy to spot, unless you mouse over and look at the bottom left corner of the screen where the URL where the image or text redirects is displayed. **To Reproduce Steps to reproduce the behavior:** 1. Go to '...' 2. Click on '....' 3. Scroll down to '....' 4. See error **OPTIONAL: Expected behavior** Maybe underline when a text that is an hyperlink? **Screenshots** If applicable, add screenshots to help explain your problem (drag-and-drop the image directly here). **Desktop (please complete the following information):** - OS: macOS 10.14.6 (18G103), French - Browser Firefox - Version 70.01.1 **Additional context** Add any other context about the problem here.
test
homepage external links are not easy to spot describe the bug i find the link to external websites are not very easy to spot unless you mouse over and look at the bottom left corner of the screen where the url where the image or text redirects is displayed to reproduce steps to reproduce the behavior go to click on scroll down to see error optional expected behavior maybe underline when a text that is an hyperlink screenshots if applicable add screenshots to help explain your problem drag and drop the image directly here desktop please complete the following information os macos french browser firefox version additional context add any other context about the problem here
1
161,222
12,534,179,299
IssuesEvent
2020-06-04 18:57:05
astropy/astropy
https://api.github.com/repos/astropy/astropy
closed
tox -e codestyle fails on astropy/_erfa/core.py
Bug Docs testing
<!-- This comments are hidden when you submit the issue, so you do not need to remove them! --> <!-- Please be sure to check out our contributing guidelines, https://github.com/astropy/astropy/blob/master/CONTRIBUTING.md . Please be sure to check out our code of conduct, https://github.com/astropy/astropy/blob/master/CODE_OF_CONDUCT.md . --> <!-- Please have a search on our GitHub repository to see if a similar issue has already been posted. If a similar issue is closed, have a quick look to see if you are satisfied by the resolution. If not please go ahead and open an issue! --> <!-- Please check that the development version still produces the same bug. You can install development version with pip install git+https://github.com/astropy/astropy command. --> ### Description <!-- Provide a general description of the bug. --> `tox -e codestyle` always fails with a huge stream of code style violations in `astropy/_erfa/core.py`. This renders it largely useless for checking my own code for style violations. I assume the file is auto-generated, so simply telling the job to ignore code style violations in that file might solve the problem. If not, a once-over with `black` or `autopep8` would probably resolve the issue. ### Expected behavior <!-- What did you expect to happen. --> `tox -e codestyle` succeeds on a clean master. ### Actual behavior <!-- What actually happened. --> <!-- Was the output confusing or poorly described? --> ``` ... astropy/_erfa/core.py:20058:37: W291 trailing whitespace astropy/_erfa/core.py:20101:37: W291 trailing whitespace astropy/_erfa/core.py:20146:37: W291 trailing whitespace astropy/_erfa/core.py:20187:37: W291 trailing whitespace astropy/_erfa/core.py:20231:37: W291 trailing whitespace astropy/_erfa/core.py:20272:37: W291 trailing whitespace astropy/_erfa/core.py:20342:22: W292 no newline at end of file 244 ERROR: InvocationError for command /home/archibald/projects/astropy/astropy/.tox/codestyle/bin/flake8 astropy --count --select=E101,W191,W291,W292,W293,W391,E111,E112,E113,E30,E502,E722,E901,E902,E999,F822,F823 (exited with code 1) _______________________________________________________ summary _______________________________________________________ ERROR: codestyle: commands failed ``` ### Steps to Reproduce <!-- Ideally a code example could be provided so we can run it ourselves. --> <!-- If you are pasting code, use triple backticks (```) around your code snippet. --> <!-- If necessary, sanitize your screen output to be pasted so you do not reveal secrets like tokens and passwords. --> ```sh tox -e codestyle ``` ### System Details <!-- Even if you do not think this is necessary, it is useful information for the maintainers. Please run the following snippet and paste the output below: import platform; print(platform.platform()) import sys; print("Python", sys.version) import numpy; print("Numpy", numpy.__version__) import astropy; print("astropy", astropy.__version__) import scipy; print("Scipy", scipy.__version__) import matplotlib; print("Matplotlib", matplotlib.__version__) --> ``` >>> import platform; print(platform.platform()) Linux-5.4.0-33-generic-x86_64-with-glibc2.29 >>> import sys; print("Python", sys.version) Python 3.8.2 (default, Apr 27 2020, 15:53:34) [GCC 9.3.0] >>> import numpy; print("Numpy", numpy.__version__) Numpy 1.18.4 >>> import astropy; print("astropy", astropy.__version__) astropy 4.2.dev162+gd0a1e469f >>> import scipy; print("Scipy", scipy.__version__) Scipy 1.4.1 >>> import matplotlib; print("Matplotlib", matplotlib.__version__) Matplotlib 3.2.1 ```
1.0
tox -e codestyle fails on astropy/_erfa/core.py - <!-- This comments are hidden when you submit the issue, so you do not need to remove them! --> <!-- Please be sure to check out our contributing guidelines, https://github.com/astropy/astropy/blob/master/CONTRIBUTING.md . Please be sure to check out our code of conduct, https://github.com/astropy/astropy/blob/master/CODE_OF_CONDUCT.md . --> <!-- Please have a search on our GitHub repository to see if a similar issue has already been posted. If a similar issue is closed, have a quick look to see if you are satisfied by the resolution. If not please go ahead and open an issue! --> <!-- Please check that the development version still produces the same bug. You can install development version with pip install git+https://github.com/astropy/astropy command. --> ### Description <!-- Provide a general description of the bug. --> `tox -e codestyle` always fails with a huge stream of code style violations in `astropy/_erfa/core.py`. This renders it largely useless for checking my own code for style violations. I assume the file is auto-generated, so simply telling the job to ignore code style violations in that file might solve the problem. If not, a once-over with `black` or `autopep8` would probably resolve the issue. ### Expected behavior <!-- What did you expect to happen. --> `tox -e codestyle` succeeds on a clean master. ### Actual behavior <!-- What actually happened. --> <!-- Was the output confusing or poorly described? --> ``` ... astropy/_erfa/core.py:20058:37: W291 trailing whitespace astropy/_erfa/core.py:20101:37: W291 trailing whitespace astropy/_erfa/core.py:20146:37: W291 trailing whitespace astropy/_erfa/core.py:20187:37: W291 trailing whitespace astropy/_erfa/core.py:20231:37: W291 trailing whitespace astropy/_erfa/core.py:20272:37: W291 trailing whitespace astropy/_erfa/core.py:20342:22: W292 no newline at end of file 244 ERROR: InvocationError for command /home/archibald/projects/astropy/astropy/.tox/codestyle/bin/flake8 astropy --count --select=E101,W191,W291,W292,W293,W391,E111,E112,E113,E30,E502,E722,E901,E902,E999,F822,F823 (exited with code 1) _______________________________________________________ summary _______________________________________________________ ERROR: codestyle: commands failed ``` ### Steps to Reproduce <!-- Ideally a code example could be provided so we can run it ourselves. --> <!-- If you are pasting code, use triple backticks (```) around your code snippet. --> <!-- If necessary, sanitize your screen output to be pasted so you do not reveal secrets like tokens and passwords. --> ```sh tox -e codestyle ``` ### System Details <!-- Even if you do not think this is necessary, it is useful information for the maintainers. Please run the following snippet and paste the output below: import platform; print(platform.platform()) import sys; print("Python", sys.version) import numpy; print("Numpy", numpy.__version__) import astropy; print("astropy", astropy.__version__) import scipy; print("Scipy", scipy.__version__) import matplotlib; print("Matplotlib", matplotlib.__version__) --> ``` >>> import platform; print(platform.platform()) Linux-5.4.0-33-generic-x86_64-with-glibc2.29 >>> import sys; print("Python", sys.version) Python 3.8.2 (default, Apr 27 2020, 15:53:34) [GCC 9.3.0] >>> import numpy; print("Numpy", numpy.__version__) Numpy 1.18.4 >>> import astropy; print("astropy", astropy.__version__) astropy 4.2.dev162+gd0a1e469f >>> import scipy; print("Scipy", scipy.__version__) Scipy 1.4.1 >>> import matplotlib; print("Matplotlib", matplotlib.__version__) Matplotlib 3.2.1 ```
test
tox e codestyle fails on astropy erfa core py this comments are hidden when you submit the issue so you do not need to remove them please be sure to check out our contributing guidelines please be sure to check out our code of conduct please have a search on our github repository to see if a similar issue has already been posted if a similar issue is closed have a quick look to see if you are satisfied by the resolution if not please go ahead and open an issue please check that the development version still produces the same bug you can install development version with pip install git command description tox e codestyle always fails with a huge stream of code style violations in astropy erfa core py this renders it largely useless for checking my own code for style violations i assume the file is auto generated so simply telling the job to ignore code style violations in that file might solve the problem if not a once over with black or would probably resolve the issue expected behavior tox e codestyle succeeds on a clean master actual behavior astropy erfa core py trailing whitespace astropy erfa core py trailing whitespace astropy erfa core py trailing whitespace astropy erfa core py trailing whitespace astropy erfa core py trailing whitespace astropy erfa core py trailing whitespace astropy erfa core py no newline at end of file error invocationerror for command home archibald projects astropy astropy tox codestyle bin astropy count select exited with code summary error codestyle commands failed steps to reproduce if you are pasting code use triple backticks around your code snippet if necessary sanitize your screen output to be pasted so you do not reveal secrets like tokens and passwords sh tox e codestyle system details even if you do not think this is necessary it is useful information for the maintainers please run the following snippet and paste the output below import platform print platform platform import sys print python sys version import numpy print numpy numpy version import astropy print astropy astropy version import scipy print scipy scipy version import matplotlib print matplotlib matplotlib version import platform print platform platform linux generic with import sys print python sys version python default apr import numpy print numpy numpy version numpy import astropy print astropy astropy version astropy import scipy print scipy scipy version scipy import matplotlib print matplotlib matplotlib version matplotlib
1
183,425
14,230,121,603
IssuesEvent
2020-11-18 07:33:15
remirror/remirror
https://api.github.com/repos/remirror/remirror
opened
Reactivate e2e tests for SSR
status: experimental :test_tube: test: e2e :heavy_check_mark: type: feature :sparkles:
### Description Some issues with server-side rendering (**SSR**) have been raised in third-party codebases. This issue should track progress on reactivating end-to-end tests for SSR to ensure that all extensions contained within this repository work as expected.
2.0
Reactivate e2e tests for SSR - ### Description Some issues with server-side rendering (**SSR**) have been raised in third-party codebases. This issue should track progress on reactivating end-to-end tests for SSR to ensure that all extensions contained within this repository work as expected.
test
reactivate tests for ssr description some issues with server side rendering ssr have been raised in third party codebases this issue should track progress on reactivating end to end tests for ssr to ensure that all extensions contained within this repository work as expected
1
725,804
24,976,359,539
IssuesEvent
2022-11-02 08:11:04
Lightning-AI/lightning
https://api.github.com/repos/Lightning-AI/lightning
closed
Testing custom work dependencies example fails due to timeout in CI
ci priority: 1 app
### Bug description One of our app example tests fails due to timeout. - CI file: https://github.com/Lightning-AI/lightning/blob/c39c8eb2e4f49e35ef38a2ae6de2765ec71623db/.azure/app-cloud-e2e.yml#L66 - CI run: https://dev.azure.com/Lightning-AI/lightning/_build/results?buildId=111050&view=logs&j=4e1b52d3-134d-51bf-b10a-6dde7facebae&t=60ace6bb-5e20-5573-9db7-5550ec53d684&l=1463 ### How to reproduce the bug _No response_ ### Error messages and logs ``` ... flow 10/30/2022 17:05:43 DEBUG: The url http://10.1.113.254:47157/ is not ready. flow 10/30/2022 17:05:43 DEBUG: The url http://10.1.113.254:47157/ is not ready. flow 10/30/2022 17:05:43 DEBUG: The url http://10.1.113.254:47157/ is not ready. (node:446) MaxListenersExceededWarning: Possible EventEmitter memory leak detected. 11 end listeners added to [ReadStream]. Use emitter.setMaxListeners() to increase limit (Use `node --trace-warnings ...` to show where the warning was created) (node:446) MaxListenersExceededWarning: Possible EventEmitter memory leak detected. 11 error listeners added to [ReadStream]. Use emitter.setMaxListeners() to increase limit /bin/sh: 1: ***: not found FAILED =================================== FAILURES =================================== _________________ test_custom_work_dependencies_example_cloud __________________ @pytest.mark.cloud def test_custom_work_dependencies_example_cloud() -> None: # if requirements not installed, the app will fail with run_app_in_cloud( os.path.join(_PROJECT_ROOT, "tests/tests_app_examples/custom_work_dependencies/"), app_name="app.py", ) as (_, _, fetch_logs, _): has_logs = False while not has_logs: > for log in fetch_logs(["flow"]): tests/tests_app_examples/test_custom_work_dependencies.py:19: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ src/***/app/testing/testing.py:440: in fetch_logs for log_event in gen: src/***/app/utilities/app_logs.py:96: in _app_logs_reader log_event: _LogEvent = read_queue.get(timeout=None if follow else 1.0) /usr/lib/python3.8/queue.py:179: in get self.not_empty.wait(remaining) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <Condition(<unlocked _thread.lock object at 0x7fe4bb27f300>, 0)> timeout = 0.9999995000000581 def wait(self, timeout=None): """Wait until notified or until a timeout occurs. If the calling thread has not acquired the lock when this method is called, a RuntimeError is raised. This method releases the underlying lock, and then blocks until it is awakened by a notify() or notify_all() call for the same condition variable in another thread, or until the optional timeout occurs. Once awakened or timed out, it re-acquires the lock and returns. When the timeout argument is present and not None, it should be a floating point number specifying a timeout for the operation in seconds (or fractions thereof). When the underlying lock is an RLock, it is not released using its release() method, since this may not actually unlock the lock when it was acquired multiple times recursively. Instead, an internal interface of the RLock class is used, which really unlocks it even when it has been recursively acquired several times. Another internal interface is then used to restore the recursion level when the lock is reacquired. """ if not self._is_owned(): raise RuntimeError("cannot wait on un-acquired lock") waiter = _allocate_lock() waiter.acquire() self._waiters.append(waiter) saved_state = self._release_save() gotit = False try: # restore state no matter what (e.g., KeyboardInterrupt) if timeout is None: waiter.acquire() gotit = True else: if timeout > 0: > gotit = waiter.acquire(True, timeout) E Failed: Timeout >1200.0s /usr/lib/python3.8/threading.py:306: Failed ``` ``` =========================== short test summary info ============================ FAILED tests/tests_app_examples/test_custom_work_dependencies.py::test_custom_work_dependencies_example_cloud ======================== 1 failed in 1373.06s (0:22:53) ======================== +++++++++++++++++++++++++++++++++++ Timeout ++++++++++++++++++++++++++++++++++++ ~~~~~~~~~~~~~~~~~~~~ Stack of Thread-241 (140620377880320) ~~~~~~~~~~~~~~~~~~~~~ File "/usr/lib/python3.8/threading.py", line 890, in _bootstrap self._bootstrap_inner() File "/usr/lib/python3.8/threading.py", line 932, in _bootstrap_inner self.run() File "/usr/lib/python3.8/threading.py", line 870, in run self._target(*self._args, **self._kwargs) File "/home/AzDevOps_azpcontainer/.local/lib/python3.8/site-packages/websocket/_app.py", line 449, in run_forever setSock() File "/home/AzDevOps_azpcontainer/.local/lib/python3.8/site-packages/websocket/_app.py", line 378, in setSock dispatcher.read(self.sock.sock, read, check) File "/home/AzDevOps_azpcontainer/.local/lib/python3.8/site-packages/websocket/_app.py", line 72, in read r = sel.select(self.ping_timeout) File "/usr/lib/python3.8/selectors.py", line 468, in select fd_event_list = self._selector.poll(timeout, max_ev) ~~~~~~~~~~~~~~~~~~~~~ Stack of waitpid-0 (140621050640128) ~~~~~~~~~~~~~~~~~~~~~ File "/usr/lib/python3.8/threading.py", line 890, in _bootstrap self._bootstrap_inner() File "/usr/lib/python3.8/threading.py", line 932, in _bootstrap_inner self.run() File "/usr/lib/python3.8/threading.py", line 870, in run self._target(*self._args, **self._kwargs) File "/usr/lib/python3.8/asyncio/unix_events.py", line 1318, in _do_waitpid pid, status = os.waitpid(expected_pid, 0) +++++++++++++++++++++++++++++++++++ Timeout ++++++++++++++++++++++++++++++++++++ ##[error]Bash exited with code '1'. ``` ### Environment _No response_ ### More info _No response_ cc @tchaton @rohitgr7 @carmocca @akihironitta @borda
1.0
Testing custom work dependencies example fails due to timeout in CI - ### Bug description One of our app example tests fails due to timeout. - CI file: https://github.com/Lightning-AI/lightning/blob/c39c8eb2e4f49e35ef38a2ae6de2765ec71623db/.azure/app-cloud-e2e.yml#L66 - CI run: https://dev.azure.com/Lightning-AI/lightning/_build/results?buildId=111050&view=logs&j=4e1b52d3-134d-51bf-b10a-6dde7facebae&t=60ace6bb-5e20-5573-9db7-5550ec53d684&l=1463 ### How to reproduce the bug _No response_ ### Error messages and logs ``` ... flow 10/30/2022 17:05:43 DEBUG: The url http://10.1.113.254:47157/ is not ready. flow 10/30/2022 17:05:43 DEBUG: The url http://10.1.113.254:47157/ is not ready. flow 10/30/2022 17:05:43 DEBUG: The url http://10.1.113.254:47157/ is not ready. (node:446) MaxListenersExceededWarning: Possible EventEmitter memory leak detected. 11 end listeners added to [ReadStream]. Use emitter.setMaxListeners() to increase limit (Use `node --trace-warnings ...` to show where the warning was created) (node:446) MaxListenersExceededWarning: Possible EventEmitter memory leak detected. 11 error listeners added to [ReadStream]. Use emitter.setMaxListeners() to increase limit /bin/sh: 1: ***: not found FAILED =================================== FAILURES =================================== _________________ test_custom_work_dependencies_example_cloud __________________ @pytest.mark.cloud def test_custom_work_dependencies_example_cloud() -> None: # if requirements not installed, the app will fail with run_app_in_cloud( os.path.join(_PROJECT_ROOT, "tests/tests_app_examples/custom_work_dependencies/"), app_name="app.py", ) as (_, _, fetch_logs, _): has_logs = False while not has_logs: > for log in fetch_logs(["flow"]): tests/tests_app_examples/test_custom_work_dependencies.py:19: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ src/***/app/testing/testing.py:440: in fetch_logs for log_event in gen: src/***/app/utilities/app_logs.py:96: in _app_logs_reader log_event: _LogEvent = read_queue.get(timeout=None if follow else 1.0) /usr/lib/python3.8/queue.py:179: in get self.not_empty.wait(remaining) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <Condition(<unlocked _thread.lock object at 0x7fe4bb27f300>, 0)> timeout = 0.9999995000000581 def wait(self, timeout=None): """Wait until notified or until a timeout occurs. If the calling thread has not acquired the lock when this method is called, a RuntimeError is raised. This method releases the underlying lock, and then blocks until it is awakened by a notify() or notify_all() call for the same condition variable in another thread, or until the optional timeout occurs. Once awakened or timed out, it re-acquires the lock and returns. When the timeout argument is present and not None, it should be a floating point number specifying a timeout for the operation in seconds (or fractions thereof). When the underlying lock is an RLock, it is not released using its release() method, since this may not actually unlock the lock when it was acquired multiple times recursively. Instead, an internal interface of the RLock class is used, which really unlocks it even when it has been recursively acquired several times. Another internal interface is then used to restore the recursion level when the lock is reacquired. """ if not self._is_owned(): raise RuntimeError("cannot wait on un-acquired lock") waiter = _allocate_lock() waiter.acquire() self._waiters.append(waiter) saved_state = self._release_save() gotit = False try: # restore state no matter what (e.g., KeyboardInterrupt) if timeout is None: waiter.acquire() gotit = True else: if timeout > 0: > gotit = waiter.acquire(True, timeout) E Failed: Timeout >1200.0s /usr/lib/python3.8/threading.py:306: Failed ``` ``` =========================== short test summary info ============================ FAILED tests/tests_app_examples/test_custom_work_dependencies.py::test_custom_work_dependencies_example_cloud ======================== 1 failed in 1373.06s (0:22:53) ======================== +++++++++++++++++++++++++++++++++++ Timeout ++++++++++++++++++++++++++++++++++++ ~~~~~~~~~~~~~~~~~~~~ Stack of Thread-241 (140620377880320) ~~~~~~~~~~~~~~~~~~~~~ File "/usr/lib/python3.8/threading.py", line 890, in _bootstrap self._bootstrap_inner() File "/usr/lib/python3.8/threading.py", line 932, in _bootstrap_inner self.run() File "/usr/lib/python3.8/threading.py", line 870, in run self._target(*self._args, **self._kwargs) File "/home/AzDevOps_azpcontainer/.local/lib/python3.8/site-packages/websocket/_app.py", line 449, in run_forever setSock() File "/home/AzDevOps_azpcontainer/.local/lib/python3.8/site-packages/websocket/_app.py", line 378, in setSock dispatcher.read(self.sock.sock, read, check) File "/home/AzDevOps_azpcontainer/.local/lib/python3.8/site-packages/websocket/_app.py", line 72, in read r = sel.select(self.ping_timeout) File "/usr/lib/python3.8/selectors.py", line 468, in select fd_event_list = self._selector.poll(timeout, max_ev) ~~~~~~~~~~~~~~~~~~~~~ Stack of waitpid-0 (140621050640128) ~~~~~~~~~~~~~~~~~~~~~ File "/usr/lib/python3.8/threading.py", line 890, in _bootstrap self._bootstrap_inner() File "/usr/lib/python3.8/threading.py", line 932, in _bootstrap_inner self.run() File "/usr/lib/python3.8/threading.py", line 870, in run self._target(*self._args, **self._kwargs) File "/usr/lib/python3.8/asyncio/unix_events.py", line 1318, in _do_waitpid pid, status = os.waitpid(expected_pid, 0) +++++++++++++++++++++++++++++++++++ Timeout ++++++++++++++++++++++++++++++++++++ ##[error]Bash exited with code '1'. ``` ### Environment _No response_ ### More info _No response_ cc @tchaton @rohitgr7 @carmocca @akihironitta @borda
non_test
testing custom work dependencies example fails due to timeout in ci bug description one of our app example tests fails due to timeout ci file ci run how to reproduce the bug no response error messages and logs flow debug the url is not ready flow debug the url is not ready flow debug the url is not ready node maxlistenersexceededwarning possible eventemitter memory leak detected end listeners added to use emitter setmaxlisteners to increase limit use node trace warnings to show where the warning was created node maxlistenersexceededwarning possible eventemitter memory leak detected error listeners added to use emitter setmaxlisteners to increase limit bin sh not found failed failures test custom work dependencies example cloud pytest mark cloud def test custom work dependencies example cloud none if requirements not installed the app will fail with run app in cloud os path join project root tests tests app examples custom work dependencies app name app py as fetch logs has logs false while not has logs for log in fetch logs tests tests app examples test custom work dependencies py src app testing testing py in fetch logs for log event in gen src app utilities app logs py in app logs reader log event logevent read queue get timeout none if follow else usr lib queue py in get self not empty wait remaining self timeout def wait self timeout none wait until notified or until a timeout occurs if the calling thread has not acquired the lock when this method is called a runtimeerror is raised this method releases the underlying lock and then blocks until it is awakened by a notify or notify all call for the same condition variable in another thread or until the optional timeout occurs once awakened or timed out it re acquires the lock and returns when the timeout argument is present and not none it should be a floating point number specifying a timeout for the operation in seconds or fractions thereof when the underlying lock is an rlock it is not released using its release method since this may not actually unlock the lock when it was acquired multiple times recursively instead an internal interface of the rlock class is used which really unlocks it even when it has been recursively acquired several times another internal interface is then used to restore the recursion level when the lock is reacquired if not self is owned raise runtimeerror cannot wait on un acquired lock waiter allocate lock waiter acquire self waiters append waiter saved state self release save gotit false try restore state no matter what e g keyboardinterrupt if timeout is none waiter acquire gotit true else if timeout gotit waiter acquire true timeout e failed timeout usr lib threading py failed short test summary info failed tests tests app examples test custom work dependencies py test custom work dependencies example cloud failed in timeout stack of thread file usr lib threading py line in bootstrap self bootstrap inner file usr lib threading py line in bootstrap inner self run file usr lib threading py line in run self target self args self kwargs file home azdevops azpcontainer local lib site packages websocket app py line in run forever setsock file home azdevops azpcontainer local lib site packages websocket app py line in setsock dispatcher read self sock sock read check file home azdevops azpcontainer local lib site packages websocket app py line in read r sel select self ping timeout file usr lib selectors py line in select fd event list self selector poll timeout max ev stack of waitpid file usr lib threading py line in bootstrap self bootstrap inner file usr lib threading py line in bootstrap inner self run file usr lib threading py line in run self target self args self kwargs file usr lib asyncio unix events py line in do waitpid pid status os waitpid expected pid timeout bash exited with code environment no response more info no response cc tchaton carmocca akihironitta borda
0
102,875
8,868,648,076
IssuesEvent
2019-01-11 00:59:19
rancher/rancher
https://api.github.com/repos/rancher/rancher
closed
AD - Group changes not reflected in user attributes even after auth-user-info-max-age-seconds passed or forcing rancher login using cli
area/authentication status/reopened status/resolved status/to-test team/az version/2.0
Original issue: https://github.com/rancherlabs/rancher-security/issues/40 Original text: Rancher server version - build from master Steps to reproduce the problem: AD auth enabled setup Create a cluster with 3 projects: p1 - member access given to "group1" p2 - member access given to "group2" p3 - member access given to "group3" Log in to rancher UI as "user1" who is part of "group1" and "group2" Create API keys. Set "auth-user-info-max-age-seconds" to 120 seconds. Use the above created keys to log in using rancher-cli. User is provided access to both projects p1 and p2 as expected (./rancher project) From AD console, remove p1 from being part of "group2". Wait for more than > "auth-user-info-max-age-seconds" to pass. scenario 1: In the current logged in rancher-cli session , execute ```./rancher project``` . Both projects p1 and p2 are listed. scenario 2: Log into rancher-cli using the same token , execute ```./rancher project``` . Both projects p1 and p2 are listed. In this case , following error logs seen in rancher-server logs: ``` 2018/12/14 01:22:48 [DEBUG] REST UPDATE apis/management.cattle.io/v3//userattributes/u-66zjeognlk 2018/12/14 01:22:48 [DEBUG] UserAttributeController calling handler mgmt-auth-userattributes-controller u-66zjeognlk 2018/12/14 01:22:48 [DEBUG] REST GET apis/management.cattle.io/v3//authconfigs/activedirectory 2018/12/14 01:22:48 [DEBUG] REST GET api//v1/mgmt-secrets/secrets/activedirectoryconfig-serviceaccountpassword 2018/12/14 01:22:48 [DEBUG] Now creating Ldap connection 2018/12/14 01:22:48 [DEBUG] Binding service account username password 2018/12/14 01:22:48 [DEBUG] LDAP Search query: {(sAMAccountName=activedirectory_user://CN=testguser1,CN=Users,DC=testad,DC=rancher,DC=io)} 2018/12/14 01:22:48 [ERROR] Error refreshing token principals, skipping: Cannot locate user information for (sAMAccountName=activedirectory_user://CN=testguser1,CN=Users,DC=testad,DC=rancher,DC=io) 2018/12/14 01:22:48 [DEBUG] REST UPDATE apis/management.cattle.io/v3//userattributes/u-66zjeognlk 2018/12/14 01:22:48 [DEBUG] UserAttributeController calling handler mgmt-auth-userattributes-controller u-66zjeognlk ``` In both scenario1 and scenario2 , we expect user to have access to only project p1 since they are not part of "group2" anymore. Also the user-attributes does not reflect the group change (```kubectl get user attribute <user-id> -o json```) Note: When "user1" logs in using the UI , I see the user-attributes reflect the group changes as expected. Further log ins done from rancher-cli using the older token shows only project p1.
1.0
AD - Group changes not reflected in user attributes even after auth-user-info-max-age-seconds passed or forcing rancher login using cli - Original issue: https://github.com/rancherlabs/rancher-security/issues/40 Original text: Rancher server version - build from master Steps to reproduce the problem: AD auth enabled setup Create a cluster with 3 projects: p1 - member access given to "group1" p2 - member access given to "group2" p3 - member access given to "group3" Log in to rancher UI as "user1" who is part of "group1" and "group2" Create API keys. Set "auth-user-info-max-age-seconds" to 120 seconds. Use the above created keys to log in using rancher-cli. User is provided access to both projects p1 and p2 as expected (./rancher project) From AD console, remove p1 from being part of "group2". Wait for more than > "auth-user-info-max-age-seconds" to pass. scenario 1: In the current logged in rancher-cli session , execute ```./rancher project``` . Both projects p1 and p2 are listed. scenario 2: Log into rancher-cli using the same token , execute ```./rancher project``` . Both projects p1 and p2 are listed. In this case , following error logs seen in rancher-server logs: ``` 2018/12/14 01:22:48 [DEBUG] REST UPDATE apis/management.cattle.io/v3//userattributes/u-66zjeognlk 2018/12/14 01:22:48 [DEBUG] UserAttributeController calling handler mgmt-auth-userattributes-controller u-66zjeognlk 2018/12/14 01:22:48 [DEBUG] REST GET apis/management.cattle.io/v3//authconfigs/activedirectory 2018/12/14 01:22:48 [DEBUG] REST GET api//v1/mgmt-secrets/secrets/activedirectoryconfig-serviceaccountpassword 2018/12/14 01:22:48 [DEBUG] Now creating Ldap connection 2018/12/14 01:22:48 [DEBUG] Binding service account username password 2018/12/14 01:22:48 [DEBUG] LDAP Search query: {(sAMAccountName=activedirectory_user://CN=testguser1,CN=Users,DC=testad,DC=rancher,DC=io)} 2018/12/14 01:22:48 [ERROR] Error refreshing token principals, skipping: Cannot locate user information for (sAMAccountName=activedirectory_user://CN=testguser1,CN=Users,DC=testad,DC=rancher,DC=io) 2018/12/14 01:22:48 [DEBUG] REST UPDATE apis/management.cattle.io/v3//userattributes/u-66zjeognlk 2018/12/14 01:22:48 [DEBUG] UserAttributeController calling handler mgmt-auth-userattributes-controller u-66zjeognlk ``` In both scenario1 and scenario2 , we expect user to have access to only project p1 since they are not part of "group2" anymore. Also the user-attributes does not reflect the group change (```kubectl get user attribute <user-id> -o json```) Note: When "user1" logs in using the UI , I see the user-attributes reflect the group changes as expected. Further log ins done from rancher-cli using the older token shows only project p1.
test
ad group changes not reflected in user attributes even after auth user info max age seconds passed or forcing rancher login using cli original issue original text rancher server version build from master steps to reproduce the problem ad auth enabled setup create a cluster with projects member access given to member access given to member access given to log in to rancher ui as who is part of and create api keys set auth user info max age seconds to seconds use the above created keys to log in using rancher cli user is provided access to both projects and as expected rancher project from ad console remove from being part of wait for more than auth user info max age seconds to pass scenario in the current logged in rancher cli session execute rancher project both projects and are listed scenario log into rancher cli using the same token execute rancher project both projects and are listed in this case following error logs seen in rancher server logs rest update apis management cattle io userattributes u userattributecontroller calling handler mgmt auth userattributes controller u rest get apis management cattle io authconfigs activedirectory rest get api mgmt secrets secrets activedirectoryconfig serviceaccountpassword now creating ldap connection binding service account username password ldap search query samaccountname activedirectory user cn cn users dc testad dc rancher dc io error refreshing token principals skipping cannot locate user information for samaccountname activedirectory user cn cn users dc testad dc rancher dc io rest update apis management cattle io userattributes u userattributecontroller calling handler mgmt auth userattributes controller u in both and we expect user to have access to only project since they are not part of anymore also the user attributes does not reflect the group change kubectl get user attribute o json note when logs in using the ui i see the user attributes reflect the group changes as expected further log ins done from rancher cli using the older token shows only project
1
345,688
30,833,304,188
IssuesEvent
2023-08-02 04:49:35
teamhanko/hanko
https://api.github.com/repos/teamhanko/hanko
closed
Make e2e test check for component ids instead of button texts etc.
javascript tests stale
Changing e.g. a button text in `hanko-auth` currently breaks e2e tests because the tests are based on strings ("see if button with text 'Sign in with a passkey" appears"). Wouldn't it be better to set ids for all relevant components in the element and make the tests check for the ids?
1.0
Make e2e test check for component ids instead of button texts etc. - Changing e.g. a button text in `hanko-auth` currently breaks e2e tests because the tests are based on strings ("see if button with text 'Sign in with a passkey" appears"). Wouldn't it be better to set ids for all relevant components in the element and make the tests check for the ids?
test
make test check for component ids instead of button texts etc changing e g a button text in hanko auth currently breaks tests because the tests are based on strings see if button with text sign in with a passkey appears wouldn t it be better to set ids for all relevant components in the element and make the tests check for the ids
1
222,736
17,470,661,078
IssuesEvent
2021-08-07 04:17:27
dotnet/aspnetcore
https://api.github.com/repos/dotnet/aspnetcore
opened
Microsoft.AspNetCore.Hosting.Tests.HostingApplicationTests.IHttpActivityFeatureIsPopulated
test-failure
## Failing Test(s) - Microsoft.AspNetCore.Hosting.Tests.HostingApplicationTests.IHttpActivityFeatureIsPopulated ## Error Message <!-- Provide the error message associated with the test failure, if applicable. --> ```text Assert.NotEqual() Failure Expected: Not Activity { ActivityTraceFlags = None, Baggage = [], Context = System.Diagnostics.ActivityContext, DisplayName = "Microsoft.AspNetCore.Hosting.HttpRequestIn", Duration = 00:00:00, ... } Actual: Activity { ActivityTraceFlags = None, Baggage = [], Context = System.Diagnostics.ActivityContext, DisplayName = "Microsoft.AspNetCore.Hosting.HttpRequestIn", Duration = 00:00:00, ... } ``` ## Stacktrace <details> <!-- Provide the stack trace associated with the test failure, if applicable. --> ```text at Microsoft.AspNetCore.Hosting.Tests.HostingApplicationTests.IHttpActivityFeatureIsPopulated() in /_/src/Hosting/Hosting/test/HostingApplicationTests.cs:line 122 ``` </details> ## Logs <details> <!-- Provide the (helix) logs associated with the test failure, if applicable. --> ```text [xUnit.net 00:00:01.21] Microsoft.AspNetCore.Hosting.Tests.HostingApplicationTests.IHttpActivityFeatureIsPopulated [FAIL] [xUnit.net 00:00:01.21] Assert.NotEqual() Failure [xUnit.net 00:00:01.21] Expected: Not Activity { ActivityTraceFlags = None, Baggage = [], Context = System.Diagnostics.ActivityContext, DisplayName = "Microsoft.AspNetCore.Hosting.HttpRequestIn", Duration = 00:00:00, ... } [xUnit.net 00:00:01.21] Actual: Activity { ActivityTraceFlags = None, Baggage = [], Context = System.Diagnostics.ActivityContext, DisplayName = "Microsoft.AspNetCore.Hosting.HttpRequestIn", Duration = 00:00:00, ... } [xUnit.net 00:00:01.21] Stack Trace: [xUnit.net 00:00:01.21] /_/src/Hosting/Hosting/test/HostingApplicationTests.cs(122,0): at Microsoft.AspNetCore.Hosting.Tests.HostingApplicationTests.IHttpActivityFeatureIsPopulated() ``` </details> ## Build <!-- Provide a link to the build where the test failure occurred. --> https://dev.azure.com/dnceng/public/_build/results?buildId=1280595&view=ms.vss-test-web.build-test-results-tab&runId=37844528&resultId=103322&paneView=debug
1.0
Microsoft.AspNetCore.Hosting.Tests.HostingApplicationTests.IHttpActivityFeatureIsPopulated - ## Failing Test(s) - Microsoft.AspNetCore.Hosting.Tests.HostingApplicationTests.IHttpActivityFeatureIsPopulated ## Error Message <!-- Provide the error message associated with the test failure, if applicable. --> ```text Assert.NotEqual() Failure Expected: Not Activity { ActivityTraceFlags = None, Baggage = [], Context = System.Diagnostics.ActivityContext, DisplayName = "Microsoft.AspNetCore.Hosting.HttpRequestIn", Duration = 00:00:00, ... } Actual: Activity { ActivityTraceFlags = None, Baggage = [], Context = System.Diagnostics.ActivityContext, DisplayName = "Microsoft.AspNetCore.Hosting.HttpRequestIn", Duration = 00:00:00, ... } ``` ## Stacktrace <details> <!-- Provide the stack trace associated with the test failure, if applicable. --> ```text at Microsoft.AspNetCore.Hosting.Tests.HostingApplicationTests.IHttpActivityFeatureIsPopulated() in /_/src/Hosting/Hosting/test/HostingApplicationTests.cs:line 122 ``` </details> ## Logs <details> <!-- Provide the (helix) logs associated with the test failure, if applicable. --> ```text [xUnit.net 00:00:01.21] Microsoft.AspNetCore.Hosting.Tests.HostingApplicationTests.IHttpActivityFeatureIsPopulated [FAIL] [xUnit.net 00:00:01.21] Assert.NotEqual() Failure [xUnit.net 00:00:01.21] Expected: Not Activity { ActivityTraceFlags = None, Baggage = [], Context = System.Diagnostics.ActivityContext, DisplayName = "Microsoft.AspNetCore.Hosting.HttpRequestIn", Duration = 00:00:00, ... } [xUnit.net 00:00:01.21] Actual: Activity { ActivityTraceFlags = None, Baggage = [], Context = System.Diagnostics.ActivityContext, DisplayName = "Microsoft.AspNetCore.Hosting.HttpRequestIn", Duration = 00:00:00, ... } [xUnit.net 00:00:01.21] Stack Trace: [xUnit.net 00:00:01.21] /_/src/Hosting/Hosting/test/HostingApplicationTests.cs(122,0): at Microsoft.AspNetCore.Hosting.Tests.HostingApplicationTests.IHttpActivityFeatureIsPopulated() ``` </details> ## Build <!-- Provide a link to the build where the test failure occurred. --> https://dev.azure.com/dnceng/public/_build/results?buildId=1280595&view=ms.vss-test-web.build-test-results-tab&runId=37844528&resultId=103322&paneView=debug
test
microsoft aspnetcore hosting tests hostingapplicationtests ihttpactivityfeatureispopulated failing test s microsoft aspnetcore hosting tests hostingapplicationtests ihttpactivityfeatureispopulated error message provide the error message associated with the test failure if applicable text assert notequal failure expected not activity activitytraceflags none baggage context system diagnostics activitycontext displayname microsoft aspnetcore hosting httprequestin duration actual activity activitytraceflags none baggage context system diagnostics activitycontext displayname microsoft aspnetcore hosting httprequestin duration stacktrace provide the stack trace associated with the test failure if applicable text at microsoft aspnetcore hosting tests hostingapplicationtests ihttpactivityfeatureispopulated in src hosting hosting test hostingapplicationtests cs line logs provide the helix logs associated with the test failure if applicable text microsoft aspnetcore hosting tests hostingapplicationtests ihttpactivityfeatureispopulated assert notequal failure expected not activity activitytraceflags none baggage context system diagnostics activitycontext displayname microsoft aspnetcore hosting httprequestin duration actual activity activitytraceflags none baggage context system diagnostics activitycontext displayname microsoft aspnetcore hosting httprequestin duration stack trace src hosting hosting test hostingapplicationtests cs at microsoft aspnetcore hosting tests hostingapplicationtests ihttpactivityfeatureispopulated build provide a link to the build where the test failure occurred
1
81,089
23,386,144,926
IssuesEvent
2022-08-11 13:54:54
xamarin/xamarin-android
https://api.github.com/repos/xamarin/xamarin-android
opened
BroadcastReceiver with targetSdkVersion="31" causes java.exe exited with code 1
Area: App+Library Build needs-triage
### Android application type Classic Xamarin.Android (MonoAndroid12.0, etc.) ### Affected platform version VS 2022 ### Description Placing this class in a Xamarin.Android app that targets sdk 31 (Android 12) and compiles with Android 12 causes a build error java.exe exited with code 1 ` [BroadcastReceiver] [IntentFilter(new[] { Android.Content.Intent.ActionBatteryChanged })] public class MyReceiver : BroadcastReceiver { public override void OnReceive(Context context, Intent intent) { } } ` Including example project [XamApp1.zip](https://github.com/xamarin/xamarin-android/files/9309213/XamApp1.zip) ### Steps to Reproduce 1. Create standard Xamarin.Android app using VS2022 2. Add the Broadcast reciever 3. Make sure your android:targetSdkVersion="31" in AndroidManifest.xml 4. Make sure the Compile Using Android Version: (Target Framework) setting is set to Android 12 (S) 5. Compile Debug->AnyCPU ### Did you find any workaround? Target Android 11 or below ### Relevant log output _No response_
1.0
BroadcastReceiver with targetSdkVersion="31" causes java.exe exited with code 1 - ### Android application type Classic Xamarin.Android (MonoAndroid12.0, etc.) ### Affected platform version VS 2022 ### Description Placing this class in a Xamarin.Android app that targets sdk 31 (Android 12) and compiles with Android 12 causes a build error java.exe exited with code 1 ` [BroadcastReceiver] [IntentFilter(new[] { Android.Content.Intent.ActionBatteryChanged })] public class MyReceiver : BroadcastReceiver { public override void OnReceive(Context context, Intent intent) { } } ` Including example project [XamApp1.zip](https://github.com/xamarin/xamarin-android/files/9309213/XamApp1.zip) ### Steps to Reproduce 1. Create standard Xamarin.Android app using VS2022 2. Add the Broadcast reciever 3. Make sure your android:targetSdkVersion="31" in AndroidManifest.xml 4. Make sure the Compile Using Android Version: (Target Framework) setting is set to Android 12 (S) 5. Compile Debug->AnyCPU ### Did you find any workaround? Target Android 11 or below ### Relevant log output _No response_
non_test
broadcastreceiver with targetsdkversion causes java exe exited with code android application type classic xamarin android etc affected platform version vs description placing this class in a xamarin android app that targets sdk android and compiles with android causes a build error java exe exited with code android content intent actionbatterychanged public class myreceiver broadcastreceiver public override void onreceive context context intent intent including example project steps to reproduce create standard xamarin android app using add the broadcast reciever make sure your android targetsdkversion in androidmanifest xml make sure the compile using android version target framework setting is set to android s compile debug anycpu did you find any workaround target android or below relevant log output no response
0
177,042
13,675,858,965
IssuesEvent
2020-09-29 13:13:24
brave/brave-browser
https://api.github.com/repos/brave/brave-browser
closed
[Desktop] Sync v2 - Apps are not automatically installed
OS/Desktop QA/Test-Plan-Specified QA/Yes feature/sync priority/P4
App is synced but the App is not installed According to @darkdh this is an upstream bug ## Steps to Reproduce <!--Please add a series of steps to reproduce the issue--> 1. Clean install on `DeviceA` and `DeviceB` 1. Establish sync chain between `DeviceA` and `DeviceB` 1. Enable Apps sync on `DeviceA` and `DeviceB` 1. `DeviceA`: Open interia.pl 1. `DeviceA`: Click `Install Interia` in the URL field ![image](https://user-images.githubusercontent.com/34715963/89920328-8d794d00-dbfc-11ea-8517-d82a1102acdc.png) 1. Wait for sync update 1. `DeviceB`: Open interia.pl ## Actual result: <!--Please add screenshots if needed--> `DeviceB`: App is synced ![image](https://user-images.githubusercontent.com/34715963/90018085-b6591b00-dcac-11ea-88b3-4c29b6bd9880.png) `DeviceB` : App is not installed ![image](https://user-images.githubusercontent.com/34715963/90018248-ebfe0400-dcac-11ea-8d0d-0fdcff279e14.png) ## Expected result: `DeviceB` : App is installed ![image](https://user-images.githubusercontent.com/34715963/90018315-02a45b00-dcad-11ea-8bae-93d5b52284d4.png) ## Reproduces how often: <!--[Easily reproduced/Intermittent issue/No steps to reproduce]--> 100% repro rate ## Brave version (brave://version info) <!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details--> Brave | 1.12.108 Chromium: 84.0.4147.105 (Official Build) (64-bit) -- | -- Revision | a6b12dfad6663f13a7e16e9a42a6a4975374096b-refs/branch-heads/4147@{#943} OS | Ubuntu 18.04 LTS cc @brave/legacy_qa @darkdh @jsecretan @AlexeyBarabash @rebron
1.0
[Desktop] Sync v2 - Apps are not automatically installed - App is synced but the App is not installed According to @darkdh this is an upstream bug ## Steps to Reproduce <!--Please add a series of steps to reproduce the issue--> 1. Clean install on `DeviceA` and `DeviceB` 1. Establish sync chain between `DeviceA` and `DeviceB` 1. Enable Apps sync on `DeviceA` and `DeviceB` 1. `DeviceA`: Open interia.pl 1. `DeviceA`: Click `Install Interia` in the URL field ![image](https://user-images.githubusercontent.com/34715963/89920328-8d794d00-dbfc-11ea-8517-d82a1102acdc.png) 1. Wait for sync update 1. `DeviceB`: Open interia.pl ## Actual result: <!--Please add screenshots if needed--> `DeviceB`: App is synced ![image](https://user-images.githubusercontent.com/34715963/90018085-b6591b00-dcac-11ea-88b3-4c29b6bd9880.png) `DeviceB` : App is not installed ![image](https://user-images.githubusercontent.com/34715963/90018248-ebfe0400-dcac-11ea-8d0d-0fdcff279e14.png) ## Expected result: `DeviceB` : App is installed ![image](https://user-images.githubusercontent.com/34715963/90018315-02a45b00-dcad-11ea-8bae-93d5b52284d4.png) ## Reproduces how often: <!--[Easily reproduced/Intermittent issue/No steps to reproduce]--> 100% repro rate ## Brave version (brave://version info) <!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details--> Brave | 1.12.108 Chromium: 84.0.4147.105 (Official Build) (64-bit) -- | -- Revision | a6b12dfad6663f13a7e16e9a42a6a4975374096b-refs/branch-heads/4147@{#943} OS | Ubuntu 18.04 LTS cc @brave/legacy_qa @darkdh @jsecretan @AlexeyBarabash @rebron
test
sync apps are not automatically installed app is synced but the app is not installed according to darkdh this is an upstream bug steps to reproduce clean install on devicea and deviceb establish sync chain between devicea and deviceb enable apps sync on devicea and deviceb devicea open interia pl devicea click install interia in the url field wait for sync update deviceb open interia pl actual result deviceb app is synced deviceb app is not installed expected result deviceb app is installed reproduces how often repro rate brave version brave version info brave chromium   official build   bit revision refs branch heads os ubuntu lts cc brave legacy qa darkdh jsecretan alexeybarabash rebron
1
147,274
23,192,583,755
IssuesEvent
2022-08-01 13:50:32
team-ribbon/ribborn-frontend
https://api.github.com/repos/team-ribbon/ribborn-frontend
closed
[공유] 공유하기 위치 조정이 필요합니다
design
<img src="https://user-images.githubusercontent.com/105181604/181536396-92705f5c-5f60-40c5-bd3c-505351fd9854.jpg" /> PC > 공유하기 클릭 시 좌우 스크롤 생기는 문제 모바일 > 공유하기 클릭하면 플로팅 버튼 위치 이동되는 문제 => 모바일에서는 공유하기 누르면 좌우로가 아니라, 위아래로 뜨면 좋을 것 같습니다.
1.0
[공유] 공유하기 위치 조정이 필요합니다 - <img src="https://user-images.githubusercontent.com/105181604/181536396-92705f5c-5f60-40c5-bd3c-505351fd9854.jpg" /> PC > 공유하기 클릭 시 좌우 스크롤 생기는 문제 모바일 > 공유하기 클릭하면 플로팅 버튼 위치 이동되는 문제 => 모바일에서는 공유하기 누르면 좌우로가 아니라, 위아래로 뜨면 좋을 것 같습니다.
non_test
공유하기 위치 조정이 필요합니다 pc 공유하기 클릭 시 좌우 스크롤 생기는 문제 모바일 공유하기 클릭하면 플로팅 버튼 위치 이동되는 문제 모바일에서는 공유하기 누르면 좌우로가 아니라 위아래로 뜨면 좋을 것 같습니다
0
221,986
17,380,614,979
IssuesEvent
2021-07-31 16:26:59
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
roachtest: clearrange/zfs/checks=true failed
C-test-failure O-roachtest O-robot branch-master release-blocker
roachtest.clearrange/zfs/checks=true [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=3246869&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=3246869&tab=artifacts#/clearrange/zfs/checks=true) on master @ [701b177d8f4b81d8654dfb4090a2cd3cf82e63a7](https://github.com/cockroachdb/cockroach/commits/701b177d8f4b81d8654dfb4090a2cd3cf82e63a7): ``` The test failed on branch=master, cloud=gce: test timed out (see artifacts for details) ``` <details><summary>Reproduce</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/roachtest) </p> </details> /cc @cockroachdb/storage <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*clearrange/zfs/checks=true.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
2.0
roachtest: clearrange/zfs/checks=true failed - roachtest.clearrange/zfs/checks=true [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=3246869&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=3246869&tab=artifacts#/clearrange/zfs/checks=true) on master @ [701b177d8f4b81d8654dfb4090a2cd3cf82e63a7](https://github.com/cockroachdb/cockroach/commits/701b177d8f4b81d8654dfb4090a2cd3cf82e63a7): ``` The test failed on branch=master, cloud=gce: test timed out (see artifacts for details) ``` <details><summary>Reproduce</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/roachtest) </p> </details> /cc @cockroachdb/storage <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*clearrange/zfs/checks=true.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
test
roachtest clearrange zfs checks true failed roachtest clearrange zfs checks true with on master the test failed on branch master cloud gce test timed out see artifacts for details reproduce see cc cockroachdb storage
1
18,372
10,948,170,205
IssuesEvent
2019-11-26 08:17:08
microsoft/botbuilder-dotnet
https://api.github.com/repos/microsoft/botbuilder-dotnet
closed
TurnContext.SendActivitiesAsync changes the activities' ID
Bot Services R7 bug customer-replied-to customer-reported
## Version master ## Describe the bug On [this line](https://github.com/microsoft/botbuilder-dotnet/blob/master/libraries/Microsoft.Bot.Builder/TurnContext.cs#L274), we can see where the turn context modifies the ID of each outgoing activity: ```c# activity.Id = responses[index].Id; ``` While outgoing activities are necessarily modified when a conversation reference is applied, modifying the ID based on the resource response creates strange behavior when the activity is saved and used again, like when it's a retry prompt. The Emulator for example will not display an activity if the ID is the same as an activity already in the conversation history. Several customers are running into issues because of this: https://github.com/microsoft/BotFramework-Emulator/issues/1684 https://stackoverflow.com/questions/58082537/why-does-the-retry-prompt-only-show-once The turn context in the Node SDK does not modify an outgoing activity's ID. ## Expected behavior There is reason to think that simply sending an activity shouldn't modify it at all, even by applying the conversation reference. That is to say, the necessary modifications should be done on a *copy* of the activity that gets passed to `SendActivityAsync` for example. However, changing this now could be breaking behavior if some bots expect the activity to be modified. Also, an easy workaround is to send an activity through the adapter or the connector, so bot developers have a choice about whether they want outgoing activities to be modified automatically. The ID in particular may be a special case for a few reasons. Not only is this modification causing the issues we see in Emulator, but the behavior is inconsistent between repos. Even if we don't want to change this ID-modifying behavior in the .NET SDK, we should still make sure the different SDK's work the same way. [bug]
1.0
TurnContext.SendActivitiesAsync changes the activities' ID - ## Version master ## Describe the bug On [this line](https://github.com/microsoft/botbuilder-dotnet/blob/master/libraries/Microsoft.Bot.Builder/TurnContext.cs#L274), we can see where the turn context modifies the ID of each outgoing activity: ```c# activity.Id = responses[index].Id; ``` While outgoing activities are necessarily modified when a conversation reference is applied, modifying the ID based on the resource response creates strange behavior when the activity is saved and used again, like when it's a retry prompt. The Emulator for example will not display an activity if the ID is the same as an activity already in the conversation history. Several customers are running into issues because of this: https://github.com/microsoft/BotFramework-Emulator/issues/1684 https://stackoverflow.com/questions/58082537/why-does-the-retry-prompt-only-show-once The turn context in the Node SDK does not modify an outgoing activity's ID. ## Expected behavior There is reason to think that simply sending an activity shouldn't modify it at all, even by applying the conversation reference. That is to say, the necessary modifications should be done on a *copy* of the activity that gets passed to `SendActivityAsync` for example. However, changing this now could be breaking behavior if some bots expect the activity to be modified. Also, an easy workaround is to send an activity through the adapter or the connector, so bot developers have a choice about whether they want outgoing activities to be modified automatically. The ID in particular may be a special case for a few reasons. Not only is this modification causing the issues we see in Emulator, but the behavior is inconsistent between repos. Even if we don't want to change this ID-modifying behavior in the .NET SDK, we should still make sure the different SDK's work the same way. [bug]
non_test
turncontext sendactivitiesasync changes the activities id version master describe the bug on we can see where the turn context modifies the id of each outgoing activity c activity id responses id while outgoing activities are necessarily modified when a conversation reference is applied modifying the id based on the resource response creates strange behavior when the activity is saved and used again like when it s a retry prompt the emulator for example will not display an activity if the id is the same as an activity already in the conversation history several customers are running into issues because of this the turn context in the node sdk does not modify an outgoing activity s id expected behavior there is reason to think that simply sending an activity shouldn t modify it at all even by applying the conversation reference that is to say the necessary modifications should be done on a copy of the activity that gets passed to sendactivityasync for example however changing this now could be breaking behavior if some bots expect the activity to be modified also an easy workaround is to send an activity through the adapter or the connector so bot developers have a choice about whether they want outgoing activities to be modified automatically the id in particular may be a special case for a few reasons not only is this modification causing the issues we see in emulator but the behavior is inconsistent between repos even if we don t want to change this id modifying behavior in the net sdk we should still make sure the different sdk s work the same way
0
68,538
21,699,680,193
IssuesEvent
2022-05-10 01:39:59
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
"your settings" in the search initialisation fail message must be styled as a link
T-Defect S-Tolerable Help Wanted A-Timeline-Search A-Indexing O-Occasional good first issue
![image](https://user-images.githubusercontent.com/14617699/126571507-c5e65e8e-e905-477b-8443-f07a5e63bec6.png) "your settings" must be styled as a link. Right now it's highlighten only on mouse over. That link should open the user setting, jump to the "Security & Privacy" tab and scroll down to the "Message search" section element-desktop 1.7.33 debian sid
1.0
"your settings" in the search initialisation fail message must be styled as a link - ![image](https://user-images.githubusercontent.com/14617699/126571507-c5e65e8e-e905-477b-8443-f07a5e63bec6.png) "your settings" must be styled as a link. Right now it's highlighten only on mouse over. That link should open the user setting, jump to the "Security & Privacy" tab and scroll down to the "Message search" section element-desktop 1.7.33 debian sid
non_test
your settings in the search initialisation fail message must be styled as a link your settings must be styled as a link right now it s highlighten only on mouse over that link should open the user setting jump to the security privacy tab and scroll down to the message search section element desktop debian sid
0
229,382
18,294,436,303
IssuesEvent
2021-10-05 18:53:31
elastic/kibana
https://api.github.com/repos/elastic/kibana
opened
Failing test: Jest Tests.x-pack/plugins/fleet/storybook - Storyshots Sections/EPM/Package List Grid List
failed-test
A test failed on a tracked branch ``` Error: expect(received).not.toContain(expected) // indexOf Expected substring: not "euiErrorBoundary" Received string: "<div class=\"kbnRedirectCrossAppLinks\"><div class=\"euiErrorBoundary\"><div class=\"euiErrorBoundary__text\"><div class=\"euiText euiText--extraSmall\"><h1>Error</h1><pre class=\"euiErrorBoundary__stack\"><p>Error: Cannot read property 'startsWith' of undefined at PackageCard (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/x-pack/plugins/fleet/public/applications/integrations/sections/epm/components/package_card.tsx:55:19) at renderWithHooks (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:16260:18) at mountIndeterminateComponent (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:18794:13) at beginWork$1 (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:20162:16) at HTMLUnknownElement.callCallback (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:336:14) at HTMLUnknownElement.callTheUserObjectsOperation (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-environment-jsdom/node_modules/jsdom/lib/jsdom/living/generated/EventListener.js:26:30) at innerInvokeEventListeners (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-environment-jsdom/node_modules/jsdom/lib/jsdom/living/events/EventTarget-impl.js:318:25) at invokeEventListeners (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-environment-jsdom/node_modules/jsdom/lib/jsdom/living/events/EventTarget-impl.js:274:3) at HTMLUnknownElementImpl._dispatch (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-environment-jsdom/node_modules/jsdom/lib/jsdom/living/events/EventTarget-impl.js:221:9) at HTMLUnknownElementImpl.dispatchEvent (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-environment-jsdom/node_modules/jsdom/lib/jsdom/living/events/EventTarget-impl.js:94:17) at HTMLUnknownElement.dispatchEvent (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-environment-jsdom/node_modules/jsdom/lib/jsdom/living/generated/EventTarget.js:231:34) at Object.invokeGuardedCallbackDev (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:385:16) at invokeGuardedCallback (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:440:31) at beginWork$$1 (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:25780:7) at performUnitOfWork (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:24698:12) at workLoopSync (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:24671:22) at performSyncWorkOnRoot (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:24270:11) at /opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:12199:24 at unstable_runWithPriority (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/scheduler/cjs/scheduler.development.js:697:12) at runWithPriority$2 (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:12149:10) at flushSyncCallbackQueueImpl (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:12194:7) at flushSyncCallbackQueue (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:12182:3) at batchedUpdates$1 (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:24392:7) at Object.act (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom-test-utils.development.js:1092:14) at wrapAct (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/enzyme-adapter-react-16/src/ReactSixteenAdapter.js:405:13) at Object.render (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/enzyme-adapter-react-16/src/ReactSixteenAdapter.js:474:16) at new ReactWrapper (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/enzyme/src/ReactWrapper.js:115:16) at mount (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/enzyme/src/mount.js:10:10) at test (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/x-pack/plugins/fleet/storybook/smoke.test.tsx:17:22) at Object.&lt;anonymous&gt; (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/@storybook/addon-storyshots/dist/api/snapshotsTestsTemplate.js:42:20) at Promise.then.completed (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/utils.js:276:28) at new Promise (&lt;anonymous&gt;) at callAsyncCircusFn (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/utils.js:216:10) at _callCircusTest (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:212:40) at _runTest (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:149:3) at _runTestsForDescribeBlock (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:63:9) at _runTestsForDescribeBlock (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:57:9) at _runTestsForDescribeBlock (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:57:9) at run (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:25:3) at runAndTransformResultsToJestFormat (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/legacy-code-todo-rewrite/jestAdapterInit.js:176:21) at jestAdapter (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/legacy-code-todo-rewrite/jestAdapter.js:109:19) at runTestInternal (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-runner/build/runTest.js:380:16) at runTest (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-runner/build/runTest.js:472:34) at Object.worker (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-runner/build/testWorker.js:133:12)</p></pre></div></div></div></div>" at test (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/x-pack/plugins/fleet/storybook/smoke.test.tsx:20:33) at _callCircusTest (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:212:5) at _runTest (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:149:3) at _runTestsForDescribeBlock (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:63:9) at _runTestsForDescribeBlock (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:57:9) at _runTestsForDescribeBlock (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:57:9) at run (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:25:3) at runAndTransformResultsToJestFormat (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/legacy-code-todo-rewrite/jestAdapterInit.js:176:21) at jestAdapter (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/legacy-code-todo-rewrite/jestAdapter.js:109:19) at runTestInternal (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-runner/build/runTest.js:380:16) at runTest (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-runner/build/runTest.js:472:34) at Object.worker (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-runner/build/testWorker.js:133:12) ``` First failure: [CI Build - master](https://buildkite.com/elastic/kibana-hourly/builds/946#ff42e543-5004-441b-887b-0e66b9018892) <!-- kibanaCiData = {"failed-test":{"test.class":"Jest Tests.x-pack/plugins/fleet/storybook","test.name":"Storyshots Sections/EPM/Package List Grid List","test.failCount":1}} -->
1.0
Failing test: Jest Tests.x-pack/plugins/fleet/storybook - Storyshots Sections/EPM/Package List Grid List - A test failed on a tracked branch ``` Error: expect(received).not.toContain(expected) // indexOf Expected substring: not "euiErrorBoundary" Received string: "<div class=\"kbnRedirectCrossAppLinks\"><div class=\"euiErrorBoundary\"><div class=\"euiErrorBoundary__text\"><div class=\"euiText euiText--extraSmall\"><h1>Error</h1><pre class=\"euiErrorBoundary__stack\"><p>Error: Cannot read property 'startsWith' of undefined at PackageCard (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/x-pack/plugins/fleet/public/applications/integrations/sections/epm/components/package_card.tsx:55:19) at renderWithHooks (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:16260:18) at mountIndeterminateComponent (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:18794:13) at beginWork$1 (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:20162:16) at HTMLUnknownElement.callCallback (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:336:14) at HTMLUnknownElement.callTheUserObjectsOperation (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-environment-jsdom/node_modules/jsdom/lib/jsdom/living/generated/EventListener.js:26:30) at innerInvokeEventListeners (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-environment-jsdom/node_modules/jsdom/lib/jsdom/living/events/EventTarget-impl.js:318:25) at invokeEventListeners (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-environment-jsdom/node_modules/jsdom/lib/jsdom/living/events/EventTarget-impl.js:274:3) at HTMLUnknownElementImpl._dispatch (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-environment-jsdom/node_modules/jsdom/lib/jsdom/living/events/EventTarget-impl.js:221:9) at HTMLUnknownElementImpl.dispatchEvent (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-environment-jsdom/node_modules/jsdom/lib/jsdom/living/events/EventTarget-impl.js:94:17) at HTMLUnknownElement.dispatchEvent (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-environment-jsdom/node_modules/jsdom/lib/jsdom/living/generated/EventTarget.js:231:34) at Object.invokeGuardedCallbackDev (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:385:16) at invokeGuardedCallback (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:440:31) at beginWork$$1 (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:25780:7) at performUnitOfWork (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:24698:12) at workLoopSync (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:24671:22) at performSyncWorkOnRoot (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:24270:11) at /opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:12199:24 at unstable_runWithPriority (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/scheduler/cjs/scheduler.development.js:697:12) at runWithPriority$2 (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:12149:10) at flushSyncCallbackQueueImpl (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:12194:7) at flushSyncCallbackQueue (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:12182:3) at batchedUpdates$1 (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom.development.js:24392:7) at Object.act (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/react-dom/cjs/react-dom-test-utils.development.js:1092:14) at wrapAct (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/enzyme-adapter-react-16/src/ReactSixteenAdapter.js:405:13) at Object.render (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/enzyme-adapter-react-16/src/ReactSixteenAdapter.js:474:16) at new ReactWrapper (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/enzyme/src/ReactWrapper.js:115:16) at mount (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/enzyme/src/mount.js:10:10) at test (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/x-pack/plugins/fleet/storybook/smoke.test.tsx:17:22) at Object.&lt;anonymous&gt; (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/@storybook/addon-storyshots/dist/api/snapshotsTestsTemplate.js:42:20) at Promise.then.completed (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/utils.js:276:28) at new Promise (&lt;anonymous&gt;) at callAsyncCircusFn (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/utils.js:216:10) at _callCircusTest (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:212:40) at _runTest (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:149:3) at _runTestsForDescribeBlock (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:63:9) at _runTestsForDescribeBlock (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:57:9) at _runTestsForDescribeBlock (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:57:9) at run (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:25:3) at runAndTransformResultsToJestFormat (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/legacy-code-todo-rewrite/jestAdapterInit.js:176:21) at jestAdapter (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/legacy-code-todo-rewrite/jestAdapter.js:109:19) at runTestInternal (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-runner/build/runTest.js:380:16) at runTest (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-runner/build/runTest.js:472:34) at Object.worker (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-runner/build/testWorker.js:133:12)</p></pre></div></div></div></div>" at test (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/x-pack/plugins/fleet/storybook/smoke.test.tsx:20:33) at _callCircusTest (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:212:5) at _runTest (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:149:3) at _runTestsForDescribeBlock (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:63:9) at _runTestsForDescribeBlock (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:57:9) at _runTestsForDescribeBlock (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:57:9) at run (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/run.js:25:3) at runAndTransformResultsToJestFormat (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/legacy-code-todo-rewrite/jestAdapterInit.js:176:21) at jestAdapter (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-circus/build/legacy-code-todo-rewrite/jestAdapter.js:109:19) at runTestInternal (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-runner/build/runTest.js:380:16) at runTest (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-runner/build/runTest.js:472:34) at Object.worker (/opt/local-ssd/buildkite/builds/kb-c2-16-3c4a15400202cd9b/elastic/kibana-hourly/kibana/node_modules/jest-runner/build/testWorker.js:133:12) ``` First failure: [CI Build - master](https://buildkite.com/elastic/kibana-hourly/builds/946#ff42e543-5004-441b-887b-0e66b9018892) <!-- kibanaCiData = {"failed-test":{"test.class":"Jest Tests.x-pack/plugins/fleet/storybook","test.name":"Storyshots Sections/EPM/Package List Grid List","test.failCount":1}} -->
test
failing test jest tests x pack plugins fleet storybook storyshots sections epm package list grid list a test failed on a tracked branch error expect received not tocontain expected indexof expected substring not euierrorboundary received string error error cannot read property startswith of undefined at packagecard opt local ssd buildkite builds kb elastic kibana hourly kibana x pack plugins fleet public applications integrations sections epm components package card tsx at renderwithhooks opt local ssd buildkite builds kb elastic kibana hourly kibana node modules react dom cjs react dom development js at mountindeterminatecomponent opt local ssd buildkite builds kb elastic kibana hourly kibana node modules react dom cjs react dom development js at beginwork opt local ssd buildkite builds kb elastic kibana hourly kibana node modules react dom cjs react dom development js at htmlunknownelement callcallback opt local ssd buildkite builds kb elastic kibana hourly kibana node modules react dom cjs react dom development js at htmlunknownelement calltheuserobjectsoperation opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest environment jsdom node modules jsdom lib jsdom living generated eventlistener js at innerinvokeeventlisteners opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest environment jsdom node modules jsdom lib jsdom living events eventtarget impl js at invokeeventlisteners opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest environment jsdom node modules jsdom lib jsdom living events eventtarget impl js at htmlunknownelementimpl dispatch opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest environment jsdom node modules jsdom lib jsdom living events eventtarget impl js at htmlunknownelementimpl dispatchevent opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest environment jsdom node modules jsdom lib jsdom living events eventtarget impl js at htmlunknownelement dispatchevent opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest environment jsdom node modules jsdom lib jsdom living generated eventtarget js at object invokeguardedcallbackdev opt local ssd buildkite builds kb elastic kibana hourly kibana node modules react dom cjs react dom development js at invokeguardedcallback opt local ssd buildkite builds kb elastic kibana hourly kibana node modules react dom cjs react dom development js at beginwork opt local ssd buildkite builds kb elastic kibana hourly kibana node modules react dom cjs react dom development js at performunitofwork opt local ssd buildkite builds kb elastic kibana hourly kibana node modules react dom cjs react dom development js at workloopsync opt local ssd buildkite builds kb elastic kibana hourly kibana node modules react dom cjs react dom development js at performsyncworkonroot opt local ssd buildkite builds kb elastic kibana hourly kibana node modules react dom cjs react dom development js at opt local ssd buildkite builds kb elastic kibana hourly kibana node modules react dom cjs react dom development js at unstable runwithpriority opt local ssd buildkite builds kb elastic kibana hourly kibana node modules scheduler cjs scheduler development js at runwithpriority opt local ssd buildkite builds kb elastic kibana hourly kibana node modules react dom cjs react dom development js at flushsynccallbackqueueimpl opt local ssd buildkite builds kb elastic kibana hourly kibana node modules react dom cjs react dom development js at flushsynccallbackqueue opt local ssd buildkite builds kb elastic kibana hourly kibana node modules react dom cjs react dom development js at batchedupdates opt local ssd buildkite builds kb elastic kibana hourly kibana node modules react dom cjs react dom development js at object act opt local ssd buildkite builds kb elastic kibana hourly kibana node modules react dom cjs react dom test utils development js at wrapact opt local ssd buildkite builds kb elastic kibana hourly kibana node modules enzyme adapter react src reactsixteenadapter js at object render opt local ssd buildkite builds kb elastic kibana hourly kibana node modules enzyme adapter react src reactsixteenadapter js at new reactwrapper opt local ssd buildkite builds kb elastic kibana hourly kibana node modules enzyme src reactwrapper js at mount opt local ssd buildkite builds kb elastic kibana hourly kibana node modules enzyme src mount js at test opt local ssd buildkite builds kb elastic kibana hourly kibana x pack plugins fleet storybook smoke test tsx at object lt anonymous gt opt local ssd buildkite builds kb elastic kibana hourly kibana node modules storybook addon storyshots dist api snapshotsteststemplate js at promise then completed opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build utils js at new promise lt anonymous gt at callasynccircusfn opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build utils js at callcircustest opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build run js at runtest opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build run js at runtestsfordescribeblock opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build run js at runtestsfordescribeblock opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build run js at runtestsfordescribeblock opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build run js at run opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build run js at runandtransformresultstojestformat opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build legacy code todo rewrite jestadapterinit js at jestadapter opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build legacy code todo rewrite jestadapter js at runtestinternal opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest runner build runtest js at runtest opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest runner build runtest js at object worker opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest runner build testworker js at test opt local ssd buildkite builds kb elastic kibana hourly kibana x pack plugins fleet storybook smoke test tsx at callcircustest opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build run js at runtest opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build run js at runtestsfordescribeblock opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build run js at runtestsfordescribeblock opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build run js at runtestsfordescribeblock opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build run js at run opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build run js at runandtransformresultstojestformat opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build legacy code todo rewrite jestadapterinit js at jestadapter opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest circus build legacy code todo rewrite jestadapter js at runtestinternal opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest runner build runtest js at runtest opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest runner build runtest js at object worker opt local ssd buildkite builds kb elastic kibana hourly kibana node modules jest runner build testworker js first failure
1
36,916
2,813,567,513
IssuesEvent
2015-05-18 15:20:07
CruxFramework/crux
https://api.github.com/repos/CruxFramework/crux
closed
Error DateBox mask format.
bug imported invalid Module-CruxWidgets Priority-Medium
_From [guilhe...@triggolabs.com](https://code.google.com/u/113083826916919097787/) on August 27, 2014 10:03:43_ Was Identified que DateBox you assume the unexpected behavior in cases for the Following: 1 - If the component is declared without any "pattern" attribute, the mask will be unconfigured. 2 - When is passed the "pattern" attribute, the mask will be unconfigured Following in the cases: 2.1- DateBox in to the "storyboard" 2.2 - DateBox off the panel obs: DateBox Within the \<gwt: HTMLPanel> or \<crux: styledPanel> Correctly only works WHEN reporting the pattern. Version: 5.2.0-SNAPSHOT WIDGETS. Attached a project using the DateBox in several cases. **Attachment:** [test-components.rar](http://code.google.com/p/crux-framework/issues/detail?id=492) _Original issue: http://code.google.com/p/crux-framework/issues/detail?id=492_
1.0
Error DateBox mask format. - _From [guilhe...@triggolabs.com](https://code.google.com/u/113083826916919097787/) on August 27, 2014 10:03:43_ Was Identified que DateBox you assume the unexpected behavior in cases for the Following: 1 - If the component is declared without any "pattern" attribute, the mask will be unconfigured. 2 - When is passed the "pattern" attribute, the mask will be unconfigured Following in the cases: 2.1- DateBox in to the "storyboard" 2.2 - DateBox off the panel obs: DateBox Within the \<gwt: HTMLPanel> or \<crux: styledPanel> Correctly only works WHEN reporting the pattern. Version: 5.2.0-SNAPSHOT WIDGETS. Attached a project using the DateBox in several cases. **Attachment:** [test-components.rar](http://code.google.com/p/crux-framework/issues/detail?id=492) _Original issue: http://code.google.com/p/crux-framework/issues/detail?id=492_
non_test
error datebox mask format from on august was identified que datebox you assume the unexpected behavior in cases for the following if the component is declared without any pattern attribute the mask will be unconfigured when is passed the pattern attribute the mask will be unconfigured following in the cases datebox in to the storyboard datebox off the panel obs datebox within the or correctly only works when reporting the pattern version snapshot widgets attached a project using the datebox in several cases attachment original issue
0
190,127
15,218,236,534
IssuesEvent
2021-02-17 17:33:59
Carisk/Carisk-backend
https://api.github.com/repos/Carisk/Carisk-backend
closed
Novo problema a se resolver com aplicativo
Arquitetura documentation
**Descrição** O escopo do aplicativo atual é demasiadamente simple e pode ser exapandido de forma a se tornar mais útil ao usuário final. Isso é possível de ser feito com a detecção de novos problemas a serem resolvidos. **Critérios de aceitação** - [ ] Novo problema a ser resolvido, com sua possível solução e utilidade, bem como comunicação com o time e documentação.
1.0
Novo problema a se resolver com aplicativo - **Descrição** O escopo do aplicativo atual é demasiadamente simple e pode ser exapandido de forma a se tornar mais útil ao usuário final. Isso é possível de ser feito com a detecção de novos problemas a serem resolvidos. **Critérios de aceitação** - [ ] Novo problema a ser resolvido, com sua possível solução e utilidade, bem como comunicação com o time e documentação.
non_test
novo problema a se resolver com aplicativo descrição o escopo do aplicativo atual é demasiadamente simple e pode ser exapandido de forma a se tornar mais útil ao usuário final isso é possível de ser feito com a detecção de novos problemas a serem resolvidos critérios de aceitação novo problema a ser resolvido com sua possível solução e utilidade bem como comunicação com o time e documentação
0
25,383
18,670,745,421
IssuesEvent
2021-10-30 17:02:10
battlecode/battlecode21
https://api.github.com/repos/battlecode/battlecode21
closed
In tournament result publishers, make tournament url a command line arg
infrastructure
tour url is a var right now. this makes it really hard to run concurrently :(( also avoids funny pushes and things
1.0
In tournament result publishers, make tournament url a command line arg - tour url is a var right now. this makes it really hard to run concurrently :(( also avoids funny pushes and things
non_test
in tournament result publishers make tournament url a command line arg tour url is a var right now this makes it really hard to run concurrently also avoids funny pushes and things
0
73,657
3,419,254,383
IssuesEvent
2015-12-08 08:47:15
movabletype/smartphone-app
https://api.github.com/repos/movabletype/smartphone-app
opened
Would be great if user can configure image width freely.
enhancement Priority: MID
## Specification * The user must choose the width from a list, or enter a value. * The system must be provided default settings.
1.0
Would be great if user can configure image width freely. - ## Specification * The user must choose the width from a list, or enter a value. * The system must be provided default settings.
non_test
would be great if user can configure image width freely specification the user must choose the width from a list or enter a value the system must be provided default settings
0
577,808
17,135,390,342
IssuesEvent
2021-07-13 00:54:29
internetarchive/openlibrary
https://api.github.com/repos/internetarchive/openlibrary
opened
Upgrade to HTTP/2
Lead: @jamesachamp Needs: Investigation Priority: 3 Type: Feature Request
<!-- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] --> ### Describe the problem that you'd like solved <!-- A clear and concise description of what you want to happen. --> While page load times have improved over the past year, they could still be faster. ### Proposal & Constraints <!-- What is the proposed solution / implementation? Is there a precedent of this approach succeeding elsewhere? --> Upgrade our web and covers servers to use HTTP/2. Doing so will greatly limit the amount of TCP connections needed to serve a page to a patron. Server push may also be utilized to further improve performance. <!-- Which suggestions or requirements should be considered for how feature needs to appear or be implemented? --> ### Additional context <!-- Add any other context or screenshots about the feature request here. --> ### Stakeholders <!-- @ tag stakeholders of this bug -->
1.0
Upgrade to HTTP/2 - <!-- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] --> ### Describe the problem that you'd like solved <!-- A clear and concise description of what you want to happen. --> While page load times have improved over the past year, they could still be faster. ### Proposal & Constraints <!-- What is the proposed solution / implementation? Is there a precedent of this approach succeeding elsewhere? --> Upgrade our web and covers servers to use HTTP/2. Doing so will greatly limit the amount of TCP connections needed to serve a page to a patron. Server push may also be utilized to further improve performance. <!-- Which suggestions or requirements should be considered for how feature needs to appear or be implemented? --> ### Additional context <!-- Add any other context or screenshots about the feature request here. --> ### Stakeholders <!-- @ tag stakeholders of this bug -->
non_test
upgrade to http describe the problem that you d like solved while page load times have improved over the past year they could still be faster proposal constraints upgrade our web and covers servers to use http doing so will greatly limit the amount of tcp connections needed to serve a page to a patron server push may also be utilized to further improve performance additional context stakeholders
0
44,918
11,545,257,740
IssuesEvent
2020-02-18 13:07:15
imageio/imageio
https://api.github.com/repos/imageio/imageio
closed
Add proxy support to ```fetching.py```
building/installing plugin: ffmpeg
Installing ffmpeg automatically causes some problems behind a firewall using the ```imageio.plugins.ffmpeg.download()``` method. Proxy support can be enhanced by adding/changing the following within [```_fetch_file```](https://github.com/imageio/imageio/blob/368c724957cc67652f9c24b32628e4fdc19ecd3c/imageio/core/fetching.py#L159). ``` python authinfo = urllib.request.HTTPBasicAuthHandler() proxy_support = urllib.request.ProxyHandler({'https': 'http://yourproxy.corporate.zombie.net:2011'}) opener = urllib.request.build_opener(proxy_support, authinfo,urllib.request.CacheFTPHandler) urllib.request.install_opener(opener) remote_file = urllib.request.urlopen(url) ``` Perhaps it would be good to pass a proxy location through ```imageio.plugins.ffmpeg.download()``` to make the process suitable when needing a proxy?
1.0
Add proxy support to ```fetching.py``` - Installing ffmpeg automatically causes some problems behind a firewall using the ```imageio.plugins.ffmpeg.download()``` method. Proxy support can be enhanced by adding/changing the following within [```_fetch_file```](https://github.com/imageio/imageio/blob/368c724957cc67652f9c24b32628e4fdc19ecd3c/imageio/core/fetching.py#L159). ``` python authinfo = urllib.request.HTTPBasicAuthHandler() proxy_support = urllib.request.ProxyHandler({'https': 'http://yourproxy.corporate.zombie.net:2011'}) opener = urllib.request.build_opener(proxy_support, authinfo,urllib.request.CacheFTPHandler) urllib.request.install_opener(opener) remote_file = urllib.request.urlopen(url) ``` Perhaps it would be good to pass a proxy location through ```imageio.plugins.ffmpeg.download()``` to make the process suitable when needing a proxy?
non_test
add proxy support to fetching py installing ffmpeg automatically causes some problems behind a firewall using the imageio plugins ffmpeg download method proxy support can be enhanced by adding changing the following within python authinfo urllib request httpbasicauthhandler proxy support urllib request proxyhandler https opener urllib request build opener proxy support authinfo urllib request cacheftphandler urllib request install opener opener remote file urllib request urlopen url perhaps it would be good to pass a proxy location through imageio plugins ffmpeg download to make the process suitable when needing a proxy
0
142,519
11,483,933,125
IssuesEvent
2020-02-11 01:37:56
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
roachtest: pgjdbc failed
C-test-failure O-roachtest O-robot branch-provisional_202002062136_v19.1.8 release-blocker
[(roachtest).pgjdbc failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=1742234&tab=buildLog) on [provisional_202002062136_v19.1.8@407017cad14dfa63f19578055082dc10f3283cc4](https://github.com/cockroachdb/cockroach/commits/407017cad14dfa63f19578055082dc10f3283cc4): ``` The test failed on branch=provisional_202002062136_v19.1.8, cloud=gce: test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/pgjdbc/run_1 java_helpers.go:121,java_helpers.go:213,pgjdbc.go:172,pgjdbc.go:184,test_runner.go:741: XML syntax error on line 15002: illegal character code U+0001 ``` <details><summary>More</summary><p> Artifacts: [/pgjdbc](https://teamcity.cockroachdb.com/viewLog.html?buildId=1742234&tab=artifacts#/pgjdbc) Related: - #44932 roachtest: pgjdbc failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202002062122_v19.2.4](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202002062122_v19.2.4) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) - #44844 roachtest: pgjdbc failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-19.2](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-19.2) - #44284 roachtest: pgjdbc failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-19.1](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-19.1) - #44283 roachtest: pgjdbc failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-2.1](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-2.1) - #44254 roachtest: pgjdbc failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202001221958_v2.1.11](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202001221958_v2.1.11) - #42905 roachtest: pgjdbc failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-master](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-master) [See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Apgjdbc.%2A&sort=title&restgroup=false&display=lastcommented+project) <sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
2.0
roachtest: pgjdbc failed - [(roachtest).pgjdbc failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=1742234&tab=buildLog) on [provisional_202002062136_v19.1.8@407017cad14dfa63f19578055082dc10f3283cc4](https://github.com/cockroachdb/cockroach/commits/407017cad14dfa63f19578055082dc10f3283cc4): ``` The test failed on branch=provisional_202002062136_v19.1.8, cloud=gce: test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/pgjdbc/run_1 java_helpers.go:121,java_helpers.go:213,pgjdbc.go:172,pgjdbc.go:184,test_runner.go:741: XML syntax error on line 15002: illegal character code U+0001 ``` <details><summary>More</summary><p> Artifacts: [/pgjdbc](https://teamcity.cockroachdb.com/viewLog.html?buildId=1742234&tab=artifacts#/pgjdbc) Related: - #44932 roachtest: pgjdbc failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202002062122_v19.2.4](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202002062122_v19.2.4) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) - #44844 roachtest: pgjdbc failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-19.2](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-19.2) - #44284 roachtest: pgjdbc failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-19.1](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-19.1) - #44283 roachtest: pgjdbc failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-2.1](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-2.1) - #44254 roachtest: pgjdbc failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202001221958_v2.1.11](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202001221958_v2.1.11) - #42905 roachtest: pgjdbc failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-master](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-master) [See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Apgjdbc.%2A&sort=title&restgroup=false&display=lastcommented+project) <sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
test
roachtest pgjdbc failed on the test failed on branch provisional cloud gce test artifacts and logs in home agent work go src github com cockroachdb cockroach artifacts pgjdbc run java helpers go java helpers go pgjdbc go pgjdbc go test runner go xml syntax error on line illegal character code u more artifacts related roachtest pgjdbc failed roachtest pgjdbc failed roachtest pgjdbc failed roachtest pgjdbc failed roachtest pgjdbc failed roachtest pgjdbc failed powered by
1
336,707
30,217,337,629
IssuesEvent
2023-07-05 16:30:20
BoBAdministration/QA-Bug-Reports
https://api.github.com/repos/BoBAdministration/QA-Bug-Reports
closed
Acro doesnt have its idle animation in its emote wheel
Fixed-PendingTesting
**Describe the Bug** Acro doesnt have its idle animation in the emote wheel **To Reproduce** 1. log into tester server 2. spawn in as acro 3. open emote menu **Expected behavior** emote menu has idle anim **Actual behavior** emote menu does not have idle anim **Screenshots & Video** ![image](https://github.com/BoBAdministration/QA-Bug-Reports/assets/133143622/88b67e05-fb76-4d0a-86ce-90a9680a5d26) **Branch Version** Tester **Character Information** 1.2 Acro **Additional Information** Tested on titania
1.0
Acro doesnt have its idle animation in its emote wheel - **Describe the Bug** Acro doesnt have its idle animation in the emote wheel **To Reproduce** 1. log into tester server 2. spawn in as acro 3. open emote menu **Expected behavior** emote menu has idle anim **Actual behavior** emote menu does not have idle anim **Screenshots & Video** ![image](https://github.com/BoBAdministration/QA-Bug-Reports/assets/133143622/88b67e05-fb76-4d0a-86ce-90a9680a5d26) **Branch Version** Tester **Character Information** 1.2 Acro **Additional Information** Tested on titania
test
acro doesnt have its idle animation in its emote wheel describe the bug acro doesnt have its idle animation in the emote wheel to reproduce log into tester server spawn in as acro open emote menu expected behavior emote menu has idle anim actual behavior emote menu does not have idle anim screenshots video branch version tester character information acro additional information tested on titania
1
230,489
17,618,833,591
IssuesEvent
2021-08-18 13:09:41
UnBArqDsw2021-1/2021.1_G02_TaNaMesa_docs
https://api.github.com/repos/UnBArqDsw2021-1/2021.1_G02_TaNaMesa_docs
opened
Protótipo de Alta Fidelidade
documentation
## Descrição da issue Issue destinada a criação e documentação do protótipo de alta fidelidade, baseado no protótipo de baixa fidelidade e nos requisitos elicitados para o projeto. ## Contexto adicional https://github.com/UnBArqDsw2021-1/2021.1_G02_TaNaMesa_docs/issues/69 ## Tópicos: - [ ] Criar documento - [ ] Criar protótipo - [ ] Documentar telas - [ ] Revisar documento
1.0
Protótipo de Alta Fidelidade - ## Descrição da issue Issue destinada a criação e documentação do protótipo de alta fidelidade, baseado no protótipo de baixa fidelidade e nos requisitos elicitados para o projeto. ## Contexto adicional https://github.com/UnBArqDsw2021-1/2021.1_G02_TaNaMesa_docs/issues/69 ## Tópicos: - [ ] Criar documento - [ ] Criar protótipo - [ ] Documentar telas - [ ] Revisar documento
non_test
protótipo de alta fidelidade descrição da issue issue destinada a criação e documentação do protótipo de alta fidelidade baseado no protótipo de baixa fidelidade e nos requisitos elicitados para o projeto contexto adicional tópicos criar documento criar protótipo documentar telas revisar documento
0
91,999
18,758,275,952
IssuesEvent
2021-11-05 13:40:07
Onelinerhub/onelinerhub
https://api.github.com/repos/Onelinerhub/onelinerhub
closed
Write shortest possible code: How to import module from another directory (python)
help wanted good first issue code python
Please write shortest code example for this question: **How to import module from another directory** in python ### How to do it: 1. Go to [python codes](https://github.com/Onelinerhub/onelinerhub/tree/main/python) 2. Create new file (named in underscore case, should contain key words from title) with `md` extension (markdown file). 3. Propose new file with following content (please use all three blocks if possible - title, code itself and explanations list): ~~~ # How to import module from another directory ```python code part1 part2 part3 ... ``` - part1 - explain code part 1 - part2 - explain code part 2 - ... ~~~ More [advanced template](https://github.com/Onelinerhub/onelinerhub/blob/main/template.md) for examples and linked solutions. More [docs here](https://github.com/Onelinerhub/onelinerhub#onelinerhub).
1.0
Write shortest possible code: How to import module from another directory (python) - Please write shortest code example for this question: **How to import module from another directory** in python ### How to do it: 1. Go to [python codes](https://github.com/Onelinerhub/onelinerhub/tree/main/python) 2. Create new file (named in underscore case, should contain key words from title) with `md` extension (markdown file). 3. Propose new file with following content (please use all three blocks if possible - title, code itself and explanations list): ~~~ # How to import module from another directory ```python code part1 part2 part3 ... ``` - part1 - explain code part 1 - part2 - explain code part 2 - ... ~~~ More [advanced template](https://github.com/Onelinerhub/onelinerhub/blob/main/template.md) for examples and linked solutions. More [docs here](https://github.com/Onelinerhub/onelinerhub#onelinerhub).
non_test
write shortest possible code how to import module from another directory python please write shortest code example for this question how to import module from another directory in python how to do it go to create new file named in underscore case should contain key words from title with md extension markdown file propose new file with following content please use all three blocks if possible title code itself and explanations list how to import module from another directory python code explain code part explain code part more for examples and linked solutions more
0
328,972
28,143,241,456
IssuesEvent
2023-04-02 07:07:37
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
opened
Fix backend_handler.test_variables
Sub Task Failing Test
| | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4586867454/jobs/8100010916" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/4586867454/jobs/8100010808" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4586867454/jobs/8100007678" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/4586867454/jobs/8100009708" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> <details> <summary>Not found</summary> Not found </details> <details> <summary>Not found</summary> Not found </details> <details> <summary>Not found</summary> Not found </details> <details> <summary>Not found</summary> Not found </details>
1.0
Fix backend_handler.test_variables - | | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4586867454/jobs/8100010916" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/4586867454/jobs/8100010808" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4586867454/jobs/8100007678" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/4586867454/jobs/8100009708" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> <details> <summary>Not found</summary> Not found </details> <details> <summary>Not found</summary> Not found </details> <details> <summary>Not found</summary> Not found </details> <details> <summary>Not found</summary> Not found </details>
test
fix backend handler test variables tensorflow img src torch img src numpy img src jax img src not found not found not found not found not found not found not found not found
1
149,941
11,940,788,156
IssuesEvent
2020-04-02 17:18:12
rancher/rancher
https://api.github.com/repos/rancher/rancher
opened
test_catalog_refresh flaky test
kind/flaky-test
https://drone-publish.rancher.io/rancher/rancher/2676/1/2 https://drone-pr.rancher.io/rancher/rancher/7191/2/2 ``` =================================== FAILURES =================================== -- 904 | _____________________________ test_catalog_refresh _____________________________ 905 | [gw5] linux -- Python 3.7.5 /go/src/github.com/rancher/rancher/tests/integration/.tox/py37/bin/python3.7 906 |   907 | admin_mc = <suite.conftest.ManagementContext object at 0x7f813b3d1ed0> 908 |   909 | def test_catalog_refresh(admin_mc): 910 | """Test that on refresh the response includes the names of the catalogs 911 | that are being refreshed""" 912 | client = admin_mc.client 913 | catalog = client.by_id_catalog("library") 914 | out = client.action(obj=catalog, action_name="refresh") 915 | assert out['catalogs'][0] == "library" 916 |   917 | catalogs = client.list_catalog() 918 | > out = client.action(obj=catalogs, action_name="refresh") 919 |   920 | test_catalog.py:282: 921 | _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 922 | ../.tox/py37/lib/python3.7/site-packages/rancher.py:471: in action 923 | return self._post_and_retry(url, *args, **kw) 924 | ../.tox/py37/lib/python3.7/site-packages/rancher.py:430: in _post_and_retry 925 | raise e 926 | ../.tox/py37/lib/python3.7/site-packages/rancher.py:423: in _post_and_retry 927 | return self._post(url, data=self._to_dict(*args, **kw)) 928 | ../.tox/py37/lib/python3.7/site-packages/rancher.py:68: in wrapped 929 | return fn(*args, **kw) 930 | ../.tox/py37/lib/python3.7/site-packages/rancher.py:319: in _post 931 | self._error(r.text) 932 | _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 933 |   934 | self = <rancher.Client object at 0x7f813b745e90> 935 | text = '{"baseType":"error","code":"ServerError","message":"catalogs.management.cattle.io \\"random-284755-1585846570\\" not found","status":500,"type":"error"}\n' 936 |   937 | def _error(self, text): 938 | > raise ApiError(self._unmarshall(text)) 939 | E rancher.ApiError: (ApiError(...), 'ServerError : catalogs.management.cattle.io "random-284755-1585846570" not found\n\t{\'baseType\': \'error\', \'code\': \'ServerError\', \'message\': \'catalogs.management.cattle.io "random-284755-1585846570" not found\', \'status\': 500, \'type\': \'error\'}') 940 |   941 | ../.tox/py37/lib/python3.7/site-packages/rancher.py:299: ApiError ```
1.0
test_catalog_refresh flaky test - https://drone-publish.rancher.io/rancher/rancher/2676/1/2 https://drone-pr.rancher.io/rancher/rancher/7191/2/2 ``` =================================== FAILURES =================================== -- 904 | _____________________________ test_catalog_refresh _____________________________ 905 | [gw5] linux -- Python 3.7.5 /go/src/github.com/rancher/rancher/tests/integration/.tox/py37/bin/python3.7 906 |   907 | admin_mc = <suite.conftest.ManagementContext object at 0x7f813b3d1ed0> 908 |   909 | def test_catalog_refresh(admin_mc): 910 | """Test that on refresh the response includes the names of the catalogs 911 | that are being refreshed""" 912 | client = admin_mc.client 913 | catalog = client.by_id_catalog("library") 914 | out = client.action(obj=catalog, action_name="refresh") 915 | assert out['catalogs'][0] == "library" 916 |   917 | catalogs = client.list_catalog() 918 | > out = client.action(obj=catalogs, action_name="refresh") 919 |   920 | test_catalog.py:282: 921 | _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 922 | ../.tox/py37/lib/python3.7/site-packages/rancher.py:471: in action 923 | return self._post_and_retry(url, *args, **kw) 924 | ../.tox/py37/lib/python3.7/site-packages/rancher.py:430: in _post_and_retry 925 | raise e 926 | ../.tox/py37/lib/python3.7/site-packages/rancher.py:423: in _post_and_retry 927 | return self._post(url, data=self._to_dict(*args, **kw)) 928 | ../.tox/py37/lib/python3.7/site-packages/rancher.py:68: in wrapped 929 | return fn(*args, **kw) 930 | ../.tox/py37/lib/python3.7/site-packages/rancher.py:319: in _post 931 | self._error(r.text) 932 | _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 933 |   934 | self = <rancher.Client object at 0x7f813b745e90> 935 | text = '{"baseType":"error","code":"ServerError","message":"catalogs.management.cattle.io \\"random-284755-1585846570\\" not found","status":500,"type":"error"}\n' 936 |   937 | def _error(self, text): 938 | > raise ApiError(self._unmarshall(text)) 939 | E rancher.ApiError: (ApiError(...), 'ServerError : catalogs.management.cattle.io "random-284755-1585846570" not found\n\t{\'baseType\': \'error\', \'code\': \'ServerError\', \'message\': \'catalogs.management.cattle.io "random-284755-1585846570" not found\', \'status\': 500, \'type\': \'error\'}') 940 |   941 | ../.tox/py37/lib/python3.7/site-packages/rancher.py:299: ApiError ```
test
test catalog refresh flaky test failures test catalog refresh linux python go src github com rancher rancher tests integration tox bin   admin mc   def test catalog refresh admin mc test that on refresh the response includes the names of the catalogs that are being refreshed client admin mc client catalog client by id catalog library out client action obj catalog action name refresh assert out library   catalogs client list catalog out client action obj catalogs action name refresh   test catalog py tox lib site packages rancher py in action return self post and retry url args kw tox lib site packages rancher py in post and retry raise e tox lib site packages rancher py in post and retry return self post url data self to dict args kw tox lib site packages rancher py in wrapped return fn args kw tox lib site packages rancher py in post self error r text   self text basetype error code servererror message catalogs management cattle io random not found status type error n   def error self text raise apierror self unmarshall text e rancher apierror apierror servererror catalogs management cattle io random not found n t basetype error code servererror message catalogs management cattle io random not found status type error   tox lib site packages rancher py apierror
1
70,212
23,051,623,561
IssuesEvent
2022-07-24 18:11:11
vector-im/element-android
https://api.github.com/repos/vector-im/element-android
closed
Voice message lock mechanism is incompatible with gesture navigation
T-Defect A-Composer A-Voice-Messages S-Major O-Occasional Z-Platform Specific
#### Describe the bug If your phone uses gesture navigation, it is impossible to use the voice message slide-to-lock feature, since sliding upwards from the voice message button is interpreted as trying to close the app. #### To Reproduce 1. Enable gesture navigation on your phone 2. Press and hold the voice message button 3. Attempt to slide upwards to lock the recording #### Expected behavior An upwards swipe should lock the recording instead of trying to close the app #### Screenshots https://user-images.githubusercontent.com/48614497/130335718-ac906310-2be0-4a5e-ae9a-64f3babebd85.mp4 #### Smartphone (please complete the following information): - Device: Google Pixel 3a - OS: LineageOS 15.1 #### Additional context - App version and store: 1.2.0 - F-Droid
1.0
Voice message lock mechanism is incompatible with gesture navigation - #### Describe the bug If your phone uses gesture navigation, it is impossible to use the voice message slide-to-lock feature, since sliding upwards from the voice message button is interpreted as trying to close the app. #### To Reproduce 1. Enable gesture navigation on your phone 2. Press and hold the voice message button 3. Attempt to slide upwards to lock the recording #### Expected behavior An upwards swipe should lock the recording instead of trying to close the app #### Screenshots https://user-images.githubusercontent.com/48614497/130335718-ac906310-2be0-4a5e-ae9a-64f3babebd85.mp4 #### Smartphone (please complete the following information): - Device: Google Pixel 3a - OS: LineageOS 15.1 #### Additional context - App version and store: 1.2.0 - F-Droid
non_test
voice message lock mechanism is incompatible with gesture navigation describe the bug if your phone uses gesture navigation it is impossible to use the voice message slide to lock feature since sliding upwards from the voice message button is interpreted as trying to close the app to reproduce enable gesture navigation on your phone press and hold the voice message button attempt to slide upwards to lock the recording expected behavior an upwards swipe should lock the recording instead of trying to close the app screenshots smartphone please complete the following information device google pixel os lineageos additional context app version and store f droid
0