Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
4
112
repo_url
stringlengths
33
141
action
stringclasses
3 values
title
stringlengths
1
1.02k
labels
stringlengths
4
1.54k
body
stringlengths
1
262k
index
stringclasses
17 values
text_combine
stringlengths
95
262k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
146,985
11,765,547,426
IssuesEvent
2020-03-14 17:54:23
IEMLdev/Intlekt-editor
https://api.github.com/repos/IEMLdev/Intlekt-editor
closed
après avoir marché pendant une heure, l'éditeur de dictionnaire a encore bloqué
bug to retest
Peut-être que c'est normal parce que j'ai fait beaucoup de modif et qu'il faut attendre demain 😬 <img width="1386" alt="Screenshot 2019-10-19 14 47 04" src="https://user-images.githubusercontent.com/19435943/67150013-2743c800-f280-11e9-95cb-6d0648e8b152.png">
1.0
après avoir marché pendant une heure, l'éditeur de dictionnaire a encore bloqué - Peut-être que c'est normal parce que j'ai fait beaucoup de modif et qu'il faut attendre demain 😬 <img width="1386" alt="Screenshot 2019-10-19 14 47 04" src="https://user-images.githubusercontent.com/19435943/67150013-2743c800-f280-11e9-95cb-6d0648e8b152.png">
test
après avoir marché pendant une heure l éditeur de dictionnaire a encore bloqué peut être que c est normal parce que j ai fait beaucoup de modif et qu il faut attendre demain 😬 img width alt screenshot src
1
257,516
27,563,796,894
IssuesEvent
2023-03-08 01:07:09
billmcchesney1/superagent
https://api.github.com/repos/billmcchesney1/superagent
opened
CVE-2022-21681 (High) detected in marked-1.2.7.tgz
security vulnerability
## CVE-2022-21681 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>marked-1.2.7.tgz</b></p></summary> <p>A markdown parser built for speed</p> <p>Library home page: <a href="https://registry.npmjs.org/marked/-/marked-1.2.7.tgz">https://registry.npmjs.org/marked/-/marked-1.2.7.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/marked/package.json</p> <p> Dependency Hierarchy: - :x: **marked-1.2.7.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Marked is a markdown parser and compiler. Prior to version 4.0.10, the regular expression `inline.reflinkSearch` may cause catastrophic backtracking against some strings and lead to a denial of service (DoS). Anyone who runs untrusted markdown through a vulnerable version of marked and does not use a worker with a time limit may be affected. This issue is patched in version 4.0.10. As a workaround, avoid running untrusted markdown through marked or run marked on a worker thread and set a reasonable time limit to prevent draining resources. <p>Publish Date: 2022-01-14 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-21681>CVE-2022-21681</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-5v2h-r2cx-5xgj">https://github.com/advisories/GHSA-5v2h-r2cx-5xgj</a></p> <p>Release Date: 2022-01-14</p> <p>Fix Resolution: 4.0.10</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue
True
CVE-2022-21681 (High) detected in marked-1.2.7.tgz - ## CVE-2022-21681 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>marked-1.2.7.tgz</b></p></summary> <p>A markdown parser built for speed</p> <p>Library home page: <a href="https://registry.npmjs.org/marked/-/marked-1.2.7.tgz">https://registry.npmjs.org/marked/-/marked-1.2.7.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/marked/package.json</p> <p> Dependency Hierarchy: - :x: **marked-1.2.7.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Marked is a markdown parser and compiler. Prior to version 4.0.10, the regular expression `inline.reflinkSearch` may cause catastrophic backtracking against some strings and lead to a denial of service (DoS). Anyone who runs untrusted markdown through a vulnerable version of marked and does not use a worker with a time limit may be affected. This issue is patched in version 4.0.10. As a workaround, avoid running untrusted markdown through marked or run marked on a worker thread and set a reasonable time limit to prevent draining resources. <p>Publish Date: 2022-01-14 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-21681>CVE-2022-21681</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-5v2h-r2cx-5xgj">https://github.com/advisories/GHSA-5v2h-r2cx-5xgj</a></p> <p>Release Date: 2022-01-14</p> <p>Fix Resolution: 4.0.10</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue
non_test
cve high detected in marked tgz cve high severity vulnerability vulnerable library marked tgz a markdown parser built for speed library home page a href path to dependency file package json path to vulnerable library node modules marked package json dependency hierarchy x marked tgz vulnerable library found in base branch master vulnerability details marked is a markdown parser and compiler prior to version the regular expression inline reflinksearch may cause catastrophic backtracking against some strings and lead to a denial of service dos anyone who runs untrusted markdown through a vulnerable version of marked and does not use a worker with a time limit may be affected this issue is patched in version as a workaround avoid running untrusted markdown through marked or run marked on a worker thread and set a reasonable time limit to prevent draining resources publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution rescue worker helmet automatic remediation is available for this issue
0
172,072
6,498,708,315
IssuesEvent
2017-08-22 18:29:54
FrogTheFrog/steam-rom-manager
https://api.github.com/repos/FrogTheFrog/steam-rom-manager
closed
Unsaved parser settings are lost when clicking "Test parser"
bug enhancement priority
1. Open a previously saved parser 1. Change settings such as ROMs directory and User's glob 1. Click the "Test parser" button. 1. Click back to the parser you were testing. All setting changes are gone. I would expect the behavior would be to leave the modified settings in place. The primary reason I would ever hit "Test parser" would be to see if my new parser settings work well. I wouldn't want to have to save my parser settings before I got them right. If you do allow navigating away without losing parsing settings, I can see how this bug report would actually turn into more of an enhancement. A more robust save system which shows each modified parser with a * next to its name so that you are aware it has been edited would then make it apparent that it is not yet saved.
1.0
Unsaved parser settings are lost when clicking "Test parser" - 1. Open a previously saved parser 1. Change settings such as ROMs directory and User's glob 1. Click the "Test parser" button. 1. Click back to the parser you were testing. All setting changes are gone. I would expect the behavior would be to leave the modified settings in place. The primary reason I would ever hit "Test parser" would be to see if my new parser settings work well. I wouldn't want to have to save my parser settings before I got them right. If you do allow navigating away without losing parsing settings, I can see how this bug report would actually turn into more of an enhancement. A more robust save system which shows each modified parser with a * next to its name so that you are aware it has been edited would then make it apparent that it is not yet saved.
non_test
unsaved parser settings are lost when clicking test parser open a previously saved parser change settings such as roms directory and user s glob click the test parser button click back to the parser you were testing all setting changes are gone i would expect the behavior would be to leave the modified settings in place the primary reason i would ever hit test parser would be to see if my new parser settings work well i wouldn t want to have to save my parser settings before i got them right if you do allow navigating away without losing parsing settings i can see how this bug report would actually turn into more of an enhancement a more robust save system which shows each modified parser with a next to its name so that you are aware it has been edited would then make it apparent that it is not yet saved
0
71,634
8,671,799,041
IssuesEvent
2018-11-29 20:11:06
poanetwork/blockscout
https://api.github.com/repos/poanetwork/blockscout
opened
Make the requirement for verifying a smart contract less harsh
design enhancement priority: medium team: developer
On the transaction details page, the user is prompted to verify a contract to view the decoded input data. This message is currently using the `danger` style. This should be changed to `info` to appear less like an error. <img width="1111" alt="screen shot 2018-11-29 at 3 07 57 pm" src="https://user-images.githubusercontent.com/17620007/49248911-e5e8c680-f3e8-11e8-9335-9ce212de0537.png">
1.0
Make the requirement for verifying a smart contract less harsh - On the transaction details page, the user is prompted to verify a contract to view the decoded input data. This message is currently using the `danger` style. This should be changed to `info` to appear less like an error. <img width="1111" alt="screen shot 2018-11-29 at 3 07 57 pm" src="https://user-images.githubusercontent.com/17620007/49248911-e5e8c680-f3e8-11e8-9335-9ce212de0537.png">
non_test
make the requirement for verifying a smart contract less harsh on the transaction details page the user is prompted to verify a contract to view the decoded input data this message is currently using the danger style this should be changed to info to appear less like an error img width alt screen shot at pm src
0
277,575
24,086,141,779
IssuesEvent
2022-09-19 11:07:08
streamnative/pulsar
https://api.github.com/repos/streamnative/pulsar
opened
ISSUE-17713: Flaky-test: SqliteJdbcSinkTest.tearDown
component/test flaky-tests
Original Issue: apache/pulsar#17713 --- ### Search before asking - [X] I searched in the [issues](https://github.com/apache/pulsar/issues) and found nothing similar. ### Example failure https://github.com/apache/pulsar/actions/runs/3069039857/jobs/4977568132#step:10:4694 ### Exception stacktrace ``` Error: Tests run: 19, Failures: 1, Errors: 0, Skipped: 11, Time elapsed: 4.036 s <<< FAILURE! - in org.apache.pulsar.io.jdbc.SqliteJdbcSinkTest Error: tearDown(org.apache.pulsar.io.jdbc.SqliteJdbcSinkTest) Time elapsed: 0.11 s <<< FAILURE! org.sqlite.SQLiteException: [SQLITE_ERROR] SQL error or missing database (cannot commit - no transaction is active) at org.sqlite.core.DB.newSQLException(DB.java:1030) at org.sqlite.core.DB.newSQLException(DB.java:1042) at org.sqlite.core.DB.throwex(DB.java:1007) at org.sqlite.core.DB.exec(DB.java:178) at org.sqlite.SQLiteConnection.commit(SQLiteConnection.java:421) at org.apache.pulsar.io.jdbc.JdbcAbstractSink.close(JdbcAbstractSink.java:141) at org.apache.pulsar.io.jdbc.SqliteJdbcSinkTest.tearDown(SqliteJdbcSinkTest.java:122) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.testng.internal.MethodInvocationHelper.invokeMethod(MethodInvocationHelper.java:132) at org.testng.internal.MethodInvocationHelper.invokeMethodConsideringTimeout(MethodInvocationHelper.java:61) at org.testng.internal.ConfigInvoker.invokeConfigurationMethod(ConfigInvoker.java:366) at org.testng.internal.ConfigInvoker.invokeConfigurations(ConfigInvoker.java:320) at org.testng.internal.TestInvoker.runConfigMethods(TestInvoker.java:701) at org.testng.internal.TestInvoker.runAfterGroupsConfigurations(TestInvoker.java:677) at org.testng.internal.TestInvoker.invokeMethod(TestInvoker.java:661) at org.testng.internal.TestInvoker.invokeTestMethod(TestInvoker.java:174) at org.testng.internal.MethodRunner.runInSequence(MethodRunner.java:46) at org.testng.internal.TestInvoker$MethodInvocationAgent.invoke(TestInvoker.java:822) at org.testng.internal.TestInvoker.invokeTestMethods(TestInvoker.java:147) at org.testng.internal.TestMethodWorker.invokeTestMethods(TestMethodWorker.java:146) at org.testng.internal.TestMethodWorker.run(TestMethodWorker.java:128) at java.base/java.util.ArrayList.forEach(ArrayList.java:1511) at org.testng.TestRunner.privateRun(TestRunner.java:764) at org.testng.TestRunner.run(TestRunner.java:585) at org.testng.SuiteRunner.runTest(SuiteRunner.java:384) at org.testng.SuiteRunner.runSequentially(SuiteRunner.java:378) at org.testng.SuiteRunner.privateRun(SuiteRunner.java:337) at org.testng.SuiteRunner.run(SuiteRunner.java:286) at org.testng.SuiteRunnerWorker.runSuite(SuiteRunnerWorker.java:53) at org.testng.SuiteRunnerWorker.run(SuiteRunnerWorker.java:96) at org.testng.TestNG.runSuitesSequentially(TestNG.java:1218) at org.testng.TestNG.runSuitesLocally(TestNG.java:1140) at org.testng.TestNG.runSuites(TestNG.java:1069) at org.testng.TestNG.run(TestNG.java:1037) at org.apache.maven.surefire.testng.TestNGExecutor.run(TestNGExecutor.java:135) at org.apache.maven.surefire.testng.TestNGDirectoryTestSuite.executeSingleClass(TestNGDirectoryTestSuite.java:112) at org.apache.maven.surefire.testng.TestNGDirectoryTestSuite.executeLazy(TestNGDirectoryTestSuite.java:123) at org.apache.maven.surefire.testng.TestNGDirectoryTestSuite.execute(TestNGDirectoryTestSuite.java:90) at org.apache.maven.surefire.testng.TestNGProvider.invoke(TestNGProvider.java:146) at org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:384) at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:345) at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:126) at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:418) ``` ### Are you willing to submit a PR? - [ ] I'm willing to submit a PR!
2.0
ISSUE-17713: Flaky-test: SqliteJdbcSinkTest.tearDown - Original Issue: apache/pulsar#17713 --- ### Search before asking - [X] I searched in the [issues](https://github.com/apache/pulsar/issues) and found nothing similar. ### Example failure https://github.com/apache/pulsar/actions/runs/3069039857/jobs/4977568132#step:10:4694 ### Exception stacktrace ``` Error: Tests run: 19, Failures: 1, Errors: 0, Skipped: 11, Time elapsed: 4.036 s <<< FAILURE! - in org.apache.pulsar.io.jdbc.SqliteJdbcSinkTest Error: tearDown(org.apache.pulsar.io.jdbc.SqliteJdbcSinkTest) Time elapsed: 0.11 s <<< FAILURE! org.sqlite.SQLiteException: [SQLITE_ERROR] SQL error or missing database (cannot commit - no transaction is active) at org.sqlite.core.DB.newSQLException(DB.java:1030) at org.sqlite.core.DB.newSQLException(DB.java:1042) at org.sqlite.core.DB.throwex(DB.java:1007) at org.sqlite.core.DB.exec(DB.java:178) at org.sqlite.SQLiteConnection.commit(SQLiteConnection.java:421) at org.apache.pulsar.io.jdbc.JdbcAbstractSink.close(JdbcAbstractSink.java:141) at org.apache.pulsar.io.jdbc.SqliteJdbcSinkTest.tearDown(SqliteJdbcSinkTest.java:122) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.testng.internal.MethodInvocationHelper.invokeMethod(MethodInvocationHelper.java:132) at org.testng.internal.MethodInvocationHelper.invokeMethodConsideringTimeout(MethodInvocationHelper.java:61) at org.testng.internal.ConfigInvoker.invokeConfigurationMethod(ConfigInvoker.java:366) at org.testng.internal.ConfigInvoker.invokeConfigurations(ConfigInvoker.java:320) at org.testng.internal.TestInvoker.runConfigMethods(TestInvoker.java:701) at org.testng.internal.TestInvoker.runAfterGroupsConfigurations(TestInvoker.java:677) at org.testng.internal.TestInvoker.invokeMethod(TestInvoker.java:661) at org.testng.internal.TestInvoker.invokeTestMethod(TestInvoker.java:174) at org.testng.internal.MethodRunner.runInSequence(MethodRunner.java:46) at org.testng.internal.TestInvoker$MethodInvocationAgent.invoke(TestInvoker.java:822) at org.testng.internal.TestInvoker.invokeTestMethods(TestInvoker.java:147) at org.testng.internal.TestMethodWorker.invokeTestMethods(TestMethodWorker.java:146) at org.testng.internal.TestMethodWorker.run(TestMethodWorker.java:128) at java.base/java.util.ArrayList.forEach(ArrayList.java:1511) at org.testng.TestRunner.privateRun(TestRunner.java:764) at org.testng.TestRunner.run(TestRunner.java:585) at org.testng.SuiteRunner.runTest(SuiteRunner.java:384) at org.testng.SuiteRunner.runSequentially(SuiteRunner.java:378) at org.testng.SuiteRunner.privateRun(SuiteRunner.java:337) at org.testng.SuiteRunner.run(SuiteRunner.java:286) at org.testng.SuiteRunnerWorker.runSuite(SuiteRunnerWorker.java:53) at org.testng.SuiteRunnerWorker.run(SuiteRunnerWorker.java:96) at org.testng.TestNG.runSuitesSequentially(TestNG.java:1218) at org.testng.TestNG.runSuitesLocally(TestNG.java:1140) at org.testng.TestNG.runSuites(TestNG.java:1069) at org.testng.TestNG.run(TestNG.java:1037) at org.apache.maven.surefire.testng.TestNGExecutor.run(TestNGExecutor.java:135) at org.apache.maven.surefire.testng.TestNGDirectoryTestSuite.executeSingleClass(TestNGDirectoryTestSuite.java:112) at org.apache.maven.surefire.testng.TestNGDirectoryTestSuite.executeLazy(TestNGDirectoryTestSuite.java:123) at org.apache.maven.surefire.testng.TestNGDirectoryTestSuite.execute(TestNGDirectoryTestSuite.java:90) at org.apache.maven.surefire.testng.TestNGProvider.invoke(TestNGProvider.java:146) at org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:384) at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:345) at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:126) at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:418) ``` ### Are you willing to submit a PR? - [ ] I'm willing to submit a PR!
test
issue flaky test sqlitejdbcsinktest teardown original issue apache pulsar search before asking i searched in the and found nothing similar example failure exception stacktrace error tests run failures errors skipped time elapsed s failure in org apache pulsar io jdbc sqlitejdbcsinktest error teardown org apache pulsar io jdbc sqlitejdbcsinktest time elapsed s failure org sqlite sqliteexception sql error or missing database cannot commit no transaction is active at org sqlite core db newsqlexception db java at org sqlite core db newsqlexception db java at org sqlite core db throwex db java at org sqlite core db exec db java at org sqlite sqliteconnection commit sqliteconnection java at org apache pulsar io jdbc jdbcabstractsink close jdbcabstractsink java at org apache pulsar io jdbc sqlitejdbcsinktest teardown sqlitejdbcsinktest java at java base jdk internal reflect nativemethodaccessorimpl native method at java base jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at java base jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java base java lang reflect method invoke method java at org testng internal methodinvocationhelper invokemethod methodinvocationhelper java at org testng internal methodinvocationhelper invokemethodconsideringtimeout methodinvocationhelper java at org testng internal configinvoker invokeconfigurationmethod configinvoker java at org testng internal configinvoker invokeconfigurations configinvoker java at org testng internal testinvoker runconfigmethods testinvoker java at org testng internal testinvoker runaftergroupsconfigurations testinvoker java at org testng internal testinvoker invokemethod testinvoker java at org testng internal testinvoker invoketestmethod testinvoker java at org testng internal methodrunner runinsequence methodrunner java at org testng internal testinvoker methodinvocationagent invoke testinvoker java at org testng internal testinvoker invoketestmethods testinvoker java at org testng internal testmethodworker invoketestmethods testmethodworker java at org testng internal testmethodworker run testmethodworker java at java base java util arraylist foreach arraylist java at org testng testrunner privaterun testrunner java at org testng testrunner run testrunner java at org testng suiterunner runtest suiterunner java at org testng suiterunner runsequentially suiterunner java at org testng suiterunner privaterun suiterunner java at org testng suiterunner run suiterunner java at org testng suiterunnerworker runsuite suiterunnerworker java at org testng suiterunnerworker run suiterunnerworker java at org testng testng runsuitessequentially testng java at org testng testng runsuiteslocally testng java at org testng testng runsuites testng java at org testng testng run testng java at org apache maven surefire testng testngexecutor run testngexecutor java at org apache maven surefire testng testngdirectorytestsuite executesingleclass testngdirectorytestsuite java at org apache maven surefire testng testngdirectorytestsuite executelazy testngdirectorytestsuite java at org apache maven surefire testng testngdirectorytestsuite execute testngdirectorytestsuite java at org apache maven surefire testng testngprovider invoke testngprovider java at org apache maven surefire booter forkedbooter invokeproviderinsameclassloader forkedbooter java at org apache maven surefire booter forkedbooter runsuitesinprocess forkedbooter java at org apache maven surefire booter forkedbooter execute forkedbooter java at org apache maven surefire booter forkedbooter main forkedbooter java are you willing to submit a pr i m willing to submit a pr
1
313,192
26,908,518,082
IssuesEvent
2023-02-06 21:15:30
louis-langholtz/PlayRho
https://api.github.com/repos/louis-langholtz/PlayRho
closed
UnderlyingValue deprecated warnings in release 1.1.0 Testbed build
Bug Testbed
### Expected/Desired Behavior or Experience: Testbed builds without warnings about `UnderlyingValue` being deprecated in release 1.1.*. ### Actual Behavior: Seeing warnings like: ``` [ 91%] Building CXX object Testbed/CMakeFiles/Testbed.dir/Framework/Main.cpp.o PlayRho/Testbed/Framework/Main.cpp:1182:28: warning: 'UnderlyingValue<unsigned short, playrho::BodyIdentifier>' is deprecated: Use to_underlying instead [-Wdeprecated-declarations] ImGui::IdContext idCtx(UnderlyingValue(b)); ^ PlayRho/PlayRho/Common/IndexingNamedType.hpp:142:3: note: 'UnderlyingValue<unsigned short, playrho::BodyIdentifier>' has been explicitly marked deprecated here [[deprecated("Use to_underlying instead")]] ^ PlayRho/Testbed/Framework/Main.cpp:1353:28: warning: 'UnderlyingValue<unsigned short, playrho::FixtureIdentifier>' is deprecated: Use to_underlying instead [-Wdeprecated-declarations] ImGui::IdContext idCtx(UnderlyingValue(fixture)); ^ PlayRho/PlayRho/Common/IndexingNamedType.hpp:142:3: note: 'UnderlyingValue<unsigned short, playrho::FixtureIdentifier>' has been explicitly marked deprecated here [[deprecated("Use to_underlying instead")]] ^ ``` ### Steps to Reproduce the Actual Behavior: ```sh git clone --recurse-submodules --branch release-1.1.0 https://github.com/louis-langholtz/PlayRho.git mkdir PlayRho-build cd PlayRho-build cmake -DPLAYRHO_BUILD_TESTBED=ON ../PlayRho make ```
1.0
UnderlyingValue deprecated warnings in release 1.1.0 Testbed build - ### Expected/Desired Behavior or Experience: Testbed builds without warnings about `UnderlyingValue` being deprecated in release 1.1.*. ### Actual Behavior: Seeing warnings like: ``` [ 91%] Building CXX object Testbed/CMakeFiles/Testbed.dir/Framework/Main.cpp.o PlayRho/Testbed/Framework/Main.cpp:1182:28: warning: 'UnderlyingValue<unsigned short, playrho::BodyIdentifier>' is deprecated: Use to_underlying instead [-Wdeprecated-declarations] ImGui::IdContext idCtx(UnderlyingValue(b)); ^ PlayRho/PlayRho/Common/IndexingNamedType.hpp:142:3: note: 'UnderlyingValue<unsigned short, playrho::BodyIdentifier>' has been explicitly marked deprecated here [[deprecated("Use to_underlying instead")]] ^ PlayRho/Testbed/Framework/Main.cpp:1353:28: warning: 'UnderlyingValue<unsigned short, playrho::FixtureIdentifier>' is deprecated: Use to_underlying instead [-Wdeprecated-declarations] ImGui::IdContext idCtx(UnderlyingValue(fixture)); ^ PlayRho/PlayRho/Common/IndexingNamedType.hpp:142:3: note: 'UnderlyingValue<unsigned short, playrho::FixtureIdentifier>' has been explicitly marked deprecated here [[deprecated("Use to_underlying instead")]] ^ ``` ### Steps to Reproduce the Actual Behavior: ```sh git clone --recurse-submodules --branch release-1.1.0 https://github.com/louis-langholtz/PlayRho.git mkdir PlayRho-build cd PlayRho-build cmake -DPLAYRHO_BUILD_TESTBED=ON ../PlayRho make ```
test
underlyingvalue deprecated warnings in release testbed build expected desired behavior or experience testbed builds without warnings about underlyingvalue being deprecated in release actual behavior seeing warnings like building cxx object testbed cmakefiles testbed dir framework main cpp o playrho testbed framework main cpp warning underlyingvalue is deprecated use to underlying instead imgui idcontext idctx underlyingvalue b playrho playrho common indexingnamedtype hpp note underlyingvalue has been explicitly marked deprecated here playrho testbed framework main cpp warning underlyingvalue is deprecated use to underlying instead imgui idcontext idctx underlyingvalue fixture playrho playrho common indexingnamedtype hpp note underlyingvalue has been explicitly marked deprecated here steps to reproduce the actual behavior sh git clone recurse submodules branch release mkdir playrho build cd playrho build cmake dplayrho build testbed on playrho make
1
71,176
15,184,767,647
IssuesEvent
2021-02-15 10:01:38
devikab2b/whites
https://api.github.com/repos/devikab2b/whites
opened
CVE-2020-8908 (Low) detected in guava-16.0.1.jar
security vulnerability
## CVE-2020-8908 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>guava-16.0.1.jar</b></p></summary> <p>Guava is a suite of core and expanded libraries that include utility classes, google's collections, io classes, and much much more. Guava has only one code dependency - javax.annotation, per the JSR-305 spec.</p> <p>Library home page: <a href="http://code.google.com/p/guava-libraries">http://code.google.com/p/guava-libraries</a></p> <p>Path to dependency file: whites/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/google/guava/guava/16.0.1/guava-16.0.1.jar</p> <p> Dependency Hierarchy: - spark-sql_2.12-3.0.1.jar (Root Library) - spark-core_2.12-3.0.1.jar - curator-recipes-2.7.1.jar - :x: **guava-16.0.1.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/devikab2b/whites/commit/4c5e641103e08d86e4119a1d3808eea3ebca1665">4c5e641103e08d86e4119a1d3808eea3ebca1665</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A temp directory creation vulnerability exist in Guava versions prior to 30.0 allowing an attacker with access to the machine to potentially access data in a temporary directory created by the Guava com.google.common.io.Files.createTempDir(). The permissions granted to the directory created default to the standard unix-like /tmp ones, leaving the files open. We recommend updating Guava to version 30.0 or later, or update to Java 7 or later, or to explicitly change the permissions after the creation of the directory if neither are possible. <p>Publish Date: 2020-12-10 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8908>CVE-2020-8908</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8908">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8908</a></p> <p>Release Date: 2020-12-10</p> <p>Fix Resolution: v30.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-8908 (Low) detected in guava-16.0.1.jar - ## CVE-2020-8908 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>guava-16.0.1.jar</b></p></summary> <p>Guava is a suite of core and expanded libraries that include utility classes, google's collections, io classes, and much much more. Guava has only one code dependency - javax.annotation, per the JSR-305 spec.</p> <p>Library home page: <a href="http://code.google.com/p/guava-libraries">http://code.google.com/p/guava-libraries</a></p> <p>Path to dependency file: whites/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/google/guava/guava/16.0.1/guava-16.0.1.jar</p> <p> Dependency Hierarchy: - spark-sql_2.12-3.0.1.jar (Root Library) - spark-core_2.12-3.0.1.jar - curator-recipes-2.7.1.jar - :x: **guava-16.0.1.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/devikab2b/whites/commit/4c5e641103e08d86e4119a1d3808eea3ebca1665">4c5e641103e08d86e4119a1d3808eea3ebca1665</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A temp directory creation vulnerability exist in Guava versions prior to 30.0 allowing an attacker with access to the machine to potentially access data in a temporary directory created by the Guava com.google.common.io.Files.createTempDir(). The permissions granted to the directory created default to the standard unix-like /tmp ones, leaving the files open. We recommend updating Guava to version 30.0 or later, or update to Java 7 or later, or to explicitly change the permissions after the creation of the directory if neither are possible. <p>Publish Date: 2020-12-10 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8908>CVE-2020-8908</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8908">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8908</a></p> <p>Release Date: 2020-12-10</p> <p>Fix Resolution: v30.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve low detected in guava jar cve low severity vulnerability vulnerable library guava jar guava is a suite of core and expanded libraries that include utility classes google s collections io classes and much much more guava has only one code dependency javax annotation per the jsr spec library home page a href path to dependency file whites pom xml path to vulnerable library home wss scanner repository com google guava guava guava jar dependency hierarchy spark sql jar root library spark core jar curator recipes jar x guava jar vulnerable library found in head commit a href found in base branch main vulnerability details a temp directory creation vulnerability exist in guava versions prior to allowing an attacker with access to the machine to potentially access data in a temporary directory created by the guava com google common io files createtempdir the permissions granted to the directory created default to the standard unix like tmp ones leaving the files open we recommend updating guava to version or later or update to java or later or to explicitly change the permissions after the creation of the directory if neither are possible publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
308,366
26,602,583,930
IssuesEvent
2023-01-23 16:47:23
microsoft/vscode-remote-release
https://api.github.com/repos/microsoft/vscode-remote-release
opened
Test: Unfiltered template list for Create Dev Container
testplan-item
Refs: https://github.com/microsoft/vscode-remote-release/issues/7629 - [ ] anyOS Complexity: 1 --- Check that the `Create Dev Container` command does not filter templates. E.g., there is no `Show All Definitions` entry in the first QuickPick, but there is a `Python` entry.
1.0
Test: Unfiltered template list for Create Dev Container - Refs: https://github.com/microsoft/vscode-remote-release/issues/7629 - [ ] anyOS Complexity: 1 --- Check that the `Create Dev Container` command does not filter templates. E.g., there is no `Show All Definitions` entry in the first QuickPick, but there is a `Python` entry.
test
test unfiltered template list for create dev container refs anyos complexity check that the create dev container command does not filter templates e g there is no show all definitions entry in the first quickpick but there is a python entry
1
258,798
22,347,997,042
IssuesEvent
2022-06-15 09:29:09
UserOfficeProject/user-office-project-issue-tracker
https://api.github.com/repos/UserOfficeProject/user-office-project-issue-tracker
closed
Create test ISIS Rapid questionnaire on dev
origin: project type: test ops: comms area: uop/stfc
We have an proposed questionnaire documented in the [project documentation](https://stfc365.sharepoint.com/:w:/r/sites/ISISProject-1102/_layouts/15/Doc.aspx?sourcedoc=%7B7C2AAE0A-BC21-4849-ADB0-964FCDDBBEF3%7D&file=Proposed+questionnaires+V+1.0.docx&action=default&mobileredirect=true&isSPOFile=1).
1.0
Create test ISIS Rapid questionnaire on dev - We have an proposed questionnaire documented in the [project documentation](https://stfc365.sharepoint.com/:w:/r/sites/ISISProject-1102/_layouts/15/Doc.aspx?sourcedoc=%7B7C2AAE0A-BC21-4849-ADB0-964FCDDBBEF3%7D&file=Proposed+questionnaires+V+1.0.docx&action=default&mobileredirect=true&isSPOFile=1).
test
create test isis rapid questionnaire on dev we have an proposed questionnaire documented in the
1
290,449
25,068,825,353
IssuesEvent
2022-11-07 10:28:10
airbytehq/airbyte
https://api.github.com/repos/airbytehq/airbyte
closed
Source AlloyDB for PostgreSQL: enable `high` test strictness level in SAT
type/enhancement area/connectors team/connectors-python test-strictness-level
## What A `test_strictness_level` field was introduced to Source Acceptance Tests (SAT). AlloyDB for PostgreSQL is a generally_available connector, we want it to have a `high` test strictness level. This will help: - maximize the SAT coverage on this connector. - document its potential weaknesses in term of test coverage. ## How 1. Migrate the existing `acceptance-test-config.yml` file to the latest configuration format. (See instructions [here](https://github.com/airbytehq/airbyte/blob/master/airbyte-integrations/bases/source-acceptance-test/README.md#L61)) 2. Enable `high` test strictness level in `acceptance-test-config.yml`. (See instructions [here](https://github.com/airbytehq/airbyte/blob/master/docs/connector-development/testing-connectors/source-acceptance-tests-reference.md#L240)) 3. Commit changes on `acceptance-test-config.yml` and open a PR 4. Run SAT with the `/test` command on the branch. 5. If tests are failing please fix the failing test or use `bypass_reason` fields to explain why a specific test can't be run.
1.0
Source AlloyDB for PostgreSQL: enable `high` test strictness level in SAT - ## What A `test_strictness_level` field was introduced to Source Acceptance Tests (SAT). AlloyDB for PostgreSQL is a generally_available connector, we want it to have a `high` test strictness level. This will help: - maximize the SAT coverage on this connector. - document its potential weaknesses in term of test coverage. ## How 1. Migrate the existing `acceptance-test-config.yml` file to the latest configuration format. (See instructions [here](https://github.com/airbytehq/airbyte/blob/master/airbyte-integrations/bases/source-acceptance-test/README.md#L61)) 2. Enable `high` test strictness level in `acceptance-test-config.yml`. (See instructions [here](https://github.com/airbytehq/airbyte/blob/master/docs/connector-development/testing-connectors/source-acceptance-tests-reference.md#L240)) 3. Commit changes on `acceptance-test-config.yml` and open a PR 4. Run SAT with the `/test` command on the branch. 5. If tests are failing please fix the failing test or use `bypass_reason` fields to explain why a specific test can't be run.
test
source alloydb for postgresql enable high test strictness level in sat what a test strictness level field was introduced to source acceptance tests sat alloydb for postgresql is a generally available connector we want it to have a high test strictness level this will help maximize the sat coverage on this connector document its potential weaknesses in term of test coverage how migrate the existing acceptance test config yml file to the latest configuration format see instructions enable high test strictness level in acceptance test config yml see instructions commit changes on acceptance test config yml and open a pr run sat with the test command on the branch if tests are failing please fix the failing test or use bypass reason fields to explain why a specific test can t be run
1
115,635
17,332,539,466
IssuesEvent
2021-07-28 05:45:23
panasalap/OpenJPEG-2.3.0_After-27841_27845_Fix
https://api.github.com/repos/panasalap/OpenJPEG-2.3.0_After-27841_27845_Fix
opened
CVE-2020-27841 (Medium) detected in openjpegv2.3.0
security vulnerability
## CVE-2020-27841 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>openjpegv2.3.0</b></p></summary> <p> <p>Official repository of the OpenJPEG project</p> <p>Library home page: <a href=https://github.com/uclouvain/openjpeg.git>https://github.com/uclouvain/openjpeg.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/panasalap/OpenJPEG-2.3.0_After-27841_27845_Fix/commit/be713d2d2bd1d1324ebad976a00e4f9fef506436">be713d2d2bd1d1324ebad976a00e4f9fef506436</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> There's a flaw in openjpeg in versions prior to 2.4.0 in src/lib/openjp2/pi.c. When an attacker is able to provide crafted input to be processed by the openjpeg encoder, this could cause an out-of-bounds read. The greatest impact from this flaw is to application availability. <p>Publish Date: 2021-01-05 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-27841>CVE-2020-27841</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://security.gentoo.org/glsa/202101-29">https://security.gentoo.org/glsa/202101-29</a></p> <p>Fix Resolution: All OpenJPEG 2 users should upgrade to the latest version # emerge --sync # emerge --ask --oneshot --verbose >=media-libs/openjpeg-2.4.02 Gentoo has discontinued support OpenJPEG 1.x and any dependent packages should now be using OpenJPEG 2 or have dropped support for the library. We recommend that users unmerge OpenJPEG 1.x # emerge --unmerge media-libs/openjpeg1 >= </p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-27841 (Medium) detected in openjpegv2.3.0 - ## CVE-2020-27841 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>openjpegv2.3.0</b></p></summary> <p> <p>Official repository of the OpenJPEG project</p> <p>Library home page: <a href=https://github.com/uclouvain/openjpeg.git>https://github.com/uclouvain/openjpeg.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/panasalap/OpenJPEG-2.3.0_After-27841_27845_Fix/commit/be713d2d2bd1d1324ebad976a00e4f9fef506436">be713d2d2bd1d1324ebad976a00e4f9fef506436</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> There's a flaw in openjpeg in versions prior to 2.4.0 in src/lib/openjp2/pi.c. When an attacker is able to provide crafted input to be processed by the openjpeg encoder, this could cause an out-of-bounds read. The greatest impact from this flaw is to application availability. <p>Publish Date: 2021-01-05 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-27841>CVE-2020-27841</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://security.gentoo.org/glsa/202101-29">https://security.gentoo.org/glsa/202101-29</a></p> <p>Fix Resolution: All OpenJPEG 2 users should upgrade to the latest version # emerge --sync # emerge --ask --oneshot --verbose >=media-libs/openjpeg-2.4.02 Gentoo has discontinued support OpenJPEG 1.x and any dependent packages should now be using OpenJPEG 2 or have dropped support for the library. We recommend that users unmerge OpenJPEG 1.x # emerge --unmerge media-libs/openjpeg1 >= </p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve medium detected in cve medium severity vulnerability vulnerable library official repository of the openjpeg project library home page a href found in head commit a href found in base branch master vulnerable source files vulnerability details there s a flaw in openjpeg in versions prior to in src lib pi c when an attacker is able to provide crafted input to be processed by the openjpeg encoder this could cause an out of bounds read the greatest impact from this flaw is to application availability publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href fix resolution all openjpeg users should upgrade to the latest version emerge sync emerge ask oneshot verbose media libs openjpeg gentoo has discontinued support openjpeg x and any dependent packages should now be using openjpeg or have dropped support for the library we recommend that users unmerge openjpeg x emerge unmerge media libs step up your open source security game with whitesource
0
257,557
22,194,358,938
IssuesEvent
2022-06-07 04:47:13
Azure/azure-sdk-for-net
https://api.github.com/repos/Azure/azure-sdk-for-net
closed
[Form Recognizer] StartRecognizeCustomFormsWithLabelsCanParseBlankPage failing in nightly runs
Client Cognitive - Form Recognizer test-reliability
Form Recognizer nightly test runs are failing with: > Error message > System.InvalidOperationException : The requested operation requires an element of type 'Number', but the target element has type 'String'. > > > Stack trace > at System.Text.Json.JsonDocument.TryGetValue(Int32 index, Single& value) > at System.Text.Json.JsonElement.GetSingle() > at Azure.AI.FormRecognizer.Models.FieldValue_internal.DeserializeFieldValue_internal(JsonElement element) in /mnt/vss/_work/1/s/sdk/formrecognizer/Azure.AI.FormRecognizer/src/Generated/Models/FieldValue_internal.Serialization.cs:line 171 > at Azure.AI.FormRecognizer.Models.DocumentResult.DeserializeDocumentResult(JsonElement element) in /mnt/vss/_work/1/s/sdk/formrecognizer/Azure.AI.FormRecognizer/src/Generated/Models/DocumentResult.Serialization.cs:line 72 > at Azure.AI.FormRecognizer.Models.V2AnalyzeResult.DeserializeV2AnalyzeResult(JsonElement element) in /mnt/vss/_work/1/s/sdk/formrecognizer/Azure.AI.FormRecognizer/src/Generated/Models/V2AnalyzeResult.Serialization.cs:line 70 > at Azure.AI.FormRecognizer.Models.AnalyzeOperationResult.DeserializeAnalyzeOperationResult(JsonElement element) in /mnt/vss/_work/1/s/sdk/formrecognizer/Azure.AI.FormRecognizer/src/Generated/Models/AnalyzeOperationResult.Serialization.cs:line 46 > at Azure.AI.FormRecognizer.FormRecognizerRestClient.GetAnalyzeFormResultAsync(Guid modelId, Guid resultId, CancellationToken cancellationToken) in /mnt/vss/_work/1/s/sdk/formrecognizer/Azure.AI.FormRecognizer/src/Generated/FormRecognizerRestClient.cs:line 397 > at Azure.AI.FormRecognizer.Models.RecognizeCustomFormsOperation.Azure.Core.IOperation<Azure.AI.FormRecognizer.Models.RecognizedFormCollection>.UpdateStateAsync(Boolean async, CancellationToken cancellationToken) in /mnt/vss/_work/1/s/sdk/formrecognizer/Azure.AI.FormRecognizer/src/FormRecognizerClient/RecognizeCustomFormsOperation.cs:line 187 > at Azure.Core.OperationInternal`1.UpdateStateAsync(Boolean async, CancellationToken cancellationToken) in /mnt/vss/_work/1/s/sdk/core/Azure.Core/src/Shared/OperationInternalOfT.cs:line 241 > at Azure.Core.OperationInternalBase.UpdateStatusAsync(Boolean async, CancellationToken cancellationToken) in /mnt/vss/_work/1/s/sdk/core/Azure.Core/src/Shared/OperationInternalBase.cs:line 201 > at Azure.Core.OperationInternalBase.UpdateStatusAsync(CancellationToken cancellationToken) in /mnt/vss/_work/1/s/sdk/core/Azure.Core/src/Shared/OperationInternalBase.cs:line 72 > at Azure.Core.OperationPoller.WaitForCompletionResponseAsync(UpdateStatusAsync updateStatusAsync, HasCompleted hasCompleted, GetRawResponse getRawResponse, Nullable`1 suggestedInterval, CancellationToken cancellationToken) in /mnt/vss/_work/1/s/sdk/core/Azure.Core/src/Shared/OperationPoller.cs:line 37 > at Azure.Core.OperationInternalBase.WaitForCompletionResponseAsync(CancellationToken cancellationToken) in /mnt/vss/_work/1/s/sdk/core/Azure.Core/src/Shared/OperationInternalBase.cs:line 111 > at Azure.Core.OperationInternal`1.WaitForCompletionAsync(CancellationToken cancellationToken) in /mnt/vss/_work/1/s/sdk/core/Azure.Core/src/Shared/OperationInternalOfT.cs:line 152 > at Azure.AI.FormRecognizer.Models.RecognizeCustomFormsOperation.WaitForCompletionAsync(CancellationToken cancellationToken) in /mnt/vss/_work/1/s/sdk/formrecognizer/Azure.AI.FormRecognizer/src/FormRecognizerClient/RecognizeCustomFormsOperation.cs:line 77 > at Azure.AI.FormRecognizer.Tests.RecognizeCustomFormsLiveTests.StartRecognizeCustomFormsWithLabelsCanParseBlankPage() in /mnt/vss/_work/1/s/sdk/formrecognizer/Azure.AI.FormRecognizer/tests/FormRecognizerClient/RecognizeCustomFormsLiveTests.cs:line 248 > at Azure.AI.FormRecognizer.Tests.RecognizeCustomFormsLiveTests.StartRecognizeCustomFormsWithLabelsCanParseBlankPage() in /mnt/vss/_work/1/s/sdk/formrecognizer/Azure.AI.FormRecognizer/tests/FormRecognizerClient/RecognizeCustomFormsLiveTests.cs:line 263 > at NUnit.Framework.Internal.TaskAwaitAdapter.GenericAdapter`1.BlockUntilCompleted() > at NUnit.Framework.Internal.MessagePumpStrategy.NoMessagePumpStrategy.WaitForCompletion(AwaitAdapter awaiter) > at NUnit.Framework.Internal.AsyncToSyncAdapter.Await(Func`1 invoke) > at NUnit.Framework.Internal.Com For more details check here: - https://dev.azure.com/azure-sdk/internal/_build/results?buildId=1507193&view=results @jsquire for notification.
1.0
[Form Recognizer] StartRecognizeCustomFormsWithLabelsCanParseBlankPage failing in nightly runs - Form Recognizer nightly test runs are failing with: > Error message > System.InvalidOperationException : The requested operation requires an element of type 'Number', but the target element has type 'String'. > > > Stack trace > at System.Text.Json.JsonDocument.TryGetValue(Int32 index, Single& value) > at System.Text.Json.JsonElement.GetSingle() > at Azure.AI.FormRecognizer.Models.FieldValue_internal.DeserializeFieldValue_internal(JsonElement element) in /mnt/vss/_work/1/s/sdk/formrecognizer/Azure.AI.FormRecognizer/src/Generated/Models/FieldValue_internal.Serialization.cs:line 171 > at Azure.AI.FormRecognizer.Models.DocumentResult.DeserializeDocumentResult(JsonElement element) in /mnt/vss/_work/1/s/sdk/formrecognizer/Azure.AI.FormRecognizer/src/Generated/Models/DocumentResult.Serialization.cs:line 72 > at Azure.AI.FormRecognizer.Models.V2AnalyzeResult.DeserializeV2AnalyzeResult(JsonElement element) in /mnt/vss/_work/1/s/sdk/formrecognizer/Azure.AI.FormRecognizer/src/Generated/Models/V2AnalyzeResult.Serialization.cs:line 70 > at Azure.AI.FormRecognizer.Models.AnalyzeOperationResult.DeserializeAnalyzeOperationResult(JsonElement element) in /mnt/vss/_work/1/s/sdk/formrecognizer/Azure.AI.FormRecognizer/src/Generated/Models/AnalyzeOperationResult.Serialization.cs:line 46 > at Azure.AI.FormRecognizer.FormRecognizerRestClient.GetAnalyzeFormResultAsync(Guid modelId, Guid resultId, CancellationToken cancellationToken) in /mnt/vss/_work/1/s/sdk/formrecognizer/Azure.AI.FormRecognizer/src/Generated/FormRecognizerRestClient.cs:line 397 > at Azure.AI.FormRecognizer.Models.RecognizeCustomFormsOperation.Azure.Core.IOperation<Azure.AI.FormRecognizer.Models.RecognizedFormCollection>.UpdateStateAsync(Boolean async, CancellationToken cancellationToken) in /mnt/vss/_work/1/s/sdk/formrecognizer/Azure.AI.FormRecognizer/src/FormRecognizerClient/RecognizeCustomFormsOperation.cs:line 187 > at Azure.Core.OperationInternal`1.UpdateStateAsync(Boolean async, CancellationToken cancellationToken) in /mnt/vss/_work/1/s/sdk/core/Azure.Core/src/Shared/OperationInternalOfT.cs:line 241 > at Azure.Core.OperationInternalBase.UpdateStatusAsync(Boolean async, CancellationToken cancellationToken) in /mnt/vss/_work/1/s/sdk/core/Azure.Core/src/Shared/OperationInternalBase.cs:line 201 > at Azure.Core.OperationInternalBase.UpdateStatusAsync(CancellationToken cancellationToken) in /mnt/vss/_work/1/s/sdk/core/Azure.Core/src/Shared/OperationInternalBase.cs:line 72 > at Azure.Core.OperationPoller.WaitForCompletionResponseAsync(UpdateStatusAsync updateStatusAsync, HasCompleted hasCompleted, GetRawResponse getRawResponse, Nullable`1 suggestedInterval, CancellationToken cancellationToken) in /mnt/vss/_work/1/s/sdk/core/Azure.Core/src/Shared/OperationPoller.cs:line 37 > at Azure.Core.OperationInternalBase.WaitForCompletionResponseAsync(CancellationToken cancellationToken) in /mnt/vss/_work/1/s/sdk/core/Azure.Core/src/Shared/OperationInternalBase.cs:line 111 > at Azure.Core.OperationInternal`1.WaitForCompletionAsync(CancellationToken cancellationToken) in /mnt/vss/_work/1/s/sdk/core/Azure.Core/src/Shared/OperationInternalOfT.cs:line 152 > at Azure.AI.FormRecognizer.Models.RecognizeCustomFormsOperation.WaitForCompletionAsync(CancellationToken cancellationToken) in /mnt/vss/_work/1/s/sdk/formrecognizer/Azure.AI.FormRecognizer/src/FormRecognizerClient/RecognizeCustomFormsOperation.cs:line 77 > at Azure.AI.FormRecognizer.Tests.RecognizeCustomFormsLiveTests.StartRecognizeCustomFormsWithLabelsCanParseBlankPage() in /mnt/vss/_work/1/s/sdk/formrecognizer/Azure.AI.FormRecognizer/tests/FormRecognizerClient/RecognizeCustomFormsLiveTests.cs:line 248 > at Azure.AI.FormRecognizer.Tests.RecognizeCustomFormsLiveTests.StartRecognizeCustomFormsWithLabelsCanParseBlankPage() in /mnt/vss/_work/1/s/sdk/formrecognizer/Azure.AI.FormRecognizer/tests/FormRecognizerClient/RecognizeCustomFormsLiveTests.cs:line 263 > at NUnit.Framework.Internal.TaskAwaitAdapter.GenericAdapter`1.BlockUntilCompleted() > at NUnit.Framework.Internal.MessagePumpStrategy.NoMessagePumpStrategy.WaitForCompletion(AwaitAdapter awaiter) > at NUnit.Framework.Internal.AsyncToSyncAdapter.Await(Func`1 invoke) > at NUnit.Framework.Internal.Com For more details check here: - https://dev.azure.com/azure-sdk/internal/_build/results?buildId=1507193&view=results @jsquire for notification.
test
startrecognizecustomformswithlabelscanparseblankpage failing in nightly runs form recognizer nightly test runs are failing with error message system invalidoperationexception the requested operation requires an element of type number but the target element has type string stack trace at system text json jsondocument trygetvalue index single value at system text json jsonelement getsingle at azure ai formrecognizer models fieldvalue internal deserializefieldvalue internal jsonelement element in mnt vss work s sdk formrecognizer azure ai formrecognizer src generated models fieldvalue internal serialization cs line at azure ai formrecognizer models documentresult deserializedocumentresult jsonelement element in mnt vss work s sdk formrecognizer azure ai formrecognizer src generated models documentresult serialization cs line at azure ai formrecognizer models jsonelement element in mnt vss work s sdk formrecognizer azure ai formrecognizer src generated models serialization cs line at azure ai formrecognizer models analyzeoperationresult deserializeanalyzeoperationresult jsonelement element in mnt vss work s sdk formrecognizer azure ai formrecognizer src generated models analyzeoperationresult serialization cs line at azure ai formrecognizer formrecognizerrestclient getanalyzeformresultasync guid modelid guid resultid cancellationtoken cancellationtoken in mnt vss work s sdk formrecognizer azure ai formrecognizer src generated formrecognizerrestclient cs line at azure ai formrecognizer models recognizecustomformsoperation azure core ioperation updatestateasync boolean async cancellationtoken cancellationtoken in mnt vss work s sdk formrecognizer azure ai formrecognizer src formrecognizerclient recognizecustomformsoperation cs line at azure core operationinternal updatestateasync boolean async cancellationtoken cancellationtoken in mnt vss work s sdk core azure core src shared operationinternaloft cs line at azure core operationinternalbase updatestatusasync boolean async cancellationtoken cancellationtoken in mnt vss work s sdk core azure core src shared operationinternalbase cs line at azure core operationinternalbase updatestatusasync cancellationtoken cancellationtoken in mnt vss work s sdk core azure core src shared operationinternalbase cs line at azure core operationpoller waitforcompletionresponseasync updatestatusasync updatestatusasync hascompleted hascompleted getrawresponse getrawresponse nullable suggestedinterval cancellationtoken cancellationtoken in mnt vss work s sdk core azure core src shared operationpoller cs line at azure core operationinternalbase waitforcompletionresponseasync cancellationtoken cancellationtoken in mnt vss work s sdk core azure core src shared operationinternalbase cs line at azure core operationinternal waitforcompletionasync cancellationtoken cancellationtoken in mnt vss work s sdk core azure core src shared operationinternaloft cs line at azure ai formrecognizer models recognizecustomformsoperation waitforcompletionasync cancellationtoken cancellationtoken in mnt vss work s sdk formrecognizer azure ai formrecognizer src formrecognizerclient recognizecustomformsoperation cs line at azure ai formrecognizer tests recognizecustomformslivetests startrecognizecustomformswithlabelscanparseblankpage in mnt vss work s sdk formrecognizer azure ai formrecognizer tests formrecognizerclient recognizecustomformslivetests cs line at azure ai formrecognizer tests recognizecustomformslivetests startrecognizecustomformswithlabelscanparseblankpage in mnt vss work s sdk formrecognizer azure ai formrecognizer tests formrecognizerclient recognizecustomformslivetests cs line at nunit framework internal taskawaitadapter genericadapter blockuntilcompleted at nunit framework internal messagepumpstrategy nomessagepumpstrategy waitforcompletion awaitadapter awaiter at nunit framework internal asynctosyncadapter await func invoke at nunit framework internal com for more details check here jsquire for notification
1
43,799
11,850,040,171
IssuesEvent
2020-03-24 16:04:02
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
opened
Incorrect LocalCacheStats metrics computation
Type: Defect
`LocalCacheStatsImpl` states that it reports the following metrics in **milli**seconds: ```java @Probe(name = CACHE_METRIC_AVERAGE_GET_TIME, unit = MS) private float averageGetTime; @Probe(name = CACHE_METRIC_AVERAGE_PUT_TIME, unit = MS) private float averagePutTime; @Probe(name = CACHE_METRIC_AVERAGE_REMOVAL_TIME, unit = MS) private float averageRemoveTime; ``` It is instantiated from `CacheStatisticsImpl` which computes them in **micro**seconds https://github.com/hazelcast/hazelcast/blob/97ff57ba92dc1fbfa77720885e7db40757999080/hazelcast/src/main/java/com/hazelcast/cache/impl/CacheStatisticsImpl.java#L195 **Expected behavior** Consistent measurement of the above metric. (However, I assume we need to check if other metrics are computed in the correct time units)
1.0
Incorrect LocalCacheStats metrics computation - `LocalCacheStatsImpl` states that it reports the following metrics in **milli**seconds: ```java @Probe(name = CACHE_METRIC_AVERAGE_GET_TIME, unit = MS) private float averageGetTime; @Probe(name = CACHE_METRIC_AVERAGE_PUT_TIME, unit = MS) private float averagePutTime; @Probe(name = CACHE_METRIC_AVERAGE_REMOVAL_TIME, unit = MS) private float averageRemoveTime; ``` It is instantiated from `CacheStatisticsImpl` which computes them in **micro**seconds https://github.com/hazelcast/hazelcast/blob/97ff57ba92dc1fbfa77720885e7db40757999080/hazelcast/src/main/java/com/hazelcast/cache/impl/CacheStatisticsImpl.java#L195 **Expected behavior** Consistent measurement of the above metric. (However, I assume we need to check if other metrics are computed in the correct time units)
non_test
incorrect localcachestats metrics computation localcachestatsimpl states that it reports the following metrics in milli seconds java probe name cache metric average get time unit ms private float averagegettime probe name cache metric average put time unit ms private float averageputtime probe name cache metric average removal time unit ms private float averageremovetime it is instantiated from cachestatisticsimpl which computes them in micro seconds expected behavior consistent measurement of the above metric however i assume we need to check if other metrics are computed in the correct time units
0
152,045
5,831,977,737
IssuesEvent
2017-05-08 20:40:52
openshift/origin
https://api.github.com/repos/openshift/origin
closed
[POST_REBASE] numeric ordering of `oc get pods` output
component/cli help-wanted kind/enhancement kind/post-rebase priority/P3
It will be nice to have a numerically ordered output for `oc get pods` : Here is the current output : ``` foobar-1-build foobar-10-build foobar-2-build foobar-3-build ```
1.0
[POST_REBASE] numeric ordering of `oc get pods` output - It will be nice to have a numerically ordered output for `oc get pods` : Here is the current output : ``` foobar-1-build foobar-10-build foobar-2-build foobar-3-build ```
non_test
numeric ordering of oc get pods output it will be nice to have a numerically ordered output for oc get pods here is the current output foobar build foobar build foobar build foobar build
0
275,043
23,890,815,248
IssuesEvent
2022-09-08 11:19:47
stores-cedcommerce/HSL-Home-page-design
https://api.github.com/repos/stores-cedcommerce/HSL-Home-page-design
closed
In address page, when the phone number is starting from 0 then the error message is coming.
Account pages Desktop Functional / bug Ready to test fixed
**Actual result:** In address page, when the phone number is starting from 0 then the error message is coming. https://www.hsldiamonds.com/account/addresses ![image](https://user-images.githubusercontent.com/102131636/188584497-3602df07-6421-46f2-b150-60dc372615c6.png) **Expected result:** The phone number validation needed to be checked.
1.0
In address page, when the phone number is starting from 0 then the error message is coming. - **Actual result:** In address page, when the phone number is starting from 0 then the error message is coming. https://www.hsldiamonds.com/account/addresses ![image](https://user-images.githubusercontent.com/102131636/188584497-3602df07-6421-46f2-b150-60dc372615c6.png) **Expected result:** The phone number validation needed to be checked.
test
in address page when the phone number is starting from then the error message is coming actual result in address page when the phone number is starting from then the error message is coming expected result the phone number validation needed to be checked
1
564,757
16,740,459,523
IssuesEvent
2021-06-11 09:08:24
codetapacademy/codetap.academy
https://api.github.com/repos/codetapacademy/codetap.academy
closed
feat: display completed percentage in the home page
Priority: High Status: Available Type: Enhancement
display completed percentage in the home page
1.0
feat: display completed percentage in the home page - display completed percentage in the home page
non_test
feat display completed percentage in the home page display completed percentage in the home page
0
177,198
13,686,269,143
IssuesEvent
2020-09-30 08:27:26
photoprism/photoprism
https://api.github.com/repos/photoprism/photoprism
closed
Hidden settings still accessible through URL route
please-test todo
Apart from the actual topic of this issue I'd like to thank all creators and maintainers of this fantastic project. I've searched some years ago for a well thought and non-cluttered self-host solution for my image and photo libraries, but unfortunately I didn't find anything at the time that matched my needs. Maybe I also stumbled over PhotoPrism, but it was in a too early development state. The development process of PhotoPrism reflects my personal workflow and is well structured, the documentation is extensive including complete user & developer guides and the communication of project goal against community requests is handled with care. And it is written in my most favorite programming language of all time 😜 I will definitely try to contribute when I have a little more time again. I've set PhotoPrism up on my server with [Traefik][] as reverse proxy and [Authelia][] as authenticator (patiently waiting with anticipation for #98) so I can host it for my family and friends. It works great and smoothly imported ~100GB in an astonishingly quick time, without any errors and a perfectly indexed gallery. Last thing I must note before going into the actual issue topic: I feel a little flattered that [my _Nord_ project is listed as design inspiration for the web UI][c], it's nice to see that it is used in so many different projects and areas. ➜ **Start to read here to skip the off-topic content** While setting up my PhotoPrism instance, I came across the `PHOTOPRISM_SETTINGS_HIDDEN` environment variable that allows to hide the “Settings“ sidebar entry that also sets settings into read-only. This is a simple way to hide UI elements that are not relevant for “normal“ users while there is no explicit user role management available, but a user that knows about the `/settings` route can also simply access the settings again. Even though the settings are read-only, I guess this is not intended and the route should not be served at all. If I'm wrong with my assumption this issue can be closed immediately because it then works as designed. [authelia]: https://www.authelia.com [c]: https://docs.photoprism.org/developer-guide/frontend/design/#colors [traefik]: https://containo.us/traefik
1.0
Hidden settings still accessible through URL route - Apart from the actual topic of this issue I'd like to thank all creators and maintainers of this fantastic project. I've searched some years ago for a well thought and non-cluttered self-host solution for my image and photo libraries, but unfortunately I didn't find anything at the time that matched my needs. Maybe I also stumbled over PhotoPrism, but it was in a too early development state. The development process of PhotoPrism reflects my personal workflow and is well structured, the documentation is extensive including complete user & developer guides and the communication of project goal against community requests is handled with care. And it is written in my most favorite programming language of all time 😜 I will definitely try to contribute when I have a little more time again. I've set PhotoPrism up on my server with [Traefik][] as reverse proxy and [Authelia][] as authenticator (patiently waiting with anticipation for #98) so I can host it for my family and friends. It works great and smoothly imported ~100GB in an astonishingly quick time, without any errors and a perfectly indexed gallery. Last thing I must note before going into the actual issue topic: I feel a little flattered that [my _Nord_ project is listed as design inspiration for the web UI][c], it's nice to see that it is used in so many different projects and areas. ➜ **Start to read here to skip the off-topic content** While setting up my PhotoPrism instance, I came across the `PHOTOPRISM_SETTINGS_HIDDEN` environment variable that allows to hide the “Settings“ sidebar entry that also sets settings into read-only. This is a simple way to hide UI elements that are not relevant for “normal“ users while there is no explicit user role management available, but a user that knows about the `/settings` route can also simply access the settings again. Even though the settings are read-only, I guess this is not intended and the route should not be served at all. If I'm wrong with my assumption this issue can be closed immediately because it then works as designed. [authelia]: https://www.authelia.com [c]: https://docs.photoprism.org/developer-guide/frontend/design/#colors [traefik]: https://containo.us/traefik
test
hidden settings still accessible through url route apart from the actual topic of this issue i d like to thank all creators and maintainers of this fantastic project i ve searched some years ago for a well thought and non cluttered self host solution for my image and photo libraries but unfortunately i didn t find anything at the time that matched my needs maybe i also stumbled over photoprism but it was in a too early development state the development process of photoprism reflects my personal workflow and is well structured the documentation is extensive including complete user developer guides and the communication of project goal against community requests is handled with care and it is written in my most favorite programming language of all time 😜 i will definitely try to contribute when i have a little more time again i ve set photoprism up on my server with as reverse proxy and as authenticator patiently waiting with anticipation for so i can host it for my family and friends it works great and smoothly imported in an astonishingly quick time without any errors and a perfectly indexed gallery last thing i must note before going into the actual issue topic i feel a little flattered that it s nice to see that it is used in so many different projects and areas ➜ start to read here to skip the off topic content while setting up my photoprism instance i came across the photoprism settings hidden environment variable that allows to hide the “settings“ sidebar entry that also sets settings into read only this is a simple way to hide ui elements that are not relevant for “normal“ users while there is no explicit user role management available but a user that knows about the settings route can also simply access the settings again even though the settings are read only i guess this is not intended and the route should not be served at all if i m wrong with my assumption this issue can be closed immediately because it then works as designed
1
100,640
30,748,075,099
IssuesEvent
2023-07-28 16:37:47
apache/daffodil-vscode
https://api.github.com/repos/apache/daffodil-vscode
closed
Merge extension and svelte package.json
build typescript extension
All the dependencies used of for `src/svelte/package.json` are all dev dependencies. So, these should be added to the main `package.json` and then remove `src/svelte/package.json`. This will allow for us to have a single `package.json` instead of multiple. This will also probably include updating some other root files and removing them from `src/svelte` since they are identical. This will include adding yarn scripts to probably compile svelte code using the root package.json
1.0
Merge extension and svelte package.json - All the dependencies used of for `src/svelte/package.json` are all dev dependencies. So, these should be added to the main `package.json` and then remove `src/svelte/package.json`. This will allow for us to have a single `package.json` instead of multiple. This will also probably include updating some other root files and removing them from `src/svelte` since they are identical. This will include adding yarn scripts to probably compile svelte code using the root package.json
non_test
merge extension and svelte package json all the dependencies used of for src svelte package json are all dev dependencies so these should be added to the main package json and then remove src svelte package json this will allow for us to have a single package json instead of multiple this will also probably include updating some other root files and removing them from src svelte since they are identical this will include adding yarn scripts to probably compile svelte code using the root package json
0
520,737
15,091,982,738
IssuesEvent
2021-02-06 17:42:36
KoderKow/twitchr
https://api.github.com/repos/KoderKow/twitchr
closed
Search Categories
Difficulty: [2] Intermediate Effort: [2] Medium Priority: [1] Low Type: ★ Enhancement
Returns a list of games or categories that match the query via name either entirely or partially. https://dev.twitch.tv/docs/api/reference#search-categories
1.0
Search Categories - Returns a list of games or categories that match the query via name either entirely or partially. https://dev.twitch.tv/docs/api/reference#search-categories
non_test
search categories returns a list of games or categories that match the query via name either entirely or partially
0
18,454
10,231,594,723
IssuesEvent
2019-08-18 10:57:28
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Could you please provide more information on the physical data encryption at rest?
Pri1 cxp product-question security-fundamentals/subsvc security/svc triaged
Hi, What is the policy on physical data encryption at rest? As in, when data is stored be it IaaS, PaaS, SaaS, does Microsoft by default do it by using their own managed keys? For example, if I create a Web App and store something on it, is it encrypted automatically at rest? Thanks, Pranav --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 16c7901b-497c-38bd-3615-defab17b572b * Version Independent ID: 9855d796-3469-8be5-c9a7-75516d9b8856 * Content: [Azure encryption overview](https://docs.microsoft.com/en-us/azure/security/fundamentals/encryption-overview) * Content Source: [articles/security/fundamentals/encryption-overview.md](https://github.com/Microsoft/azure-docs/blob/master/articles/security/fundamentals/encryption-overview.md) * Service: **security** * Sub-service: **security-fundamentals** * GitHub Login: @barclayn * Microsoft Alias: **barclayn**
True
Could you please provide more information on the physical data encryption at rest? - Hi, What is the policy on physical data encryption at rest? As in, when data is stored be it IaaS, PaaS, SaaS, does Microsoft by default do it by using their own managed keys? For example, if I create a Web App and store something on it, is it encrypted automatically at rest? Thanks, Pranav --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 16c7901b-497c-38bd-3615-defab17b572b * Version Independent ID: 9855d796-3469-8be5-c9a7-75516d9b8856 * Content: [Azure encryption overview](https://docs.microsoft.com/en-us/azure/security/fundamentals/encryption-overview) * Content Source: [articles/security/fundamentals/encryption-overview.md](https://github.com/Microsoft/azure-docs/blob/master/articles/security/fundamentals/encryption-overview.md) * Service: **security** * Sub-service: **security-fundamentals** * GitHub Login: @barclayn * Microsoft Alias: **barclayn**
non_test
could you please provide more information on the physical data encryption at rest hi what is the policy on physical data encryption at rest as in when data is stored be it iaas paas saas does microsoft by default do it by using their own managed keys for example if i create a web app and store something on it is it encrypted automatically at rest thanks pranav document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service security sub service security fundamentals github login barclayn microsoft alias barclayn
0
29,063
23,701,699,246
IssuesEvent
2022-08-29 19:39:30
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
opened
Cleanup of third-party-notices files
area-Infrastructure-installer
There are inconsistencies, potential duplicates and unclear notices. There should be a cleanup across all contributing repos.
1.0
Cleanup of third-party-notices files - There are inconsistencies, potential duplicates and unclear notices. There should be a cleanup across all contributing repos.
non_test
cleanup of third party notices files there are inconsistencies potential duplicates and unclear notices there should be a cleanup across all contributing repos
0
134,221
29,931,078,949
IssuesEvent
2023-06-22 09:31:00
eclipse/che
https://api.github.com/repos/eclipse/che
reopened
Auto-open project's `README.MD` file in VS Code
kind/enhancement severity/P2 lifecycle/stale area/editor/vscode
### Is your enhancement related to a problem? Please describe When the user starts a new DevWorkspace to work on a new project she usually starts familiarizing herself with the `README.md` file. Today, VS Code doesn't open a `README.md` file automatically. This would be one more small contribution to improving the user onboarding experience. ### Describe the solution you'd like To provide a better onboarding experience, we should make VS Code open a project's `README.md` file automatically when a new DevWorkspace is started. ### Describe alternatives you've considered Not to make this feature too intrusive, we can consider: - opening the `README.md` file only at the first start, implying that the user usually does not need to familiarize himself with the `README.md` each time they start an existing workspace; - adding a VS Code setting with which the user could tell if they are interested in that feature at all. ### Additional context _No response_
1.0
Auto-open project's `README.MD` file in VS Code - ### Is your enhancement related to a problem? Please describe When the user starts a new DevWorkspace to work on a new project she usually starts familiarizing herself with the `README.md` file. Today, VS Code doesn't open a `README.md` file automatically. This would be one more small contribution to improving the user onboarding experience. ### Describe the solution you'd like To provide a better onboarding experience, we should make VS Code open a project's `README.md` file automatically when a new DevWorkspace is started. ### Describe alternatives you've considered Not to make this feature too intrusive, we can consider: - opening the `README.md` file only at the first start, implying that the user usually does not need to familiarize himself with the `README.md` each time they start an existing workspace; - adding a VS Code setting with which the user could tell if they are interested in that feature at all. ### Additional context _No response_
non_test
auto open project s readme md file in vs code is your enhancement related to a problem please describe when the user starts a new devworkspace to work on a new project she usually starts familiarizing herself with the readme md file today vs code doesn t open a readme md file automatically this would be one more small contribution to improving the user onboarding experience describe the solution you d like to provide a better onboarding experience we should make vs code open a project s readme md file automatically when a new devworkspace is started describe alternatives you ve considered not to make this feature too intrusive we can consider opening the readme md file only at the first start implying that the user usually does not need to familiarize himself with the readme md each time they start an existing workspace adding a vs code setting with which the user could tell if they are interested in that feature at all additional context no response
0
186,931
14,426,868,379
IssuesEvent
2020-12-06 00:28:36
kalexmills/github-vet-tests-dec2020
https://api.github.com/repos/kalexmills/github-vet-tests-dec2020
closed
vtrbtf/recruitment-challenge-parse-apache-log-files: collect/src/collect/vendor/github.com/mattn/go-runewidth/runewidth_test.go; 5 LoC
fresh test tiny
Found a possible issue in [vtrbtf/recruitment-challenge-parse-apache-log-files](https://www.github.com/vtrbtf/recruitment-challenge-parse-apache-log-files) at [collect/src/collect/vendor/github.com/mattn/go-runewidth/runewidth_test.go](https://github.com/vtrbtf/recruitment-challenge-parse-apache-log-files/blob/99328d3fc1322a815f78edc9f62b4265db2491e9/collect/src/collect/vendor/github.com/mattn/go-runewidth/runewidth_test.go#L34-L38) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call which takes a reference to tbl at line 35 may start a goroutine [Click here to see the code in its original context.](https://github.com/vtrbtf/recruitment-challenge-parse-apache-log-files/blob/99328d3fc1322a815f78edc9f62b4265db2491e9/collect/src/collect/vendor/github.com/mattn/go-runewidth/runewidth_test.go#L34-L38) <details> <summary>Click here to show the 5 line(s) of Go which triggered the analyzer.</summary> ```go for _, tbl := range tables { if !sort.IsSorted(&tbl) { t.Errorf("not sorted") } } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 99328d3fc1322a815f78edc9f62b4265db2491e9
1.0
vtrbtf/recruitment-challenge-parse-apache-log-files: collect/src/collect/vendor/github.com/mattn/go-runewidth/runewidth_test.go; 5 LoC - Found a possible issue in [vtrbtf/recruitment-challenge-parse-apache-log-files](https://www.github.com/vtrbtf/recruitment-challenge-parse-apache-log-files) at [collect/src/collect/vendor/github.com/mattn/go-runewidth/runewidth_test.go](https://github.com/vtrbtf/recruitment-challenge-parse-apache-log-files/blob/99328d3fc1322a815f78edc9f62b4265db2491e9/collect/src/collect/vendor/github.com/mattn/go-runewidth/runewidth_test.go#L34-L38) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call which takes a reference to tbl at line 35 may start a goroutine [Click here to see the code in its original context.](https://github.com/vtrbtf/recruitment-challenge-parse-apache-log-files/blob/99328d3fc1322a815f78edc9f62b4265db2491e9/collect/src/collect/vendor/github.com/mattn/go-runewidth/runewidth_test.go#L34-L38) <details> <summary>Click here to show the 5 line(s) of Go which triggered the analyzer.</summary> ```go for _, tbl := range tables { if !sort.IsSorted(&tbl) { t.Errorf("not sorted") } } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 99328d3fc1322a815f78edc9f62b4265db2491e9
test
vtrbtf recruitment challenge parse apache log files collect src collect vendor github com mattn go runewidth runewidth test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message function call which takes a reference to tbl at line may start a goroutine click here to show the line s of go which triggered the analyzer go for tbl range tables if sort issorted tbl t errorf not sorted leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
1
231,272
18,757,125,338
IssuesEvent
2021-11-05 12:18:30
mozilla-mobile/focus-android
https://api.github.com/repos/mozilla-mobile/focus-android
reopened
Intermittent UI test crash - <CustomTabTest.openCustomTabInFocusTest>
crash 🔥 eng:ui-test eng:intermittent-test
### Firebase Test Run: [Firebase link](https://console.firebase.google.com/u/0/project/moz-fx-mobile-firebase-testlab/testlab/histories/bh.2b4ca2c7fdb2f0f0/matrices/6163529191173936926/executions/bs.7a1e1b9bc2148f99/testcases/1/logs) [Firebase link](https://console.firebase.google.com/u/0/project/moz-fx-mobile-firebase-testlab/testlab/histories/bh.2b4ca2c7fdb2f0f0/matrices/6163529191173936926/executions/bs.1fca83bd5319af66/testcases/1/logs) [Firebase link](https://console.firebase.google.com/u/0/project/moz-fx-mobile-firebase-testlab/testlab/histories/bh.2b4ca2c7fdb2f0f0/matrices/4718985546851459947/executions/bs.feb269f1016d22a5/testcases/1/logs) [Firebase link](https://console.firebase.google.com/u/0/project/moz-fx-mobile-firebase-testlab/testlab/histories/bh.2b4ca2c7fdb2f0f0/matrices/6292561295356878379/executions/bs.d78bbed7c9cac82b/testcases/1/logs) ### Stacktrace: 08-30 08:54:13.252: I/eportingProces(8879): --------- beginning of crash 08-30 08:54:13.261: E/AndroidRuntime(7492): FATAL EXCEPTION: main 08-30 08:54:13.261: E/AndroidRuntime(7492): Process: org.mozilla.focus.debug, PID: 7492 08-30 08:54:13.261: E/AndroidRuntime(7492): java.lang.IllegalStateException: SET SESSION: Current activity: CustomTabActivity hashcode 5303933 Other activity: MainActivity hashcode 16940721 08-30 08:54:13.261: E/AndroidRuntime(7492): at mozilla.components.browser.engine.gecko.GeckoEngineView.render(GeckoEngineView.kt:129) 08-30 08:54:13.261: E/AndroidRuntime(7492): at mozilla.components.feature.session.engine.EngineViewPresenter.renderTab(EngineViewPresenter.kt:85) 08-30 08:54:13.261: E/AndroidRuntime(7492): at mozilla.components.feature.session.engine.EngineViewPresenter.onTabToRender(EngineViewPresenter.kt:61) 08-30 08:54:13.261: E/AndroidRuntime(7492): at mozilla.components.feature.session.engine.EngineViewPresenter.access$onTabToRender(EngineViewPresenter.kt:23) 08-30 08:54:13.261: E/AndroidRuntime(7492): at mozilla.components.feature.session.engine.EngineViewPresenter$start$1$invokeSuspend$$inlined$collect$1.emit(Collect.kt:135) 08-30 08:54:13.261: E/AndroidRuntime(7492): at mozilla.components.support.ktx.kotlinx.coroutines.flow.FlowKt$ifAnyChanged$$inlined$filter$1$2.emit(Collect.kt:143) 08-30 08:54:13.261: E/AndroidRuntime(7492): at mozilla.components.feature.session.engine.EngineViewPresenter$start$1$invokeSuspend$$inlined$map$1$2.emit(Collect.kt:137) 08-30 08:54:13.261: E/AndroidRuntime(7492): at kotlinx.coroutines.flow.FlowKt__ChannelsKt.emitAllImpl$FlowKt__ChannelsKt(Channels.kt:62) 08-30 08:54:13.261: E/AndroidRuntime(7492): at kotlinx.coroutines.flow.FlowKt__ChannelsKt.access$emitAllImpl(Channels.kt:1) 08-30 08:54:13.261: E/AndroidRuntime(7492): at kotlinx.coroutines.flow.FlowKt__ChannelsKt$emitAllImpl$1.invokeSuspend(Unknown Source:14) 08-30 08:54:13.261: E/AndroidRuntime(7492): at kotlin.coroutines.jvm.internal.BaseContinuationImpl.resumeWith(ContinuationImpl.kt:33) 08-30 08:54:13.261: E/AndroidRuntime(7492): at kotlinx.coroutines.DispatchedTask.run(DispatchedTask.kt:106) 08-30 08:54:13.261: E/AndroidRuntime(7492): at android.os.Handler.handleCallback(Handler.java:873) 08-30 08:54:13.261: E/AndroidRuntime(7492): at android.os.Handler.dispatchMessage(Handler.java:99) 08-30 08:54:13.261: E/AndroidRuntime(7492): at androidx.test.espresso.base.Interrogator.loopAndInterrogate(Interrogator.java:148) 08-30 08:54:13.261: E/AndroidRuntime(7492): at androidx.test.espresso.base.UiControllerImpl.loopUntil(UiControllerImpl.java:525) 08-30 08:54:13.261: E/AndroidRuntime(7492): at androidx.test.espresso.base.UiControllerImpl.loopUntil(UiControllerImpl.java:484) 08-30 08:54:13.261: E/AndroidRuntime(7492): at androidx.test.espresso.base.UiControllerImpl.injectMotionEvent(UiControllerImpl.java:236) 08-30 08:54:13.261: E/AndroidRuntime(7492): at androidx.test.espresso.action.MotionEvents.sendUp(MotionEvents.java:162) 08-30 08:54:13.261: E/AndroidRuntime(7492): at androidx.test.espresso.action.MotionEvents.sendUp(MotionEvents.java:139) 08-30 08:54:13.261: E/AndroidRuntime(7492): at androidx.test.espresso.action.Tap.sendSingleTap(Tap.java:170) 08-30 08:54:13.261: E/AndroidRuntime(7492): at androidx.test.espresso.action.Tap.access$100(Tap.java:31) 08-30 08:54:13.261: E/AndroidRuntime(7492): at androidx.test.espresso.action.Tap$1.sendTap(Tap.java:47) 08-30 08:54:13.261: E/AndroidRuntime(7492): at androidx.test.espresso.action.GeneralClickAction.perform(GeneralClickAction.java:137) 08-30 08:54:13.261: E/AndroidRuntime(7492): at androidx.test.espresso.ViewInteraction$SingleExecutionViewAction.perform(ViewInteraction.java:366) 08-30 08:54:13.261: E/AndroidRuntime(7492): at androidx.test.espresso.ViewInteraction.doPerform(ViewInteraction.java:255) 08-30 08:54:13.261: E/AndroidRuntime(7492): at androidx.test.espresso.ViewInteraction.access$100(ViewInteraction.java:65) 08-30 08:54:13.261: E/AndroidRuntime(7492): at androidx.test.espresso.ViewInteraction$1.call(ViewInteraction.java:158) 08-30 08:54:13.261: E/AndroidRuntime(7492): at androidx.test.espresso.ViewInteraction$1.call(ViewInteraction.java:155) 08-30 08:54:13.261: E/AndroidRuntime(7492): at java.util.concurrent.FutureTask.run(FutureTask.java:266) 08-30 08:54:13.261: E/AndroidRuntime(7492): at android.os.Handler.handleCallback(Handler.java:873) 08-30 08:54:13.261: E/AndroidRuntime(7492): at android.os.Handler.dispatchMessage(Handler.java:99) 08-30 08:54:13.261: E/AndroidRuntime(7492): at android.os.Looper.loop(Looper.java:193) 08-30 08:54:13.261: E/AndroidRuntime(7492): at android.app.ActivityThread.main(ActivityThread.java:6669) 08-30 08:54:13.261: E/AndroidRuntime(7492): at java.lang.reflect.Method.invoke(Native Method) 08-30 08:54:13.261: E/AndroidRuntime(7492): at com.android.internal.os.RuntimeInit$MethodAndArgsCaller.run(RuntimeInit.java:493) 08-30 08:54:13.261: E/AndroidRuntime(7492): at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:858) 08-30 08:54:13.261: E/AndroidRuntime(7492): Caused by: java.lang.IllegalStateException: Display already acquired 08-30 08:54:13.261: E/AndroidRuntime(7492): at org.mozilla.geckoview.GeckoSession.acquireDisplay(GeckoSession.java:2348) 08-30 08:54:13.261: E/AndroidRuntime(7492): at org.mozilla.geckoview.GeckoView.setSession(GeckoView.java:460) 08-30 08:54:13.261: E/AndroidRuntime(7492): at mozilla.components.browser.engine.gecko.GeckoEngineView.render(GeckoEngineView.kt:116) 08-30 08:54:13.261: E/AndroidRuntime(7492): ... 36 more 08-30 08:54:13.265: I/Process(7492): Sending signal. PID: 7492 SIG: 9 08-30 08:54:13.267: W/BroadcastQueue(2071): Background execution not allowed: receiving Intent { act=android.intent.action.DROPBOX_ENTRY_ADDED flg=0x10 (has extras) } to com.google.android.gms/.stats.service.DropBoxEntryAddedReceiver 08-30 08:54:13.267: W/BroadcastQueue(2071): Background execution not allowed: receiving Intent { act=android.intent.action.DROPBOX_ENTRY_ADDED flg=0x10 (has extras) } to com.google.android.gms/.chimera.GmsIntentOperationService$PersistentTrustedReceiver 08-30 08:54:13.304: E/memtrack(2071): Couldn't load memtrack module 08-30 08:54:13.304: W/android.os.Debug(2071): failed to get memory consumption info: -1 08-30 08:54:13.313: W/eportingProces(8879): Unexpected CPU variant for X86 using defaults: x86 08-30 08:54:13.320: E/memtrack(2071): Couldn't load memtrack module 08-30 08:54:13.320: W/android.os.Debug(2071): failed to get memory consumption info: -1 08-30 08:54:13.347: I/Zygote(8909): seccomp disabled by setenforce 0 08-30 08:54:13.347: I/h.CrashReporte(8909): Late-enabling -Xcheck:jni 08-30 08:54:13.350: I/ActivityManager(2071): Start proc 8909:org.mozilla.focus.debug:mozilla.components.lib.crash.CrashReporter/u0a95 for activity org.mozilla.focus.debug/mozilla.components.lib.crash.prompt.CrashReporterActivity 08-30 08:54:13.357: I/BackdropActivity(6200): onResume() ### Build: 8/30 Main
2.0
Intermittent UI test crash - <CustomTabTest.openCustomTabInFocusTest> - ### Firebase Test Run: [Firebase link](https://console.firebase.google.com/u/0/project/moz-fx-mobile-firebase-testlab/testlab/histories/bh.2b4ca2c7fdb2f0f0/matrices/6163529191173936926/executions/bs.7a1e1b9bc2148f99/testcases/1/logs) [Firebase link](https://console.firebase.google.com/u/0/project/moz-fx-mobile-firebase-testlab/testlab/histories/bh.2b4ca2c7fdb2f0f0/matrices/6163529191173936926/executions/bs.1fca83bd5319af66/testcases/1/logs) [Firebase link](https://console.firebase.google.com/u/0/project/moz-fx-mobile-firebase-testlab/testlab/histories/bh.2b4ca2c7fdb2f0f0/matrices/4718985546851459947/executions/bs.feb269f1016d22a5/testcases/1/logs) [Firebase link](https://console.firebase.google.com/u/0/project/moz-fx-mobile-firebase-testlab/testlab/histories/bh.2b4ca2c7fdb2f0f0/matrices/6292561295356878379/executions/bs.d78bbed7c9cac82b/testcases/1/logs) ### Stacktrace: 08-30 08:54:13.252: I/eportingProces(8879): --------- beginning of crash 08-30 08:54:13.261: E/AndroidRuntime(7492): FATAL EXCEPTION: main 08-30 08:54:13.261: E/AndroidRuntime(7492): Process: org.mozilla.focus.debug, PID: 7492 08-30 08:54:13.261: E/AndroidRuntime(7492): java.lang.IllegalStateException: SET SESSION: Current activity: CustomTabActivity hashcode 5303933 Other activity: MainActivity hashcode 16940721 08-30 08:54:13.261: E/AndroidRuntime(7492): at mozilla.components.browser.engine.gecko.GeckoEngineView.render(GeckoEngineView.kt:129) 08-30 08:54:13.261: E/AndroidRuntime(7492): at mozilla.components.feature.session.engine.EngineViewPresenter.renderTab(EngineViewPresenter.kt:85) 08-30 08:54:13.261: E/AndroidRuntime(7492): at mozilla.components.feature.session.engine.EngineViewPresenter.onTabToRender(EngineViewPresenter.kt:61) 08-30 08:54:13.261: E/AndroidRuntime(7492): at mozilla.components.feature.session.engine.EngineViewPresenter.access$onTabToRender(EngineViewPresenter.kt:23) 08-30 08:54:13.261: E/AndroidRuntime(7492): at mozilla.components.feature.session.engine.EngineViewPresenter$start$1$invokeSuspend$$inlined$collect$1.emit(Collect.kt:135) 08-30 08:54:13.261: E/AndroidRuntime(7492): at mozilla.components.support.ktx.kotlinx.coroutines.flow.FlowKt$ifAnyChanged$$inlined$filter$1$2.emit(Collect.kt:143) 08-30 08:54:13.261: E/AndroidRuntime(7492): at mozilla.components.feature.session.engine.EngineViewPresenter$start$1$invokeSuspend$$inlined$map$1$2.emit(Collect.kt:137) 08-30 08:54:13.261: E/AndroidRuntime(7492): at kotlinx.coroutines.flow.FlowKt__ChannelsKt.emitAllImpl$FlowKt__ChannelsKt(Channels.kt:62) 08-30 08:54:13.261: E/AndroidRuntime(7492): at kotlinx.coroutines.flow.FlowKt__ChannelsKt.access$emitAllImpl(Channels.kt:1) 08-30 08:54:13.261: E/AndroidRuntime(7492): at kotlinx.coroutines.flow.FlowKt__ChannelsKt$emitAllImpl$1.invokeSuspend(Unknown Source:14) 08-30 08:54:13.261: E/AndroidRuntime(7492): at kotlin.coroutines.jvm.internal.BaseContinuationImpl.resumeWith(ContinuationImpl.kt:33) 08-30 08:54:13.261: E/AndroidRuntime(7492): at kotlinx.coroutines.DispatchedTask.run(DispatchedTask.kt:106) 08-30 08:54:13.261: E/AndroidRuntime(7492): at android.os.Handler.handleCallback(Handler.java:873) 08-30 08:54:13.261: E/AndroidRuntime(7492): at android.os.Handler.dispatchMessage(Handler.java:99) 08-30 08:54:13.261: E/AndroidRuntime(7492): at androidx.test.espresso.base.Interrogator.loopAndInterrogate(Interrogator.java:148) 08-30 08:54:13.261: E/AndroidRuntime(7492): at androidx.test.espresso.base.UiControllerImpl.loopUntil(UiControllerImpl.java:525) 08-30 08:54:13.261: E/AndroidRuntime(7492): at androidx.test.espresso.base.UiControllerImpl.loopUntil(UiControllerImpl.java:484) 08-30 08:54:13.261: E/AndroidRuntime(7492): at androidx.test.espresso.base.UiControllerImpl.injectMotionEvent(UiControllerImpl.java:236) 08-30 08:54:13.261: E/AndroidRuntime(7492): at androidx.test.espresso.action.MotionEvents.sendUp(MotionEvents.java:162) 08-30 08:54:13.261: E/AndroidRuntime(7492): at androidx.test.espresso.action.MotionEvents.sendUp(MotionEvents.java:139) 08-30 08:54:13.261: E/AndroidRuntime(7492): at androidx.test.espresso.action.Tap.sendSingleTap(Tap.java:170) 08-30 08:54:13.261: E/AndroidRuntime(7492): at androidx.test.espresso.action.Tap.access$100(Tap.java:31) 08-30 08:54:13.261: E/AndroidRuntime(7492): at androidx.test.espresso.action.Tap$1.sendTap(Tap.java:47) 08-30 08:54:13.261: E/AndroidRuntime(7492): at androidx.test.espresso.action.GeneralClickAction.perform(GeneralClickAction.java:137) 08-30 08:54:13.261: E/AndroidRuntime(7492): at androidx.test.espresso.ViewInteraction$SingleExecutionViewAction.perform(ViewInteraction.java:366) 08-30 08:54:13.261: E/AndroidRuntime(7492): at androidx.test.espresso.ViewInteraction.doPerform(ViewInteraction.java:255) 08-30 08:54:13.261: E/AndroidRuntime(7492): at androidx.test.espresso.ViewInteraction.access$100(ViewInteraction.java:65) 08-30 08:54:13.261: E/AndroidRuntime(7492): at androidx.test.espresso.ViewInteraction$1.call(ViewInteraction.java:158) 08-30 08:54:13.261: E/AndroidRuntime(7492): at androidx.test.espresso.ViewInteraction$1.call(ViewInteraction.java:155) 08-30 08:54:13.261: E/AndroidRuntime(7492): at java.util.concurrent.FutureTask.run(FutureTask.java:266) 08-30 08:54:13.261: E/AndroidRuntime(7492): at android.os.Handler.handleCallback(Handler.java:873) 08-30 08:54:13.261: E/AndroidRuntime(7492): at android.os.Handler.dispatchMessage(Handler.java:99) 08-30 08:54:13.261: E/AndroidRuntime(7492): at android.os.Looper.loop(Looper.java:193) 08-30 08:54:13.261: E/AndroidRuntime(7492): at android.app.ActivityThread.main(ActivityThread.java:6669) 08-30 08:54:13.261: E/AndroidRuntime(7492): at java.lang.reflect.Method.invoke(Native Method) 08-30 08:54:13.261: E/AndroidRuntime(7492): at com.android.internal.os.RuntimeInit$MethodAndArgsCaller.run(RuntimeInit.java:493) 08-30 08:54:13.261: E/AndroidRuntime(7492): at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:858) 08-30 08:54:13.261: E/AndroidRuntime(7492): Caused by: java.lang.IllegalStateException: Display already acquired 08-30 08:54:13.261: E/AndroidRuntime(7492): at org.mozilla.geckoview.GeckoSession.acquireDisplay(GeckoSession.java:2348) 08-30 08:54:13.261: E/AndroidRuntime(7492): at org.mozilla.geckoview.GeckoView.setSession(GeckoView.java:460) 08-30 08:54:13.261: E/AndroidRuntime(7492): at mozilla.components.browser.engine.gecko.GeckoEngineView.render(GeckoEngineView.kt:116) 08-30 08:54:13.261: E/AndroidRuntime(7492): ... 36 more 08-30 08:54:13.265: I/Process(7492): Sending signal. PID: 7492 SIG: 9 08-30 08:54:13.267: W/BroadcastQueue(2071): Background execution not allowed: receiving Intent { act=android.intent.action.DROPBOX_ENTRY_ADDED flg=0x10 (has extras) } to com.google.android.gms/.stats.service.DropBoxEntryAddedReceiver 08-30 08:54:13.267: W/BroadcastQueue(2071): Background execution not allowed: receiving Intent { act=android.intent.action.DROPBOX_ENTRY_ADDED flg=0x10 (has extras) } to com.google.android.gms/.chimera.GmsIntentOperationService$PersistentTrustedReceiver 08-30 08:54:13.304: E/memtrack(2071): Couldn't load memtrack module 08-30 08:54:13.304: W/android.os.Debug(2071): failed to get memory consumption info: -1 08-30 08:54:13.313: W/eportingProces(8879): Unexpected CPU variant for X86 using defaults: x86 08-30 08:54:13.320: E/memtrack(2071): Couldn't load memtrack module 08-30 08:54:13.320: W/android.os.Debug(2071): failed to get memory consumption info: -1 08-30 08:54:13.347: I/Zygote(8909): seccomp disabled by setenforce 0 08-30 08:54:13.347: I/h.CrashReporte(8909): Late-enabling -Xcheck:jni 08-30 08:54:13.350: I/ActivityManager(2071): Start proc 8909:org.mozilla.focus.debug:mozilla.components.lib.crash.CrashReporter/u0a95 for activity org.mozilla.focus.debug/mozilla.components.lib.crash.prompt.CrashReporterActivity 08-30 08:54:13.357: I/BackdropActivity(6200): onResume() ### Build: 8/30 Main
test
intermittent ui test crash firebase test run stacktrace i eportingproces beginning of crash e androidruntime fatal exception main e androidruntime process org mozilla focus debug pid e androidruntime java lang illegalstateexception set session current activity customtabactivity hashcode other activity mainactivity hashcode e androidruntime at mozilla components browser engine gecko geckoengineview render geckoengineview kt e androidruntime at mozilla components feature session engine engineviewpresenter rendertab engineviewpresenter kt e androidruntime at mozilla components feature session engine engineviewpresenter ontabtorender engineviewpresenter kt e androidruntime at mozilla components feature session engine engineviewpresenter access ontabtorender engineviewpresenter kt e androidruntime at mozilla components feature session engine engineviewpresenter start invokesuspend inlined collect emit collect kt e androidruntime at mozilla components support ktx kotlinx coroutines flow flowkt ifanychanged inlined filter emit collect kt e androidruntime at mozilla components feature session engine engineviewpresenter start invokesuspend inlined map emit collect kt e androidruntime at kotlinx coroutines flow flowkt channelskt emitallimpl flowkt channelskt channels kt e androidruntime at kotlinx coroutines flow flowkt channelskt access emitallimpl channels kt e androidruntime at kotlinx coroutines flow flowkt channelskt emitallimpl invokesuspend unknown source e androidruntime at kotlin coroutines jvm internal basecontinuationimpl resumewith continuationimpl kt e androidruntime at kotlinx coroutines dispatchedtask run dispatchedtask kt e androidruntime at android os handler handlecallback handler java e androidruntime at android os handler dispatchmessage handler java e androidruntime at androidx test espresso base interrogator loopandinterrogate interrogator java e androidruntime at androidx test espresso base uicontrollerimpl loopuntil uicontrollerimpl java e androidruntime at androidx test espresso base uicontrollerimpl loopuntil uicontrollerimpl java e androidruntime at androidx test espresso base uicontrollerimpl injectmotionevent uicontrollerimpl java e androidruntime at androidx test espresso action motionevents sendup motionevents java e androidruntime at androidx test espresso action motionevents sendup motionevents java e androidruntime at androidx test espresso action tap sendsingletap tap java e androidruntime at androidx test espresso action tap access tap java e androidruntime at androidx test espresso action tap sendtap tap java e androidruntime at androidx test espresso action generalclickaction perform generalclickaction java e androidruntime at androidx test espresso viewinteraction singleexecutionviewaction perform viewinteraction java e androidruntime at androidx test espresso viewinteraction doperform viewinteraction java e androidruntime at androidx test espresso viewinteraction access viewinteraction java e androidruntime at androidx test espresso viewinteraction call viewinteraction java e androidruntime at androidx test espresso viewinteraction call viewinteraction java e androidruntime at java util concurrent futuretask run futuretask java e androidruntime at android os handler handlecallback handler java e androidruntime at android os handler dispatchmessage handler java e androidruntime at android os looper loop looper java e androidruntime at android app activitythread main activitythread java e androidruntime at java lang reflect method invoke native method e androidruntime at com android internal os runtimeinit methodandargscaller run runtimeinit java e androidruntime at com android internal os zygoteinit main zygoteinit java e androidruntime caused by java lang illegalstateexception display already acquired e androidruntime at org mozilla geckoview geckosession acquiredisplay geckosession java e androidruntime at org mozilla geckoview geckoview setsession geckoview java e androidruntime at mozilla components browser engine gecko geckoengineview render geckoengineview kt e androidruntime more i process sending signal pid sig w broadcastqueue background execution not allowed receiving intent act android intent action dropbox entry added flg has extras to com google android gms stats service dropboxentryaddedreceiver w broadcastqueue background execution not allowed receiving intent act android intent action dropbox entry added flg has extras to com google android gms chimera gmsintentoperationservice persistenttrustedreceiver e memtrack couldn t load memtrack module w android os debug failed to get memory consumption info w eportingproces unexpected cpu variant for using defaults e memtrack couldn t load memtrack module w android os debug failed to get memory consumption info i zygote seccomp disabled by setenforce i h crashreporte late enabling xcheck jni i activitymanager start proc org mozilla focus debug mozilla components lib crash crashreporter for activity org mozilla focus debug mozilla components lib crash prompt crashreporteractivity i backdropactivity onresume build main
1
206,504
15,732,165,845
IssuesEvent
2021-03-29 17:57:49
elastic/kibana
https://api.github.com/repos/elastic/kibana
closed
Failing test: Jest Tests.x-pack/plugins/security/public/management/users/edit_user - CreateUserPage validates form
Team:Security failed-test
A test failed on a tracked branch ``` Error: thrown: "Exceeded timeout of 5000 ms for a test. Use jest.setTimeout(newTimeout) to increase the timeout value, if this is a long-running test." at /var/lib/jenkins/workspace/elastic+kibana+master/kibana/x-pack/plugins/security/public/management/users/edit_user/create_user_page.test.tsx:56:3 at _dispatchDescribe (/var/lib/jenkins/workspace/elastic+kibana+master/kibana/node_modules/jest-circus/build/index.js:67:26) at describe (/var/lib/jenkins/workspace/elastic+kibana+master/kibana/node_modules/jest-circus/build/index.js:30:5) at Object.<anonymous> (/var/lib/jenkins/workspace/elastic+kibana+master/kibana/x-pack/plugins/security/public/management/users/edit_user/create_user_page.test.tsx:22:1) at Runtime._execModule (/var/lib/jenkins/workspace/elastic+kibana+master/kibana/node_modules/jest-runtime/build/index.js:1299:24) at Runtime._loadModule (/var/lib/jenkins/workspace/elastic+kibana+master/kibana/node_modules/jest-runtime/build/index.js:898:12) at Runtime.requireModule (/var/lib/jenkins/workspace/elastic+kibana+master/kibana/node_modules/jest-runtime/build/index.js:746:10) at jestAdapter (/var/lib/jenkins/workspace/elastic+kibana+master/kibana/node_modules/jest-circus/build/legacy-code-todo-rewrite/jestAdapter.js:106:13) at processTicksAndRejections (internal/process/task_queues.js:93:5) at runTestInternal (/var/lib/jenkins/workspace/elastic+kibana+master/kibana/node_modules/jest-runner/build/runTest.js:380:16) at runTest (/var/lib/jenkins/workspace/elastic+kibana+master/kibana/node_modules/jest-runner/build/runTest.js:472:34) at Object.worker (/var/lib/jenkins/workspace/elastic+kibana+master/kibana/node_modules/jest-runner/build/testWorker.js:133:12) ``` First failure: [Jenkins Build](https://kibana-ci.elastic.co/job/elastic+kibana+master/12822/) <!-- kibanaCiData = {"failed-test":{"test.class":"Jest Tests.x-pack/plugins/security/public/management/users/edit_user","test.name":"CreateUserPage validates form","test.failCount":2}} -->
1.0
Failing test: Jest Tests.x-pack/plugins/security/public/management/users/edit_user - CreateUserPage validates form - A test failed on a tracked branch ``` Error: thrown: "Exceeded timeout of 5000 ms for a test. Use jest.setTimeout(newTimeout) to increase the timeout value, if this is a long-running test." at /var/lib/jenkins/workspace/elastic+kibana+master/kibana/x-pack/plugins/security/public/management/users/edit_user/create_user_page.test.tsx:56:3 at _dispatchDescribe (/var/lib/jenkins/workspace/elastic+kibana+master/kibana/node_modules/jest-circus/build/index.js:67:26) at describe (/var/lib/jenkins/workspace/elastic+kibana+master/kibana/node_modules/jest-circus/build/index.js:30:5) at Object.<anonymous> (/var/lib/jenkins/workspace/elastic+kibana+master/kibana/x-pack/plugins/security/public/management/users/edit_user/create_user_page.test.tsx:22:1) at Runtime._execModule (/var/lib/jenkins/workspace/elastic+kibana+master/kibana/node_modules/jest-runtime/build/index.js:1299:24) at Runtime._loadModule (/var/lib/jenkins/workspace/elastic+kibana+master/kibana/node_modules/jest-runtime/build/index.js:898:12) at Runtime.requireModule (/var/lib/jenkins/workspace/elastic+kibana+master/kibana/node_modules/jest-runtime/build/index.js:746:10) at jestAdapter (/var/lib/jenkins/workspace/elastic+kibana+master/kibana/node_modules/jest-circus/build/legacy-code-todo-rewrite/jestAdapter.js:106:13) at processTicksAndRejections (internal/process/task_queues.js:93:5) at runTestInternal (/var/lib/jenkins/workspace/elastic+kibana+master/kibana/node_modules/jest-runner/build/runTest.js:380:16) at runTest (/var/lib/jenkins/workspace/elastic+kibana+master/kibana/node_modules/jest-runner/build/runTest.js:472:34) at Object.worker (/var/lib/jenkins/workspace/elastic+kibana+master/kibana/node_modules/jest-runner/build/testWorker.js:133:12) ``` First failure: [Jenkins Build](https://kibana-ci.elastic.co/job/elastic+kibana+master/12822/) <!-- kibanaCiData = {"failed-test":{"test.class":"Jest Tests.x-pack/plugins/security/public/management/users/edit_user","test.name":"CreateUserPage validates form","test.failCount":2}} -->
test
failing test jest tests x pack plugins security public management users edit user createuserpage validates form a test failed on a tracked branch error thrown exceeded timeout of ms for a test use jest settimeout newtimeout to increase the timeout value if this is a long running test at var lib jenkins workspace elastic kibana master kibana x pack plugins security public management users edit user create user page test tsx at dispatchdescribe var lib jenkins workspace elastic kibana master kibana node modules jest circus build index js at describe var lib jenkins workspace elastic kibana master kibana node modules jest circus build index js at object var lib jenkins workspace elastic kibana master kibana x pack plugins security public management users edit user create user page test tsx at runtime execmodule var lib jenkins workspace elastic kibana master kibana node modules jest runtime build index js at runtime loadmodule var lib jenkins workspace elastic kibana master kibana node modules jest runtime build index js at runtime requiremodule var lib jenkins workspace elastic kibana master kibana node modules jest runtime build index js at jestadapter var lib jenkins workspace elastic kibana master kibana node modules jest circus build legacy code todo rewrite jestadapter js at processticksandrejections internal process task queues js at runtestinternal var lib jenkins workspace elastic kibana master kibana node modules jest runner build runtest js at runtest var lib jenkins workspace elastic kibana master kibana node modules jest runner build runtest js at object worker var lib jenkins workspace elastic kibana master kibana node modules jest runner build testworker js first failure
1
703,200
24,150,182,710
IssuesEvent
2022-09-21 23:16:39
tymkiwdylan/cs321_milestone_2
https://api.github.com/repos/tymkiwdylan/cs321_milestone_2
opened
As a player, I want to see my own stats so that I have access to my own information
high priority
Allow players to see their own information
1.0
As a player, I want to see my own stats so that I have access to my own information - Allow players to see their own information
non_test
as a player i want to see my own stats so that i have access to my own information allow players to see their own information
0
98,384
8,675,496,483
IssuesEvent
2018-11-30 11:03:23
shahkhan40/shantestrep
https://api.github.com/repos/shahkhan40/shantestrep
closed
fxscantest : ApiV1EnvsProjectIdIdGetQueryParamPageInvalidDatatype
fxscantest
Project : fxscantest Job : uatenv Env : uatenv Region : US_WEST Result : fail Status Code : 404 Headers : {X-Content-Type-Options=[nosniff], X-XSS-Protection=[1; mode=block], Cache-Control=[no-cache, no-store, max-age=0, must-revalidate], Pragma=[no-cache], Expires=[0], X-Frame-Options=[DENY], Set-Cookie=[SESSION=MmJkZjhlODAtNTI4ZS00MGNkLWE4MjAtZGRhMGQxNmQ0MGE0; Path=/; HttpOnly], Content-Type=[application/json;charset=UTF-8], Transfer-Encoding=[chunked], Date=[Fri, 30 Nov 2018 10:13:48 GMT]} Endpoint : http://13.56.210.25/api/v1/api/v1/envs/project-id/xYCNEayw?page=LjokR1 Request : Response : { "timestamp" : "2018-11-30T10:13:49.039+0000", "status" : 404, "error" : "Not Found", "message" : "No message available", "path" : "/api/v1/api/v1/envs/project-id/xYCNEayw" } Logs : Assertion [@StatusCode != 401] resolved-to [404 != 401] result [Passed]Assertion [@StatusCode != 404] resolved-to [404 != 404] result [Failed] --- FX Bot ---
1.0
fxscantest : ApiV1EnvsProjectIdIdGetQueryParamPageInvalidDatatype - Project : fxscantest Job : uatenv Env : uatenv Region : US_WEST Result : fail Status Code : 404 Headers : {X-Content-Type-Options=[nosniff], X-XSS-Protection=[1; mode=block], Cache-Control=[no-cache, no-store, max-age=0, must-revalidate], Pragma=[no-cache], Expires=[0], X-Frame-Options=[DENY], Set-Cookie=[SESSION=MmJkZjhlODAtNTI4ZS00MGNkLWE4MjAtZGRhMGQxNmQ0MGE0; Path=/; HttpOnly], Content-Type=[application/json;charset=UTF-8], Transfer-Encoding=[chunked], Date=[Fri, 30 Nov 2018 10:13:48 GMT]} Endpoint : http://13.56.210.25/api/v1/api/v1/envs/project-id/xYCNEayw?page=LjokR1 Request : Response : { "timestamp" : "2018-11-30T10:13:49.039+0000", "status" : 404, "error" : "Not Found", "message" : "No message available", "path" : "/api/v1/api/v1/envs/project-id/xYCNEayw" } Logs : Assertion [@StatusCode != 401] resolved-to [404 != 401] result [Passed]Assertion [@StatusCode != 404] resolved-to [404 != 404] result [Failed] --- FX Bot ---
test
fxscantest project fxscantest job uatenv env uatenv region us west result fail status code headers x content type options x xss protection cache control pragma expires x frame options set cookie content type transfer encoding date endpoint request response timestamp status error not found message no message available path api api envs project id xycneayw logs assertion resolved to result assertion resolved to result fx bot
1
69,113
7,125,166,716
IssuesEvent
2018-01-19 21:47:42
brunobuzzi/BpmFlow
https://api.github.com/repos/brunobuzzi/BpmFlow
closed
Sequence Test Error
SUnit Test bug testing
VisaApplianceProcessExecutionTest>>testExecuteProcess_visaAccepted VisaApplianceProcessExecutionTest>>testExecuteProcess_visaAcceptedPaused01 VisaApplianceProcessExecutionTest>>testExecuteProcess_visaAborted_02 (ERROR) The error is caused by previous test (testExecuteProcess_visaAcceptedPaused01)
2.0
Sequence Test Error - VisaApplianceProcessExecutionTest>>testExecuteProcess_visaAccepted VisaApplianceProcessExecutionTest>>testExecuteProcess_visaAcceptedPaused01 VisaApplianceProcessExecutionTest>>testExecuteProcess_visaAborted_02 (ERROR) The error is caused by previous test (testExecuteProcess_visaAcceptedPaused01)
test
sequence test error visaapplianceprocessexecutiontest testexecuteprocess visaaccepted visaapplianceprocessexecutiontest testexecuteprocess visaapplianceprocessexecutiontest testexecuteprocess visaaborted error the error is caused by previous test testexecuteprocess
1
22,400
3,955,872,739
IssuesEvent
2016-04-29 23:02:23
SynBioDex/libSBOLj
https://api.github.com/repos/SynBioDex/libSBOLj
closed
Complemented join/order in GenBank conversion
change needs testing
In the GenBank documentation (http://www.ddbj.nig.ac.jp/FT/full_index.html), it says the following: Note : location operator "complement" can be used in combination with either " join" or "order" within the same location; combinations of "join" and "order" within the same location (nested operators) are illegal. complement(join(2691..4571,4918..5163)) Joins regions 2691 to 4571 and 4918 to 5163, then complements the joined segments (the feature is on the strand complementary to the presented strand) join(complement(4918..5163),complement(2691..4571)) Complements regions 4918 to 5163 and 2691 to 4571, then joins the complemented segments (the feature is on the strand complementary to the presented strand) The later one is not too difficult to support in SBOL 2.0, since we now allow a SequenceAnnotation (SA) to have a list of locations, and each location in this list is allowed to have its own orientation. However, the former is tricky, since it is essentially saying that each range is “inline” but the set of ranges is “reverseComplement”. We do not have an orientation field on an SA though. One solution I can think of would be to add a level of hierarchy to complement the joined range. For example, assume the feature is defined in “originalCompDef” and the id of the top-level CD is “foo”, CD foo Component complement Definition -> complementDef SA complementSA component -> complement GenericLocation Orientation -> reverseComplement CD complementDef Component originalComp Definition -> originalCompDef SA originalSA component -> originalComp Range Start -> 2691 End -> 4571 Orientation -> inline Range Start -> 4918 End -> 5163 Orientation -> inline A simpler solution without hierarchy would be to simply add an annotation to the SA that indicates that the set of ranges should be complemented.
1.0
Complemented join/order in GenBank conversion - In the GenBank documentation (http://www.ddbj.nig.ac.jp/FT/full_index.html), it says the following: Note : location operator "complement" can be used in combination with either " join" or "order" within the same location; combinations of "join" and "order" within the same location (nested operators) are illegal. complement(join(2691..4571,4918..5163)) Joins regions 2691 to 4571 and 4918 to 5163, then complements the joined segments (the feature is on the strand complementary to the presented strand) join(complement(4918..5163),complement(2691..4571)) Complements regions 4918 to 5163 and 2691 to 4571, then joins the complemented segments (the feature is on the strand complementary to the presented strand) The later one is not too difficult to support in SBOL 2.0, since we now allow a SequenceAnnotation (SA) to have a list of locations, and each location in this list is allowed to have its own orientation. However, the former is tricky, since it is essentially saying that each range is “inline” but the set of ranges is “reverseComplement”. We do not have an orientation field on an SA though. One solution I can think of would be to add a level of hierarchy to complement the joined range. For example, assume the feature is defined in “originalCompDef” and the id of the top-level CD is “foo”, CD foo Component complement Definition -> complementDef SA complementSA component -> complement GenericLocation Orientation -> reverseComplement CD complementDef Component originalComp Definition -> originalCompDef SA originalSA component -> originalComp Range Start -> 2691 End -> 4571 Orientation -> inline Range Start -> 4918 End -> 5163 Orientation -> inline A simpler solution without hierarchy would be to simply add an annotation to the SA that indicates that the set of ranges should be complemented.
test
complemented join order in genbank conversion in the genbank documentation it says the following note location operator complement can be used in combination with either join or order within the same location combinations of join and order within the same location nested operators are illegal complement join joins regions to and to then complements the joined segments the feature is on the strand complementary to the presented strand join complement complement complements regions to and to then joins the complemented segments the feature is on the strand complementary to the presented strand the later one is not too difficult to support in sbol since we now allow a sequenceannotation sa to have a list of locations and each location in this list is allowed to have its own orientation however the former is tricky since it is essentially saying that each range is “inline” but the set of ranges is “reversecomplement” we do not have an orientation field on an sa though one solution i can think of would be to add a level of hierarchy to complement the joined range for example assume the feature is defined in “originalcompdef” and the id of the top level cd is “foo” cd foo component complement definition complementdef sa complementsa component complement genericlocation orientation reversecomplement cd complementdef component originalcomp definition originalcompdef sa originalsa component originalcomp range start end orientation inline range start end orientation inline a simpler solution without hierarchy would be to simply add an annotation to the sa that indicates that the set of ranges should be complemented
1
318,153
27,290,749,185
IssuesEvent
2023-02-23 16:26:17
harvester/harvester
https://api.github.com/repos/harvester/harvester
closed
[BUG] The namespace and name field of the created page disappear
kind/bug area/ui severity/1 reproduce/always not-require/test-plan
**Describe the bug** <!-- A clear and concise description of what the bug is. --> Import `v1.2-head` harvester in `2.7.1` rancher. ![image.png](https://images.zenhubusercontent.com/60345555ec1db310c78aa2b8/055e5953-d07e-4ad2-8190-b4fa21d0638c) **To Reproduce** Steps to reproduce the behavior: 1. Create a v2.7.1 rancher 2. Create a v1.1.2-head harvester 3. Import the `harvester` into the `rancher` 4. Go to the `harvester` page 5. Click `Volume` tab 6. Click `Create` button 7. name and namespace filed are not visible in the header of the creation page **Expected behavior** <!-- A clear and concise description of what you expected to happen. --> The name and namespace filed are displayed normally **Support bundle** <!-- You can generate a support bundle in the bottom of Harvester UI (https://docs.harvesterhci.io/v1.0/troubleshooting/harvester/#generate-a-support-bundle). It includes logs and configurations that help diagnose the issue. Tokens, passwords, and secrets are automatically removed from support bundles. If you feel it's not appropriate to share the bundle files publicly, please consider: - Wait for a developer to reach you and provide the bundle file by any secure methods. - Join our Slack community (https://rancher-users.slack.com/archives/C01GKHKAG0K) to provide the bundle. - Send the bundle to harvester-support-bundle@suse.com with the correct issue ID. --> **Environment** - Harvester ISO version: - Underlying Infrastructure (e.g. Baremetal with Dell PowerEdge R630): **Additional context** Add any other context about the problem here.
1.0
[BUG] The namespace and name field of the created page disappear - **Describe the bug** <!-- A clear and concise description of what the bug is. --> Import `v1.2-head` harvester in `2.7.1` rancher. ![image.png](https://images.zenhubusercontent.com/60345555ec1db310c78aa2b8/055e5953-d07e-4ad2-8190-b4fa21d0638c) **To Reproduce** Steps to reproduce the behavior: 1. Create a v2.7.1 rancher 2. Create a v1.1.2-head harvester 3. Import the `harvester` into the `rancher` 4. Go to the `harvester` page 5. Click `Volume` tab 6. Click `Create` button 7. name and namespace filed are not visible in the header of the creation page **Expected behavior** <!-- A clear and concise description of what you expected to happen. --> The name and namespace filed are displayed normally **Support bundle** <!-- You can generate a support bundle in the bottom of Harvester UI (https://docs.harvesterhci.io/v1.0/troubleshooting/harvester/#generate-a-support-bundle). It includes logs and configurations that help diagnose the issue. Tokens, passwords, and secrets are automatically removed from support bundles. If you feel it's not appropriate to share the bundle files publicly, please consider: - Wait for a developer to reach you and provide the bundle file by any secure methods. - Join our Slack community (https://rancher-users.slack.com/archives/C01GKHKAG0K) to provide the bundle. - Send the bundle to harvester-support-bundle@suse.com with the correct issue ID. --> **Environment** - Harvester ISO version: - Underlying Infrastructure (e.g. Baremetal with Dell PowerEdge R630): **Additional context** Add any other context about the problem here.
test
the namespace and name field of the created page disappear describe the bug import head harvester in rancher to reproduce steps to reproduce the behavior create a rancher create a head harvester import the harvester into the rancher go to the harvester page click volume tab click create button name and namespace filed are not visible in the header of the creation page expected behavior the name and namespace filed are displayed normally support bundle you can generate a support bundle in the bottom of harvester ui it includes logs and configurations that help diagnose the issue tokens passwords and secrets are automatically removed from support bundles if you feel it s not appropriate to share the bundle files publicly please consider wait for a developer to reach you and provide the bundle file by any secure methods join our slack community to provide the bundle send the bundle to harvester support bundle suse com with the correct issue id environment harvester iso version underlying infrastructure e g baremetal with dell poweredge additional context add any other context about the problem here
1
5,002
2,757,762,925
IssuesEvent
2015-04-27 16:30:13
mozilla/readability
https://api.github.com/repos/mozilla/readability
closed
Reader View is available for wordpress.com/fresh although it shouldn't
has-website-testcase should-not-be-readerable
**Reproducible on**: Beta 38.0b2-build1 (20150406174117) **Sample page showing this issue**: https://wordpress.com/fresh/
1.0
Reader View is available for wordpress.com/fresh although it shouldn't - **Reproducible on**: Beta 38.0b2-build1 (20150406174117) **Sample page showing this issue**: https://wordpress.com/fresh/
test
reader view is available for wordpress com fresh although it shouldn t reproducible on beta sample page showing this issue
1
87,999
8,129,063,602
IssuesEvent
2018-08-17 13:59:22
ValveSoftware/steam-for-linux
https://api.github.com/repos/ValveSoftware/steam-for-linux
closed
SteamCMD segmentation fault on exit
Need Retest SteamCMD reviewed
The current Linux SteamCMD executable always segfaults when quitting. **Steps to reproduce:** 1. get a clean SteamCMD installation (from tarball) 2. run it so it can update itself 3. stumble upon the following output: _Loading Steam API...Process $PID failed to shm_open /steam-ValveIPCSharedObjects5_ 4. type _quit_ when the _Steam>_ prompt appears 5. get a segmentation fault This happens on an up-to-date Debian Jessie. **Output with gdb attached:** ``` GNU gdb (Debian 7.7.1+dfsg-5) 7.7.1 Copyright (C) 2014 Free Software Foundation, Inc. License GPLv3+: GNU GPL version 3 or later <http://gnu.org/licenses/gpl.html> This is free software: you are free to change and redistribute it. There is NO WARRANTY, to the extent permitted by law. Type "show copying" and "show warranty" for details. This GDB was configured as "i586-linux-gnu". Type "show configuration" for configuration details. For bug reporting instructions, please see: <http://www.gnu.org/software/gdb/bugs/>. Find the GDB manual and other documentation resources online at: <http://www.gnu.org/software/gdb/documentation/>. For help, type "help". Type "apropos word" to search for commands related to "word"... Reading symbols from /tmp/linux32/steamcmd...(no debugging symbols found)...done. (gdb) run Starting program: /tmp/linux32/steamcmd [Thread debugging using libthread_db enabled] Using host libthread_db library "/lib/i386-linux-gnu/libthread_db.so.1". Redirecting stderr to '/home/steam/Steam/logs/stderr.txt' Looks like steam didn't shutdown cleanly, scheduling immediate update check [New Thread 0xf7766b40 (LWP 8496)] [ 0%] Checking for available updates... [New Thread 0xf74ffb40 (LWP 8497)] [----] Verifying installation... [Thread 0xf7766b40 (LWP 8496) exited] Steam Console Client (c) Valve Corporation -- type 'quit' to exit -- [New Thread 0xf5caeb40 (LWP 8499)] [New Thread 0xf5baeb40 (LWP 8500)] [New Thread 0xf58ffb40 (LWP 8501)] [New Thread 0xf57ffb40 (LWP 8502)] [New Thread 0xf55ffb40 (LWP 8503)] [Thread 0xf58ffb40 (LWP 8501) exited] [Thread 0xf57ffb40 (LWP 8502) exited] Loading Steam API...Process 8492 failed to shm_open /steam-ValveIPCSharedObjects5 OK. [Thread 0xf5baeb40 (LWP 8500) exited] [New Thread 0xf57ffb40 (LWP 8504)] Steam>quit Steam>[Thread 0xf57ffb40 (LWP 8504) exited] Program received signal SIGSEGV, Segmentation fault. 0xf6297e11 in ?? () from /tmp/linux32/steamclient.so (gdb) bt #0 0xf6297e11 in ?? () from /tmp/linux32/steamclient.so #1 0xf6407a43 in ?? () from /tmp/linux32/steamclient.so #2 0xf6402522 in ?? () from /tmp/linux32/steamclient.so #3 0xf6405010 in ?? () from /tmp/linux32/steamclient.so #4 0xf64051d8 in ?? () from /tmp/linux32/steamclient.so #5 0xf73a44c8 in ?? () from /tmp/linux32/steamconsole.so #6 0xf73b6b7d in SteamDllMainEx () from /tmp/linux32/steamconsole.so #7 0x5659c9b9 in RunSteam(int, char**, bool) () #8 0x5658ae7d in ?? () #9 0xf7d3a723 in __libc_start_main (main=0x5658a0a0, argc=1, argv=0xffffd784, init=0x5681dbe4 <__libc_csu_init>, fini=0x5681dc43 <__libc_csu_fini>, rtld_fini=0xf7febc90 <_dl_fini>, stack_end=0xffffd77c) at libc-start.c:287 #10 0x5658d639 in _start () (gdb) x/i $pc => 0xf6297e11: mov (%eax),%edx (gdb) info registers eax 0x10 16 ecx 0x0 0 edx 0xf71811c8 -149417528 ebx 0xf71b8dd0 -149189168 esp 0xffff8b00 0xffff8b00 ebp 0x56932ed0 0x56932ed0 esi 0x569bfe90 1453063824 edi 0x569bfe90 1453063824 eip 0xf6297e11 0xf6297e11 eflags 0x10202 [ IF RF ] cs 0x23 35 ss 0x2b 43 ds 0x2b 43 es 0x2b 43 fs 0x0 0 gs 0x63 99 ```
1.0
SteamCMD segmentation fault on exit - The current Linux SteamCMD executable always segfaults when quitting. **Steps to reproduce:** 1. get a clean SteamCMD installation (from tarball) 2. run it so it can update itself 3. stumble upon the following output: _Loading Steam API...Process $PID failed to shm_open /steam-ValveIPCSharedObjects5_ 4. type _quit_ when the _Steam>_ prompt appears 5. get a segmentation fault This happens on an up-to-date Debian Jessie. **Output with gdb attached:** ``` GNU gdb (Debian 7.7.1+dfsg-5) 7.7.1 Copyright (C) 2014 Free Software Foundation, Inc. License GPLv3+: GNU GPL version 3 or later <http://gnu.org/licenses/gpl.html> This is free software: you are free to change and redistribute it. There is NO WARRANTY, to the extent permitted by law. Type "show copying" and "show warranty" for details. This GDB was configured as "i586-linux-gnu". Type "show configuration" for configuration details. For bug reporting instructions, please see: <http://www.gnu.org/software/gdb/bugs/>. Find the GDB manual and other documentation resources online at: <http://www.gnu.org/software/gdb/documentation/>. For help, type "help". Type "apropos word" to search for commands related to "word"... Reading symbols from /tmp/linux32/steamcmd...(no debugging symbols found)...done. (gdb) run Starting program: /tmp/linux32/steamcmd [Thread debugging using libthread_db enabled] Using host libthread_db library "/lib/i386-linux-gnu/libthread_db.so.1". Redirecting stderr to '/home/steam/Steam/logs/stderr.txt' Looks like steam didn't shutdown cleanly, scheduling immediate update check [New Thread 0xf7766b40 (LWP 8496)] [ 0%] Checking for available updates... [New Thread 0xf74ffb40 (LWP 8497)] [----] Verifying installation... [Thread 0xf7766b40 (LWP 8496) exited] Steam Console Client (c) Valve Corporation -- type 'quit' to exit -- [New Thread 0xf5caeb40 (LWP 8499)] [New Thread 0xf5baeb40 (LWP 8500)] [New Thread 0xf58ffb40 (LWP 8501)] [New Thread 0xf57ffb40 (LWP 8502)] [New Thread 0xf55ffb40 (LWP 8503)] [Thread 0xf58ffb40 (LWP 8501) exited] [Thread 0xf57ffb40 (LWP 8502) exited] Loading Steam API...Process 8492 failed to shm_open /steam-ValveIPCSharedObjects5 OK. [Thread 0xf5baeb40 (LWP 8500) exited] [New Thread 0xf57ffb40 (LWP 8504)] Steam>quit Steam>[Thread 0xf57ffb40 (LWP 8504) exited] Program received signal SIGSEGV, Segmentation fault. 0xf6297e11 in ?? () from /tmp/linux32/steamclient.so (gdb) bt #0 0xf6297e11 in ?? () from /tmp/linux32/steamclient.so #1 0xf6407a43 in ?? () from /tmp/linux32/steamclient.so #2 0xf6402522 in ?? () from /tmp/linux32/steamclient.so #3 0xf6405010 in ?? () from /tmp/linux32/steamclient.so #4 0xf64051d8 in ?? () from /tmp/linux32/steamclient.so #5 0xf73a44c8 in ?? () from /tmp/linux32/steamconsole.so #6 0xf73b6b7d in SteamDllMainEx () from /tmp/linux32/steamconsole.so #7 0x5659c9b9 in RunSteam(int, char**, bool) () #8 0x5658ae7d in ?? () #9 0xf7d3a723 in __libc_start_main (main=0x5658a0a0, argc=1, argv=0xffffd784, init=0x5681dbe4 <__libc_csu_init>, fini=0x5681dc43 <__libc_csu_fini>, rtld_fini=0xf7febc90 <_dl_fini>, stack_end=0xffffd77c) at libc-start.c:287 #10 0x5658d639 in _start () (gdb) x/i $pc => 0xf6297e11: mov (%eax),%edx (gdb) info registers eax 0x10 16 ecx 0x0 0 edx 0xf71811c8 -149417528 ebx 0xf71b8dd0 -149189168 esp 0xffff8b00 0xffff8b00 ebp 0x56932ed0 0x56932ed0 esi 0x569bfe90 1453063824 edi 0x569bfe90 1453063824 eip 0xf6297e11 0xf6297e11 eflags 0x10202 [ IF RF ] cs 0x23 35 ss 0x2b 43 ds 0x2b 43 es 0x2b 43 fs 0x0 0 gs 0x63 99 ```
test
steamcmd segmentation fault on exit the current linux steamcmd executable always segfaults when quitting steps to reproduce get a clean steamcmd installation from tarball run it so it can update itself stumble upon the following output loading steam api process pid failed to shm open steam type quit when the steam prompt appears get a segmentation fault this happens on an up to date debian jessie output with gdb attached gnu gdb debian dfsg copyright c free software foundation inc license gnu gpl version or later this is free software you are free to change and redistribute it there is no warranty to the extent permitted by law type show copying and show warranty for details this gdb was configured as linux gnu type show configuration for configuration details for bug reporting instructions please see find the gdb manual and other documentation resources online at for help type help type apropos word to search for commands related to word reading symbols from tmp steamcmd no debugging symbols found done gdb run starting program tmp steamcmd using host libthread db library lib linux gnu libthread db so redirecting stderr to home steam steam logs stderr txt looks like steam didn t shutdown cleanly scheduling immediate update check checking for available updates verifying installation steam console client c valve corporation type quit to exit loading steam api process failed to shm open steam ok steam quit steam program received signal sigsegv segmentation fault in from tmp steamclient so gdb bt in from tmp steamclient so in from tmp steamclient so in from tmp steamclient so in from tmp steamclient so in from tmp steamclient so in from tmp steamconsole so in steamdllmainex from tmp steamconsole so in runsteam int char bool in in libc start main main argc argv init fini rtld fini stack end at libc start c in start gdb x i pc mov eax edx gdb info registers eax ecx edx ebx esp ebp esi edi eip eflags cs ss ds es fs gs
1
44,753
5,642,847,621
IssuesEvent
2017-04-06 22:12:24
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
closed
Reduce Deployment replicas in e2e tests
area/test area/workload-api/deployment kind/flake sig/apps
Most of the Deployment tests in the [Kubernetes 24-Hour Test Report](http://storage.googleapis.com/kubernetes-test-history/static/index.html) flake because the pods used by the Deployment are not ending up [running after 5 minutes](https://github.com/kubernetes/kubernetes/blob/3527153426581d48804dd0073ab254f209142ade/test/e2e/framework/util.go#L1217) which is the [timeout used for waiting pods to run](https://github.com/kubernetes/kubernetes/blob/3527153426581d48804dd0073ab254f209142ade/test/e2e/framework/util.go#L102)[1]. Funny enough there is a TODO for decreasing that timeout back to 30s. I am opening this issue for tracking the changes needed to drop unecessary replicas in Deployment e2e tests (except for stress-testing, most Deployment-related tests should use 1 replica and a combination of specific strategy parameters) but it would be useful to also discuss why 5m timeouts are not enough for waiting for pods to run. [1] Look out for "failed to wait for pods running: [timed out waiting for the condition]" @kubernetes/sig-apps-misc @kubernetes/sig-node-bugs
1.0
Reduce Deployment replicas in e2e tests - Most of the Deployment tests in the [Kubernetes 24-Hour Test Report](http://storage.googleapis.com/kubernetes-test-history/static/index.html) flake because the pods used by the Deployment are not ending up [running after 5 minutes](https://github.com/kubernetes/kubernetes/blob/3527153426581d48804dd0073ab254f209142ade/test/e2e/framework/util.go#L1217) which is the [timeout used for waiting pods to run](https://github.com/kubernetes/kubernetes/blob/3527153426581d48804dd0073ab254f209142ade/test/e2e/framework/util.go#L102)[1]. Funny enough there is a TODO for decreasing that timeout back to 30s. I am opening this issue for tracking the changes needed to drop unecessary replicas in Deployment e2e tests (except for stress-testing, most Deployment-related tests should use 1 replica and a combination of specific strategy parameters) but it would be useful to also discuss why 5m timeouts are not enough for waiting for pods to run. [1] Look out for "failed to wait for pods running: [timed out waiting for the condition]" @kubernetes/sig-apps-misc @kubernetes/sig-node-bugs
test
reduce deployment replicas in tests most of the deployment tests in the flake because the pods used by the deployment are not ending up which is the funny enough there is a todo for decreasing that timeout back to i am opening this issue for tracking the changes needed to drop unecessary replicas in deployment tests except for stress testing most deployment related tests should use replica and a combination of specific strategy parameters but it would be useful to also discuss why timeouts are not enough for waiting for pods to run look out for failed to wait for pods running kubernetes sig apps misc kubernetes sig node bugs
1
710,009
24,400,615,702
IssuesEvent
2022-10-05 00:45:51
satyrnidae/foxcraft-dungeon-loot-2
https://api.github.com/repos/satyrnidae/foxcraft-dungeon-loot-2
closed
Nerf Trader Spawners
bug high priority
Disallow traders spawned via spawner blocks from dropping copper, notifying players, or providing trades
1.0
Nerf Trader Spawners - Disallow traders spawned via spawner blocks from dropping copper, notifying players, or providing trades
non_test
nerf trader spawners disallow traders spawned via spawner blocks from dropping copper notifying players or providing trades
0
220,110
24,562,365,651
IssuesEvent
2022-10-12 21:39:24
opensearch-project/OpenSearch-Dashboards
https://api.github.com/repos/opensearch-project/OpenSearch-Dashboards
opened
CVE-2022-37599 (Medium) detected in loader-utils-1.4.0.tgz, loader-utils-2.0.2.tgz
security vulnerability
## CVE-2022-37599 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>loader-utils-1.4.0.tgz</b>, <b>loader-utils-2.0.2.tgz</b></p></summary> <p> <details><summary><b>loader-utils-1.4.0.tgz</b></p></summary> <p>utils for webpack loaders</p> <p>Library home page: <a href="https://registry.npmjs.org/loader-utils/-/loader-utils-1.4.0.tgz">https://registry.npmjs.org/loader-utils/-/loader-utils-1.4.0.tgz</a></p> <p> Dependency Hierarchy: - @osd/pm-1.0.0.tgz (Root Library) - string-replace-loader-2.3.0.tgz - :x: **loader-utils-1.4.0.tgz** (Vulnerable Library) </details> <details><summary><b>loader-utils-2.0.2.tgz</b></p></summary> <p>utils for webpack loaders</p> <p>Library home page: <a href="https://registry.npmjs.org/loader-utils/-/loader-utils-2.0.2.tgz">https://registry.npmjs.org/loader-utils/-/loader-utils-2.0.2.tgz</a></p> <p> Dependency Hierarchy: - @osd/pm-1.0.0.tgz (Root Library) - babel-loader-8.2.4.tgz - :x: **loader-utils-2.0.2.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/opensearch-project/OpenSearch-Dashboards/commit/cba076465f44b6a819e3cff7986ff4cd21a66371">cba076465f44b6a819e3cff7986ff4cd21a66371</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A Regular expression denial of service (ReDoS) flaw was found in Function interpolateName in interpolateName.js in webpack loader-utils 2.0.0 via the resourcePath variable in interpolateName.js. <p>Publish Date: 2022-10-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-37599>CVE-2022-37599</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p>
True
CVE-2022-37599 (Medium) detected in loader-utils-1.4.0.tgz, loader-utils-2.0.2.tgz - ## CVE-2022-37599 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>loader-utils-1.4.0.tgz</b>, <b>loader-utils-2.0.2.tgz</b></p></summary> <p> <details><summary><b>loader-utils-1.4.0.tgz</b></p></summary> <p>utils for webpack loaders</p> <p>Library home page: <a href="https://registry.npmjs.org/loader-utils/-/loader-utils-1.4.0.tgz">https://registry.npmjs.org/loader-utils/-/loader-utils-1.4.0.tgz</a></p> <p> Dependency Hierarchy: - @osd/pm-1.0.0.tgz (Root Library) - string-replace-loader-2.3.0.tgz - :x: **loader-utils-1.4.0.tgz** (Vulnerable Library) </details> <details><summary><b>loader-utils-2.0.2.tgz</b></p></summary> <p>utils for webpack loaders</p> <p>Library home page: <a href="https://registry.npmjs.org/loader-utils/-/loader-utils-2.0.2.tgz">https://registry.npmjs.org/loader-utils/-/loader-utils-2.0.2.tgz</a></p> <p> Dependency Hierarchy: - @osd/pm-1.0.0.tgz (Root Library) - babel-loader-8.2.4.tgz - :x: **loader-utils-2.0.2.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/opensearch-project/OpenSearch-Dashboards/commit/cba076465f44b6a819e3cff7986ff4cd21a66371">cba076465f44b6a819e3cff7986ff4cd21a66371</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A Regular expression denial of service (ReDoS) flaw was found in Function interpolateName in interpolateName.js in webpack loader-utils 2.0.0 via the resourcePath variable in interpolateName.js. <p>Publish Date: 2022-10-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-37599>CVE-2022-37599</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p>
non_test
cve medium detected in loader utils tgz loader utils tgz cve medium severity vulnerability vulnerable libraries loader utils tgz loader utils tgz loader utils tgz utils for webpack loaders library home page a href dependency hierarchy osd pm tgz root library string replace loader tgz x loader utils tgz vulnerable library loader utils tgz utils for webpack loaders library home page a href dependency hierarchy osd pm tgz root library babel loader tgz x loader utils tgz vulnerable library found in head commit a href found in base branch main vulnerability details a regular expression denial of service redos flaw was found in function interpolatename in interpolatename js in webpack loader utils via the resourcepath variable in interpolatename js publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href
0
69,320
3,297,213,755
IssuesEvent
2015-11-02 07:07:31
otavanopisto/muikku
https://api.github.com/repos/otavanopisto/muikku
reopened
Combining environment and workspace discussions js files
bug in progress priority
Environment and workspace discussions js files are not synced, meaning that they offer different user experience in environment and workspace discussions.
1.0
Combining environment and workspace discussions js files - Environment and workspace discussions js files are not synced, meaning that they offer different user experience in environment and workspace discussions.
non_test
combining environment and workspace discussions js files environment and workspace discussions js files are not synced meaning that they offer different user experience in environment and workspace discussions
0
320,617
9,784,057,599
IssuesEvent
2019-06-08 15:54:53
xiph/rav1e
https://api.github.com/repos/xiph/rav1e
closed
Very Bad Edge Artifacting on Sintel
bug high priority
Output file: http://chromashift.org/sintel.720.mp4 - look at ~11:52. Encoding parameters: Encoded via crav1e API, with: `speed=2:tile_cols_log2=1:tile_rows_log2=1:quantizer=255:bitrate=2252800:key_frame_interval=72:min_key_frame_interval=72` I can provide the exact lossless input sequence (with our zimg-based resizing, etc.) if that is useful. Just didn't yet because Big Files.
1.0
Very Bad Edge Artifacting on Sintel - Output file: http://chromashift.org/sintel.720.mp4 - look at ~11:52. Encoding parameters: Encoded via crav1e API, with: `speed=2:tile_cols_log2=1:tile_rows_log2=1:quantizer=255:bitrate=2252800:key_frame_interval=72:min_key_frame_interval=72` I can provide the exact lossless input sequence (with our zimg-based resizing, etc.) if that is useful. Just didn't yet because Big Files.
non_test
very bad edge artifacting on sintel output file look at encoding parameters encoded via api with speed tile cols tile rows quantizer bitrate key frame interval min key frame interval i can provide the exact lossless input sequence with our zimg based resizing etc if that is useful just didn t yet because big files
0
296,629
25,564,493,796
IssuesEvent
2022-11-30 13:21:14
airbytehq/airbyte
https://api.github.com/repos/airbytehq/airbyte
closed
Source Freshdesk: enable `high` test strictness level in SAT
type/enhancement area/connectors team/connectors-python test-strictness-level
## What A `test_strictness_level` field was introduced to Source Acceptance Tests (SAT). Freshdesk is a generally_available connector, we want it to have a `high` test strictness level. **This will help**: - maximize the SAT coverage on this connector. - document its potential weaknesses in term of test coverage. ## How 1. Migrate the existing `acceptance-test-config.yml` file to the latest configuration format. (See instructions [here](https://github.com/airbytehq/airbyte/blob/master/airbyte-integrations/bases/source-acceptance-test/README.md#L61)) 2. Enable `high` test strictness level in `acceptance-test-config.yml`. (See instructions [here](https://github.com/airbytehq/airbyte/blob/master/docs/connector-development/testing-connectors/source-acceptance-tests-reference.md#L240)) 3. Commit changes on `acceptance-test-config.yml` and open a PR. 4. Run SAT with the `/test` command on the branch. 5. If tests are failing please fix the failing test or use `bypass_reason` fields to explain why a specific test can't be run.
1.0
Source Freshdesk: enable `high` test strictness level in SAT - ## What A `test_strictness_level` field was introduced to Source Acceptance Tests (SAT). Freshdesk is a generally_available connector, we want it to have a `high` test strictness level. **This will help**: - maximize the SAT coverage on this connector. - document its potential weaknesses in term of test coverage. ## How 1. Migrate the existing `acceptance-test-config.yml` file to the latest configuration format. (See instructions [here](https://github.com/airbytehq/airbyte/blob/master/airbyte-integrations/bases/source-acceptance-test/README.md#L61)) 2. Enable `high` test strictness level in `acceptance-test-config.yml`. (See instructions [here](https://github.com/airbytehq/airbyte/blob/master/docs/connector-development/testing-connectors/source-acceptance-tests-reference.md#L240)) 3. Commit changes on `acceptance-test-config.yml` and open a PR. 4. Run SAT with the `/test` command on the branch. 5. If tests are failing please fix the failing test or use `bypass_reason` fields to explain why a specific test can't be run.
test
source freshdesk enable high test strictness level in sat what a test strictness level field was introduced to source acceptance tests sat freshdesk is a generally available connector we want it to have a high test strictness level this will help maximize the sat coverage on this connector document its potential weaknesses in term of test coverage how migrate the existing acceptance test config yml file to the latest configuration format see instructions enable high test strictness level in acceptance test config yml see instructions commit changes on acceptance test config yml and open a pr run sat with the test command on the branch if tests are failing please fix the failing test or use bypass reason fields to explain why a specific test can t be run
1
140,009
11,301,406,768
IssuesEvent
2020-01-17 15:31:58
stevenschader/kabanero-foundation
https://api.github.com/repos/stevenschader/kabanero-foundation
closed
SVT: TER: Kabanero Automation test execution master branch - Setup: svtcrc-838996-1.fyre.ibm.com
SVT Kabanero Test Execution Test Execution Record
"msg": "******************** crc_fyresetup_start ********************" } TASK [load_secrets : Load all secrets] ******************************************************************************************** Thursday 16 January 2020 16:30:22 -0500 (0:00:00.085) 0:00:00.382 ****** ok: [localhost] TASK [crc_fyrevm : include_tasks] ************************************************************************************************* Thursday 16 January 2020 16:30:22 -0500 (0:00:00.088) 0:00:00.470 ****** included: /home/nest/git/icpa-system-test/automation/ansible-playbooks/roles/crc_fyrevm/tasks/crc_fyrevm.yml for localhost TASK [crc_fyrevm : set_fact] ****************************************************************************************************** Thursday 16 January 2020 16:30:23 -0500 (0:00:00.214) 0:00:00.685 ****** ok: [localhost] TASK [crc_fyrevm : check status of the named cluster] ***************************************************************************** Thursday 16 January 2020 16:30:23 -0500 (0:00:00.080) 0:00:00.766 ****** changed: [localhost -> localhost] TASK [crc_fyrevm : debug] ********************************************************************************************************* Thursday 16 January 2020 16:30:24 -0500 (0:00:01.489) 0:00:02.256 ****** skipping: [localhost] TASK [crc_fyrevm : get the public ssh id] ***************************************************************************************** Thursday 16 January 2020 16:30:24 -0500 (0:00:00.088) 0:00:02.345 ****** changed: [localhost -> localhost] TASK [crc_fyrevm : create input json file] **************************************************************************************** Thursday 16 January 2020 16:30:25 -0500 (0:00:00.500) 0:00:02.845 ****** changed: [localhost] TASK [crc_fyrevm : Create Fyre stack] ********************************************************************************************* Thursday 16 January 2020 16:30:26 -0500 (0:00:01.020) 0:00:03.866 ****** changed: [localhost] TASK [crc_fyrevm : remove temp json] ********************************************************************************************** Thursday 16 January 2020 16:30:28 -0500 (0:00:01.821) 0:00:05.687 ****** changed: [localhost -> localhost] TASK [crc_fyrevm : check Fyre request status] ************************************************************************************* Thursday 16 January 2020 16:30:28 -0500 (0:00:00.545) 0:00:06.233 ****** changed: [localhost] TASK [crc_fyrevm : debug] ********************************************************************************************************* Thursday 16 January 2020 16:30:29 -0500 (0:00:01.066) 0:00:07.299 ****** ok: [localhost] => { "fyreStatus.stdout": { "request": [ { "cluster_name": "", "created": "2020-01-16 13:30:27", "displayname": "steven r. schader", "error_details": "0", "id": "2750591", "status": "building", "type": "build", "updated": "2020-01-16 13:30:28", "username": "schader" } ] } } TASK [crc_fyrevm : pause] ********************************************************************************************************* Thursday 16 January 2020 16:30:29 -0500 (0:00:00.035) 0:00:07.335 ****** Pausing for 120 seconds (ctrl+C then 'C' = continue early, ctrl+C then 'A' = abort) ok: [localhost] TASK [crc_fyrevm : get Fyre request status] *************************************************************************************** Thursday 16 January 2020 16:32:29 -0500 (0:02:00.071) 0:02:07.407 ****** FAILED - RETRYING: get Fyre request status (999 retries left). FAILED - RETRYING: get Fyre request status (998 retries left). FAILED - RETRYING: get Fyre request status (997 retries left). FAILED - RETRYING: get Fyre request status (996 retries left). FAILED - RETRYING: get Fyre request status (995 retries left). FAILED - RETRYING: get Fyre request status (994 retries left). FAILED - RETRYING: get Fyre request status (993 retries left). FAILED - RETRYING: get Fyre request status (992 retries left). FAILED - RETRYING: get Fyre request status (991 retries left). FAILED - RETRYING: get Fyre request status (990 retries left). FAILED - RETRYING: get Fyre request status (989 retries left). FAILED - RETRYING: get Fyre request status (988 retries left). FAILED - RETRYING: get Fyre request status (987 retries left). FAILED - RETRYING: get Fyre request status (986 retries left). FAILED - RETRYING: get Fyre request status (985 retries left). FAILED - RETRYING: get Fyre request status (984 retries left). FAILED - RETRYING: get Fyre request status (983 retries left). FAILED - RETRYING: get Fyre request status (982 retries left). FAILED - RETRYING: get Fyre request status (981 retries left). FAILED - RETRYING: get Fyre request status (980 retries left). FAILED - RETRYING: get Fyre request status (979 retries left). FAILED - RETRYING: get Fyre request status (978 retries left). FAILED - RETRYING: get Fyre request status (977 retries left). FAILED - RETRYING: get Fyre request status (976 retries left). FAILED - RETRYING: get Fyre request status (975 retries left). FAILED - RETRYING: get Fyre request status (974 retries left). FAILED - RETRYING: get Fyre request status (973 retries left). FAILED - RETRYING: get Fyre request status (972 retries left). FAILED - RETRYING: get Fyre request status (971 retries left). FAILED - RETRYING: get Fyre request status (970 retries left). FAILED - RETRYING: get Fyre request status (969 retries left). FAILED - RETRYING: get Fyre request status (968 retries left). FAILED - RETRYING: get Fyre request status (967 retries left). FAILED - RETRYING: get Fyre request status (966 retries left). FAILED - RETRYING: get Fyre request status (965 retries left). FAILED - RETRYING: get Fyre request status (964 retries left). FAILED - RETRYING: get Fyre request status (963 retries left). FAILED - RETRYING: get Fyre request status (962 retries left). FAILED - RETRYING: get Fyre request status (961 retries left). FAILED - RETRYING: get Fyre request status (960 retries left). FAILED - RETRYING: get Fyre request status (959 retries left). FAILED - RETRYING: get Fyre request status (958 retries left). FAILED - RETRYING: get Fyre request status (957 retries left). FAILED - RETRYING: get Fyre request status (956 retries left). FAILED - RETRYING: get Fyre request status (955 retries left). FAILED - RETRYING: get Fyre request status (954 retries left). FAILED - RETRYING: get Fyre request status (953 retries left). FAILED - RETRYING: get Fyre request status (952 retries left). FAILED - RETRYING: get Fyre request status (951 retries left). FAILED - RETRYING: get Fyre request status (950 retries left). FAILED - RETRYING: get Fyre request status (949 retries left). FAILED - RETRYING: get Fyre request status (948 retries left). FAILED - RETRYING: get Fyre request status (947 retries left). FAILED - RETRYING: get Fyre request status (946 retries left). changed: [localhost] TASK [crc_fyrevm : check Fyre request status] ************************************************************************************* Thursday 16 January 2020 16:37:52 -0500 (0:05:22.785) 0:07:30.192 ****** skipping: [localhost] TASK [crc_fyrevm : check the host for a active ssh] ******************************************************************************* Thursday 16 January 2020 16:37:52 -0500 (0:00:00.041) 0:07:30.234 ****** changed: [localhost -> localhost] TASK [crc_fyrevm : remove new host from localhost known_hosts] ******************************************************************** Thursday 16 January 2020 16:37:53 -0500 (0:00:00.729) 0:07:30.963 ****** changed: [localhost -> localhost] TASK [crc_fyrevm : add host to known_hosts] *************************************************************************************** Thursday 16 January 2020 16:37:53 -0500 (0:00:00.333) 0:07:31.297 ****** changed: [localhost -> localhost] TASK [crc_fyrevm : add_host] ****************************************************************************************************** Thursday 16 January 2020 16:37:54 -0500 (0:00:00.425) 0:07:31.722 ****** changed: [localhost] TASK [crc_fyrevm : create host inventory file for debugging] ********************************************************************** Thursday 16 January 2020 16:37:54 -0500 (0:00:00.059) 0:07:31.782 ****** changed: [localhost -> localhost] TASK [crc_fyrevm : create plain file with host] *********************************************************************************** Thursday 16 January 2020 16:37:54 -0500 (0:00:00.575) 0:07:32.358 ****** changed: [localhost -> localhost] TASK [crc_fyrevm : create plain file with cluster name] *************************************************************************** Thursday 16 January 2020 16:37:55 -0500 (0:00:00.590) 0:07:32.948 ****** changed: [localhost -> localhost] PLAY [kabanero_crc] *************************************************************************************************************** TASK [load_secrets : Load all secrets] ******************************************************************************************** Thursday 16 January 2020 16:37:55 -0500 (0:00:00.647) 0:07:33.596 ****** ok: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_host_prereqs : set timezone to to regional tz] ************************************************************************** Thursday 16 January 2020 16:37:56 -0500 (0:00:00.090) 0:07:33.686 ****** changed: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_host_prereqs : check if jq is installed] ******************************************************************************** Thursday 16 January 2020 16:37:57 -0500 (0:00:01.126) 0:07:34.813 ****** ok: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_host_prereqs : Copy JQ to VM] ******************************************************************************************* Thursday 16 January 2020 16:37:57 -0500 (0:00:00.473) 0:07:35.287 ****** changed: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_host_prereqs : Change jq permissions] *********************************************************************************** Thursday 16 January 2020 16:37:58 -0500 (0:00:01.062) 0:07:36.349 ****** changed: [svtcrc-838996-1.fyre.ibm.com] PLAY [kabanero_crc] *************************************************************************************************************** TASK [Gathering Facts] ************************************************************************************************************ Thursday 16 January 2020 16:37:59 -0500 (0:00:00.526) 0:07:36.876 ****** ok: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_user : include_tasks] *************************************************************************************************** Thursday 16 January 2020 16:38:00 -0500 (0:00:01.732) 0:07:38.609 ****** included: /home/nest/git/icpa-system-test/automation/ansible-playbooks/roles/crc_user/tasks/crc_user.yml for svtcrc-838996-1.fyre.ibm.com TASK [crc_user : Ensure User is Present] ****************************************************************************************** Thursday 16 January 2020 16:38:01 -0500 (0:00:00.252) 0:07:38.861 ****** changed: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_user : Add user to sudoers] ********************************************************************************************* Thursday 16 January 2020 16:38:02 -0500 (0:00:01.399) 0:07:40.260 ****** changed: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_user : update user hard / soft ulimit nofile] *************************************************************************** Thursday 16 January 2020 16:38:03 -0500 (0:00:00.841) 0:07:41.101 ****** changed: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_user : authorized_key] ************************************************************************************************** Thursday 16 January 2020 16:38:04 -0500 (0:00:00.733) 0:07:41.834 ****** changed: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_user : add bin to path] ************************************************************************************************* Thursday 16 January 2020 16:38:05 -0500 (0:00:00.894) 0:07:42.729 ****** changed: [svtcrc-838996-1.fyre.ibm.com] TASK [log : include_tasks] ******************************************************************************************************** Thursday 16 January 2020 16:38:05 -0500 (0:00:00.570) 0:07:43.300 ****** included: /home/nest/git/icpa-system-test/automation/ansible-playbooks/roles/log/tasks/log.yml for svtcrc-838996-1.fyre.ibm.com TASK [log : debug] **************************************************************************************************************** Thursday 16 January 2020 16:38:05 -0500 (0:00:00.192) 0:07:43.492 ****** ok: [svtcrc-838996-1.fyre.ibm.com] => { "msg": "******************** crc_fyresetup_end ********************" } PLAY [kabanero_crc] *************************************************************************************************************** TASK [Gathering Facts] ************************************************************************************************************ Thursday 16 January 2020 16:38:06 -0500 (0:00:00.164) 0:07:43.657 ****** ok: [svtcrc-838996-1.fyre.ibm.com] TASK [log : include_tasks] ******************************************************************************************************** Thursday 16 January 2020 16:38:08 -0500 (0:00:02.220) 0:07:45.877 ****** included: /home/nest/git/icpa-system-test/automation/ansible-playbooks/roles/log/tasks/log.yml for svtcrc-838996-1.fyre.ibm.com TASK [log : debug] **************************************************************************************************************** Thursday 16 January 2020 16:38:08 -0500 (0:00:00.171) 0:07:46.049 ****** ok: [svtcrc-838996-1.fyre.ibm.com] => { "msg": "******************** crc_install_start ********************" } TASK [load_secrets : Load all secrets] ******************************************************************************************** Thursday 16 January 2020 16:38:08 -0500 (0:00:00.113) 0:07:46.162 ****** ok: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_install : include_tasks] ************************************************************************************************ Thursday 16 January 2020 16:38:08 -0500 (0:00:00.118) 0:07:46.281 ****** included: /home/nest/git/icpa-system-test/automation/ansible-playbooks/roles/crc_install/tasks/crc_install.yml for svtcrc-838996-1.fyre.ibm.com TASK [crc_install : check if crc is installed] ************************************************************************************ Thursday 16 January 2020 16:38:08 -0500 (0:00:00.100) 0:07:46.382 ****** ok: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_install : Download CRC Archive] ***************************************************************************************** Thursday 16 January 2020 16:38:09 -0500 (0:00:00.510) 0:07:46.893 ****** changed: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_install : archive dir structure] **************************************************************************************** Thursday 16 January 2020 16:39:03 -0500 (0:00:54.692) 0:08:41.585 ****** changed: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_install : debug] ******************************************************************************************************** Thursday 16 January 2020 16:39:12 -0500 (0:00:08.280) 0:08:49.866 ****** ok: [svtcrc-838996-1.fyre.ibm.com] => { "tarlist.stdout_lines": [ "crc-linux-1.4.0-amd64/", "crc-linux-1.4.0-amd64/LICENSE", "crc-linux-1.4.0-amd64/doc.pdf", "crc-linux-1.4.0-amd64/crc" ] } TASK [crc_install : install crc] ************************************************************************************************** Thursday 16 January 2020 16:39:12 -0500 (0:00:00.151) 0:08:50.018 ****** changed: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_install : home bin] ***************************************************************************************************** Thursday 16 January 2020 16:39:41 -0500 (0:00:28.921) 0:09:18.945 ****** changed: [svtcrc-838996-1.fyre.ibm.com] => (item=/home/kevin/bin) TASK [crc_install : link crc] ***************************************************************************************************** Thursday 16 January 2020 16:39:42 -0500 (0:00:00.757) 0:09:19.702 ****** changed: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_install : Install NetworkManager FireFox] ******************************************************************************* Thursday 16 January 2020 16:39:42 -0500 (0:00:00.652) 0:09:20.355 ****** changed: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_install : enable and start NetworkManager] ****************************************************************************** Thursday 16 January 2020 16:41:31 -0500 (0:01:48.925) 0:11:09.280 ****** changed: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_start : include_tasks] ************************************************************************************************** Thursday 16 January 2020 16:41:33 -0500 (0:00:01.501) 0:11:10.782 ****** included: /home/nest/git/icpa-system-test/automation/ansible-playbooks/roles/crc_start/tasks/crc_start.yml for svtcrc-838996-1.fyre.ibm.com TASK [crc_start : check if .crc is created] *************************************************************************************** Thursday 16 January 2020 16:41:33 -0500 (0:00:00.262) 0:11:11.044 ****** ok: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_start : crc stop] ******************************************************************************************************* Thursday 16 January 2020 16:41:34 -0500 (0:00:00.660) 0:11:11.705 ****** skipping: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_start : crcstop stdout] ************************************************************************************************* Thursday 16 January 2020 16:41:34 -0500 (0:00:00.135) 0:11:11.840 ****** skipping: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_start : crc delete] ***************************************************************************************************** Thursday 16 January 2020 16:41:34 -0500 (0:00:00.056) 0:11:11.897 ****** skipping: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_start : crc delete stdout] ********************************************************************************************** Thursday 16 January 2020 16:41:34 -0500 (0:00:00.107) 0:11:12.005 ****** skipping: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_start : crc setup] ****************************************************************************************************** Thursday 16 January 2020 16:41:34 -0500 (0:00:00.043) 0:11:12.048 ****** changed: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_start : crc setup stdout] *********************************************************************************************** Thursday 16 January 2020 16:42:04 -0500 (0:00:29.895) 0:11:41.944 ****** ok: [svtcrc-838996-1.fyre.ibm.com] => { "msg": [ "level=info msg=\"Checking if oc binary is cached\"", "level=info msg=\"Caching oc binary\"", "level=info msg=\"Checking if CRC bundle is cached in '$HOME/.crc'\"", "level=info msg=\"Unpacking bundle from the CRC binary\"", "level=info msg=\"Checking if running as non-root\"", "level=info msg=\"Checking if Virtualization is enabled\"", "level=info msg=\"Checking if KVM is enabled\"", "level=info msg=\"Checking if libvirt is installed\"", "level=info msg=\"Installing libvirt service and dependencies\"", "level=info msg=\"Will use root access: install virtualization related packages\"", "level=info msg=\"Checking if user is part of libvirt group\"", "level=info msg=\"Adding user to libvirt group\"", "level=info msg=\"Will use root access: add user to libvirt group\"", "level=info msg=\"Checking if libvirt is enabled\"", "level=info msg=\"Checking if libvirt daemon is running\"", "level=info msg=\"Starting libvirt service\"", "level=info msg=\"Will use root access: start libvirtd service\"", "level=info msg=\"Checking if a supported libvirt version is installed\"", "level=info msg=\"Checking if crc-driver-libvirt is installed\"", "level=info msg=\"Installing crc-driver-libvirt\"", "level=info msg=\"Checking for obsolete crc-driver-libvirt\"", "level=info msg=\"Checking if libvirt 'crc' network is available\"", "level=info msg=\"Setting up libvirt 'crc' network\"", "level=info msg=\"Checking if libvirt 'crc' network is active\"", "level=info msg=\"Starting libvirt 'crc' network\"", "level=info msg=\"Checking if NetworkManager is installed\"", "level=info msg=\"Checking if NetworkManager service is running\"", "level=info msg=\"Checking if /etc/NetworkManager/conf.d/crc-nm-dnsmasq.conf exists\"", "level=info msg=\"Writing Network Manager config for crc\"", "level=info msg=\"Will use root access: write NetworkManager config in /etc/NetworkManager/conf.d/crc-nm-dnsmasq.conf\"", "level=info msg=\"Will use root access: execute systemctl daemon-reload command\"", "level=info msg=\"Will use root access: execute systemctl stop/start command\"", "level=info msg=\"Checking if /etc/NetworkManager/dnsmasq.d/crc.conf exists\"", "level=info msg=\"Writing dnsmasq config for crc\"", "level=info msg=\"Will use root access: write dnsmasq configuration in /etc/NetworkManager/dnsmasq.d/crc.conf\"", "level=info msg=\"Will use root access: execute systemctl daemon-reload command\"", "level=info msg=\"Will use root access: execute systemctl stop/start command\"", "Setup is complete, you can now run 'crc start' to start the OpenShift cluster" ] } TASK [crc_start : crc pull secret] ************************************************************************************************ Thursday 16 January 2020 16:42:04 -0500 (0:00:00.052) 0:11:41.996 ****** changed: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_start : crc start] ****************************************************************************************************** Thursday 16 January 2020 16:42:05 -0500 (0:00:00.938) 0:11:42.934 ****** changed: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_start : crc start stdout] *********************************************************************************************** Thursday 16 January 2020 16:56:32 -0500 (0:14:27.687) 0:26:10.621 ****** ok: [svtcrc-838996-1.fyre.ibm.com] => { "msg": [ "level=info msg=\"Checking if oc binary is cached\"", "level=info msg=\"Checking if running as non-root\"", "level=info msg=\"Checking if Virtualization is enabled\"", "level=info msg=\"Checking if KVM is enabled\"", "level=info msg=\"Checking if libvirt is installed\"", "level=info msg=\"Checking if user is part of libvirt group\"", "level=info msg=\"Checking if libvirt is enabled\"", "level=info msg=\"Checking if libvirt daemon is running\"", "level=info msg=\"Checking if a supported libvirt version is installed\"", "level=info msg=\"Checking if crc-driver-libvirt is installed\"", "level=info msg=\"Checking if libvirt 'crc' network is available\"", "level=info msg=\"Checking if libvirt 'crc' network is active\"", "level=info msg=\"Checking if NetworkManager is installed\"", "level=info msg=\"Checking if NetworkManager service is running\"", "level=info msg=\"Checking if /etc/NetworkManager/conf.d/crc-nm-dnsmasq.conf exists\"", "level=info msg=\"Checking if /etc/NetworkManager/dnsmasq.d/crc.conf exists\"", "level=info msg=\"Extracting bundle: crc_libvirt_4.2.13.crcbundle ...\"", "level=info msg=\"Creating CodeReady Containers VM for OpenShift 4.2.13...\"", "level=info msg=\"Verifying validity of the cluster certificates ...\"", "level=info msg=\"Check internal and public DNS query ...\"", "level=info msg=\"Check DNS query from host ...\"", "level=info msg=\"Copying kubeconfig file to instance dir ...\"", "level=info msg=\"Adding user's pull secret ...\"", "level=info msg=\"Updating cluster ID ...\"", "level=info msg=\"Starting OpenShift cluster ... [waiting 3m]\"", "level=info", "level=info msg=\"To access the cluster, first set up your environment by following 'crc oc-env' instructions\"", "level=info msg=\"Then you can access it by running 'oc login -u developer -p developer https://api.crc.testing:6443'\"", "level=info msg=\"To login as an admin, run 'oc login -u kubeadmin -p xx-password-xx https://api.crc.testing:6443'\"", "level=info", "level=info msg=\"You can now run 'crc console' and use these credentials to access the OpenShift web console\"", "Started the OpenShift cluster", "level=warning msg=\"The cluster might report a degraded or error state. This is expected since several operators have been disabled to lower the resource usage. For more information, please consult the documentation\"" ] } TASK [crc_start : crc status] ***************************************************************************************************** Thursday 16 January 2020 16:56:33 -0500 (0:00:00.135) 0:26:10.757 ****** FAILED - RETRYING: crc status (100 retries left). FAILED - RETRYING: crc status (99 retries left). FAILED - RETRYING: crc status (98 retries left). FAILED - RETRYING: crc status (97 retries left). FAILED - RETRYING: crc status (96 retries left). FAILED - RETRYING: crc status (95 retries left). FAILED - RETRYING: crc status (94 retries left). FAILED - RETRYING: crc status (93 retries left). FAILED - RETRYING: crc status (92 retries left). FAILED - RETRYING: crc status (91 retries left). FAILED - RETRYING: crc status (90 retries left). FAILED - RETRYING: crc status (89 retries left). FAILED - RETRYING: crc status (88 retries left). FAILED - RETRYING: crc status (87 retries left). FAILED - RETRYING: crc status (86 retries left). FAILED - RETRYING: crc status (85 retries left). FAILED - RETRYING: crc status (84 retries left). FAILED - RETRYING: crc status (83 retries left). changed: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_start : crc status stdout] ********************************************************************************************** Thursday 16 January 2020 17:01:33 -0500 (0:05:00.366) 0:31:11.123 ****** ok: [svtcrc-838996-1.fyre.ibm.com] => { "msg": [ "CRC VM: Running", "OpenShift: Running (v4.2.13)", "Disk Usage: 11.23GB of 32.2GB (Inside the CRC VM)", "Cache Usage: 13.23GB", "Cache Directory: /home/kevin/.crc/cache" ] } TASK [crc_oc_cli : include_tasks] ************************************************************************************************* Thursday 16 January 2020 17:01:33 -0500 (0:00:00.132) 0:31:11.256 ****** included: /home/nest/git/icpa-system-test/automation/ansible-playbooks/roles/crc_oc_cli/tasks/crc_oc_cli.yml for svtcrc-838996-1.fyre.ibm.com TASK [crc_oc_cli : oc in user path] *********************************************************************************************** Thursday 16 January 2020 17:01:33 -0500 (0:00:00.198) 0:31:11.454 ****** changed: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_oc_cli : link oc kubectl] *********************************************************************************************** Thursday 16 January 2020 17:01:34 -0500 (0:00:00.680) 0:31:12.135 ****** changed: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_oc_cli : crc creds] ***************************************************************************************************** Thursday 16 January 2020 17:01:35 -0500 (0:00:00.575) 0:31:12.710 ****** changed: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_oc_cli : crc creds stdout] ********************************************************************************************** Thursday 16 January 2020 17:01:35 -0500 (0:00:00.586) 0:31:13.296 ****** ok: [svtcrc-838996-1.fyre.ibm.com] => { "msg": [ "To login as a regular user, run 'oc login -u developer -p developer https://api.crc.testing:6443'.", "To login as an admin, run 'oc login -u kubeadmin -p xx-password-xx https://api.crc.testing:6443'" ] } TASK [crc_oc_cli : set_fact] ****************************************************************************************************** Thursday 16 January 2020 17:01:35 -0500 (0:00:00.132) 0:31:13.429 ****** ok: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_oc_cli : set_fact] ****************************************************************************************************** Thursday 16 January 2020 17:01:35 -0500 (0:00:00.111) 0:31:13.541 ****** ok: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_oc_cli : set_fact] ****************************************************************************************************** Thursday 16 January 2020 17:01:36 -0500 (0:00:00.119) 0:31:13.661 ****** ok: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_oc_cli : set_fact] ****************************************************************************************************** Thursday 16 January 2020 17:01:36 -0500 (0:00:00.133) 0:31:13.794 ****** ok: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_oc_cli : set_fact] ****************************************************************************************************** Thursday 16 January 2020 17:01:36 -0500 (0:00:00.112) 0:31:13.906 ****** ok: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_oc_cli : oc login] ****************************************************************************************************** Thursday 16 January 2020 17:01:36 -0500 (0:00:00.124) 0:31:14.031 ****** changed: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_oc_cli : oc login stdout] *********************************************************************************************** Thursday 16 January 2020 17:01:37 -0500 (0:00:01.183) 0:31:15.215 ****** ok: [svtcrc-838996-1.fyre.ibm.com] => { "msg": [ "Login successful.", "", "You have access to 51 projects, the list has been suppressed. You can list all projects with 'oc projects'", "", "Using project \"default\".", "Welcome! See 'oc help' to get started." ] } TASK [crc_oc_cli : password masking process] ************************************************************************************** Thursday 16 January 2020 17:01:37 -0500 (0:00:00.146) 0:31:15.368 ****** changed: [svtcrc-838996-1.fyre.ibm.com -> localhost] TASK [log : include_tasks] ******************************************************************************************************** Thursday 16 January 2020 17:01:38 -0500 (0:00:00.655) 0:31:16.023 ****** included: /home/nest/git/icpa-system-test/automation/ansible-playbooks/roles/log/tasks/log.yml for svtcrc-838996-1.fyre.ibm.com TASK [log : debug] **************************************************************************************************************** Thursday 16 January 2020 17:01:38 -0500 (0:00:00.164) 0:31:16.188 ****** ok: [svtcrc-838996-1.fyre.ibm.com] => { "msg": "******************** crc_install_end ********************"
2.0
SVT: TER: Kabanero Automation test execution master branch - Setup: svtcrc-838996-1.fyre.ibm.com - "msg": "******************** crc_fyresetup_start ********************" } TASK [load_secrets : Load all secrets] ******************************************************************************************** Thursday 16 January 2020 16:30:22 -0500 (0:00:00.085) 0:00:00.382 ****** ok: [localhost] TASK [crc_fyrevm : include_tasks] ************************************************************************************************* Thursday 16 January 2020 16:30:22 -0500 (0:00:00.088) 0:00:00.470 ****** included: /home/nest/git/icpa-system-test/automation/ansible-playbooks/roles/crc_fyrevm/tasks/crc_fyrevm.yml for localhost TASK [crc_fyrevm : set_fact] ****************************************************************************************************** Thursday 16 January 2020 16:30:23 -0500 (0:00:00.214) 0:00:00.685 ****** ok: [localhost] TASK [crc_fyrevm : check status of the named cluster] ***************************************************************************** Thursday 16 January 2020 16:30:23 -0500 (0:00:00.080) 0:00:00.766 ****** changed: [localhost -> localhost] TASK [crc_fyrevm : debug] ********************************************************************************************************* Thursday 16 January 2020 16:30:24 -0500 (0:00:01.489) 0:00:02.256 ****** skipping: [localhost] TASK [crc_fyrevm : get the public ssh id] ***************************************************************************************** Thursday 16 January 2020 16:30:24 -0500 (0:00:00.088) 0:00:02.345 ****** changed: [localhost -> localhost] TASK [crc_fyrevm : create input json file] **************************************************************************************** Thursday 16 January 2020 16:30:25 -0500 (0:00:00.500) 0:00:02.845 ****** changed: [localhost] TASK [crc_fyrevm : Create Fyre stack] ********************************************************************************************* Thursday 16 January 2020 16:30:26 -0500 (0:00:01.020) 0:00:03.866 ****** changed: [localhost] TASK [crc_fyrevm : remove temp json] ********************************************************************************************** Thursday 16 January 2020 16:30:28 -0500 (0:00:01.821) 0:00:05.687 ****** changed: [localhost -> localhost] TASK [crc_fyrevm : check Fyre request status] ************************************************************************************* Thursday 16 January 2020 16:30:28 -0500 (0:00:00.545) 0:00:06.233 ****** changed: [localhost] TASK [crc_fyrevm : debug] ********************************************************************************************************* Thursday 16 January 2020 16:30:29 -0500 (0:00:01.066) 0:00:07.299 ****** ok: [localhost] => { "fyreStatus.stdout": { "request": [ { "cluster_name": "", "created": "2020-01-16 13:30:27", "displayname": "steven r. schader", "error_details": "0", "id": "2750591", "status": "building", "type": "build", "updated": "2020-01-16 13:30:28", "username": "schader" } ] } } TASK [crc_fyrevm : pause] ********************************************************************************************************* Thursday 16 January 2020 16:30:29 -0500 (0:00:00.035) 0:00:07.335 ****** Pausing for 120 seconds (ctrl+C then 'C' = continue early, ctrl+C then 'A' = abort) ok: [localhost] TASK [crc_fyrevm : get Fyre request status] *************************************************************************************** Thursday 16 January 2020 16:32:29 -0500 (0:02:00.071) 0:02:07.407 ****** FAILED - RETRYING: get Fyre request status (999 retries left). FAILED - RETRYING: get Fyre request status (998 retries left). FAILED - RETRYING: get Fyre request status (997 retries left). FAILED - RETRYING: get Fyre request status (996 retries left). FAILED - RETRYING: get Fyre request status (995 retries left). FAILED - RETRYING: get Fyre request status (994 retries left). FAILED - RETRYING: get Fyre request status (993 retries left). FAILED - RETRYING: get Fyre request status (992 retries left). FAILED - RETRYING: get Fyre request status (991 retries left). FAILED - RETRYING: get Fyre request status (990 retries left). FAILED - RETRYING: get Fyre request status (989 retries left). FAILED - RETRYING: get Fyre request status (988 retries left). FAILED - RETRYING: get Fyre request status (987 retries left). FAILED - RETRYING: get Fyre request status (986 retries left). FAILED - RETRYING: get Fyre request status (985 retries left). FAILED - RETRYING: get Fyre request status (984 retries left). FAILED - RETRYING: get Fyre request status (983 retries left). FAILED - RETRYING: get Fyre request status (982 retries left). FAILED - RETRYING: get Fyre request status (981 retries left). FAILED - RETRYING: get Fyre request status (980 retries left). FAILED - RETRYING: get Fyre request status (979 retries left). FAILED - RETRYING: get Fyre request status (978 retries left). FAILED - RETRYING: get Fyre request status (977 retries left). FAILED - RETRYING: get Fyre request status (976 retries left). FAILED - RETRYING: get Fyre request status (975 retries left). FAILED - RETRYING: get Fyre request status (974 retries left). FAILED - RETRYING: get Fyre request status (973 retries left). FAILED - RETRYING: get Fyre request status (972 retries left). FAILED - RETRYING: get Fyre request status (971 retries left). FAILED - RETRYING: get Fyre request status (970 retries left). FAILED - RETRYING: get Fyre request status (969 retries left). FAILED - RETRYING: get Fyre request status (968 retries left). FAILED - RETRYING: get Fyre request status (967 retries left). FAILED - RETRYING: get Fyre request status (966 retries left). FAILED - RETRYING: get Fyre request status (965 retries left). FAILED - RETRYING: get Fyre request status (964 retries left). FAILED - RETRYING: get Fyre request status (963 retries left). FAILED - RETRYING: get Fyre request status (962 retries left). FAILED - RETRYING: get Fyre request status (961 retries left). FAILED - RETRYING: get Fyre request status (960 retries left). FAILED - RETRYING: get Fyre request status (959 retries left). FAILED - RETRYING: get Fyre request status (958 retries left). FAILED - RETRYING: get Fyre request status (957 retries left). FAILED - RETRYING: get Fyre request status (956 retries left). FAILED - RETRYING: get Fyre request status (955 retries left). FAILED - RETRYING: get Fyre request status (954 retries left). FAILED - RETRYING: get Fyre request status (953 retries left). FAILED - RETRYING: get Fyre request status (952 retries left). FAILED - RETRYING: get Fyre request status (951 retries left). FAILED - RETRYING: get Fyre request status (950 retries left). FAILED - RETRYING: get Fyre request status (949 retries left). FAILED - RETRYING: get Fyre request status (948 retries left). FAILED - RETRYING: get Fyre request status (947 retries left). FAILED - RETRYING: get Fyre request status (946 retries left). changed: [localhost] TASK [crc_fyrevm : check Fyre request status] ************************************************************************************* Thursday 16 January 2020 16:37:52 -0500 (0:05:22.785) 0:07:30.192 ****** skipping: [localhost] TASK [crc_fyrevm : check the host for a active ssh] ******************************************************************************* Thursday 16 January 2020 16:37:52 -0500 (0:00:00.041) 0:07:30.234 ****** changed: [localhost -> localhost] TASK [crc_fyrevm : remove new host from localhost known_hosts] ******************************************************************** Thursday 16 January 2020 16:37:53 -0500 (0:00:00.729) 0:07:30.963 ****** changed: [localhost -> localhost] TASK [crc_fyrevm : add host to known_hosts] *************************************************************************************** Thursday 16 January 2020 16:37:53 -0500 (0:00:00.333) 0:07:31.297 ****** changed: [localhost -> localhost] TASK [crc_fyrevm : add_host] ****************************************************************************************************** Thursday 16 January 2020 16:37:54 -0500 (0:00:00.425) 0:07:31.722 ****** changed: [localhost] TASK [crc_fyrevm : create host inventory file for debugging] ********************************************************************** Thursday 16 January 2020 16:37:54 -0500 (0:00:00.059) 0:07:31.782 ****** changed: [localhost -> localhost] TASK [crc_fyrevm : create plain file with host] *********************************************************************************** Thursday 16 January 2020 16:37:54 -0500 (0:00:00.575) 0:07:32.358 ****** changed: [localhost -> localhost] TASK [crc_fyrevm : create plain file with cluster name] *************************************************************************** Thursday 16 January 2020 16:37:55 -0500 (0:00:00.590) 0:07:32.948 ****** changed: [localhost -> localhost] PLAY [kabanero_crc] *************************************************************************************************************** TASK [load_secrets : Load all secrets] ******************************************************************************************** Thursday 16 January 2020 16:37:55 -0500 (0:00:00.647) 0:07:33.596 ****** ok: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_host_prereqs : set timezone to to regional tz] ************************************************************************** Thursday 16 January 2020 16:37:56 -0500 (0:00:00.090) 0:07:33.686 ****** changed: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_host_prereqs : check if jq is installed] ******************************************************************************** Thursday 16 January 2020 16:37:57 -0500 (0:00:01.126) 0:07:34.813 ****** ok: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_host_prereqs : Copy JQ to VM] ******************************************************************************************* Thursday 16 January 2020 16:37:57 -0500 (0:00:00.473) 0:07:35.287 ****** changed: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_host_prereqs : Change jq permissions] *********************************************************************************** Thursday 16 January 2020 16:37:58 -0500 (0:00:01.062) 0:07:36.349 ****** changed: [svtcrc-838996-1.fyre.ibm.com] PLAY [kabanero_crc] *************************************************************************************************************** TASK [Gathering Facts] ************************************************************************************************************ Thursday 16 January 2020 16:37:59 -0500 (0:00:00.526) 0:07:36.876 ****** ok: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_user : include_tasks] *************************************************************************************************** Thursday 16 January 2020 16:38:00 -0500 (0:00:01.732) 0:07:38.609 ****** included: /home/nest/git/icpa-system-test/automation/ansible-playbooks/roles/crc_user/tasks/crc_user.yml for svtcrc-838996-1.fyre.ibm.com TASK [crc_user : Ensure User is Present] ****************************************************************************************** Thursday 16 January 2020 16:38:01 -0500 (0:00:00.252) 0:07:38.861 ****** changed: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_user : Add user to sudoers] ********************************************************************************************* Thursday 16 January 2020 16:38:02 -0500 (0:00:01.399) 0:07:40.260 ****** changed: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_user : update user hard / soft ulimit nofile] *************************************************************************** Thursday 16 January 2020 16:38:03 -0500 (0:00:00.841) 0:07:41.101 ****** changed: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_user : authorized_key] ************************************************************************************************** Thursday 16 January 2020 16:38:04 -0500 (0:00:00.733) 0:07:41.834 ****** changed: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_user : add bin to path] ************************************************************************************************* Thursday 16 January 2020 16:38:05 -0500 (0:00:00.894) 0:07:42.729 ****** changed: [svtcrc-838996-1.fyre.ibm.com] TASK [log : include_tasks] ******************************************************************************************************** Thursday 16 January 2020 16:38:05 -0500 (0:00:00.570) 0:07:43.300 ****** included: /home/nest/git/icpa-system-test/automation/ansible-playbooks/roles/log/tasks/log.yml for svtcrc-838996-1.fyre.ibm.com TASK [log : debug] **************************************************************************************************************** Thursday 16 January 2020 16:38:05 -0500 (0:00:00.192) 0:07:43.492 ****** ok: [svtcrc-838996-1.fyre.ibm.com] => { "msg": "******************** crc_fyresetup_end ********************" } PLAY [kabanero_crc] *************************************************************************************************************** TASK [Gathering Facts] ************************************************************************************************************ Thursday 16 January 2020 16:38:06 -0500 (0:00:00.164) 0:07:43.657 ****** ok: [svtcrc-838996-1.fyre.ibm.com] TASK [log : include_tasks] ******************************************************************************************************** Thursday 16 January 2020 16:38:08 -0500 (0:00:02.220) 0:07:45.877 ****** included: /home/nest/git/icpa-system-test/automation/ansible-playbooks/roles/log/tasks/log.yml for svtcrc-838996-1.fyre.ibm.com TASK [log : debug] **************************************************************************************************************** Thursday 16 January 2020 16:38:08 -0500 (0:00:00.171) 0:07:46.049 ****** ok: [svtcrc-838996-1.fyre.ibm.com] => { "msg": "******************** crc_install_start ********************" } TASK [load_secrets : Load all secrets] ******************************************************************************************** Thursday 16 January 2020 16:38:08 -0500 (0:00:00.113) 0:07:46.162 ****** ok: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_install : include_tasks] ************************************************************************************************ Thursday 16 January 2020 16:38:08 -0500 (0:00:00.118) 0:07:46.281 ****** included: /home/nest/git/icpa-system-test/automation/ansible-playbooks/roles/crc_install/tasks/crc_install.yml for svtcrc-838996-1.fyre.ibm.com TASK [crc_install : check if crc is installed] ************************************************************************************ Thursday 16 January 2020 16:38:08 -0500 (0:00:00.100) 0:07:46.382 ****** ok: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_install : Download CRC Archive] ***************************************************************************************** Thursday 16 January 2020 16:38:09 -0500 (0:00:00.510) 0:07:46.893 ****** changed: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_install : archive dir structure] **************************************************************************************** Thursday 16 January 2020 16:39:03 -0500 (0:00:54.692) 0:08:41.585 ****** changed: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_install : debug] ******************************************************************************************************** Thursday 16 January 2020 16:39:12 -0500 (0:00:08.280) 0:08:49.866 ****** ok: [svtcrc-838996-1.fyre.ibm.com] => { "tarlist.stdout_lines": [ "crc-linux-1.4.0-amd64/", "crc-linux-1.4.0-amd64/LICENSE", "crc-linux-1.4.0-amd64/doc.pdf", "crc-linux-1.4.0-amd64/crc" ] } TASK [crc_install : install crc] ************************************************************************************************** Thursday 16 January 2020 16:39:12 -0500 (0:00:00.151) 0:08:50.018 ****** changed: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_install : home bin] ***************************************************************************************************** Thursday 16 January 2020 16:39:41 -0500 (0:00:28.921) 0:09:18.945 ****** changed: [svtcrc-838996-1.fyre.ibm.com] => (item=/home/kevin/bin) TASK [crc_install : link crc] ***************************************************************************************************** Thursday 16 January 2020 16:39:42 -0500 (0:00:00.757) 0:09:19.702 ****** changed: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_install : Install NetworkManager FireFox] ******************************************************************************* Thursday 16 January 2020 16:39:42 -0500 (0:00:00.652) 0:09:20.355 ****** changed: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_install : enable and start NetworkManager] ****************************************************************************** Thursday 16 January 2020 16:41:31 -0500 (0:01:48.925) 0:11:09.280 ****** changed: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_start : include_tasks] ************************************************************************************************** Thursday 16 January 2020 16:41:33 -0500 (0:00:01.501) 0:11:10.782 ****** included: /home/nest/git/icpa-system-test/automation/ansible-playbooks/roles/crc_start/tasks/crc_start.yml for svtcrc-838996-1.fyre.ibm.com TASK [crc_start : check if .crc is created] *************************************************************************************** Thursday 16 January 2020 16:41:33 -0500 (0:00:00.262) 0:11:11.044 ****** ok: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_start : crc stop] ******************************************************************************************************* Thursday 16 January 2020 16:41:34 -0500 (0:00:00.660) 0:11:11.705 ****** skipping: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_start : crcstop stdout] ************************************************************************************************* Thursday 16 January 2020 16:41:34 -0500 (0:00:00.135) 0:11:11.840 ****** skipping: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_start : crc delete] ***************************************************************************************************** Thursday 16 January 2020 16:41:34 -0500 (0:00:00.056) 0:11:11.897 ****** skipping: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_start : crc delete stdout] ********************************************************************************************** Thursday 16 January 2020 16:41:34 -0500 (0:00:00.107) 0:11:12.005 ****** skipping: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_start : crc setup] ****************************************************************************************************** Thursday 16 January 2020 16:41:34 -0500 (0:00:00.043) 0:11:12.048 ****** changed: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_start : crc setup stdout] *********************************************************************************************** Thursday 16 January 2020 16:42:04 -0500 (0:00:29.895) 0:11:41.944 ****** ok: [svtcrc-838996-1.fyre.ibm.com] => { "msg": [ "level=info msg=\"Checking if oc binary is cached\"", "level=info msg=\"Caching oc binary\"", "level=info msg=\"Checking if CRC bundle is cached in '$HOME/.crc'\"", "level=info msg=\"Unpacking bundle from the CRC binary\"", "level=info msg=\"Checking if running as non-root\"", "level=info msg=\"Checking if Virtualization is enabled\"", "level=info msg=\"Checking if KVM is enabled\"", "level=info msg=\"Checking if libvirt is installed\"", "level=info msg=\"Installing libvirt service and dependencies\"", "level=info msg=\"Will use root access: install virtualization related packages\"", "level=info msg=\"Checking if user is part of libvirt group\"", "level=info msg=\"Adding user to libvirt group\"", "level=info msg=\"Will use root access: add user to libvirt group\"", "level=info msg=\"Checking if libvirt is enabled\"", "level=info msg=\"Checking if libvirt daemon is running\"", "level=info msg=\"Starting libvirt service\"", "level=info msg=\"Will use root access: start libvirtd service\"", "level=info msg=\"Checking if a supported libvirt version is installed\"", "level=info msg=\"Checking if crc-driver-libvirt is installed\"", "level=info msg=\"Installing crc-driver-libvirt\"", "level=info msg=\"Checking for obsolete crc-driver-libvirt\"", "level=info msg=\"Checking if libvirt 'crc' network is available\"", "level=info msg=\"Setting up libvirt 'crc' network\"", "level=info msg=\"Checking if libvirt 'crc' network is active\"", "level=info msg=\"Starting libvirt 'crc' network\"", "level=info msg=\"Checking if NetworkManager is installed\"", "level=info msg=\"Checking if NetworkManager service is running\"", "level=info msg=\"Checking if /etc/NetworkManager/conf.d/crc-nm-dnsmasq.conf exists\"", "level=info msg=\"Writing Network Manager config for crc\"", "level=info msg=\"Will use root access: write NetworkManager config in /etc/NetworkManager/conf.d/crc-nm-dnsmasq.conf\"", "level=info msg=\"Will use root access: execute systemctl daemon-reload command\"", "level=info msg=\"Will use root access: execute systemctl stop/start command\"", "level=info msg=\"Checking if /etc/NetworkManager/dnsmasq.d/crc.conf exists\"", "level=info msg=\"Writing dnsmasq config for crc\"", "level=info msg=\"Will use root access: write dnsmasq configuration in /etc/NetworkManager/dnsmasq.d/crc.conf\"", "level=info msg=\"Will use root access: execute systemctl daemon-reload command\"", "level=info msg=\"Will use root access: execute systemctl stop/start command\"", "Setup is complete, you can now run 'crc start' to start the OpenShift cluster" ] } TASK [crc_start : crc pull secret] ************************************************************************************************ Thursday 16 January 2020 16:42:04 -0500 (0:00:00.052) 0:11:41.996 ****** changed: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_start : crc start] ****************************************************************************************************** Thursday 16 January 2020 16:42:05 -0500 (0:00:00.938) 0:11:42.934 ****** changed: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_start : crc start stdout] *********************************************************************************************** Thursday 16 January 2020 16:56:32 -0500 (0:14:27.687) 0:26:10.621 ****** ok: [svtcrc-838996-1.fyre.ibm.com] => { "msg": [ "level=info msg=\"Checking if oc binary is cached\"", "level=info msg=\"Checking if running as non-root\"", "level=info msg=\"Checking if Virtualization is enabled\"", "level=info msg=\"Checking if KVM is enabled\"", "level=info msg=\"Checking if libvirt is installed\"", "level=info msg=\"Checking if user is part of libvirt group\"", "level=info msg=\"Checking if libvirt is enabled\"", "level=info msg=\"Checking if libvirt daemon is running\"", "level=info msg=\"Checking if a supported libvirt version is installed\"", "level=info msg=\"Checking if crc-driver-libvirt is installed\"", "level=info msg=\"Checking if libvirt 'crc' network is available\"", "level=info msg=\"Checking if libvirt 'crc' network is active\"", "level=info msg=\"Checking if NetworkManager is installed\"", "level=info msg=\"Checking if NetworkManager service is running\"", "level=info msg=\"Checking if /etc/NetworkManager/conf.d/crc-nm-dnsmasq.conf exists\"", "level=info msg=\"Checking if /etc/NetworkManager/dnsmasq.d/crc.conf exists\"", "level=info msg=\"Extracting bundle: crc_libvirt_4.2.13.crcbundle ...\"", "level=info msg=\"Creating CodeReady Containers VM for OpenShift 4.2.13...\"", "level=info msg=\"Verifying validity of the cluster certificates ...\"", "level=info msg=\"Check internal and public DNS query ...\"", "level=info msg=\"Check DNS query from host ...\"", "level=info msg=\"Copying kubeconfig file to instance dir ...\"", "level=info msg=\"Adding user's pull secret ...\"", "level=info msg=\"Updating cluster ID ...\"", "level=info msg=\"Starting OpenShift cluster ... [waiting 3m]\"", "level=info", "level=info msg=\"To access the cluster, first set up your environment by following 'crc oc-env' instructions\"", "level=info msg=\"Then you can access it by running 'oc login -u developer -p developer https://api.crc.testing:6443'\"", "level=info msg=\"To login as an admin, run 'oc login -u kubeadmin -p xx-password-xx https://api.crc.testing:6443'\"", "level=info", "level=info msg=\"You can now run 'crc console' and use these credentials to access the OpenShift web console\"", "Started the OpenShift cluster", "level=warning msg=\"The cluster might report a degraded or error state. This is expected since several operators have been disabled to lower the resource usage. For more information, please consult the documentation\"" ] } TASK [crc_start : crc status] ***************************************************************************************************** Thursday 16 January 2020 16:56:33 -0500 (0:00:00.135) 0:26:10.757 ****** FAILED - RETRYING: crc status (100 retries left). FAILED - RETRYING: crc status (99 retries left). FAILED - RETRYING: crc status (98 retries left). FAILED - RETRYING: crc status (97 retries left). FAILED - RETRYING: crc status (96 retries left). FAILED - RETRYING: crc status (95 retries left). FAILED - RETRYING: crc status (94 retries left). FAILED - RETRYING: crc status (93 retries left). FAILED - RETRYING: crc status (92 retries left). FAILED - RETRYING: crc status (91 retries left). FAILED - RETRYING: crc status (90 retries left). FAILED - RETRYING: crc status (89 retries left). FAILED - RETRYING: crc status (88 retries left). FAILED - RETRYING: crc status (87 retries left). FAILED - RETRYING: crc status (86 retries left). FAILED - RETRYING: crc status (85 retries left). FAILED - RETRYING: crc status (84 retries left). FAILED - RETRYING: crc status (83 retries left). changed: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_start : crc status stdout] ********************************************************************************************** Thursday 16 January 2020 17:01:33 -0500 (0:05:00.366) 0:31:11.123 ****** ok: [svtcrc-838996-1.fyre.ibm.com] => { "msg": [ "CRC VM: Running", "OpenShift: Running (v4.2.13)", "Disk Usage: 11.23GB of 32.2GB (Inside the CRC VM)", "Cache Usage: 13.23GB", "Cache Directory: /home/kevin/.crc/cache" ] } TASK [crc_oc_cli : include_tasks] ************************************************************************************************* Thursday 16 January 2020 17:01:33 -0500 (0:00:00.132) 0:31:11.256 ****** included: /home/nest/git/icpa-system-test/automation/ansible-playbooks/roles/crc_oc_cli/tasks/crc_oc_cli.yml for svtcrc-838996-1.fyre.ibm.com TASK [crc_oc_cli : oc in user path] *********************************************************************************************** Thursday 16 January 2020 17:01:33 -0500 (0:00:00.198) 0:31:11.454 ****** changed: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_oc_cli : link oc kubectl] *********************************************************************************************** Thursday 16 January 2020 17:01:34 -0500 (0:00:00.680) 0:31:12.135 ****** changed: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_oc_cli : crc creds] ***************************************************************************************************** Thursday 16 January 2020 17:01:35 -0500 (0:00:00.575) 0:31:12.710 ****** changed: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_oc_cli : crc creds stdout] ********************************************************************************************** Thursday 16 January 2020 17:01:35 -0500 (0:00:00.586) 0:31:13.296 ****** ok: [svtcrc-838996-1.fyre.ibm.com] => { "msg": [ "To login as a regular user, run 'oc login -u developer -p developer https://api.crc.testing:6443'.", "To login as an admin, run 'oc login -u kubeadmin -p xx-password-xx https://api.crc.testing:6443'" ] } TASK [crc_oc_cli : set_fact] ****************************************************************************************************** Thursday 16 January 2020 17:01:35 -0500 (0:00:00.132) 0:31:13.429 ****** ok: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_oc_cli : set_fact] ****************************************************************************************************** Thursday 16 January 2020 17:01:35 -0500 (0:00:00.111) 0:31:13.541 ****** ok: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_oc_cli : set_fact] ****************************************************************************************************** Thursday 16 January 2020 17:01:36 -0500 (0:00:00.119) 0:31:13.661 ****** ok: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_oc_cli : set_fact] ****************************************************************************************************** Thursday 16 January 2020 17:01:36 -0500 (0:00:00.133) 0:31:13.794 ****** ok: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_oc_cli : set_fact] ****************************************************************************************************** Thursday 16 January 2020 17:01:36 -0500 (0:00:00.112) 0:31:13.906 ****** ok: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_oc_cli : oc login] ****************************************************************************************************** Thursday 16 January 2020 17:01:36 -0500 (0:00:00.124) 0:31:14.031 ****** changed: [svtcrc-838996-1.fyre.ibm.com] TASK [crc_oc_cli : oc login stdout] *********************************************************************************************** Thursday 16 January 2020 17:01:37 -0500 (0:00:01.183) 0:31:15.215 ****** ok: [svtcrc-838996-1.fyre.ibm.com] => { "msg": [ "Login successful.", "", "You have access to 51 projects, the list has been suppressed. You can list all projects with 'oc projects'", "", "Using project \"default\".", "Welcome! See 'oc help' to get started." ] } TASK [crc_oc_cli : password masking process] ************************************************************************************** Thursday 16 January 2020 17:01:37 -0500 (0:00:00.146) 0:31:15.368 ****** changed: [svtcrc-838996-1.fyre.ibm.com -> localhost] TASK [log : include_tasks] ******************************************************************************************************** Thursday 16 January 2020 17:01:38 -0500 (0:00:00.655) 0:31:16.023 ****** included: /home/nest/git/icpa-system-test/automation/ansible-playbooks/roles/log/tasks/log.yml for svtcrc-838996-1.fyre.ibm.com TASK [log : debug] **************************************************************************************************************** Thursday 16 January 2020 17:01:38 -0500 (0:00:00.164) 0:31:16.188 ****** ok: [svtcrc-838996-1.fyre.ibm.com] => { "msg": "******************** crc_install_end ********************"
test
svt ter kabanero automation test execution master branch setup svtcrc fyre ibm com msg crc fyresetup start task thursday january ok task thursday january included home nest git icpa system test automation ansible playbooks roles crc fyrevm tasks crc fyrevm yml for localhost task thursday january ok task thursday january changed task thursday january skipping task thursday january changed task thursday january changed task thursday january changed task thursday january changed task thursday january changed task thursday january ok fyrestatus stdout request cluster name created displayname steven r schader error details id status building type build updated username schader task thursday january pausing for seconds ctrl c then c continue early ctrl c then a abort ok task thursday january failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left failed retrying get fyre request status retries left changed task thursday january skipping task thursday january changed task thursday january changed task thursday january changed task thursday january changed task thursday january changed task thursday january changed task thursday january changed play task thursday january ok task thursday january changed task thursday january ok task thursday january changed task thursday january changed play task thursday january ok task thursday january included home nest git icpa system test automation ansible playbooks roles crc user tasks crc user yml for svtcrc fyre ibm com task thursday january changed task thursday january changed task thursday january changed task thursday january changed task thursday january changed task thursday january included home nest git icpa system test automation ansible playbooks roles log tasks log yml for svtcrc fyre ibm com task thursday january ok msg crc fyresetup end play task thursday january ok task thursday january included home nest git icpa system test automation ansible playbooks roles log tasks log yml for svtcrc fyre ibm com task thursday january ok msg crc install start task thursday january ok task thursday january included home nest git icpa system test automation ansible playbooks roles crc install tasks crc install yml for svtcrc fyre ibm com task thursday january ok task thursday january changed task thursday january changed task thursday january ok tarlist stdout lines crc linux crc linux license crc linux doc pdf crc linux crc task thursday january changed task thursday january changed item home kevin bin task thursday january changed task thursday january changed task thursday january changed task thursday january included home nest git icpa system test automation ansible playbooks roles crc start tasks crc start yml for svtcrc fyre ibm com task thursday january ok task thursday january skipping task thursday january skipping task thursday january skipping task thursday january skipping task thursday january changed task thursday january ok msg level info msg checking if oc binary is cached level info msg caching oc binary level info msg checking if crc bundle is cached in home crc level info msg unpacking bundle from the crc binary level info msg checking if running as non root level info msg checking if virtualization is enabled level info msg checking if kvm is enabled level info msg checking if libvirt is installed level info msg installing libvirt service and dependencies level info msg will use root access install virtualization related packages level info msg checking if user is part of libvirt group level info msg adding user to libvirt group level info msg will use root access add user to libvirt group level info msg checking if libvirt is enabled level info msg checking if libvirt daemon is running level info msg starting libvirt service level info msg will use root access start libvirtd service level info msg checking if a supported libvirt version is installed level info msg checking if crc driver libvirt is installed level info msg installing crc driver libvirt level info msg checking for obsolete crc driver libvirt level info msg checking if libvirt crc network is available level info msg setting up libvirt crc network level info msg checking if libvirt crc network is active level info msg starting libvirt crc network level info msg checking if networkmanager is installed level info msg checking if networkmanager service is running level info msg checking if etc networkmanager conf d crc nm dnsmasq conf exists level info msg writing network manager config for crc level info msg will use root access write networkmanager config in etc networkmanager conf d crc nm dnsmasq conf level info msg will use root access execute systemctl daemon reload command level info msg will use root access execute systemctl stop start command level info msg checking if etc networkmanager dnsmasq d crc conf exists level info msg writing dnsmasq config for crc level info msg will use root access write dnsmasq configuration in etc networkmanager dnsmasq d crc conf level info msg will use root access execute systemctl daemon reload command level info msg will use root access execute systemctl stop start command setup is complete you can now run crc start to start the openshift cluster task thursday january changed task thursday january changed task thursday january ok msg level info msg checking if oc binary is cached level info msg checking if running as non root level info msg checking if virtualization is enabled level info msg checking if kvm is enabled level info msg checking if libvirt is installed level info msg checking if user is part of libvirt group level info msg checking if libvirt is enabled level info msg checking if libvirt daemon is running level info msg checking if a supported libvirt version is installed level info msg checking if crc driver libvirt is installed level info msg checking if libvirt crc network is available level info msg checking if libvirt crc network is active level info msg checking if networkmanager is installed level info msg checking if networkmanager service is running level info msg checking if etc networkmanager conf d crc nm dnsmasq conf exists level info msg checking if etc networkmanager dnsmasq d crc conf exists level info msg extracting bundle crc libvirt crcbundle level info msg creating codeready containers vm for openshift level info msg verifying validity of the cluster certificates level info msg check internal and public dns query level info msg check dns query from host level info msg copying kubeconfig file to instance dir level info msg adding user s pull secret level info msg updating cluster id level info msg starting openshift cluster level info level info msg to access the cluster first set up your environment by following crc oc env instructions level info msg then you can access it by running oc login u developer p developer level info msg to login as an admin run oc login u kubeadmin p xx password xx level info level info msg you can now run crc console and use these credentials to access the openshift web console started the openshift cluster level warning msg the cluster might report a degraded or error state this is expected since several operators have been disabled to lower the resource usage for more information please consult the documentation task thursday january failed retrying crc status retries left failed retrying crc status retries left failed retrying crc status retries left failed retrying crc status retries left failed retrying crc status retries left failed retrying crc status retries left failed retrying crc status retries left failed retrying crc status retries left failed retrying crc status retries left failed retrying crc status retries left failed retrying crc status retries left failed retrying crc status retries left failed retrying crc status retries left failed retrying crc status retries left failed retrying crc status retries left failed retrying crc status retries left failed retrying crc status retries left failed retrying crc status retries left changed task thursday january ok msg crc vm running openshift running disk usage of inside the crc vm cache usage cache directory home kevin crc cache task thursday january included home nest git icpa system test automation ansible playbooks roles crc oc cli tasks crc oc cli yml for svtcrc fyre ibm com task thursday january changed task thursday january changed task thursday january changed task thursday january ok msg to login as a regular user run oc login u developer p developer to login as an admin run oc login u kubeadmin p xx password xx task thursday january ok task thursday january ok task thursday january ok task thursday january ok task thursday january ok task thursday january changed task thursday january ok msg login successful you have access to projects the list has been suppressed you can list all projects with oc projects using project default welcome see oc help to get started task thursday january changed task thursday january included home nest git icpa system test automation ansible playbooks roles log tasks log yml for svtcrc fyre ibm com task thursday january ok msg crc install end
1
312,561
26,871,897,256
IssuesEvent
2023-02-04 15:20:05
MPMG-DCC-UFMG/F01
https://api.github.com/repos/MPMG-DCC-UFMG/F01
closed
Teste de generalizacao para a tag Informações Institucionais - Registro das Competências - Soledade de Minas
generalization test development template - ABO (21) tag - Informações Institucionais subtag - Registro das Competências
DoD: Realizar o teste de Generalização do validador da tag Informações Institucionais - Registro das Competências para o Município de Soledade de Minas.
1.0
Teste de generalizacao para a tag Informações Institucionais - Registro das Competências - Soledade de Minas - DoD: Realizar o teste de Generalização do validador da tag Informações Institucionais - Registro das Competências para o Município de Soledade de Minas.
test
teste de generalizacao para a tag informações institucionais registro das competências soledade de minas dod realizar o teste de generalização do validador da tag informações institucionais registro das competências para o município de soledade de minas
1
76,442
21,415,206,494
IssuesEvent
2022-04-22 10:06:16
appsmithorg/appsmith
https://api.github.com/repos/appsmithorg/appsmith
opened
[Task]: Send Analytics events for Property pane keyboard navigation
UI Builders Pod Task
### Is there an existing issue for this? - [X] I have searched the existing issues ### SubTasks meta data should include - key pressed - source property control
1.0
[Task]: Send Analytics events for Property pane keyboard navigation - ### Is there an existing issue for this? - [X] I have searched the existing issues ### SubTasks meta data should include - key pressed - source property control
non_test
send analytics events for property pane keyboard navigation is there an existing issue for this i have searched the existing issues subtasks meta data should include key pressed source property control
0
167,268
13,018,061,691
IssuesEvent
2020-07-26 15:31:51
IntellectualSites/FastAsyncWorldEdit
https://api.github.com/repos/IntellectualSites/FastAsyncWorldEdit
opened
jumpto + coordinate not working
Requires Testing
# Bug report for FastAsyncWorldEdit 1.16.1 If I use `/jumpto` all works fine. But with `/jumpto [world,x,y,z]` I get the error: "error while handling a WorldEdit command" https://pastebin.com/KsEfh6uP **[REQUIRED] FastAsyncWorldEdit Configuration Files:** https://athion.net/ISPaster/paste/view/9d90979f779e4bb3bda8285de5b0d17a **[REQUIRED] FastAsyncWorldEdit Version Number:** <!--- Enter /fawe version in game or in your console and copy the full output here --> **[REQUIRED] Spigot/Paper Version Number:** > fawe 1.16 - 306 > PaperMC 1.15.2 **How to replicate**: `/jumpto world;12;12;12` `/jumpto world,12,12,12` **Checklist**: <!--- Make sure you've completed the following steps (put an "X" between of brackets): --> - [X] I included all information required in the sections above - [X] I made sure there are no duplicates of this report [(Use Search)](https://github.com/IntellectualSites/FastAsyncWorldEdit/issues?q=is%3Aissue) - [X] I made sure I am using an up-to-date version of [FastAsyncWorldEdit for 1.16.1](https://ci.athion.net/job/FastAsyncWorldEdit-1.16/) - [X] I made sure the bug/error is not caused by any other plugin
1.0
jumpto + coordinate not working - # Bug report for FastAsyncWorldEdit 1.16.1 If I use `/jumpto` all works fine. But with `/jumpto [world,x,y,z]` I get the error: "error while handling a WorldEdit command" https://pastebin.com/KsEfh6uP **[REQUIRED] FastAsyncWorldEdit Configuration Files:** https://athion.net/ISPaster/paste/view/9d90979f779e4bb3bda8285de5b0d17a **[REQUIRED] FastAsyncWorldEdit Version Number:** <!--- Enter /fawe version in game or in your console and copy the full output here --> **[REQUIRED] Spigot/Paper Version Number:** > fawe 1.16 - 306 > PaperMC 1.15.2 **How to replicate**: `/jumpto world;12;12;12` `/jumpto world,12,12,12` **Checklist**: <!--- Make sure you've completed the following steps (put an "X" between of brackets): --> - [X] I included all information required in the sections above - [X] I made sure there are no duplicates of this report [(Use Search)](https://github.com/IntellectualSites/FastAsyncWorldEdit/issues?q=is%3Aissue) - [X] I made sure I am using an up-to-date version of [FastAsyncWorldEdit for 1.16.1](https://ci.athion.net/job/FastAsyncWorldEdit-1.16/) - [X] I made sure the bug/error is not caused by any other plugin
test
jumpto coordinate not working bug report for fastasyncworldedit if i use jumpto all works fine but with jumpto i get the error error while handling a worldedit command fastasyncworldedit configuration files fastasyncworldedit version number spigot paper version number fawe papermc how to replicate jumpto world jumpto world checklist i included all information required in the sections above i made sure there are no duplicates of this report i made sure i am using an up to date version of i made sure the bug error is not caused by any other plugin
1
79,462
7,716,550,328
IssuesEvent
2018-05-23 11:03:32
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
acceptance: TestDockerCLI/test_log_config_msg.tcl failed
A-cli C-test-failure
Seen in https://teamcity.cockroachdb.com/viewLog.html?buildId=641386&buildTypeId=Cockroach_UnitTests TestDockerCLI/test_log_config_msg.tcl [artifacts.log.zip](https://github.com/cockroachdb/cockroach/files/1976610/artifacts.log.zip) ``` .180504 11:14:10.925984999 EXPECT TEST: START TEST: Check that the cluster ID is reported at the start of new log files. child process exited abnormally while executing "system "grep '\\\[config\\\] clusterID:' logs/db/logs/cockroach.log"" (file "/go/src/github.com/cockroachdb/cockroach/cli/interactive_tests/test_log_config_msg.tcl" line 17) non-zero exit code: 1 cli_test.go:92: non-zero exit code: 1 ```
1.0
acceptance: TestDockerCLI/test_log_config_msg.tcl failed - Seen in https://teamcity.cockroachdb.com/viewLog.html?buildId=641386&buildTypeId=Cockroach_UnitTests TestDockerCLI/test_log_config_msg.tcl [artifacts.log.zip](https://github.com/cockroachdb/cockroach/files/1976610/artifacts.log.zip) ``` .180504 11:14:10.925984999 EXPECT TEST: START TEST: Check that the cluster ID is reported at the start of new log files. child process exited abnormally while executing "system "grep '\\\[config\\\] clusterID:' logs/db/logs/cockroach.log"" (file "/go/src/github.com/cockroachdb/cockroach/cli/interactive_tests/test_log_config_msg.tcl" line 17) non-zero exit code: 1 cli_test.go:92: non-zero exit code: 1 ```
test
acceptance testdockercli test log config msg tcl failed seen in testdockercli test log config msg tcl expect test start test check that the cluster id is reported at the start of new log files child process exited abnormally while executing system grep clusterid logs db logs cockroach log file go src github com cockroachdb cockroach cli interactive tests test log config msg tcl line non zero exit code cli test go non zero exit code
1
167,965
14,134,867,744
IssuesEvent
2020-11-10 00:17:27
directus/next
https://api.github.com/repos/directus/next
closed
Building Directus results in a ton of errors
Documentation Question
Hi there, I wanted to see what v9 looks like and maybe even help on some things, but when I tried to "npm run dev", I ran in some trouble. I downloaded the ZIP from this repo, installed node v14.15.0, then ran npm install (which succeeded), then tried to "npm run dev", which resulted in: ``` Valentino@VALENTINO-PC MINGW64 ~/Desktop/next-main $ npm run dev > directus-monorepo@ dev C:\Users\Valentino\Desktop\next-main > lerna run dev --stream --parallel lerna notice cli v3.22.1 lerna info Executing command in 4 packages: "npm run dev" directus: > directus@9.0.0-rc.8 dev C:\Users\Valentino\Desktop\next-main\api directus: > cross-env NODE_ENV=development LOG_LEVEL=trace ts-node-dev --files src/start.ts --respawn --watch "src/**/*.ts" --transpile-only @directus/app: > @directus/app@9.0.0-rc.8 dev C:\Users\Valentino\Desktop\next-main\app @directus/app: > vue-cli-service serve @directus/docs: > @directus/docs@9.0.0-rc.8 dev C:\Users\Valentino\Desktop\next-main\docs @directus/docs: > npm-watch build @directus/specs: > @directus/specs@9.0.0-rc.8 dev C:\Users\Valentino\Desktop\next-main\packages\spec @directus/specs: > npm-watch build directus: [INFO] 10:32:04 ts-node-dev ver. 1.0.0 (using ts-node ver. 9.0.0, typescript ver. 3.9.7) @directus/app: INFO Starting development server... @directus/docs: Building docs... @directus/docs: Built docs in 121 ms @directus/specs: Created dist\openapi.json from specs/openapi.yaml @directus/app: Starting type checking service... @directus/app: Using 1 worker with 2048MB memory limit directus: (node:9860) UnhandledPromiseRejectionWarning: Error: knex: Required configuration option 'client' is missing. directus: at new Client (C:\Users\Valentino\Desktop\next-main\node_modules\knex\lib\client.js:54:11) directus: at Object.Knex [as default] (C:\Users\Valentino\Desktop\next-main\node_modules\knex\lib\knex.js:53:28) directus: at Object.<anonymous> (C:\Users\Valentino\Desktop\next-main\api\src\database\index.ts:47:22) directus: at Module._compile (internal/modules/cjs/loader.js:1063:30) directus: at Module._compile (C:\Users\Valentino\Desktop\next-main\node_modules\source-map-support\source-map-support.js:547:25) directus: at Module.m._compile (C:\Users\VALENT~1\AppData\Local\Temp\ts-node-dev-hook-8387129263146487.js:57:33) directus: at Module._extensions..js (internal/modules/cjs/loader.js:1092:10) directus: at require.extensions.<computed> (C:\Users\VALENT~1\AppData\Local\Temp\ts-node-dev-hook-8387129263146487.js:59:20) directus: at Object.nodeDevHook [as .ts] (C:\Users\Valentino\Desktop\next-main\node_modules\ts-node-dev\lib\hook.js:63:13) directus: at Module.load (internal/modules/cjs/loader.js:928:32) directus: (Use `node --trace-warnings ...` to show where the warning was created) directus: (node:9860) UnhandledPromiseRejectionWarning: Unhandled promise rejection. This error originated either by throwing inside of an async function without a catch block, or by rejecting a promise which was not handled with .catch(). To terminate the node process on unhandled promise rejection, use the CLI flag `--unhandled-rejections=strict` (see https://nodejs.org/api/cli.html#cli_unhandled_rejections_mode). (rejection id: 1) directus: (node:9860) [DEP0018] DeprecationWarning: Unhandled promise rejections are deprecated. In the future, promise rejections that are not handled will terminate the Node.js process with a non-zero exit code. @directus/app: ERROR Failed to compile with 8 errors10:32:19 @directus/app: This dependency was not found: @directus/app: * @directus/format-title in ../node_modules/cache-loader/dist/cjs.js??ref--13-0!../node_modules/babel-loader/lib!../node_modules/ts-loader??ref--13-2!../node_modules/cache-loader/dist/cjs.js??ref--0-0!../node_modules/vue-loader/lib??vue-loader-options!./src/components/v-field-template/field-list-item.vue?vue&type=script&lang=ts&, ../node_modules/cache-loader/dist/cjs.js??ref--13-0!../node_modules/babel-loader/lib!../node_modules/ts-loader??ref--13-2!../node_modules/cache-loader/dist/cjs.js??ref--0-0!../node_modules/vue-loader/lib??vue-loader-options!./src/displays/formatted-value/formatted-value.vue?vue&type=script&lang=ts& and 6 others @directus/app: To install it, you can run: npm install --save @directus/format-title @directus/app: Type checking in progress... @directus/app: WAIT Compiling...10:32:20 @directus/app: Error from chokidar (C:\): Error: EBUSY: resource busy or locked, lstat 'C:\DumpStack.log.tmp' @directus/app: ERROR Failed to compile with 8 errors10:32:24 @directus/app: This dependency was not found: @directus/app: * @directus/format-title in ../node_modules/cache-loader/dist/cjs.js??ref--13-0!../node_modules/babel-loader/lib!../node_modules/ts-loader??ref--13-2!../node_modules/cache-loader/dist/cjs.js??ref--0-0!../node_modules/vue-loader/lib??vue-loader-options!./src/components/v-field-template/field-list-item.vue?vue&type=script&lang=ts&, ../node_modules/cache-loader/dist/cjs.js??ref--13-0!../node_modules/babel-loader/lib!../node_modules/ts-loader??ref--13-2!../node_modules/cache-loader/dist/cjs.js??ref--0-0!../node_modules/vue-loader/lib??vue-loader-options!./src/displays/formatted-value/formatted-value.vue?vue&type=script&lang=ts& and 6 others @directus/app: To install it, you can run: npm install --save @directus/format-title @directus/app: Type checking in progress... @directus/app: WAIT Compiling...10:32:24 @directus/app: ERROR Failed to compile with 8 errors10:32:25 @directus/app: This dependency was not found: @directus/app: * @directus/format-title in ../node_modules/cache-loader/dist/cjs.js??ref--13-0!../node_modules/babel-loader/lib!../node_modules/ts-loader??ref--13-2!../node_modules/cache-loader/dist/cjs.js??ref--0-0!../node_modules/vue-loader/lib??vue-loader-options!./src/components/v-field-template/field-list-item.vue?vue&type=script&lang=ts&, ../node_modules/cache-loader/dist/cjs.js??ref--13-0!../node_modules/babel-loader/lib!../node_modules/ts-loader??ref--13-2!../node_modules/cache-loader/dist/cjs.js??ref--0-0!../node_modules/vue-loader/lib??vue-loader-options!./src/displays/formatted-value/formatted-value.vue?vue&type=script&lang=ts& and 6 others @directus/app: To install it, you can run: npm install --save @directus/format-title @directus/app: Type checking in progress... @directus/app: ERROR in C:/Users/Valentino/Desktop/next-main/app/src/components/v-field-template/field-list-item.vue(25,25): @directus/app: 25:25 Cannot find module '@directus/format-title' or its corresponding type declarations. @directus/app: 23 | import { defineComponent, PropType } from '@vue/composition-api'; @directus/app: 24 | import { FieldTree } from './types'; @directus/app: > 25 | import formatTitle from '@directus/format-title'; @directus/app: | ^ @directus/app: 26 | @directus/app: 27 | export default defineComponent({ @directus/app: 28 | name: 'field-list-item', @directus/app: Version: typescript 3.9.7 @directus/app: Time: 14808ms @directus/app: ERROR in C:/Users/Valentino/Desktop/next-main/app/src/displays/color/color.vue(10,25): @directus/app: 10:25 Cannot find module '@directus/format-title' or its corresponding type declarations. @directus/app: 8 | <script lang="ts"> @directus/app: 9 | import { defineComponent, computed, PropType } from '@vue/composition-api'; @directus/app: > 10 | import formatTitle from '@directus/format-title'; @directus/app: | ^ @directus/app: 11 | import Color from 'color'; @directus/app: 12 | import { isHex } from '@/utils/color'; @directus/app: 13 | @directus/app: ERROR in C:/Users/Valentino/Desktop/next-main/app/src/displays/formatted-value/formatted-value.vue(11,25): @directus/app: 11:25 Cannot find module '@directus/format-title' or its corresponding type declarations. @directus/app: 9 | <script lang="ts"> @directus/app: 10 | import { defineComponent, computed } from '@vue/composition-api'; @directus/app: > 11 | import formatTitle from '@directus/format-title'; @directus/app: | ^ @directus/app: 12 | @directus/app: 13 | export default defineComponent({ @directus/app: 14 | props: { @directus/app: ERROR in C:/Users/Valentino/Desktop/next-main/app/src/displays/labels/labels.vue(32,25): @directus/app: 32:25 Cannot find module '@directus/format-title' or its corresponding type declarations. @directus/app: 30 | <script lang="ts"> @directus/app: 31 | import { defineComponent, computed, PropType } from '@vue/composition-api'; @directus/app: > 32 | import formatTitle from '@directus/format-title'; @directus/app: | ^ @directus/app: 33 | @directus/app: 34 | type Choice = { @directus/app: 35 | value: string; @directus/app: ERROR in C:/Users/Valentino/Desktop/next-main/app/src/interfaces/code/code.vue(28,25): @directus/app: 28:25 Cannot find module '@directus/format-title' or its corresponding type declarations. @directus/app: 26 | import 'codemirror/keymap/sublime.js'; @directus/app: 27 | @directus/app: > 28 | import formatTitle from '@directus/format-title'; @directus/app: | ^ @directus/app: 29 | @directus/app: 30 | export default defineComponent({ @directus/app: 31 | props: { @directus/app: ERROR in C:/Users/Valentino/Desktop/next-main/app/src/interfaces/icon/icon.vue(51,25): @directus/app: 51:25 Cannot find module '@directus/format-title' or its corresponding type declarations. @directus/app: 49 | import icons from './icons.json'; @directus/app: 50 | import { defineComponent, ref, computed } from '@vue/composition-api'; @directus/app: > 51 | import formatTitle from '@directus/format-title'; @directus/app: | ^ @directus/app: 52 | @directus/app: 53 | export default defineComponent({ @directus/app: 54 | props: { @directus/app: ERROR in C:/Users/Valentino/Desktop/next-main/app/src/interfaces/tags/tags.vue(46,25): @directus/app: 46:25 Cannot find module '@directus/format-title' or its corresponding type declarations. @directus/app: 44 | <script lang="ts"> @directus/app: 45 | import { defineComponent, PropType, ref, computed, watch } from '@vue/composition-api'; @directus/app: > 46 | import formatTitle from '@directus/format-title'; @directus/app: | ^ @directus/app: 47 | @directus/app: 48 | export default defineComponent({ @directus/app: 49 | props: { @directus/app: ERROR in C:/Users/Valentino/Desktop/next-main/app/src/stores/collections.ts(7,25): @directus/app: 7:25 Cannot find module '@directus/format-title' or its corresponding type declarations. @directus/app: 5 | import { notEmpty } from '@/utils/is-empty/'; @directus/app: 6 | import VueI18n from 'vue-i18n'; @directus/app: > 7 | import formatTitle from '@directus/format-title'; @directus/app: | ^ @directus/app: 8 | import { notify } from '@/utils/notify'; @directus/app: 9 | import { unexpectedError } from '@/utils/unexpected-error'; @directus/app: 10 | @directus/app: ERROR in C:/Users/Valentino/Desktop/next-main/app/src/stores/fields.ts(6,25): @directus/app: 6:25 Cannot find module '@directus/format-title' or its corresponding type declarations. @directus/app: 4 | import { notEmpty } from '@/utils/is-empty/'; @directus/app: 5 | import { i18n } from '@/lang'; @directus/app: > 6 | import formatTitle from '@directus/format-title'; @directus/app: | ^ @directus/app: 7 | import { useRelationsStore } from '@/stores/'; @directus/app: 8 | import { Relation, FieldRaw, Field } from '@/types'; @directus/app: 9 | import { merge } from 'lodash'; ``` Any clue on what I'm doing wrong here? EDIT: What I'm seeing a lot is: `Cannot find module '@directus/format-title' or its corresponding type declarations.` Should I try and install this module(?) separately? And if yes, why doesn't it come specified in package.json (dev-deps) so it gets installed automatically? I mean, I don't even know if that would solve it. There are a lot of other errors too. Also, if more information is needed, let me know. EDIT: `git clone <repo>` instead of downloading the ZIP results in the same errors. Thanks in advance!!
1.0
Building Directus results in a ton of errors - Hi there, I wanted to see what v9 looks like and maybe even help on some things, but when I tried to "npm run dev", I ran in some trouble. I downloaded the ZIP from this repo, installed node v14.15.0, then ran npm install (which succeeded), then tried to "npm run dev", which resulted in: ``` Valentino@VALENTINO-PC MINGW64 ~/Desktop/next-main $ npm run dev > directus-monorepo@ dev C:\Users\Valentino\Desktop\next-main > lerna run dev --stream --parallel lerna notice cli v3.22.1 lerna info Executing command in 4 packages: "npm run dev" directus: > directus@9.0.0-rc.8 dev C:\Users\Valentino\Desktop\next-main\api directus: > cross-env NODE_ENV=development LOG_LEVEL=trace ts-node-dev --files src/start.ts --respawn --watch "src/**/*.ts" --transpile-only @directus/app: > @directus/app@9.0.0-rc.8 dev C:\Users\Valentino\Desktop\next-main\app @directus/app: > vue-cli-service serve @directus/docs: > @directus/docs@9.0.0-rc.8 dev C:\Users\Valentino\Desktop\next-main\docs @directus/docs: > npm-watch build @directus/specs: > @directus/specs@9.0.0-rc.8 dev C:\Users\Valentino\Desktop\next-main\packages\spec @directus/specs: > npm-watch build directus: [INFO] 10:32:04 ts-node-dev ver. 1.0.0 (using ts-node ver. 9.0.0, typescript ver. 3.9.7) @directus/app: INFO Starting development server... @directus/docs: Building docs... @directus/docs: Built docs in 121 ms @directus/specs: Created dist\openapi.json from specs/openapi.yaml @directus/app: Starting type checking service... @directus/app: Using 1 worker with 2048MB memory limit directus: (node:9860) UnhandledPromiseRejectionWarning: Error: knex: Required configuration option 'client' is missing. directus: at new Client (C:\Users\Valentino\Desktop\next-main\node_modules\knex\lib\client.js:54:11) directus: at Object.Knex [as default] (C:\Users\Valentino\Desktop\next-main\node_modules\knex\lib\knex.js:53:28) directus: at Object.<anonymous> (C:\Users\Valentino\Desktop\next-main\api\src\database\index.ts:47:22) directus: at Module._compile (internal/modules/cjs/loader.js:1063:30) directus: at Module._compile (C:\Users\Valentino\Desktop\next-main\node_modules\source-map-support\source-map-support.js:547:25) directus: at Module.m._compile (C:\Users\VALENT~1\AppData\Local\Temp\ts-node-dev-hook-8387129263146487.js:57:33) directus: at Module._extensions..js (internal/modules/cjs/loader.js:1092:10) directus: at require.extensions.<computed> (C:\Users\VALENT~1\AppData\Local\Temp\ts-node-dev-hook-8387129263146487.js:59:20) directus: at Object.nodeDevHook [as .ts] (C:\Users\Valentino\Desktop\next-main\node_modules\ts-node-dev\lib\hook.js:63:13) directus: at Module.load (internal/modules/cjs/loader.js:928:32) directus: (Use `node --trace-warnings ...` to show where the warning was created) directus: (node:9860) UnhandledPromiseRejectionWarning: Unhandled promise rejection. This error originated either by throwing inside of an async function without a catch block, or by rejecting a promise which was not handled with .catch(). To terminate the node process on unhandled promise rejection, use the CLI flag `--unhandled-rejections=strict` (see https://nodejs.org/api/cli.html#cli_unhandled_rejections_mode). (rejection id: 1) directus: (node:9860) [DEP0018] DeprecationWarning: Unhandled promise rejections are deprecated. In the future, promise rejections that are not handled will terminate the Node.js process with a non-zero exit code. @directus/app: ERROR Failed to compile with 8 errors10:32:19 @directus/app: This dependency was not found: @directus/app: * @directus/format-title in ../node_modules/cache-loader/dist/cjs.js??ref--13-0!../node_modules/babel-loader/lib!../node_modules/ts-loader??ref--13-2!../node_modules/cache-loader/dist/cjs.js??ref--0-0!../node_modules/vue-loader/lib??vue-loader-options!./src/components/v-field-template/field-list-item.vue?vue&type=script&lang=ts&, ../node_modules/cache-loader/dist/cjs.js??ref--13-0!../node_modules/babel-loader/lib!../node_modules/ts-loader??ref--13-2!../node_modules/cache-loader/dist/cjs.js??ref--0-0!../node_modules/vue-loader/lib??vue-loader-options!./src/displays/formatted-value/formatted-value.vue?vue&type=script&lang=ts& and 6 others @directus/app: To install it, you can run: npm install --save @directus/format-title @directus/app: Type checking in progress... @directus/app: WAIT Compiling...10:32:20 @directus/app: Error from chokidar (C:\): Error: EBUSY: resource busy or locked, lstat 'C:\DumpStack.log.tmp' @directus/app: ERROR Failed to compile with 8 errors10:32:24 @directus/app: This dependency was not found: @directus/app: * @directus/format-title in ../node_modules/cache-loader/dist/cjs.js??ref--13-0!../node_modules/babel-loader/lib!../node_modules/ts-loader??ref--13-2!../node_modules/cache-loader/dist/cjs.js??ref--0-0!../node_modules/vue-loader/lib??vue-loader-options!./src/components/v-field-template/field-list-item.vue?vue&type=script&lang=ts&, ../node_modules/cache-loader/dist/cjs.js??ref--13-0!../node_modules/babel-loader/lib!../node_modules/ts-loader??ref--13-2!../node_modules/cache-loader/dist/cjs.js??ref--0-0!../node_modules/vue-loader/lib??vue-loader-options!./src/displays/formatted-value/formatted-value.vue?vue&type=script&lang=ts& and 6 others @directus/app: To install it, you can run: npm install --save @directus/format-title @directus/app: Type checking in progress... @directus/app: WAIT Compiling...10:32:24 @directus/app: ERROR Failed to compile with 8 errors10:32:25 @directus/app: This dependency was not found: @directus/app: * @directus/format-title in ../node_modules/cache-loader/dist/cjs.js??ref--13-0!../node_modules/babel-loader/lib!../node_modules/ts-loader??ref--13-2!../node_modules/cache-loader/dist/cjs.js??ref--0-0!../node_modules/vue-loader/lib??vue-loader-options!./src/components/v-field-template/field-list-item.vue?vue&type=script&lang=ts&, ../node_modules/cache-loader/dist/cjs.js??ref--13-0!../node_modules/babel-loader/lib!../node_modules/ts-loader??ref--13-2!../node_modules/cache-loader/dist/cjs.js??ref--0-0!../node_modules/vue-loader/lib??vue-loader-options!./src/displays/formatted-value/formatted-value.vue?vue&type=script&lang=ts& and 6 others @directus/app: To install it, you can run: npm install --save @directus/format-title @directus/app: Type checking in progress... @directus/app: ERROR in C:/Users/Valentino/Desktop/next-main/app/src/components/v-field-template/field-list-item.vue(25,25): @directus/app: 25:25 Cannot find module '@directus/format-title' or its corresponding type declarations. @directus/app: 23 | import { defineComponent, PropType } from '@vue/composition-api'; @directus/app: 24 | import { FieldTree } from './types'; @directus/app: > 25 | import formatTitle from '@directus/format-title'; @directus/app: | ^ @directus/app: 26 | @directus/app: 27 | export default defineComponent({ @directus/app: 28 | name: 'field-list-item', @directus/app: Version: typescript 3.9.7 @directus/app: Time: 14808ms @directus/app: ERROR in C:/Users/Valentino/Desktop/next-main/app/src/displays/color/color.vue(10,25): @directus/app: 10:25 Cannot find module '@directus/format-title' or its corresponding type declarations. @directus/app: 8 | <script lang="ts"> @directus/app: 9 | import { defineComponent, computed, PropType } from '@vue/composition-api'; @directus/app: > 10 | import formatTitle from '@directus/format-title'; @directus/app: | ^ @directus/app: 11 | import Color from 'color'; @directus/app: 12 | import { isHex } from '@/utils/color'; @directus/app: 13 | @directus/app: ERROR in C:/Users/Valentino/Desktop/next-main/app/src/displays/formatted-value/formatted-value.vue(11,25): @directus/app: 11:25 Cannot find module '@directus/format-title' or its corresponding type declarations. @directus/app: 9 | <script lang="ts"> @directus/app: 10 | import { defineComponent, computed } from '@vue/composition-api'; @directus/app: > 11 | import formatTitle from '@directus/format-title'; @directus/app: | ^ @directus/app: 12 | @directus/app: 13 | export default defineComponent({ @directus/app: 14 | props: { @directus/app: ERROR in C:/Users/Valentino/Desktop/next-main/app/src/displays/labels/labels.vue(32,25): @directus/app: 32:25 Cannot find module '@directus/format-title' or its corresponding type declarations. @directus/app: 30 | <script lang="ts"> @directus/app: 31 | import { defineComponent, computed, PropType } from '@vue/composition-api'; @directus/app: > 32 | import formatTitle from '@directus/format-title'; @directus/app: | ^ @directus/app: 33 | @directus/app: 34 | type Choice = { @directus/app: 35 | value: string; @directus/app: ERROR in C:/Users/Valentino/Desktop/next-main/app/src/interfaces/code/code.vue(28,25): @directus/app: 28:25 Cannot find module '@directus/format-title' or its corresponding type declarations. @directus/app: 26 | import 'codemirror/keymap/sublime.js'; @directus/app: 27 | @directus/app: > 28 | import formatTitle from '@directus/format-title'; @directus/app: | ^ @directus/app: 29 | @directus/app: 30 | export default defineComponent({ @directus/app: 31 | props: { @directus/app: ERROR in C:/Users/Valentino/Desktop/next-main/app/src/interfaces/icon/icon.vue(51,25): @directus/app: 51:25 Cannot find module '@directus/format-title' or its corresponding type declarations. @directus/app: 49 | import icons from './icons.json'; @directus/app: 50 | import { defineComponent, ref, computed } from '@vue/composition-api'; @directus/app: > 51 | import formatTitle from '@directus/format-title'; @directus/app: | ^ @directus/app: 52 | @directus/app: 53 | export default defineComponent({ @directus/app: 54 | props: { @directus/app: ERROR in C:/Users/Valentino/Desktop/next-main/app/src/interfaces/tags/tags.vue(46,25): @directus/app: 46:25 Cannot find module '@directus/format-title' or its corresponding type declarations. @directus/app: 44 | <script lang="ts"> @directus/app: 45 | import { defineComponent, PropType, ref, computed, watch } from '@vue/composition-api'; @directus/app: > 46 | import formatTitle from '@directus/format-title'; @directus/app: | ^ @directus/app: 47 | @directus/app: 48 | export default defineComponent({ @directus/app: 49 | props: { @directus/app: ERROR in C:/Users/Valentino/Desktop/next-main/app/src/stores/collections.ts(7,25): @directus/app: 7:25 Cannot find module '@directus/format-title' or its corresponding type declarations. @directus/app: 5 | import { notEmpty } from '@/utils/is-empty/'; @directus/app: 6 | import VueI18n from 'vue-i18n'; @directus/app: > 7 | import formatTitle from '@directus/format-title'; @directus/app: | ^ @directus/app: 8 | import { notify } from '@/utils/notify'; @directus/app: 9 | import { unexpectedError } from '@/utils/unexpected-error'; @directus/app: 10 | @directus/app: ERROR in C:/Users/Valentino/Desktop/next-main/app/src/stores/fields.ts(6,25): @directus/app: 6:25 Cannot find module '@directus/format-title' or its corresponding type declarations. @directus/app: 4 | import { notEmpty } from '@/utils/is-empty/'; @directus/app: 5 | import { i18n } from '@/lang'; @directus/app: > 6 | import formatTitle from '@directus/format-title'; @directus/app: | ^ @directus/app: 7 | import { useRelationsStore } from '@/stores/'; @directus/app: 8 | import { Relation, FieldRaw, Field } from '@/types'; @directus/app: 9 | import { merge } from 'lodash'; ``` Any clue on what I'm doing wrong here? EDIT: What I'm seeing a lot is: `Cannot find module '@directus/format-title' or its corresponding type declarations.` Should I try and install this module(?) separately? And if yes, why doesn't it come specified in package.json (dev-deps) so it gets installed automatically? I mean, I don't even know if that would solve it. There are a lot of other errors too. Also, if more information is needed, let me know. EDIT: `git clone <repo>` instead of downloading the ZIP results in the same errors. Thanks in advance!!
non_test
building directus results in a ton of errors hi there i wanted to see what looks like and maybe even help on some things but when i tried to npm run dev i ran in some trouble i downloaded the zip from this repo installed node then ran npm install which succeeded then tried to npm run dev which resulted in valentino valentino pc desktop next main npm run dev directus monorepo dev c users valentino desktop next main lerna run dev stream parallel lerna notice cli lerna info executing command in packages npm run dev directus directus rc dev c users valentino desktop next main api directus cross env node env development log level trace ts node dev files src start ts respawn watch src ts transpile only directus app directus app rc dev c users valentino desktop next main app directus app vue cli service serve directus docs directus docs rc dev c users valentino desktop next main docs directus docs npm watch build directus specs directus specs rc dev c users valentino desktop next main packages spec directus specs npm watch build directus ts node dev ver using ts node ver typescript ver directus app info starting development server directus docs building docs directus docs built docs in ms directus specs created dist openapi json from specs openapi yaml directus app starting type checking service directus app using worker with memory limit directus node unhandledpromiserejectionwarning error knex required configuration option client is missing directus at new client c users valentino desktop next main node modules knex lib client js directus at object knex c users valentino desktop next main node modules knex lib knex js directus at object c users valentino desktop next main api src database index ts directus at module compile internal modules cjs loader js directus at module compile c users valentino desktop next main node modules source map support source map support js directus at module m compile c users valent appdata local temp ts node dev hook js directus at module extensions js internal modules cjs loader js directus at require extensions c users valent appdata local temp ts node dev hook js directus at object nodedevhook c users valentino desktop next main node modules ts node dev lib hook js directus at module load internal modules cjs loader js directus use node trace warnings to show where the warning was created directus node unhandledpromiserejectionwarning unhandled promise rejection this error originated either by throwing inside of an async function without a catch block or by rejecting a promise which was not handled with catch to terminate the node process on unhandled promise rejection use the cli flag unhandled rejections strict see rejection id directus node deprecationwarning unhandled promise rejections are deprecated in the future promise rejections that are not handled will terminate the node js process with a non zero exit code directus app error failed to compile with directus app this dependency was not found directus app directus format title in node modules cache loader dist cjs js ref node modules babel loader lib node modules ts loader ref node modules cache loader dist cjs js ref node modules vue loader lib vue loader options src components v field template field list item vue vue type script lang ts node modules cache loader dist cjs js ref node modules babel loader lib node modules ts loader ref node modules cache loader dist cjs js ref node modules vue loader lib vue loader options src displays formatted value formatted value vue vue type script lang ts and others directus app to install it you can run npm install save directus format title directus app type checking in progress directus app wait compiling directus app error from chokidar c error ebusy resource busy or locked lstat c dumpstack log tmp directus app error failed to compile with directus app this dependency was not found directus app directus format title in node modules cache loader dist cjs js ref node modules babel loader lib node modules ts loader ref node modules cache loader dist cjs js ref node modules vue loader lib vue loader options src components v field template field list item vue vue type script lang ts node modules cache loader dist cjs js ref node modules babel loader lib node modules ts loader ref node modules cache loader dist cjs js ref node modules vue loader lib vue loader options src displays formatted value formatted value vue vue type script lang ts and others directus app to install it you can run npm install save directus format title directus app type checking in progress directus app wait compiling directus app error failed to compile with directus app this dependency was not found directus app directus format title in node modules cache loader dist cjs js ref node modules babel loader lib node modules ts loader ref node modules cache loader dist cjs js ref node modules vue loader lib vue loader options src components v field template field list item vue vue type script lang ts node modules cache loader dist cjs js ref node modules babel loader lib node modules ts loader ref node modules cache loader dist cjs js ref node modules vue loader lib vue loader options src displays formatted value formatted value vue vue type script lang ts and others directus app to install it you can run npm install save directus format title directus app type checking in progress directus app error in c users valentino desktop next main app src components v field template field list item vue directus app cannot find module directus format title or its corresponding type declarations directus app import definecomponent proptype from vue composition api directus app import fieldtree from types directus app import formattitle from directus format title directus app directus app directus app export default definecomponent directus app name field list item directus app version typescript directus app time directus app error in c users valentino desktop next main app src displays color color vue directus app cannot find module directus format title or its corresponding type declarations directus app directus app import definecomponent computed proptype from vue composition api directus app import formattitle from directus format title directus app directus app import color from color directus app import ishex from utils color directus app directus app error in c users valentino desktop next main app src displays formatted value formatted value vue directus app cannot find module directus format title or its corresponding type declarations directus app directus app import definecomponent computed from vue composition api directus app import formattitle from directus format title directus app directus app directus app export default definecomponent directus app props directus app error in c users valentino desktop next main app src displays labels labels vue directus app cannot find module directus format title or its corresponding type declarations directus app directus app import definecomponent computed proptype from vue composition api directus app import formattitle from directus format title directus app directus app directus app type choice directus app value string directus app error in c users valentino desktop next main app src interfaces code code vue directus app cannot find module directus format title or its corresponding type declarations directus app import codemirror keymap sublime js directus app directus app import formattitle from directus format title directus app directus app directus app export default definecomponent directus app props directus app error in c users valentino desktop next main app src interfaces icon icon vue directus app cannot find module directus format title or its corresponding type declarations directus app import icons from icons json directus app import definecomponent ref computed from vue composition api directus app import formattitle from directus format title directus app directus app directus app export default definecomponent directus app props directus app error in c users valentino desktop next main app src interfaces tags tags vue directus app cannot find module directus format title or its corresponding type declarations directus app directus app import definecomponent proptype ref computed watch from vue composition api directus app import formattitle from directus format title directus app directus app directus app export default definecomponent directus app props directus app error in c users valentino desktop next main app src stores collections ts directus app cannot find module directus format title or its corresponding type declarations directus app import notempty from utils is empty directus app import from vue directus app import formattitle from directus format title directus app directus app import notify from utils notify directus app import unexpectederror from utils unexpected error directus app directus app error in c users valentino desktop next main app src stores fields ts directus app cannot find module directus format title or its corresponding type declarations directus app import notempty from utils is empty directus app import from lang directus app import formattitle from directus format title directus app directus app import userelationsstore from stores directus app import relation fieldraw field from types directus app import merge from lodash any clue on what i m doing wrong here edit what i m seeing a lot is cannot find module directus format title or its corresponding type declarations should i try and install this module separately and if yes why doesn t it come specified in package json dev deps so it gets installed automatically i mean i don t even know if that would solve it there are a lot of other errors too also if more information is needed let me know edit git clone instead of downloading the zip results in the same errors thanks in advance
0
136,859
11,089,847,842
IssuesEvent
2019-12-14 21:33:20
nodejs/node
https://api.github.com/repos/nodejs/node
closed
Investigate flaky test parallel/test-worker-message-port-transfer-terminate
CI / flaky test worker
<!-- Thank you for reporting a possible bug in Node.js. Please fill in as much of the template below as you can. Version: output of `node -v` Platform: output of `uname -a` (UNIX), or version and 32 or 64-bit (Windows) Subsystem: if known, please specify the affected core module name If possible, please provide code that demonstrates the problem, keeping it as simple and free of external dependencies as you can. --> * **Version**: master * **Platform**: Windows * **Subsystem**: test <!-- Please provide more details below this comment. --> ``` foo[7636]: C:\workspace\node-compile-windows\node\src\node_platform.cc:275: Assertion `!flush_tasks_' failed. 1: 00007FF6FAA6E2BF v8::internal::wasm::DisjointAllocationPool::~DisjointAllocationPool+89903 2: 00007FF6FAA0C2B6 v8::base::CPU::has_sse+35974 3: 00007FF6FAA0C5D3 v8::base::CPU::has_sse+36771 4: 00007FF6FA9B84D2 v8::internal::wasm::JSToWasmWrapperCompilationUnit::~JSToWasmWrapperCompilationUnit+148338 5: 00007FF6FA9B977B v8::internal::wasm::JSToWasmWrapperCompilationUnit::~JSToWasmWrapperCompilationUnit+153115 6: 00007FF6FA9902BB v8::internal::interpreter::BytecodeLabel::bind+4379 7: 00007FF6FA9928E0 v8::internal::interpreter::BytecodeLabel::bind+14144 8: 00007FF6FA98F169 v8::internal::LookupIterator::index+92601 9: 00007FF6FAAA8F0D uv_poll_stop+765 10: 00007FF6FB728230 v8::internal::SetupIsolateDelegate::SetupHeap+1526544 11: 00007FF96DF57BD4 BaseThreadInitThunk+20 12: 00007FF96FB0CED1 RtlUserThreadStart+33 ``` On: https://ci.nodejs.org/computer/test-azure_msft-win10_vcbt2015-x64-4/ Ref: https://ci.nodejs.org/job/node-test-binary-windows-js-suites/69/RUN_SUBSET=1,nodes=win10-COMPILED_BY-vs2019/testReport/junit/(root)/test/parallel_test_worker_message_port_transfer_terminate/
1.0
Investigate flaky test parallel/test-worker-message-port-transfer-terminate - <!-- Thank you for reporting a possible bug in Node.js. Please fill in as much of the template below as you can. Version: output of `node -v` Platform: output of `uname -a` (UNIX), or version and 32 or 64-bit (Windows) Subsystem: if known, please specify the affected core module name If possible, please provide code that demonstrates the problem, keeping it as simple and free of external dependencies as you can. --> * **Version**: master * **Platform**: Windows * **Subsystem**: test <!-- Please provide more details below this comment. --> ``` foo[7636]: C:\workspace\node-compile-windows\node\src\node_platform.cc:275: Assertion `!flush_tasks_' failed. 1: 00007FF6FAA6E2BF v8::internal::wasm::DisjointAllocationPool::~DisjointAllocationPool+89903 2: 00007FF6FAA0C2B6 v8::base::CPU::has_sse+35974 3: 00007FF6FAA0C5D3 v8::base::CPU::has_sse+36771 4: 00007FF6FA9B84D2 v8::internal::wasm::JSToWasmWrapperCompilationUnit::~JSToWasmWrapperCompilationUnit+148338 5: 00007FF6FA9B977B v8::internal::wasm::JSToWasmWrapperCompilationUnit::~JSToWasmWrapperCompilationUnit+153115 6: 00007FF6FA9902BB v8::internal::interpreter::BytecodeLabel::bind+4379 7: 00007FF6FA9928E0 v8::internal::interpreter::BytecodeLabel::bind+14144 8: 00007FF6FA98F169 v8::internal::LookupIterator::index+92601 9: 00007FF6FAAA8F0D uv_poll_stop+765 10: 00007FF6FB728230 v8::internal::SetupIsolateDelegate::SetupHeap+1526544 11: 00007FF96DF57BD4 BaseThreadInitThunk+20 12: 00007FF96FB0CED1 RtlUserThreadStart+33 ``` On: https://ci.nodejs.org/computer/test-azure_msft-win10_vcbt2015-x64-4/ Ref: https://ci.nodejs.org/job/node-test-binary-windows-js-suites/69/RUN_SUBSET=1,nodes=win10-COMPILED_BY-vs2019/testReport/junit/(root)/test/parallel_test_worker_message_port_transfer_terminate/
test
investigate flaky test parallel test worker message port transfer terminate thank you for reporting a possible bug in node js please fill in as much of the template below as you can version output of node v platform output of uname a unix or version and or bit windows subsystem if known please specify the affected core module name if possible please provide code that demonstrates the problem keeping it as simple and free of external dependencies as you can version master platform windows subsystem test foo c workspace node compile windows node src node platform cc assertion flush tasks failed internal wasm disjointallocationpool disjointallocationpool base cpu has sse base cpu has sse internal wasm jstowasmwrappercompilationunit jstowasmwrappercompilationunit internal wasm jstowasmwrappercompilationunit jstowasmwrappercompilationunit internal interpreter bytecodelabel bind internal interpreter bytecodelabel bind internal lookupiterator index uv poll stop internal setupisolatedelegate setupheap basethreadinitthunk rtluserthreadstart on ref
1
404,006
27,446,229,200
IssuesEvent
2023-03-02 14:28:01
alphagov/govuk-prototype-kit
https://api.github.com/repos/alphagov/govuk-prototype-kit
closed
Document how to update a prototype that is already on GitHub
documentation 🕔 days
## Related documentation https://govuk-prototype-kit.herokuapp.com/docs/updating-the-kit ## Suggestion We should consider adding some extra instructions to the update tutorial if the prototype was already on GitHub so that other people who have a local copy of the prototype get the changed files. ### Steps to add (draft) - Branch off the `main` (or equivalent) branch and pull down the latest changes [...] - Once the update has completed locally, commit the files and open a PR etc., depending on the process that the team working on the prototype follows ## Evidence (where applicable) A user who was updating prototypes that were already on GitHub said that it wasn't clear that after the update was complete, they needed to commit and push up the changed files.
1.0
Document how to update a prototype that is already on GitHub - ## Related documentation https://govuk-prototype-kit.herokuapp.com/docs/updating-the-kit ## Suggestion We should consider adding some extra instructions to the update tutorial if the prototype was already on GitHub so that other people who have a local copy of the prototype get the changed files. ### Steps to add (draft) - Branch off the `main` (or equivalent) branch and pull down the latest changes [...] - Once the update has completed locally, commit the files and open a PR etc., depending on the process that the team working on the prototype follows ## Evidence (where applicable) A user who was updating prototypes that were already on GitHub said that it wasn't clear that after the update was complete, they needed to commit and push up the changed files.
non_test
document how to update a prototype that is already on github related documentation suggestion we should consider adding some extra instructions to the update tutorial if the prototype was already on github so that other people who have a local copy of the prototype get the changed files steps to add draft branch off the main or equivalent branch and pull down the latest changes once the update has completed locally commit the files and open a pr etc depending on the process that the team working on the prototype follows evidence where applicable a user who was updating prototypes that were already on github said that it wasn t clear that after the update was complete they needed to commit and push up the changed files
0
155,570
12,261,436,402
IssuesEvent
2020-05-06 20:05:59
ansible/awx
https://api.github.com/repos/ansible/awx
closed
[Foreman Inventory] satellite6_group_patterns
component:api priority:high state:needs_test type:task
I have talked with @wtome about this and we have an idea of how this can be done. The solution is gnarly, and there may still be problems with it, so not trivial. Probably a few hours of work to confirm that the approach does in fact work and actually add it in.
1.0
[Foreman Inventory] satellite6_group_patterns - I have talked with @wtome about this and we have an idea of how this can be done. The solution is gnarly, and there may still be problems with it, so not trivial. Probably a few hours of work to confirm that the approach does in fact work and actually add it in.
test
group patterns i have talked with wtome about this and we have an idea of how this can be done the solution is gnarly and there may still be problems with it so not trivial probably a few hours of work to confirm that the approach does in fact work and actually add it in
1
247,389
26,706,245,243
IssuesEvent
2023-01-27 18:26:22
CDCgov/prime-simplereport
https://api.github.com/repos/CDCgov/prime-simplereport
closed
SPIKE Improve secret management
DevOps Security Better Data P3 Eng Health
Currently our secrets are pulled our of azure key vault and then passed through to the application via environment variables. This process can be improved by giving our application access to azure key vault directly Advantages - Less development work to pass variables through to the application layer - Local development secrets can be managed through the azure CLI authentication instead of `application-local.yml` After we move to azure key vault for local dev we should refresh all secrets that have been on developer machines ### Reading https://docs.microsoft.com/en-us/azure/developer/java/spring-framework/configure-spring-boot-starter-java-app-with-azure-key-vault Note: the article uses a package `azure-webapp-maven-plugin`. We will need investigate the third party gradle plugin https://github.com/lenala/azure-gradle-plugins or find another way ### Outcome have a plan and detailed ticket on how the Azure CLI should work with our application locally. Time box spend no more then 1 day
True
SPIKE Improve secret management - Currently our secrets are pulled our of azure key vault and then passed through to the application via environment variables. This process can be improved by giving our application access to azure key vault directly Advantages - Less development work to pass variables through to the application layer - Local development secrets can be managed through the azure CLI authentication instead of `application-local.yml` After we move to azure key vault for local dev we should refresh all secrets that have been on developer machines ### Reading https://docs.microsoft.com/en-us/azure/developer/java/spring-framework/configure-spring-boot-starter-java-app-with-azure-key-vault Note: the article uses a package `azure-webapp-maven-plugin`. We will need investigate the third party gradle plugin https://github.com/lenala/azure-gradle-plugins or find another way ### Outcome have a plan and detailed ticket on how the Azure CLI should work with our application locally. Time box spend no more then 1 day
non_test
spike improve secret management currently our secrets are pulled our of azure key vault and then passed through to the application via environment variables this process can be improved by giving our application access to azure key vault directly advantages less development work to pass variables through to the application layer local development secrets can be managed through the azure cli authentication instead of application local yml after we move to azure key vault for local dev we should refresh all secrets that have been on developer machines reading note the article uses a package azure webapp maven plugin we will need investigate the third party gradle plugin or find another way outcome have a plan and detailed ticket on how the azure cli should work with our application locally time box spend no more then day
0
287,655
8,817,982,631
IssuesEvent
2018-12-31 07:41:17
telerik/kendo-ui-core
https://api.github.com/repos/telerik/kendo-ui-core
opened
Resizing RTL Grid with filter row throws JavaScript error
Bug C: Grid Kendo1 Priority 1 SEV: Low
### Reproduction of the problem 1. Run dojo - https://dojo.telerik.com/UyOroxin 2. Try to resize a column using the handler in the filter row. ### Current behavior JavaScript error is shown. https://www.screencast.com/t/D60Ndpgss9pT ### Expected/desired behavior Resizing should work the same as in non-rtl mode. ### Environment * **Kendo UI version:** 2018.3.1017
1.0
Resizing RTL Grid with filter row throws JavaScript error - ### Reproduction of the problem 1. Run dojo - https://dojo.telerik.com/UyOroxin 2. Try to resize a column using the handler in the filter row. ### Current behavior JavaScript error is shown. https://www.screencast.com/t/D60Ndpgss9pT ### Expected/desired behavior Resizing should work the same as in non-rtl mode. ### Environment * **Kendo UI version:** 2018.3.1017
non_test
resizing rtl grid with filter row throws javascript error reproduction of the problem run dojo try to resize a column using the handler in the filter row current behavior javascript error is shown expected desired behavior resizing should work the same as in non rtl mode environment kendo ui version
0
209,044
15,964,478,096
IssuesEvent
2021-04-16 06:17:58
bosagora/agora
https://api.github.com/repos/bosagora/agora
closed
Test-suite sometimes fails with an odd error
C.General type-bug type-testing
``` Run ./ci/run.sh + set -o pipefail + dub test -b unittest-cov --skip-registry=all --compiler=ldc2 Executable configuration "unittest" of package agora defines no main source file, this may cause certain build modes to fail. Add an explicit "mainSourceFile" to the package description to fix this. Running custom 'unittest' configuration. Running pre-generate commands for vibe-d:tls... Running pre-generate commands for agora... Failed to execute program (No such file or directory) Command failed with exit code 2: /Users/runner/hostedtoolcache/dc/ldc2-4eaa2fd8/x64/ldc2-4eaa2fd8-osx-x86_64/bin/dub source/scpp/build.d ``` This is a strange one, it seems like there's a race condition somewhere where we try to run an executable which doesn't exist yet. I see it happen more and more often.
1.0
Test-suite sometimes fails with an odd error - ``` Run ./ci/run.sh + set -o pipefail + dub test -b unittest-cov --skip-registry=all --compiler=ldc2 Executable configuration "unittest" of package agora defines no main source file, this may cause certain build modes to fail. Add an explicit "mainSourceFile" to the package description to fix this. Running custom 'unittest' configuration. Running pre-generate commands for vibe-d:tls... Running pre-generate commands for agora... Failed to execute program (No such file or directory) Command failed with exit code 2: /Users/runner/hostedtoolcache/dc/ldc2-4eaa2fd8/x64/ldc2-4eaa2fd8-osx-x86_64/bin/dub source/scpp/build.d ``` This is a strange one, it seems like there's a race condition somewhere where we try to run an executable which doesn't exist yet. I see it happen more and more often.
test
test suite sometimes fails with an odd error run ci run sh set o pipefail dub test b unittest cov skip registry all compiler executable configuration unittest of package agora defines no main source file this may cause certain build modes to fail add an explicit mainsourcefile to the package description to fix this running custom unittest configuration running pre generate commands for vibe d tls running pre generate commands for agora failed to execute program no such file or directory command failed with exit code users runner hostedtoolcache dc osx bin dub source scpp build d this is a strange one it seems like there s a race condition somewhere where we try to run an executable which doesn t exist yet i see it happen more and more often
1
345,925
30,853,412,569
IssuesEvent
2023-08-02 18:32:43
microsoft/AzureStorageExplorer
https://api.github.com/repos/microsoft/AzureStorageExplorer
closed
There is no activity log when downloading audio files in preview panel
🧪 testing
**Storage Explorer Version:** 1.31.0-dev **Build Number:** 20230802.3 **Branch:** main **Platform/OS:** Windows 10/Linux Ubuntu 20.04/MacOS Ventura 13.5 (Apple M1 Pro) **Architecture:** x64/x64/arm64 **How Found:** From running test cases **Regression From:** Not a regression ## Steps to Reproduce ## 1. Expand one storage account -> Blob Containers. 2. Create a blob container -> Upload an audio file -> Click 'Preview'. 3. Go to the preview panel -> Click '**⋮**' button. 4. Select 'Download' -> Click 'Save'. 5. Check whether there is a download activity log. ## Expected Experience ## There is a download activity log. ## Actual Experience ## There is no activity log.
1.0
There is no activity log when downloading audio files in preview panel - **Storage Explorer Version:** 1.31.0-dev **Build Number:** 20230802.3 **Branch:** main **Platform/OS:** Windows 10/Linux Ubuntu 20.04/MacOS Ventura 13.5 (Apple M1 Pro) **Architecture:** x64/x64/arm64 **How Found:** From running test cases **Regression From:** Not a regression ## Steps to Reproduce ## 1. Expand one storage account -> Blob Containers. 2. Create a blob container -> Upload an audio file -> Click 'Preview'. 3. Go to the preview panel -> Click '**⋮**' button. 4. Select 'Download' -> Click 'Save'. 5. Check whether there is a download activity log. ## Expected Experience ## There is a download activity log. ## Actual Experience ## There is no activity log.
test
there is no activity log when downloading audio files in preview panel storage explorer version dev build number branch main platform os windows linux ubuntu macos ventura apple pro architecture how found from running test cases regression from not a regression steps to reproduce expand one storage account blob containers create a blob container upload an audio file click preview go to the preview panel click ⋮ button select download click save check whether there is a download activity log expected experience there is a download activity log actual experience there is no activity log
1
99,432
8,700,534,621
IssuesEvent
2018-12-05 09:01:58
SME-Issues/issues
https://api.github.com/repos/SME-Issues/issues
closed
Query Payment Tests Comprehension None - 04/12/2018 - 5004
NLP Api pulse_tests
**Query Payment Tests Comprehension None** - Total: 26 - Passed: 10 - **Pass: 10 (38%)** - Not Understood: 0 - Error (not understood): 0 - Failed but Understood: 16 (62%)
1.0
Query Payment Tests Comprehension None - 04/12/2018 - 5004 - **Query Payment Tests Comprehension None** - Total: 26 - Passed: 10 - **Pass: 10 (38%)** - Not Understood: 0 - Error (not understood): 0 - Failed but Understood: 16 (62%)
test
query payment tests comprehension none query payment tests comprehension none total passed pass not understood error not understood failed but understood
1
10,090
7,893,086,166
IssuesEvent
2018-06-28 16:52:48
globaleaks/GlobaLeaks
https://api.github.com/repos/globaleaks/GlobaLeaks
opened
The administrator should be able to reset the password for a user
C: Backend C: Client F: Security F: UserInterface T: Enhancement
Current password reset enable the admin to manually set the password for the user. This easy implementation is become more and more inadequate: - on creation the admin has to manually notify users the password. - on change the admin has to manually choose and set a password and notify users the password. This ticket is about removing the possibility for the admin to set a password and replace the functionality with a Reset Password button. The button should trigger the same automatic workflow of the user password reset sending and authentication token to the configured email. This change impacts both usability and security of the plaftorm.
True
The administrator should be able to reset the password for a user - Current password reset enable the admin to manually set the password for the user. This easy implementation is become more and more inadequate: - on creation the admin has to manually notify users the password. - on change the admin has to manually choose and set a password and notify users the password. This ticket is about removing the possibility for the admin to set a password and replace the functionality with a Reset Password button. The button should trigger the same automatic workflow of the user password reset sending and authentication token to the configured email. This change impacts both usability and security of the plaftorm.
non_test
the administrator should be able to reset the password for a user current password reset enable the admin to manually set the password for the user this easy implementation is become more and more inadequate on creation the admin has to manually notify users the password on change the admin has to manually choose and set a password and notify users the password this ticket is about removing the possibility for the admin to set a password and replace the functionality with a reset password button the button should trigger the same automatic workflow of the user password reset sending and authentication token to the configured email this change impacts both usability and security of the plaftorm
0
147,080
11,771,115,308
IssuesEvent
2020-03-15 22:22:18
SCE-Development/Core-v4
https://api.github.com/repos/SCE-Development/Core-v4
closed
Add Cross OS Testing to Travis
Testing enhancement
Development is conducted on various Linux distros and OSX. We should ensure that all tests pass on these OS's.
1.0
Add Cross OS Testing to Travis - Development is conducted on various Linux distros and OSX. We should ensure that all tests pass on these OS's.
test
add cross os testing to travis development is conducted on various linux distros and osx we should ensure that all tests pass on these os s
1
319,795
9,755,133,486
IssuesEvent
2019-06-04 13:17:29
inverse-inc/packetfence
https://api.github.com/repos/inverse-inc/packetfence
opened
Inconsistent use of pf_management_ip in docs
Priority: Medium Type: Bug
**Describe the bug** There are currently 6 different uses of the packetfence hostname/ip throughout our documentation in /static/docs. - %3Chostname%3E - your_portal_hostname - your_portal_ip - pf_management_ip - %3Cyour_captive_portal_ip%3E - _ip_address_of_packetfence **Expected behavior** A single definition should be used for all of the above for consistency. **Additional context** Consistency is needed in order to reduce the number of cases needed to rewrite the URLs when embedding the documentation in the new admin.
1.0
Inconsistent use of pf_management_ip in docs - **Describe the bug** There are currently 6 different uses of the packetfence hostname/ip throughout our documentation in /static/docs. - %3Chostname%3E - your_portal_hostname - your_portal_ip - pf_management_ip - %3Cyour_captive_portal_ip%3E - _ip_address_of_packetfence **Expected behavior** A single definition should be used for all of the above for consistency. **Additional context** Consistency is needed in order to reduce the number of cases needed to rewrite the URLs when embedding the documentation in the new admin.
non_test
inconsistent use of pf management ip in docs describe the bug there are currently different uses of the packetfence hostname ip throughout our documentation in static docs your portal hostname your portal ip pf management ip captive portal ip ip address of packetfence expected behavior a single definition should be used for all of the above for consistency additional context consistency is needed in order to reduce the number of cases needed to rewrite the urls when embedding the documentation in the new admin
0
55,683
6,489,315,780
IssuesEvent
2017-08-21 00:49:39
FireFly-WoW/FireFly-IssueTracker
https://api.github.com/repos/FireFly-WoW/FireFly-IssueTracker
closed
Quest: As the Crow Flies
Confirmed Status: Needs Testing
After the automated flight around zangramarsh completes you are locked into crow form until you die. This is purely cosmetic as you can cast and mount properly, but you are a crow.
1.0
Quest: As the Crow Flies - After the automated flight around zangramarsh completes you are locked into crow form until you die. This is purely cosmetic as you can cast and mount properly, but you are a crow.
test
quest as the crow flies after the automated flight around zangramarsh completes you are locked into crow form until you die this is purely cosmetic as you can cast and mount properly but you are a crow
1
201,538
15,212,101,552
IssuesEvent
2021-02-17 09:58:24
QubesOS/updates-status
https://api.github.com/repos/QubesOS/updates-status
closed
linux-kernel v5.4.90-1 (r4.1)
r4.1-dom0-cur-test
Update of linux-kernel to v5.4.90-1 for Qubes r4.1, see comments below for details. Built from: https://github.com/QubesOS/qubes-linux-kernel/commit/31155347d4cb1bf54e5c30577f58346d175c821a [Changes since previous version](https://github.com/QubesOS/qubes-linux-kernel/compare/v5.4.88-1...v5.4.90-1): QubesOS/qubes-linux-kernel@3115534 version 5.4.90-1 QubesOS/qubes-linux-kernel@fe16e71 Always include u2mfn module, also for R4.1 builds QubesOS/qubes-linux-kernel@489d784 Update to kernel-5.4.89 Referenced issues: QubesOS/qubes-issues#4867 If you're release manager, you can issue GPG-inline signed command: * `Upload linux-kernel 31155347d4cb1bf54e5c30577f58346d175c821a r4.1 current repo` (available 7 days from now) * `Upload linux-kernel 31155347d4cb1bf54e5c30577f58346d175c821a r4.1 current (dists) repo`, you can choose subset of distributions, like `vm-fc24 vm-fc25` (available 7 days from now) * `Upload linux-kernel 31155347d4cb1bf54e5c30577f58346d175c821a r4.1 security-testing repo` Above commands will work only if packages in current-testing repository were built from given commit (i.e. no new version superseded it).
1.0
linux-kernel v5.4.90-1 (r4.1) - Update of linux-kernel to v5.4.90-1 for Qubes r4.1, see comments below for details. Built from: https://github.com/QubesOS/qubes-linux-kernel/commit/31155347d4cb1bf54e5c30577f58346d175c821a [Changes since previous version](https://github.com/QubesOS/qubes-linux-kernel/compare/v5.4.88-1...v5.4.90-1): QubesOS/qubes-linux-kernel@3115534 version 5.4.90-1 QubesOS/qubes-linux-kernel@fe16e71 Always include u2mfn module, also for R4.1 builds QubesOS/qubes-linux-kernel@489d784 Update to kernel-5.4.89 Referenced issues: QubesOS/qubes-issues#4867 If you're release manager, you can issue GPG-inline signed command: * `Upload linux-kernel 31155347d4cb1bf54e5c30577f58346d175c821a r4.1 current repo` (available 7 days from now) * `Upload linux-kernel 31155347d4cb1bf54e5c30577f58346d175c821a r4.1 current (dists) repo`, you can choose subset of distributions, like `vm-fc24 vm-fc25` (available 7 days from now) * `Upload linux-kernel 31155347d4cb1bf54e5c30577f58346d175c821a r4.1 security-testing repo` Above commands will work only if packages in current-testing repository were built from given commit (i.e. no new version superseded it).
test
linux kernel update of linux kernel to for qubes see comments below for details built from qubesos qubes linux kernel version qubesos qubes linux kernel always include module also for builds qubesos qubes linux kernel update to kernel referenced issues qubesos qubes issues if you re release manager you can issue gpg inline signed command upload linux kernel current repo available days from now upload linux kernel current dists repo you can choose subset of distributions like vm vm available days from now upload linux kernel security testing repo above commands will work only if packages in current testing repository were built from given commit i e no new version superseded it
1
416,795
12,151,630,162
IssuesEvent
2020-04-24 20:21:54
imazen/imageflow
https://api.github.com/repos/imazen/imageflow
closed
Transparent png show black background
priority-high
Hi, Thanks for your amazing software, it works really well!! We are having a weird problem with these images: ## Original ![imo-1-original](https://user-images.githubusercontent.com/614413/31891458-188e5fb4-b7cc-11e7-8c65-57c745070e59.png) (by the way, its the same image of #153, different problem) I optimized the original image using [Tinypng](https://tinypng.com): ## Optimized ![imo-1-optimized](https://user-images.githubusercontent.com/614413/31891594-7ae85034-b7cc-11e7-8893-682dffa275e9.png) In both cases the image has a transparent background. Everything look fine if you see the image directly. But when asked through imageflow appears with a different background: ## Served by imageflow ![imo-1-optimized-imageflow](https://user-images.githubusercontent.com/614413/31891803-0cf3dd36-b7cd-11e7-834a-205c218a6612.png) In the optimization the image size decrease 75%, which is great for our home page. It is something about the Tinypng optimization? or perhaps is a profile mishandling in imageflow? Thanks for all your help!!
1.0
Transparent png show black background - Hi, Thanks for your amazing software, it works really well!! We are having a weird problem with these images: ## Original ![imo-1-original](https://user-images.githubusercontent.com/614413/31891458-188e5fb4-b7cc-11e7-8c65-57c745070e59.png) (by the way, its the same image of #153, different problem) I optimized the original image using [Tinypng](https://tinypng.com): ## Optimized ![imo-1-optimized](https://user-images.githubusercontent.com/614413/31891594-7ae85034-b7cc-11e7-8893-682dffa275e9.png) In both cases the image has a transparent background. Everything look fine if you see the image directly. But when asked through imageflow appears with a different background: ## Served by imageflow ![imo-1-optimized-imageflow](https://user-images.githubusercontent.com/614413/31891803-0cf3dd36-b7cd-11e7-834a-205c218a6612.png) In the optimization the image size decrease 75%, which is great for our home page. It is something about the Tinypng optimization? or perhaps is a profile mishandling in imageflow? Thanks for all your help!!
non_test
transparent png show black background hi thanks for your amazing software it works really well we are having a weird problem with these images original by the way its the same image of different problem i optimized the original image using optimized in both cases the image has a transparent background everything look fine if you see the image directly but when asked through imageflow appears with a different background served by imageflow in the optimization the image size decrease which is great for our home page it is something about the tinypng optimization or perhaps is a profile mishandling in imageflow thanks for all your help
0
120,976
25,907,100,201
IssuesEvent
2022-12-15 11:04:27
arduino/arduino-ide
https://api.github.com/repos/arduino/arduino-ide
opened
"Preferences" dialog scrollability not easily discoverable by user
topic: code type: imperfection
### Describe the problem The dimensions of the "**Preferences**" dialog is limited by the Arduino IDE window dimensions. If the window dimensions are small, the contents of the dialog may not fit within the available space. In this case, the dialog is scrollable, which allows the user to access the settings that fall outside the visible area of the dialog. 🐛 The ability to scroll the dialog is difficult for the user to discover. These users are given the impression that some preferences are missing from the dialog due to a bug in Arduino IDE. ### To reproduce 1. Resize the Arduino IDE window to the minimum vertical size. 1. Select **File > Preferences** from the Arduino IDE menus. The preferences farthest down in the dialog are not visible. 1. Move the mouse pointer to an area of the screen not over the dialog: ![image](https://user-images.githubusercontent.com/8572152/207843298-6ed8d0ab-6494-4c38-9927-33b56fc80578.png) 🐛 There is no indication at all that the dialog is scrollable: 1. Hover the mouse pointer over the dialog, but not over the scrollbar area: ![image](https://user-images.githubusercontent.com/8572152/207843329-0534b2e4-3849-43df-9de3-6cd28cbcf27e.png) 🐛 A scrollbar appears, but the non-standard design causes the nature of the UI element to be non-obvious to the average user: - Rounded "thumb" - Thumb is significantly less than standard width - No "trough" - Scrollbar overlaps the horizontal rule of the tab bar - There is no spacing between the right side of the "**BROWSE**" button and the scrollbar 1. Hover the mouse pointer over the the scrollbar area of the dialog: ![image](https://user-images.githubusercontent.com/8572152/207843352-e522e201-11fa-4a21-9cb2-dc432736bc72.png) 🙂 The nature of the UI element is now fairly obvious to the average user: - There is a trough - The scrollbar has a standard width ### Expected behavior It is obvious to the user when a dialog is scrollable. ### Arduino IDE version 7c86f1f ### Operating system Windows ### Operating system version 10 ### Additional context Originally reported at https://forum.arduino.cc/t/missing-additional-board-manager-urls-field-in-preferences/1055935/1 Additional reports: - https://forum.arduino.cc/t/arduino-core-available-for-the-ai-thinker-bl602/1064372/5 - https://forum.arduino.cc/t/missing-additional-board-manager-urls-field-in-preferences/1055935/7 ### Issue checklist - [X] I searched for previous reports in [the issue tracker](https://github.com/arduino/arduino-ide/issues?q=) - [X] I verified the problem still occurs when using the latest [nightly build](https://www.arduino.cc/en/software#nightly-builds) - [X] My report contains all necessary details
1.0
"Preferences" dialog scrollability not easily discoverable by user - ### Describe the problem The dimensions of the "**Preferences**" dialog is limited by the Arduino IDE window dimensions. If the window dimensions are small, the contents of the dialog may not fit within the available space. In this case, the dialog is scrollable, which allows the user to access the settings that fall outside the visible area of the dialog. 🐛 The ability to scroll the dialog is difficult for the user to discover. These users are given the impression that some preferences are missing from the dialog due to a bug in Arduino IDE. ### To reproduce 1. Resize the Arduino IDE window to the minimum vertical size. 1. Select **File > Preferences** from the Arduino IDE menus. The preferences farthest down in the dialog are not visible. 1. Move the mouse pointer to an area of the screen not over the dialog: ![image](https://user-images.githubusercontent.com/8572152/207843298-6ed8d0ab-6494-4c38-9927-33b56fc80578.png) 🐛 There is no indication at all that the dialog is scrollable: 1. Hover the mouse pointer over the dialog, but not over the scrollbar area: ![image](https://user-images.githubusercontent.com/8572152/207843329-0534b2e4-3849-43df-9de3-6cd28cbcf27e.png) 🐛 A scrollbar appears, but the non-standard design causes the nature of the UI element to be non-obvious to the average user: - Rounded "thumb" - Thumb is significantly less than standard width - No "trough" - Scrollbar overlaps the horizontal rule of the tab bar - There is no spacing between the right side of the "**BROWSE**" button and the scrollbar 1. Hover the mouse pointer over the the scrollbar area of the dialog: ![image](https://user-images.githubusercontent.com/8572152/207843352-e522e201-11fa-4a21-9cb2-dc432736bc72.png) 🙂 The nature of the UI element is now fairly obvious to the average user: - There is a trough - The scrollbar has a standard width ### Expected behavior It is obvious to the user when a dialog is scrollable. ### Arduino IDE version 7c86f1f ### Operating system Windows ### Operating system version 10 ### Additional context Originally reported at https://forum.arduino.cc/t/missing-additional-board-manager-urls-field-in-preferences/1055935/1 Additional reports: - https://forum.arduino.cc/t/arduino-core-available-for-the-ai-thinker-bl602/1064372/5 - https://forum.arduino.cc/t/missing-additional-board-manager-urls-field-in-preferences/1055935/7 ### Issue checklist - [X] I searched for previous reports in [the issue tracker](https://github.com/arduino/arduino-ide/issues?q=) - [X] I verified the problem still occurs when using the latest [nightly build](https://www.arduino.cc/en/software#nightly-builds) - [X] My report contains all necessary details
non_test
preferences dialog scrollability not easily discoverable by user describe the problem the dimensions of the preferences dialog is limited by the arduino ide window dimensions if the window dimensions are small the contents of the dialog may not fit within the available space in this case the dialog is scrollable which allows the user to access the settings that fall outside the visible area of the dialog 🐛 the ability to scroll the dialog is difficult for the user to discover these users are given the impression that some preferences are missing from the dialog due to a bug in arduino ide to reproduce resize the arduino ide window to the minimum vertical size select file preferences from the arduino ide menus the preferences farthest down in the dialog are not visible move the mouse pointer to an area of the screen not over the dialog 🐛 there is no indication at all that the dialog is scrollable hover the mouse pointer over the dialog but not over the scrollbar area 🐛 a scrollbar appears but the non standard design causes the nature of the ui element to be non obvious to the average user rounded thumb thumb is significantly less than standard width no trough scrollbar overlaps the horizontal rule of the tab bar there is no spacing between the right side of the browse button and the scrollbar hover the mouse pointer over the the scrollbar area of the dialog 🙂 the nature of the ui element is now fairly obvious to the average user there is a trough the scrollbar has a standard width expected behavior it is obvious to the user when a dialog is scrollable arduino ide version operating system windows operating system version additional context originally reported at additional reports issue checklist i searched for previous reports in i verified the problem still occurs when using the latest my report contains all necessary details
0
13,903
23,934,408,032
IssuesEvent
2022-09-11 02:16:43
CS3219-AY2223S1/cs3219-project-ay2223s1-g5
https://api.github.com/repos/CS3219-AY2223S1/cs3219-project-ay2223s1-g5
closed
[AUTH-FR-3] The system should allow users to log into their accounts by entering their email and password.
functional requirement P1
- [x] #71 - [x] #72 - [x] #73 - [x] #74
1.0
[AUTH-FR-3] The system should allow users to log into their accounts by entering their email and password. - - [x] #71 - [x] #72 - [x] #73 - [x] #74
non_test
the system should allow users to log into their accounts by entering their email and password
0
339,816
30,477,799,230
IssuesEvent
2023-07-17 17:50:55
alcionai/corso
https://api.github.com/repos/alcionai/corso
closed
Long-running tests for immutable backups
duplicate tests ci
Immutable backups will require long-running test coverage to ensure things like renewing object locks functions as expected. These tests can be made a part of the new longevity test
1.0
Long-running tests for immutable backups - Immutable backups will require long-running test coverage to ensure things like renewing object locks functions as expected. These tests can be made a part of the new longevity test
test
long running tests for immutable backups immutable backups will require long running test coverage to ensure things like renewing object locks functions as expected these tests can be made a part of the new longevity test
1
177,806
13,748,685,011
IssuesEvent
2020-10-06 09:24:18
home-assistant/core
https://api.github.com/repos/home-assistant/core
closed
Rewrite statsd unittest tests to pytest style test functions
Hacktoberfest Help wanted integration: statsd unittest.TestCase
## Problem The statsd integration still contains `unittest.TestCase` based unit tests. We want to rewrite them to standalone pytest test functions. ## Background The Home Assistant core standard is to write tests as standalone pytest test functions. We still have some old tests that are based on `unittest.TestCase`. We want all these tests to be rewritten as pytest test functions. Here are the docs for pytest: https://docs.pytest.org/en/stable/ Here's an example of an async pytest test function in Home Assistant core: https://github.com/home-assistant/core/blob/4cce724473233d4fb32c08bd251940b1ce2ba570/tests/components/tradfri/test_light.py#L156-L176 There are many pytest fixtures to help writing the tests. See: - https://docs.pytest.org/en/stable/reference.html#fixtures - https://github.com/home-assistant/core/blob/dev/tests/conftest.py - The most common fixture to use is [`hass`](https://github.com/home-assistant/core/blob/4cce724473233d4fb32c08bd251940b1ce2ba570/tests/conftest.py#L107) which will set up a [`HomeAssistant`](https://github.com/home-assistant/core/blob/4cce724473233d4fb32c08bd251940b1ce2ba570/homeassistant/core.py#L166) instance and start it. Here's an example of a pull request that rewrote a module of `unittest.TestCase` tests to standalone pytest test functions: https://github.com/home-assistant/core/pull/40749 Here's an example command to run a single test module with pytest inside tox on Python 3.8: ```sh tox -e py38 -- --cov-report term-missing --cov=homeassistant.components.command_line.switch tests/components/command_line/test_switch.py ``` It will print coverage information with lines that are missing coverage. ## Prerequisites - A working [development environment](https://developers.home-assistant.io/docs/development_environment). ## Task - Rewrite the tests one module at a time and submit the changes as a pull request to this repository. - We want to limit the change scope to a single module to not have the pull request be too long, which would take longer time to review. - Remember to reference this issue in your pull request, but don't close or fix the issue until all tests for the integration are updated.
2.0
Rewrite statsd unittest tests to pytest style test functions - ## Problem The statsd integration still contains `unittest.TestCase` based unit tests. We want to rewrite them to standalone pytest test functions. ## Background The Home Assistant core standard is to write tests as standalone pytest test functions. We still have some old tests that are based on `unittest.TestCase`. We want all these tests to be rewritten as pytest test functions. Here are the docs for pytest: https://docs.pytest.org/en/stable/ Here's an example of an async pytest test function in Home Assistant core: https://github.com/home-assistant/core/blob/4cce724473233d4fb32c08bd251940b1ce2ba570/tests/components/tradfri/test_light.py#L156-L176 There are many pytest fixtures to help writing the tests. See: - https://docs.pytest.org/en/stable/reference.html#fixtures - https://github.com/home-assistant/core/blob/dev/tests/conftest.py - The most common fixture to use is [`hass`](https://github.com/home-assistant/core/blob/4cce724473233d4fb32c08bd251940b1ce2ba570/tests/conftest.py#L107) which will set up a [`HomeAssistant`](https://github.com/home-assistant/core/blob/4cce724473233d4fb32c08bd251940b1ce2ba570/homeassistant/core.py#L166) instance and start it. Here's an example of a pull request that rewrote a module of `unittest.TestCase` tests to standalone pytest test functions: https://github.com/home-assistant/core/pull/40749 Here's an example command to run a single test module with pytest inside tox on Python 3.8: ```sh tox -e py38 -- --cov-report term-missing --cov=homeassistant.components.command_line.switch tests/components/command_line/test_switch.py ``` It will print coverage information with lines that are missing coverage. ## Prerequisites - A working [development environment](https://developers.home-assistant.io/docs/development_environment). ## Task - Rewrite the tests one module at a time and submit the changes as a pull request to this repository. - We want to limit the change scope to a single module to not have the pull request be too long, which would take longer time to review. - Remember to reference this issue in your pull request, but don't close or fix the issue until all tests for the integration are updated.
test
rewrite statsd unittest tests to pytest style test functions problem the statsd integration still contains unittest testcase based unit tests we want to rewrite them to standalone pytest test functions background the home assistant core standard is to write tests as standalone pytest test functions we still have some old tests that are based on unittest testcase we want all these tests to be rewritten as pytest test functions here are the docs for pytest here s an example of an async pytest test function in home assistant core there are many pytest fixtures to help writing the tests see the most common fixture to use is which will set up a instance and start it here s an example of a pull request that rewrote a module of unittest testcase tests to standalone pytest test functions here s an example command to run a single test module with pytest inside tox on python sh tox e cov report term missing cov homeassistant components command line switch tests components command line test switch py it will print coverage information with lines that are missing coverage prerequisites a working task rewrite the tests one module at a time and submit the changes as a pull request to this repository we want to limit the change scope to a single module to not have the pull request be too long which would take longer time to review remember to reference this issue in your pull request but don t close or fix the issue until all tests for the integration are updated
1
392,747
26,956,943,916
IssuesEvent
2023-02-08 15:30:18
ORNL/ipyanchorviz
https://api.github.com/repos/ORNL/ipyanchorviz
opened
Add sphinx documentation
documentation
Random thing to ensure is covered in the documentation: the `autoNorm` option is specifically row normalization to 1.
1.0
Add sphinx documentation - Random thing to ensure is covered in the documentation: the `autoNorm` option is specifically row normalization to 1.
non_test
add sphinx documentation random thing to ensure is covered in the documentation the autonorm option is specifically row normalization to
0
159,142
12,464,821,936
IssuesEvent
2020-05-28 13:07:41
dasch-swiss/knora-app
https://api.github.com/repos/dasch-swiss/knora-app
closed
Homepage : make the name of the app more visible
bug user-testing
**Describe the bug** If Knora-app is going to be the name of Knora GUI, then this name should be displayed in a more prominent way. Currently it only appears on the left bottom corner. **To Reproduce Steps to reproduce the behavior:** 1. Go to '...' 2. Click on '....' 3. Scroll down to '....' 4. See error **OPTIONAL: Expected behavior** Maybe "**Knora app** is the generic web application of the Data and Service Center for Humanities DaSCH. It's the user interface research layer of Knora..." **Screenshots** If applicable, add screenshots to help explain your problem (drag-and-drop the image directly here). **Desktop (please complete the following information):** - OS: macOS 10.14.6 (18G103), French - Browser Firefox - Version 70.01.1 **Additional context** Add any other context about the problem here.
1.0
Homepage : make the name of the app more visible - **Describe the bug** If Knora-app is going to be the name of Knora GUI, then this name should be displayed in a more prominent way. Currently it only appears on the left bottom corner. **To Reproduce Steps to reproduce the behavior:** 1. Go to '...' 2. Click on '....' 3. Scroll down to '....' 4. See error **OPTIONAL: Expected behavior** Maybe "**Knora app** is the generic web application of the Data and Service Center for Humanities DaSCH. It's the user interface research layer of Knora..." **Screenshots** If applicable, add screenshots to help explain your problem (drag-and-drop the image directly here). **Desktop (please complete the following information):** - OS: macOS 10.14.6 (18G103), French - Browser Firefox - Version 70.01.1 **Additional context** Add any other context about the problem here.
test
homepage make the name of the app more visible describe the bug if knora app is going to be the name of knora gui then this name should be displayed in a more prominent way currently it only appears on the left bottom corner to reproduce steps to reproduce the behavior go to click on scroll down to see error optional expected behavior maybe knora app is the generic web application of the data and service center for humanities dasch it s the user interface research layer of knora screenshots if applicable add screenshots to help explain your problem drag and drop the image directly here desktop please complete the following information os macos french browser firefox version additional context add any other context about the problem here
1
516,965
14,991,811,677
IssuesEvent
2021-01-29 08:57:31
microsoft/PowerToys
https://api.github.com/repos/microsoft/PowerToys
closed
[Keyboard Navigation: Settings>Remap key>Add button]: Focus order is not logical on activating 'Add remap key' button.
Accessibility [E+D] Area-Accessibility Issue-Bug Priority-2 Product-Keyboard Shortcut Manager Resolution-Fix-Committed Severity-Regression
[Power Toys Settings-Keyboard Manager>Remap Key] User Experience: This will impact the keyboard users if the keyboard focus is not landing on the triggered controls after activating Add button. Test Environment: "OS Version: 20221.1000 App Name: Power Toy Preview App Version: v0.23.0 Screen Reader: Narrator" Repro-steps: "1. Open Power Toys Settings App. 2. Navigate to Keyboard Manager list item present in left pane and activate it. 3. Navigate to Remap a key button present in right pane and activate it. 4. Remap Key window will open. 5. Navigate to 'Add' button present in window and activate it and verify the issue. " Actual Result: "Focus order is not logical on activating 'Add remap key' button. Focus stays on Add Remap key which is not correct Note: Same issue is repro in 'Remap a Shortcut'." Expected Result: Keyboard focus should jump to 'Type' button on activating Add remap key which is Expected and logical. MAS Reference: https://microsoft.sharepoint.com/:w:/r/teams/msenable/_layouts/15/WopiFrame.aspx?sourcedoc={0de7fbe1-ad7e-48e5-bcbb-8d986691e2b9} [16_Remap key_MAS2.4.3_After activating Add button focus is not landing on added controls.zip](https://github.com/microsoft/PowerToys/files/5328949/16_Remap.key_MAS2.4.3_After.activating.Add.button.focus.is.not.landing.on.added.controls.zip)
1.0
[Keyboard Navigation: Settings>Remap key>Add button]: Focus order is not logical on activating 'Add remap key' button. - [Power Toys Settings-Keyboard Manager>Remap Key] User Experience: This will impact the keyboard users if the keyboard focus is not landing on the triggered controls after activating Add button. Test Environment: "OS Version: 20221.1000 App Name: Power Toy Preview App Version: v0.23.0 Screen Reader: Narrator" Repro-steps: "1. Open Power Toys Settings App. 2. Navigate to Keyboard Manager list item present in left pane and activate it. 3. Navigate to Remap a key button present in right pane and activate it. 4. Remap Key window will open. 5. Navigate to 'Add' button present in window and activate it and verify the issue. " Actual Result: "Focus order is not logical on activating 'Add remap key' button. Focus stays on Add Remap key which is not correct Note: Same issue is repro in 'Remap a Shortcut'." Expected Result: Keyboard focus should jump to 'Type' button on activating Add remap key which is Expected and logical. MAS Reference: https://microsoft.sharepoint.com/:w:/r/teams/msenable/_layouts/15/WopiFrame.aspx?sourcedoc={0de7fbe1-ad7e-48e5-bcbb-8d986691e2b9} [16_Remap key_MAS2.4.3_After activating Add button focus is not landing on added controls.zip](https://github.com/microsoft/PowerToys/files/5328949/16_Remap.key_MAS2.4.3_After.activating.Add.button.focus.is.not.landing.on.added.controls.zip)
non_test
focus order is not logical on activating add remap key button user experience this will impact the keyboard users if the keyboard focus is not landing on the triggered controls after activating add button test environment os version app name power toy preview app version screen reader narrator repro steps open power toys settings app navigate to keyboard manager list item present in left pane and activate it navigate to remap a key button present in right pane and activate it remap key window will open navigate to add button present in window and activate it and verify the issue actual result focus order is not logical on activating add remap key button focus stays on add remap key which is not correct note same issue is repro in remap a shortcut expected result keyboard focus should jump to type button on activating add remap key which is expected and logical mas reference
0
149,262
13,261,022,036
IssuesEvent
2020-08-20 19:13:27
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
closed
documentation - Latest simulation release docs. (Trac #770)
Migrated from Trac documentation enhancement
I'd like simulation and icerec to also have links to the doc builds of the latest release. For simulation this is currently http://code.icecube.wisc.edu/svn/meta-projects/simulation/releases/V04-01-07 . !IceRec may want two links : one for the latest and greatest release and another for the latest L2 processing release. <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/770">https://code.icecube.wisc.edu/projects/icecube/ticket/770</a>, reported by olivasand owned by nega</em></summary> <p> ```json { "status": "closed", "changetime": "2015-02-11T19:09:46", "_ts": "1423681786427517", "description": "I'd like simulation and icerec to also have links to the doc builds of the latest release. For simulation this is currently http://code.icecube.wisc.edu/svn/meta-projects/simulation/releases/V04-01-07 .\n\n!IceRec may want two links : one for the latest and greatest release and another for the latest L2 processing release.", "reporter": "olivas", "cc": "meike.dewith", "resolution": "wontfix", "time": "2014-10-02T23:57:33", "component": "documentation", "summary": "documentation - Latest simulation release docs.", "priority": "normal", "keywords": "", "milestone": "", "owner": "nega", "type": "enhancement" } ``` </p> </details>
1.0
documentation - Latest simulation release docs. (Trac #770) - I'd like simulation and icerec to also have links to the doc builds of the latest release. For simulation this is currently http://code.icecube.wisc.edu/svn/meta-projects/simulation/releases/V04-01-07 . !IceRec may want two links : one for the latest and greatest release and another for the latest L2 processing release. <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/770">https://code.icecube.wisc.edu/projects/icecube/ticket/770</a>, reported by olivasand owned by nega</em></summary> <p> ```json { "status": "closed", "changetime": "2015-02-11T19:09:46", "_ts": "1423681786427517", "description": "I'd like simulation and icerec to also have links to the doc builds of the latest release. For simulation this is currently http://code.icecube.wisc.edu/svn/meta-projects/simulation/releases/V04-01-07 .\n\n!IceRec may want two links : one for the latest and greatest release and another for the latest L2 processing release.", "reporter": "olivas", "cc": "meike.dewith", "resolution": "wontfix", "time": "2014-10-02T23:57:33", "component": "documentation", "summary": "documentation - Latest simulation release docs.", "priority": "normal", "keywords": "", "milestone": "", "owner": "nega", "type": "enhancement" } ``` </p> </details>
non_test
documentation latest simulation release docs trac i d like simulation and icerec to also have links to the doc builds of the latest release for simulation this is currently icerec may want two links one for the latest and greatest release and another for the latest processing release migrated from json status closed changetime ts description i d like simulation and icerec to also have links to the doc builds of the latest release for simulation this is currently n n icerec may want two links one for the latest and greatest release and another for the latest processing release reporter olivas cc meike dewith resolution wontfix time component documentation summary documentation latest simulation release docs priority normal keywords milestone owner nega type enhancement
0
138,125
11,184,268,189
IssuesEvent
2019-12-31 17:15:04
dbrownukk/EFD_v2
https://api.github.com/repos/dbrownukk/EFD_v2
closed
Unable to delete Transfers
For Testing bug
Instance: EFD_HM App: OHEA Module: WGI Project: NIMFRU LZ: Katakwi Fishing & Livestock Zone District: Katakwi Village: Okorio WG: Poor 1. In detail view, click Transfers tab 1. Click delete icon for first transfer > Note the record disappears as expected 1. Click Save > Error: Value for Unit in Transfer is required > Error: Value for Transfer type in Transfer is required > Error: Value for Status in Transfer is required 1. Exit the detail for this Village then return to the Transfers > Note the deleted record is displayed. I created further Transfers & attempted to delete them with the same result, so it does not appear to be a data problem with old inconsistent data entered before the latest change was made.
1.0
Unable to delete Transfers - Instance: EFD_HM App: OHEA Module: WGI Project: NIMFRU LZ: Katakwi Fishing & Livestock Zone District: Katakwi Village: Okorio WG: Poor 1. In detail view, click Transfers tab 1. Click delete icon for first transfer > Note the record disappears as expected 1. Click Save > Error: Value for Unit in Transfer is required > Error: Value for Transfer type in Transfer is required > Error: Value for Status in Transfer is required 1. Exit the detail for this Village then return to the Transfers > Note the deleted record is displayed. I created further Transfers & attempted to delete them with the same result, so it does not appear to be a data problem with old inconsistent data entered before the latest change was made.
test
unable to delete transfers instance efd hm app ohea module wgi project nimfru lz katakwi fishing livestock zone district katakwi village okorio wg poor in detail view click transfers tab click delete icon for first transfer note the record disappears as expected click save error value for unit in transfer is required error value for transfer type in transfer is required error value for status in transfer is required exit the detail for this village then return to the transfers note the deleted record is displayed i created further transfers attempted to delete them with the same result so it does not appear to be a data problem with old inconsistent data entered before the latest change was made
1
42,727
17,270,596,611
IssuesEvent
2021-07-22 19:16:03
cityofaustin/atd-data-tech
https://api.github.com/repos/cityofaustin/atd-data-tech
closed
Figure out which email address to use for account creation
Provider: CTM Service: Apps Type: Other Workgroup: HR
<!-- Email --> <!-- linda.escamilla@austintexas.gov --> > What application are you using? Other / Not Sure > Describe the problem. Ramona A has returned to work at ATD after working at the Convention Center. CTM has given her an alias email address that includes a "2". What email address should I have entered in Knack when she was hired. Thanks, Linda > How soon do you need this? Soon — This week > Requested By Linda E. Request ID: DTS21-102235
1.0
Figure out which email address to use for account creation - <!-- Email --> <!-- linda.escamilla@austintexas.gov --> > What application are you using? Other / Not Sure > Describe the problem. Ramona A has returned to work at ATD after working at the Convention Center. CTM has given her an alias email address that includes a "2". What email address should I have entered in Knack when she was hired. Thanks, Linda > How soon do you need this? Soon — This week > Requested By Linda E. Request ID: DTS21-102235
non_test
figure out which email address to use for account creation what application are you using other not sure describe the problem ramona a has returned to work at atd after working at the convention center ctm has given her an alias email address that includes a what email address should i have entered in knack when she was hired thanks linda how soon do you need this soon — this week requested by linda e request id
0
327,822
28,084,311,825
IssuesEvent
2023-03-30 08:45:58
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
roachtest: kv95/enc=false/nodes=3/tracing failed
C-test-failure O-robot O-roachtest branch-master release-blocker
roachtest.kv95/enc=false/nodes=3/tracing [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyAwsBazel/9329888?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyAwsBazel/9329888?buildTab=artifacts#/kv95/enc=false/nodes=3/tracing) on master @ [1f8024bf14433ca169e5a8c3768c5d223dc5018c](https://github.com/cockroachdb/cockroach/commits/1f8024bf14433ca169e5a8c3768c5d223dc5018c): ``` test artifacts and logs in: /artifacts/kv95/enc=false/nodes=3/tracing/run_1 (cluster.go:1977).Run: output in run_071450.060861652_n4_workload-run-kv-tole: ./workload run kv --tolerate-errors --init --histograms=perf/stats.json --concurrency=192 --splits=1000 --duration=30m0s --read-percent=95 {pgurl:1-3} returned: COMMAND_PROBLEM: ssh verbose log retained in ssh_071450.653797077_n4_workload-run-kv-tole.log: exit status 1 (monitor.go:127).Wait: monitor failure: monitor task failed: t.Fatal() was called ``` <p>Parameters: <code>ROACHTEST_cloud=aws</code> , <code>ROACHTEST_cpu=8</code> , <code>ROACHTEST_encrypted=false</code> , <code>ROACHTEST_fs=ext4</code> , <code>ROACHTEST_localSSD=true</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/obs-inf-prs <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*kv95/enc=false/nodes=3/tracing.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
2.0
roachtest: kv95/enc=false/nodes=3/tracing failed - roachtest.kv95/enc=false/nodes=3/tracing [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyAwsBazel/9329888?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyAwsBazel/9329888?buildTab=artifacts#/kv95/enc=false/nodes=3/tracing) on master @ [1f8024bf14433ca169e5a8c3768c5d223dc5018c](https://github.com/cockroachdb/cockroach/commits/1f8024bf14433ca169e5a8c3768c5d223dc5018c): ``` test artifacts and logs in: /artifacts/kv95/enc=false/nodes=3/tracing/run_1 (cluster.go:1977).Run: output in run_071450.060861652_n4_workload-run-kv-tole: ./workload run kv --tolerate-errors --init --histograms=perf/stats.json --concurrency=192 --splits=1000 --duration=30m0s --read-percent=95 {pgurl:1-3} returned: COMMAND_PROBLEM: ssh verbose log retained in ssh_071450.653797077_n4_workload-run-kv-tole.log: exit status 1 (monitor.go:127).Wait: monitor failure: monitor task failed: t.Fatal() was called ``` <p>Parameters: <code>ROACHTEST_cloud=aws</code> , <code>ROACHTEST_cpu=8</code> , <code>ROACHTEST_encrypted=false</code> , <code>ROACHTEST_fs=ext4</code> , <code>ROACHTEST_localSSD=true</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/obs-inf-prs <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*kv95/enc=false/nodes=3/tracing.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
test
roachtest enc false nodes tracing failed roachtest enc false nodes tracing with on master test artifacts and logs in artifacts enc false nodes tracing run cluster go run output in run workload run kv tole workload run kv tolerate errors init histograms perf stats json concurrency splits duration read percent pgurl returned command problem ssh verbose log retained in ssh workload run kv tole log exit status monitor go wait monitor failure monitor task failed t fatal was called parameters roachtest cloud aws roachtest cpu roachtest encrypted false roachtest fs roachtest localssd true roachtest ssd help see see cc cockroachdb obs inf prs
1
298,136
9,196,195,706
IssuesEvent
2019-03-07 06:06:15
BentoBoxWorld/BentoBox
https://api.github.com/repos/BentoBoxWorld/BentoBox
closed
When PvP disabled on island, projectiles disappear
Priority: High Status: Under investigation Type: Bug
### Description #### Describe the bug <!-- A clear and concise description of the problem you're encountering. --> <!-- /!\ Leaving this section blank will result in your ticket being closed without further explanation. --> <!-- Please type below this line. --> When PvP is off, ff you shoot someone in the island with a bow/trident/..., the arrow or the trident disappear. I think it comes from this https://github.com/BentoBoxWorld/BentoBox/blob/587e400287a92dd28e8989fb837c01e39c92c292/src/main/java/world/bentobox/bentobox/listeners/flags/settings/PVPListener.java#L92 ##### Configuration <!-- /!\ Leaving this section blank will result in your ticket being closed without further explanation. --> <!-- Please replace the underscores with your answer. Do not remove the '*' characters. --> - BentoBox: 1.3.0
1.0
When PvP disabled on island, projectiles disappear - ### Description #### Describe the bug <!-- A clear and concise description of the problem you're encountering. --> <!-- /!\ Leaving this section blank will result in your ticket being closed without further explanation. --> <!-- Please type below this line. --> When PvP is off, ff you shoot someone in the island with a bow/trident/..., the arrow or the trident disappear. I think it comes from this https://github.com/BentoBoxWorld/BentoBox/blob/587e400287a92dd28e8989fb837c01e39c92c292/src/main/java/world/bentobox/bentobox/listeners/flags/settings/PVPListener.java#L92 ##### Configuration <!-- /!\ Leaving this section blank will result in your ticket being closed without further explanation. --> <!-- Please replace the underscores with your answer. Do not remove the '*' characters. --> - BentoBox: 1.3.0
non_test
when pvp disabled on island projectiles disappear description describe the bug when pvp is off ff you shoot someone in the island with a bow trident the arrow or the trident disappear i think it comes from this configuration bentobox
0
201,599
15,214,737,461
IssuesEvent
2021-02-17 13:35:49
artalar/reatom
https://api.github.com/repos/artalar/reatom
closed
Using dtslint for write tests for types
CI&CD tests types
https://github.com/SamVerschueren/tsd https://github.com/microsoft/dtslint Alternative https://github.com/dsherret/conditional-type-checks or (`@toriningen` prompted) https://github.com/microsoft/TypeScript/blob/master/src/harness/fourslash.ts
1.0
Using dtslint for write tests for types - https://github.com/SamVerschueren/tsd https://github.com/microsoft/dtslint Alternative https://github.com/dsherret/conditional-type-checks or (`@toriningen` prompted) https://github.com/microsoft/TypeScript/blob/master/src/harness/fourslash.ts
test
using dtslint for write tests for types alternative or toriningen prompted
1
127,670
10,477,839,385
IssuesEvent
2019-09-23 21:55:43
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachtest: tpcc/w=100/nodes=3/chaos=true failed
C-test-failure O-roachtest O-robot
SHA: https://github.com/cockroachdb/cockroach/commits/4784fe3c51545db5fb5d411937ec1db2ef2b9761 Parameters: To repro, try: ``` # Don't forget to check out a clean suitable branch and experiment with the # stress invocation until the desired results present themselves. For example, # using stress instead of stressrace and passing the '-p' stressflag which # controls concurrency. ./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh cd ~/go/src/github.com/cockroachdb/cockroach && \ stdbuf -oL -eL \ make stressrace TESTS=tpcc/w=100/nodes=3/chaos=true PKG=roachtest TESTTIMEOUT=5m STRESSFLAGS='-maxtime 20m -timeout 10m' 2>&1 | tee /tmp/stress.log ``` Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=1472753&tab=buildLog ``` The test failed on branch=provisional_201909060000_v19.2.0-beta.20190910, cloud=gce: test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/20190906-1472753/tpcc/w=100/nodes=3/chaos=true/run_1 cluster.go:1735,tpcc.go:162,tpcc.go:164,tpcc.go:175,tpcc.go:330,test_runner.go:688: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-1567786968-16-n4cpu4:4 -- ./workload fixtures load tpcc --warehouses=100 {pgurl:1} returned: stderr: stdout: I190906 16:28:18.477370 1 ccl/workloadccl/cliccl/fixtures.go:283 starting restore of 9 tables I190906 16:28:35.685660 86 ccl/workloadccl/fixture.go:553 loaded 11 MiB table new_order in 17.200659374s (900000 rows, 0 index entries, 664 KiB) Error: exit status 255 : exit status 1 ```
2.0
roachtest: tpcc/w=100/nodes=3/chaos=true failed - SHA: https://github.com/cockroachdb/cockroach/commits/4784fe3c51545db5fb5d411937ec1db2ef2b9761 Parameters: To repro, try: ``` # Don't forget to check out a clean suitable branch and experiment with the # stress invocation until the desired results present themselves. For example, # using stress instead of stressrace and passing the '-p' stressflag which # controls concurrency. ./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh cd ~/go/src/github.com/cockroachdb/cockroach && \ stdbuf -oL -eL \ make stressrace TESTS=tpcc/w=100/nodes=3/chaos=true PKG=roachtest TESTTIMEOUT=5m STRESSFLAGS='-maxtime 20m -timeout 10m' 2>&1 | tee /tmp/stress.log ``` Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=1472753&tab=buildLog ``` The test failed on branch=provisional_201909060000_v19.2.0-beta.20190910, cloud=gce: test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/20190906-1472753/tpcc/w=100/nodes=3/chaos=true/run_1 cluster.go:1735,tpcc.go:162,tpcc.go:164,tpcc.go:175,tpcc.go:330,test_runner.go:688: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-1567786968-16-n4cpu4:4 -- ./workload fixtures load tpcc --warehouses=100 {pgurl:1} returned: stderr: stdout: I190906 16:28:18.477370 1 ccl/workloadccl/cliccl/fixtures.go:283 starting restore of 9 tables I190906 16:28:35.685660 86 ccl/workloadccl/fixture.go:553 loaded 11 MiB table new_order in 17.200659374s (900000 rows, 0 index entries, 664 KiB) Error: exit status 255 : exit status 1 ```
test
roachtest tpcc w nodes chaos true failed sha parameters to repro try don t forget to check out a clean suitable branch and experiment with the stress invocation until the desired results present themselves for example using stress instead of stressrace and passing the p stressflag which controls concurrency scripts gceworker sh start scripts gceworker sh mosh cd go src github com cockroachdb cockroach stdbuf ol el make stressrace tests tpcc w nodes chaos true pkg roachtest testtimeout stressflags maxtime timeout tee tmp stress log failed test the test failed on branch provisional beta cloud gce test artifacts and logs in home agent work go src github com cockroachdb cockroach artifacts tpcc w nodes chaos true run cluster go tpcc go tpcc go tpcc go tpcc go test runner go home agent work go src github com cockroachdb cockroach bin roachprod run teamcity workload fixtures load tpcc warehouses pgurl returned stderr stdout ccl workloadccl cliccl fixtures go starting restore of tables ccl workloadccl fixture go loaded mib table new order in rows index entries kib error exit status exit status
1
299,526
25,909,517,814
IssuesEvent
2022-12-15 12:56:36
eclipse-openj9/openj9
https://api.github.com/repos/eclipse-openj9/openj9
opened
JDK19 serviceability_jvmti_j9_0_FAILED j9mm.107 * ** ASSERTION FAILED ** at gc_base/GCExtensions.cpp:308: ((false && ((0) == continuation)))
comp:gc test failure
Failure link ------------ From [an internal build](https://hyc-runtimes-jenkins.swg-devops.com/job/Test_openjdk19_j9_extended.openjdk_ppc64_aix/27/)(`paix820`): ``` openjdk version "19.0.1" 2022-10-18 IBM Semeru Runtime Open Edition 19.0.1+10 (build 19.0.1+10) Eclipse OpenJ9 VM 19.0.1+10 (build master-f0fef59ea, JRE 19 AIX ppc64-64-Bit Compressed References 20221018_7 (JIT enabled, AOT enabled) OpenJ9 - f0fef59ea OMR - 8f5212d52 JCL - d78ba9640f based on jdk-19.0.1+10) ``` [Rerun in Grinder](https://hyc-runtimes-jenkins.swg-devops.com/job/Grinder/parambuild/?SDK_RESOURCE=customized&amp;TARGET=serviceability_jvmti_j9_0&amp;TEST_FLAG=&amp;UPSTREAM_TEST_JOB_NAME=&amp;DOCKER_REQUIRED=false&amp;ACTIVE_NODE_TIMEOUT=0&amp;VENDOR_TEST_DIRS=&amp;EXTRA_DOCKER_ARGS=&amp;TKG_OWNER_BRANCH=adoptium%3Amaster&amp;OPENJ9_SYSTEMTEST_OWNER_BRANCH=eclipse%3Amaster&amp;PLATFORM=ppc64_aix&amp;GENERATE_JOBS=true&amp;KEEP_REPORTDIR=true&amp;PERSONAL_BUILD=false&amp;DOCKER_REGISTRY_DIR=&amp;ADOPTOPENJDK_REPO=https%3A%2F%2Fgithub.com%2Fadoptium%2Faqa-tests.git&amp;RERUN_ITERATIONS=0&amp;DOCKER_REGISTRY_URL_CREDENTIAL_ID=&amp;LABEL=&amp;EXTRA_OPTIONS=&amp;CUSTOMIZED_SDK_URL=+https%3A%2F%2Fna.artifactory.swg-devops.com%2Fartifactory%2Fsys-rt-generic-local%2Fhyc-runtimes-jenkins.swg-devops.com%2Fbuild-scripts%2Fjobs%2Fjdk19u%2Fjdk19u-aix-ppc64-openj9%2F7%2Fibm-semeru-open-jre_ppc64_aix_JDK19U_2022-12-14-21-57.tar.gz+https%3A%2F%2Fna.artifactory.swg-devops.com%2Fartifactory%2Fsys-rt-generic-local%2Fhyc-runtimes-jenkins.swg-devops.com%2Fbuild-scripts%2Fjobs%2Fjdk19u%2Fjdk19u-aix-ppc64-openj9%2F7%2Fibm-semeru-open-testimage_ppc64_aix_JDK19U_2022-12-14-21-57.tar.gz+https%3A%2F%2Fna.artifactory.swg-devops.com%2Fartifactory%2Fsys-rt-generic-local%2Fhyc-runtimes-jenkins.swg-devops.com%2Fbuild-scripts%2Fjobs%2Fjdk19u%2Fjdk19u-aix-ppc64-openj9%2F7%2Fibm-semeru-open-jdk_ppc64_aix_JDK19U_2022-12-14-21-57.tar.gz+https%3A%2F%2Fna.artifactory.swg-devops.com%2Fartifactory%2Fsys-rt-generic-local%2Fhyc-runtimes-jenkins.swg-devops.com%2Fbuild-scripts%2Fjobs%2Fjdk19u%2Fjdk19u-aix-ppc64-openj9%2F7%2Fibm-semeru-open-debugimage_ppc64_aix_JDK19U_2022-12-14-21-57.tar.gz&amp;BUILD_IDENTIFIER=&amp;ADOPTOPENJDK_BRANCH=master&amp;LIGHT_WEIGHT_CHECKOUT=false&amp;USE_JRE=false&amp;ARTIFACTORY_SERVER=na.artifactory.swg-devops&amp;KEEP_WORKSPACE=false&amp;USER_CREDENTIALS_ID=83181e25-eea4-4f55-8b3e-e79615733226&amp;JDK_VERSION=19&amp;DOCKER_REGISTRY_URL=&amp;ITERATIONS=1&amp;VENDOR_TEST_REPOS=&amp;JDK_REPO=https%3A%2F%2Fgithub.com%2Fibmruntimes%2Fopenj9-openjdk-jdk19&amp;JCK_GIT_BRANCH=main&amp;RELEASE_TAG=&amp;OPENJ9_BRANCH=master&amp;OPENJ9_SHA=&amp;JCK_GIT_REPO=&amp;VENDOR_TEST_BRANCHES=&amp;OPENJ9_REPO=https%3A%2F%2Fgithub.com%2Feclipse-openj9%2Fopenj9.git&amp;UPSTREAM_JOB_NAME=&amp;CLOUD_PROVIDER=&amp;CUSTOM_TARGET=&amp;VENDOR_TEST_SHAS=&amp;JDK_BRANCH=openj9&amp;LABEL_ADDITION=&amp;ARTIFACTORY_REPO=&amp;ARTIFACTORY_ROOT_DIR=&amp;UPSTREAM_TEST_JOB_NUMBER=&amp;DOCKERIMAGE_TAG=&amp;JDK_IMPL=openj9&amp;TEST_TIME=&amp;SSH_AGENT_CREDENTIAL=83181e25-eea4-4f55-8b3e-e79615733226&amp;AUTO_DETECT=true&amp;SLACK_CHANNEL=%23rt-jenkins&amp;DYNAMIC_COMPILE=false&amp;ADOPTOPENJDK_SYSTEMTEST_OWNER_BRANCH=adoptium%3Amaster&amp;APPLICATION_OPTIONS=&amp;CUSTOMIZED_SDK_URL_CREDENTIAL_ID=4e18ffe7-b1b1-4272-9979-99769b68bcc2&amp;ARCHIVE_TEST_RESULTS=false&amp;NUM_MACHINES=&amp;OPENJDK_SHA=&amp;TRSS_URL=http%3A%2F%2Ftrss1.fyre.ibm.com&amp;USE_TESTENV_PROPERTIES=false&amp;BUILD_LIST=openjdk&amp;UPSTREAM_JOB_NUMBER=&amp;STF_OWNER_BRANCH=adoptium%3Amaster&amp;TIME_LIMIT=20&amp;JVM_OPTIONS=&amp;PARALLEL=None) - Change TARGET to run only the failed test targets. Optional info ------------- Failure output (captured from console output) --------------------------------------------- ``` [2022-12-14T23:52:35.576Z] variation: Mode150 [2022-12-14T23:52:35.576Z] JVM_OPTIONS: -XX:+UseCompressedOops [2022-12-15T00:01:15.949Z] TEST: serviceability/jvmti/vthread/RedefineClasses/RedefineRunningMethods.java [2022-12-15T00:01:15.950Z] STDERR: [2022-12-15T00:01:15.950Z] JVMJ9VM007W Command-line option unrecognised: -Xlog:redefine+class+iklass+add=trace,redefine+class+iklass+purge=trace,class+loader+data=debug,safepoint+cleanup,gc+phases=debug:rt.log [2022-12-15T00:01:15.953Z] 00:01:06.658 0x301b4c00 j9mm.107 * ** ASSERTION FAILED ** at /home/jenkins/workspace/build-scripts/jobs/jdk19u/jdk19u-aix-ppc64-openj9/workspace/build/src/openj9/runtime/gc_base/GCExtensions.cpp:308: ((false && ((0) == continuation))) [2022-12-15T00:01:15.953Z] JVMDUMP039I Processing dump event "traceassert", detail "" at 2022/12/14 19:01:06 - please wait. [2022-12-15T00:01:15.953Z] TEST RESULT: Failed. Unexpected exit from test [exit code: 255] [2022-12-15T00:01:15.953Z] -------------------------------------------------- [2022-12-15T00:01:25.362Z] Test results: passed: 113; failed: 1 [2022-12-15T00:01:35.843Z] Report written to /home/jenkins/workspace/Test_openjdk19_j9_extended.openjdk_ppc64_aix/jvmtest/openjdk/report/html/report.html [2022-12-15T00:01:35.843Z] Results written to /home/jenkins/workspace/Test_openjdk19_j9_extended.openjdk_ppc64_aix/aqa-tests/TKG/output_16710617857732/serviceability_jvmti_j9_0/work [2022-12-15T00:01:35.843Z] Error: Some tests failed or other problems occurred. [2022-12-15T00:01:35.843Z] ----------------------------------- [2022-12-15T00:01:35.843Z] serviceability_jvmti_j9_0_FAILED ``` [20x internal grinder](https://hyc-runtimes-jenkins.swg-devops.com/job/Grinder/30295/)
1.0
JDK19 serviceability_jvmti_j9_0_FAILED j9mm.107 * ** ASSERTION FAILED ** at gc_base/GCExtensions.cpp:308: ((false && ((0) == continuation))) - Failure link ------------ From [an internal build](https://hyc-runtimes-jenkins.swg-devops.com/job/Test_openjdk19_j9_extended.openjdk_ppc64_aix/27/)(`paix820`): ``` openjdk version "19.0.1" 2022-10-18 IBM Semeru Runtime Open Edition 19.0.1+10 (build 19.0.1+10) Eclipse OpenJ9 VM 19.0.1+10 (build master-f0fef59ea, JRE 19 AIX ppc64-64-Bit Compressed References 20221018_7 (JIT enabled, AOT enabled) OpenJ9 - f0fef59ea OMR - 8f5212d52 JCL - d78ba9640f based on jdk-19.0.1+10) ``` [Rerun in Grinder](https://hyc-runtimes-jenkins.swg-devops.com/job/Grinder/parambuild/?SDK_RESOURCE=customized&amp;TARGET=serviceability_jvmti_j9_0&amp;TEST_FLAG=&amp;UPSTREAM_TEST_JOB_NAME=&amp;DOCKER_REQUIRED=false&amp;ACTIVE_NODE_TIMEOUT=0&amp;VENDOR_TEST_DIRS=&amp;EXTRA_DOCKER_ARGS=&amp;TKG_OWNER_BRANCH=adoptium%3Amaster&amp;OPENJ9_SYSTEMTEST_OWNER_BRANCH=eclipse%3Amaster&amp;PLATFORM=ppc64_aix&amp;GENERATE_JOBS=true&amp;KEEP_REPORTDIR=true&amp;PERSONAL_BUILD=false&amp;DOCKER_REGISTRY_DIR=&amp;ADOPTOPENJDK_REPO=https%3A%2F%2Fgithub.com%2Fadoptium%2Faqa-tests.git&amp;RERUN_ITERATIONS=0&amp;DOCKER_REGISTRY_URL_CREDENTIAL_ID=&amp;LABEL=&amp;EXTRA_OPTIONS=&amp;CUSTOMIZED_SDK_URL=+https%3A%2F%2Fna.artifactory.swg-devops.com%2Fartifactory%2Fsys-rt-generic-local%2Fhyc-runtimes-jenkins.swg-devops.com%2Fbuild-scripts%2Fjobs%2Fjdk19u%2Fjdk19u-aix-ppc64-openj9%2F7%2Fibm-semeru-open-jre_ppc64_aix_JDK19U_2022-12-14-21-57.tar.gz+https%3A%2F%2Fna.artifactory.swg-devops.com%2Fartifactory%2Fsys-rt-generic-local%2Fhyc-runtimes-jenkins.swg-devops.com%2Fbuild-scripts%2Fjobs%2Fjdk19u%2Fjdk19u-aix-ppc64-openj9%2F7%2Fibm-semeru-open-testimage_ppc64_aix_JDK19U_2022-12-14-21-57.tar.gz+https%3A%2F%2Fna.artifactory.swg-devops.com%2Fartifactory%2Fsys-rt-generic-local%2Fhyc-runtimes-jenkins.swg-devops.com%2Fbuild-scripts%2Fjobs%2Fjdk19u%2Fjdk19u-aix-ppc64-openj9%2F7%2Fibm-semeru-open-jdk_ppc64_aix_JDK19U_2022-12-14-21-57.tar.gz+https%3A%2F%2Fna.artifactory.swg-devops.com%2Fartifactory%2Fsys-rt-generic-local%2Fhyc-runtimes-jenkins.swg-devops.com%2Fbuild-scripts%2Fjobs%2Fjdk19u%2Fjdk19u-aix-ppc64-openj9%2F7%2Fibm-semeru-open-debugimage_ppc64_aix_JDK19U_2022-12-14-21-57.tar.gz&amp;BUILD_IDENTIFIER=&amp;ADOPTOPENJDK_BRANCH=master&amp;LIGHT_WEIGHT_CHECKOUT=false&amp;USE_JRE=false&amp;ARTIFACTORY_SERVER=na.artifactory.swg-devops&amp;KEEP_WORKSPACE=false&amp;USER_CREDENTIALS_ID=83181e25-eea4-4f55-8b3e-e79615733226&amp;JDK_VERSION=19&amp;DOCKER_REGISTRY_URL=&amp;ITERATIONS=1&amp;VENDOR_TEST_REPOS=&amp;JDK_REPO=https%3A%2F%2Fgithub.com%2Fibmruntimes%2Fopenj9-openjdk-jdk19&amp;JCK_GIT_BRANCH=main&amp;RELEASE_TAG=&amp;OPENJ9_BRANCH=master&amp;OPENJ9_SHA=&amp;JCK_GIT_REPO=&amp;VENDOR_TEST_BRANCHES=&amp;OPENJ9_REPO=https%3A%2F%2Fgithub.com%2Feclipse-openj9%2Fopenj9.git&amp;UPSTREAM_JOB_NAME=&amp;CLOUD_PROVIDER=&amp;CUSTOM_TARGET=&amp;VENDOR_TEST_SHAS=&amp;JDK_BRANCH=openj9&amp;LABEL_ADDITION=&amp;ARTIFACTORY_REPO=&amp;ARTIFACTORY_ROOT_DIR=&amp;UPSTREAM_TEST_JOB_NUMBER=&amp;DOCKERIMAGE_TAG=&amp;JDK_IMPL=openj9&amp;TEST_TIME=&amp;SSH_AGENT_CREDENTIAL=83181e25-eea4-4f55-8b3e-e79615733226&amp;AUTO_DETECT=true&amp;SLACK_CHANNEL=%23rt-jenkins&amp;DYNAMIC_COMPILE=false&amp;ADOPTOPENJDK_SYSTEMTEST_OWNER_BRANCH=adoptium%3Amaster&amp;APPLICATION_OPTIONS=&amp;CUSTOMIZED_SDK_URL_CREDENTIAL_ID=4e18ffe7-b1b1-4272-9979-99769b68bcc2&amp;ARCHIVE_TEST_RESULTS=false&amp;NUM_MACHINES=&amp;OPENJDK_SHA=&amp;TRSS_URL=http%3A%2F%2Ftrss1.fyre.ibm.com&amp;USE_TESTENV_PROPERTIES=false&amp;BUILD_LIST=openjdk&amp;UPSTREAM_JOB_NUMBER=&amp;STF_OWNER_BRANCH=adoptium%3Amaster&amp;TIME_LIMIT=20&amp;JVM_OPTIONS=&amp;PARALLEL=None) - Change TARGET to run only the failed test targets. Optional info ------------- Failure output (captured from console output) --------------------------------------------- ``` [2022-12-14T23:52:35.576Z] variation: Mode150 [2022-12-14T23:52:35.576Z] JVM_OPTIONS: -XX:+UseCompressedOops [2022-12-15T00:01:15.949Z] TEST: serviceability/jvmti/vthread/RedefineClasses/RedefineRunningMethods.java [2022-12-15T00:01:15.950Z] STDERR: [2022-12-15T00:01:15.950Z] JVMJ9VM007W Command-line option unrecognised: -Xlog:redefine+class+iklass+add=trace,redefine+class+iklass+purge=trace,class+loader+data=debug,safepoint+cleanup,gc+phases=debug:rt.log [2022-12-15T00:01:15.953Z] 00:01:06.658 0x301b4c00 j9mm.107 * ** ASSERTION FAILED ** at /home/jenkins/workspace/build-scripts/jobs/jdk19u/jdk19u-aix-ppc64-openj9/workspace/build/src/openj9/runtime/gc_base/GCExtensions.cpp:308: ((false && ((0) == continuation))) [2022-12-15T00:01:15.953Z] JVMDUMP039I Processing dump event "traceassert", detail "" at 2022/12/14 19:01:06 - please wait. [2022-12-15T00:01:15.953Z] TEST RESULT: Failed. Unexpected exit from test [exit code: 255] [2022-12-15T00:01:15.953Z] -------------------------------------------------- [2022-12-15T00:01:25.362Z] Test results: passed: 113; failed: 1 [2022-12-15T00:01:35.843Z] Report written to /home/jenkins/workspace/Test_openjdk19_j9_extended.openjdk_ppc64_aix/jvmtest/openjdk/report/html/report.html [2022-12-15T00:01:35.843Z] Results written to /home/jenkins/workspace/Test_openjdk19_j9_extended.openjdk_ppc64_aix/aqa-tests/TKG/output_16710617857732/serviceability_jvmti_j9_0/work [2022-12-15T00:01:35.843Z] Error: Some tests failed or other problems occurred. [2022-12-15T00:01:35.843Z] ----------------------------------- [2022-12-15T00:01:35.843Z] serviceability_jvmti_j9_0_FAILED ``` [20x internal grinder](https://hyc-runtimes-jenkins.swg-devops.com/job/Grinder/30295/)
test
serviceability jvmti failed assertion failed at gc base gcextensions cpp false continuation failure link from openjdk version ibm semeru runtime open edition build eclipse vm build master jre aix bit compressed references jit enabled aot enabled omr jcl based on jdk change target to run only the failed test targets optional info failure output captured from console output variation jvm options xx usecompressedoops test serviceability jvmti vthread redefineclasses redefinerunningmethods java stderr command line option unrecognised xlog redefine class iklass add trace redefine class iklass purge trace class loader data debug safepoint cleanup gc phases debug rt log assertion failed at home jenkins workspace build scripts jobs aix workspace build src runtime gc base gcextensions cpp false continuation processing dump event traceassert detail at please wait test result failed unexpected exit from test test results passed failed report written to home jenkins workspace test extended openjdk aix jvmtest openjdk report html report html results written to home jenkins workspace test extended openjdk aix aqa tests tkg output serviceability jvmti work error some tests failed or other problems occurred serviceability jvmti failed
1
441,562
12,719,751,067
IssuesEvent
2020-06-24 09:48:22
wso2/devstudio-tooling-ei
https://api.github.com/repos/wso2/devstudio-tooling-ei
closed
Cannot add onError sequence for sequence templates
Priority/Highest Severity/Critical
**Description:** The added onError sequence is not persisted. ![image](https://user-images.githubusercontent.com/4967046/84590147-9f837d00-ae51-11ea-87c5-d34b066da7c8.png) As per doc it is supported ![image](https://user-images.githubusercontent.com/4967046/84590164-c80b7700-ae51-11ea-99fb-96c92828bc36.png) **Related Issues:** https://github.com/wso2/devstudio-tooling-ei/issues/1089
1.0
Cannot add onError sequence for sequence templates - **Description:** The added onError sequence is not persisted. ![image](https://user-images.githubusercontent.com/4967046/84590147-9f837d00-ae51-11ea-87c5-d34b066da7c8.png) As per doc it is supported ![image](https://user-images.githubusercontent.com/4967046/84590164-c80b7700-ae51-11ea-99fb-96c92828bc36.png) **Related Issues:** https://github.com/wso2/devstudio-tooling-ei/issues/1089
non_test
cannot add onerror sequence for sequence templates description the added onerror sequence is not persisted as per doc it is supported related issues
0
151,988
12,067,632,602
IssuesEvent
2020-04-16 13:40:51
w3c/webrtc-pc
https://api.github.com/repos/w3c/webrtc-pc
closed
Test coverage chapter 7
Test suite issue
This is a sub-issue of #2442, covering whether sections of chapter 7 (DTMF) are marked correctly as testable (red/green) or nontestable (white). - 7.1 attribute dtmf is certainly testable - 7.4 attribute tone of RTCDTMFToneChangeEvent and -init are certainly testable
1.0
Test coverage chapter 7 - This is a sub-issue of #2442, covering whether sections of chapter 7 (DTMF) are marked correctly as testable (red/green) or nontestable (white). - 7.1 attribute dtmf is certainly testable - 7.4 attribute tone of RTCDTMFToneChangeEvent and -init are certainly testable
test
test coverage chapter this is a sub issue of covering whether sections of chapter dtmf are marked correctly as testable red green or nontestable white attribute dtmf is certainly testable attribute tone of rtcdtmftonechangeevent and init are certainly testable
1
71,335
7,242,521,919
IssuesEvent
2018-02-14 08:10:14
healthlocker/oxleas-adhd
https://api.github.com/repos/healthlocker/oxleas-adhd
closed
Display correct formatting on coping strategies, to clinicians
T25m enhancement please-test
See: https://github.com/healthlocker/healthlocker/issues/1131 - [ ] As a clinician I see the any formatting a SU has entered displaying on their goals (unlike the healthlocker issue above, coping strategies seem fine on HeadScape) ## **I entered:** ![screenshot 2018-02-05 14 59 29](https://user-images.githubusercontent.com/24604903/35810949-3144e3dc-0a85-11e8-9125-030cfd8b4bca.png) ## **Which displayed to me as a service user:** ![screenshot 2018-02-05 15 00 06](https://user-images.githubusercontent.com/24604903/35810976-45deb99e-0a85-11e8-8264-ec66e2a420b6.png) ## **But looks like this for a clinician:** ![screenshot 2018-02-05 14 56 02](https://user-images.githubusercontent.com/24604903/35810858-f52b1ed4-0a84-11e8-813a-270835e9f758.png) (time estimate based on the time estimate given to the HL issue about the same)
1.0
Display correct formatting on coping strategies, to clinicians - See: https://github.com/healthlocker/healthlocker/issues/1131 - [ ] As a clinician I see the any formatting a SU has entered displaying on their goals (unlike the healthlocker issue above, coping strategies seem fine on HeadScape) ## **I entered:** ![screenshot 2018-02-05 14 59 29](https://user-images.githubusercontent.com/24604903/35810949-3144e3dc-0a85-11e8-9125-030cfd8b4bca.png) ## **Which displayed to me as a service user:** ![screenshot 2018-02-05 15 00 06](https://user-images.githubusercontent.com/24604903/35810976-45deb99e-0a85-11e8-8264-ec66e2a420b6.png) ## **But looks like this for a clinician:** ![screenshot 2018-02-05 14 56 02](https://user-images.githubusercontent.com/24604903/35810858-f52b1ed4-0a84-11e8-813a-270835e9f758.png) (time estimate based on the time estimate given to the HL issue about the same)
test
display correct formatting on coping strategies to clinicians see as a clinician i see the any formatting a su has entered displaying on their goals unlike the healthlocker issue above coping strategies seem fine on headscape i entered which displayed to me as a service user but looks like this for a clinician time estimate based on the time estimate given to the hl issue about the same
1
333,441
10,122,077,448
IssuesEvent
2019-07-31 17:04:07
pulumi/pulumi-cloud
https://api.github.com/repos/pulumi/pulumi-cloud
closed
CI is failing
priority/P1
Three tests are currently failing with: ``` pulumi:pulumi:Stack table-p-it-travis-job-table-b0c07e08 creating Error: Secret outputs cannot be captured by a closure. ``` This was caused by https://github.com/pulumi/pulumi-terraform/pull/427 (and exacerbated by https://github.com/pulumi/pulumi/issues/2718).
1.0
CI is failing - Three tests are currently failing with: ``` pulumi:pulumi:Stack table-p-it-travis-job-table-b0c07e08 creating Error: Secret outputs cannot be captured by a closure. ``` This was caused by https://github.com/pulumi/pulumi-terraform/pull/427 (and exacerbated by https://github.com/pulumi/pulumi/issues/2718).
non_test
ci is failing three tests are currently failing with pulumi pulumi stack table p it travis job table creating error secret outputs cannot be captured by a closure this was caused by and exacerbated by
0
47,010
5,845,625,073
IssuesEvent
2017-05-10 14:36:34
openshift/origin
https://api.github.com/repos/openshift/origin
closed
3.6 Upgrade Failure: Imagebuilder e2e tests fail on policy
component/install kind/bug kind/test-flake priority/P1
When installing Origin 1.5 and upgrading to Origin 3.6, we see 2 `imagebuilder` policy tests fail on the e2e suite: - Extended.[builds] Optimized image builds should fail as a normal user [Conformance] - Extended.[builds] Optimized image builds should succeed as an admin [Conformance] These can be seen in all `test_branch_origin_extended_conformance_install_upgrade` jobs until [299](https://ci.openshift.redhat.com/jenkins/job/test_branch_origin_extended_conformance_install_update/299/testReport/)
1.0
3.6 Upgrade Failure: Imagebuilder e2e tests fail on policy - When installing Origin 1.5 and upgrading to Origin 3.6, we see 2 `imagebuilder` policy tests fail on the e2e suite: - Extended.[builds] Optimized image builds should fail as a normal user [Conformance] - Extended.[builds] Optimized image builds should succeed as an admin [Conformance] These can be seen in all `test_branch_origin_extended_conformance_install_upgrade` jobs until [299](https://ci.openshift.redhat.com/jenkins/job/test_branch_origin_extended_conformance_install_update/299/testReport/)
test
upgrade failure imagebuilder tests fail on policy when installing origin and upgrading to origin we see imagebuilder policy tests fail on the suite extended optimized image builds should fail as a normal user extended optimized image builds should succeed as an admin these can be seen in all test branch origin extended conformance install upgrade jobs until
1
107,779
9,223,213,723
IssuesEvent
2019-03-12 02:26:21
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachtest: schemachange/index/tpcc/w=100 failed
C-test-failure O-roachtest O-robot
SHA: https://github.com/cockroachdb/cockroach/commits/959dcf7de0f94cfcfa0062387b109adebd1f11da Parameters: To repro, try: ``` # Don't forget to check out a clean suitable branch and experiment with the # stress invocation until the desired results present themselves. For example, # using stress instead of stressrace and passing the '-p' stressflag which # controls concurrency. ./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh cd ~/go/src/github.com/cockroachdb/cockroach && \ stdbuf -oL -eL \ make stressrace TESTS=schemachange/index/tpcc/w=100 PKG=roachtest TESTTIMEOUT=5m STRESSFLAGS='-maxtime 20m -timeout 10m' 2>&1 | tee /tmp/stress.log ``` Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=1163702&tab=buildLog ``` The test failed on master: cluster.go:1244,tpcc.go:115,tpcc.go:117,schemachange.go:310,test.go:1214: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-1163702-schemachange-index-tpcc-w-100:5 -- ./workload fixtures load tpcc --checks=true --warehouses=100 {pgurl:1} returned: stderr: stdout: iB) I190306 13:25:33.517885 30 ccl/workloadccl/fixture.go:527 loaded item (1m6s, 100000 rows, 0 index entries, 7.8 MiB) I190306 13:26:08.582050 28 ccl/workloadccl/fixture.go:527 loaded order (1m41s, 3000000 rows, 6000000 index entries, 166 MiB) I190306 13:27:04.580047 24 ccl/workloadccl/fixture.go:527 loaded warehouse (2m37s, 100 rows, 0 index entries, 5.1 KiB) I190306 13:27:54.841091 27 ccl/workloadccl/fixture.go:527 loaded history (3m27s, 3000000 rows, 6000000 index entries, 380 MiB) I190306 13:28:12.861772 32 ccl/workloadccl/fixture.go:527 loaded order_line (3m45s, 29995875 rows, 29995875 index entries, 2.0 GiB) I190306 13:28:22.754536 26 ccl/workloadccl/fixture.go:527 loaded customer (3m55s, 3000000 rows, 3000000 index entries, 1.7 GiB) I190306 13:28:54.672729 31 ccl/workloadccl/fixture.go:527 loaded stock (4m27s, 10000000 rows, 10000000 index entries, 3.1 GiB) Error: restoring fixture: PostLoad hook: pq: count-leases: AS OF SYSTEM TIME: cannot specify timestamp in the future Error: exit status 1 : exit status 1 ```
2.0
roachtest: schemachange/index/tpcc/w=100 failed - SHA: https://github.com/cockroachdb/cockroach/commits/959dcf7de0f94cfcfa0062387b109adebd1f11da Parameters: To repro, try: ``` # Don't forget to check out a clean suitable branch and experiment with the # stress invocation until the desired results present themselves. For example, # using stress instead of stressrace and passing the '-p' stressflag which # controls concurrency. ./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh cd ~/go/src/github.com/cockroachdb/cockroach && \ stdbuf -oL -eL \ make stressrace TESTS=schemachange/index/tpcc/w=100 PKG=roachtest TESTTIMEOUT=5m STRESSFLAGS='-maxtime 20m -timeout 10m' 2>&1 | tee /tmp/stress.log ``` Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=1163702&tab=buildLog ``` The test failed on master: cluster.go:1244,tpcc.go:115,tpcc.go:117,schemachange.go:310,test.go:1214: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-1163702-schemachange-index-tpcc-w-100:5 -- ./workload fixtures load tpcc --checks=true --warehouses=100 {pgurl:1} returned: stderr: stdout: iB) I190306 13:25:33.517885 30 ccl/workloadccl/fixture.go:527 loaded item (1m6s, 100000 rows, 0 index entries, 7.8 MiB) I190306 13:26:08.582050 28 ccl/workloadccl/fixture.go:527 loaded order (1m41s, 3000000 rows, 6000000 index entries, 166 MiB) I190306 13:27:04.580047 24 ccl/workloadccl/fixture.go:527 loaded warehouse (2m37s, 100 rows, 0 index entries, 5.1 KiB) I190306 13:27:54.841091 27 ccl/workloadccl/fixture.go:527 loaded history (3m27s, 3000000 rows, 6000000 index entries, 380 MiB) I190306 13:28:12.861772 32 ccl/workloadccl/fixture.go:527 loaded order_line (3m45s, 29995875 rows, 29995875 index entries, 2.0 GiB) I190306 13:28:22.754536 26 ccl/workloadccl/fixture.go:527 loaded customer (3m55s, 3000000 rows, 3000000 index entries, 1.7 GiB) I190306 13:28:54.672729 31 ccl/workloadccl/fixture.go:527 loaded stock (4m27s, 10000000 rows, 10000000 index entries, 3.1 GiB) Error: restoring fixture: PostLoad hook: pq: count-leases: AS OF SYSTEM TIME: cannot specify timestamp in the future Error: exit status 1 : exit status 1 ```
test
roachtest schemachange index tpcc w failed sha parameters to repro try don t forget to check out a clean suitable branch and experiment with the stress invocation until the desired results present themselves for example using stress instead of stressrace and passing the p stressflag which controls concurrency scripts gceworker sh start scripts gceworker sh mosh cd go src github com cockroachdb cockroach stdbuf ol el make stressrace tests schemachange index tpcc w pkg roachtest testtimeout stressflags maxtime timeout tee tmp stress log failed test the test failed on master cluster go tpcc go tpcc go schemachange go test go home agent work go src github com cockroachdb cockroach bin roachprod run teamcity schemachange index tpcc w workload fixtures load tpcc checks true warehouses pgurl returned stderr stdout ib ccl workloadccl fixture go loaded item rows index entries mib ccl workloadccl fixture go loaded order rows index entries mib ccl workloadccl fixture go loaded warehouse rows index entries kib ccl workloadccl fixture go loaded history rows index entries mib ccl workloadccl fixture go loaded order line rows index entries gib ccl workloadccl fixture go loaded customer rows index entries gib ccl workloadccl fixture go loaded stock rows index entries gib error restoring fixture postload hook pq count leases as of system time cannot specify timestamp in the future error exit status exit status
1
15,946
11,779,835,835
IssuesEvent
2020-03-16 18:47:48
dotnet/roslyn
https://api.github.com/repos/dotnet/roslyn
closed
A bunch of files in the Roslyn folder are mistakenly queued for ngen
Area-Infrastructure Investigation Required
The Roslyn bits included in 4.8 SDK installed from the VS 16.3 Preview Build Tools are causing several files to be queued for ngen that are not valid files to be ngen'd. **Steps to Reproduce**: 1. Run the following steps on a Windows machine with Docker for Windows installed, targeting Windows containers. 2. Create a new folder on the desktop. 3. Download https://download.visualstudio.microsoft.com/download/pr/6bd8729b-a836-4721-8988-6b9b23649a10/fc200a5e2b37cf4052a63f0b7b66fd1b/vs_buildtools.exe to the new folder. This URL is what https://aka.ms/vs/16/pre/vs_buildtools.exe is pointing to at the time of creating this issue. 4. Create a file named "Dockerfile" in the same folder with the following contents: ``` # escape=` FROM mcr.microsoft.com/dotnet/framework/runtime:4.8-windowsservercore-ltsc2019 COPY vs_buildtools.exe . RUN start /w vs_BuildTools.exe ^ ` --add Microsoft.Net.Component.4.8.SDK ^ ` --quiet --norestart --nocache --wait RUN \Windows\Microsoft.NET\Framework\v4.0.30319\ngen update ``` 5. From command line, change the directory to the new folder that you created. Run the following command: `docker build .` **Expected Behavior**: The `docker` command should succeed. **Actual Behavior**: The `docker` command fails due to the `ngen` call in the Dockerfile. Ngen fails due to 2 sets of errors: 1. `X is not a valid Win32 application. (Exception from HRESULT: 0x800700C1)`. Caused by the following files: ``` C:\Program Files (x86)\Microsoft Visual Studio\2019\Preview\MSBuild\Current\Bin\Roslyn\csc.exe.config C:\Program Files (x86)\Microsoft Visual Studio\2019\Preview\MSBuild\Current\Bin\Roslyn\csc.rsp C:\Program Files (x86)\Microsoft Visual Studio\2019\Preview\MSBuild\Current\Bin\Roslyn\csi.exe.config C:\Program Files (x86)\Microsoft Visual Studio\2019\Preview\MSBuild\Current\Bin\Roslyn\csi.rsp C:\Program Files (x86)\Microsoft Visual Studio\2019\Preview\MSBuild\Current\Bin\Roslyn\Microsoft.CSharp.Core.targets C:\Program Files (x86)\Microsoft Visual Studio\2019\Preview\MSBuild\Current\Bin\Roslyn\Microsoft.Managed.Core.targets C:\Program Files (x86)\Microsoft Visual Studio\2019\Preview\MSBuild\Current\Bin\Roslyn\Microsoft.Managed.EditorConfig.targets C:\Program Files (x86)\Microsoft Visual Studio\2019\Preview\MSBuild\Current\Bin\Roslyn\Microsoft.VisualBasic.Core.targets C:\Program Files (x86)\Microsoft Visual Studio\2019\Preview\MSBuild\Current\Bin\Roslyn\vbc.exe.config C:\Program Files (x86)\Microsoft Visual Studio\2019\Preview\MSBuild\Current\Bin\Roslyn\vbc.rsp C:\Program Files (x86)\Microsoft Visual Studio\2019\Preview\MSBuild\Current\Bin\Roslyn\VBCSCompiler.exe.config ``` 2. `The module was expected to contain an assembly manifest. (Exception from HRESULT: 0x80131018)`. Caused by the following files: ``` C:\Program Files (x86)\Microsoft Visual Studio\2019\Preview\MSBuild\Current\Bin\Roslyn\Microsoft.DiaSymReader.Native.amd64.dll C:\Program Files (x86)\Microsoft Visual Studio\2019\Preview\MSBuild\Current\Bin\Roslyn\Microsoft.DiaSymReader.Native.x86.dll ``` Full ngen log file is attached: [log.txt](https://github.com/dotnet/roslyn/files/3542580/log.txt) This is related to #30341 but that issue may be fixed now because I don't see any cases of .pkgdef files causing ngen errors.
1.0
A bunch of files in the Roslyn folder are mistakenly queued for ngen - The Roslyn bits included in 4.8 SDK installed from the VS 16.3 Preview Build Tools are causing several files to be queued for ngen that are not valid files to be ngen'd. **Steps to Reproduce**: 1. Run the following steps on a Windows machine with Docker for Windows installed, targeting Windows containers. 2. Create a new folder on the desktop. 3. Download https://download.visualstudio.microsoft.com/download/pr/6bd8729b-a836-4721-8988-6b9b23649a10/fc200a5e2b37cf4052a63f0b7b66fd1b/vs_buildtools.exe to the new folder. This URL is what https://aka.ms/vs/16/pre/vs_buildtools.exe is pointing to at the time of creating this issue. 4. Create a file named "Dockerfile" in the same folder with the following contents: ``` # escape=` FROM mcr.microsoft.com/dotnet/framework/runtime:4.8-windowsservercore-ltsc2019 COPY vs_buildtools.exe . RUN start /w vs_BuildTools.exe ^ ` --add Microsoft.Net.Component.4.8.SDK ^ ` --quiet --norestart --nocache --wait RUN \Windows\Microsoft.NET\Framework\v4.0.30319\ngen update ``` 5. From command line, change the directory to the new folder that you created. Run the following command: `docker build .` **Expected Behavior**: The `docker` command should succeed. **Actual Behavior**: The `docker` command fails due to the `ngen` call in the Dockerfile. Ngen fails due to 2 sets of errors: 1. `X is not a valid Win32 application. (Exception from HRESULT: 0x800700C1)`. Caused by the following files: ``` C:\Program Files (x86)\Microsoft Visual Studio\2019\Preview\MSBuild\Current\Bin\Roslyn\csc.exe.config C:\Program Files (x86)\Microsoft Visual Studio\2019\Preview\MSBuild\Current\Bin\Roslyn\csc.rsp C:\Program Files (x86)\Microsoft Visual Studio\2019\Preview\MSBuild\Current\Bin\Roslyn\csi.exe.config C:\Program Files (x86)\Microsoft Visual Studio\2019\Preview\MSBuild\Current\Bin\Roslyn\csi.rsp C:\Program Files (x86)\Microsoft Visual Studio\2019\Preview\MSBuild\Current\Bin\Roslyn\Microsoft.CSharp.Core.targets C:\Program Files (x86)\Microsoft Visual Studio\2019\Preview\MSBuild\Current\Bin\Roslyn\Microsoft.Managed.Core.targets C:\Program Files (x86)\Microsoft Visual Studio\2019\Preview\MSBuild\Current\Bin\Roslyn\Microsoft.Managed.EditorConfig.targets C:\Program Files (x86)\Microsoft Visual Studio\2019\Preview\MSBuild\Current\Bin\Roslyn\Microsoft.VisualBasic.Core.targets C:\Program Files (x86)\Microsoft Visual Studio\2019\Preview\MSBuild\Current\Bin\Roslyn\vbc.exe.config C:\Program Files (x86)\Microsoft Visual Studio\2019\Preview\MSBuild\Current\Bin\Roslyn\vbc.rsp C:\Program Files (x86)\Microsoft Visual Studio\2019\Preview\MSBuild\Current\Bin\Roslyn\VBCSCompiler.exe.config ``` 2. `The module was expected to contain an assembly manifest. (Exception from HRESULT: 0x80131018)`. Caused by the following files: ``` C:\Program Files (x86)\Microsoft Visual Studio\2019\Preview\MSBuild\Current\Bin\Roslyn\Microsoft.DiaSymReader.Native.amd64.dll C:\Program Files (x86)\Microsoft Visual Studio\2019\Preview\MSBuild\Current\Bin\Roslyn\Microsoft.DiaSymReader.Native.x86.dll ``` Full ngen log file is attached: [log.txt](https://github.com/dotnet/roslyn/files/3542580/log.txt) This is related to #30341 but that issue may be fixed now because I don't see any cases of .pkgdef files causing ngen errors.
non_test
a bunch of files in the roslyn folder are mistakenly queued for ngen the roslyn bits included in sdk installed from the vs preview build tools are causing several files to be queued for ngen that are not valid files to be ngen d steps to reproduce run the following steps on a windows machine with docker for windows installed targeting windows containers create a new folder on the desktop download to the new folder this url is what is pointing to at the time of creating this issue create a file named dockerfile in the same folder with the following contents escape from mcr microsoft com dotnet framework runtime windowsservercore copy vs buildtools exe run start w vs buildtools exe add microsoft net component sdk quiet norestart nocache wait run windows microsoft net framework ngen update from command line change the directory to the new folder that you created run the following command docker build expected behavior the docker command should succeed actual behavior the docker command fails due to the ngen call in the dockerfile ngen fails due to sets of errors x is not a valid application exception from hresult caused by the following files c program files microsoft visual studio preview msbuild current bin roslyn csc exe config c program files microsoft visual studio preview msbuild current bin roslyn csc rsp c program files microsoft visual studio preview msbuild current bin roslyn csi exe config c program files microsoft visual studio preview msbuild current bin roslyn csi rsp c program files microsoft visual studio preview msbuild current bin roslyn microsoft csharp core targets c program files microsoft visual studio preview msbuild current bin roslyn microsoft managed core targets c program files microsoft visual studio preview msbuild current bin roslyn microsoft managed editorconfig targets c program files microsoft visual studio preview msbuild current bin roslyn microsoft visualbasic core targets c program files microsoft visual studio preview msbuild current bin roslyn vbc exe config c program files microsoft visual studio preview msbuild current bin roslyn vbc rsp c program files microsoft visual studio preview msbuild current bin roslyn vbcscompiler exe config the module was expected to contain an assembly manifest exception from hresult caused by the following files c program files microsoft visual studio preview msbuild current bin roslyn microsoft diasymreader native dll c program files microsoft visual studio preview msbuild current bin roslyn microsoft diasymreader native dll full ngen log file is attached this is related to but that issue may be fixed now because i don t see any cases of pkgdef files causing ngen errors
0
51,083
13,188,092,348
IssuesEvent
2020-08-13 05:32:00
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
closed
[simulation] error in MC afterprocessing (MCPulseSeriesParticleIDMap) (Trac #1942)
Migrated from Trac combo simulation defect
I was again on the search for simulation of some MCstudies, for which I need the complete I3MCPulse-info (MCPulseMap, ParticleIDMap and I3MCTRee); It took me some while to locate an appropriate one, which I found in dataset 11843 resp. 11642 (IC86:2013 GENIE simulation), one of the latest produced datasets This dataset was produced with icesim.V04-01-11-G4C (whatever this means). All other datasets I found which are produced with an olderor similar release candidate (icesim.V04-01-11) have the necessary objects, but the after-processing for Noise-MCPulses is flawed: there are no entries in the ParticleIDMap for those pulses as originated from ParticleID(0,0) This can be identified by inspecting the I3MCPulseSeriesParticleIDMap in the frame finding no entries for the major/minor-ID(0,0) and in general a lot of entries with only a single OMKey as key and an empty vector for the value in this I3Map<OMKey, vector<...> > So whatever is the difference between these meta-releases, somebody should have a look into that the after-processing is done correctly, otherwise MC-information is permanently lost! <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1942">https://code.icecube.wisc.edu/ticket/1942</a>, reported by mzoll and owned by juancarlos</em></summary> <p> ```json { "status": "closed", "changetime": "2017-02-21T17:32:43", "description": "I was again on the search for simulation of some MCstudies, for which I need the complete I3MCPulse-info (MCPulseMap, ParticleIDMap and I3MCTRee); It took me some while to locate an appropriate one, which I found in dataset 11843 resp. 11642 (IC86:2013 GENIE simulation), one of the latest produced datasets\n\nThis dataset was produced with icesim.V04-01-11-G4C (whatever this means). All other datasets I found which are produced with an olderor similar release candidate (icesim.V04-01-11) have the necessary objects, but the after-processing for Noise-MCPulses is flawed: there are no entries in the ParticleIDMap for those pulses as originated from ParticleID(0,0)\n\nThis can be identified by inspecting the I3MCPulseSeriesParticleIDMap in the frame finding no entries for the major/minor-ID(0,0) and in general a lot of entries with only a single OMKey as key and an empty vector for the value in this I3Map<OMKey, vector<...> >\n\nSo whatever is the difference between these meta-releases, somebody should have a look into that the after-processing is done correctly, otherwise MC-information is permanently lost!", "reporter": "mzoll", "cc": "olivas, cweaver", "resolution": "fixed", "_ts": "1487698363713061", "component": "combo simulation", "summary": "[simulation] error in MC afterprocessing (MCPulseSeriesParticleIDMap)", "priority": "critical", "keywords": "MC simulation MCPulse", "time": "2017-01-26T09:02:15", "milestone": "", "owner": "juancarlos", "type": "defect" } ``` </p> </details>
1.0
[simulation] error in MC afterprocessing (MCPulseSeriesParticleIDMap) (Trac #1942) - I was again on the search for simulation of some MCstudies, for which I need the complete I3MCPulse-info (MCPulseMap, ParticleIDMap and I3MCTRee); It took me some while to locate an appropriate one, which I found in dataset 11843 resp. 11642 (IC86:2013 GENIE simulation), one of the latest produced datasets This dataset was produced with icesim.V04-01-11-G4C (whatever this means). All other datasets I found which are produced with an olderor similar release candidate (icesim.V04-01-11) have the necessary objects, but the after-processing for Noise-MCPulses is flawed: there are no entries in the ParticleIDMap for those pulses as originated from ParticleID(0,0) This can be identified by inspecting the I3MCPulseSeriesParticleIDMap in the frame finding no entries for the major/minor-ID(0,0) and in general a lot of entries with only a single OMKey as key and an empty vector for the value in this I3Map<OMKey, vector<...> > So whatever is the difference between these meta-releases, somebody should have a look into that the after-processing is done correctly, otherwise MC-information is permanently lost! <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1942">https://code.icecube.wisc.edu/ticket/1942</a>, reported by mzoll and owned by juancarlos</em></summary> <p> ```json { "status": "closed", "changetime": "2017-02-21T17:32:43", "description": "I was again on the search for simulation of some MCstudies, for which I need the complete I3MCPulse-info (MCPulseMap, ParticleIDMap and I3MCTRee); It took me some while to locate an appropriate one, which I found in dataset 11843 resp. 11642 (IC86:2013 GENIE simulation), one of the latest produced datasets\n\nThis dataset was produced with icesim.V04-01-11-G4C (whatever this means). All other datasets I found which are produced with an olderor similar release candidate (icesim.V04-01-11) have the necessary objects, but the after-processing for Noise-MCPulses is flawed: there are no entries in the ParticleIDMap for those pulses as originated from ParticleID(0,0)\n\nThis can be identified by inspecting the I3MCPulseSeriesParticleIDMap in the frame finding no entries for the major/minor-ID(0,0) and in general a lot of entries with only a single OMKey as key and an empty vector for the value in this I3Map<OMKey, vector<...> >\n\nSo whatever is the difference between these meta-releases, somebody should have a look into that the after-processing is done correctly, otherwise MC-information is permanently lost!", "reporter": "mzoll", "cc": "olivas, cweaver", "resolution": "fixed", "_ts": "1487698363713061", "component": "combo simulation", "summary": "[simulation] error in MC afterprocessing (MCPulseSeriesParticleIDMap)", "priority": "critical", "keywords": "MC simulation MCPulse", "time": "2017-01-26T09:02:15", "milestone": "", "owner": "juancarlos", "type": "defect" } ``` </p> </details>
non_test
error in mc afterprocessing mcpulseseriesparticleidmap trac i was again on the search for simulation of some mcstudies for which i need the complete info mcpulsemap particleidmap and it took me some while to locate an appropriate one which i found in dataset resp genie simulation one of the latest produced datasets this dataset was produced with icesim whatever this means all other datasets i found which are produced with an olderor similar release candidate icesim have the necessary objects but the after processing for noise mcpulses is flawed there are no entries in the particleidmap for those pulses as originated from particleid this can be identified by inspecting the in the frame finding no entries for the major minor id and in general a lot of entries with only a single omkey as key and an empty vector for the value in this so whatever is the difference between these meta releases somebody should have a look into that the after processing is done correctly otherwise mc information is permanently lost migrated from json status closed changetime description i was again on the search for simulation of some mcstudies for which i need the complete info mcpulsemap particleidmap and it took me some while to locate an appropriate one which i found in dataset resp genie simulation one of the latest produced datasets n nthis dataset was produced with icesim whatever this means all other datasets i found which are produced with an olderor similar release candidate icesim have the necessary objects but the after processing for noise mcpulses is flawed there are no entries in the particleidmap for those pulses as originated from particleid n nthis can be identified by inspecting the in the frame finding no entries for the major minor id and in general a lot of entries with only a single omkey as key and an empty vector for the value in this n nso whatever is the difference between these meta releases somebody should have a look into that the after processing is done correctly otherwise mc information is permanently lost reporter mzoll cc olivas cweaver resolution fixed ts component combo simulation summary error in mc afterprocessing mcpulseseriesparticleidmap priority critical keywords mc simulation mcpulse time milestone owner juancarlos type defect
0
3,566
2,679,540,392
IssuesEvent
2015-03-26 17:11:26
rethinkdb/rethinkdb
https://api.github.com/repos/rethinkdb/rethinkdb
opened
Intermittent failure in ClusteringIssues.OutdatedIndexes unit test
cp:testing
This happened twice for me on gantz on v1.16.x, but not every time: ``` [ RUN ] ClusteringIssues.OutdatedIndex src/unittest/clustering_issues.cc:181: Failure Value of: rows.size() Actual: 0 Expected: static_cast<size_t>(1) Which is: 1 src/unittest/clustering_issues.cc:181: Failure Value of: rows.size() Actual: 0 Expected: static_cast<size_t>(1) Which is: 1 src/unittest/clustering_issues.cc:204: Failure Value of: issue_with_names.has() && issue_with_uuids.has() Actual: false Expected: true [ FAILED ] ClusteringIssues.OutdatedIndex (236 ms) ```
1.0
Intermittent failure in ClusteringIssues.OutdatedIndexes unit test - This happened twice for me on gantz on v1.16.x, but not every time: ``` [ RUN ] ClusteringIssues.OutdatedIndex src/unittest/clustering_issues.cc:181: Failure Value of: rows.size() Actual: 0 Expected: static_cast<size_t>(1) Which is: 1 src/unittest/clustering_issues.cc:181: Failure Value of: rows.size() Actual: 0 Expected: static_cast<size_t>(1) Which is: 1 src/unittest/clustering_issues.cc:204: Failure Value of: issue_with_names.has() && issue_with_uuids.has() Actual: false Expected: true [ FAILED ] ClusteringIssues.OutdatedIndex (236 ms) ```
test
intermittent failure in clusteringissues outdatedindexes unit test this happened twice for me on gantz on x but not every time clusteringissues outdatedindex src unittest clustering issues cc failure value of rows size actual expected static cast which is src unittest clustering issues cc failure value of rows size actual expected static cast which is src unittest clustering issues cc failure value of issue with names has issue with uuids has actual false expected true clusteringissues outdatedindex ms
1
93,162
8,402,343,918
IssuesEvent
2018-10-11 06:12:57
tendermint/tendermint
https://api.github.com/repos/tendermint/tendermint
closed
Fail to go test rpc/lib/rpc_test.go
help wanted test waiting for response
<!-- Please fill in as much of the template below as you can. Be ready for followup questions, and please respond in a timely manner. We might ask you to provide additional logs and data (tendermint & app). --> **Tendermint version** (use `tendermint version` or `git rev-parse --verify HEAD` if installed from source): $ git rev-parse --verify HEAD c1f6f6479d8ed1c9e5512f5698749bb9cb85f79a **ABCI app** (name for built-in, URL for self-written if it's publicly available): **Environment**: 3.10.0-693.el7.x86_64 - **OS** (e.g. from /etc/os-release): CentOS Linux 7 (Core) - **Install tools**: - **Others**: Go version: go version go1.9.4 linux/amd64 **What happened**: 1. go get -u github.com/tendermint/tendermint/ 2. dep ensure 3. cd rpc/lib/ 4. go test -run TestWSClientPingPong 5. --- FAIL: TestWSClientPingPong (0.00s) require.go:765: Error Trace: rpc_test.go:365 Error: Expected nil, but got: &errors.errorString{s:"Moved Permanently"} Test: TestWSClientPingPong FAIL exit status 1 FAIL github.com/tendermint/tendermint/rpc/lib 2.009s **What you expected to happen**: 1. Test PASS **Have you tried the latest version**: yes/no 1. Last Commit: c1f6f6479d8ed1c9e5512f5698749bb9cb85f79a **How to reproduce it** (as minimally and precisely as possible): 1. go test -run TestWSClientPingPong **Logs (paste a small part showing an error (< 10 lines) or link a pastebin, gist, etc. containing more of the log file)**: **Config (you can paste only the changes you've made)**: **node command runtime flags**: **`/dump_consensus_state` output for consensus bugs** **Anything else we need to know**:
1.0
Fail to go test rpc/lib/rpc_test.go - <!-- Please fill in as much of the template below as you can. Be ready for followup questions, and please respond in a timely manner. We might ask you to provide additional logs and data (tendermint & app). --> **Tendermint version** (use `tendermint version` or `git rev-parse --verify HEAD` if installed from source): $ git rev-parse --verify HEAD c1f6f6479d8ed1c9e5512f5698749bb9cb85f79a **ABCI app** (name for built-in, URL for self-written if it's publicly available): **Environment**: 3.10.0-693.el7.x86_64 - **OS** (e.g. from /etc/os-release): CentOS Linux 7 (Core) - **Install tools**: - **Others**: Go version: go version go1.9.4 linux/amd64 **What happened**: 1. go get -u github.com/tendermint/tendermint/ 2. dep ensure 3. cd rpc/lib/ 4. go test -run TestWSClientPingPong 5. --- FAIL: TestWSClientPingPong (0.00s) require.go:765: Error Trace: rpc_test.go:365 Error: Expected nil, but got: &errors.errorString{s:"Moved Permanently"} Test: TestWSClientPingPong FAIL exit status 1 FAIL github.com/tendermint/tendermint/rpc/lib 2.009s **What you expected to happen**: 1. Test PASS **Have you tried the latest version**: yes/no 1. Last Commit: c1f6f6479d8ed1c9e5512f5698749bb9cb85f79a **How to reproduce it** (as minimally and precisely as possible): 1. go test -run TestWSClientPingPong **Logs (paste a small part showing an error (< 10 lines) or link a pastebin, gist, etc. containing more of the log file)**: **Config (you can paste only the changes you've made)**: **node command runtime flags**: **`/dump_consensus_state` output for consensus bugs** **Anything else we need to know**:
test
fail to go test rpc lib rpc test go please fill in as much of the template below as you can be ready for followup questions and please respond in a timely manner we might ask you to provide additional logs and data tendermint app tendermint version use tendermint version or git rev parse verify head if installed from source git rev parse verify head abci app name for built in url for self written if it s publicly available environment os e g from etc os release centos linux core install tools others go version go version linux what happened go get u github com tendermint tendermint dep ensure cd rpc lib go test run testwsclientpingpong fail testwsclientpingpong require go error trace rpc test go error expected nil but got errors errorstring s moved permanently test testwsclientpingpong fail exit status fail github com tendermint tendermint rpc lib what you expected to happen test pass have you tried the latest version yes no last commit how to reproduce it as minimally and precisely as possible go test run testwsclientpingpong logs paste a small part showing an error lines or link a pastebin gist etc containing more of the log file config you can paste only the changes you ve made node command runtime flags dump consensus state output for consensus bugs anything else we need to know
1
157,742
12,389,599,938
IssuesEvent
2020-05-20 09:16:47
microsoft/AzureStorageExplorer
https://api.github.com/repos/microsoft/AzureStorageExplorer
opened
Display a blank line in one SAS attached blob container's context menu
:beetle: regression :gear: attach :gear: blobs :gear: sas 🧪 testing
**Storage Explorer Version:** 1.14.0-dev **Build:** [20200520.4](https://devdiv.visualstudio.com/DevDiv/_build/results?buildId=3742543&view=results) **Branch**: master **Platform/OS:** Windows 10/ Linux Ubuntu 16.04/ macOS Mojave **Architecture:** ia32/x64 **Regression From:** Previous release (1.13.0) **Steps to reproduce:** 1. Expand one non-ADLS Gen2 storage account -> Blob Containers. 2. Select one blob container -> Attach it via sas. 3. Check the attached blob container's context menu. **Expect Experience:** The context menu displays well. ![image](https://user-images.githubusercontent.com/41351993/82428520-352f2500-9abd-11ea-8a53-a822287f94ee.png) **Actual Experience:** Display a blank line in the context menu. ![image](https://user-images.githubusercontent.com/41351993/82428530-37917f00-9abd-11ea-943a-2db0582cc66c.png) **More Info:** Record: https://msit.microsoftstream.com/video/4e3aa1ff-0400-96d0-3a31-f1ea9a6b03b3
1.0
Display a blank line in one SAS attached blob container's context menu - **Storage Explorer Version:** 1.14.0-dev **Build:** [20200520.4](https://devdiv.visualstudio.com/DevDiv/_build/results?buildId=3742543&view=results) **Branch**: master **Platform/OS:** Windows 10/ Linux Ubuntu 16.04/ macOS Mojave **Architecture:** ia32/x64 **Regression From:** Previous release (1.13.0) **Steps to reproduce:** 1. Expand one non-ADLS Gen2 storage account -> Blob Containers. 2. Select one blob container -> Attach it via sas. 3. Check the attached blob container's context menu. **Expect Experience:** The context menu displays well. ![image](https://user-images.githubusercontent.com/41351993/82428520-352f2500-9abd-11ea-8a53-a822287f94ee.png) **Actual Experience:** Display a blank line in the context menu. ![image](https://user-images.githubusercontent.com/41351993/82428530-37917f00-9abd-11ea-943a-2db0582cc66c.png) **More Info:** Record: https://msit.microsoftstream.com/video/4e3aa1ff-0400-96d0-3a31-f1ea9a6b03b3
test
display a blank line in one sas attached blob container s context menu storage explorer version dev build branch master platform os windows linux ubuntu macos mojave architecture regression from previous release steps to reproduce expand one non adls storage account blob containers select one blob container attach it via sas check the attached blob container s context menu expect experience the context menu displays well actual experience display a blank line in the context menu more info record
1
219,414
16,829,742,145
IssuesEvent
2021-06-18 01:33:35
ElektraInitiative/libelektra
https://api.github.com/repos/ElektraInitiative/libelektra
closed
elektrad: improve management of kdb handles
documentation lang/go low priority stale
KDB handles are currently held by `elektrad` for 60 minutes of inactivity before being closed. Depending on the scenario this might have to be improved. One possibility is having a `MAX_SESSIONS` limit and removing the oldest sessions to mitigate e.g. denial of service. Other suggestions are welcome.
1.0
elektrad: improve management of kdb handles - KDB handles are currently held by `elektrad` for 60 minutes of inactivity before being closed. Depending on the scenario this might have to be improved. One possibility is having a `MAX_SESSIONS` limit and removing the oldest sessions to mitigate e.g. denial of service. Other suggestions are welcome.
non_test
elektrad improve management of kdb handles kdb handles are currently held by elektrad for minutes of inactivity before being closed depending on the scenario this might have to be improved one possibility is having a max sessions limit and removing the oldest sessions to mitigate e g denial of service other suggestions are welcome
0
216,221
16,748,469,358
IssuesEvent
2021-06-11 18:52:46
Quick-Event/quickbox
https://api.github.com/repos/Quick-Event/quickbox
closed
Štafety - export výsledků s mezičasy
1 done-testing feature request relays
Při štafetách zatím asi nelze snadno vyrobit výsledky s mězičasy. Přímo tisk mezičasů v programu není a export výsledků IOF nejspíše není správně, nebo ho nic neumí importovat, to jsem zatím nezjišťoval. ORIS import výsledků štafet nejspíše nepodporuje, takže jsem zkoušel importovat export výsledků do WinSplits, ale výsledkem bylo tam jen asi 7 závodníků v pár kategoriích. Zatím by bylo dobrým řešením, kdyby si export rozuměl alespoň s WinSplits, kam lze výsledky snadno nahrát. Nebo jestli někoho napadá nějaké lepší řešení (zatím bez ORISu), tak raďte. Díky, Vojta
1.0
Štafety - export výsledků s mezičasy - Při štafetách zatím asi nelze snadno vyrobit výsledky s mězičasy. Přímo tisk mezičasů v programu není a export výsledků IOF nejspíše není správně, nebo ho nic neumí importovat, to jsem zatím nezjišťoval. ORIS import výsledků štafet nejspíše nepodporuje, takže jsem zkoušel importovat export výsledků do WinSplits, ale výsledkem bylo tam jen asi 7 závodníků v pár kategoriích. Zatím by bylo dobrým řešením, kdyby si export rozuměl alespoň s WinSplits, kam lze výsledky snadno nahrát. Nebo jestli někoho napadá nějaké lepší řešení (zatím bez ORISu), tak raďte. Díky, Vojta
test
štafety export výsledků s mezičasy při štafetách zatím asi nelze snadno vyrobit výsledky s mězičasy přímo tisk mezičasů v programu není a export výsledků iof nejspíše není správně nebo ho nic neumí importovat to jsem zatím nezjišťoval oris import výsledků štafet nejspíše nepodporuje takže jsem zkoušel importovat export výsledků do winsplits ale výsledkem bylo tam jen asi závodníků v pár kategoriích zatím by bylo dobrým řešením kdyby si export rozuměl alespoň s winsplits kam lze výsledky snadno nahrát nebo jestli někoho napadá nějaké lepší řešení zatím bez orisu tak raďte díky vojta
1
57,800
14,219,789,633
IssuesEvent
2020-11-17 13:45:40
LalithK90/nandanaMotors
https://api.github.com/repos/LalithK90/nandanaMotors
opened
CVE-2019-14900 (Medium) detected in hibernate-core-5.4.10.Final.jar
security vulnerability
## CVE-2019-14900 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hibernate-core-5.4.10.Final.jar</b></p></summary> <p>Hibernate's core ORM functionality</p> <p>Library home page: <a href="http://hibernate.org/orm">http://hibernate.org/orm</a></p> <p>Path to dependency file: nandanaMotors/build.gradle</p> <p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.hibernate/hibernate-core/5.4.10.Final/365690f87b040dc5d22cc24a4daff76d1cffef23/hibernate-core-5.4.10.Final.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-data-jpa-2.2.4.RELEASE.jar (Root Library) - :x: **hibernate-core-5.4.10.Final.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/LalithK90/nandanaMotors/commit/9ebf8ba435ba9756d5a6f1ff78590adcf1ee8487">9ebf8ba435ba9756d5a6f1ff78590adcf1ee8487</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A flaw was found in Hibernate ORM in versions before 5.3.18, 5.4.18 and 5.5.0.Beta1. A SQL injection in the implementation of the JPA Criteria API can permit unsanitized literals when a literal is used in the SELECT or GROUP BY parts of the query. This flaw could allow an attacker to access unauthorized information or possibly conduct further attacks. <p>Publish Date: 2020-07-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14900>CVE-2019-14900</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-14900">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-14900</a></p> <p>Release Date: 2020-07-06</p> <p>Fix Resolution: org.hibernate:hibernate-core:5.4.18.Final</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-14900 (Medium) detected in hibernate-core-5.4.10.Final.jar - ## CVE-2019-14900 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hibernate-core-5.4.10.Final.jar</b></p></summary> <p>Hibernate's core ORM functionality</p> <p>Library home page: <a href="http://hibernate.org/orm">http://hibernate.org/orm</a></p> <p>Path to dependency file: nandanaMotors/build.gradle</p> <p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.hibernate/hibernate-core/5.4.10.Final/365690f87b040dc5d22cc24a4daff76d1cffef23/hibernate-core-5.4.10.Final.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-data-jpa-2.2.4.RELEASE.jar (Root Library) - :x: **hibernate-core-5.4.10.Final.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/LalithK90/nandanaMotors/commit/9ebf8ba435ba9756d5a6f1ff78590adcf1ee8487">9ebf8ba435ba9756d5a6f1ff78590adcf1ee8487</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A flaw was found in Hibernate ORM in versions before 5.3.18, 5.4.18 and 5.5.0.Beta1. A SQL injection in the implementation of the JPA Criteria API can permit unsanitized literals when a literal is used in the SELECT or GROUP BY parts of the query. This flaw could allow an attacker to access unauthorized information or possibly conduct further attacks. <p>Publish Date: 2020-07-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14900>CVE-2019-14900</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-14900">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-14900</a></p> <p>Release Date: 2020-07-06</p> <p>Fix Resolution: org.hibernate:hibernate-core:5.4.18.Final</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve medium detected in hibernate core final jar cve medium severity vulnerability vulnerable library hibernate core final jar hibernate s core orm functionality library home page a href path to dependency file nandanamotors build gradle path to vulnerable library home wss scanner gradle caches modules files org hibernate hibernate core final hibernate core final jar dependency hierarchy spring boot starter data jpa release jar root library x hibernate core final jar vulnerable library found in head commit a href found in base branch master vulnerability details a flaw was found in hibernate orm in versions before and a sql injection in the implementation of the jpa criteria api can permit unsanitized literals when a literal is used in the select or group by parts of the query this flaw could allow an attacker to access unauthorized information or possibly conduct further attacks publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org hibernate hibernate core final step up your open source security game with whitesource
0
40,038
5,267,113,556
IssuesEvent
2017-02-04 19:20:59
paperjs/paper.js
https://api.github.com/repos/paperjs/paper.js
closed
Optimize boolean operations when there are no crossings.
cat: boolean-operations status: needs-tests type: improvemnet
When there are no crossings, the result can already be known ahead of `tracePaths()`, probably leading to a massive speed-up: - intersect: return `null` - unite: return a compound path with both operands - subtract: return the first operand - exclude: same as unite. - divide: no change needed, since it redirects to the other methods.
1.0
Optimize boolean operations when there are no crossings. - When there are no crossings, the result can already be known ahead of `tracePaths()`, probably leading to a massive speed-up: - intersect: return `null` - unite: return a compound path with both operands - subtract: return the first operand - exclude: same as unite. - divide: no change needed, since it redirects to the other methods.
test
optimize boolean operations when there are no crossings when there are no crossings the result can already be known ahead of tracepaths probably leading to a massive speed up intersect return null unite return a compound path with both operands subtract return the first operand exclude same as unite divide no change needed since it redirects to the other methods
1
48,812
7,459,220,646
IssuesEvent
2018-03-30 14:24:49
nim-lang/Nim
https://api.github.com/repos/nim-lang/Nim
closed
Missing documentation for asyncfutures
Documentation
https://nim-lang.org/docs/asyncfutures.html I would provide a PR, but I am not sure what is there to do - isn't the documentation for the stdlib auto-generated from the modules themselves?
1.0
Missing documentation for asyncfutures - https://nim-lang.org/docs/asyncfutures.html I would provide a PR, but I am not sure what is there to do - isn't the documentation for the stdlib auto-generated from the modules themselves?
non_test
missing documentation for asyncfutures i would provide a pr but i am not sure what is there to do isn t the documentation for the stdlib auto generated from the modules themselves
0
319,650
9,747,685,857
IssuesEvent
2019-06-03 14:52:40
gamificalostudio/Tankerfield
https://api.github.com/repos/gamificalostudio/Tankerfield
closed
[BUG]: Boxes dissapear when on the margins of the camera
Frequency: Permanent Priority: Cosmetic fixed
**Bug description** Boxes dissapear when on the margins of the camera Probably because of the camera culling Object::frame should be bigger, to fit the whole box and its shadow. **Steps to reproduce** **Expected behaviour** **Actual behaviour** **Game version** **Screenshots** ![image](https://user-images.githubusercontent.com/36292291/57981959-cdc03e80-7a3e-11e9-979f-bf7536e9563b.png) Move a little. ![image](https://user-images.githubusercontent.com/36292291/57981963-d7e23d00-7a3e-11e9-884d-14fade8ea2fa.png) **Additional context**
1.0
[BUG]: Boxes dissapear when on the margins of the camera - **Bug description** Boxes dissapear when on the margins of the camera Probably because of the camera culling Object::frame should be bigger, to fit the whole box and its shadow. **Steps to reproduce** **Expected behaviour** **Actual behaviour** **Game version** **Screenshots** ![image](https://user-images.githubusercontent.com/36292291/57981959-cdc03e80-7a3e-11e9-979f-bf7536e9563b.png) Move a little. ![image](https://user-images.githubusercontent.com/36292291/57981963-d7e23d00-7a3e-11e9-884d-14fade8ea2fa.png) **Additional context**
non_test
boxes dissapear when on the margins of the camera bug description boxes dissapear when on the margins of the camera probably because of the camera culling object frame should be bigger to fit the whole box and its shadow steps to reproduce expected behaviour actual behaviour game version screenshots move a little additional context
0
428,066
12,402,276,225
IssuesEvent
2020-05-21 11:37:59
Broken-Gem-Studio/Broken-Engine
https://api.github.com/repos/Broken-Gem-Studio/Broken-Engine
closed
When the engine crashes some GameObjects are deactivated
Bug Low Priority Stale
## Bug Description Sometimes when the engine crahses and I reopen it the most of the scene has been deactivates, normally everything except the last GameObject I was editing ## Type of Bug Select the type of bug with and "x" ([x]) * [ ] Visual * [ ] Physics * [ ] Audio * [ ] Particles * [ ] Resource Management & Save/Load * [ ] Materials * [ ] Components * [X] Game Objects * [ ] UI/UX * [ ] Scripting * [ ] Other ## Severity Select the severity of bug affection and mark with "x" ([x]) - [ ] Crash - [ ] Game stopper/slower - [X] Cosmetic ## Reproduction Steps to reproduce the behavior: I don't know how to reproduce it, because it only happens sometimes and it's very random 1. 2. 3. 4. ## Frequency Select the frequency with which the bug appears and mark it "x" ([x]) * [ ] Always * [ ] Very Often * [X] Usually * [ ] Few Times * [ ] Few Times under specific conditions ## Conduct ### Expected result: I was expecting that when I reopened the engine the scene would load normally ### Actual result: The scene load with no problem but almost every gameobject was unactive ## Screenshots and Illustrations: ## Build - **Please specify the build:** ``0.5.3`` ## Observations and Additional Information
1.0
When the engine crashes some GameObjects are deactivated - ## Bug Description Sometimes when the engine crahses and I reopen it the most of the scene has been deactivates, normally everything except the last GameObject I was editing ## Type of Bug Select the type of bug with and "x" ([x]) * [ ] Visual * [ ] Physics * [ ] Audio * [ ] Particles * [ ] Resource Management & Save/Load * [ ] Materials * [ ] Components * [X] Game Objects * [ ] UI/UX * [ ] Scripting * [ ] Other ## Severity Select the severity of bug affection and mark with "x" ([x]) - [ ] Crash - [ ] Game stopper/slower - [X] Cosmetic ## Reproduction Steps to reproduce the behavior: I don't know how to reproduce it, because it only happens sometimes and it's very random 1. 2. 3. 4. ## Frequency Select the frequency with which the bug appears and mark it "x" ([x]) * [ ] Always * [ ] Very Often * [X] Usually * [ ] Few Times * [ ] Few Times under specific conditions ## Conduct ### Expected result: I was expecting that when I reopened the engine the scene would load normally ### Actual result: The scene load with no problem but almost every gameobject was unactive ## Screenshots and Illustrations: ## Build - **Please specify the build:** ``0.5.3`` ## Observations and Additional Information
non_test
when the engine crashes some gameobjects are deactivated bug description sometimes when the engine crahses and i reopen it the most of the scene has been deactivates normally everything except the last gameobject i was editing type of bug select the type of bug with and x visual physics audio particles resource management save load materials components game objects ui ux scripting other severity select the severity of bug affection and mark with x crash game stopper slower cosmetic reproduction steps to reproduce the behavior i don t know how to reproduce it because it only happens sometimes and it s very random frequency select the frequency with which the bug appears and mark it x always very often usually few times few times under specific conditions conduct expected result i was expecting that when i reopened the engine the scene would load normally actual result the scene load with no problem but almost every gameobject was unactive screenshots and illustrations build please specify the build observations and additional information
0
487,958
14,072,553,217
IssuesEvent
2020-11-04 02:10:52
microsoft/AdaptiveCards
https://api.github.com/repos/microsoft/AdaptiveCards
closed
[Android][Accessibility] [On input validation failure focus does not go to the first invalid input field when the first invalid field is date/time/compact choiceset]
AdaptiveCards v20.10.1 Area-Inconsistency Bug Msft-TeamsMobile MsftTeams-Integration Priority-Now Status-Fixed
# Platform What platform is your issue or question related to? (Delete other platforms). - [ ] Android # Author or host Microsoft Teams # Version of SDK 2.3.0 # Details On input validation failure focus does not go to the first invalid input field when the first invalid input field is date/time/compact choiceset.
1.0
[Android][Accessibility] [On input validation failure focus does not go to the first invalid input field when the first invalid field is date/time/compact choiceset] - # Platform What platform is your issue or question related to? (Delete other platforms). - [ ] Android # Author or host Microsoft Teams # Version of SDK 2.3.0 # Details On input validation failure focus does not go to the first invalid input field when the first invalid input field is date/time/compact choiceset.
non_test
platform what platform is your issue or question related to delete other platforms android author or host microsoft teams version of sdk details on input validation failure focus does not go to the first invalid input field when the first invalid input field is date time compact choiceset
0
185,498
6,722,440,655
IssuesEvent
2017-10-16 15:08:48
JeffersonLab/hallc_replay
https://api.github.com/repos/JeffersonLab/hallc_replay
opened
Broken CONFIG Files
Medium Priority
The directory restructuring of #230 requires that all the files in CONFIG be updated to point to the corresponding macro(s) now residing in the utilities (UTIL) directory.
1.0
Broken CONFIG Files - The directory restructuring of #230 requires that all the files in CONFIG be updated to point to the corresponding macro(s) now residing in the utilities (UTIL) directory.
non_test
broken config files the directory restructuring of requires that all the files in config be updated to point to the corresponding macro s now residing in the utilities util directory
0
24,232
23,527,322,490
IssuesEvent
2022-08-19 12:12:20
Ericsson/codechecker
https://api.github.com/repos/Ericsson/codechecker
opened
The actual error in a report message should be at the diagnostic message, not at the last bug path event.
GUI :art: usability :+1: refactoring :angry: ➡️ :slightly_smiling_face:
The diagnostic message should be used when indicating an error, as this report element indicates an error in the code semantically. Right now the last bug path event is treated as the error because Clang static analyzer emits reports in format where diagnostic message is duplicated in the last position, this is a clangsa quirk and Codechecker should not rely on this behavior. For Cppcheck to work correctly the bug path event list is needed to be extended with the diagnostic message, which is a redundant step, as the diagnostic message also has a location. The position of an error is critical from the point of report suppression, so the impact of this modification should be carefully examined, but a clearer error presentation would be desirable in the future. This is a follow up to #3680 Also a follow up to #3721
True
The actual error in a report message should be at the diagnostic message, not at the last bug path event. - The diagnostic message should be used when indicating an error, as this report element indicates an error in the code semantically. Right now the last bug path event is treated as the error because Clang static analyzer emits reports in format where diagnostic message is duplicated in the last position, this is a clangsa quirk and Codechecker should not rely on this behavior. For Cppcheck to work correctly the bug path event list is needed to be extended with the diagnostic message, which is a redundant step, as the diagnostic message also has a location. The position of an error is critical from the point of report suppression, so the impact of this modification should be carefully examined, but a clearer error presentation would be desirable in the future. This is a follow up to #3680 Also a follow up to #3721
non_test
the actual error in a report message should be at the diagnostic message not at the last bug path event the diagnostic message should be used when indicating an error as this report element indicates an error in the code semantically right now the last bug path event is treated as the error because clang static analyzer emits reports in format where diagnostic message is duplicated in the last position this is a clangsa quirk and codechecker should not rely on this behavior for cppcheck to work correctly the bug path event list is needed to be extended with the diagnostic message which is a redundant step as the diagnostic message also has a location the position of an error is critical from the point of report suppression so the impact of this modification should be carefully examined but a clearer error presentation would be desirable in the future this is a follow up to also a follow up to
0
142,938
19,142,255,923
IssuesEvent
2021-12-02 01:05:02
BrianMcDonaldWS/genie
https://api.github.com/repos/BrianMcDonaldWS/genie
opened
CVE-2021-37137 (High) detected in netty-codec-4.1.43.Final.jar
security vulnerability
## CVE-2021-37137 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>netty-codec-4.1.43.Final.jar</b></p></summary> <p>Netty is an asynchronous event-driven network application framework for rapid development of maintainable high performance protocol servers and clients.</p> <p>Library home page: <a href="https://netty.io/">https://netty.io/</a></p> <p>Path to dependency file: genie/genie-agent/build.gradle</p> <p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/io.netty/netty-codec/4.1.43.Final/44731a5739c40a8434c43e1303e9196f23da67fa/netty-codec-4.1.43.Final.jar,/root/.gradle/caches/modules-2/files-2.1/io.netty/netty-codec/4.1.43.Final/44731a5739c40a8434c43e1303e9196f23da67fa/netty-codec-4.1.43.Final.jar</p> <p> Dependency Hierarchy: - grpc-netty-1.20.0.jar (Root Library) - netty-codec-http2-4.1.43.Final.jar - netty-codec-http-4.1.43.Final.jar - :x: **netty-codec-4.1.43.Final.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The Snappy frame decoder function doesn't restrict the chunk length which may lead to excessive memory usage. Beside this it also may buffer reserved skippable chunks until the whole chunk was received which may lead to excessive memory usage as well. This vulnerability can be triggered by supplying malicious input that decompresses to a very big size (via a network stream or a file) or by sending a huge skippable chunk. <p>Publish Date: 2021-10-19 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37137>CVE-2021-37137</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-9vjp-v76f-g363">https://github.com/advisories/GHSA-9vjp-v76f-g363</a></p> <p>Release Date: 2021-10-19</p> <p>Fix Resolution: io.netty:netty-codec:4.1.68.Final;io.netty:netty-all:4.1.68.Final</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"io.netty","packageName":"netty-codec","packageVersion":"4.1.43.Final","packageFilePaths":["/genie-agent/build.gradle"],"isTransitiveDependency":true,"dependencyTree":"io.grpc:grpc-netty:1.20.0;io.netty:netty-codec-http2:4.1.43.Final;io.netty:netty-codec-http:4.1.43.Final;io.netty:netty-codec:4.1.43.Final","isMinimumFixVersionAvailable":true,"minimumFixVersion":"io.netty:netty-codec:4.1.68.Final;io.netty:netty-all:4.1.68.Final","isBinary":false}],"baseBranches":[],"vulnerabilityIdentifier":"CVE-2021-37137","vulnerabilityDetails":"The Snappy frame decoder function doesn\u0027t restrict the chunk length which may lead to excessive memory usage. Beside this it also may buffer reserved skippable chunks until the whole chunk was received which may lead to excessive memory usage as well. This vulnerability can be triggered by supplying malicious input that decompresses to a very big size (via a network stream or a file) or by sending a huge skippable chunk.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37137","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2021-37137 (High) detected in netty-codec-4.1.43.Final.jar - ## CVE-2021-37137 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>netty-codec-4.1.43.Final.jar</b></p></summary> <p>Netty is an asynchronous event-driven network application framework for rapid development of maintainable high performance protocol servers and clients.</p> <p>Library home page: <a href="https://netty.io/">https://netty.io/</a></p> <p>Path to dependency file: genie/genie-agent/build.gradle</p> <p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/io.netty/netty-codec/4.1.43.Final/44731a5739c40a8434c43e1303e9196f23da67fa/netty-codec-4.1.43.Final.jar,/root/.gradle/caches/modules-2/files-2.1/io.netty/netty-codec/4.1.43.Final/44731a5739c40a8434c43e1303e9196f23da67fa/netty-codec-4.1.43.Final.jar</p> <p> Dependency Hierarchy: - grpc-netty-1.20.0.jar (Root Library) - netty-codec-http2-4.1.43.Final.jar - netty-codec-http-4.1.43.Final.jar - :x: **netty-codec-4.1.43.Final.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The Snappy frame decoder function doesn't restrict the chunk length which may lead to excessive memory usage. Beside this it also may buffer reserved skippable chunks until the whole chunk was received which may lead to excessive memory usage as well. This vulnerability can be triggered by supplying malicious input that decompresses to a very big size (via a network stream or a file) or by sending a huge skippable chunk. <p>Publish Date: 2021-10-19 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37137>CVE-2021-37137</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-9vjp-v76f-g363">https://github.com/advisories/GHSA-9vjp-v76f-g363</a></p> <p>Release Date: 2021-10-19</p> <p>Fix Resolution: io.netty:netty-codec:4.1.68.Final;io.netty:netty-all:4.1.68.Final</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"io.netty","packageName":"netty-codec","packageVersion":"4.1.43.Final","packageFilePaths":["/genie-agent/build.gradle"],"isTransitiveDependency":true,"dependencyTree":"io.grpc:grpc-netty:1.20.0;io.netty:netty-codec-http2:4.1.43.Final;io.netty:netty-codec-http:4.1.43.Final;io.netty:netty-codec:4.1.43.Final","isMinimumFixVersionAvailable":true,"minimumFixVersion":"io.netty:netty-codec:4.1.68.Final;io.netty:netty-all:4.1.68.Final","isBinary":false}],"baseBranches":[],"vulnerabilityIdentifier":"CVE-2021-37137","vulnerabilityDetails":"The Snappy frame decoder function doesn\u0027t restrict the chunk length which may lead to excessive memory usage. Beside this it also may buffer reserved skippable chunks until the whole chunk was received which may lead to excessive memory usage as well. This vulnerability can be triggered by supplying malicious input that decompresses to a very big size (via a network stream or a file) or by sending a huge skippable chunk.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37137","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_test
cve high detected in netty codec final jar cve high severity vulnerability vulnerable library netty codec final jar netty is an asynchronous event driven network application framework for rapid development of maintainable high performance protocol servers and clients library home page a href path to dependency file genie genie agent build gradle path to vulnerable library root gradle caches modules files io netty netty codec final netty codec final jar root gradle caches modules files io netty netty codec final netty codec final jar dependency hierarchy grpc netty jar root library netty codec final jar netty codec http final jar x netty codec final jar vulnerable library vulnerability details the snappy frame decoder function doesn t restrict the chunk length which may lead to excessive memory usage beside this it also may buffer reserved skippable chunks until the whole chunk was received which may lead to excessive memory usage as well this vulnerability can be triggered by supplying malicious input that decompresses to a very big size via a network stream or a file or by sending a huge skippable chunk publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution io netty netty codec final io netty netty all final isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree io grpc grpc netty io netty netty codec final io netty netty codec http final io netty netty codec final isminimumfixversionavailable true minimumfixversion io netty netty codec final io netty netty all final isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails the snappy frame decoder function doesn restrict the chunk length which may lead to excessive memory usage beside this it also may buffer reserved skippable chunks until the whole chunk was received which may lead to excessive memory usage as well this vulnerability can be triggered by supplying malicious input that decompresses to a very big size via a network stream or a file or by sending a huge skippable chunk vulnerabilityurl
0
278,249
21,058,231,927
IssuesEvent
2022-04-01 06:54:14
9teMare/ped
https://api.github.com/repos/9teMare/ped
opened
Mismatch of DG and ResultDisplay description regarding the `within` command
type.DocumentationBug severity.Medium
![image.png](https://raw.githubusercontent.com/9teMare/ped/main/files/b2816f1e-76c1-4338-b4a6-02bb875d919b.png) ![image.png](https://raw.githubusercontent.com/9teMare/ped/main/files/ce7a1cff-9e68-4ded-b37e-1e2b9ce6021e.png) `Days` can actually be `0`, make sure to update UG accordingly. <!--session: 1648792343681-a11b53de-f6cf-4799-94ca-28879ffefa05--> <!--Version: Web v3.4.2-->
1.0
Mismatch of DG and ResultDisplay description regarding the `within` command - ![image.png](https://raw.githubusercontent.com/9teMare/ped/main/files/b2816f1e-76c1-4338-b4a6-02bb875d919b.png) ![image.png](https://raw.githubusercontent.com/9teMare/ped/main/files/ce7a1cff-9e68-4ded-b37e-1e2b9ce6021e.png) `Days` can actually be `0`, make sure to update UG accordingly. <!--session: 1648792343681-a11b53de-f6cf-4799-94ca-28879ffefa05--> <!--Version: Web v3.4.2-->
non_test
mismatch of dg and resultdisplay description regarding the within command days can actually be make sure to update ug accordingly
0
697,888
23,957,509,651
IssuesEvent
2022-09-12 16:04:01
o3de/o3de
https://api.github.com/repos/o3de/o3de
closed
Feature Request: Constexpr AZ::Uuid Updates
sig/core triage/accepted priority/minor
**Is your feature request related to a problem? Please describe.** Through the O3DE codebase, we use the `AZ::Uuid` class for mapping C++ types to an identifier value for ad-hoc static reflection support. Every class reflected to the the SerializeContext requires a [TypeId](https://github.com/o3de/o3de/blob/development/Code/Framework/AzCore/AzCore/Serialization/SerializeContext.h#L1916-L1917) to be associated with it. Currently all of those Typeids are hardcoded values that are initialized into an `AZ::Uuid` structure at static initialization time. A small runtime performance benefit is to initialize those `AZ::Uuid` at compile time, removing the need to initialize Uuids at static init time. Furthermore if the Uuid class supports constexpr, it can be used as arguments to non-type template parameters. This would open up the possibility of writing a deprecated converter for a deleted class by just specializing a template with an `auto` parameter. Ex. ``` template <auto TypeId> bool DeprecationConverter(AZ::SerializeContext& sc, AZ::SerializeContext::DataElement& rootElement); // ... // Specialization for class that has been deleted from the code base template <> bool DeprecationConverter<AZ::Uuid("{C0F1AFAD-5CB3-450E-B0F5-ADB5D46B0E22}")>(AZ::SerializeContext& sc, AZ::SerializeContext::DataElement& rootElement); ``` **Describe the solution you'd like** Update the AZ::Uuid struct constructors, Create\* functions, ToStrings and operator+ to be constexpr in [Uuid.h](https://github.com/o3de/o3de/blob/development/Code/Framework/AzCore/AzCore/Math/Uuid.h#L49-L76) **Describe alternatives you've considered** None **Additional context** Add any other context or screenshots about the feature request here.
1.0
Feature Request: Constexpr AZ::Uuid Updates - **Is your feature request related to a problem? Please describe.** Through the O3DE codebase, we use the `AZ::Uuid` class for mapping C++ types to an identifier value for ad-hoc static reflection support. Every class reflected to the the SerializeContext requires a [TypeId](https://github.com/o3de/o3de/blob/development/Code/Framework/AzCore/AzCore/Serialization/SerializeContext.h#L1916-L1917) to be associated with it. Currently all of those Typeids are hardcoded values that are initialized into an `AZ::Uuid` structure at static initialization time. A small runtime performance benefit is to initialize those `AZ::Uuid` at compile time, removing the need to initialize Uuids at static init time. Furthermore if the Uuid class supports constexpr, it can be used as arguments to non-type template parameters. This would open up the possibility of writing a deprecated converter for a deleted class by just specializing a template with an `auto` parameter. Ex. ``` template <auto TypeId> bool DeprecationConverter(AZ::SerializeContext& sc, AZ::SerializeContext::DataElement& rootElement); // ... // Specialization for class that has been deleted from the code base template <> bool DeprecationConverter<AZ::Uuid("{C0F1AFAD-5CB3-450E-B0F5-ADB5D46B0E22}")>(AZ::SerializeContext& sc, AZ::SerializeContext::DataElement& rootElement); ``` **Describe the solution you'd like** Update the AZ::Uuid struct constructors, Create\* functions, ToStrings and operator+ to be constexpr in [Uuid.h](https://github.com/o3de/o3de/blob/development/Code/Framework/AzCore/AzCore/Math/Uuid.h#L49-L76) **Describe alternatives you've considered** None **Additional context** Add any other context or screenshots about the feature request here.
non_test
feature request constexpr az uuid updates is your feature request related to a problem please describe through the codebase we use the az uuid class for mapping c types to an identifier value for ad hoc static reflection support every class reflected to the the serializecontext requires a to be associated with it currently all of those typeids are hardcoded values that are initialized into an az uuid structure at static initialization time a small runtime performance benefit is to initialize those az uuid at compile time removing the need to initialize uuids at static init time furthermore if the uuid class supports constexpr it can be used as arguments to non type template parameters this would open up the possibility of writing a deprecated converter for a deleted class by just specializing a template with an auto parameter ex template bool deprecationconverter az serializecontext sc az serializecontext dataelement rootelement specialization for class that has been deleted from the code base template bool deprecationconverter az serializecontext sc az serializecontext dataelement rootelement describe the solution you d like update the az uuid struct constructors create functions tostrings and operator to be constexpr in describe alternatives you ve considered none additional context add any other context or screenshots about the feature request here
0
314,031
26,971,287,842
IssuesEvent
2023-02-09 05:14:30
ToolJet/ToolJet
https://api.github.com/repos/ToolJet/ToolJet
closed
Add data-cy to fix table regression
test cypress
### Specify the kind of test <!-- Provide the kind of test --> (Cypress integration-test) ### Describe the test <!-- Provide a clear description of the test --> Not a test, Adding data-cy only
1.0
Add data-cy to fix table regression - ### Specify the kind of test <!-- Provide the kind of test --> (Cypress integration-test) ### Describe the test <!-- Provide a clear description of the test --> Not a test, Adding data-cy only
test
add data cy to fix table regression specify the kind of test provide the kind of test cypress integration test describe the test provide a clear description of the test not a test adding data cy only
1
225,599
17,867,345,974
IssuesEvent
2021-09-06 11:06:43
milvus-io/milvus
https://api.github.com/repos/milvus-io/milvus
closed
Nightly ci Pod crashed due to ephemeral storage limit
kind/improvement area/test
<!-- Please state your issue using the following template and, most importantly, in English. --> #### Steps/Code to reproduce: Run the nightly ci with the following code: ``` cd tests/scripts MILVUS_CLUSTER_ENABLED=true ./e2e-k8s.sh --test-extra-arg --tags L0 L1 L2 --test-timeout 18000 ``` #### Expected result: Expected the nightly ci pod runs normally until all the tests finished. #### Actual results: ``` [2021-09-06T03:45:41.747Z] testcases/entity/test_search.py::TestSearchBase::test_search_ip_flat[get_simple_index7-10-1100] Cannot contact milvus-e2e-test-kind-nightly-546tr-cmn0v: hudson.remoting.RequestAbortedException: java.nio.channels.ClosedChannelException [2021-09-06T03:50:41.763Z] Could not connect to milvus-e2e-test-kind-nightly-546tr-cmn0v to send interrupt signal to process ``` #### Environment: - Milvus version(e.g. v2.0.0-RC2 or 8b23a93): - Deployment mode(standalone or cluster): - SDK version(e.g. pymilvus v2.0.0rc2): - OS(Ubuntu or CentOS): - CPU/Memory: - GPU: - Others: #### Configuration file: #### Additional context:
1.0
Nightly ci Pod crashed due to ephemeral storage limit - <!-- Please state your issue using the following template and, most importantly, in English. --> #### Steps/Code to reproduce: Run the nightly ci with the following code: ``` cd tests/scripts MILVUS_CLUSTER_ENABLED=true ./e2e-k8s.sh --test-extra-arg --tags L0 L1 L2 --test-timeout 18000 ``` #### Expected result: Expected the nightly ci pod runs normally until all the tests finished. #### Actual results: ``` [2021-09-06T03:45:41.747Z] testcases/entity/test_search.py::TestSearchBase::test_search_ip_flat[get_simple_index7-10-1100] Cannot contact milvus-e2e-test-kind-nightly-546tr-cmn0v: hudson.remoting.RequestAbortedException: java.nio.channels.ClosedChannelException [2021-09-06T03:50:41.763Z] Could not connect to milvus-e2e-test-kind-nightly-546tr-cmn0v to send interrupt signal to process ``` #### Environment: - Milvus version(e.g. v2.0.0-RC2 or 8b23a93): - Deployment mode(standalone or cluster): - SDK version(e.g. pymilvus v2.0.0rc2): - OS(Ubuntu or CentOS): - CPU/Memory: - GPU: - Others: #### Configuration file: #### Additional context:
test
nightly ci pod crashed due to ephemeral storage limit steps code to reproduce run the nightly ci with the following code cd tests scripts milvus cluster enabled true sh test extra arg tags test timeout expected result expected the nightly ci pod runs normally until all the tests finished actual results testcases entity test search py testsearchbase test search ip flat cannot contact milvus test kind nightly hudson remoting requestabortedexception java nio channels closedchannelexception could not connect to milvus test kind nightly to send interrupt signal to process environment milvus version e g or deployment mode standalone or cluster sdk version e g pymilvus os ubuntu or centos cpu memory gpu others configuration file additional context
1
338,833
30,323,084,369
IssuesEvent
2023-07-10 20:53:41
Bears-R-Us/arkouda
https://api.github.com/repos/Bears-R-Us/arkouda
closed
`setops_test.py` Conversion to new test framework
In Progress Testing
Part of #2499 Add `setops_test.py` to the new testing framework and make any configuration updates necessary.
1.0
`setops_test.py` Conversion to new test framework - Part of #2499 Add `setops_test.py` to the new testing framework and make any configuration updates necessary.
test
setops test py conversion to new test framework part of add setops test py to the new testing framework and make any configuration updates necessary
1