Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
4
112
repo_url
stringlengths
33
141
action
stringclasses
3 values
title
stringlengths
1
1.02k
labels
stringlengths
4
1.54k
body
stringlengths
1
262k
index
stringclasses
17 values
text_combine
stringlengths
95
262k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
112,423
9,566,938,083
IssuesEvent
2019-05-06 00:24:29
zulip/zulip
https://api.github.com/repos/zulip/zulip
opened
Fix broken LDAP test when running tests in reverse order
area: authentication area: testing-coverage bug priority: high
We had to disabled one of our LDAP tests in https://github.com/zulip/zulip/commit/b45a7a828ec914d64694db20c44ef7a36e1bfe1b, because it was broken; we should fix it so we can re-enable the test.
1.0
Fix broken LDAP test when running tests in reverse order - We had to disabled one of our LDAP tests in https://github.com/zulip/zulip/commit/b45a7a828ec914d64694db20c44ef7a36e1bfe1b, because it was broken; we should fix it so we can re-enable the test.
test
fix broken ldap test when running tests in reverse order we had to disabled one of our ldap tests in because it was broken we should fix it so we can re enable the test
1
620,489
19,563,402,607
IssuesEvent
2022-01-03 19:38:24
bounswe/2021SpringGroup1
https://api.github.com/repos/bounswe/2021SpringGroup1
closed
Converting all Backend functions to Json-LD format
Type: Enhancement Priority: Medium Platform: Backend
Previously, necessary changes were made to return it in Json-LD format for Community functions. Now this format should be added to other functions as well. Here, relatively static fields such as "@context", "@type" should be added first. If there is no problem here, necessary changes should be made on the database and the field names in the Models.
1.0
Converting all Backend functions to Json-LD format - Previously, necessary changes were made to return it in Json-LD format for Community functions. Now this format should be added to other functions as well. Here, relatively static fields such as "@context", "@type" should be added first. If there is no problem here, necessary changes should be made on the database and the field names in the Models.
non_test
converting all backend functions to json ld format previously necessary changes were made to return it in json ld format for community functions now this format should be added to other functions as well here relatively static fields such as context type should be added first if there is no problem here necessary changes should be made on the database and the field names in the models
0
750,631
26,208,912,167
IssuesEvent
2023-01-04 03:15:58
open-telemetry/opentelemetry-collector
https://api.github.com/repos/open-telemetry/opentelemetry-collector
closed
Retry link checking on network failures
enhancement Stale priority:p3 release:after-ga area:miscellaneous
The CI currently checks for links by going over the network to check if the destination exists. The problem is that the network isn't reliable ([fallacy 1](https://en.wikipedia.org/wiki/Fallacies_of_distributed_computing)). Failure: ``` [✖] https://godoc.org/google.golang.org/grpc#WithInsecure → Status: 0 Error: ESOCKETTIMEDOUT ``` https://app.circleci.com/pipelines/github/open-telemetry/opentelemetry-collector/4166/workflows/f4d8eb48-9c89-4032-b08e-b7fbba3da81f/jobs/47015
1.0
Retry link checking on network failures - The CI currently checks for links by going over the network to check if the destination exists. The problem is that the network isn't reliable ([fallacy 1](https://en.wikipedia.org/wiki/Fallacies_of_distributed_computing)). Failure: ``` [✖] https://godoc.org/google.golang.org/grpc#WithInsecure → Status: 0 Error: ESOCKETTIMEDOUT ``` https://app.circleci.com/pipelines/github/open-telemetry/opentelemetry-collector/4166/workflows/f4d8eb48-9c89-4032-b08e-b7fbba3da81f/jobs/47015
non_test
retry link checking on network failures the ci currently checks for links by going over the network to check if the destination exists the problem is that the network isn t reliable failure → status error esockettimedout
0
90,332
18,108,697,851
IssuesEvent
2021-09-22 22:50:04
aiekick/ImGuiFileDialog
https://api.github.com/repos/aiekick/ImGuiFileDialog
reopened
Windows OS can't display Chinese filename, also can't select those file
bug unicode
OS: Windows 10 Chinese version ImGui 1.80 ImFileDialog git master Can NOT display Chinese file name(I already added Chinese fonts, and can show Chinese char will on other interface) Can NOT select Chinese file name or folder name.
1.0
Windows OS can't display Chinese filename, also can't select those file - OS: Windows 10 Chinese version ImGui 1.80 ImFileDialog git master Can NOT display Chinese file name(I already added Chinese fonts, and can show Chinese char will on other interface) Can NOT select Chinese file name or folder name.
non_test
windows os can t display chinese filename also can t select those file os windows chinese version imgui imfiledialog git master can not display chinese file name i already added chinese fonts and can show chinese char will on other interface can not select chinese file name or folder name
0
114,263
9,694,704,304
IssuesEvent
2019-05-24 19:47:07
eclipse/openj9
https://api.github.com/repos/eclipse/openj9
closed
Test quick start instructions do not work for Java 8
comp:test
Looking at the Java 8 build page, https://www.eclipse.org/openj9/oj9_build.html, it refers to the test quick start guide https://github.com/eclipse/openj9/blob/master/test/README.md - These instructions do not default to Java 8, and do not specify the JAVA_VERSION. Attempting to follow the instructions after finding them from the Java 8 build page does not work, the "make compile" step fails. - as Java 9 is out of support, the tests should default to something else. https://github.com/eclipse/openj9/blob/master/test/docs/OpenJ9TestUserGuide.md shows the default as SE90. - after setting JAVA_VERSION=SE80, I reran the configure step and make compile and got an error because I had set JAVA_BIN to the j2sdk-image/bin directory and the compile was looking for j2sdk-image/bin/../../bin/javac, which doesn't work.
1.0
Test quick start instructions do not work for Java 8 - Looking at the Java 8 build page, https://www.eclipse.org/openj9/oj9_build.html, it refers to the test quick start guide https://github.com/eclipse/openj9/blob/master/test/README.md - These instructions do not default to Java 8, and do not specify the JAVA_VERSION. Attempting to follow the instructions after finding them from the Java 8 build page does not work, the "make compile" step fails. - as Java 9 is out of support, the tests should default to something else. https://github.com/eclipse/openj9/blob/master/test/docs/OpenJ9TestUserGuide.md shows the default as SE90. - after setting JAVA_VERSION=SE80, I reran the configure step and make compile and got an error because I had set JAVA_BIN to the j2sdk-image/bin directory and the compile was looking for j2sdk-image/bin/../../bin/javac, which doesn't work.
test
test quick start instructions do not work for java looking at the java build page it refers to the test quick start guide these instructions do not default to java and do not specify the java version attempting to follow the instructions after finding them from the java build page does not work the make compile step fails as java is out of support the tests should default to something else shows the default as after setting java version i reran the configure step and make compile and got an error because i had set java bin to the image bin directory and the compile was looking for image bin bin javac which doesn t work
1
280,222
24,285,141,456
IssuesEvent
2022-09-28 21:14:58
Northeastern-Electric-Racing/FinishLine
https://api.github.com/repos/Northeastern-Electric-Racing/FinishLine
opened
Backend - Test addProposedSolution (part 2)
back-end straightforward testing
### Description Write unit tests for some of the cases of the addProposedSolution endpoint. To work on this ticket, you will need to read through the endpoint to learn how it works. ### Acceptance Criteria The following cases should be covered: * the associated scope cr does not exist * it works as intended ### Proposed Solution Check out this epic https://github.com/Northeastern-Electric-Racing/FinishLine/issues/238 for resources and tips on how to write these tests. ### Mocks _No response_
1.0
Backend - Test addProposedSolution (part 2) - ### Description Write unit tests for some of the cases of the addProposedSolution endpoint. To work on this ticket, you will need to read through the endpoint to learn how it works. ### Acceptance Criteria The following cases should be covered: * the associated scope cr does not exist * it works as intended ### Proposed Solution Check out this epic https://github.com/Northeastern-Electric-Racing/FinishLine/issues/238 for resources and tips on how to write these tests. ### Mocks _No response_
test
backend test addproposedsolution part description write unit tests for some of the cases of the addproposedsolution endpoint to work on this ticket you will need to read through the endpoint to learn how it works acceptance criteria the following cases should be covered the associated scope cr does not exist it works as intended proposed solution check out this epic for resources and tips on how to write these tests mocks no response
1
220,286
17,185,569,807
IssuesEvent
2021-07-16 00:58:20
elastic/kibana
https://api.github.com/repos/elastic/kibana
opened
[test-failed]: Chrome X-Pack UI Functional Tests1.x-pack/test/functional/apps/monitoring/elasticsearch/index_detail_mb·js - Monitoring app Elasticsearch index detail mb Active Indices "before all" hook for "should have an index summary with green status index with full shard allocation"
failed-test test-cloud
**Version: 7.14.0** **Class: Chrome X-Pack UI Functional Tests1.x-pack/test/functional/apps/monitoring/elasticsearch/index_detail_mb·js** **Stack Trace:** ``` Error: retry.try timeout: TimeoutError: Waiting for element to be located By(css selector, [data-test-subj="alerts-modal-button"]) Wait timed out after 10044ms at /var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/node_modules/selenium-webdriver/lib/webdriver.js:842:17 at runMicrotasks (<anonymous>) at processTicksAndRejections (internal/process/task_queues.js:95:5) at onFailure (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/common/services/retry/retry_for_success.ts:17:9) at retryForSuccess (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/common/services/retry/retry_for_success.ts:57:13) at RetryService.try (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/common/services/retry/retry.ts:31:12) at Proxy.clickByCssSelector (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/services/common/find.ts:360:5) at TestSubjects.click (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/services/common/test_subjects.ts:105:5) at Context.<anonymous> (test/functional/apps/monitoring/elasticsearch/index_detail_mb.js:35:9) at Object.apply (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/node_modules/@kbn/test/target_node/functional_test_runner/lib/mocha/wrap_function.js:87:16) ``` **Other test failures:** _Test Report: https://internal-ci.elastic.co/view/Stack%20Tests/job/elastic+estf-cloud-kibana-tests/2041/testReport/_
2.0
[test-failed]: Chrome X-Pack UI Functional Tests1.x-pack/test/functional/apps/monitoring/elasticsearch/index_detail_mb·js - Monitoring app Elasticsearch index detail mb Active Indices "before all" hook for "should have an index summary with green status index with full shard allocation" - **Version: 7.14.0** **Class: Chrome X-Pack UI Functional Tests1.x-pack/test/functional/apps/monitoring/elasticsearch/index_detail_mb·js** **Stack Trace:** ``` Error: retry.try timeout: TimeoutError: Waiting for element to be located By(css selector, [data-test-subj="alerts-modal-button"]) Wait timed out after 10044ms at /var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/node_modules/selenium-webdriver/lib/webdriver.js:842:17 at runMicrotasks (<anonymous>) at processTicksAndRejections (internal/process/task_queues.js:95:5) at onFailure (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/common/services/retry/retry_for_success.ts:17:9) at retryForSuccess (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/common/services/retry/retry_for_success.ts:57:13) at RetryService.try (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/common/services/retry/retry.ts:31:12) at Proxy.clickByCssSelector (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/services/common/find.ts:360:5) at TestSubjects.click (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/test/functional/services/common/test_subjects.ts:105:5) at Context.<anonymous> (test/functional/apps/monitoring/elasticsearch/index_detail_mb.js:35:9) at Object.apply (/var/lib/jenkins/workspace/elastic+estf-cloud-kibana-tests/JOB/xpackGrp4/TASK/saas_run_kibana_tests/node/ess-testing/ci/cloud/common/build/kibana/node_modules/@kbn/test/target_node/functional_test_runner/lib/mocha/wrap_function.js:87:16) ``` **Other test failures:** _Test Report: https://internal-ci.elastic.co/view/Stack%20Tests/job/elastic+estf-cloud-kibana-tests/2041/testReport/_
test
chrome x pack ui functional x pack test functional apps monitoring elasticsearch index detail mb·js monitoring app elasticsearch index detail mb active indices before all hook for should have an index summary with green status index with full shard allocation version class chrome x pack ui functional x pack test functional apps monitoring elasticsearch index detail mb·js stack trace error retry try timeout timeouterror waiting for element to be located by css selector wait timed out after at var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana node modules selenium webdriver lib webdriver js at runmicrotasks at processticksandrejections internal process task queues js at onfailure var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test common services retry retry for success ts at retryforsuccess var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test common services retry retry for success ts at retryservice try var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test common services retry retry ts at proxy clickbycssselector var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test functional services common find ts at testsubjects click var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana test functional services common test subjects ts at context test functional apps monitoring elasticsearch index detail mb js at object apply var lib jenkins workspace elastic estf cloud kibana tests job task saas run kibana tests node ess testing ci cloud common build kibana node modules kbn test target node functional test runner lib mocha wrap function js other test failures test report
1
161,539
13,852,142,670
IssuesEvent
2020-10-15 05:53:39
odoo/odoo
https://api.github.com/repos/odoo/odoo
closed
The Javascript API Doc page has not been ported to 13.0
13.0 Documentation
Impacted versions: 13.0 Documentation Steps to reproduce: go to https://www.odoo.com/documentation/13.0/reference/javascript_api.html# Expected behavior: https://www.odoo.com/documentation/12.0/reference/javascript_api.html#
1.0
The Javascript API Doc page has not been ported to 13.0 - Impacted versions: 13.0 Documentation Steps to reproduce: go to https://www.odoo.com/documentation/13.0/reference/javascript_api.html# Expected behavior: https://www.odoo.com/documentation/12.0/reference/javascript_api.html#
non_test
the javascript api doc page has not been ported to impacted versions documentation steps to reproduce go to expected behavior
0
60,198
6,675,712,964
IssuesEvent
2017-10-05 00:01:45
Indemnity83/solder
https://api.github.com/repos/Indemnity83/solder
closed
Browser testing fails on Travis-CI
PRs plz! type: test
There doesn't seem to be a clear solution to getting chrome working on Travis-CI build environments. There's either a lot of extra steps that need to be added to the test script (seems fragile) or users just simply changing to a different driver (like Firefox or PhantomJS). I'd really like to get the tests working with Chrome since it works so seamlessly in local development so for now I'm disabling the automated tests and inviting the community to help me troubleshoot getting the automated tests running so I can concentrate on finishing the GUI instead of spinning my wheels debugging Travis.
1.0
Browser testing fails on Travis-CI - There doesn't seem to be a clear solution to getting chrome working on Travis-CI build environments. There's either a lot of extra steps that need to be added to the test script (seems fragile) or users just simply changing to a different driver (like Firefox or PhantomJS). I'd really like to get the tests working with Chrome since it works so seamlessly in local development so for now I'm disabling the automated tests and inviting the community to help me troubleshoot getting the automated tests running so I can concentrate on finishing the GUI instead of spinning my wheels debugging Travis.
test
browser testing fails on travis ci there doesn t seem to be a clear solution to getting chrome working on travis ci build environments there s either a lot of extra steps that need to be added to the test script seems fragile or users just simply changing to a different driver like firefox or phantomjs i d really like to get the tests working with chrome since it works so seamlessly in local development so for now i m disabling the automated tests and inviting the community to help me troubleshoot getting the automated tests running so i can concentrate on finishing the gui instead of spinning my wheels debugging travis
1
64,855
6,924,959,277
IssuesEvent
2017-11-30 14:34:08
dwyl/bestevidence
https://api.github.com/repos/dwyl/bestevidence
closed
Bug - Filtering by evidence type no longer working
bug please-test priority-1 T4h
The dropdown menu, that enables users to filter by evidence type is not working on the app. I tested it when it was first developed and it worked but that functionality appears to have dropped off in the production version. This needs to be fixed urgently as it is pretty basic to the use of the app.
1.0
Bug - Filtering by evidence type no longer working - The dropdown menu, that enables users to filter by evidence type is not working on the app. I tested it when it was first developed and it worked but that functionality appears to have dropped off in the production version. This needs to be fixed urgently as it is pretty basic to the use of the app.
test
bug filtering by evidence type no longer working the dropdown menu that enables users to filter by evidence type is not working on the app i tested it when it was first developed and it worked but that functionality appears to have dropped off in the production version this needs to be fixed urgently as it is pretty basic to the use of the app
1
108,148
9,276,397,669
IssuesEvent
2019-03-20 02:45:01
dotnet/roslyn
https://api.github.com/repos/dotnet/roslyn
opened
Deadlock in RoamingVisualStudioProfileOptionPersister
Area-IDE Bug Integration-Test help wanted
**Version Used**: Stacks: ``` [Managed to Native Transition] mscorlib.dll!System.Threading.Monitor.Enter(object obj, ref bool lockTaken) Line 62 C# > Microsoft.VisualStudio.LanguageServices.dll!Microsoft.VisualStudio.LanguageServices.Implementation.Options.RoamingVisualStudioProfileOptionPersister.RecordObservedValueToWatchForChanges(Microsoft.CodeAnalysis.Options.OptionKey optionKey, string storageKey) Line 222 C# Microsoft.VisualStudio.LanguageServices.dll!Microsoft.VisualStudio.LanguageServices.Implementation.Options.RoamingVisualStudioProfileOptionPersister.GetFirstOrDefaultValue(Microsoft.CodeAnalysis.Options.OptionKey optionKey, System.Collections.Generic.IEnumerable<Microsoft.CodeAnalysis.Options.RoamingProfileStorageLocation> roamingSerializations) Line 99 C# Microsoft.VisualStudio.LanguageServices.dll!Microsoft.VisualStudio.LanguageServices.Implementation.Options.RoamingVisualStudioProfileOptionPersister.TryFetch(Microsoft.CodeAnalysis.Options.OptionKey optionKey, out object value) Line 128 C# Microsoft.CodeAnalysis.Workspaces.dll!Microsoft.CodeAnalysis.Options.GlobalOptionService.LoadOptionFromSerializerOrGetDefault(Microsoft.CodeAnalysis.Options.OptionKey optionKey) Line 48 C# Microsoft.CodeAnalysis.Workspaces.dll!Microsoft.CodeAnalysis.Options.GlobalOptionService.GetOption(Microsoft.CodeAnalysis.Options.OptionKey optionKey) Line 83 C# Microsoft.CodeAnalysis.Workspaces.dll!Microsoft.CodeAnalysis.Options.OptionServiceFactory.OptionService.GetOption(Microsoft.CodeAnalysis.Options.OptionKey optionKey) Line 121 C# Microsoft.CodeAnalysis.Workspaces.dll!Microsoft.CodeAnalysis.Options.WorkspaceOptionSet.GetOption(Microsoft.CodeAnalysis.Options.OptionKey optionKey) Line 39 C# Microsoft.CodeAnalysis.Workspaces.dll!Microsoft.CodeAnalysis.Options.WorkspaceOptionSet.WithChangedOption(Microsoft.CodeAnalysis.Options.OptionKey optionAndLanguage, object value) Line 46 C# Microsoft.VisualStudio.IntegrationTest.Utilities.dll!Microsoft.VisualStudio.IntegrationTest.Utilities.InProcess.VisualStudioWorkspace_InProc.SetPerLanguageOption(string optionName, string feature, string language, object value) Line 82 C# [Native to Managed Transition] [Managed to Native Transition] mscorlib.dll!System.Runtime.Remoting.Messaging.StackBuilderSink.SyncProcessMessage(System.Runtime.Remoting.Messaging.IMessage msg) Line 189 C# mscorlib.dll!System.Runtime.Remoting.Messaging.ServerObjectTerminatorSink.SyncProcessMessage(System.Runtime.Remoting.Messaging.IMessage reqMsg) Line 780 C# mscorlib.dll!System.Runtime.Remoting.Messaging.ServerContextTerminatorSink.SyncProcessMessage(System.Runtime.Remoting.Messaging.IMessage reqMsg) Line 616 C# mscorlib.dll!System.Runtime.Remoting.Channels.CrossContextChannel.SyncProcessMessageCallback(object[] args) Line 102 C# mscorlib.dll!System.Runtime.Remoting.Channels.ChannelServices.DispatchMessage(System.Runtime.Remoting.Channels.IServerChannelSinkStack sinkStack, System.Runtime.Remoting.Messaging.IMessage msg, out System.Runtime.Remoting.Messaging.IMessage replyMsg) Line 767 C# mscorlib.dll!System.Runtime.Remoting.Channels.DispatchChannelSink.ProcessMessage(System.Runtime.Remoting.Channels.IServerChannelSinkStack sinkStack, System.Runtime.Remoting.Messaging.IMessage requestMsg, System.Runtime.Remoting.Channels.ITransportHeaders requestHeaders, System.IO.Stream requestStream, out System.Runtime.Remoting.Messaging.IMessage responseMsg, out System.Runtime.Remoting.Channels.ITransportHeaders responseHeaders, out System.IO.Stream responseStream) Line 77 C# System.Runtime.Remoting.dll!System.Runtime.Remoting.Channels.BinaryServerFormatterSink.ProcessMessage(System.Runtime.Remoting.Channels.IServerChannelSinkStack sinkStack, System.Runtime.Remoting.Messaging.IMessage requestMsg, System.Runtime.Remoting.Channels.ITransportHeaders requestHeaders, System.IO.Stream requestStream, out System.Runtime.Remoting.Messaging.IMessage responseMsg, out System.Runtime.Remoting.Channels.ITransportHeaders responseHeaders, out System.IO.Stream responseStream) Unknown System.Runtime.Remoting.dll!System.Runtime.Remoting.Channels.Ipc.IpcServerTransportSink.ServiceRequest(object state) Unknown System.Runtime.Remoting.dll!System.Runtime.Remoting.Channels.SocketHandler.ProcessRequestNow() Unknown System.Runtime.Remoting.dll!System.Runtime.Remoting.Channels.RequestQueue.ProcessNextRequest(System.Runtime.Remoting.Channels.SocketHandler sh) Unknown System.Runtime.Remoting.dll!System.Runtime.Remoting.Channels.SocketHandler.BeginReadMessageCallback(System.IAsyncResult ar) Unknown System.Runtime.Remoting.dll!System.Runtime.Remoting.Channels.Ipc.IpcPort.AsyncFSCallback(uint errorCode, uint numBytes, System.Threading.NativeOverlapped* pOverlapped) Unknown mscorlib.dll!System.Threading._IOCompletionCallback.PerformIOCompletionCallback(uint errorCode, uint numBytes, System.Threading.NativeOverlapped* pOVERLAP) Line 135 C# [Native to Managed Transition] ``` ``` [Managed to Native Transition] mscorlib.dll!System.Threading.Monitor.Enter(object obj, ref bool lockTaken) Line 62 C# Microsoft.CodeAnalysis.Workspaces.dll!Microsoft.CodeAnalysis.Options.GlobalOptionService.RefreshOption(Microsoft.CodeAnalysis.Options.OptionKey optionKey, object newValue) Line 141 C# > Microsoft.VisualStudio.LanguageServices.dll!Microsoft.VisualStudio.LanguageServices.Implementation.Options.RoamingVisualStudioProfileOptionPersister.OnSettingChangedAsync(object sender, System.ComponentModel.PropertyChangedEventArgs args) Line 76 C# Microsoft.VisualStudio.Utilities.dll!Microsoft.VisualStudio.Settings.SettingsManager.AsyncHandler.Invoke(Microsoft.VisualStudio.Settings.SettingsManager sender, System.ComponentModel.PropertyChangedEventArgs args) Unknown Microsoft.VisualStudio.Utilities.dll!Microsoft.VisualStudio.Settings.SettingsManager.FireLocalSettingChangeEventAsync(System.ComponentModel.PropertyChangedEventArgs args, System.Collections.Generic.List<Microsoft.VisualStudio.Settings.SettingsManager.ScopedEventHandler> handlers) Unknown mscorlib.dll!System.Runtime.CompilerServices.AsyncMethodBuilderCore.MoveNextRunner.InvokeMoveNext(object stateMachine) Line 1090 C# mscorlib.dll!System.Threading.ExecutionContext.RunInternal(System.Threading.ExecutionContext executionContext, System.Threading.ContextCallback callback, object state, bool preserveSyncCtx) Line 954 C# mscorlib.dll!System.Threading.ExecutionContext.Run(System.Threading.ExecutionContext executionContext, System.Threading.ContextCallback callback, object state, bool preserveSyncCtx) Line 902 C# mscorlib.dll!System.Runtime.CompilerServices.AsyncMethodBuilderCore.MoveNextRunner.Run() Line 1070 C# mscorlib.dll!System.Threading.Tasks.AwaitTaskContinuation.System.Threading.IThreadPoolWorkItem.ExecuteWorkItem() Line 715 C# mscorlib.dll!System.Threading.ThreadPoolWorkQueue.Dispatch() Line 820 C# mscorlib.dll!System.Threading._ThreadPoolWaitCallback.PerformWaitCallback() Line 1161 C# [Native to Managed Transition] ```
1.0
Deadlock in RoamingVisualStudioProfileOptionPersister - **Version Used**: Stacks: ``` [Managed to Native Transition] mscorlib.dll!System.Threading.Monitor.Enter(object obj, ref bool lockTaken) Line 62 C# > Microsoft.VisualStudio.LanguageServices.dll!Microsoft.VisualStudio.LanguageServices.Implementation.Options.RoamingVisualStudioProfileOptionPersister.RecordObservedValueToWatchForChanges(Microsoft.CodeAnalysis.Options.OptionKey optionKey, string storageKey) Line 222 C# Microsoft.VisualStudio.LanguageServices.dll!Microsoft.VisualStudio.LanguageServices.Implementation.Options.RoamingVisualStudioProfileOptionPersister.GetFirstOrDefaultValue(Microsoft.CodeAnalysis.Options.OptionKey optionKey, System.Collections.Generic.IEnumerable<Microsoft.CodeAnalysis.Options.RoamingProfileStorageLocation> roamingSerializations) Line 99 C# Microsoft.VisualStudio.LanguageServices.dll!Microsoft.VisualStudio.LanguageServices.Implementation.Options.RoamingVisualStudioProfileOptionPersister.TryFetch(Microsoft.CodeAnalysis.Options.OptionKey optionKey, out object value) Line 128 C# Microsoft.CodeAnalysis.Workspaces.dll!Microsoft.CodeAnalysis.Options.GlobalOptionService.LoadOptionFromSerializerOrGetDefault(Microsoft.CodeAnalysis.Options.OptionKey optionKey) Line 48 C# Microsoft.CodeAnalysis.Workspaces.dll!Microsoft.CodeAnalysis.Options.GlobalOptionService.GetOption(Microsoft.CodeAnalysis.Options.OptionKey optionKey) Line 83 C# Microsoft.CodeAnalysis.Workspaces.dll!Microsoft.CodeAnalysis.Options.OptionServiceFactory.OptionService.GetOption(Microsoft.CodeAnalysis.Options.OptionKey optionKey) Line 121 C# Microsoft.CodeAnalysis.Workspaces.dll!Microsoft.CodeAnalysis.Options.WorkspaceOptionSet.GetOption(Microsoft.CodeAnalysis.Options.OptionKey optionKey) Line 39 C# Microsoft.CodeAnalysis.Workspaces.dll!Microsoft.CodeAnalysis.Options.WorkspaceOptionSet.WithChangedOption(Microsoft.CodeAnalysis.Options.OptionKey optionAndLanguage, object value) Line 46 C# Microsoft.VisualStudio.IntegrationTest.Utilities.dll!Microsoft.VisualStudio.IntegrationTest.Utilities.InProcess.VisualStudioWorkspace_InProc.SetPerLanguageOption(string optionName, string feature, string language, object value) Line 82 C# [Native to Managed Transition] [Managed to Native Transition] mscorlib.dll!System.Runtime.Remoting.Messaging.StackBuilderSink.SyncProcessMessage(System.Runtime.Remoting.Messaging.IMessage msg) Line 189 C# mscorlib.dll!System.Runtime.Remoting.Messaging.ServerObjectTerminatorSink.SyncProcessMessage(System.Runtime.Remoting.Messaging.IMessage reqMsg) Line 780 C# mscorlib.dll!System.Runtime.Remoting.Messaging.ServerContextTerminatorSink.SyncProcessMessage(System.Runtime.Remoting.Messaging.IMessage reqMsg) Line 616 C# mscorlib.dll!System.Runtime.Remoting.Channels.CrossContextChannel.SyncProcessMessageCallback(object[] args) Line 102 C# mscorlib.dll!System.Runtime.Remoting.Channels.ChannelServices.DispatchMessage(System.Runtime.Remoting.Channels.IServerChannelSinkStack sinkStack, System.Runtime.Remoting.Messaging.IMessage msg, out System.Runtime.Remoting.Messaging.IMessage replyMsg) Line 767 C# mscorlib.dll!System.Runtime.Remoting.Channels.DispatchChannelSink.ProcessMessage(System.Runtime.Remoting.Channels.IServerChannelSinkStack sinkStack, System.Runtime.Remoting.Messaging.IMessage requestMsg, System.Runtime.Remoting.Channels.ITransportHeaders requestHeaders, System.IO.Stream requestStream, out System.Runtime.Remoting.Messaging.IMessage responseMsg, out System.Runtime.Remoting.Channels.ITransportHeaders responseHeaders, out System.IO.Stream responseStream) Line 77 C# System.Runtime.Remoting.dll!System.Runtime.Remoting.Channels.BinaryServerFormatterSink.ProcessMessage(System.Runtime.Remoting.Channels.IServerChannelSinkStack sinkStack, System.Runtime.Remoting.Messaging.IMessage requestMsg, System.Runtime.Remoting.Channels.ITransportHeaders requestHeaders, System.IO.Stream requestStream, out System.Runtime.Remoting.Messaging.IMessage responseMsg, out System.Runtime.Remoting.Channels.ITransportHeaders responseHeaders, out System.IO.Stream responseStream) Unknown System.Runtime.Remoting.dll!System.Runtime.Remoting.Channels.Ipc.IpcServerTransportSink.ServiceRequest(object state) Unknown System.Runtime.Remoting.dll!System.Runtime.Remoting.Channels.SocketHandler.ProcessRequestNow() Unknown System.Runtime.Remoting.dll!System.Runtime.Remoting.Channels.RequestQueue.ProcessNextRequest(System.Runtime.Remoting.Channels.SocketHandler sh) Unknown System.Runtime.Remoting.dll!System.Runtime.Remoting.Channels.SocketHandler.BeginReadMessageCallback(System.IAsyncResult ar) Unknown System.Runtime.Remoting.dll!System.Runtime.Remoting.Channels.Ipc.IpcPort.AsyncFSCallback(uint errorCode, uint numBytes, System.Threading.NativeOverlapped* pOverlapped) Unknown mscorlib.dll!System.Threading._IOCompletionCallback.PerformIOCompletionCallback(uint errorCode, uint numBytes, System.Threading.NativeOverlapped* pOVERLAP) Line 135 C# [Native to Managed Transition] ``` ``` [Managed to Native Transition] mscorlib.dll!System.Threading.Monitor.Enter(object obj, ref bool lockTaken) Line 62 C# Microsoft.CodeAnalysis.Workspaces.dll!Microsoft.CodeAnalysis.Options.GlobalOptionService.RefreshOption(Microsoft.CodeAnalysis.Options.OptionKey optionKey, object newValue) Line 141 C# > Microsoft.VisualStudio.LanguageServices.dll!Microsoft.VisualStudio.LanguageServices.Implementation.Options.RoamingVisualStudioProfileOptionPersister.OnSettingChangedAsync(object sender, System.ComponentModel.PropertyChangedEventArgs args) Line 76 C# Microsoft.VisualStudio.Utilities.dll!Microsoft.VisualStudio.Settings.SettingsManager.AsyncHandler.Invoke(Microsoft.VisualStudio.Settings.SettingsManager sender, System.ComponentModel.PropertyChangedEventArgs args) Unknown Microsoft.VisualStudio.Utilities.dll!Microsoft.VisualStudio.Settings.SettingsManager.FireLocalSettingChangeEventAsync(System.ComponentModel.PropertyChangedEventArgs args, System.Collections.Generic.List<Microsoft.VisualStudio.Settings.SettingsManager.ScopedEventHandler> handlers) Unknown mscorlib.dll!System.Runtime.CompilerServices.AsyncMethodBuilderCore.MoveNextRunner.InvokeMoveNext(object stateMachine) Line 1090 C# mscorlib.dll!System.Threading.ExecutionContext.RunInternal(System.Threading.ExecutionContext executionContext, System.Threading.ContextCallback callback, object state, bool preserveSyncCtx) Line 954 C# mscorlib.dll!System.Threading.ExecutionContext.Run(System.Threading.ExecutionContext executionContext, System.Threading.ContextCallback callback, object state, bool preserveSyncCtx) Line 902 C# mscorlib.dll!System.Runtime.CompilerServices.AsyncMethodBuilderCore.MoveNextRunner.Run() Line 1070 C# mscorlib.dll!System.Threading.Tasks.AwaitTaskContinuation.System.Threading.IThreadPoolWorkItem.ExecuteWorkItem() Line 715 C# mscorlib.dll!System.Threading.ThreadPoolWorkQueue.Dispatch() Line 820 C# mscorlib.dll!System.Threading._ThreadPoolWaitCallback.PerformWaitCallback() Line 1161 C# [Native to Managed Transition] ```
test
deadlock in roamingvisualstudioprofileoptionpersister version used stacks mscorlib dll system threading monitor enter object obj ref bool locktaken line c microsoft visualstudio languageservices dll microsoft visualstudio languageservices implementation options roamingvisualstudioprofileoptionpersister recordobservedvaluetowatchforchanges microsoft codeanalysis options optionkey optionkey string storagekey line c microsoft visualstudio languageservices dll microsoft visualstudio languageservices implementation options roamingvisualstudioprofileoptionpersister getfirstordefaultvalue microsoft codeanalysis options optionkey optionkey system collections generic ienumerable roamingserializations line c microsoft visualstudio languageservices dll microsoft visualstudio languageservices implementation options roamingvisualstudioprofileoptionpersister tryfetch microsoft codeanalysis options optionkey optionkey out object value line c microsoft codeanalysis workspaces dll microsoft codeanalysis options globaloptionservice loadoptionfromserializerorgetdefault microsoft codeanalysis options optionkey optionkey line c microsoft codeanalysis workspaces dll microsoft codeanalysis options globaloptionservice getoption microsoft codeanalysis options optionkey optionkey line c microsoft codeanalysis workspaces dll microsoft codeanalysis options optionservicefactory optionservice getoption microsoft codeanalysis options optionkey optionkey line c microsoft codeanalysis workspaces dll microsoft codeanalysis options workspaceoptionset getoption microsoft codeanalysis options optionkey optionkey line c microsoft codeanalysis workspaces dll microsoft codeanalysis options workspaceoptionset withchangedoption microsoft codeanalysis options optionkey optionandlanguage object value line c microsoft visualstudio integrationtest utilities dll microsoft visualstudio integrationtest utilities inprocess visualstudioworkspace inproc setperlanguageoption string optionname string feature string language object value line c mscorlib dll system runtime remoting messaging stackbuildersink syncprocessmessage system runtime remoting messaging imessage msg line c mscorlib dll system runtime remoting messaging serverobjectterminatorsink syncprocessmessage system runtime remoting messaging imessage reqmsg line c mscorlib dll system runtime remoting messaging servercontextterminatorsink syncprocessmessage system runtime remoting messaging imessage reqmsg line c mscorlib dll system runtime remoting channels crosscontextchannel syncprocessmessagecallback object args line c mscorlib dll system runtime remoting channels channelservices dispatchmessage system runtime remoting channels iserverchannelsinkstack sinkstack system runtime remoting messaging imessage msg out system runtime remoting messaging imessage replymsg line c mscorlib dll system runtime remoting channels dispatchchannelsink processmessage system runtime remoting channels iserverchannelsinkstack sinkstack system runtime remoting messaging imessage requestmsg system runtime remoting channels itransportheaders requestheaders system io stream requeststream out system runtime remoting messaging imessage responsemsg out system runtime remoting channels itransportheaders responseheaders out system io stream responsestream line c system runtime remoting dll system runtime remoting channels binaryserverformattersink processmessage system runtime remoting channels iserverchannelsinkstack sinkstack system runtime remoting messaging imessage requestmsg system runtime remoting channels itransportheaders requestheaders system io stream requeststream out system runtime remoting messaging imessage responsemsg out system runtime remoting channels itransportheaders responseheaders out system io stream responsestream unknown system runtime remoting dll system runtime remoting channels ipc ipcservertransportsink servicerequest object state unknown system runtime remoting dll system runtime remoting channels sockethandler processrequestnow unknown system runtime remoting dll system runtime remoting channels requestqueue processnextrequest system runtime remoting channels sockethandler sh unknown system runtime remoting dll system runtime remoting channels sockethandler beginreadmessagecallback system iasyncresult ar unknown system runtime remoting dll system runtime remoting channels ipc ipcport asyncfscallback uint errorcode uint numbytes system threading nativeoverlapped poverlapped unknown mscorlib dll system threading iocompletioncallback performiocompletioncallback uint errorcode uint numbytes system threading nativeoverlapped poverlap line c mscorlib dll system threading monitor enter object obj ref bool locktaken line c microsoft codeanalysis workspaces dll microsoft codeanalysis options globaloptionservice refreshoption microsoft codeanalysis options optionkey optionkey object newvalue line c microsoft visualstudio languageservices dll microsoft visualstudio languageservices implementation options roamingvisualstudioprofileoptionpersister onsettingchangedasync object sender system componentmodel propertychangedeventargs args line c microsoft visualstudio utilities dll microsoft visualstudio settings settingsmanager asynchandler invoke microsoft visualstudio settings settingsmanager sender system componentmodel propertychangedeventargs args unknown microsoft visualstudio utilities dll microsoft visualstudio settings settingsmanager firelocalsettingchangeeventasync system componentmodel propertychangedeventargs args system collections generic list handlers unknown mscorlib dll system runtime compilerservices asyncmethodbuildercore movenextrunner invokemovenext object statemachine line c mscorlib dll system threading executioncontext runinternal system threading executioncontext executioncontext system threading contextcallback callback object state bool preservesyncctx line c mscorlib dll system threading executioncontext run system threading executioncontext executioncontext system threading contextcallback callback object state bool preservesyncctx line c mscorlib dll system runtime compilerservices asyncmethodbuildercore movenextrunner run line c mscorlib dll system threading tasks awaittaskcontinuation system threading ithreadpoolworkitem executeworkitem line c mscorlib dll system threading threadpoolworkqueue dispatch line c mscorlib dll system threading threadpoolwaitcallback performwaitcallback line c
1
126,084
10,382,913,924
IssuesEvent
2019-09-10 08:33:40
magento/async-import
https://api.github.com/repos/magento/async-import
closed
Create api-functional tests for DELETE source
Contribution Day Progress: dev in progress Topic: Asynchronous Import good first issue tests
### Summary (*) We need to cover with tests DELETE source functionality: Related PR: https://github.com/magento/async-import/pull/88 ### Proposed solution Please create Integration-Api tests for Delete source endpoint
1.0
Create api-functional tests for DELETE source - ### Summary (*) We need to cover with tests DELETE source functionality: Related PR: https://github.com/magento/async-import/pull/88 ### Proposed solution Please create Integration-Api tests for Delete source endpoint
test
create api functional tests for delete source summary we need to cover with tests delete source functionality related pr proposed solution please create integration api tests for delete source endpoint
1
63,410
12,313,274,714
IssuesEvent
2020-05-12 15:05:18
phetsims/build-a-molecule
https://api.github.com/repos/phetsims/build-a-molecule
closed
Common prefix name for source files
dev:code-review
From the code review in https://github.com/phetsims/build-a-molecule/issues/173: > Do filenames use an appropriate prefix? Some filenames may be prefixed with the repository name, e.g. MolarityConstants.js in molarity. If the repository name is long, the developer may choose to abbreviate the repository name, e.g. EEConstants.js in expression-exchange. If the abbreviation is already used by another respository, then the full name must be used. For example, if the "EE" abbreviation is already used by expression-exchange, then it should not be used in equality-explorer. Whichever convention is used, it should be used consistently within a repository - don't mix abbreviations and full names. It looks like BAM is the prefix, but we have BuildAMoleculeQueryParameters... can we rename that to be consistent?
1.0
Common prefix name for source files - From the code review in https://github.com/phetsims/build-a-molecule/issues/173: > Do filenames use an appropriate prefix? Some filenames may be prefixed with the repository name, e.g. MolarityConstants.js in molarity. If the repository name is long, the developer may choose to abbreviate the repository name, e.g. EEConstants.js in expression-exchange. If the abbreviation is already used by another respository, then the full name must be used. For example, if the "EE" abbreviation is already used by expression-exchange, then it should not be used in equality-explorer. Whichever convention is used, it should be used consistently within a repository - don't mix abbreviations and full names. It looks like BAM is the prefix, but we have BuildAMoleculeQueryParameters... can we rename that to be consistent?
non_test
common prefix name for source files from the code review in do filenames use an appropriate prefix some filenames may be prefixed with the repository name e g molarityconstants js in molarity if the repository name is long the developer may choose to abbreviate the repository name e g eeconstants js in expression exchange if the abbreviation is already used by another respository then the full name must be used for example if the ee abbreviation is already used by expression exchange then it should not be used in equality explorer whichever convention is used it should be used consistently within a repository don t mix abbreviations and full names it looks like bam is the prefix but we have buildamoleculequeryparameters can we rename that to be consistent
0
178,036
13,758,386,558
IssuesEvent
2020-10-06 23:54:20
rancher/rke2
https://api.github.com/repos/rancher/rke2
closed
--data-dir flag is ignored
[zube]: To Test
Starting rke2 with the --data-dir option does not alter the data directory. It always uses the default value of /var/lib/rancher/rke2.
1.0
--data-dir flag is ignored - Starting rke2 with the --data-dir option does not alter the data directory. It always uses the default value of /var/lib/rancher/rke2.
test
data dir flag is ignored starting with the data dir option does not alter the data directory it always uses the default value of var lib rancher
1
216,702
16,796,145,761
IssuesEvent
2021-06-16 04:02:33
AleoHQ/leo
https://api.github.com/repos/AleoHQ/leo
closed
[Tests] more tests for characters and strings
tests
## 🐛 Bug Report There is no separate issue label for tests that are needed, so I will call this a bug. There should be more tests for characters and strings. In `leo/tests/parser/expression/literal/char.leo` more kinds of literal characters should be tested. The principle here is to test at least one kind of each special category. Take a look at `rfc/001-initial-strings.md` to see what is special. Currently need to be added are: - '\'', '\\', '\n', 'u{..}' with 1, 2, or 3 hex digits in the braces; - 'å' (or any unicode character that has a 2-byte utf-8 encoding, in this example it is C3 A5 and the name is LATIN SMALL LETTER A WITH RING ABOVE); - '😭' (or any unicode character that has a 4-byte utf-8 encoding, in this example it is F0 9F 98 AD and the name is LOUDLY CRYING FACE); - for the \x and \u{} escapes, add tests test lowercase letter hex digits - Also, test all the nonprinting ascii characters as utf-8 chars. In other words, literal chars with code points 0 through 31. (If any of these are not supported, that should be documented.) In `leo/tests/compiler/char/` every special case that you see in `parser/expression/literal/char.leo` should be tested: * as an input literal * as a literal within the program * as an output literal (make sure the right thing shows up in the output) * sent to console.log() For string tests, there should be at least as good coverage of literals that there is with characters, both under tests/parser and tests/compiler There should be more failure tests, too. - Any backslash-<something> that is not valid should have a must-fail test. For example, \? where ? stands for any byte other than ascii \"'trn0xu should all fail. It would be good to test all such bytes, but especially important are ones that have an obvious reason to fail, like 1TRNXU - \x followed by anything other than 0-7 should fail - \x? followed by anything other than 0-9a-fA-F should fail - \u followed by anything other than { should fail - \u{ followed by anything other than 0-9a-fA-F should fail - \u{? followed by anything other than 0-9a-fA-F} should fail - \u{xxxxxx if the exes are all hex digits, then anything following other than } should fail - Sequences of bytes in literal chars and strings that are invalid UTF-8 should fail. To get good coverage of invalid UTF-8 byte sequences you can look at the well-formed sequences in table 3.7 of the unicode specification: https://www.unicode.org/versions/Unicode13.0.0/ch03.pdf#G7404 to see where the ill-formed sequences are. For example, here are some that should fail (where brackets represent the char or string quotes, spaces are ignored, and the hex values are the byte values): [80], [C1], [C2 7F], [DF C0], [E0 9F], etc. For tests that should fail parsing, it might be OK to test them only under tests/parser and not bother with tests/compiler. Although there is the issue of how the compiler handles parse tests, that might not need to be demonstrated or every kind of parse failure. ## Your Environment Leo 1.5.0 rustc 1.51.0 linux mint 19.3 (on ubuntu 18.04)
1.0
[Tests] more tests for characters and strings - ## 🐛 Bug Report There is no separate issue label for tests that are needed, so I will call this a bug. There should be more tests for characters and strings. In `leo/tests/parser/expression/literal/char.leo` more kinds of literal characters should be tested. The principle here is to test at least one kind of each special category. Take a look at `rfc/001-initial-strings.md` to see what is special. Currently need to be added are: - '\'', '\\', '\n', 'u{..}' with 1, 2, or 3 hex digits in the braces; - 'å' (or any unicode character that has a 2-byte utf-8 encoding, in this example it is C3 A5 and the name is LATIN SMALL LETTER A WITH RING ABOVE); - '😭' (or any unicode character that has a 4-byte utf-8 encoding, in this example it is F0 9F 98 AD and the name is LOUDLY CRYING FACE); - for the \x and \u{} escapes, add tests test lowercase letter hex digits - Also, test all the nonprinting ascii characters as utf-8 chars. In other words, literal chars with code points 0 through 31. (If any of these are not supported, that should be documented.) In `leo/tests/compiler/char/` every special case that you see in `parser/expression/literal/char.leo` should be tested: * as an input literal * as a literal within the program * as an output literal (make sure the right thing shows up in the output) * sent to console.log() For string tests, there should be at least as good coverage of literals that there is with characters, both under tests/parser and tests/compiler There should be more failure tests, too. - Any backslash-<something> that is not valid should have a must-fail test. For example, \? where ? stands for any byte other than ascii \"'trn0xu should all fail. It would be good to test all such bytes, but especially important are ones that have an obvious reason to fail, like 1TRNXU - \x followed by anything other than 0-7 should fail - \x? followed by anything other than 0-9a-fA-F should fail - \u followed by anything other than { should fail - \u{ followed by anything other than 0-9a-fA-F should fail - \u{? followed by anything other than 0-9a-fA-F} should fail - \u{xxxxxx if the exes are all hex digits, then anything following other than } should fail - Sequences of bytes in literal chars and strings that are invalid UTF-8 should fail. To get good coverage of invalid UTF-8 byte sequences you can look at the well-formed sequences in table 3.7 of the unicode specification: https://www.unicode.org/versions/Unicode13.0.0/ch03.pdf#G7404 to see where the ill-formed sequences are. For example, here are some that should fail (where brackets represent the char or string quotes, spaces are ignored, and the hex values are the byte values): [80], [C1], [C2 7F], [DF C0], [E0 9F], etc. For tests that should fail parsing, it might be OK to test them only under tests/parser and not bother with tests/compiler. Although there is the issue of how the compiler handles parse tests, that might not need to be demonstrated or every kind of parse failure. ## Your Environment Leo 1.5.0 rustc 1.51.0 linux mint 19.3 (on ubuntu 18.04)
test
more tests for characters and strings 🐛 bug report there is no separate issue label for tests that are needed so i will call this a bug there should be more tests for characters and strings in leo tests parser expression literal char leo more kinds of literal characters should be tested the principle here is to test at least one kind of each special category take a look at rfc initial strings md to see what is special currently need to be added are n u with or hex digits in the braces å or any unicode character that has a byte utf encoding in this example it is and the name is latin small letter a with ring above 😭 or any unicode character that has a byte utf encoding in this example it is ad and the name is loudly crying face for the x and u escapes add tests test lowercase letter hex digits also test all the nonprinting ascii characters as utf chars in other words literal chars with code points through if any of these are not supported that should be documented in leo tests compiler char every special case that you see in parser expression literal char leo should be tested as an input literal as a literal within the program as an output literal make sure the right thing shows up in the output sent to console log for string tests there should be at least as good coverage of literals that there is with characters both under tests parser and tests compiler there should be more failure tests too any backslash that is not valid should have a must fail test for example where stands for any byte other than ascii should all fail it would be good to test all such bytes but especially important are ones that have an obvious reason to fail like x followed by anything other than should fail x followed by anything other than fa f should fail u followed by anything other than should fail u followed by anything other than fa f should fail u followed by anything other than fa f should fail u xxxxxx if the exes are all hex digits then anything following other than should fail sequences of bytes in literal chars and strings that are invalid utf should fail to get good coverage of invalid utf byte sequences you can look at the well formed sequences in table of the unicode specification to see where the ill formed sequences are for example here are some that should fail where brackets represent the char or string quotes spaces are ignored and the hex values are the byte values etc for tests that should fail parsing it might be ok to test them only under tests parser and not bother with tests compiler although there is the issue of how the compiler handles parse tests that might not need to be demonstrated or every kind of parse failure your environment leo rustc linux mint on ubuntu
1
246,387
20,862,339,690
IssuesEvent
2022-03-22 00:57:13
backend-br/vagas
https://api.github.com/repos/backend-br/vagas
closed
[Remoto] Back-end Developer Ruby on Rails @Red Pill Pontomais
CLT Pleno TDD Ruby AWS Testes automatizados MongoDB Redis SQL Rest Stale
A Red Pill juntamente a Pontomais está em busca de um novo Desenvolvedor Ruby on Rails. Como Pessoa Desenvolvedora Back-End você terá a missão de desenvolver tecnicamente o produto visando a melhor experiência para o usuário, mantendo a estabilidade, SLA e entregas de qualidade, usando as melhores práticas de código. No seu dia a dia você vai: Desenvolvimento de aplicações com o uso das melhores práticas de engenharia de software, com foco no produto, com aderência nos processos da empresa e com envolvimento de diversas áreas durante a execução, incluindo a entrega; Garantir desenvolvimento de códigos com qualidade e cobertura de testes automatizados; Implementar features e automações através de ferramentas homologadas e trarão valor para nossos clientes e nossa operação; Garantir a estabilidade do produto visando alcançar a SLA de CES; Fazer code review; Contribuir para atingir os ORKS da empresa, desdobrados na área. O que você precisa ter: Experiência profissional a partir de 3 anos; Desenvolvimento de softwares utilizando banco de dados e Ruby on Rails Ferramenta de controle de versionamento de código; API Rest; Implantação End-to-End; Seria legal qye você tenha coenhecimento básico em: Automação de testes; Desenvolvimento orientado ao TDD; AWS - EC2, SQS, SNS, S3; Conhecimento de bancos de dados no-sql como Redis e MongoDB; Postgrsql; Benefícios Se liga aqui que tem muita coisa boa: Vale Refeição de R$30 por dia; Vale Alimentação de R$300 (quando vc sair de férias ainda recebe o VA, viu). VR e VA não tem aquele desconto na folha de pagamento; Plano de Saúde sem custo na mensalidade; Plano Odontológico sem custo na mensalidade; Serviço de ambulância para emergências médicas; Vale Farmácia; Totall pass - Seguindo a linha da saúde temos mais esse beneficio para você; R$100,00 de vale home office; Plano de Carreiras, Cargos e Salários; PAE - Programa de Assistência Jurídica, Psicológica, Previdenciária e mais; Licença Maternidade e Paternidade estendida; Treinamentos e workshops (internos e externos); Parcerias com Instituições de Ensino; Day-off em 1 dia do mês do seu aniversário - pq é um dia especial e vc merece esse mimo de escolher qual dia vc quer folgar no mês do seu aniversário (só não esquece de trazer um pedaço de bolo pra gente!); CLT Nível Pleno Faixa de 10k Currículos para crystal.g@redpillrh.com.br ou msg no whats 51985593215.
1.0
[Remoto] Back-end Developer Ruby on Rails @Red Pill Pontomais - A Red Pill juntamente a Pontomais está em busca de um novo Desenvolvedor Ruby on Rails. Como Pessoa Desenvolvedora Back-End você terá a missão de desenvolver tecnicamente o produto visando a melhor experiência para o usuário, mantendo a estabilidade, SLA e entregas de qualidade, usando as melhores práticas de código. No seu dia a dia você vai: Desenvolvimento de aplicações com o uso das melhores práticas de engenharia de software, com foco no produto, com aderência nos processos da empresa e com envolvimento de diversas áreas durante a execução, incluindo a entrega; Garantir desenvolvimento de códigos com qualidade e cobertura de testes automatizados; Implementar features e automações através de ferramentas homologadas e trarão valor para nossos clientes e nossa operação; Garantir a estabilidade do produto visando alcançar a SLA de CES; Fazer code review; Contribuir para atingir os ORKS da empresa, desdobrados na área. O que você precisa ter: Experiência profissional a partir de 3 anos; Desenvolvimento de softwares utilizando banco de dados e Ruby on Rails Ferramenta de controle de versionamento de código; API Rest; Implantação End-to-End; Seria legal qye você tenha coenhecimento básico em: Automação de testes; Desenvolvimento orientado ao TDD; AWS - EC2, SQS, SNS, S3; Conhecimento de bancos de dados no-sql como Redis e MongoDB; Postgrsql; Benefícios Se liga aqui que tem muita coisa boa: Vale Refeição de R$30 por dia; Vale Alimentação de R$300 (quando vc sair de férias ainda recebe o VA, viu). VR e VA não tem aquele desconto na folha de pagamento; Plano de Saúde sem custo na mensalidade; Plano Odontológico sem custo na mensalidade; Serviço de ambulância para emergências médicas; Vale Farmácia; Totall pass - Seguindo a linha da saúde temos mais esse beneficio para você; R$100,00 de vale home office; Plano de Carreiras, Cargos e Salários; PAE - Programa de Assistência Jurídica, Psicológica, Previdenciária e mais; Licença Maternidade e Paternidade estendida; Treinamentos e workshops (internos e externos); Parcerias com Instituições de Ensino; Day-off em 1 dia do mês do seu aniversário - pq é um dia especial e vc merece esse mimo de escolher qual dia vc quer folgar no mês do seu aniversário (só não esquece de trazer um pedaço de bolo pra gente!); CLT Nível Pleno Faixa de 10k Currículos para crystal.g@redpillrh.com.br ou msg no whats 51985593215.
test
back end developer ruby on rails red pill pontomais a red pill juntamente a pontomais está em busca de um novo desenvolvedor ruby on rails como pessoa desenvolvedora back end você terá a missão de desenvolver tecnicamente o produto visando a melhor experiência para o usuário mantendo a estabilidade sla e entregas de qualidade usando as melhores práticas de código no seu dia a dia você vai desenvolvimento de aplicações com o uso das melhores práticas de engenharia de software com foco no produto com aderência nos processos da empresa e com envolvimento de diversas áreas durante a execução incluindo a entrega garantir desenvolvimento de códigos com qualidade e cobertura de testes automatizados implementar features e automações através de ferramentas homologadas e trarão valor para nossos clientes e nossa operação garantir a estabilidade do produto visando alcançar a sla de ces fazer code review contribuir para atingir os orks da empresa desdobrados na área o que você precisa ter experiência profissional a partir de anos desenvolvimento de softwares utilizando banco de dados e ruby on rails ferramenta de controle de versionamento de código api rest implantação end to end seria legal qye você tenha coenhecimento básico em automação de testes desenvolvimento orientado ao tdd aws sqs sns conhecimento de bancos de dados no sql como redis e mongodb postgrsql benefícios se liga aqui que tem muita coisa boa vale refeição de r por dia vale alimentação de r quando vc sair de férias ainda recebe o va viu vr e va não tem aquele desconto na folha de pagamento plano de saúde sem custo na mensalidade plano odontológico sem custo na mensalidade serviço de ambulância para emergências médicas vale farmácia totall pass seguindo a linha da saúde temos mais esse beneficio para você r de vale home office plano de carreiras cargos e salários pae programa de assistência jurídica psicológica previdenciária e mais licença maternidade e paternidade estendida treinamentos e workshops internos e externos parcerias com instituições de ensino day off em dia do mês do seu aniversário pq é um dia especial e vc merece esse mimo de escolher qual dia vc quer folgar no mês do seu aniversário só não esquece de trazer um pedaço de bolo pra gente clt nível pleno faixa de currículos para crystal g redpillrh com br ou msg no whats
1
66,055
6,987,327,493
IssuesEvent
2017-12-14 08:48:46
nodejs/node
https://api.github.com/repos/nodejs/node
closed
make test triggers "node quit unexpectedly" on Fedora 27
question test
* **Version**: master * **Platform**: Linux 4.14.3-300.fc27.x86_64 #1 SMP Mon Dec 4 17:18:27 UTC 2017 x86_64 x86_64 x86_64 GNU/Linux * **Subsystem**: test <!-- Enter your issue details below this comment. --> Last week I formatted my main SSD and did a fresh installation of Fedora 27. Now, when I run `make test`, there are always three crashes reported by the system as "node quit unexpectedly". This did not happen with my old installation. It seems those are part of the tests, but I wonder if it's possible to avoid having them caught by the system? ![screenshot from 2017-12-12 13-08-49](https://user-images.githubusercontent.com/2352663/33883880-d6a8c154-df3d-11e7-8590-c8aafcf01e6c.png) /cc @bnoordhuis ?
1.0
make test triggers "node quit unexpectedly" on Fedora 27 - * **Version**: master * **Platform**: Linux 4.14.3-300.fc27.x86_64 #1 SMP Mon Dec 4 17:18:27 UTC 2017 x86_64 x86_64 x86_64 GNU/Linux * **Subsystem**: test <!-- Enter your issue details below this comment. --> Last week I formatted my main SSD and did a fresh installation of Fedora 27. Now, when I run `make test`, there are always three crashes reported by the system as "node quit unexpectedly". This did not happen with my old installation. It seems those are part of the tests, but I wonder if it's possible to avoid having them caught by the system? ![screenshot from 2017-12-12 13-08-49](https://user-images.githubusercontent.com/2352663/33883880-d6a8c154-df3d-11e7-8590-c8aafcf01e6c.png) /cc @bnoordhuis ?
test
make test triggers node quit unexpectedly on fedora version master platform linux smp mon dec utc gnu linux subsystem test last week i formatted my main ssd and did a fresh installation of fedora now when i run make test there are always three crashes reported by the system as node quit unexpectedly this did not happen with my old installation it seems those are part of the tests but i wonder if it s possible to avoid having them caught by the system cc bnoordhuis
1
386,564
11,441,083,129
IssuesEvent
2020-02-05 10:55:42
craftercms/craftercms
https://api.github.com/repos/craftercms/craftercms
closed
[studio] Error Site context is not valid anymore when creating first site in Studio
bug priority: high
## Describe the bug On a fresh install or when you delete the `data` folder, there is an error that the site context is not valid anymore when creating the very first site in Studio ``` [ERROR] 2020-01-06T11:11:16,696 [http-nio-8080-exec-6] [] [context.SiteContextManager] | Site context SiteContext{siteName='newsite', context=FileSystemContext{id='977d6174f06786497c2d88f76bf315b0', rootFolderPath='file:/Users/vita/temp/test3/craftercms/crafter-authoring/data/repos/sites/newsite/sandbox/'}, fallback=false, staticAssetsPath='/static-assets', templatesPath='/', restScriptsPath='/scripts/rest', controllerScriptsPath='/scripts/controllers'} is not valid anymore ``` ## To Reproduce Steps to reproduce the behavior: 1. Clone craftercms or, delete the `crafter-authoring/data` folder 2. Start Studio 3. Create a site using the website editorial bp 4. Watch the logs as the site is being created and notice after your site is created, thesite preview is not available and instead you get a `Could not resolve site for the current request.` ![Screen Shot 2020-01-06 at 11 51 09 AM](https://user-images.githubusercontent.com/25483966/71833460-f30bc680-307a-11ea-8c83-0f7dbdde53d7.png) ## Expected behavior There should be no errors when creating the first site in Studio ## Screenshots {{If applicable, add screenshots to help explain your problem.}} ## Logs ``` [ERROR] 2020-01-06T11:11:16,696 [http-nio-8080-exec-6] [] [context.SiteContextManager] | Site context SiteContext{siteName='newsite', context=FileSystemContext{id='977d6174f06786497c2d88f76bf315b0', rootFolderPath='file:/Users/vita/temp/test3/craftercms/crafter-authoring/data/repos/sites/newsite/sandbox/'}, fallback=false, staticAssetsPath='/static-assets', templatesPath='/', restScriptsPath='/scripts/rest', controllerScriptsPath='/scripts/controllers'} is not valid anymore [INFO] 2020-01-06T11:11:16,696 [http-nio-8080-exec-6] [] [context.SiteContextManager] | ================================================== [INFO] 2020-01-06T11:11:16,696 [http-nio-8080-exec-6] [] [context.SiteContextManager] | <Destroying site context: newsite> [INFO] 2020-01-06T11:11:16,697 [http-nio-8080-exec-6] [] [context.SiteContextManager] | ================================================== [INFO] 2020-01-06T11:11:16,697 [http-nio-8080-exec-6] [] [support.GenericApplicationContext] | Closing org.springframework.context.support.GenericApplicationContext@441fa52e: startup date [Mon Jan 06 11:11:15 EST 2020]; parent: Root WebApplicationContext [INFO] 2020-01-06T11:11:16,698 [http-nio-8080-exec-6] [] [context.SiteContextManager] | Site context destroyed: SiteContext{siteName='newsite', context=FileSystemContext{id='977d6174f06786497c2d88f76bf315b0', rootFolderPath='file:/Users/vita/temp/test3/craftercms/crafter-authoring/data/repos/sites/newsite/sandbox/'}, fallback=false, staticAssetsPath='/static-assets', templatesPath='/', restScriptsPath='/scripts/rest', controllerScriptsPath='/scripts/controllers'} [INFO] 2020-01-06T11:11:16,698 [http-nio-8080-exec-6] [] [context.SiteContextManager] | ================================================== [INFO] 2020-01-06T11:11:16,699 [http-nio-8080-exec-6] [] [context.SiteContextManager] | </Destroying site context: newsite> [INFO] 2020-01-06T11:11:16,699 [http-nio-8080-exec-6] [] [context.SiteContextManager] | ================================================== [ERROR] 2020-01-06T11:11:16,699 [http-nio-8080-exec-6] [] [filter.SiteContextResolvingFilter] | Error while resolving site context for current request java.lang.IllegalStateException: Unable to resolve context for site name 'newsite' at org.craftercms.engine.service.context.SiteContextResolverImpl.getContext(SiteContextResolverImpl.java:81) ~[classes/:3.1.5-SNAPSHOT] at org.craftercms.engine.servlet.filter.SiteContextResolvingFilter.getContext(SiteContextResolvingFilter.java:95) [classes/:3.1.5-SNAPSHOT] at org.craftercms.engine.servlet.filter.SiteContextResolvingFilter.doFilter(SiteContextResolvingFilter.java:79) [classes/:3.1.5-SNAPSHOT] at org.springframework.web.filter.DelegatingFilterProxy.invokeDelegate(DelegatingFilterProxy.java:347) [spring-web-4.3.18.RELEASE.jar:4.3.18.RELEASE] at org.springframework.web.filter.DelegatingFilterProxy.doFilter(DelegatingFilterProxy.java:263) [spring-web-4.3.18.RELEASE.jar:4.3.18.RELEASE] at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193) [catalina.jar:8.5.24] at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) [catalina.jar:8.5.24] at org.craftercms.commons.http.RequestContextBindingFilter.doFilter(RequestContextBindingFilter.java:79) [crafter-commons-utilities-3.1.5-SNAPSHOT.jar:3.1.5-SNAPSHOT] at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193) [catalina.jar:8.5.24] at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) [catalina.jar:8.5.24] at org.springframework.web.filter.CharacterEncodingFilter.doFilterInternal(CharacterEncodingFilter.java:197) [spring-web-4.3.18.RELEASE.jar:4.3.18.RELEASE] at org.springframework.web.filter.OncePerRequestFilter.doFilter(OncePerRequestFilter.java:107) [spring-web-4.3.18.RELEASE.jar:4.3.18.RELEASE] at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193) [catalina.jar:8.5.24] at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) [catalina.jar:8.5.24] at org.apache.logging.log4j.web.Log4jServletFilter.doFilter(Log4jServletFilter.java:71) [log4j-web-2.11.2.jar:2.11.2] at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193) [catalina.jar:8.5.24] at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) [catalina.jar:8.5.24] at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:198) [catalina.jar:8.5.24] at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:96) [catalina.jar:8.5.24] at org.apache.catalina.authenticator.AuthenticatorBase.invoke(AuthenticatorBase.java:504) [catalina.jar:8.5.24] at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:140) [catalina.jar:8.5.24] at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:81) [catalina.jar:8.5.24] at org.apache.catalina.valves.AbstractAccessLogValve.invoke(AbstractAccessLogValve.java:650) [catalina.jar:8.5.24] at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:87) [catalina.jar:8.5.24] at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:342) [catalina.jar:8.5.24] at org.apache.coyote.http11.Http11Processor.service(Http11Processor.java:803) [tomcat-coyote.jar:8.5.24] at org.apache.coyote.AbstractProcessorLight.process(AbstractProcessorLight.java:66) [tomcat-coyote.jar:8.5.24] at org.apache.coyote.AbstractProtocol$ConnectionHandler.process(AbstractProtocol.java:790) [tomcat-coyote.jar:8.5.24] at org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.doRun(NioEndpoint.java:1459) [tomcat-coyote.jar:8.5.24] at org.apache.tomcat.util.net.SocketProcessorBase.run(SocketProcessorBase.java:49) [tomcat-coyote.jar:8.5.24] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) [?:1.8.0_162] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) [?:1.8.0_162] at org.apache.tomcat.util.threads.TaskThread$WrappingRunnable.run(TaskThread.java:61) [tomcat-util.jar:8.5.24] at java.lang.Thread.run(Thread.java:748) [?:1.8.0_162] ``` ## Specs ### Version Studio Version Number: 3.1.5-SNAPSHOT-189958 Build Number: 189958a33a56a46094df29c2c9675f709331ceba Build Date/Time: 01-06-2020 11:02:57 -0500 ### OS OS X ### Browser Chrome browser ## Additional context The error only appears the first time you create a site. Also, after getting the error `Could not resolve site for the current request.` in the UI, if you click on `Main Menu` -> `Sites` then click on your site, the site is now available for preview.
1.0
[studio] Error Site context is not valid anymore when creating first site in Studio - ## Describe the bug On a fresh install or when you delete the `data` folder, there is an error that the site context is not valid anymore when creating the very first site in Studio ``` [ERROR] 2020-01-06T11:11:16,696 [http-nio-8080-exec-6] [] [context.SiteContextManager] | Site context SiteContext{siteName='newsite', context=FileSystemContext{id='977d6174f06786497c2d88f76bf315b0', rootFolderPath='file:/Users/vita/temp/test3/craftercms/crafter-authoring/data/repos/sites/newsite/sandbox/'}, fallback=false, staticAssetsPath='/static-assets', templatesPath='/', restScriptsPath='/scripts/rest', controllerScriptsPath='/scripts/controllers'} is not valid anymore ``` ## To Reproduce Steps to reproduce the behavior: 1. Clone craftercms or, delete the `crafter-authoring/data` folder 2. Start Studio 3. Create a site using the website editorial bp 4. Watch the logs as the site is being created and notice after your site is created, thesite preview is not available and instead you get a `Could not resolve site for the current request.` ![Screen Shot 2020-01-06 at 11 51 09 AM](https://user-images.githubusercontent.com/25483966/71833460-f30bc680-307a-11ea-8c83-0f7dbdde53d7.png) ## Expected behavior There should be no errors when creating the first site in Studio ## Screenshots {{If applicable, add screenshots to help explain your problem.}} ## Logs ``` [ERROR] 2020-01-06T11:11:16,696 [http-nio-8080-exec-6] [] [context.SiteContextManager] | Site context SiteContext{siteName='newsite', context=FileSystemContext{id='977d6174f06786497c2d88f76bf315b0', rootFolderPath='file:/Users/vita/temp/test3/craftercms/crafter-authoring/data/repos/sites/newsite/sandbox/'}, fallback=false, staticAssetsPath='/static-assets', templatesPath='/', restScriptsPath='/scripts/rest', controllerScriptsPath='/scripts/controllers'} is not valid anymore [INFO] 2020-01-06T11:11:16,696 [http-nio-8080-exec-6] [] [context.SiteContextManager] | ================================================== [INFO] 2020-01-06T11:11:16,696 [http-nio-8080-exec-6] [] [context.SiteContextManager] | <Destroying site context: newsite> [INFO] 2020-01-06T11:11:16,697 [http-nio-8080-exec-6] [] [context.SiteContextManager] | ================================================== [INFO] 2020-01-06T11:11:16,697 [http-nio-8080-exec-6] [] [support.GenericApplicationContext] | Closing org.springframework.context.support.GenericApplicationContext@441fa52e: startup date [Mon Jan 06 11:11:15 EST 2020]; parent: Root WebApplicationContext [INFO] 2020-01-06T11:11:16,698 [http-nio-8080-exec-6] [] [context.SiteContextManager] | Site context destroyed: SiteContext{siteName='newsite', context=FileSystemContext{id='977d6174f06786497c2d88f76bf315b0', rootFolderPath='file:/Users/vita/temp/test3/craftercms/crafter-authoring/data/repos/sites/newsite/sandbox/'}, fallback=false, staticAssetsPath='/static-assets', templatesPath='/', restScriptsPath='/scripts/rest', controllerScriptsPath='/scripts/controllers'} [INFO] 2020-01-06T11:11:16,698 [http-nio-8080-exec-6] [] [context.SiteContextManager] | ================================================== [INFO] 2020-01-06T11:11:16,699 [http-nio-8080-exec-6] [] [context.SiteContextManager] | </Destroying site context: newsite> [INFO] 2020-01-06T11:11:16,699 [http-nio-8080-exec-6] [] [context.SiteContextManager] | ================================================== [ERROR] 2020-01-06T11:11:16,699 [http-nio-8080-exec-6] [] [filter.SiteContextResolvingFilter] | Error while resolving site context for current request java.lang.IllegalStateException: Unable to resolve context for site name 'newsite' at org.craftercms.engine.service.context.SiteContextResolverImpl.getContext(SiteContextResolverImpl.java:81) ~[classes/:3.1.5-SNAPSHOT] at org.craftercms.engine.servlet.filter.SiteContextResolvingFilter.getContext(SiteContextResolvingFilter.java:95) [classes/:3.1.5-SNAPSHOT] at org.craftercms.engine.servlet.filter.SiteContextResolvingFilter.doFilter(SiteContextResolvingFilter.java:79) [classes/:3.1.5-SNAPSHOT] at org.springframework.web.filter.DelegatingFilterProxy.invokeDelegate(DelegatingFilterProxy.java:347) [spring-web-4.3.18.RELEASE.jar:4.3.18.RELEASE] at org.springframework.web.filter.DelegatingFilterProxy.doFilter(DelegatingFilterProxy.java:263) [spring-web-4.3.18.RELEASE.jar:4.3.18.RELEASE] at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193) [catalina.jar:8.5.24] at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) [catalina.jar:8.5.24] at org.craftercms.commons.http.RequestContextBindingFilter.doFilter(RequestContextBindingFilter.java:79) [crafter-commons-utilities-3.1.5-SNAPSHOT.jar:3.1.5-SNAPSHOT] at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193) [catalina.jar:8.5.24] at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) [catalina.jar:8.5.24] at org.springframework.web.filter.CharacterEncodingFilter.doFilterInternal(CharacterEncodingFilter.java:197) [spring-web-4.3.18.RELEASE.jar:4.3.18.RELEASE] at org.springframework.web.filter.OncePerRequestFilter.doFilter(OncePerRequestFilter.java:107) [spring-web-4.3.18.RELEASE.jar:4.3.18.RELEASE] at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193) [catalina.jar:8.5.24] at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) [catalina.jar:8.5.24] at org.apache.logging.log4j.web.Log4jServletFilter.doFilter(Log4jServletFilter.java:71) [log4j-web-2.11.2.jar:2.11.2] at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193) [catalina.jar:8.5.24] at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) [catalina.jar:8.5.24] at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:198) [catalina.jar:8.5.24] at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:96) [catalina.jar:8.5.24] at org.apache.catalina.authenticator.AuthenticatorBase.invoke(AuthenticatorBase.java:504) [catalina.jar:8.5.24] at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:140) [catalina.jar:8.5.24] at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:81) [catalina.jar:8.5.24] at org.apache.catalina.valves.AbstractAccessLogValve.invoke(AbstractAccessLogValve.java:650) [catalina.jar:8.5.24] at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:87) [catalina.jar:8.5.24] at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:342) [catalina.jar:8.5.24] at org.apache.coyote.http11.Http11Processor.service(Http11Processor.java:803) [tomcat-coyote.jar:8.5.24] at org.apache.coyote.AbstractProcessorLight.process(AbstractProcessorLight.java:66) [tomcat-coyote.jar:8.5.24] at org.apache.coyote.AbstractProtocol$ConnectionHandler.process(AbstractProtocol.java:790) [tomcat-coyote.jar:8.5.24] at org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.doRun(NioEndpoint.java:1459) [tomcat-coyote.jar:8.5.24] at org.apache.tomcat.util.net.SocketProcessorBase.run(SocketProcessorBase.java:49) [tomcat-coyote.jar:8.5.24] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) [?:1.8.0_162] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) [?:1.8.0_162] at org.apache.tomcat.util.threads.TaskThread$WrappingRunnable.run(TaskThread.java:61) [tomcat-util.jar:8.5.24] at java.lang.Thread.run(Thread.java:748) [?:1.8.0_162] ``` ## Specs ### Version Studio Version Number: 3.1.5-SNAPSHOT-189958 Build Number: 189958a33a56a46094df29c2c9675f709331ceba Build Date/Time: 01-06-2020 11:02:57 -0500 ### OS OS X ### Browser Chrome browser ## Additional context The error only appears the first time you create a site. Also, after getting the error `Could not resolve site for the current request.` in the UI, if you click on `Main Menu` -> `Sites` then click on your site, the site is now available for preview.
non_test
error site context is not valid anymore when creating first site in studio describe the bug on a fresh install or when you delete the data folder there is an error that the site context is not valid anymore when creating the very first site in studio site context sitecontext sitename newsite context filesystemcontext id rootfolderpath file users vita temp craftercms crafter authoring data repos sites newsite sandbox fallback false staticassetspath static assets templatespath restscriptspath scripts rest controllerscriptspath scripts controllers is not valid anymore to reproduce steps to reproduce the behavior clone craftercms or delete the crafter authoring data folder start studio create a site using the website editorial bp watch the logs as the site is being created and notice after your site is created thesite preview is not available and instead you get a could not resolve site for the current request expected behavior there should be no errors when creating the first site in studio screenshots if applicable add screenshots to help explain your problem logs site context sitecontext sitename newsite context filesystemcontext id rootfolderpath file users vita temp craftercms crafter authoring data repos sites newsite sandbox fallback false staticassetspath static assets templatespath restscriptspath scripts rest controllerscriptspath scripts controllers is not valid anymore closing org springframework context support genericapplicationcontext startup date parent root webapplicationcontext site context destroyed sitecontext sitename newsite context filesystemcontext id rootfolderpath file users vita temp craftercms crafter authoring data repos sites newsite sandbox fallback false staticassetspath static assets templatespath restscriptspath scripts rest controllerscriptspath scripts controllers error while resolving site context for current request java lang illegalstateexception unable to resolve context for site name newsite at org craftercms engine service context sitecontextresolverimpl getcontext sitecontextresolverimpl java at org craftercms engine servlet filter sitecontextresolvingfilter getcontext sitecontextresolvingfilter java at org craftercms engine servlet filter sitecontextresolvingfilter dofilter sitecontextresolvingfilter java at org springframework web filter delegatingfilterproxy invokedelegate delegatingfilterproxy java at org springframework web filter delegatingfilterproxy dofilter delegatingfilterproxy java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org craftercms commons http requestcontextbindingfilter dofilter requestcontextbindingfilter java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org springframework web filter characterencodingfilter dofilterinternal characterencodingfilter java at org springframework web filter onceperrequestfilter dofilter onceperrequestfilter java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org apache logging web dofilter java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org apache catalina core standardwrappervalve invoke standardwrappervalve java at org apache catalina core standardcontextvalve invoke standardcontextvalve java at org apache catalina authenticator authenticatorbase invoke authenticatorbase java at org apache catalina core standardhostvalve invoke standardhostvalve java at org apache catalina valves errorreportvalve invoke errorreportvalve java at org apache catalina valves abstractaccesslogvalve invoke abstractaccesslogvalve java at org apache catalina core standardenginevalve invoke standardenginevalve java at org apache catalina connector coyoteadapter service coyoteadapter java at org apache coyote service java at org apache coyote abstractprocessorlight process abstractprocessorlight java at org apache coyote abstractprotocol connectionhandler process abstractprotocol java at org apache tomcat util net nioendpoint socketprocessor dorun nioendpoint java at org apache tomcat util net socketprocessorbase run socketprocessorbase java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at org apache tomcat util threads taskthread wrappingrunnable run taskthread java at java lang thread run thread java specs version studio version number snapshot build number build date time os os x browser chrome browser additional context the error only appears the first time you create a site also after getting the error could not resolve site for the current request in the ui if you click on main menu sites then click on your site the site is now available for preview
0
48,016
2,990,121,378
IssuesEvent
2015-07-21 07:04:46
jayway/rest-assured
https://api.github.com/repos/jayway/rest-assured
closed
Custom cookie parsing fails when request contains redirect
bug imported invalid Priority-Medium
_From [everf...@gmail.com](https://code.google.com/u/100456722887713625293/) on May 11, 2013 20:15:23_ Given: A request with a redirect as a result, f.e. @Test public void logAll() { RestAssured.reset(); given().log().all().when().get("/demo"); //setting cookies, HTTP 302 } expected: request can be used like all others. actual result: Request method: GET Request path: http://localhost:8080/demo Request params: \<none> Query params: \<none> Form params: \<none> Path params: \<none> Headers: Content-Type=*/* Cookies: \<none> Body: \<none> Assumption: Since the Apache httpclient internally handles the redirect, rest assured is not noticing what happens before the final result. This leads to lost cookies and possibly other problems. _Original issue: http://code.google.com/p/rest-assured/issues/detail?id=235_
1.0
Custom cookie parsing fails when request contains redirect - _From [everf...@gmail.com](https://code.google.com/u/100456722887713625293/) on May 11, 2013 20:15:23_ Given: A request with a redirect as a result, f.e. @Test public void logAll() { RestAssured.reset(); given().log().all().when().get("/demo"); //setting cookies, HTTP 302 } expected: request can be used like all others. actual result: Request method: GET Request path: http://localhost:8080/demo Request params: \<none> Query params: \<none> Form params: \<none> Path params: \<none> Headers: Content-Type=*/* Cookies: \<none> Body: \<none> Assumption: Since the Apache httpclient internally handles the redirect, rest assured is not noticing what happens before the final result. This leads to lost cookies and possibly other problems. _Original issue: http://code.google.com/p/rest-assured/issues/detail?id=235_
non_test
custom cookie parsing fails when request contains redirect from on may given a request with a redirect as a result f e test public void logall restassured reset given log all when get demo setting cookies http expected request can be used like all others actual result request method get request path request params query params form params path params headers content type cookies body assumption since the apache httpclient internally handles the redirect rest assured is not noticing what happens before the final result this leads to lost cookies and possibly other problems original issue
0
202,300
15,280,460,293
IssuesEvent
2021-02-23 06:22:21
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
ccl/importccl: TestCSVImportCanBeResumed failed
C-test-failure O-robot branch-master
[(ccl/importccl).TestCSVImportCanBeResumed failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2707940&tab=buildLog) on [master@2572200f7612c6508a52735a6a18767cfb7cc09d](https://github.com/cockroachdb/cockroach/commits/2572200f7612c6508a52735a6a18767cfb7cc09d): ``` github.com/cockroachdb/cockroach/pkg/ccl/importccl.(*readImportDataProcessor).Start() /go/src/github.com/cockroachdb/cockroach/pkg/ccl/importccl/import_processor.go:105 +0xa4 github.com/cockroachdb/cockroach/pkg/sql/execinfra.(*ProcessorBase).Run() /go/src/github.com/cockroachdb/cockroach/pkg/sql/execinfra/processorsbase.go:764 +0x8e github.com/cockroachdb/cockroach/pkg/ccl/importccl.(*readImportDataProcessor).Run() <autogenerated>:1 +0x57 github.com/cockroachdb/cockroach/pkg/sql/flowinfra.(*FlowBase).Run() /go/src/github.com/cockroachdb/cockroach/pkg/sql/flowinfra/flow.go:382 +0x2c6 github.com/cockroachdb/cockroach/pkg/sql/rowflow.(*rowBasedFlow).Run() <autogenerated>:1 +0x84 github.com/cockroachdb/cockroach/pkg/sql.(*DistSQLPlanner).Run() /go/src/github.com/cockroachdb/cockroach/pkg/sql/distsql_running.go:367 +0x7a7 github.com/cockroachdb/cockroach/pkg/sql.DistIngest.func6() /go/src/github.com/cockroachdb/cockroach/pkg/sql/distsql_plan_csv.go:332 +0x1b2 github.com/cockroachdb/cockroach/pkg/util/ctxgroup.Group.GoCtx.func1() /go/src/github.com/cockroachdb/cockroach/pkg/util/ctxgroup/ctxgroup.go:166 +0x4e golang.org/x/sync/errgroup.(*Group).Go.func1() /go/src/github.com/cockroachdb/cockroach/vendor/golang.org/x/sync/errgroup/errgroup.go:57 +0x94 Goroutine 26 (running) created at: golang.org/x/sync/errgroup.(*Group).Go() /go/src/github.com/cockroachdb/cockroach/vendor/golang.org/x/sync/errgroup/errgroup.go:54 +0x73 github.com/cockroachdb/cockroach/pkg/util/ctxgroup.Group.GoCtx() /go/src/github.com/cockroachdb/cockroach/pkg/util/ctxgroup/ctxgroup.go:165 +0xe4 github.com/cockroachdb/cockroach/pkg/sql.DistIngest() /go/src/github.com/cockroachdb/cockroach/pkg/sql/distsql_plan_csv.go:328 +0x128d github.com/cockroachdb/cockroach/pkg/ccl/importccl.(*importResumer).Resume() /go/src/github.com/cockroachdb/cockroach/pkg/ccl/importccl/import_stmt.go:1632 +0xe38 github.com/cockroachdb/cockroach/pkg/ccl/importccl.(*cancellableImportResumer).Resume() /go/src/github.com/cockroachdb/cockroach/pkg/ccl/importccl/import_processor_test.go:554 +0x1cc github.com/cockroachdb/cockroach/pkg/jobs.(*Registry).stepThroughStateMachine.func1() /go/src/github.com/cockroachdb/cockroach/pkg/jobs/registry.go:1170 +0x141 github.com/cockroachdb/cockroach/pkg/jobs.(*Registry).stepThroughStateMachine() /go/src/github.com/cockroachdb/cockroach/pkg/jobs/registry.go:1171 +0xa85 github.com/cockroachdb/cockroach/pkg/jobs.(*Registry).runJob() /go/src/github.com/cockroachdb/cockroach/pkg/jobs/adopt.go:250 +0x43e github.com/cockroachdb/cockroach/pkg/jobs.(*StartableJob).Start.func3() /go/src/github.com/cockroachdb/cockroach/pkg/jobs/jobs.go:867 +0x25c github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask.func1() /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:351 +0x149 ================== E210223 06:16:33.560091 1505 jobs/adopt.go:255 [n1] 1 job 635668376058003457: adoption completed with error job 635668376058003457: could not mark as reverting: context canceled: job with status pause-requested cannot be reverted import_processor_test.go:733: Resume pos: 30 E210223 06:16:34.408378 635 jobs/registry.go:672 [-] 2 error claiming jobs: could not query jobs table: claim-jobs: context canceled E210223 06:16:34.408913 633 jobs/registry.go:662 [-] 3 failed to serve pause and cancel requests: could not query jobs table: cancel/pause-requested: context canceled E210223 06:16:34.410822 635 jobs/registry.go:682 [-] 4 error processing claimed jobs: could not query for claimed jobs: select-running/get-claimed-jobs: context canceled import_processor_test.go:747: -- test log scope end -- test logs left over in: /go/src/github.com/cockroachdb/cockroach/artifacts/logTestCSVImportCanBeResumed385659822 testing.go:1038: race detected during execution of test --- FAIL: TestCSVImportCanBeResumed (1.85s) ``` <details><summary>More</summary><p> Parameters: - TAGS= - GOFLAGS=-race -parallel=4 ``` make stressrace TESTS=TestCSVImportCanBeResumed PKG=./pkg/ccl/importccl TESTTIMEOUT=5m STRESSFLAGS='-timeout 5m' 2>&1 ``` [See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2ATestCSVImportCanBeResumed.%2A&sort=title&restgroup=false&display=lastcommented+project) <sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
1.0
ccl/importccl: TestCSVImportCanBeResumed failed - [(ccl/importccl).TestCSVImportCanBeResumed failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2707940&tab=buildLog) on [master@2572200f7612c6508a52735a6a18767cfb7cc09d](https://github.com/cockroachdb/cockroach/commits/2572200f7612c6508a52735a6a18767cfb7cc09d): ``` github.com/cockroachdb/cockroach/pkg/ccl/importccl.(*readImportDataProcessor).Start() /go/src/github.com/cockroachdb/cockroach/pkg/ccl/importccl/import_processor.go:105 +0xa4 github.com/cockroachdb/cockroach/pkg/sql/execinfra.(*ProcessorBase).Run() /go/src/github.com/cockroachdb/cockroach/pkg/sql/execinfra/processorsbase.go:764 +0x8e github.com/cockroachdb/cockroach/pkg/ccl/importccl.(*readImportDataProcessor).Run() <autogenerated>:1 +0x57 github.com/cockroachdb/cockroach/pkg/sql/flowinfra.(*FlowBase).Run() /go/src/github.com/cockroachdb/cockroach/pkg/sql/flowinfra/flow.go:382 +0x2c6 github.com/cockroachdb/cockroach/pkg/sql/rowflow.(*rowBasedFlow).Run() <autogenerated>:1 +0x84 github.com/cockroachdb/cockroach/pkg/sql.(*DistSQLPlanner).Run() /go/src/github.com/cockroachdb/cockroach/pkg/sql/distsql_running.go:367 +0x7a7 github.com/cockroachdb/cockroach/pkg/sql.DistIngest.func6() /go/src/github.com/cockroachdb/cockroach/pkg/sql/distsql_plan_csv.go:332 +0x1b2 github.com/cockroachdb/cockroach/pkg/util/ctxgroup.Group.GoCtx.func1() /go/src/github.com/cockroachdb/cockroach/pkg/util/ctxgroup/ctxgroup.go:166 +0x4e golang.org/x/sync/errgroup.(*Group).Go.func1() /go/src/github.com/cockroachdb/cockroach/vendor/golang.org/x/sync/errgroup/errgroup.go:57 +0x94 Goroutine 26 (running) created at: golang.org/x/sync/errgroup.(*Group).Go() /go/src/github.com/cockroachdb/cockroach/vendor/golang.org/x/sync/errgroup/errgroup.go:54 +0x73 github.com/cockroachdb/cockroach/pkg/util/ctxgroup.Group.GoCtx() /go/src/github.com/cockroachdb/cockroach/pkg/util/ctxgroup/ctxgroup.go:165 +0xe4 github.com/cockroachdb/cockroach/pkg/sql.DistIngest() /go/src/github.com/cockroachdb/cockroach/pkg/sql/distsql_plan_csv.go:328 +0x128d github.com/cockroachdb/cockroach/pkg/ccl/importccl.(*importResumer).Resume() /go/src/github.com/cockroachdb/cockroach/pkg/ccl/importccl/import_stmt.go:1632 +0xe38 github.com/cockroachdb/cockroach/pkg/ccl/importccl.(*cancellableImportResumer).Resume() /go/src/github.com/cockroachdb/cockroach/pkg/ccl/importccl/import_processor_test.go:554 +0x1cc github.com/cockroachdb/cockroach/pkg/jobs.(*Registry).stepThroughStateMachine.func1() /go/src/github.com/cockroachdb/cockroach/pkg/jobs/registry.go:1170 +0x141 github.com/cockroachdb/cockroach/pkg/jobs.(*Registry).stepThroughStateMachine() /go/src/github.com/cockroachdb/cockroach/pkg/jobs/registry.go:1171 +0xa85 github.com/cockroachdb/cockroach/pkg/jobs.(*Registry).runJob() /go/src/github.com/cockroachdb/cockroach/pkg/jobs/adopt.go:250 +0x43e github.com/cockroachdb/cockroach/pkg/jobs.(*StartableJob).Start.func3() /go/src/github.com/cockroachdb/cockroach/pkg/jobs/jobs.go:867 +0x25c github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask.func1() /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:351 +0x149 ================== E210223 06:16:33.560091 1505 jobs/adopt.go:255 [n1] 1 job 635668376058003457: adoption completed with error job 635668376058003457: could not mark as reverting: context canceled: job with status pause-requested cannot be reverted import_processor_test.go:733: Resume pos: 30 E210223 06:16:34.408378 635 jobs/registry.go:672 [-] 2 error claiming jobs: could not query jobs table: claim-jobs: context canceled E210223 06:16:34.408913 633 jobs/registry.go:662 [-] 3 failed to serve pause and cancel requests: could not query jobs table: cancel/pause-requested: context canceled E210223 06:16:34.410822 635 jobs/registry.go:682 [-] 4 error processing claimed jobs: could not query for claimed jobs: select-running/get-claimed-jobs: context canceled import_processor_test.go:747: -- test log scope end -- test logs left over in: /go/src/github.com/cockroachdb/cockroach/artifacts/logTestCSVImportCanBeResumed385659822 testing.go:1038: race detected during execution of test --- FAIL: TestCSVImportCanBeResumed (1.85s) ``` <details><summary>More</summary><p> Parameters: - TAGS= - GOFLAGS=-race -parallel=4 ``` make stressrace TESTS=TestCSVImportCanBeResumed PKG=./pkg/ccl/importccl TESTTIMEOUT=5m STRESSFLAGS='-timeout 5m' 2>&1 ``` [See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2ATestCSVImportCanBeResumed.%2A&sort=title&restgroup=false&display=lastcommented+project) <sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
test
ccl importccl testcsvimportcanberesumed failed on github com cockroachdb cockroach pkg ccl importccl readimportdataprocessor start go src github com cockroachdb cockroach pkg ccl importccl import processor go github com cockroachdb cockroach pkg sql execinfra processorbase run go src github com cockroachdb cockroach pkg sql execinfra processorsbase go github com cockroachdb cockroach pkg ccl importccl readimportdataprocessor run github com cockroachdb cockroach pkg sql flowinfra flowbase run go src github com cockroachdb cockroach pkg sql flowinfra flow go github com cockroachdb cockroach pkg sql rowflow rowbasedflow run github com cockroachdb cockroach pkg sql distsqlplanner run go src github com cockroachdb cockroach pkg sql distsql running go github com cockroachdb cockroach pkg sql distingest go src github com cockroachdb cockroach pkg sql distsql plan csv go github com cockroachdb cockroach pkg util ctxgroup group goctx go src github com cockroachdb cockroach pkg util ctxgroup ctxgroup go golang org x sync errgroup group go go src github com cockroachdb cockroach vendor golang org x sync errgroup errgroup go goroutine running created at golang org x sync errgroup group go go src github com cockroachdb cockroach vendor golang org x sync errgroup errgroup go github com cockroachdb cockroach pkg util ctxgroup group goctx go src github com cockroachdb cockroach pkg util ctxgroup ctxgroup go github com cockroachdb cockroach pkg sql distingest go src github com cockroachdb cockroach pkg sql distsql plan csv go github com cockroachdb cockroach pkg ccl importccl importresumer resume go src github com cockroachdb cockroach pkg ccl importccl import stmt go github com cockroachdb cockroach pkg ccl importccl cancellableimportresumer resume go src github com cockroachdb cockroach pkg ccl importccl import processor test go github com cockroachdb cockroach pkg jobs registry stepthroughstatemachine go src github com cockroachdb cockroach pkg jobs registry go github com cockroachdb cockroach pkg jobs registry stepthroughstatemachine go src github com cockroachdb cockroach pkg jobs registry go github com cockroachdb cockroach pkg jobs registry runjob go src github com cockroachdb cockroach pkg jobs adopt go github com cockroachdb cockroach pkg jobs startablejob start go src github com cockroachdb cockroach pkg jobs jobs go github com cockroachdb cockroach pkg util stop stopper runasynctask go src github com cockroachdb cockroach pkg util stop stopper go jobs adopt go job adoption completed with error job could not mark as reverting context canceled job with status pause requested cannot be reverted import processor test go resume pos jobs registry go error claiming jobs could not query jobs table claim jobs context canceled jobs registry go failed to serve pause and cancel requests could not query jobs table cancel pause requested context canceled jobs registry go error processing claimed jobs could not query for claimed jobs select running get claimed jobs context canceled import processor test go test log scope end test logs left over in go src github com cockroachdb cockroach artifacts testing go race detected during execution of test fail testcsvimportcanberesumed more parameters tags goflags race parallel make stressrace tests testcsvimportcanberesumed pkg pkg ccl importccl testtimeout stressflags timeout powered by
1
9,156
2,607,928,336
IssuesEvent
2015-02-26 00:25:40
chrsmithdemos/minify
https://api.github.com/repos/chrsmithdemos/minify
closed
Move URI rewriting from Minify_CSS
auto-migrated Priority-Medium Release-2.1.1 Type-Enhancement
``` Олег Черній has contributed some code dealing with URI rewriting and I think it'd be a good goal to remove the rewriting functionality from Minify_CSS since rewriting is not a goal of minification. UriRewriter.php will have methods like "fileToRootRelative" and "rootToFileRelative" and a new CSS minifier will be composed from Minify_CSS and UriRewriter to serve most of Minify::serve's needs, just as Minify_Javascript wraps JSMin. ``` ----- Original issue reported on code.google.com by `mrclay....@gmail.com` on 3 Oct 2008 at 8:33
1.0
Move URI rewriting from Minify_CSS - ``` Олег Черній has contributed some code dealing with URI rewriting and I think it'd be a good goal to remove the rewriting functionality from Minify_CSS since rewriting is not a goal of minification. UriRewriter.php will have methods like "fileToRootRelative" and "rootToFileRelative" and a new CSS minifier will be composed from Minify_CSS and UriRewriter to serve most of Minify::serve's needs, just as Minify_Javascript wraps JSMin. ``` ----- Original issue reported on code.google.com by `mrclay....@gmail.com` on 3 Oct 2008 at 8:33
non_test
move uri rewriting from minify css олег черній has contributed some code dealing with uri rewriting and i think it d be a good goal to remove the rewriting functionality from minify css since rewriting is not a goal of minification urirewriter php will have methods like filetorootrelative and roottofilerelative and a new css minifier will be composed from minify css and urirewriter to serve most of minify serve s needs just as minify javascript wraps jsmin original issue reported on code google com by mrclay gmail com on oct at
0
17,735
10,762,594,838
IssuesEvent
2019-11-01 00:16:20
Azure/azure-sdk-for-js
https://api.github.com/repos/Azure/azure-sdk-for-js
closed
[Service Bus] Browser tests are failing on CI
Client Service Bus
Some of the browser tests are failing on live test runs on CI. This issue is to track investigation and related fix for it.
1.0
[Service Bus] Browser tests are failing on CI - Some of the browser tests are failing on live test runs on CI. This issue is to track investigation and related fix for it.
non_test
browser tests are failing on ci some of the browser tests are failing on live test runs on ci this issue is to track investigation and related fix for it
0
121,508
10,170,955,952
IssuesEvent
2019-08-08 07:08:12
cseeger-epages/mail2most
https://api.github.com/repos/cseeger-epages/mail2most
closed
invalid e-mail content stops mail2most
add-unittests bug
**Describe the bug** Having the option `SubjectOnly = false`, `mail2most` does not post any e-mail. Instead, the log just says `invalid message,` and `mail2most` stops. Unfortunately, there is no more information, although I set the log to `debug`. Setting `SubjectOnly = true` posts all e-mails without any error. Setting `SubjectOnly = false` back again posts new e-mails with all contents just fine. I reckon, one of the e-mails contains some invalid content but I don't know which one (so I don't know what's invalid). **To Reproduce** Steps to reproduce the behavior: 1. Send an e-mail with invalid content to your watched e-mail account. **Expected behavior** All e-mails are posted completely, invalid content is ignored instead of stopping `mail2most`. **Release version** v1.1.0, arm version. **Additional context** I'm running mattermost on a raspberry pi, with [these unofficial releases](https://github.com/SmartHoneybee/ubiquitous-memory).
1.0
invalid e-mail content stops mail2most - **Describe the bug** Having the option `SubjectOnly = false`, `mail2most` does not post any e-mail. Instead, the log just says `invalid message,` and `mail2most` stops. Unfortunately, there is no more information, although I set the log to `debug`. Setting `SubjectOnly = true` posts all e-mails without any error. Setting `SubjectOnly = false` back again posts new e-mails with all contents just fine. I reckon, one of the e-mails contains some invalid content but I don't know which one (so I don't know what's invalid). **To Reproduce** Steps to reproduce the behavior: 1. Send an e-mail with invalid content to your watched e-mail account. **Expected behavior** All e-mails are posted completely, invalid content is ignored instead of stopping `mail2most`. **Release version** v1.1.0, arm version. **Additional context** I'm running mattermost on a raspberry pi, with [these unofficial releases](https://github.com/SmartHoneybee/ubiquitous-memory).
test
invalid e mail content stops describe the bug having the option subjectonly false does not post any e mail instead the log just says invalid message and stops unfortunately there is no more information although i set the log to debug setting subjectonly true posts all e mails without any error setting subjectonly false back again posts new e mails with all contents just fine i reckon one of the e mails contains some invalid content but i don t know which one so i don t know what s invalid to reproduce steps to reproduce the behavior send an e mail with invalid content to your watched e mail account expected behavior all e mails are posted completely invalid content is ignored instead of stopping release version arm version additional context i m running mattermost on a raspberry pi with
1
119,522
10,057,122,288
IssuesEvent
2019-07-22 10:47:38
scylladb/scylla
https://api.github.com/repos/scylladb/scylla
closed
sporadic failures in TestMigrationUpgradeSSTables dtests
bug dtest
This is Scylla's bug tracker, to be used for reporting bugs only. If you have a question about Scylla, and not a bug, please ask it in our mailing-list at scylladb-dev@googlegroups.com or in our slack channel. - [X] I have read the disclaimer above, and I am reporting a suspected malfunction in Scylla. *Installation details* Scylla version (or git commit hash): c987397e529a6f3838da37199ca277a48cda4634 Cluster size: 1 Several similar dtest cases expect only sstables version mc after `nodetool upgradesstable` and they fail when seeing also sstables from an older version that were uploaded to the node. For example: [dtest-release/166/testReport/junit/migration_test/TestMigrationUpgradeSSTables_with_2_2_x/migrate_sstable_with_collection_map_test](https://jenkins.scylladb.com/view/master/job/scylla-master/job/dtest-release/166/testReport/junit/migration_test/TestMigrationUpgradeSSTables_with_2_2_x/migrate_sstable_with_collection_map_test/) [dtest-release/167/testReport/junit/migration_test/TestMigrationUpgradeSSTables_with_2_2_x/migrate_sstable_with_compact_storage_and_composite_key_test](https://jenkins.scylladb.com/view/master/job/scylla-master/job/dtest-release/167/testReport/junit/migration_test/TestMigrationUpgradeSSTables_with_2_2_x/migrate_sstable_with_compact_storage_and_composite_key_test/) [dtest-release/169/testReport/junit/migration_test/TestMigrationUpgradeSSTables_with_2_2_x/migrate_sstable_with_static_cell_test](https://jenkins.scylladb.com/view/master/job/scylla-master/job/dtest-release/169/testReport/junit/migration_test/TestMigrationUpgradeSSTables_with_2_2_x/migrate_sstable_with_static_cell_test/) [dtest-release/170/testReport/junit/migration_test/TestMigrationUpgradeSSTables_with_2_1_x/migrate_sstable_with_lz4_compression_test](https://jenkins.scylladb.com/view/master/job/scylla-master/job/dtest-release/170/testReport/junit/migration_test/TestMigrationUpgradeSSTables_with_2_1_x/migrate_sstable_with_lz4_compression_test/) The test looks correct in the sense that it checks the data directory right after `nodetool upgradesstables` completes as the latter should be synchronous and return after compaction upgraded the loaded sstables and deleted the old ones.
1.0
sporadic failures in TestMigrationUpgradeSSTables dtests - This is Scylla's bug tracker, to be used for reporting bugs only. If you have a question about Scylla, and not a bug, please ask it in our mailing-list at scylladb-dev@googlegroups.com or in our slack channel. - [X] I have read the disclaimer above, and I am reporting a suspected malfunction in Scylla. *Installation details* Scylla version (or git commit hash): c987397e529a6f3838da37199ca277a48cda4634 Cluster size: 1 Several similar dtest cases expect only sstables version mc after `nodetool upgradesstable` and they fail when seeing also sstables from an older version that were uploaded to the node. For example: [dtest-release/166/testReport/junit/migration_test/TestMigrationUpgradeSSTables_with_2_2_x/migrate_sstable_with_collection_map_test](https://jenkins.scylladb.com/view/master/job/scylla-master/job/dtest-release/166/testReport/junit/migration_test/TestMigrationUpgradeSSTables_with_2_2_x/migrate_sstable_with_collection_map_test/) [dtest-release/167/testReport/junit/migration_test/TestMigrationUpgradeSSTables_with_2_2_x/migrate_sstable_with_compact_storage_and_composite_key_test](https://jenkins.scylladb.com/view/master/job/scylla-master/job/dtest-release/167/testReport/junit/migration_test/TestMigrationUpgradeSSTables_with_2_2_x/migrate_sstable_with_compact_storage_and_composite_key_test/) [dtest-release/169/testReport/junit/migration_test/TestMigrationUpgradeSSTables_with_2_2_x/migrate_sstable_with_static_cell_test](https://jenkins.scylladb.com/view/master/job/scylla-master/job/dtest-release/169/testReport/junit/migration_test/TestMigrationUpgradeSSTables_with_2_2_x/migrate_sstable_with_static_cell_test/) [dtest-release/170/testReport/junit/migration_test/TestMigrationUpgradeSSTables_with_2_1_x/migrate_sstable_with_lz4_compression_test](https://jenkins.scylladb.com/view/master/job/scylla-master/job/dtest-release/170/testReport/junit/migration_test/TestMigrationUpgradeSSTables_with_2_1_x/migrate_sstable_with_lz4_compression_test/) The test looks correct in the sense that it checks the data directory right after `nodetool upgradesstables` completes as the latter should be synchronous and return after compaction upgraded the loaded sstables and deleted the old ones.
test
sporadic failures in testmigrationupgradesstables dtests this is scylla s bug tracker to be used for reporting bugs only if you have a question about scylla and not a bug please ask it in our mailing list at scylladb dev googlegroups com or in our slack channel i have read the disclaimer above and i am reporting a suspected malfunction in scylla installation details scylla version or git commit hash cluster size several similar dtest cases expect only sstables version mc after nodetool upgradesstable and they fail when seeing also sstables from an older version that were uploaded to the node for example the test looks correct in the sense that it checks the data directory right after nodetool upgradesstables completes as the latter should be synchronous and return after compaction upgraded the loaded sstables and deleted the old ones
1
328,742
28,134,171,340
IssuesEvent
2023-04-01 07:02:04
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
closed
Fix general_functions.test_tensorflow_pad
TensorFlow Frontend Sub Task Failing Test
| | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4580338848/jobs/8088986022" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/4580338848/jobs/8088986022" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4580338848/jobs/8088986022" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/4580338848/jobs/8088986022" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
1.0
Fix general_functions.test_tensorflow_pad - | | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4580338848/jobs/8088986022" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/4580338848/jobs/8088986022" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4580338848/jobs/8088986022" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/4580338848/jobs/8088986022" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
test
fix general functions test tensorflow pad tensorflow img src torch img src numpy img src jax img src
1
105,816
9,101,237,306
IssuesEvent
2019-02-20 10:34:40
banzaicloud/bank-vaults
https://api.github.com/repos/banzaicloud/bank-vaults
opened
Operator acceptance test fails occasionally
bug operator testing
**Describe the bug** ``` Failing command: Operator acceptance test Exit code: 1 Output: .go:128\ngithub.com/banzaicloud/bank-vaults/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:215\ngithub.com/banzaicloud/bank-vaults/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func1\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:158\ngithub.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil.func1\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:133\ngithub.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:134\ngithub.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait.Until\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:88"} {"level":"info","ts":1550658627.9762602,"logger":"controller_vault","msg":"Reconciling Vault","Request.Namespace":"default","Request.Name":"vault"} {"level":"error","ts":1550658628.810033,"logger":"kubebuilder.controller","msg":"Reconciler error","controller":"vault-controller","request":"default/vault","error":"failed to create/update etcd cluster: etcdclusters.etcd.database.coreos.com \"etcd-cluster\" is invalid: metadata.resourceVersion: Invalid value: 0x0: must be specified for an update","stacktrace":"github.com/banzaicloud/bank-vaults/vendor/github.com/go-logr/zapr.(*zapLogger).Error\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/github.com/go-logr/zapr/zapr.go:128\ngithub.com/banzaicloud/bank-vaults/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:215\ngithub.com/banzaicloud/bank-vaults/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func1\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:158\ngithub.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil.func1\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:133\ngithub.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:134\ngithub.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait.Until\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:88"} {"level":"info","ts":1550658631.370171,"logger":"controller_vault","msg":"Reconciling Vault","Request.Namespace":"default","Request.Name":"vault"} {"level":"error","ts":1550658632.1957633,"logger":"kubebuilder.controller","msg":"Reconciler error","controller":"vault-controller","request":"default/vault","error":"failed to create/update etcd cluster: etcdclusters.etcd.database.coreos.com \"etcd-cluster\" is invalid: metadata.resourceVersion: Invalid value: 0x0: must be specified for an update","stacktrace":"github.com/banzaicloud/bank-vaults/vendor/github.com/go-logr/zapr.(*zapLogger).Error\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/github.com/go-logr/zapr/zapr.go:128\ngithub.com/banzaicloud/bank-vaults/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:215\ngithub.com/banzaicloud/bank-vaults/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func1\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:158\ngithub.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil.func1\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:133\ngithub.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:134\ngithub.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait.Until\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:88"} {"level":"info","ts":1550658637.315934,"logger":"controller_vault","msg":"Reconciling Vault","Request.Namespace":"default","Request.Name":"vault"} {"level":"error","ts":1550658638.5753064,"logger":"kubebuilder.controller","msg":"Reconciler error","controller":"vault-controller","request":"default/vault","error":"failed to create/update etcd cluster: etcdclusters.etcd.database.coreos.com \"etcd-cluster\" is invalid: metadata.resourceVersion: Invalid value: 0x0: must be specified for an update","stacktrace":"github.com/banzaicloud/bank-vaults/vendor/github.com/go-logr/zapr.(*zapLogger).Error\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/github.com/go-logr/zapr/zapr.go:128\ngithub.com/banzaicloud/bank-vaults/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:215\ngithub.com/banzaicloud/bank-vaults/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func1\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:158\ngithub.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil.func1\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:133\ngithub.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:134\ngithub.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait.Until\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:88"} {"level":"info","ts":1550658645.1819189,"logger":"controller_vault","msg":"Reconciling Vault","Request.Namespace":"default","Request.Name":"vault"} {"level":"error","ts":1550658646.418675,"logger":"kubebuilder.controller","msg":"Reconciler error","controller":"vault-controller","request":"default/vault","error":"failed to create/update etcd cluster: etcdclusters.etcd.database.coreos.com \"etcd-cluster\" is invalid: metadata.resourceVersion: Invalid value: 0x0: must be specified for an update","stacktrace":"github.com/banzaicloud/bank-vaults/vendor/github.com/go-logr/zapr.(*zapLogger).Error\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/github.com/go-logr/zapr/zapr.go:128\ngithub.com/banzaicloud/bank-vaults/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:215\ngithub.com/banzaicloud/bank-vaults/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func1\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:158\ngithub.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil.func1\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:133\ngithub.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:134\ngithub.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait.Until\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:88"} {"level":"info","ts":1550658648.81547,"logger":"controller_vault","msg":"Reconciling Vault","Request.Namespace":"default","Request.Name":"vault"} {"level":"error","ts":1550658649.6408687,"logger":"kubebuilder.controller","msg":"Reconciler error","controller":"vault-controller","request":"default/vault","error":"failed to create/update etcd cluster: etcdclusters.etcd.database.coreos.com \"etcd-cluster\" is invalid: metadata.resourceVersion: Invalid value: 0x0: must be specified for an update","stacktrace":"github.com/banzaicloud/bank-vaults/vendor/github.com/go-logr/zapr.(*zapLogger).Error\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/github.com/go-logr/zapr/zapr.go:128\ngithub.com/banzaicloud/bank-vaults/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:215\ngithub.com/banzaicloud/bank-vaults/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func1\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:158\ngithub.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil.func1\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:133\ngithub.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:134\ngithub.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait.Until\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:88"} + kubectl describe pod vault-0 vault-1 Error from server (NotFound): pods "vault-0" not found Exited with code 1 ``` **To Reproduce** Run CircleCI acceptance test. **Expected behavior** This shouldn't happen.
1.0
Operator acceptance test fails occasionally - **Describe the bug** ``` Failing command: Operator acceptance test Exit code: 1 Output: .go:128\ngithub.com/banzaicloud/bank-vaults/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:215\ngithub.com/banzaicloud/bank-vaults/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func1\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:158\ngithub.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil.func1\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:133\ngithub.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:134\ngithub.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait.Until\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:88"} {"level":"info","ts":1550658627.9762602,"logger":"controller_vault","msg":"Reconciling Vault","Request.Namespace":"default","Request.Name":"vault"} {"level":"error","ts":1550658628.810033,"logger":"kubebuilder.controller","msg":"Reconciler error","controller":"vault-controller","request":"default/vault","error":"failed to create/update etcd cluster: etcdclusters.etcd.database.coreos.com \"etcd-cluster\" is invalid: metadata.resourceVersion: Invalid value: 0x0: must be specified for an update","stacktrace":"github.com/banzaicloud/bank-vaults/vendor/github.com/go-logr/zapr.(*zapLogger).Error\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/github.com/go-logr/zapr/zapr.go:128\ngithub.com/banzaicloud/bank-vaults/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:215\ngithub.com/banzaicloud/bank-vaults/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func1\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:158\ngithub.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil.func1\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:133\ngithub.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:134\ngithub.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait.Until\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:88"} {"level":"info","ts":1550658631.370171,"logger":"controller_vault","msg":"Reconciling Vault","Request.Namespace":"default","Request.Name":"vault"} {"level":"error","ts":1550658632.1957633,"logger":"kubebuilder.controller","msg":"Reconciler error","controller":"vault-controller","request":"default/vault","error":"failed to create/update etcd cluster: etcdclusters.etcd.database.coreos.com \"etcd-cluster\" is invalid: metadata.resourceVersion: Invalid value: 0x0: must be specified for an update","stacktrace":"github.com/banzaicloud/bank-vaults/vendor/github.com/go-logr/zapr.(*zapLogger).Error\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/github.com/go-logr/zapr/zapr.go:128\ngithub.com/banzaicloud/bank-vaults/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:215\ngithub.com/banzaicloud/bank-vaults/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func1\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:158\ngithub.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil.func1\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:133\ngithub.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:134\ngithub.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait.Until\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:88"} {"level":"info","ts":1550658637.315934,"logger":"controller_vault","msg":"Reconciling Vault","Request.Namespace":"default","Request.Name":"vault"} {"level":"error","ts":1550658638.5753064,"logger":"kubebuilder.controller","msg":"Reconciler error","controller":"vault-controller","request":"default/vault","error":"failed to create/update etcd cluster: etcdclusters.etcd.database.coreos.com \"etcd-cluster\" is invalid: metadata.resourceVersion: Invalid value: 0x0: must be specified for an update","stacktrace":"github.com/banzaicloud/bank-vaults/vendor/github.com/go-logr/zapr.(*zapLogger).Error\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/github.com/go-logr/zapr/zapr.go:128\ngithub.com/banzaicloud/bank-vaults/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:215\ngithub.com/banzaicloud/bank-vaults/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func1\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:158\ngithub.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil.func1\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:133\ngithub.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:134\ngithub.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait.Until\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:88"} {"level":"info","ts":1550658645.1819189,"logger":"controller_vault","msg":"Reconciling Vault","Request.Namespace":"default","Request.Name":"vault"} {"level":"error","ts":1550658646.418675,"logger":"kubebuilder.controller","msg":"Reconciler error","controller":"vault-controller","request":"default/vault","error":"failed to create/update etcd cluster: etcdclusters.etcd.database.coreos.com \"etcd-cluster\" is invalid: metadata.resourceVersion: Invalid value: 0x0: must be specified for an update","stacktrace":"github.com/banzaicloud/bank-vaults/vendor/github.com/go-logr/zapr.(*zapLogger).Error\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/github.com/go-logr/zapr/zapr.go:128\ngithub.com/banzaicloud/bank-vaults/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:215\ngithub.com/banzaicloud/bank-vaults/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func1\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:158\ngithub.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil.func1\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:133\ngithub.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:134\ngithub.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait.Until\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:88"} {"level":"info","ts":1550658648.81547,"logger":"controller_vault","msg":"Reconciling Vault","Request.Namespace":"default","Request.Name":"vault"} {"level":"error","ts":1550658649.6408687,"logger":"kubebuilder.controller","msg":"Reconciler error","controller":"vault-controller","request":"default/vault","error":"failed to create/update etcd cluster: etcdclusters.etcd.database.coreos.com \"etcd-cluster\" is invalid: metadata.resourceVersion: Invalid value: 0x0: must be specified for an update","stacktrace":"github.com/banzaicloud/bank-vaults/vendor/github.com/go-logr/zapr.(*zapLogger).Error\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/github.com/go-logr/zapr/zapr.go:128\ngithub.com/banzaicloud/bank-vaults/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:215\ngithub.com/banzaicloud/bank-vaults/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func1\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:158\ngithub.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil.func1\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:133\ngithub.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait.JitterUntil\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:134\ngithub.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait.Until\n\t/go/src/github.com/banzaicloud/bank-vaults/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:88"} + kubectl describe pod vault-0 vault-1 Error from server (NotFound): pods "vault-0" not found Exited with code 1 ``` **To Reproduce** Run CircleCI acceptance test. **Expected behavior** This shouldn't happen.
test
operator acceptance test fails occasionally describe the bug failing command operator acceptance test exit code output go ngithub com banzaicloud bank vaults vendor sigs io controller runtime pkg internal controller controller processnextworkitem n t go src github com banzaicloud bank vaults vendor sigs io controller runtime pkg internal controller controller go ngithub com banzaicloud bank vaults vendor sigs io controller runtime pkg internal controller controller start n t go src github com banzaicloud bank vaults vendor sigs io controller runtime pkg internal controller controller go ngithub com banzaicloud bank vaults vendor io apimachinery pkg util wait jitteruntil n t go src github com banzaicloud bank vaults vendor io apimachinery pkg util wait wait go ngithub com banzaicloud bank vaults vendor io apimachinery pkg util wait jitteruntil n t go src github com banzaicloud bank vaults vendor io apimachinery pkg util wait wait go ngithub com banzaicloud bank vaults vendor io apimachinery pkg util wait until n t go src github com banzaicloud bank vaults vendor io apimachinery pkg util wait wait go level info ts logger controller vault msg reconciling vault request namespace default request name vault level error ts logger kubebuilder controller msg reconciler error controller vault controller request default vault error failed to create update etcd cluster etcdclusters etcd database coreos com etcd cluster is invalid metadata resourceversion invalid value must be specified for an update stacktrace github com banzaicloud bank vaults vendor github com go logr zapr zaplogger error n t go src github com banzaicloud bank vaults vendor github com go logr zapr zapr go ngithub com banzaicloud bank vaults vendor sigs io controller runtime pkg internal controller controller processnextworkitem n t go src github com banzaicloud bank vaults vendor sigs io controller runtime pkg internal controller controller go ngithub com banzaicloud bank vaults vendor sigs io controller runtime pkg internal controller controller start n t go src github com banzaicloud bank vaults vendor sigs io controller runtime pkg internal controller controller go ngithub com banzaicloud bank vaults vendor io apimachinery pkg util wait jitteruntil n t go src github com banzaicloud bank vaults vendor io apimachinery pkg util wait wait go ngithub com banzaicloud bank vaults vendor io apimachinery pkg util wait jitteruntil n t go src github com banzaicloud bank vaults vendor io apimachinery pkg util wait wait go ngithub com banzaicloud bank vaults vendor io apimachinery pkg util wait until n t go src github com banzaicloud bank vaults vendor io apimachinery pkg util wait wait go level info ts logger controller vault msg reconciling vault request namespace default request name vault level error ts logger kubebuilder controller msg reconciler error controller vault controller request default vault error failed to create update etcd cluster etcdclusters etcd database coreos com etcd cluster is invalid metadata resourceversion invalid value must be specified for an update stacktrace github com banzaicloud bank vaults vendor github com go logr zapr zaplogger error n t go src github com banzaicloud bank vaults vendor github com go logr zapr zapr go ngithub com banzaicloud bank vaults vendor sigs io controller runtime pkg internal controller controller processnextworkitem n t go src github com banzaicloud bank vaults vendor sigs io controller runtime pkg internal controller controller go ngithub com banzaicloud bank vaults vendor sigs io controller runtime pkg internal controller controller start n t go src github com banzaicloud bank vaults vendor sigs io controller runtime pkg internal controller controller go ngithub com banzaicloud bank vaults vendor io apimachinery pkg util wait jitteruntil n t go src github com banzaicloud bank vaults vendor io apimachinery pkg util wait wait go ngithub com banzaicloud bank vaults vendor io apimachinery pkg util wait jitteruntil n t go src github com banzaicloud bank vaults vendor io apimachinery pkg util wait wait go ngithub com banzaicloud bank vaults vendor io apimachinery pkg util wait until n t go src github com banzaicloud bank vaults vendor io apimachinery pkg util wait wait go level info ts logger controller vault msg reconciling vault request namespace default request name vault level error ts logger kubebuilder controller msg reconciler error controller vault controller request default vault error failed to create update etcd cluster etcdclusters etcd database coreos com etcd cluster is invalid metadata resourceversion invalid value must be specified for an update stacktrace github com banzaicloud bank vaults vendor github com go logr zapr zaplogger error n t go src github com banzaicloud bank vaults vendor github com go logr zapr zapr go ngithub com banzaicloud bank vaults vendor sigs io controller runtime pkg internal controller controller processnextworkitem n t go src github com banzaicloud bank vaults vendor sigs io controller runtime pkg internal controller controller go ngithub com banzaicloud bank vaults vendor sigs io controller runtime pkg internal controller controller start n t go src github com banzaicloud bank vaults vendor sigs io controller runtime pkg internal controller controller go ngithub com banzaicloud bank vaults vendor io apimachinery pkg util wait jitteruntil n t go src github com banzaicloud bank vaults vendor io apimachinery pkg util wait wait go ngithub com banzaicloud bank vaults vendor io apimachinery pkg util wait jitteruntil n t go src github com banzaicloud bank vaults vendor io apimachinery pkg util wait wait go ngithub com banzaicloud bank vaults vendor io apimachinery pkg util wait until n t go src github com banzaicloud bank vaults vendor io apimachinery pkg util wait wait go level info ts logger controller vault msg reconciling vault request namespace default request name vault level error ts logger kubebuilder controller msg reconciler error controller vault controller request default vault error failed to create update etcd cluster etcdclusters etcd database coreos com etcd cluster is invalid metadata resourceversion invalid value must be specified for an update stacktrace github com banzaicloud bank vaults vendor github com go logr zapr zaplogger error n t go src github com banzaicloud bank vaults vendor github com go logr zapr zapr go ngithub com banzaicloud bank vaults vendor sigs io controller runtime pkg internal controller controller processnextworkitem n t go src github com banzaicloud bank vaults vendor sigs io controller runtime pkg internal controller controller go ngithub com banzaicloud bank vaults vendor sigs io controller runtime pkg internal controller controller start n t go src github com banzaicloud bank vaults vendor sigs io controller runtime pkg internal controller controller go ngithub com banzaicloud bank vaults vendor io apimachinery pkg util wait jitteruntil n t go src github com banzaicloud bank vaults vendor io apimachinery pkg util wait wait go ngithub com banzaicloud bank vaults vendor io apimachinery pkg util wait jitteruntil n t go src github com banzaicloud bank vaults vendor io apimachinery pkg util wait wait go ngithub com banzaicloud bank vaults vendor io apimachinery pkg util wait until n t go src github com banzaicloud bank vaults vendor io apimachinery pkg util wait wait go level info ts logger controller vault msg reconciling vault request namespace default request name vault level error ts logger kubebuilder controller msg reconciler error controller vault controller request default vault error failed to create update etcd cluster etcdclusters etcd database coreos com etcd cluster is invalid metadata resourceversion invalid value must be specified for an update stacktrace github com banzaicloud bank vaults vendor github com go logr zapr zaplogger error n t go src github com banzaicloud bank vaults vendor github com go logr zapr zapr go ngithub com banzaicloud bank vaults vendor sigs io controller runtime pkg internal controller controller processnextworkitem n t go src github com banzaicloud bank vaults vendor sigs io controller runtime pkg internal controller controller go ngithub com banzaicloud bank vaults vendor sigs io controller runtime pkg internal controller controller start n t go src github com banzaicloud bank vaults vendor sigs io controller runtime pkg internal controller controller go ngithub com banzaicloud bank vaults vendor io apimachinery pkg util wait jitteruntil n t go src github com banzaicloud bank vaults vendor io apimachinery pkg util wait wait go ngithub com banzaicloud bank vaults vendor io apimachinery pkg util wait jitteruntil n t go src github com banzaicloud bank vaults vendor io apimachinery pkg util wait wait go ngithub com banzaicloud bank vaults vendor io apimachinery pkg util wait until n t go src github com banzaicloud bank vaults vendor io apimachinery pkg util wait wait go kubectl describe pod vault vault error from server notfound pods vault not found exited with code to reproduce run circleci acceptance test expected behavior this shouldn t happen
1
42,621
5,446,556,183
IssuesEvent
2017-03-07 10:56:58
red/red
https://api.github.com/repos/red/red
closed
Command line argument inconsistency
Epic status.built status.tested type.review
I'd really like to start using Red to write useful programs, and for me, that means command line tools. Unfortunately, there's inconsistency with how command line arguments work at the moment that makes it difficult or painfully slow to test anything. First, here's what I expect, as demonstrated by the below Rebol 2 code: ```rebol Rebol [] print ["system/options/args:" mold system/options/args] print ["system/script/args:" mold system/script/args] ``` ```shell $ ./rebol -qws hello-args.r hi there 'double "quotes"' system/options/args: ["hi" "there" {double "quotes"}] system/script/args: {hi there double "quotes"} ``` So the expected value for `system/options/args` is a block of strings, while `system/script/args` is equivalent to `reform system/options/args`. Red's behavior is different than the above, and is even different between interpreted and compiled code. To demonstrate: ```red Red [] print ["system/options/args:" mold system/options/args] print ["system/script/args:" mold system/script/args] ``` When interpreted, `system/options/args` is a string containing the quoted name of the script file, while `system/script/args` is `none`: ```shell $ ./red-061 hello-args.red hi there 'double "quotes"' system/options/args: {"hello-args.red"} system/script/args: none ``` When compiled, `system/options/args` is a string that contains all the arguments surrounded by double quotes and `reform`ed without escaping double quotes that are passed to it, while `system/script/args` is still `none`: ```shell $ ./red-061 -c hello-args.red -=== Red Compiler 0.6.1 ===- Compiling /home/cromo/scratch/red/hello-args.red ... ...compilation time : 500 ms Compiling to native code... ...compilation time : 15568 ms ...linking time : 143 ms ...output file size : 455168 bytes ...output file : /home/cromo/scratch/red/hello-args $ ./hello-args hi there 'double "quotes"' system/options/args: {"hi" "there" "double "quotes""} system/script/args: none ``` The interpreted values are consistent with issues #1985 and #2242 - only the script name is passed in `system/options/args`. This issue includes those along with this inconsistency between the interpreted mode and the compiled mode. The main problem I have here is that interpreting scripts is quite speedy but gives the wrong inputs, while the compiled version takes 15+ seconds to compile, but gives a potentially workable value in `system/options/args`. I picked up Rebol in the last week or so and absolutely love the language family. I'd be happy to work on a pull request to provide a similar output to the Rebol 2 source above if you could point me in the right direction.
1.0
Command line argument inconsistency - I'd really like to start using Red to write useful programs, and for me, that means command line tools. Unfortunately, there's inconsistency with how command line arguments work at the moment that makes it difficult or painfully slow to test anything. First, here's what I expect, as demonstrated by the below Rebol 2 code: ```rebol Rebol [] print ["system/options/args:" mold system/options/args] print ["system/script/args:" mold system/script/args] ``` ```shell $ ./rebol -qws hello-args.r hi there 'double "quotes"' system/options/args: ["hi" "there" {double "quotes"}] system/script/args: {hi there double "quotes"} ``` So the expected value for `system/options/args` is a block of strings, while `system/script/args` is equivalent to `reform system/options/args`. Red's behavior is different than the above, and is even different between interpreted and compiled code. To demonstrate: ```red Red [] print ["system/options/args:" mold system/options/args] print ["system/script/args:" mold system/script/args] ``` When interpreted, `system/options/args` is a string containing the quoted name of the script file, while `system/script/args` is `none`: ```shell $ ./red-061 hello-args.red hi there 'double "quotes"' system/options/args: {"hello-args.red"} system/script/args: none ``` When compiled, `system/options/args` is a string that contains all the arguments surrounded by double quotes and `reform`ed without escaping double quotes that are passed to it, while `system/script/args` is still `none`: ```shell $ ./red-061 -c hello-args.red -=== Red Compiler 0.6.1 ===- Compiling /home/cromo/scratch/red/hello-args.red ... ...compilation time : 500 ms Compiling to native code... ...compilation time : 15568 ms ...linking time : 143 ms ...output file size : 455168 bytes ...output file : /home/cromo/scratch/red/hello-args $ ./hello-args hi there 'double "quotes"' system/options/args: {"hi" "there" "double "quotes""} system/script/args: none ``` The interpreted values are consistent with issues #1985 and #2242 - only the script name is passed in `system/options/args`. This issue includes those along with this inconsistency between the interpreted mode and the compiled mode. The main problem I have here is that interpreting scripts is quite speedy but gives the wrong inputs, while the compiled version takes 15+ seconds to compile, but gives a potentially workable value in `system/options/args`. I picked up Rebol in the last week or so and absolutely love the language family. I'd be happy to work on a pull request to provide a similar output to the Rebol 2 source above if you could point me in the right direction.
test
command line argument inconsistency i d really like to start using red to write useful programs and for me that means command line tools unfortunately there s inconsistency with how command line arguments work at the moment that makes it difficult or painfully slow to test anything first here s what i expect as demonstrated by the below rebol code rebol rebol print print shell rebol qws hello args r hi there double quotes system options args system script args hi there double quotes so the expected value for system options args is a block of strings while system script args is equivalent to reform system options args red s behavior is different than the above and is even different between interpreted and compiled code to demonstrate red red print print when interpreted system options args is a string containing the quoted name of the script file while system script args is none shell red hello args red hi there double quotes system options args hello args red system script args none when compiled system options args is a string that contains all the arguments surrounded by double quotes and reform ed without escaping double quotes that are passed to it while system script args is still none shell red c hello args red red compiler compiling home cromo scratch red hello args red compilation time ms compiling to native code compilation time ms linking time ms output file size bytes output file home cromo scratch red hello args hello args hi there double quotes system options args hi there double quotes system script args none the interpreted values are consistent with issues and only the script name is passed in system options args this issue includes those along with this inconsistency between the interpreted mode and the compiled mode the main problem i have here is that interpreting scripts is quite speedy but gives the wrong inputs while the compiled version takes seconds to compile but gives a potentially workable value in system options args i picked up rebol in the last week or so and absolutely love the language family i d be happy to work on a pull request to provide a similar output to the rebol source above if you could point me in the right direction
1
244,027
18,737,110,372
IssuesEvent
2021-11-04 09:09:00
AY2122S1-CS2103-F09-1/tp
https://api.github.com/repos/AY2122S1-CS2103-F09-1/tp
closed
Add in PUML for my part in DG
documentation
Look through DG, learn about what/how to write, make improvements
1.0
Add in PUML for my part in DG - Look through DG, learn about what/how to write, make improvements
non_test
add in puml for my part in dg look through dg learn about what how to write make improvements
0
60,365
6,688,978,958
IssuesEvent
2017-10-08 20:41:08
dhermes/bezier
https://api.github.com/repos/dhermes/bezier
opened
Images from doctests should be opt-in, not opt-out
hygiene testing
[Currently][1] it's necessary to set `NO_IMAGES=True` (or any value) for running doctests. This should be the opposite (i.e. `nox -s docs_images` should have to set an environment variable). (It's also worth looking into `pytest` support for doctests.) [1]: https://github.com/dhermes/bezier/blob/eaec7cfbba84cfb27247647be52286551df9e619/nox.py#L199
1.0
Images from doctests should be opt-in, not opt-out - [Currently][1] it's necessary to set `NO_IMAGES=True` (or any value) for running doctests. This should be the opposite (i.e. `nox -s docs_images` should have to set an environment variable). (It's also worth looking into `pytest` support for doctests.) [1]: https://github.com/dhermes/bezier/blob/eaec7cfbba84cfb27247647be52286551df9e619/nox.py#L199
test
images from doctests should be opt in not opt out it s necessary to set no images true or any value for running doctests this should be the opposite i e nox s docs images should have to set an environment variable it s also worth looking into pytest support for doctests
1
51,774
6,197,176,425
IssuesEvent
2017-07-05 16:12:46
phetsims/friction
https://api.github.com/repos/phetsims/friction
opened
Invalid value: Property{0}###
type:automated-testing type:bug
``` Error: Assertion failed: invalid value: Property{0}450.5 at window.assertions.assertFunction (https://bayes.colorado.edu/continuous-testing/snapshot-1499268192217/assert/js/assert.js:21:13) at NumberProperty.set (https://bayes.colorado.edu/continuous-testing/snapshot-1499268192217/axon/js/Property.js?bust=1499268992000:138:40) at FrictionModel.move (https://bayes.colorado.edu/continuous-testing/snapshot-1499268192217/friction/js/friction/model/FrictionModel.js?bust=1499268992000:240:35) at translate (https://bayes.colorado.edu/continuous-testing/snapshot-1499268192217/friction/js/friction/model/FrictionModel.js?bust=1499268992000:261:16) at Object.move (https://bayes.colorado.edu/continuous-testing/snapshot-1499268192217/scenery/js/input/SimpleDragHandler.js?bust=1499268992000:118:34) at Input.dispatchToPointer (https://bayes.colorado.edu/continuous-testing/snapshot-1499268192217/scenery/js/input/Input.js?bust=1499268992000:774:27) at Input.dispatchEvent (https://bayes.colorado.edu/continuous-testing/snapshot-1499268192217/scenery/js/input/Input.js?bust=1499268992000:747:12) at Input.branchChangeEvents (https://bayes.colorado.edu/continuous-testing/snapshot-1499268192217/scenery/js/input/Input.js?bust=1499268992000:685:14) at Input.moveEvent (https://bayes.colorado.edu/continuous-testing/snapshot-1499268192217/scenery/js/input/Input.js?bust=1499268992000:650:26) at Input.mouseMove (https://bayes.colorado.edu/continuous-testing/snapshot-1499268192217/scenery/js/input/Input.js?bust=1499268992000:351:12) Approximately 7/5/2017, 9:23:12 AM ``` Possibly related, you can just drag the Chemistry book wherever you want. Here I dragged it down past the Physics book and out of the magnification window: <img width="1450" alt="screen shot 2017-07-05 at 10 10 51 am" src="https://user-images.githubusercontent.com/5863899/27873712-55e35684-616a-11e7-928e-1d6c97a854d6.png">
1.0
Invalid value: Property{0}### - ``` Error: Assertion failed: invalid value: Property{0}450.5 at window.assertions.assertFunction (https://bayes.colorado.edu/continuous-testing/snapshot-1499268192217/assert/js/assert.js:21:13) at NumberProperty.set (https://bayes.colorado.edu/continuous-testing/snapshot-1499268192217/axon/js/Property.js?bust=1499268992000:138:40) at FrictionModel.move (https://bayes.colorado.edu/continuous-testing/snapshot-1499268192217/friction/js/friction/model/FrictionModel.js?bust=1499268992000:240:35) at translate (https://bayes.colorado.edu/continuous-testing/snapshot-1499268192217/friction/js/friction/model/FrictionModel.js?bust=1499268992000:261:16) at Object.move (https://bayes.colorado.edu/continuous-testing/snapshot-1499268192217/scenery/js/input/SimpleDragHandler.js?bust=1499268992000:118:34) at Input.dispatchToPointer (https://bayes.colorado.edu/continuous-testing/snapshot-1499268192217/scenery/js/input/Input.js?bust=1499268992000:774:27) at Input.dispatchEvent (https://bayes.colorado.edu/continuous-testing/snapshot-1499268192217/scenery/js/input/Input.js?bust=1499268992000:747:12) at Input.branchChangeEvents (https://bayes.colorado.edu/continuous-testing/snapshot-1499268192217/scenery/js/input/Input.js?bust=1499268992000:685:14) at Input.moveEvent (https://bayes.colorado.edu/continuous-testing/snapshot-1499268192217/scenery/js/input/Input.js?bust=1499268992000:650:26) at Input.mouseMove (https://bayes.colorado.edu/continuous-testing/snapshot-1499268192217/scenery/js/input/Input.js?bust=1499268992000:351:12) Approximately 7/5/2017, 9:23:12 AM ``` Possibly related, you can just drag the Chemistry book wherever you want. Here I dragged it down past the Physics book and out of the magnification window: <img width="1450" alt="screen shot 2017-07-05 at 10 10 51 am" src="https://user-images.githubusercontent.com/5863899/27873712-55e35684-616a-11e7-928e-1d6c97a854d6.png">
test
invalid value property error assertion failed invalid value property at window assertions assertfunction at numberproperty set at frictionmodel move at translate at object move at input dispatchtopointer at input dispatchevent at input branchchangeevents at input moveevent at input mousemove approximately am possibly related you can just drag the chemistry book wherever you want here i dragged it down past the physics book and out of the magnification window img width alt screen shot at am src
1
515,944
14,972,369,529
IssuesEvent
2021-01-27 22:43:48
inverse-inc/packetfence
https://api.github.com/repos/inverse-inc/packetfence
closed
pfacct stores 00:00:00:00:00:00 MAC in DB when Calling-Station-ID is XXXX-XXXX-XXXX
Priority: Medium Type: Bug
**Describe the bug** If `pfacct` receive an Accounting Request with a Calling-Station-ID that uses following convention: XXXX-XXXX-XXXX, an entry is created in `bandwidth_accounting` table with `mac` field equals to `00:00:00:00:00:00` **To Reproduce** Steps to reproduce the behavior: 1. Forge an accounting request with specific Calling-Station-ID: ``` cat > /opt/acct << EOF NAS-Port-Type = 15 Acct-Input-Octets = 0 Acct-Session-Id = 1111111 NAS-IP-Address = 192.168.0.1 Called-Station-Id = "F4-E5-F2-FE-6B-61" Calling-Station-Id = "ecf4-bb37-ddaa" User-Name = "username@realm.com" Acct-Authentic = 2 Acct-Status-Type = 3 Acct-Output-Packets = 0 Acct-Output-Octets = 0 Event-Timestamp = "Feb 07 2020 10:30:00 CET" Acct-Input-Gigawords = 0 Acct-Input-Packets = 0 NAS-Port = 1 Acct-Delay-Time = 0 EOF cat /opt/acct | radclient -x 172.17.17.10:1813 acct `cat /usr/local/pf/conf/local_secret` ``` 2. `mysql pf -e "select * from bandwidth_accounting;` => Results: ``` +-----------------+----------------------+---------------------+-------------+----------+-----------+-------------------+-----------+---------------------+-------------+ | node_id | unique_session_id | time_bucket | source_type | in_bytes | out_bytes | mac | tenant_id | last_updated | total_bytes | +-----------------+----------------------+---------------------+-------------+----------+-----------+-------------------+-----------+---------------------+-------------+ | 281474976710656 | 14713129202345799727 | 2020-02-01 00:00:00 | radius | 0 | 0 | 00:00:00:00:00:00 | 1 | 2021-01-27 13:19:30 | 0 | +-----------------+----------------------+---------------------+-------------+----------+-----------+-------------------+-----------+---------------------+-------------+ ``` **Expected behavior** All format of MAC address should be supported. **Additional context** PacketFence is able to create an entry in cache because `pfcmd cache accounting dump MAC` returned accounting data.
1.0
pfacct stores 00:00:00:00:00:00 MAC in DB when Calling-Station-ID is XXXX-XXXX-XXXX - **Describe the bug** If `pfacct` receive an Accounting Request with a Calling-Station-ID that uses following convention: XXXX-XXXX-XXXX, an entry is created in `bandwidth_accounting` table with `mac` field equals to `00:00:00:00:00:00` **To Reproduce** Steps to reproduce the behavior: 1. Forge an accounting request with specific Calling-Station-ID: ``` cat > /opt/acct << EOF NAS-Port-Type = 15 Acct-Input-Octets = 0 Acct-Session-Id = 1111111 NAS-IP-Address = 192.168.0.1 Called-Station-Id = "F4-E5-F2-FE-6B-61" Calling-Station-Id = "ecf4-bb37-ddaa" User-Name = "username@realm.com" Acct-Authentic = 2 Acct-Status-Type = 3 Acct-Output-Packets = 0 Acct-Output-Octets = 0 Event-Timestamp = "Feb 07 2020 10:30:00 CET" Acct-Input-Gigawords = 0 Acct-Input-Packets = 0 NAS-Port = 1 Acct-Delay-Time = 0 EOF cat /opt/acct | radclient -x 172.17.17.10:1813 acct `cat /usr/local/pf/conf/local_secret` ``` 2. `mysql pf -e "select * from bandwidth_accounting;` => Results: ``` +-----------------+----------------------+---------------------+-------------+----------+-----------+-------------------+-----------+---------------------+-------------+ | node_id | unique_session_id | time_bucket | source_type | in_bytes | out_bytes | mac | tenant_id | last_updated | total_bytes | +-----------------+----------------------+---------------------+-------------+----------+-----------+-------------------+-----------+---------------------+-------------+ | 281474976710656 | 14713129202345799727 | 2020-02-01 00:00:00 | radius | 0 | 0 | 00:00:00:00:00:00 | 1 | 2021-01-27 13:19:30 | 0 | +-----------------+----------------------+---------------------+-------------+----------+-----------+-------------------+-----------+---------------------+-------------+ ``` **Expected behavior** All format of MAC address should be supported. **Additional context** PacketFence is able to create an entry in cache because `pfcmd cache accounting dump MAC` returned accounting data.
non_test
pfacct stores mac in db when calling station id is xxxx xxxx xxxx describe the bug if pfacct receive an accounting request with a calling station id that uses following convention xxxx xxxx xxxx an entry is created in bandwidth accounting table with mac field equals to to reproduce steps to reproduce the behavior forge an accounting request with specific calling station id cat opt acct eof nas port type acct input octets acct session id nas ip address called station id fe calling station id ddaa user name username realm com acct authentic acct status type acct output packets acct output octets event timestamp feb cet acct input gigawords acct input packets nas port acct delay time eof cat opt acct radclient x acct cat usr local pf conf local secret mysql pf e select from bandwidth accounting results node id unique session id time bucket source type in bytes out bytes mac tenant id last updated total bytes radius expected behavior all format of mac address should be supported additional context packetfence is able to create an entry in cache because pfcmd cache accounting dump mac returned accounting data
0
322,202
27,590,077,202
IssuesEvent
2023-03-08 23:18:05
nodejs/node
https://api.github.com/repos/nodejs/node
closed
Request: mark test runner stable in Node 20.0.0
feature request test_runner
### What is the problem this feature will solve? Having a stable test runner in core. People have been asking for it, and I feel the majority of the test runner is ready. ### What is the feature you are proposing to solve the problem? I would like to stabilize the test runner for the 20.0.0 release. This currently would not include code coverage, which is behind a separate flag. ### What alternatives have you considered? Not stabilizing the test runner.
1.0
Request: mark test runner stable in Node 20.0.0 - ### What is the problem this feature will solve? Having a stable test runner in core. People have been asking for it, and I feel the majority of the test runner is ready. ### What is the feature you are proposing to solve the problem? I would like to stabilize the test runner for the 20.0.0 release. This currently would not include code coverage, which is behind a separate flag. ### What alternatives have you considered? Not stabilizing the test runner.
test
request mark test runner stable in node what is the problem this feature will solve having a stable test runner in core people have been asking for it and i feel the majority of the test runner is ready what is the feature you are proposing to solve the problem i would like to stabilize the test runner for the release this currently would not include code coverage which is behind a separate flag what alternatives have you considered not stabilizing the test runner
1
11,431
3,490,221,348
IssuesEvent
2016-01-04 08:35:29
syl20bnr/spacemacs
https://api.github.com/repos/syl20bnr/spacemacs
closed
mu4e composing mode evilification
documentation :-> Email Key Bindings
I have just tried mu4e layer from develop branch. I liked it a lot, but it seems that keybindings in composing mode are still default i.e. non-evilified (`C-c C-s` is send for example). I noticed that `SPC m` bindings are completely missed. I am a recent vim convert so my elisp experience is around zero. Can you point me to some docs how to remap keybindings in spacemacs way?
1.0
mu4e composing mode evilification - I have just tried mu4e layer from develop branch. I liked it a lot, but it seems that keybindings in composing mode are still default i.e. non-evilified (`C-c C-s` is send for example). I noticed that `SPC m` bindings are completely missed. I am a recent vim convert so my elisp experience is around zero. Can you point me to some docs how to remap keybindings in spacemacs way?
non_test
composing mode evilification i have just tried layer from develop branch i liked it a lot but it seems that keybindings in composing mode are still default i e non evilified c c c s is send for example i noticed that spc m bindings are completely missed i am a recent vim convert so my elisp experience is around zero can you point me to some docs how to remap keybindings in spacemacs way
0
823,164
30,958,134,011
IssuesEvent
2023-08-08 00:01:27
space-wizards/space-station-14
https://api.github.com/repos/space-wizards/space-station-14
closed
Expeditions needs gamer loot removed
Priority: 2-Before Release Issue: Feature Request Difficulty: 1-Easy
dork and I are bikeshedding a better version that's less of a powergamer trap but in the interim remove the heals and the RPG and shit.
1.0
Expeditions needs gamer loot removed - dork and I are bikeshedding a better version that's less of a powergamer trap but in the interim remove the heals and the RPG and shit.
non_test
expeditions needs gamer loot removed dork and i are bikeshedding a better version that s less of a powergamer trap but in the interim remove the heals and the rpg and shit
0
352,416
10,542,009,013
IssuesEvent
2019-10-02 12:14:59
photonstorm/phaser
https://api.github.com/repos/photonstorm/phaser
closed
Load tileset from atlas
⌛️ Time Consuming 💖 Feature Request 🥉 Priority: Low
closed here but not sure what the solution is: https://github.com/photonstorm/phaser/issues/3415 https://phaser.discourse.group/t/loading-tileset-from-atlas/811 "I have a texture atlas that packs together all my tilesets(10 of them at the moment*), and I need to load them into Phaser TileSet so I can use it to generate static and dynamic layers. Seems like there is no frame option on .addTilesetImage and .createStaticLayer"
1.0
Load tileset from atlas - closed here but not sure what the solution is: https://github.com/photonstorm/phaser/issues/3415 https://phaser.discourse.group/t/loading-tileset-from-atlas/811 "I have a texture atlas that packs together all my tilesets(10 of them at the moment*), and I need to load them into Phaser TileSet so I can use it to generate static and dynamic layers. Seems like there is no frame option on .addTilesetImage and .createStaticLayer"
non_test
load tileset from atlas closed here but not sure what the solution is i have a texture atlas that packs together all my tilesets of them at the moment and i need to load them into phaser tileset so i can use it to generate static and dynamic layers seems like there is no frame option on addtilesetimage and createstaticlayer
0
237,778
19,674,673,055
IssuesEvent
2022-01-11 11:02:24
keycloak/keycloak
https://api.github.com/repos/keycloak/keycloak
opened
Verify the WebAuthn functionality and settings for authentication
area/testsuite kind/enhancement area/authentication/webauthn
### Description See: https://issues.redhat.com/browse/KEYCLOAK-19892 ### Discussion _No response_ ### Motivation _No response_ ### Details _No response_
1.0
Verify the WebAuthn functionality and settings for authentication - ### Description See: https://issues.redhat.com/browse/KEYCLOAK-19892 ### Discussion _No response_ ### Motivation _No response_ ### Details _No response_
test
verify the webauthn functionality and settings for authentication description see discussion no response motivation no response details no response
1
371,123
10,961,817,441
IssuesEvent
2019-11-27 16:03:44
Midas-Financials/midas-ios
https://api.github.com/repos/Midas-Financials/midas-ios
opened
Re-activate Idle App using Face ID
priority:high
User can re-activate idle app using Face ID (see how it works on Jenius app)
1.0
Re-activate Idle App using Face ID - User can re-activate idle app using Face ID (see how it works on Jenius app)
non_test
re activate idle app using face id user can re activate idle app using face id see how it works on jenius app
0
545,061
15,935,284,794
IssuesEvent
2021-04-14 09:40:24
buddyboss/buddyboss-platform
https://api.github.com/repos/buddyboss/buddyboss-platform
opened
When Private Website option is enable and added some specific page URL then allow RSS feed and API endpoint for those same page URL
feature: enhancement priority: medium
**Describe the feature** When Private Website option is enable from General setting and added some specific page URL to access for the user. At that time, We should allow RSS feed and API endpoint for those specific page URL. **For ex** - If we added News Feed page in public content (General -> Privacy Setting should be enable ) then user also should accessible RSS feed for the same.
1.0
When Private Website option is enable and added some specific page URL then allow RSS feed and API endpoint for those same page URL - **Describe the feature** When Private Website option is enable from General setting and added some specific page URL to access for the user. At that time, We should allow RSS feed and API endpoint for those specific page URL. **For ex** - If we added News Feed page in public content (General -> Privacy Setting should be enable ) then user also should accessible RSS feed for the same.
non_test
when private website option is enable and added some specific page url then allow rss feed and api endpoint for those same page url describe the feature when private website option is enable from general setting and added some specific page url to access for the user at that time we should allow rss feed and api endpoint for those specific page url for ex if we added news feed page in public content general privacy setting should be enable then user also should accessible rss feed for the same
0
200,510
15,109,718,999
IssuesEvent
2021-02-08 18:13:30
aio-libs/aiosmtpd
https://api.github.com/repos/aio-libs/aiosmtpd
closed
Keep using nose2 or change to pytest?
pr-available testing
As the title says. I want to see what everyone thinks about this. And I'll volunteer myself to do the heavy-lifting of rewriting all Test Cases into pytest if the community chooses too change.
1.0
Keep using nose2 or change to pytest? - As the title says. I want to see what everyone thinks about this. And I'll volunteer myself to do the heavy-lifting of rewriting all Test Cases into pytest if the community chooses too change.
test
keep using or change to pytest as the title says i want to see what everyone thinks about this and i ll volunteer myself to do the heavy lifting of rewriting all test cases into pytest if the community chooses too change
1
21,874
11,660,529,496
IssuesEvent
2020-03-03 03:39:23
cityofaustin/atd-knack-dts-portal
https://api.github.com/repos/cityofaustin/atd-knack-dts-portal
closed
Match text in Github description field to text in service request form
Need: 3-Could Have Project: DTS Service Delivery Service: Apps Service: Dev Type: Enhancement Workgroup: DTS
It would be helpful for us to see the same text in the issue description that the requester sees in the form. ![Screen Shot 2019-09-05 at 2 50 50 PM](https://user-images.githubusercontent.com/1463708/64379447-89a87800-cff4-11e9-909a-d8d22f813236.png)
3.0
Match text in Github description field to text in service request form - It would be helpful for us to see the same text in the issue description that the requester sees in the form. ![Screen Shot 2019-09-05 at 2 50 50 PM](https://user-images.githubusercontent.com/1463708/64379447-89a87800-cff4-11e9-909a-d8d22f813236.png)
non_test
match text in github description field to text in service request form it would be helpful for us to see the same text in the issue description that the requester sees in the form
0
792,828
27,975,329,573
IssuesEvent
2023-03-25 14:09:55
AY2223S2-CS2103T-T17-3/tp
https://api.github.com/repos/AY2223S2-CS2103T-T17-3/tp
closed
Add import csv file
type.Story priority.High
As an insurance agent with a csv dataset of customer information, I can import a dataset so I can easily manipulate it on the app.
1.0
Add import csv file - As an insurance agent with a csv dataset of customer information, I can import a dataset so I can easily manipulate it on the app.
non_test
add import csv file as an insurance agent with a csv dataset of customer information i can import a dataset so i can easily manipulate it on the app
0
236,392
19,537,613,178
IssuesEvent
2021-12-31 11:14:58
bounswe/2021SpringGroup7
https://api.github.com/repos/bounswe/2021SpringGroup7
closed
Unit Test for Story Text Search
Type: Enhancement Status: In Progress Type: Test
Is your proposal related to a problem? -------------------------------------- <!-- Provide a clear and concise description of what the problem is. For example, "I'm always frustrated when..." --> The text search functionality has not been tested. It should be tested to see if it works properly. Describe the solution you'd like -------------------------------- <!-- Provide a clear and concise description of what you want to happen. --> I will write tests for the text search API to be sure for its functionality.
1.0
Unit Test for Story Text Search - Is your proposal related to a problem? -------------------------------------- <!-- Provide a clear and concise description of what the problem is. For example, "I'm always frustrated when..." --> The text search functionality has not been tested. It should be tested to see if it works properly. Describe the solution you'd like -------------------------------- <!-- Provide a clear and concise description of what you want to happen. --> I will write tests for the text search API to be sure for its functionality.
test
unit test for story text search is your proposal related to a problem provide a clear and concise description of what the problem is for example i m always frustrated when the text search functionality has not been tested it should be tested to see if it works properly describe the solution you d like provide a clear and concise description of what you want to happen i will write tests for the text search api to be sure for its functionality
1
300,735
25,991,950,014
IssuesEvent
2022-12-20 08:24:51
ita-social-projects/EventsExpress
https://api.github.com/repos/ita-social-projects/EventsExpress
opened
Verify that the user is able to filter events by category
test case
**Date Time** 12-20-2022 **Priority** Medium **Description** This test case varifies that the user is able to filter events by category **Pre-requisites** Steps to be fulfilled before test: 1.Move to https://eventsexpress-test.azurewebsites.net/ 2.The user is logged on a system (Username = 'User1', Password = 'Pass1') 3.The user is on the "My Events" page 4.The event creation functionality is implemented **Test Cases** | S# | Action | Test Data | Expected Result | Actual Result | P/F | Automated | |:-----------:|:-----------:|:-----------:|:---------------:|:-------------:|:-----------:|:-----------:| | 1 | Click the 'Filters' button in the top right corner | <br> | The modal window with filters is opened | <br> | <br> | <br> | | 2 | Verify that the "Category" filter is displayed | <br> | The "Category" filter is displayed | <br> | <br> | <br> | | 3 | Click the "Category" dropdown menu | <br> | The "Category" dropdown list is opened | <br> | <br> | <br> | | 4 | Click the "Art&Craft" category list | <br> | The 'Art&Craft" check box list is opened | <br> | <br> | <br> | | 5 | Check the 'Drawing' check box | <br> | The "Drawing" check box is checked | <br> | <br> | <br> | | 6 | Click the "Apply" button | <br> | The "Dancing" events are displayed on the page | <br> | <br> | <br> | **Screenshots** ![image](https://user-images.githubusercontent.com/115736426/208610799-cc8e712b-4d02-42db-b80a-76aa9c46feb2.png) **Environment:** - OS: Win 10 - Browser Google Chrome - Version 108.0.5359.100 **Additional context** Add any other context about the test here. User story links E.g.: "User story #100 " Labels to be added "Test case", Priority ("pri: "), Type ("UI, "Functional"), "API" .
1.0
Verify that the user is able to filter events by category - **Date Time** 12-20-2022 **Priority** Medium **Description** This test case varifies that the user is able to filter events by category **Pre-requisites** Steps to be fulfilled before test: 1.Move to https://eventsexpress-test.azurewebsites.net/ 2.The user is logged on a system (Username = 'User1', Password = 'Pass1') 3.The user is on the "My Events" page 4.The event creation functionality is implemented **Test Cases** | S# | Action | Test Data | Expected Result | Actual Result | P/F | Automated | |:-----------:|:-----------:|:-----------:|:---------------:|:-------------:|:-----------:|:-----------:| | 1 | Click the 'Filters' button in the top right corner | <br> | The modal window with filters is opened | <br> | <br> | <br> | | 2 | Verify that the "Category" filter is displayed | <br> | The "Category" filter is displayed | <br> | <br> | <br> | | 3 | Click the "Category" dropdown menu | <br> | The "Category" dropdown list is opened | <br> | <br> | <br> | | 4 | Click the "Art&Craft" category list | <br> | The 'Art&Craft" check box list is opened | <br> | <br> | <br> | | 5 | Check the 'Drawing' check box | <br> | The "Drawing" check box is checked | <br> | <br> | <br> | | 6 | Click the "Apply" button | <br> | The "Dancing" events are displayed on the page | <br> | <br> | <br> | **Screenshots** ![image](https://user-images.githubusercontent.com/115736426/208610799-cc8e712b-4d02-42db-b80a-76aa9c46feb2.png) **Environment:** - OS: Win 10 - Browser Google Chrome - Version 108.0.5359.100 **Additional context** Add any other context about the test here. User story links E.g.: "User story #100 " Labels to be added "Test case", Priority ("pri: "), Type ("UI, "Functional"), "API" .
test
verify that the user is able to filter events by category date time priority medium description this test case varifies that the user is able to filter events by category pre requisites steps to be fulfilled before test move to the user is logged on a system username password the user is on the my events page the event creation functionality is implemented test cases s action test data expected result actual result p f automated click the filters button in the top right corner the modal window with filters is opened verify that the category filter is displayed the category filter is displayed click the category dropdown menu the category dropdown list is opened click the art craft category list the art craft check box list is opened check the drawing check box the drawing check box is checked click the apply button the dancing events are displayed on the page screenshots environment os win browser google chrome version additional context add any other context about the test here user story links e g user story labels to be added test case priority pri type ui functional api
1
339,996
30,489,612,127
IssuesEvent
2023-07-18 06:40:55
centrifuge/centrifuge-chain
https://api.github.com/repos/centrifuge/centrifuge-chain
opened
Loans: add integration tests
I4-tests Q5-hard P4-required
### Description Currently, there is no specific integration test for loan use cases. We need to add it and also to tests runtime APIs. Loan use-cases logic is very tied to other pallets. So it involves setting up pools, assets, registries, investors, etc...
1.0
Loans: add integration tests - ### Description Currently, there is no specific integration test for loan use cases. We need to add it and also to tests runtime APIs. Loan use-cases logic is very tied to other pallets. So it involves setting up pools, assets, registries, investors, etc...
test
loans add integration tests description currently there is no specific integration test for loan use cases we need to add it and also to tests runtime apis loan use cases logic is very tied to other pallets so it involves setting up pools assets registries investors etc
1
269,625
23,453,914,965
IssuesEvent
2022-08-16 07:05:24
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
ccl/streamingccl/streamingest: TestTenantStreamingUnavailableStreamAddress failed
C-test-failure O-robot branch-master
ccl/streamingccl/streamingest.TestTenantStreamingUnavailableStreamAddress [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/6108451?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/6108451?buildTab=artifacts#/) on master @ [f4042d47fa8062a612c38d4696eb6bee9cee7c21](https://github.com/cockroachdb/cockroach/commits/f4042d47fa8062a612c38d4696eb6bee9cee7c21): ``` === RUN TestTenantStreamingUnavailableStreamAddress test_log_scope.go:162: test logs captured to: /artifacts/tmp/_tmp/045fecfe9e18c9da6f776a53d446f2c5/logTestTenantStreamingUnavailableStreamAddress3221948880 test_log_scope.go:80: use -show-logs to present logs inline stream_replication_e2e_test.go:805: condition failed to evaluate within 2m0s: expected job status succeeded, but got paused panic.go:500: -- test log scope end -- --- FAIL: TestTenantStreamingUnavailableStreamAddress (217.79s) ``` <p>Parameters: <code>TAGS=bazel,gss</code> </p> <details><summary>Help</summary> <p> See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) </p> </details> /cc @cockroachdb/tenant-streaming <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestTenantStreamingUnavailableStreamAddress.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
1.0
ccl/streamingccl/streamingest: TestTenantStreamingUnavailableStreamAddress failed - ccl/streamingccl/streamingest.TestTenantStreamingUnavailableStreamAddress [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/6108451?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/6108451?buildTab=artifacts#/) on master @ [f4042d47fa8062a612c38d4696eb6bee9cee7c21](https://github.com/cockroachdb/cockroach/commits/f4042d47fa8062a612c38d4696eb6bee9cee7c21): ``` === RUN TestTenantStreamingUnavailableStreamAddress test_log_scope.go:162: test logs captured to: /artifacts/tmp/_tmp/045fecfe9e18c9da6f776a53d446f2c5/logTestTenantStreamingUnavailableStreamAddress3221948880 test_log_scope.go:80: use -show-logs to present logs inline stream_replication_e2e_test.go:805: condition failed to evaluate within 2m0s: expected job status succeeded, but got paused panic.go:500: -- test log scope end -- --- FAIL: TestTenantStreamingUnavailableStreamAddress (217.79s) ``` <p>Parameters: <code>TAGS=bazel,gss</code> </p> <details><summary>Help</summary> <p> See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) </p> </details> /cc @cockroachdb/tenant-streaming <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestTenantStreamingUnavailableStreamAddress.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
test
ccl streamingccl streamingest testtenantstreamingunavailablestreamaddress failed ccl streamingccl streamingest testtenantstreamingunavailablestreamaddress with on master run testtenantstreamingunavailablestreamaddress test log scope go test logs captured to artifacts tmp tmp test log scope go use show logs to present logs inline stream replication test go condition failed to evaluate within expected job status succeeded but got paused panic go test log scope end fail testtenantstreamingunavailablestreamaddress parameters tags bazel gss help see also cc cockroachdb tenant streaming
1
349,689
24,952,229,242
IssuesEvent
2022-11-01 08:28:29
bounswe/bounswe2022group1
https://api.github.com/repos/bounswe/bounswe2022group1
closed
Frontend: Project plan
Type: Documentation Status: Review Needed Frontend
**Issue Description:** I will log our project plan on ProjectLibre for it to look organized, and to put it on our Deliverables for Milestone 1 report **Tasks to Do:** - [x] create project plan - [x] upload it on deliverables *Task Deadline:* 01.11.2022 10.00 *Final Situation:* <img width="1440" alt="image" src="https://user-images.githubusercontent.com/43001709/199162409-0ce815c6-6ebf-4252-84f7-5a98413929b0.png"> *Reviewer:* @kadirgokhann , @hseyyidkaplan *Review Deadline:* 01.11.2022 11.00
1.0
Frontend: Project plan - **Issue Description:** I will log our project plan on ProjectLibre for it to look organized, and to put it on our Deliverables for Milestone 1 report **Tasks to Do:** - [x] create project plan - [x] upload it on deliverables *Task Deadline:* 01.11.2022 10.00 *Final Situation:* <img width="1440" alt="image" src="https://user-images.githubusercontent.com/43001709/199162409-0ce815c6-6ebf-4252-84f7-5a98413929b0.png"> *Reviewer:* @kadirgokhann , @hseyyidkaplan *Review Deadline:* 01.11.2022 11.00
non_test
frontend project plan issue description i will log our project plan on projectlibre for it to look organized and to put it on our deliverables for milestone report tasks to do create project plan upload it on deliverables task deadline final situation img width alt image src reviewer kadirgokhann hseyyidkaplan review deadline
0
163,789
12,744,724,993
IssuesEvent
2020-06-26 13:02:59
oamg/leapp
https://api.github.com/repos/oamg/leapp
closed
Deprecated decorated is not covered by tests
unittest
Based on coverage report the MR https://github.com/oamg/leapp/pull/640 is not covered by tests. ``` ----------- coverage: platform linux, python 3.8.3-final-0 ----------- Name Stmts Miss Cover Missing ------------------------------------------------------------------------------- leapp/utils/deprecation.py 64 52 19% 22-41, 63-114 ```
1.0
Deprecated decorated is not covered by tests - Based on coverage report the MR https://github.com/oamg/leapp/pull/640 is not covered by tests. ``` ----------- coverage: platform linux, python 3.8.3-final-0 ----------- Name Stmts Miss Cover Missing ------------------------------------------------------------------------------- leapp/utils/deprecation.py 64 52 19% 22-41, 63-114 ```
test
deprecated decorated is not covered by tests based on coverage report the mr is not covered by tests coverage platform linux python final name stmts miss cover missing leapp utils deprecation py
1
38,822
5,200,624,489
IssuesEvent
2017-01-24 00:37:51
openshift/origin
https://api.github.com/repos/openshift/origin
closed
[networking][router] weighted openshift router flake
component/networking kind/test-flake priority/P1
``` • Failure [156.277 seconds] [networking][router] weighted openshift router /data/src/github.com/openshift/origin/test/extended/router/weighted.go:112 The HAProxy router /data/src/github.com/openshift/origin/test/extended/router/weighted.go:111 should appropriately serve a route that points to two services [It] /data/src/github.com/openshift/origin/test/extended/router/weighted.go:110 Expected error: <*errors.errorString | 0xc82017a980>: { s: "timed out waiting for the condition", } timed out waiting for the condition not to have occurred /data/src/github.com/openshift/origin/test/extended/router/weighted.go:56 ``` https://ci.openshift.redhat.com/jenkins/job/test_pull_requests_origin_conformance/7997/consoleFull#39374030356cbb9a5e4b02b88ae8c2f77
1.0
[networking][router] weighted openshift router flake - ``` • Failure [156.277 seconds] [networking][router] weighted openshift router /data/src/github.com/openshift/origin/test/extended/router/weighted.go:112 The HAProxy router /data/src/github.com/openshift/origin/test/extended/router/weighted.go:111 should appropriately serve a route that points to two services [It] /data/src/github.com/openshift/origin/test/extended/router/weighted.go:110 Expected error: <*errors.errorString | 0xc82017a980>: { s: "timed out waiting for the condition", } timed out waiting for the condition not to have occurred /data/src/github.com/openshift/origin/test/extended/router/weighted.go:56 ``` https://ci.openshift.redhat.com/jenkins/job/test_pull_requests_origin_conformance/7997/consoleFull#39374030356cbb9a5e4b02b88ae8c2f77
test
weighted openshift router flake • failure weighted openshift router data src github com openshift origin test extended router weighted go the haproxy router data src github com openshift origin test extended router weighted go should appropriately serve a route that points to two services data src github com openshift origin test extended router weighted go expected error s timed out waiting for the condition timed out waiting for the condition not to have occurred data src github com openshift origin test extended router weighted go
1
507,803
14,681,068,495
IssuesEvent
2020-12-31 12:04:29
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.google.com - see bug description
browser-fenix engine-gecko ml-needsdiagnosis-false ml-probability-high priority-critical
<!-- @browser: Firefox Mobile 85.0 --> <!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:85.0) Gecko/85.0 Firefox/85.0 --> <!-- @reported_with: android-components-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/64393 --> <!-- @extra_labels: browser-fenix --> **URL**: https://www.google.com/search?q=mandy&ie=utf-8&oe=utf-8&client=firefox-b-m **Browser / Version**: Firefox Mobile 85.0 **Operating System**: Android **Tested Another Browser**: Yes Chrome **Problem type**: Something else **Description**: Refused to open the site **Steps to Reproduce**: <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2020/12/4e760e79-fba5-4fd0-a1e6-d3241bd164f6.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20201223151005</li><li>channel: beta</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2020/12/c7a7ed11-ee9c-43c1-b747-a94d2cf2a2f4) _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.google.com - see bug description - <!-- @browser: Firefox Mobile 85.0 --> <!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:85.0) Gecko/85.0 Firefox/85.0 --> <!-- @reported_with: android-components-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/64393 --> <!-- @extra_labels: browser-fenix --> **URL**: https://www.google.com/search?q=mandy&ie=utf-8&oe=utf-8&client=firefox-b-m **Browser / Version**: Firefox Mobile 85.0 **Operating System**: Android **Tested Another Browser**: Yes Chrome **Problem type**: Something else **Description**: Refused to open the site **Steps to Reproduce**: <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2020/12/4e760e79-fba5-4fd0-a1e6-d3241bd164f6.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20201223151005</li><li>channel: beta</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2020/12/c7a7ed11-ee9c-43c1-b747-a94d2cf2a2f4) _From [webcompat.com](https://webcompat.com/) with ❤️_
non_test
see bug description url browser version firefox mobile operating system android tested another browser yes chrome problem type something else description refused to open the site steps to reproduce view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel beta hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
0
110,489
23,939,784,455
IssuesEvent
2022-09-11 18:46:18
appsmithorg/appsmith
https://api.github.com/repos/appsmithorg/appsmith
opened
[Feature]: Guide the user to use equals (=) operator instead of the IS keyword in case of boolean value comparison in MySQL
Enhancement Query Execution MySQL BE Coders Pod
### Is there an existing issue for this? - [X] I have searched the existing issues ### Summary ## Background Appsmith doesn't support MySQL queries with the “IS” keyword in the prepared statement. We thought the underlying MySQL r2dbc library doesn’t support this. But the fact is MySQL r2dbc supports queries having the “IS” keyword under the prepared statement too. The only drawback with the “IS” keyword under the prepared statement is it doesn’t work as expected when the “IS” keyword is used to check some boolean values where the boolean value is coming from a binding. `SELECT * FROM table where col IS {{true}}` `SELECT * FROM table where col IS {{false}}` `SELECT * FROM table where col IS NOT {{true}}` `SELECT * FROM table where col IS NOT {{false}}` Due to this reason earlier we decided to disallow queries having the “IS” keyword under the prepared statement. ### Problem Statement Comparing boolean value using the IS keyword doesn't work as expected when the boolean value is coming from a binding. A query like the below falls into this case `SELECT * FROM table where col IS {{true}}` ### Proposed Solution We can guide the user to go with the = operator way instead of the "IS" keyword way for comparing boolean values when a query is written like the above. ### Why should this be worked on? Theoretically, there are two ways to compare boolean values with the prepared statement. But there is only **one** way to check null value according to the SQL standard i.e. with the `IS` keyword. **Ways to compare boolean values** - With the equals (=) operator - With the IS keyword **Ways to compare null values** - With IS keyword We should not close the door entirely just because IS doesn’t work as expected in case of boolean values because there’s just one way to check for null. And there are plenty of use cases where null checking is made as part of the query and we should not force the user to rewrite the whole query by turning off the prepared statement for this reason.
1.0
[Feature]: Guide the user to use equals (=) operator instead of the IS keyword in case of boolean value comparison in MySQL - ### Is there an existing issue for this? - [X] I have searched the existing issues ### Summary ## Background Appsmith doesn't support MySQL queries with the “IS” keyword in the prepared statement. We thought the underlying MySQL r2dbc library doesn’t support this. But the fact is MySQL r2dbc supports queries having the “IS” keyword under the prepared statement too. The only drawback with the “IS” keyword under the prepared statement is it doesn’t work as expected when the “IS” keyword is used to check some boolean values where the boolean value is coming from a binding. `SELECT * FROM table where col IS {{true}}` `SELECT * FROM table where col IS {{false}}` `SELECT * FROM table where col IS NOT {{true}}` `SELECT * FROM table where col IS NOT {{false}}` Due to this reason earlier we decided to disallow queries having the “IS” keyword under the prepared statement. ### Problem Statement Comparing boolean value using the IS keyword doesn't work as expected when the boolean value is coming from a binding. A query like the below falls into this case `SELECT * FROM table where col IS {{true}}` ### Proposed Solution We can guide the user to go with the = operator way instead of the "IS" keyword way for comparing boolean values when a query is written like the above. ### Why should this be worked on? Theoretically, there are two ways to compare boolean values with the prepared statement. But there is only **one** way to check null value according to the SQL standard i.e. with the `IS` keyword. **Ways to compare boolean values** - With the equals (=) operator - With the IS keyword **Ways to compare null values** - With IS keyword We should not close the door entirely just because IS doesn’t work as expected in case of boolean values because there’s just one way to check for null. And there are plenty of use cases where null checking is made as part of the query and we should not force the user to rewrite the whole query by turning off the prepared statement for this reason.
non_test
guide the user to use equals operator instead of the is keyword in case of boolean value comparison in mysql is there an existing issue for this i have searched the existing issues summary background appsmith doesn t support mysql queries with the “is” keyword in the prepared statement we thought the underlying mysql library doesn’t support this but the fact is mysql supports queries having the “is” keyword under the prepared statement too the only drawback with the “is” keyword under the prepared statement is it doesn’t work as expected when the “is” keyword is used to check some boolean values where the boolean value is coming from a binding select from table where col is true select from table where col is false select from table where col is not true select from table where col is not false due to this reason earlier we decided to disallow queries having the “is” keyword under the prepared statement problem statement comparing boolean value using the is keyword doesn t work as expected when the boolean value is coming from a binding a query like the below falls into this case select from table where col is true proposed solution we can guide the user to go with the operator way instead of the is keyword way for comparing boolean values when a query is written like the above why should this be worked on theoretically there are two ways to compare boolean values with the prepared statement but there is only one way to check null value according to the sql standard i e with the is keyword ways to compare boolean values with the equals operator with the is keyword ways to compare null values with is keyword we should not close the door entirely just because is doesn’t work as expected in case of boolean values because there’s just one way to check for null and there are plenty of use cases where null checking is made as part of the query and we should not force the user to rewrite the whole query by turning off the prepared statement for this reason
0
94,326
8,484,728,331
IssuesEvent
2018-10-26 04:18:55
owncloud/core
https://api.github.com/repos/owncloud/core
closed
fix API acceptance test steps, so that they really use old/new webdav path
1 - To develop QA-team dev:acceptance-tests
this steps only use the old webdav and not obeying what is said in `Given using <dav_version> DAV path` - [x] `@When /^the public downloads the last public shared file with range "([^"]*)" using the old WebDAV API$/` - [x] `@When /^the public downloads file "([^"]*)" from inside the last public shared folder with range "([^"]*)" using the old WebDAV API$/` - [x] ` @When /^the public downloads file "([^"]*)" from inside the last public shared folder with password "([^"]*)" with range "([^"]*)" using the old WebDAV API$/` - [x] `@Then /^the public should be able to download the range "([^"]*)" of file "([^"]*)" from inside the last public shared folder with password "([^"]*)" and the content should be "([^"]*)"$/` -> it uses `publicDownloadsTheFileInsideThePublicSharedFolderWithPassword()`
1.0
fix API acceptance test steps, so that they really use old/new webdav path - this steps only use the old webdav and not obeying what is said in `Given using <dav_version> DAV path` - [x] `@When /^the public downloads the last public shared file with range "([^"]*)" using the old WebDAV API$/` - [x] `@When /^the public downloads file "([^"]*)" from inside the last public shared folder with range "([^"]*)" using the old WebDAV API$/` - [x] ` @When /^the public downloads file "([^"]*)" from inside the last public shared folder with password "([^"]*)" with range "([^"]*)" using the old WebDAV API$/` - [x] `@Then /^the public should be able to download the range "([^"]*)" of file "([^"]*)" from inside the last public shared folder with password "([^"]*)" and the content should be "([^"]*)"$/` -> it uses `publicDownloadsTheFileInsideThePublicSharedFolderWithPassword()`
test
fix api acceptance test steps so that they really use old new webdav path this steps only use the old webdav and not obeying what is said in given using dav path when the public downloads the last public shared file with range using the old webdav api when the public downloads file from inside the last public shared folder with range using the old webdav api when the public downloads file from inside the last public shared folder with password with range using the old webdav api then the public should be able to download the range of file from inside the last public shared folder with password and the content should be it uses publicdownloadsthefileinsidethepublicsharedfolderwithpassword
1
334,386
29,833,562,131
IssuesEvent
2023-06-18 15:05:43
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
reopened
Fix linalg.test_torch_pinv
PyTorch Frontend Sub Task Failing Test
| | | |---|---| |tensorflow|<a href="null" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="null" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="null" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="null" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |paddle|<a href="null" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
1.0
Fix linalg.test_torch_pinv - | | | |---|---| |tensorflow|<a href="null" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="null" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="null" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="null" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |paddle|<a href="null" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
test
fix linalg test torch pinv tensorflow img src torch img src numpy img src jax img src paddle img src
1
37,911
15,394,622,679
IssuesEvent
2021-03-03 18:08:40
cityofaustin/atd-data-tech
https://api.github.com/repos/cityofaustin/atd-data-tech
opened
User Management | Backfill existing user roles in staging and production
Impact: 2-Major Product: Moped Project: Moped v1.0 Service: Dev
In https://github.com/cityofaustin/atd-data-tech/issues/4994, we added user roles to the DB so we can edit them through the Moped UI. Since existing users were created before this change, we need to backfill them in the DB with the new changes to the DB, API, and UI. - [ ] Check staging users and set roles - [ ] Check production users and set roles
1.0
User Management | Backfill existing user roles in staging and production - In https://github.com/cityofaustin/atd-data-tech/issues/4994, we added user roles to the DB so we can edit them through the Moped UI. Since existing users were created before this change, we need to backfill them in the DB with the new changes to the DB, API, and UI. - [ ] Check staging users and set roles - [ ] Check production users and set roles
non_test
user management backfill existing user roles in staging and production in we added user roles to the db so we can edit them through the moped ui since existing users were created before this change we need to backfill them in the db with the new changes to the db api and ui check staging users and set roles check production users and set roles
0
108,626
9,312,666,468
IssuesEvent
2019-03-26 02:16:48
xcat2/xcat2-task-management
https://api.github.com/repos/xcat2/xcat2-task-management
closed
offer bundles for physical machine test
sprint4 test
what to do: offer below bundles for physical machine test * [ ] offer a suitable method of bundle catalogue, offer the bundle names for example: ``` physical_node_openpower_openbmc.bundle physical_node_openpower_ipmi.bundle physical_node_power_hmc.bundle physical_node_x86_ipmi.bundle ...... ``` These bundles are composed of below sub-bundles and other necessary test cases(such like provision test cases) ``` hdctrl_bmc.bundle hdctrl_general.bundle hdctrl_openpower_ipmi.bundle hdctrl_openpower_openbmc.bundle hdctrl_ppc_hmc.bundle ``` * [ ] update below bundles depending on the latest test case list ``` hdctrl_bmc.bundle hdctrl_general.bundle hdctrl_openpower_ipmi.bundle hdctrl_openpower_openbmc.bundle hdctrl_ppc_hmc.bundle ``` * [ ] offer bundles for perl and python test * [ ] offer a [document](https://github.ibm.com/xcat2/team_process/blob/master/FVT/docs/how_to_test_hardware_by_xcattest.md) for how to using these bundles * [ ] make below projects in automation development environment work http://10.3.17.9:8080/view/Release/job/regression_p8le_physical_server/ http://10.3.17.9:8080/view/Release/job/regression_x86_physical_server/ http://10.3.17.9:8080/view/Release/job/Pegas1.2-P9-PostgreSQL-Perl/
1.0
offer bundles for physical machine test - what to do: offer below bundles for physical machine test * [ ] offer a suitable method of bundle catalogue, offer the bundle names for example: ``` physical_node_openpower_openbmc.bundle physical_node_openpower_ipmi.bundle physical_node_power_hmc.bundle physical_node_x86_ipmi.bundle ...... ``` These bundles are composed of below sub-bundles and other necessary test cases(such like provision test cases) ``` hdctrl_bmc.bundle hdctrl_general.bundle hdctrl_openpower_ipmi.bundle hdctrl_openpower_openbmc.bundle hdctrl_ppc_hmc.bundle ``` * [ ] update below bundles depending on the latest test case list ``` hdctrl_bmc.bundle hdctrl_general.bundle hdctrl_openpower_ipmi.bundle hdctrl_openpower_openbmc.bundle hdctrl_ppc_hmc.bundle ``` * [ ] offer bundles for perl and python test * [ ] offer a [document](https://github.ibm.com/xcat2/team_process/blob/master/FVT/docs/how_to_test_hardware_by_xcattest.md) for how to using these bundles * [ ] make below projects in automation development environment work http://10.3.17.9:8080/view/Release/job/regression_p8le_physical_server/ http://10.3.17.9:8080/view/Release/job/regression_x86_physical_server/ http://10.3.17.9:8080/view/Release/job/Pegas1.2-P9-PostgreSQL-Perl/
test
offer bundles for physical machine test what to do offer below bundles for physical machine test offer a suitable method of bundle catalogue offer the bundle names for example physical node openpower openbmc bundle physical node openpower ipmi bundle physical node power hmc bundle physical node ipmi bundle these bundles are composed of below sub bundles and other necessary test cases such like provision test cases hdctrl bmc bundle hdctrl general bundle hdctrl openpower ipmi bundle hdctrl openpower openbmc bundle hdctrl ppc hmc bundle update below bundles depending on the latest test case list hdctrl bmc bundle hdctrl general bundle hdctrl openpower ipmi bundle hdctrl openpower openbmc bundle hdctrl ppc hmc bundle offer bundles for perl and python test offer a for how to using these bundles make below projects in automation development environment work
1
88,350
17,569,449,349
IssuesEvent
2021-08-14 11:19:34
CoDemIT2021/CoDem_team_WebSite
https://api.github.com/repos/CoDemIT2021/CoDem_team_WebSite
closed
Team website section Form
Code
Column of text and column form with inputs clone or pull from GitHub use HTML & CSS make your changes and push to the GitHub create pull requests
1.0
Team website section Form - Column of text and column form with inputs clone or pull from GitHub use HTML & CSS make your changes and push to the GitHub create pull requests
non_test
team website section form column of text and column form with inputs clone or pull from github use html css make your changes and push to the github create pull requests
0
212,475
16,452,674,900
IssuesEvent
2021-05-21 08:15:03
pytorch/vision
https://api.github.com/repos/pytorch/vision
closed
Unit test failure
high priority module: tests topic: video triage review
``` self = <test_videoapi.TestVideoApi testMethod=test_frame_reading> @unittest.skipIf(av is None, "PyAV unavailable") def test_frame_reading(self): for test_video, config in test_videos.items(): full_path = os.path.join(VIDEO_DIR, test_video) av_reader = av.open(full_path) if av_reader.streams.video: video_reader = VideoReader(full_path, "video") for av_frame in av_reader.decode(av_reader.streams.video[0]): vr_frame = next(video_reader) self.assertAlmostEqual( float(av_frame.pts * av_frame.time_base), vr_frame["pts"], delta=0.1, ) av_array = torch.tensor(av_frame.to_rgb().to_ndarray()).permute( 2, 0, 1 ) vr_array = vr_frame["data"] mean_delta = torch.mean( torch.abs(av_array.float() - vr_array.float()) ) # on average the difference is very small and caused # by decoding (around 1%) # TODO: asses empirically how to set this? atm it's 1% # averaged over all frames > self.assertTrue(mean_delta.item() < 2.5) E AssertionError: False is not true test/test_videoapi.py:97: AssertionError ``` https://app.circleci.com/pipelines/github/pytorch/vision/7915/workflows/82c134ae-52a5-4b83-8dbe-22d3f30ce3e5/jobs/561557 cc @bjuncek
1.0
Unit test failure - ``` self = <test_videoapi.TestVideoApi testMethod=test_frame_reading> @unittest.skipIf(av is None, "PyAV unavailable") def test_frame_reading(self): for test_video, config in test_videos.items(): full_path = os.path.join(VIDEO_DIR, test_video) av_reader = av.open(full_path) if av_reader.streams.video: video_reader = VideoReader(full_path, "video") for av_frame in av_reader.decode(av_reader.streams.video[0]): vr_frame = next(video_reader) self.assertAlmostEqual( float(av_frame.pts * av_frame.time_base), vr_frame["pts"], delta=0.1, ) av_array = torch.tensor(av_frame.to_rgb().to_ndarray()).permute( 2, 0, 1 ) vr_array = vr_frame["data"] mean_delta = torch.mean( torch.abs(av_array.float() - vr_array.float()) ) # on average the difference is very small and caused # by decoding (around 1%) # TODO: asses empirically how to set this? atm it's 1% # averaged over all frames > self.assertTrue(mean_delta.item() < 2.5) E AssertionError: False is not true test/test_videoapi.py:97: AssertionError ``` https://app.circleci.com/pipelines/github/pytorch/vision/7915/workflows/82c134ae-52a5-4b83-8dbe-22d3f30ce3e5/jobs/561557 cc @bjuncek
test
unit test failure self unittest skipif av is none pyav unavailable def test frame reading self for test video config in test videos items full path os path join video dir test video av reader av open full path if av reader streams video video reader videoreader full path video for av frame in av reader decode av reader streams video vr frame next video reader self assertalmostequal float av frame pts av frame time base vr frame delta av array torch tensor av frame to rgb to ndarray permute vr array vr frame mean delta torch mean torch abs av array float vr array float on average the difference is very small and caused by decoding around todo asses empirically how to set this atm it s averaged over all frames self asserttrue mean delta item e assertionerror false is not true test test videoapi py assertionerror cc bjuncek
1
343,989
30,705,280,704
IssuesEvent
2023-07-27 05:25:03
chrisvaillancourt/website
https://api.github.com/repos/chrisvaillancourt/website
opened
test: listen for console.error messages
test
Listen to [`page.on('console')`](https://playwright.dev/docs/api/class-page#page-event-console) events to listen for error messages. Each page test should fail if there's a console.error message. See [ConsoleMessage object docs](https://playwright.dev/docs/api/class-consolemessage).
1.0
test: listen for console.error messages - Listen to [`page.on('console')`](https://playwright.dev/docs/api/class-page#page-event-console) events to listen for error messages. Each page test should fail if there's a console.error message. See [ConsoleMessage object docs](https://playwright.dev/docs/api/class-consolemessage).
test
test listen for console error messages listen to events to listen for error messages each page test should fail if there s a console error message see
1
348,750
31,714,859,424
IssuesEvent
2023-09-09 18:37:04
denoland/vscode_deno
https://api.github.com/repos/denoland/vscode_deno
closed
Testing view should respect `test.files`
bug testing
**Screenshots** ![include](https://user-images.githubusercontent.com/23058303/231770020-5e5e368b-abef-46b5-a8d2-645422cf6ae2.png) ![exclude](https://user-images.githubusercontent.com/23058303/231770808-a2009d3d-d8dc-4797-804d-c1aac613293c.png) **Versions** vscode: 1.77.0 deno: 1.32.2 extension: 3.17.0
1.0
Testing view should respect `test.files` - **Screenshots** ![include](https://user-images.githubusercontent.com/23058303/231770020-5e5e368b-abef-46b5-a8d2-645422cf6ae2.png) ![exclude](https://user-images.githubusercontent.com/23058303/231770808-a2009d3d-d8dc-4797-804d-c1aac613293c.png) **Versions** vscode: 1.77.0 deno: 1.32.2 extension: 3.17.0
test
testing view should respect test files screenshots versions vscode deno extension
1
15,529
3,474,733,478
IssuesEvent
2015-12-25 02:54:45
ensime/scala-debugger
https://api.github.com/repos/ensime/scala-debugger
opened
Add .drone.yml file to conform with existing Ensime testing infrastructure
for test refactor
Currently, we have a Travis CI setup for testing pull requests. Ensime uses a drone infrastructure with beefier machines; so, we need to move our setup from Travis to drone. Also involves removing the .travis.yml file. Will need to see if tests have any issue running on drone compared to local machine and Travis CI.
1.0
Add .drone.yml file to conform with existing Ensime testing infrastructure - Currently, we have a Travis CI setup for testing pull requests. Ensime uses a drone infrastructure with beefier machines; so, we need to move our setup from Travis to drone. Also involves removing the .travis.yml file. Will need to see if tests have any issue running on drone compared to local machine and Travis CI.
test
add drone yml file to conform with existing ensime testing infrastructure currently we have a travis ci setup for testing pull requests ensime uses a drone infrastructure with beefier machines so we need to move our setup from travis to drone also involves removing the travis yml file will need to see if tests have any issue running on drone compared to local machine and travis ci
1
114,166
11,839,375,260
IssuesEvent
2020-03-23 17:03:31
spring-projects/spring-framework
https://api.github.com/repos/spring-projects/spring-framework
closed
Clarify name generation responsibility for PropertySourceFactory implementations
in: core type: documentation
**Affects:** spring-context 5.2.3 --- I'd like to discuss an enhancement to the documentation based on an issue that I encountered while implementing a custom `PropertySourceFactory`. Current documentation for the `name()` method of the annotation reads: > Indicate the name of this property source. If omitted, a name will be generated based on the description of the underlying resource. I implemented my own factory not worrying about _nullable_ name parameter of method `createPropertySource`. When I used properties without explicit I encountered an issue - my resources were overridden instead of merged themselves because of the same name (null) and thus same hashCode (`ConfigurationClassParser` lines [#437](https://github.com/spring-projects/spring-framework/blob/v5.2.3.RELEASE/spring-context/src/main/java/org/springframework/context/annotation/ConfigurationClassParser.java#L437) and [#475](https://github.com/spring-projects/spring-framework/blob/v5.2.3.RELEASE/spring-context/src/main/java/org/springframework/context/annotation/ConfigurationClassParser.java#L475)). This led me to believe there is a bug and the name is not generated as the documentation states. Yet, it is - but only by `DefaultPropertySourceFactory` ([DefaultPropertySourceFactory#37](https://github.com/spring-projects/spring-framework/blob/v5.2.3.RELEASE/spring-core/src/main/java/org/springframework/core/io/support/DefaultPropertySourceFactory.java#L37) and [ResourcePropertySource#67](https://github.com/spring-projects/spring-framework/blob/v5.2.3.RELEASE/spring-core/src/main/java/org/springframework/core/io/support/ResourcePropertySource.java)). Suggestion from top of my head: 1. update the documentation to state that generation is performed only by the `DefaultPropertySourceFactory` 2. update the contract of the `PropertySourceFactory` to be responsible for the name generation 3. make the `@PropertySource.name()` required && notBlank and skip generation
1.0
Clarify name generation responsibility for PropertySourceFactory implementations - **Affects:** spring-context 5.2.3 --- I'd like to discuss an enhancement to the documentation based on an issue that I encountered while implementing a custom `PropertySourceFactory`. Current documentation for the `name()` method of the annotation reads: > Indicate the name of this property source. If omitted, a name will be generated based on the description of the underlying resource. I implemented my own factory not worrying about _nullable_ name parameter of method `createPropertySource`. When I used properties without explicit I encountered an issue - my resources were overridden instead of merged themselves because of the same name (null) and thus same hashCode (`ConfigurationClassParser` lines [#437](https://github.com/spring-projects/spring-framework/blob/v5.2.3.RELEASE/spring-context/src/main/java/org/springframework/context/annotation/ConfigurationClassParser.java#L437) and [#475](https://github.com/spring-projects/spring-framework/blob/v5.2.3.RELEASE/spring-context/src/main/java/org/springframework/context/annotation/ConfigurationClassParser.java#L475)). This led me to believe there is a bug and the name is not generated as the documentation states. Yet, it is - but only by `DefaultPropertySourceFactory` ([DefaultPropertySourceFactory#37](https://github.com/spring-projects/spring-framework/blob/v5.2.3.RELEASE/spring-core/src/main/java/org/springframework/core/io/support/DefaultPropertySourceFactory.java#L37) and [ResourcePropertySource#67](https://github.com/spring-projects/spring-framework/blob/v5.2.3.RELEASE/spring-core/src/main/java/org/springframework/core/io/support/ResourcePropertySource.java)). Suggestion from top of my head: 1. update the documentation to state that generation is performed only by the `DefaultPropertySourceFactory` 2. update the contract of the `PropertySourceFactory` to be responsible for the name generation 3. make the `@PropertySource.name()` required && notBlank and skip generation
non_test
clarify name generation responsibility for propertysourcefactory implementations affects spring context i d like to discuss an enhancement to the documentation based on an issue that i encountered while implementing a custom propertysourcefactory current documentation for the name method of the annotation reads indicate the name of this property source if omitted a name will be generated based on the description of the underlying resource i implemented my own factory not worrying about nullable name parameter of method createpropertysource when i used properties without explicit i encountered an issue my resources were overridden instead of merged themselves because of the same name null and thus same hashcode configurationclassparser lines and this led me to believe there is a bug and the name is not generated as the documentation states yet it is but only by defaultpropertysourcefactory and suggestion from top of my head update the documentation to state that generation is performed only by the defaultpropertysourcefactory update the contract of the propertysourcefactory to be responsible for the name generation make the propertysource name required notblank and skip generation
0
125,586
17,836,390,302
IssuesEvent
2021-09-03 02:07:29
Kijacode/dotfiles
https://api.github.com/repos/Kijacode/dotfiles
opened
CVE-2021-37712 (High) detected in tar-4.4.1.tgz
security vulnerability
## CVE-2021-37712 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tar-4.4.1.tgz</b></p></summary> <p>tar for node</p> <p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-4.4.1.tgz">https://registry.npmjs.org/tar/-/tar-4.4.1.tgz</a></p> <p>Path to dependency file: dotfiles/.vscode/extensions/angulardoc.angulardoc-vscode-6.1.3/package.json</p> <p>Path to vulnerable library: dotfiles/.vscode/extensions/angulardoc.angulardoc-vscode-6.1.3/node_modules/fsevents/node_modules/tar/package.json</p> <p> Dependency Hierarchy: - compiler-cli-6.0.5.tgz (Root Library) - chokidar-1.7.0.tgz - fsevents-1.2.4.tgz - node-pre-gyp-0.10.0.tgz - :x: **tar-4.4.1.tgz** (Vulnerable Library) <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The npm package "tar" (aka node-tar) before versions 4.4.18, 5.0.10, and 6.1.9 has an arbitrary file creation/overwrite and arbitrary code execution vulnerability. node-tar aims to guarantee that any file whose location would be modified by a symbolic link is not extracted. This is, in part, achieved by ensuring that extracted directories are not symlinks. Additionally, in order to prevent unnecessary stat calls to determine whether a given path is a directory, paths are cached when directories are created. This logic was insufficient when extracting tar files that contained both a directory and a symlink with names containing unicode values that normalized to the same value. Additionally, on Windows systems, long path portions would resolve to the same file system entities as their 8.3 "short path" counterparts. A specially crafted tar archive could thus include a directory with one form of the path, followed by a symbolic link with a different string that resolves to the same file system entity, followed by a file using the first form. By first creating a directory, and then replacing that directory with a symlink that had a different apparent name that resolved to the same entry in the filesystem, it was thus possible to bypass node-tar symlink checks on directories, essentially allowing an untrusted tar file to symlink into an arbitrary location and subsequently extracting arbitrary files into that location, thus allowing arbitrary file creation and overwrite. These issues were addressed in releases 4.4.18, 5.0.10 and 6.1.9. The v3 branch of node-tar has been deprecated and did not receive patches for these issues. If you are still using a v3 release we recommend you update to a more recent version of node-tar. If this is not possible, a workaround is available in the referenced GHSA-qq89-hq3f-393p. <p>Publish Date: 2021-08-31 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37712>CVE-2021-37712</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.2</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/npm/node-tar/security/advisories/GHSA-qq89-hq3f-393p">https://github.com/npm/node-tar/security/advisories/GHSA-qq89-hq3f-393p</a></p> <p>Release Date: 2021-08-31</p> <p>Fix Resolution: tar - 4.4.18, 5.0.10, 6.1.9</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-37712 (High) detected in tar-4.4.1.tgz - ## CVE-2021-37712 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tar-4.4.1.tgz</b></p></summary> <p>tar for node</p> <p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-4.4.1.tgz">https://registry.npmjs.org/tar/-/tar-4.4.1.tgz</a></p> <p>Path to dependency file: dotfiles/.vscode/extensions/angulardoc.angulardoc-vscode-6.1.3/package.json</p> <p>Path to vulnerable library: dotfiles/.vscode/extensions/angulardoc.angulardoc-vscode-6.1.3/node_modules/fsevents/node_modules/tar/package.json</p> <p> Dependency Hierarchy: - compiler-cli-6.0.5.tgz (Root Library) - chokidar-1.7.0.tgz - fsevents-1.2.4.tgz - node-pre-gyp-0.10.0.tgz - :x: **tar-4.4.1.tgz** (Vulnerable Library) <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The npm package "tar" (aka node-tar) before versions 4.4.18, 5.0.10, and 6.1.9 has an arbitrary file creation/overwrite and arbitrary code execution vulnerability. node-tar aims to guarantee that any file whose location would be modified by a symbolic link is not extracted. This is, in part, achieved by ensuring that extracted directories are not symlinks. Additionally, in order to prevent unnecessary stat calls to determine whether a given path is a directory, paths are cached when directories are created. This logic was insufficient when extracting tar files that contained both a directory and a symlink with names containing unicode values that normalized to the same value. Additionally, on Windows systems, long path portions would resolve to the same file system entities as their 8.3 "short path" counterparts. A specially crafted tar archive could thus include a directory with one form of the path, followed by a symbolic link with a different string that resolves to the same file system entity, followed by a file using the first form. By first creating a directory, and then replacing that directory with a symlink that had a different apparent name that resolved to the same entry in the filesystem, it was thus possible to bypass node-tar symlink checks on directories, essentially allowing an untrusted tar file to symlink into an arbitrary location and subsequently extracting arbitrary files into that location, thus allowing arbitrary file creation and overwrite. These issues were addressed in releases 4.4.18, 5.0.10 and 6.1.9. The v3 branch of node-tar has been deprecated and did not receive patches for these issues. If you are still using a v3 release we recommend you update to a more recent version of node-tar. If this is not possible, a workaround is available in the referenced GHSA-qq89-hq3f-393p. <p>Publish Date: 2021-08-31 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37712>CVE-2021-37712</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.2</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/npm/node-tar/security/advisories/GHSA-qq89-hq3f-393p">https://github.com/npm/node-tar/security/advisories/GHSA-qq89-hq3f-393p</a></p> <p>Release Date: 2021-08-31</p> <p>Fix Resolution: tar - 4.4.18, 5.0.10, 6.1.9</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in tar tgz cve high severity vulnerability vulnerable library tar tgz tar for node library home page a href path to dependency file dotfiles vscode extensions angulardoc angulardoc vscode package json path to vulnerable library dotfiles vscode extensions angulardoc angulardoc vscode node modules fsevents node modules tar package json dependency hierarchy compiler cli tgz root library chokidar tgz fsevents tgz node pre gyp tgz x tar tgz vulnerable library found in base branch main vulnerability details the npm package tar aka node tar before versions and has an arbitrary file creation overwrite and arbitrary code execution vulnerability node tar aims to guarantee that any file whose location would be modified by a symbolic link is not extracted this is in part achieved by ensuring that extracted directories are not symlinks additionally in order to prevent unnecessary stat calls to determine whether a given path is a directory paths are cached when directories are created this logic was insufficient when extracting tar files that contained both a directory and a symlink with names containing unicode values that normalized to the same value additionally on windows systems long path portions would resolve to the same file system entities as their short path counterparts a specially crafted tar archive could thus include a directory with one form of the path followed by a symbolic link with a different string that resolves to the same file system entity followed by a file using the first form by first creating a directory and then replacing that directory with a symlink that had a different apparent name that resolved to the same entry in the filesystem it was thus possible to bypass node tar symlink checks on directories essentially allowing an untrusted tar file to symlink into an arbitrary location and subsequently extracting arbitrary files into that location thus allowing arbitrary file creation and overwrite these issues were addressed in releases and the branch of node tar has been deprecated and did not receive patches for these issues if you are still using a release we recommend you update to a more recent version of node tar if this is not possible a workaround is available in the referenced ghsa publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact high integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tar step up your open source security game with whitesource
0
67,212
12,888,202,644
IssuesEvent
2020-07-13 12:38:14
topcoder-platform/community-app
https://api.github.com/repos/topcoder-platform/community-app
closed
[$40]User details not displayed on header
Beta Env Challenge Listing Merged for QA P2 QA Pass tcx_Assigned tcx_FixAccepted tcx_ReadyForReview v5-intgration-sub-code
User details not displayed on header https://beta-community-app.topcoder.com/challenges? <img width="1440" alt="Screenshot 2020-06-23 at 3 24 37 PM" src="https://user-images.githubusercontent.com/58783823/85389992-cdf60c00-b565-11ea-9866-4c11f9d73b85.png"> Looks good on test env, <img width="1440" alt="Screenshot 2020-06-23 at 3 26 10 PM" src="https://user-images.githubusercontent.com/58783823/85390109-f67e0600-b565-11ea-841c-708dfd6273f5.png">
1.0
[$40]User details not displayed on header - User details not displayed on header https://beta-community-app.topcoder.com/challenges? <img width="1440" alt="Screenshot 2020-06-23 at 3 24 37 PM" src="https://user-images.githubusercontent.com/58783823/85389992-cdf60c00-b565-11ea-9866-4c11f9d73b85.png"> Looks good on test env, <img width="1440" alt="Screenshot 2020-06-23 at 3 26 10 PM" src="https://user-images.githubusercontent.com/58783823/85390109-f67e0600-b565-11ea-841c-708dfd6273f5.png">
non_test
user details not displayed on header user details not displayed on header img width alt screenshot at pm src looks good on test env img width alt screenshot at pm src
0
328,706
28,131,436,728
IssuesEvent
2023-03-31 23:58:11
microsoft/ebpf-for-windows
https://api.github.com/repos/microsoft/ebpf-for-windows
reopened
Simulate NMR and OS API failures
enhancement triaged tests
This issue is to add tests for the following: Add tests to simulate systematic failure of NMR and OS API failures for both ebpfcore and netebpfext. This can use the same framework currently used for low memory simulation in user mode. **NMR APIs:** 1. NmrRegisterProvider 2. NmrDeregisterProvider 3. NmrWaitForProviderDeregisterComplete 4. NmrClientAttachProvider 5. NmrRegisterClient 6. NmrDeregisterClient 7. NmrWaitForClientDeregisterComplete OS APIs: 8. ebpf_platform_user.cpp 9. fwp_um.cpp 10.
1.0
Simulate NMR and OS API failures - This issue is to add tests for the following: Add tests to simulate systematic failure of NMR and OS API failures for both ebpfcore and netebpfext. This can use the same framework currently used for low memory simulation in user mode. **NMR APIs:** 1. NmrRegisterProvider 2. NmrDeregisterProvider 3. NmrWaitForProviderDeregisterComplete 4. NmrClientAttachProvider 5. NmrRegisterClient 6. NmrDeregisterClient 7. NmrWaitForClientDeregisterComplete OS APIs: 8. ebpf_platform_user.cpp 9. fwp_um.cpp 10.
test
simulate nmr and os api failures this issue is to add tests for the following add tests to simulate systematic failure of nmr and os api failures for both ebpfcore and netebpfext this can use the same framework currently used for low memory simulation in user mode nmr apis nmrregisterprovider nmrderegisterprovider nmrwaitforproviderderegistercomplete nmrclientattachprovider nmrregisterclient nmrderegisterclient nmrwaitforclientderegistercomplete os apis ebpf platform user cpp fwp um cpp
1
24,958
7,605,628,304
IssuesEvent
2018-04-30 09:49:42
MarlinFirmware/Marlin
https://api.github.com/repos/MarlinFirmware/Marlin
closed
[bugfix-2.0.x] Error compiling ArduinoIDE
C: Build / Toolchain
I´ve got this error trying to compile Marlin 2.0: ``` Arduino: 1.8.5 (Windows 10), Board: "Arduino/Genuino Mega or Mega 2560, ATmega2560 (Mega 2560)" fork/exec C:\Program Files (x86)\Arduino\hardware\tools\avr/bin/avr-gcc.exe: file name or extension too big. Error compiling for board Arduino/Genuino Mega or Mega 2560. ```
1.0
[bugfix-2.0.x] Error compiling ArduinoIDE - I´ve got this error trying to compile Marlin 2.0: ``` Arduino: 1.8.5 (Windows 10), Board: "Arduino/Genuino Mega or Mega 2560, ATmega2560 (Mega 2560)" fork/exec C:\Program Files (x86)\Arduino\hardware\tools\avr/bin/avr-gcc.exe: file name or extension too big. Error compiling for board Arduino/Genuino Mega or Mega 2560. ```
non_test
error compiling arduinoide i´ve got this error trying to compile marlin arduino windows board arduino genuino mega or mega mega fork exec c program files arduino hardware tools avr bin avr gcc exe file name or extension too big error compiling for board arduino genuino mega or mega
0
313,082
9,556,795,881
IssuesEvent
2019-05-03 09:29:02
bojzi/angular-sandbox
https://api.github.com/repos/bojzi/angular-sandbox
closed
Direct links are not working when deployed
Priority: Critical Status: In Progress Type: Bug
Example: https://poslek.com/angular-sandbox/ivysaur It should produce the Pokemon Details page for Ivysaur and not a 404.
1.0
Direct links are not working when deployed - Example: https://poslek.com/angular-sandbox/ivysaur It should produce the Pokemon Details page for Ivysaur and not a 404.
non_test
direct links are not working when deployed example it should produce the pokemon details page for ivysaur and not a
0
221,124
17,292,018,073
IssuesEvent
2021-07-25 00:27:52
mgba-emu/mgba
https://api.github.com/repos/mgba-emu/mgba
closed
White screen on 2Ds and screen is halfway cut off
blocked:needs retest
I homebrewed my ds and downloaded this I installed pokemon fire red and when I go into it I see a write screen with unimportant text at the bottom I click around and when I Click on the shoulder buttons I hear a weird sound, so then I open it with the app and the screen is halfway off screen to where I can only soo the left side of the screen any suggestions
1.0
White screen on 2Ds and screen is halfway cut off - I homebrewed my ds and downloaded this I installed pokemon fire red and when I go into it I see a write screen with unimportant text at the bottom I click around and when I Click on the shoulder buttons I hear a weird sound, so then I open it with the app and the screen is halfway off screen to where I can only soo the left side of the screen any suggestions
test
white screen on and screen is halfway cut off i homebrewed my ds and downloaded this i installed pokemon fire red and when i go into it i see a write screen with unimportant text at the bottom i click around and when i click on the shoulder buttons i hear a weird sound so then i open it with the app and the screen is halfway off screen to where i can only soo the left side of the screen any suggestions
1
308,521
26,610,785,059
IssuesEvent
2023-01-23 23:55:58
PalisadoesFoundation/talawa-api
https://api.github.com/repos/PalisadoesFoundation/talawa-api
closed
Resolvers: Create tests for MembershipRequest.js
good first issue unapproved points 02 test
The Talawa-API code base needs to be 100% reliable. This means we need to have 100% test code coverage. Tests need to be written for file `lib/resolvers/MembershipRequest.js` - We will need the API to be refactored for all methods, classes and/or functions found in this file for testing to be correctly executed. - When complete, all all methods, classes and/or functions in the refactored file will need to be tested. These tests must be placed in a single file with the name `tests/resolvers/MembershipRequest.spec.js`. You may need to create the appropriate directory structure to do this. ### IMPORTANT: Please refer to the parent issue on how to implement these tests correctly: - https://github.com/PalisadoesFoundation/talawa-api/issues/490 ### PR Acceptance Criteria - When complete this file must show **100%** coverage when merged into the code base. - [The current code coverage for the file can be found here](https://codecov.io/gh/PalisadoesFoundation/talawa-api/tree/develop/lib/resolvers/). If the file isn't found in this directory, then tests have not been created. - The PR will show a report for the code coverage for the file you have added. You can use that as a guide.
1.0
Resolvers: Create tests for MembershipRequest.js - The Talawa-API code base needs to be 100% reliable. This means we need to have 100% test code coverage. Tests need to be written for file `lib/resolvers/MembershipRequest.js` - We will need the API to be refactored for all methods, classes and/or functions found in this file for testing to be correctly executed. - When complete, all all methods, classes and/or functions in the refactored file will need to be tested. These tests must be placed in a single file with the name `tests/resolvers/MembershipRequest.spec.js`. You may need to create the appropriate directory structure to do this. ### IMPORTANT: Please refer to the parent issue on how to implement these tests correctly: - https://github.com/PalisadoesFoundation/talawa-api/issues/490 ### PR Acceptance Criteria - When complete this file must show **100%** coverage when merged into the code base. - [The current code coverage for the file can be found here](https://codecov.io/gh/PalisadoesFoundation/talawa-api/tree/develop/lib/resolvers/). If the file isn't found in this directory, then tests have not been created. - The PR will show a report for the code coverage for the file you have added. You can use that as a guide.
test
resolvers create tests for membershiprequest js the talawa api code base needs to be reliable this means we need to have test code coverage tests need to be written for file lib resolvers membershiprequest js we will need the api to be refactored for all methods classes and or functions found in this file for testing to be correctly executed when complete all all methods classes and or functions in the refactored file will need to be tested these tests must be placed in a single file with the name tests resolvers membershiprequest spec js you may need to create the appropriate directory structure to do this important please refer to the parent issue on how to implement these tests correctly pr acceptance criteria when complete this file must show coverage when merged into the code base if the file isn t found in this directory then tests have not been created the pr will show a report for the code coverage for the file you have added you can use that as a guide
1
263,315
23,048,256,158
IssuesEvent
2022-07-24 08:23:26
Rebirth-of-the-Night/Rebirth-Of-The-Night
https://api.github.com/repos/Rebirth-of-the-Night/Rebirth-Of-The-Night
closed
[BUG] XP Dupe with Xorcite and Biotite
mod bug needs testing
<!-- ISSUES NOT USING OUR TEMPLATE WILL NOT BE CONSIDERED AND WILL BE CLOSED! --> **Describe the bug** <!-- A clear and concise description of what the bug is. -->** Right clicking Xorcite ore after mining biotite will give you an infinite amount of experience. **To Reproduce** Steps to reproduce the behavior: Get to the end, find xorcite, wait for it to be fully matured, mine a single biotite ore, right click xorcite ore to your hearts content. **Expected behavior** <!-- A clear and concise description of what you expected to happen. --> Not sure lol **Info (please complete the following information):** - Modpack Version: 2.77.5 - Server or Singleplayer: Server - Version the world was created on: 2.77.5
1.0
[BUG] XP Dupe with Xorcite and Biotite - <!-- ISSUES NOT USING OUR TEMPLATE WILL NOT BE CONSIDERED AND WILL BE CLOSED! --> **Describe the bug** <!-- A clear and concise description of what the bug is. -->** Right clicking Xorcite ore after mining biotite will give you an infinite amount of experience. **To Reproduce** Steps to reproduce the behavior: Get to the end, find xorcite, wait for it to be fully matured, mine a single biotite ore, right click xorcite ore to your hearts content. **Expected behavior** <!-- A clear and concise description of what you expected to happen. --> Not sure lol **Info (please complete the following information):** - Modpack Version: 2.77.5 - Server or Singleplayer: Server - Version the world was created on: 2.77.5
test
xp dupe with xorcite and biotite describe the bug right clicking xorcite ore after mining biotite will give you an infinite amount of experience to reproduce steps to reproduce the behavior get to the end find xorcite wait for it to be fully matured mine a single biotite ore right click xorcite ore to your hearts content expected behavior not sure lol info please complete the following information modpack version server or singleplayer server version the world was created on
1
63,674
12,366,946,879
IssuesEvent
2020-05-18 11:24:14
ElaSparks/the15puzzle
https://api.github.com/repos/ElaSparks/the15puzzle
closed
To complete the interface.
Code Database Develop Fix Testing
<p> 1. Доделать интерфейс. </p> <p> 2. Исправить ошибку malloc_consolidate() invalid chunk size. </p> @OldBoots @ElaSparks
1.0
To complete the interface. - <p> 1. Доделать интерфейс. </p> <p> 2. Исправить ошибку malloc_consolidate() invalid chunk size. </p> @OldBoots @ElaSparks
non_test
to complete the interface доделать интерфейс исправить ошибку malloc consolidate invalid chunk size oldboots elasparks
0
270,739
23,532,197,846
IssuesEvent
2022-08-19 16:26:23
mozilla-mobile/fenix
https://api.github.com/repos/mozilla-mobile/fenix
closed
Intermittent Unit Test failure - StartupTypeTelemetryTest.WHEN record is called THEN telemetry is recorded with the appropriate label
eng:intermittent-test
### Test Run: Test failed in https://github.com/mozilla-mobile/fenix/pull/26378 - https://firefoxci.taskcluster-artifacts.net/OIThuwo1TRSMqXOucAYXqg/0/public/reports/test/testDebugUnitTest/index.html ### Stacktrace: > java.lang.AssertionError: expected:<1> but was:<null> at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.failNotEquals(Assert.java:834) at org.junit.Assert.assertEquals(Assert.java:118) at org.junit.Assert.assertEquals(Assert.java:144) at org.mozilla.fenix.perf.StartupTypeTelemetryTest.WHEN record is called THEN telemetry is recorded with the appropriate label(StartupTypeTelemetryTest.kt:96) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:55) at org.junit.rules.RunRules.evaluate(RunRules.java:20) at org.robolectric.RobolectricTestRunner$HelperTestRunner$1.evaluate(RobolectricTestRunner.java:591) at org.robolectric.internal.SandboxTestRunner$2.lambda$evaluate$0(SandboxTestRunner.java:274) at org.robolectric.internal.bytecode.Sandbox.lambda$runOnMainThread$0(Sandbox.java:88) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) at java.base/java.lang.Thread.run(Thread.java:829) Suppressed: org.robolectric.android.internal.AndroidTestEnvironment$UnExecutedRunnablesException: Main looper has queued unexecuted runnables. This might be the cause of the test failure. You might need a shadowOf(Looper.getMainLooper()).idle() call. ### Build: 105 ┆Issue is synchronized with this [Jira Task](https://mozilla-hub.atlassian.net/browse/FNXV2-21233)
1.0
Intermittent Unit Test failure - StartupTypeTelemetryTest.WHEN record is called THEN telemetry is recorded with the appropriate label - ### Test Run: Test failed in https://github.com/mozilla-mobile/fenix/pull/26378 - https://firefoxci.taskcluster-artifacts.net/OIThuwo1TRSMqXOucAYXqg/0/public/reports/test/testDebugUnitTest/index.html ### Stacktrace: > java.lang.AssertionError: expected:<1> but was:<null> at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.failNotEquals(Assert.java:834) at org.junit.Assert.assertEquals(Assert.java:118) at org.junit.Assert.assertEquals(Assert.java:144) at org.mozilla.fenix.perf.StartupTypeTelemetryTest.WHEN record is called THEN telemetry is recorded with the appropriate label(StartupTypeTelemetryTest.kt:96) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:566) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:55) at org.junit.rules.RunRules.evaluate(RunRules.java:20) at org.robolectric.RobolectricTestRunner$HelperTestRunner$1.evaluate(RobolectricTestRunner.java:591) at org.robolectric.internal.SandboxTestRunner$2.lambda$evaluate$0(SandboxTestRunner.java:274) at org.robolectric.internal.bytecode.Sandbox.lambda$runOnMainThread$0(Sandbox.java:88) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) at java.base/java.lang.Thread.run(Thread.java:829) Suppressed: org.robolectric.android.internal.AndroidTestEnvironment$UnExecutedRunnablesException: Main looper has queued unexecuted runnables. This might be the cause of the test failure. You might need a shadowOf(Looper.getMainLooper()).idle() call. ### Build: 105 ┆Issue is synchronized with this [Jira Task](https://mozilla-hub.atlassian.net/browse/FNXV2-21233)
test
intermittent unit test failure startuptypetelemetrytest when record is called then telemetry is recorded with the appropriate label test run test failed in stacktrace java lang assertionerror expected but was at org junit assert fail assert java at org junit assert failnotequals assert java at org junit assert assertequals assert java at org junit assert assertequals assert java at org mozilla fenix perf startuptypetelemetrytest when record is called then telemetry is recorded with the appropriate label startuptypetelemetrytest kt at java base jdk internal reflect nativemethodaccessorimpl native method at java base jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at java base jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java base java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at org junit internal runners statements runbefores evaluate runbefores java at org junit rules testwatcher evaluate testwatcher java at org junit rules runrules evaluate runrules java at org robolectric robolectrictestrunner helpertestrunner evaluate robolectrictestrunner java at org robolectric internal sandboxtestrunner lambda evaluate sandboxtestrunner java at org robolectric internal bytecode sandbox lambda runonmainthread sandbox java at java base java util concurrent futuretask run futuretask java at java base java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java base java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java base java lang thread run thread java suppressed org robolectric android internal androidtestenvironment unexecutedrunnablesexception main looper has queued unexecuted runnables this might be the cause of the test failure you might need a shadowof looper getmainlooper idle call build ┆issue is synchronized with this
1
32,446
4,771,791,672
IssuesEvent
2016-10-26 19:00:03
RevolutionAnalytics/checkpoint
https://api.github.com/repos/RevolutionAnalytics/checkpoint
closed
Fix failing unit test
testing
The test for `"does not display message whan scanForPackages=FALSE"` fails.
1.0
Fix failing unit test - The test for `"does not display message whan scanForPackages=FALSE"` fails.
test
fix failing unit test the test for does not display message whan scanforpackages false fails
1
109,507
4,388,646,096
IssuesEvent
2016-08-08 19:33:28
SuperTux/flexlay
https://api.github.com/repos/SuperTux/flexlay
closed
Graphics Update
priority:low type:idea
It looks like the graphics in Flexlay are a little rusty, I found this cool site with tons of free icons: https://icomoon.io/#preview-free You can download them as a number of different formats such as png, pdf, and even svg. So you can edit them!
1.0
Graphics Update - It looks like the graphics in Flexlay are a little rusty, I found this cool site with tons of free icons: https://icomoon.io/#preview-free You can download them as a number of different formats such as png, pdf, and even svg. So you can edit them!
non_test
graphics update it looks like the graphics in flexlay are a little rusty i found this cool site with tons of free icons you can download them as a number of different formats such as png pdf and even svg so you can edit them
0
10,736
3,135,924,895
IssuesEvent
2015-09-10 17:26:15
ManageIQ/manageiq
https://api.github.com/repos/ManageIQ/manageiq
closed
EmsAzure will perform a full refresh RestClient::Unauthorized: 401 Unauthorized error
bug/sporadic test failure
I saw this on a seemingly unrelated PR test run: https://github.com/ManageIQ/manageiq/pull/4250 ``` 1) EmsAzure will perform a full refresh Failure/Error: Unable to find matching line from backtrace RestClient::Unauthorized: 401 Unauthorized # /home/travis/build/ManageIQ/manageiq/vendor/bundle/ruby/2.2.0/gems/rest-client-2.0.0.rc1/lib/restclient/abstract_response.rb:197:in `exception_with_response' # /home/travis/build/ManageIQ/manageiq/vendor/bundle/ruby/2.2.0/gems/rest-client-2.0.0.rc1/lib/restclient/abstract_response.rb:83:in `return!' # /home/travis/build/ManageIQ/manageiq/vendor/bundle/ruby/2.2.0/gems/rest-client-2.0.0.rc1/lib/restclient/request.rb:594:in `process_result' # /home/travis/build/ManageIQ/manageiq/vendor/bundle/ruby/2.2.0/gems/rest-client-2.0.0.rc1/lib/restclient/request.rb:504:in `block in transmit' # /home/travis/build/ManageIQ/manageiq/vendor/bundle/ruby/2.2.0/gems/webmock-1.21.0/lib/webmock/http_lib_adapters/net_http.rb:123:in `start_without_connect' # /home/travis/build/ManageIQ/manageiq/vendor/bundle/ruby/2.2.0/gems/webmock-1.21.0/lib/webmock/http_lib_adapters/net_http.rb:150:in `start' # /home/travis/build/ManageIQ/manageiq/vendor/bundle/ruby/2.2.0/gems/rest-client-2.0.0.rc1/lib/restclient/request.rb:494:in `transmit' # /home/travis/build/ManageIQ/manageiq/vendor/bundle/ruby/2.2.0/gems/rest-client-2.0.0.rc1/lib/restclient/request.rb:202:in `execute' # /home/travis/build/ManageIQ/manageiq/vendor/bundle/ruby/2.2.0/gems/rest-client-2.0.0.rc1/lib/restclient/request.rb:52:in `execute' # /home/travis/build/ManageIQ/manageiq/vendor/bundle/ruby/2.2.0/gems/rest-client-2.0.0.rc1/lib/restclient.rb:66:in `get' # /home/travis/build/ManageIQ/manageiq/vendor/bundle/ruby/2.2.0/bundler/gems/azure-armrest-e47ff458c05a/lib/azure/armrest/armrest_service.rb:310:in `rest_get' # /home/travis/build/ManageIQ/manageiq/vendor/bundle/ruby/2.2.0/bundler/gems/azure-armrest-e47ff458c05a/lib/azure/armrest/virtual_machine_service.rb:85:in `block (2 levels) in list' ... rspec ./spec/models/ems_refresh/refreshers/azure_refresher_spec.rb:16 # EmsAzure will perform a full refresh Randomized with seed 11108 ``` cc @bronaghs @djberg96 (have you seen this before?)
1.0
EmsAzure will perform a full refresh RestClient::Unauthorized: 401 Unauthorized error - I saw this on a seemingly unrelated PR test run: https://github.com/ManageIQ/manageiq/pull/4250 ``` 1) EmsAzure will perform a full refresh Failure/Error: Unable to find matching line from backtrace RestClient::Unauthorized: 401 Unauthorized # /home/travis/build/ManageIQ/manageiq/vendor/bundle/ruby/2.2.0/gems/rest-client-2.0.0.rc1/lib/restclient/abstract_response.rb:197:in `exception_with_response' # /home/travis/build/ManageIQ/manageiq/vendor/bundle/ruby/2.2.0/gems/rest-client-2.0.0.rc1/lib/restclient/abstract_response.rb:83:in `return!' # /home/travis/build/ManageIQ/manageiq/vendor/bundle/ruby/2.2.0/gems/rest-client-2.0.0.rc1/lib/restclient/request.rb:594:in `process_result' # /home/travis/build/ManageIQ/manageiq/vendor/bundle/ruby/2.2.0/gems/rest-client-2.0.0.rc1/lib/restclient/request.rb:504:in `block in transmit' # /home/travis/build/ManageIQ/manageiq/vendor/bundle/ruby/2.2.0/gems/webmock-1.21.0/lib/webmock/http_lib_adapters/net_http.rb:123:in `start_without_connect' # /home/travis/build/ManageIQ/manageiq/vendor/bundle/ruby/2.2.0/gems/webmock-1.21.0/lib/webmock/http_lib_adapters/net_http.rb:150:in `start' # /home/travis/build/ManageIQ/manageiq/vendor/bundle/ruby/2.2.0/gems/rest-client-2.0.0.rc1/lib/restclient/request.rb:494:in `transmit' # /home/travis/build/ManageIQ/manageiq/vendor/bundle/ruby/2.2.0/gems/rest-client-2.0.0.rc1/lib/restclient/request.rb:202:in `execute' # /home/travis/build/ManageIQ/manageiq/vendor/bundle/ruby/2.2.0/gems/rest-client-2.0.0.rc1/lib/restclient/request.rb:52:in `execute' # /home/travis/build/ManageIQ/manageiq/vendor/bundle/ruby/2.2.0/gems/rest-client-2.0.0.rc1/lib/restclient.rb:66:in `get' # /home/travis/build/ManageIQ/manageiq/vendor/bundle/ruby/2.2.0/bundler/gems/azure-armrest-e47ff458c05a/lib/azure/armrest/armrest_service.rb:310:in `rest_get' # /home/travis/build/ManageIQ/manageiq/vendor/bundle/ruby/2.2.0/bundler/gems/azure-armrest-e47ff458c05a/lib/azure/armrest/virtual_machine_service.rb:85:in `block (2 levels) in list' ... rspec ./spec/models/ems_refresh/refreshers/azure_refresher_spec.rb:16 # EmsAzure will perform a full refresh Randomized with seed 11108 ``` cc @bronaghs @djberg96 (have you seen this before?)
test
emsazure will perform a full refresh restclient unauthorized unauthorized error i saw this on a seemingly unrelated pr test run emsazure will perform a full refresh failure error unable to find matching line from backtrace restclient unauthorized unauthorized home travis build manageiq manageiq vendor bundle ruby gems rest client lib restclient abstract response rb in exception with response home travis build manageiq manageiq vendor bundle ruby gems rest client lib restclient abstract response rb in return home travis build manageiq manageiq vendor bundle ruby gems rest client lib restclient request rb in process result home travis build manageiq manageiq vendor bundle ruby gems rest client lib restclient request rb in block in transmit home travis build manageiq manageiq vendor bundle ruby gems webmock lib webmock http lib adapters net http rb in start without connect home travis build manageiq manageiq vendor bundle ruby gems webmock lib webmock http lib adapters net http rb in start home travis build manageiq manageiq vendor bundle ruby gems rest client lib restclient request rb in transmit home travis build manageiq manageiq vendor bundle ruby gems rest client lib restclient request rb in execute home travis build manageiq manageiq vendor bundle ruby gems rest client lib restclient request rb in execute home travis build manageiq manageiq vendor bundle ruby gems rest client lib restclient rb in get home travis build manageiq manageiq vendor bundle ruby bundler gems azure armrest lib azure armrest armrest service rb in rest get home travis build manageiq manageiq vendor bundle ruby bundler gems azure armrest lib azure armrest virtual machine service rb in block levels in list rspec spec models ems refresh refreshers azure refresher spec rb emsazure will perform a full refresh randomized with seed cc bronaghs have you seen this before
1
346,800
31,025,387,918
IssuesEvent
2023-08-10 08:47:45
brave/brave-browser
https://api.github.com/repos/brave/brave-browser
closed
Date format for `Estimated Earnings` differs between `brave://rewards` and Rewards panel
bug feature/rewards priority/P4 QA/Yes QA/Test-Plan-Specified OS/Desktop
<!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue. PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE. INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED--> ## Description <!--Provide a brief description of the issue--> Date format for `Estimated Earnings` differs between `brave://rewards` and Rewards panel Spun off from https://github.com/brave/brave-browser/issues/27193 ## Steps to Reproduce <!--Please add a series of steps to reproduce the issue--> 1. install `1.48.148` 2. launch Brave 3. set Rewards env. to staging via `QA Preferences` 4. tap the `BAT` icon 5. tap `Start using Brave Rewards` 6. tap `Connect` 7. tap on `Uphold` 8. log in 9. wait 10. compare the date range format above `Estimated` earnings, between`brave://rewards` and the panel ## Actual result: <!--Please add screenshots if needed--> It's `Jan 1 - Jan 31` on `brave://rewards` It's `1 Jan - 31 Jan` on the Rewards panel example | example ---------|--------- ![Screenshot_20230130-212405](https://user-images.githubusercontent.com/387249/215674560-24e6afbe-6a92-4b03-8047-aad099d5eefb.png) | ![Screenshot_20230130-212548](https://user-images.githubusercontent.com/387249/215674568-fe4c9f20-b098-4a7e-8e0c-c1d4084089a6.png) ## Expected result: Probably should be `Jan 1 - Jan 31` ## Reproduces how often: <!--[Easily reproduced/Intermittent issue/No steps to reproduce]--> ## Brave version (brave://version info) <!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details--> `1.48.148`, `Chromium` `109.0.5414.119` on a `Google Pixel XL` (`arm64`) running `Android 9` ## Version/Channel Information: <!--Does this issue happen on any other channels? Or is it specific to a certain channel?--> - Can you reproduce this issue with the current release? - Can you reproduce this issue with the beta channel? - Can you reproduce this issue with the nightly channel? ## Other Additional Information: - Does the issue resolve itself when disabling Brave Shields? - Does the issue resolve itself when disabling Brave Rewards? - Is the issue reproducible on the latest version of Chrome? ## Miscellaneous Information: <!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue--> cc @szilardszaloki @Miyayes @LaurenWags @zenparsing @deeppandya
1.0
Date format for `Estimated Earnings` differs between `brave://rewards` and Rewards panel - <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue. PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE. INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED--> ## Description <!--Provide a brief description of the issue--> Date format for `Estimated Earnings` differs between `brave://rewards` and Rewards panel Spun off from https://github.com/brave/brave-browser/issues/27193 ## Steps to Reproduce <!--Please add a series of steps to reproduce the issue--> 1. install `1.48.148` 2. launch Brave 3. set Rewards env. to staging via `QA Preferences` 4. tap the `BAT` icon 5. tap `Start using Brave Rewards` 6. tap `Connect` 7. tap on `Uphold` 8. log in 9. wait 10. compare the date range format above `Estimated` earnings, between`brave://rewards` and the panel ## Actual result: <!--Please add screenshots if needed--> It's `Jan 1 - Jan 31` on `brave://rewards` It's `1 Jan - 31 Jan` on the Rewards panel example | example ---------|--------- ![Screenshot_20230130-212405](https://user-images.githubusercontent.com/387249/215674560-24e6afbe-6a92-4b03-8047-aad099d5eefb.png) | ![Screenshot_20230130-212548](https://user-images.githubusercontent.com/387249/215674568-fe4c9f20-b098-4a7e-8e0c-c1d4084089a6.png) ## Expected result: Probably should be `Jan 1 - Jan 31` ## Reproduces how often: <!--[Easily reproduced/Intermittent issue/No steps to reproduce]--> ## Brave version (brave://version info) <!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details--> `1.48.148`, `Chromium` `109.0.5414.119` on a `Google Pixel XL` (`arm64`) running `Android 9` ## Version/Channel Information: <!--Does this issue happen on any other channels? Or is it specific to a certain channel?--> - Can you reproduce this issue with the current release? - Can you reproduce this issue with the beta channel? - Can you reproduce this issue with the nightly channel? ## Other Additional Information: - Does the issue resolve itself when disabling Brave Shields? - Does the issue resolve itself when disabling Brave Rewards? - Is the issue reproducible on the latest version of Chrome? ## Miscellaneous Information: <!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue--> cc @szilardszaloki @Miyayes @LaurenWags @zenparsing @deeppandya
test
date format for estimated earnings differs between brave rewards and rewards panel have you searched for similar issues before submitting this issue please check the open issues and add a note before logging a new issue please use the template below to provide information about the issue insufficient info will get the issue closed it will only be reopened after sufficient info is provided description date format for estimated earnings differs between brave rewards and rewards panel spun off from steps to reproduce install launch brave set rewards env to staging via qa preferences tap the bat icon tap start using brave rewards tap connect tap on uphold log in wait compare the date range format above estimated earnings between brave rewards and the panel actual result it s jan jan on brave rewards it s jan jan on the rewards panel example example expected result probably should be jan jan reproduces how often brave version brave version info chromium on a google pixel xl running android version channel information can you reproduce this issue with the current release can you reproduce this issue with the beta channel can you reproduce this issue with the nightly channel other additional information does the issue resolve itself when disabling brave shields does the issue resolve itself when disabling brave rewards is the issue reproducible on the latest version of chrome miscellaneous information cc szilardszaloki miyayes laurenwags zenparsing deeppandya
1
191,435
14,594,246,424
IssuesEvent
2020-12-20 04:21:19
github-vet/rangeloop-pointer-findings
https://api.github.com/repos/github-vet/rangeloop-pointer-findings
closed
jhamon42/goinfre: Developer/goroot/src/net/timeout_test.go; 64 LoC
fresh medium test
Found a possible issue in [jhamon42/goinfre](https://www.github.com/jhamon42/goinfre) at [Developer/goroot/src/net/timeout_test.go](https://github.com/jhamon42/goinfre/blob/61860eeeeff292ed4abb4fc46032ff41271d1b0b/Developer/goroot/src/net/timeout_test.go#L843-L906) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > range-loop variable timeout used in defer or goroutine at line 886 [Click here to see the code in its original context.](https://github.com/jhamon42/goinfre/blob/61860eeeeff292ed4abb4fc46032ff41271d1b0b/Developer/goroot/src/net/timeout_test.go#L843-L906) <details> <summary>Click here to show the 64 line(s) of Go which triggered the analyzer.</summary> ```go for _, timeout := range []time.Duration{ 1 * time.Nanosecond, 2 * time.Nanosecond, 5 * time.Nanosecond, 50 * time.Nanosecond, 100 * time.Nanosecond, 200 * time.Nanosecond, 500 * time.Nanosecond, 750 * time.Nanosecond, 1 * time.Microsecond, 5 * time.Microsecond, 25 * time.Microsecond, 250 * time.Microsecond, 500 * time.Microsecond, 1 * time.Millisecond, 5 * time.Millisecond, 100 * time.Millisecond, 250 * time.Millisecond, 500 * time.Millisecond, 1 * time.Second, } { numRuns := 3 if testing.Short() { numRuns = 1 if timeout > 500*time.Microsecond { continue } } for run := 0; run < numRuns; run++ { name := fmt.Sprintf("%v %d/%d", timeout, run, numRuns) t.Log(name) tooSlow := time.NewTimer(5 * time.Second) defer tooSlow.Stop() c, err := Dial(ls.Listener.Addr().Network(), ls.Listener.Addr().String()) if err != nil { t.Fatal(err) } ch := make(chan result, 1) go func() { t0 := time.Now() if err := c.SetDeadline(t0.Add(timeout)); err != nil { t.Error(err) } n, err := io.Copy(ioutil.Discard, c) dt := time.Since(t0) c.Close() ch <- result{n, err, dt} }() select { case res := <-ch: if nerr, ok := res.err.(Error); ok && nerr.Timeout() { t.Logf("%v: good timeout after %v; %d bytes", name, res.d, res.n) } else { t.Fatalf("%v: Copy = %d, %v; want timeout", name, res.n, res.err) } case <-tooSlow.C: t.Fatalf("%v: client stuck in Dial+Copy", name) } } } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 61860eeeeff292ed4abb4fc46032ff41271d1b0b
1.0
jhamon42/goinfre: Developer/goroot/src/net/timeout_test.go; 64 LoC - Found a possible issue in [jhamon42/goinfre](https://www.github.com/jhamon42/goinfre) at [Developer/goroot/src/net/timeout_test.go](https://github.com/jhamon42/goinfre/blob/61860eeeeff292ed4abb4fc46032ff41271d1b0b/Developer/goroot/src/net/timeout_test.go#L843-L906) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > range-loop variable timeout used in defer or goroutine at line 886 [Click here to see the code in its original context.](https://github.com/jhamon42/goinfre/blob/61860eeeeff292ed4abb4fc46032ff41271d1b0b/Developer/goroot/src/net/timeout_test.go#L843-L906) <details> <summary>Click here to show the 64 line(s) of Go which triggered the analyzer.</summary> ```go for _, timeout := range []time.Duration{ 1 * time.Nanosecond, 2 * time.Nanosecond, 5 * time.Nanosecond, 50 * time.Nanosecond, 100 * time.Nanosecond, 200 * time.Nanosecond, 500 * time.Nanosecond, 750 * time.Nanosecond, 1 * time.Microsecond, 5 * time.Microsecond, 25 * time.Microsecond, 250 * time.Microsecond, 500 * time.Microsecond, 1 * time.Millisecond, 5 * time.Millisecond, 100 * time.Millisecond, 250 * time.Millisecond, 500 * time.Millisecond, 1 * time.Second, } { numRuns := 3 if testing.Short() { numRuns = 1 if timeout > 500*time.Microsecond { continue } } for run := 0; run < numRuns; run++ { name := fmt.Sprintf("%v %d/%d", timeout, run, numRuns) t.Log(name) tooSlow := time.NewTimer(5 * time.Second) defer tooSlow.Stop() c, err := Dial(ls.Listener.Addr().Network(), ls.Listener.Addr().String()) if err != nil { t.Fatal(err) } ch := make(chan result, 1) go func() { t0 := time.Now() if err := c.SetDeadline(t0.Add(timeout)); err != nil { t.Error(err) } n, err := io.Copy(ioutil.Discard, c) dt := time.Since(t0) c.Close() ch <- result{n, err, dt} }() select { case res := <-ch: if nerr, ok := res.err.(Error); ok && nerr.Timeout() { t.Logf("%v: good timeout after %v; %d bytes", name, res.d, res.n) } else { t.Fatalf("%v: Copy = %d, %v; want timeout", name, res.n, res.err) } case <-tooSlow.C: t.Fatalf("%v: client stuck in Dial+Copy", name) } } } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 61860eeeeff292ed4abb4fc46032ff41271d1b0b
test
goinfre developer goroot src net timeout test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message range loop variable timeout used in defer or goroutine at line click here to show the line s of go which triggered the analyzer go for timeout range time duration time nanosecond time nanosecond time nanosecond time nanosecond time nanosecond time nanosecond time nanosecond time nanosecond time microsecond time microsecond time microsecond time microsecond time microsecond time millisecond time millisecond time millisecond time millisecond time millisecond time second numruns if testing short numruns if timeout time microsecond continue for run run numruns run name fmt sprintf v d d timeout run numruns t log name tooslow time newtimer time second defer tooslow stop c err dial ls listener addr network ls listener addr string if err nil t fatal err ch make chan result go func time now if err c setdeadline add timeout err nil t error err n err io copy ioutil discard c dt time since c close ch result n err dt select case res ch if nerr ok res err error ok nerr timeout t logf v good timeout after v d bytes name res d res n else t fatalf v copy d v want timeout name res n res err case tooslow c t fatalf v client stuck in dial copy name leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
1
326,025
27,974,336,663
IssuesEvent
2023-03-25 11:52:23
zinclabs/zincobserve
https://api.github.com/repos/zinclabs/zincobserve
closed
unable to use a numeric field as x_axis
testing
### Which ZincObserve functionalities are the source of the bug? dashboards ### Is this a regression? Yes ### Description Setting x_axis to a numeric field gives errors in console. e.g. use code in x_axis ### Please provide a link to a minimal reproduction of the bug _No response_ ### Please provide the exception or error you saw _No response_ ### Please provide the version you discovered this bug in (check about page for version information) ```true v0.3.0 ``` ### Anything else? _No response_
1.0
unable to use a numeric field as x_axis - ### Which ZincObserve functionalities are the source of the bug? dashboards ### Is this a regression? Yes ### Description Setting x_axis to a numeric field gives errors in console. e.g. use code in x_axis ### Please provide a link to a minimal reproduction of the bug _No response_ ### Please provide the exception or error you saw _No response_ ### Please provide the version you discovered this bug in (check about page for version information) ```true v0.3.0 ``` ### Anything else? _No response_
test
unable to use a numeric field as x axis which zincobserve functionalities are the source of the bug dashboards is this a regression yes description setting x axis to a numeric field gives errors in console e g use code in x axis please provide a link to a minimal reproduction of the bug no response please provide the exception or error you saw no response please provide the version you discovered this bug in check about page for version information true anything else no response
1
15,773
3,974,842,329
IssuesEvent
2016-05-05 00:17:30
letsencrypt/letsencrypt
https://api.github.com/repos/letsencrypt/letsencrypt
closed
Include developer guide link in README.rst
documentation rename
Since we are not including a developer link on the certbot website, we should put a link to the developer guide in README.
1.0
Include developer guide link in README.rst - Since we are not including a developer link on the certbot website, we should put a link to the developer guide in README.
non_test
include developer guide link in readme rst since we are not including a developer link on the certbot website we should put a link to the developer guide in readme
0
207,505
15,820,389,021
IssuesEvent
2021-04-05 18:53:57
certbot/certbot
https://api.github.com/repos/certbot/certbot
closed
Fix lint and mypy with Python < 3.8
area: mypy area: testing area: tooling priority: significant
In https://github.com/certbot/certbot/pull/8748, we made a change that causes our lint and mypy tests to need to be run on Python 3.8+ to pass. See https://github.com/certbot/certbot/pull/8748#issuecomment-808790093 for the discussion of the problem here. I don't think we should do this. Certbot supports Python 3.6+ and I think it could cause a particularly bad experience for new devs that don't happen to know they need Python 3.8+. This change also broke our development Dockerfile as can be seen at https://dev.azure.com/certbot/certbot/_build/results?buildId=3742&view=logs&j=bea2d267-f41e-5b33-7b51-a88065a8cbb0&t=0dc90756-6888-5ee6-5a6a-5855e6b9ae76&l=1873. Instead, I think we should change our approach here so the tests work on all versions of Python we support. I'm open to other ideas, but the two ideas I had for this are: 1. Just declare a runtime dependency on `typing-extensions`. 2. Add `typing-extensions` as a dev/test dependency and try to import it, but use similar fallback code to what we current have if it's not available. What do you think @adferrand? Are you interested in working on this?
1.0
Fix lint and mypy with Python < 3.8 - In https://github.com/certbot/certbot/pull/8748, we made a change that causes our lint and mypy tests to need to be run on Python 3.8+ to pass. See https://github.com/certbot/certbot/pull/8748#issuecomment-808790093 for the discussion of the problem here. I don't think we should do this. Certbot supports Python 3.6+ and I think it could cause a particularly bad experience for new devs that don't happen to know they need Python 3.8+. This change also broke our development Dockerfile as can be seen at https://dev.azure.com/certbot/certbot/_build/results?buildId=3742&view=logs&j=bea2d267-f41e-5b33-7b51-a88065a8cbb0&t=0dc90756-6888-5ee6-5a6a-5855e6b9ae76&l=1873. Instead, I think we should change our approach here so the tests work on all versions of Python we support. I'm open to other ideas, but the two ideas I had for this are: 1. Just declare a runtime dependency on `typing-extensions`. 2. Add `typing-extensions` as a dev/test dependency and try to import it, but use similar fallback code to what we current have if it's not available. What do you think @adferrand? Are you interested in working on this?
test
fix lint and mypy with python in we made a change that causes our lint and mypy tests to need to be run on python to pass see for the discussion of the problem here i don t think we should do this certbot supports python and i think it could cause a particularly bad experience for new devs that don t happen to know they need python this change also broke our development dockerfile as can be seen at instead i think we should change our approach here so the tests work on all versions of python we support i m open to other ideas but the two ideas i had for this are just declare a runtime dependency on typing extensions add typing extensions as a dev test dependency and try to import it but use similar fallback code to what we current have if it s not available what do you think adferrand are you interested in working on this
1
13,920
10,545,737,249
IssuesEvent
2019-10-02 19:49:59
sciencehistory/scihist_digicoll
https://api.github.com/repos/sciencehistory/scihist_digicoll
closed
Add `cd` to rails app to .bash_profile for dev accounts
developer tools infrastructure
While we would like to never manually ssh to an EC2 machine, we still do it sometimes. Almost always the first thing me and @eddierubeiz (and maybe dan too?) have to do is `cd /opt/scihist_digicoll/current`. I've added that to my `~/.bash_profile` on staging jobs -- but of course, that will only last until the next time the machine is (re)built. Would it make sense to have ansible create a .bash_profile in the individual accounts it creates that just has `cd /opt/scihist_digicoll/current` in it? Not a high priority item.
1.0
Add `cd` to rails app to .bash_profile for dev accounts - While we would like to never manually ssh to an EC2 machine, we still do it sometimes. Almost always the first thing me and @eddierubeiz (and maybe dan too?) have to do is `cd /opt/scihist_digicoll/current`. I've added that to my `~/.bash_profile` on staging jobs -- but of course, that will only last until the next time the machine is (re)built. Would it make sense to have ansible create a .bash_profile in the individual accounts it creates that just has `cd /opt/scihist_digicoll/current` in it? Not a high priority item.
non_test
add cd to rails app to bash profile for dev accounts while we would like to never manually ssh to an machine we still do it sometimes almost always the first thing me and eddierubeiz and maybe dan too have to do is cd opt scihist digicoll current i ve added that to my bash profile on staging jobs but of course that will only last until the next time the machine is re built would it make sense to have ansible create a bash profile in the individual accounts it creates that just has cd opt scihist digicoll current in it not a high priority item
0
136,516
5,284,165,038
IssuesEvent
2017-02-07 23:23:38
rails-girls-summer-of-code/rgsoc-teams
https://api.github.com/repos/rails-girls-summer-of-code/rgsoc-teams
opened
Can't sign in
bug high-priority
Today we got a message from a student on Slack: "Hi. My team member is trying to log in with her GitHub account for the first time on RGSoC website. However, she is not receiving any activation email or getting signed in." I faced the same issue testing the app locally: ![cannot-sign-in](https://cloud.githubusercontent.com/assets/1745735/22716078/53bcf6b6-ed94-11e6-8ee7-0b5b93c69f81.gif) Can't reproduce it on Production though.
1.0
Can't sign in - Today we got a message from a student on Slack: "Hi. My team member is trying to log in with her GitHub account for the first time on RGSoC website. However, she is not receiving any activation email or getting signed in." I faced the same issue testing the app locally: ![cannot-sign-in](https://cloud.githubusercontent.com/assets/1745735/22716078/53bcf6b6-ed94-11e6-8ee7-0b5b93c69f81.gif) Can't reproduce it on Production though.
non_test
can t sign in today we got a message from a student on slack hi my team member is trying to log in with her github account for the first time on rgsoc website however she is not receiving any activation email or getting signed in i faced the same issue testing the app locally can t reproduce it on production though
0
129,583
10,578,717,420
IssuesEvent
2019-10-07 23:43:00
rancher/rancher
https://api.github.com/repos/rancher/rancher
closed
RHEL 7.7 hosts go to disconnected/reconnecting state
[zube]: To Test kind/bug-qa team/ca
Version: v1.6.29-rc2 Steps: 1. Create a Cattle setup with RHEL 7.7 nodes Native docker and SELINUX enabled 2. Run automation tests The hosts go to disconnected and reconnecting states frequently Rancher server logs below: ``` 2019-09-26 21:17:55,608 ERROR [:] [] [] [] [ecutorService-1] [.e.s.i.ProcessInstanceDispatcherImpl] Unknown exception running process [instance.start:209383] on [9438] java.lang.IllegalStateException: More than one volume exists for volume URI [file:///var/lib/docker/volumes/test_insp_vol/_data] or name/externalId [test_insp_vol] and storage pool [4]. at io.cattle.platform.docker.process.dao.impl.DockerComputeDaoImpl.getDockerVolumeInPool(DockerComputeDaoImpl.java:57) at sun.reflect.GeneratedMethodAccessor979.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.springframework.aop.support.AopUtils.invokeJoinpointUsingReflection(AopUtils.java:333) at org.springframework.aop.framework.ReflectiveMethodInvocation.invokeJoinpoint(ReflectiveMethodInvocation.java:190) at org.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:157) at org.springframework.transaction.interceptor.TransactionInterceptor$1.proceedWithInvocation(TransactionInterceptor.java:99) at org.springframework.transaction.interceptor.TransactionAspectSupport.invokeWithinTransaction(TransactionAspectSupport.java:280) at org.springframework.transaction.interceptor.TransactionInterceptor.invoke(TransactionInterceptor.java:96) at org.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:179) at org.springframework.aop.interceptor.ExposeInvocationInterceptor.invoke(ExposeInvocationInterceptor.java:92) at org.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:179) at org.springframework.aop.framework.JdkDynamicAopProxy.invoke(JdkDynamicAopProxy.java:213) at com.sun.proxy.$Proxy69.getDockerVolumeInPool(Unknown Source) at io.cattle.platform.docker.process.instancehostmap.DockerPostInstanceHostMapActivate.createVolumeInStoragePool(DockerPostInstanceHostMapActivate.java:214) at io.cattle.platform.docker.process.instancehostmap.DockerPostInstanceHostMapActivate.processVolumes(DockerPostInstanceHostMapActivate.java:197) at io.cattle.platform.docker.process.instancehostmap.DockerPostInstanceHostMapActivate.handle(DockerPostInstanceHostMapActivate.java:117) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl.runHandler(DefaultProcessInstanceImpl.java:470) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl$4.execute(DefaultProcessInstanceImpl.java:421) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl$4.execute(DefaultProcessInstanceImpl.java:415) at io.cattle.platform.engine.idempotent.Idempotent.execute(Idempotent.java:42) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl.runHandlers(DefaultProcessInstanceImpl.java:415) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl.runLogic(DefaultProcessInstanceImpl.java:520) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl.runWithProcessLock(DefaultProcessInstanceImpl.java:345) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl$2.doWithLockNoResult(DefaultProcessInstanceImpl.java:245) at io.cattle.platform.lock.LockCallbackNoReturn.doWithLock(LockCallbackNoReturn.java:7) at io.cattle.platform.lock.LockCallbackNoReturn.doWithLock(LockCallbackNoReturn.java:3) at io.cattle.platform.lock.impl.AbstractLockManagerImpl$3.doWithLock(AbstractLockManagerImpl.java:40) at io.cattle.platform.lock.impl.LockManagerImpl.doLock(LockManagerImpl.java:33) at io.cattle.platform.lock.impl.AbstractLockManagerImpl.lock(AbstractLockManagerImpl.java:13) at io.cattle.platform.lock.impl.AbstractLockManagerImpl.lock(AbstractLockManagerImpl.java:37) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl.acquireLockAndRun(DefaultProcessInstanceImpl.java:242) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl.runDelegateLoop(DefaultProcessInstanceImpl.java:184) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl.executeWithProcessInstanceLock(DefaultProcessInstanceImpl.java:157) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl$1.doWithLock(DefaultProcessInstanceImpl.java:116) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl$1.doWithLock(DefaultProcessInstanceImpl.java:113) at io.cattle.platform.lock.impl.AbstractLockManagerImpl$3.doWithLock(AbstractLockManagerImpl.java:40) at io.cattle.platform.lock.impl.LockManagerImpl.doLock(LockManagerImpl.java:33) at io.cattle.platform.lock.impl.AbstractLockManagerImpl.lock(AbstractLockManagerImpl.java:13) at io.cattle.platform.lock.impl.AbstractLockManagerImpl.lock(AbstractLockManagerImpl.java:37) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl.execute(DefaultProcessInstanceImpl.java:113) at io.cattle.platform.object.process.impl.DefaultObjectProcessManager.executeStandardProcess(DefaultObjectProcessManager.java:32) at io.cattle.platform.process.common.handler.AbstractObjectProcessLogic.activate(AbstractObjectProcessLogic.java:31) at io.cattle.platform.process.instance.InstanceStart.compute(InstanceStart.java:428) at io.cattle.platform.process.instance.InstanceStart.handle(InstanceStart.java:165) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl.runHandler(DefaultProcessInstanceImpl.java:470) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl$4.execute(DefaultProcessInstanceImpl.java:421) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl$4.execute(DefaultProcessInstanceImpl.java:415) at io.cattle.platform.engine.idempotent.Idempotent.execute(Idempotent.java:42) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl.runHandlers(DefaultProcessInstanceImpl.java:415) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl.runLogic(DefaultProcessInstanceImpl.java:517) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl.runWithProcessLock(DefaultProcessInstanceImpl.java:345) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl$2.doWithLockNoResult(DefaultProcessInstanceImpl.java:245) at io.cattle.platform.lock.LockCallbackNoReturn.doWithLock(LockCallbackNoReturn.java:7) at io.cattle.platform.lock.LockCallbackNoReturn.doWithLock(LockCallbackNoReturn.java:3) at io.cattle.platform.lock.impl.AbstractLockManagerImpl$3.doWithLock(AbstractLockManagerImpl.java:40) at io.cattle.platform.lock.impl.LockManagerImpl.doLock(LockManagerImpl.java:33) at io.cattle.platform.lock.impl.AbstractLockManagerImpl.lock(AbstractLockManagerImpl.java:13) at io.cattle.platform.lock.impl.AbstractLockManagerImpl.lock(AbstractLockManagerImpl.java:37) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl.acquireLockAndRun(DefaultProcessInstanceImpl.java:242) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl.runDelegateLoop(DefaultProcessInstanceImpl.java:184) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl.executeWithProcessInstanceLock(DefaultProcessInstanceImpl.java:157) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl$1.doWithLock(DefaultProcessInstanceImpl.java:116) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl$1.doWithLock(DefaultProcessInstanceImpl.java:113) at io.cattle.platform.lock.impl.AbstractLockManagerImpl$3.doWithLock(AbstractLockManagerImpl.java:40) at io.cattle.platform.lock.impl.LockManagerImpl.doLock(LockManagerImpl.java:33) at io.cattle.platform.lock.impl.AbstractLockManagerImpl.lock(AbstractLockManagerImpl.java:13) at io.cattle.platform.lock.impl.AbstractLockManagerImpl.lock(AbstractLockManagerImpl.java:37) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl.execute(DefaultProcessInstanceImpl.java:113) at io.cattle.platform.engine.server.impl.ProcessInstanceDispatcherImpl.processExecuteWithLock(ProcessInstanceDispatcherImpl.java:136) at io.cattle.platform.engine.server.impl.ProcessInstanceDispatcherImpl$1.doWithLock(ProcessInstanceDispatcherImpl.java:123) at io.cattle.platform.engine.server.impl.ProcessInstanceDispatcherImpl$1.doWithLock(ProcessInstanceDispatcherImpl.java:120) at io.cattle.platform.lock.impl.AbstractLockManagerImpl$4.doWithLock(AbstractLockManagerImpl.java:50) at io.cattle.platform.lock.impl.LockManagerImpl.doLock(LockManagerImpl.java:33) at io.cattle.platform.lock.impl.AbstractLockManagerImpl.tryLock(AbstractLockManagerImpl.java:25) at io.cattle.platform.lock.impl.AbstractLockManagerImpl.tryLock(AbstractLockManagerImpl.java:47) at io.cattle.platform.engine.server.impl.ProcessInstanceDispatcherImpl.execute(ProcessInstanceDispatcherImpl.java:120) at io.cattle.platform.engine.server.ProcessInstanceReference.doRun(ProcessInstanceReference.java:59) at org.apache.cloudstack.managed.context.NoExceptionRunnable.runInContext(NoExceptionRunnable.java:15) at org.apache.cloudstack.managed.context.ManagedContextRunnable$1.run(ManagedContextRunnable.java:49) at org.apache.cloudstack.managed.context.impl.DefaultManagedContext$1.call(DefaultManagedContext.java:55) at org.apache.cloudstack.managed.context.impl.DefaultManagedContext.callWithContext(DefaultManagedContext.java:108) at org.apache.cloudstack.managed.context.impl.DefaultManagedContext.runWithContext(DefaultManagedContext.java:52) at org.apache.cloudstack.managed.context.ManagedContextRunnable.run(ManagedContextRunnable.java:46) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745) time="2019-09-26T21:18:00Z" level=info msg="Stack Create Event Received" eventId=7cf2d829-9ac5-4a2d-accd-832f0dbad89d resourceId=1st1328 time="2019-09-26T21:18:00Z" level=info msg="Stack Create Event Done" eventId=7cf2d829-9ac5-4a2d-accd-832f0dbad89d resourceId=1st1328 time="2019-09-26T21:19:12Z" level=info msg="Shutting down backend 55a0180f-0643-48f5-5fc0-daccd241a882. Connection closed because: read tcp 172.17.0.2:8080->18.224.61.180:36880: read: connection reset by peer." time="2019-09-26T21:19:12Z" level=info msg="Removed backend. Key: 55a0180f-0643-48f5-5fc0-daccd241a882. Session ID 900ada27-b56b-4b07-ae83-9cd0b96a0266 ." 2019-09-26 21:19:12,430 ERROR [:] [] [] [] [ecutorService-1] [o.a.c.m.context.NoExceptionRunnable ] Container should have been running but is in stopped state. Check logs for more information. 2019-09-26 21:19:27,954 ERROR [5d98a1e9-05a4-470e-9ee3-c745c3b0af51:226365] [instance:10056->instanceHostMap:7251] [instance.start->(InstanceStart)->instancehostmap.activate] [] [cutorService-42] [c.p.e.p.i.DefaultProcessInstanceImpl] Agent error for [compute.instance.activate.reply;agent=27]: Error response from daemon: OCI runtime create failed: container_linux.go:345: starting container process caused "process_linux.go:430: container init caused \"process_linux.go:396: setting cgroup config for procHooks process caused \\\"kernel memory accounting disabled in this runc build\\\"\"": unknown 2019-09-26 21:19:27,954 ERROR [5d98a1e9-05a4-470e-9ee3-c745c3b0af51:226365] [instance:10056] [instance.start->(InstanceStart)] [] [cutorService-42] [i.c.p.process.instance.InstanceStart] Failed [1/5] to Starting for instance [10056] 2019-09-26 21:19:29,346 ERROR [5d98a1e9-05a4-470e-9ee3-c745c3b0af51:226365] [instance:10056->instanceHostMap:7251] [instance.start->(InstanceStart)->instancehostmap.activate] [] [cutorService-42] [c.p.e.p.i.DefaultProcessInstanceImpl] Agent error for [compute.instance.activate.reply;agent=27]: Error response from daemon: OCI runtime create failed: container_linux.go:345: starting container process caused "process_linux.go:430: container init caused \"process_linux.go:396: setting cgroup config for procHooks process caused \\\"kernel memory accounting disabled in this runc build\\\"\"": unknown 2019-09-26 21:19:29,346 ERROR [5d98a1e9-05a4-470e-9ee3-c745c3b0af51:226365] [instance:10056] [instance.start->(InstanceStart)] [] [cutorService-42] [i.c.p.process.instance.InstanceStart] Failed [2/5] to Starting for instance [10056] 2019-09-26 21:19:30,657 ERROR [5d98a1e9-05a4-470e-9ee3-c745c3b0af51:226365] [instance:10056->instanceHostMap:7251] [in ```
1.0
RHEL 7.7 hosts go to disconnected/reconnecting state - Version: v1.6.29-rc2 Steps: 1. Create a Cattle setup with RHEL 7.7 nodes Native docker and SELINUX enabled 2. Run automation tests The hosts go to disconnected and reconnecting states frequently Rancher server logs below: ``` 2019-09-26 21:17:55,608 ERROR [:] [] [] [] [ecutorService-1] [.e.s.i.ProcessInstanceDispatcherImpl] Unknown exception running process [instance.start:209383] on [9438] java.lang.IllegalStateException: More than one volume exists for volume URI [file:///var/lib/docker/volumes/test_insp_vol/_data] or name/externalId [test_insp_vol] and storage pool [4]. at io.cattle.platform.docker.process.dao.impl.DockerComputeDaoImpl.getDockerVolumeInPool(DockerComputeDaoImpl.java:57) at sun.reflect.GeneratedMethodAccessor979.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.springframework.aop.support.AopUtils.invokeJoinpointUsingReflection(AopUtils.java:333) at org.springframework.aop.framework.ReflectiveMethodInvocation.invokeJoinpoint(ReflectiveMethodInvocation.java:190) at org.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:157) at org.springframework.transaction.interceptor.TransactionInterceptor$1.proceedWithInvocation(TransactionInterceptor.java:99) at org.springframework.transaction.interceptor.TransactionAspectSupport.invokeWithinTransaction(TransactionAspectSupport.java:280) at org.springframework.transaction.interceptor.TransactionInterceptor.invoke(TransactionInterceptor.java:96) at org.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:179) at org.springframework.aop.interceptor.ExposeInvocationInterceptor.invoke(ExposeInvocationInterceptor.java:92) at org.springframework.aop.framework.ReflectiveMethodInvocation.proceed(ReflectiveMethodInvocation.java:179) at org.springframework.aop.framework.JdkDynamicAopProxy.invoke(JdkDynamicAopProxy.java:213) at com.sun.proxy.$Proxy69.getDockerVolumeInPool(Unknown Source) at io.cattle.platform.docker.process.instancehostmap.DockerPostInstanceHostMapActivate.createVolumeInStoragePool(DockerPostInstanceHostMapActivate.java:214) at io.cattle.platform.docker.process.instancehostmap.DockerPostInstanceHostMapActivate.processVolumes(DockerPostInstanceHostMapActivate.java:197) at io.cattle.platform.docker.process.instancehostmap.DockerPostInstanceHostMapActivate.handle(DockerPostInstanceHostMapActivate.java:117) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl.runHandler(DefaultProcessInstanceImpl.java:470) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl$4.execute(DefaultProcessInstanceImpl.java:421) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl$4.execute(DefaultProcessInstanceImpl.java:415) at io.cattle.platform.engine.idempotent.Idempotent.execute(Idempotent.java:42) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl.runHandlers(DefaultProcessInstanceImpl.java:415) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl.runLogic(DefaultProcessInstanceImpl.java:520) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl.runWithProcessLock(DefaultProcessInstanceImpl.java:345) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl$2.doWithLockNoResult(DefaultProcessInstanceImpl.java:245) at io.cattle.platform.lock.LockCallbackNoReturn.doWithLock(LockCallbackNoReturn.java:7) at io.cattle.platform.lock.LockCallbackNoReturn.doWithLock(LockCallbackNoReturn.java:3) at io.cattle.platform.lock.impl.AbstractLockManagerImpl$3.doWithLock(AbstractLockManagerImpl.java:40) at io.cattle.platform.lock.impl.LockManagerImpl.doLock(LockManagerImpl.java:33) at io.cattle.platform.lock.impl.AbstractLockManagerImpl.lock(AbstractLockManagerImpl.java:13) at io.cattle.platform.lock.impl.AbstractLockManagerImpl.lock(AbstractLockManagerImpl.java:37) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl.acquireLockAndRun(DefaultProcessInstanceImpl.java:242) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl.runDelegateLoop(DefaultProcessInstanceImpl.java:184) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl.executeWithProcessInstanceLock(DefaultProcessInstanceImpl.java:157) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl$1.doWithLock(DefaultProcessInstanceImpl.java:116) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl$1.doWithLock(DefaultProcessInstanceImpl.java:113) at io.cattle.platform.lock.impl.AbstractLockManagerImpl$3.doWithLock(AbstractLockManagerImpl.java:40) at io.cattle.platform.lock.impl.LockManagerImpl.doLock(LockManagerImpl.java:33) at io.cattle.platform.lock.impl.AbstractLockManagerImpl.lock(AbstractLockManagerImpl.java:13) at io.cattle.platform.lock.impl.AbstractLockManagerImpl.lock(AbstractLockManagerImpl.java:37) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl.execute(DefaultProcessInstanceImpl.java:113) at io.cattle.platform.object.process.impl.DefaultObjectProcessManager.executeStandardProcess(DefaultObjectProcessManager.java:32) at io.cattle.platform.process.common.handler.AbstractObjectProcessLogic.activate(AbstractObjectProcessLogic.java:31) at io.cattle.platform.process.instance.InstanceStart.compute(InstanceStart.java:428) at io.cattle.platform.process.instance.InstanceStart.handle(InstanceStart.java:165) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl.runHandler(DefaultProcessInstanceImpl.java:470) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl$4.execute(DefaultProcessInstanceImpl.java:421) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl$4.execute(DefaultProcessInstanceImpl.java:415) at io.cattle.platform.engine.idempotent.Idempotent.execute(Idempotent.java:42) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl.runHandlers(DefaultProcessInstanceImpl.java:415) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl.runLogic(DefaultProcessInstanceImpl.java:517) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl.runWithProcessLock(DefaultProcessInstanceImpl.java:345) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl$2.doWithLockNoResult(DefaultProcessInstanceImpl.java:245) at io.cattle.platform.lock.LockCallbackNoReturn.doWithLock(LockCallbackNoReturn.java:7) at io.cattle.platform.lock.LockCallbackNoReturn.doWithLock(LockCallbackNoReturn.java:3) at io.cattle.platform.lock.impl.AbstractLockManagerImpl$3.doWithLock(AbstractLockManagerImpl.java:40) at io.cattle.platform.lock.impl.LockManagerImpl.doLock(LockManagerImpl.java:33) at io.cattle.platform.lock.impl.AbstractLockManagerImpl.lock(AbstractLockManagerImpl.java:13) at io.cattle.platform.lock.impl.AbstractLockManagerImpl.lock(AbstractLockManagerImpl.java:37) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl.acquireLockAndRun(DefaultProcessInstanceImpl.java:242) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl.runDelegateLoop(DefaultProcessInstanceImpl.java:184) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl.executeWithProcessInstanceLock(DefaultProcessInstanceImpl.java:157) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl$1.doWithLock(DefaultProcessInstanceImpl.java:116) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl$1.doWithLock(DefaultProcessInstanceImpl.java:113) at io.cattle.platform.lock.impl.AbstractLockManagerImpl$3.doWithLock(AbstractLockManagerImpl.java:40) at io.cattle.platform.lock.impl.LockManagerImpl.doLock(LockManagerImpl.java:33) at io.cattle.platform.lock.impl.AbstractLockManagerImpl.lock(AbstractLockManagerImpl.java:13) at io.cattle.platform.lock.impl.AbstractLockManagerImpl.lock(AbstractLockManagerImpl.java:37) at io.cattle.platform.engine.process.impl.DefaultProcessInstanceImpl.execute(DefaultProcessInstanceImpl.java:113) at io.cattle.platform.engine.server.impl.ProcessInstanceDispatcherImpl.processExecuteWithLock(ProcessInstanceDispatcherImpl.java:136) at io.cattle.platform.engine.server.impl.ProcessInstanceDispatcherImpl$1.doWithLock(ProcessInstanceDispatcherImpl.java:123) at io.cattle.platform.engine.server.impl.ProcessInstanceDispatcherImpl$1.doWithLock(ProcessInstanceDispatcherImpl.java:120) at io.cattle.platform.lock.impl.AbstractLockManagerImpl$4.doWithLock(AbstractLockManagerImpl.java:50) at io.cattle.platform.lock.impl.LockManagerImpl.doLock(LockManagerImpl.java:33) at io.cattle.platform.lock.impl.AbstractLockManagerImpl.tryLock(AbstractLockManagerImpl.java:25) at io.cattle.platform.lock.impl.AbstractLockManagerImpl.tryLock(AbstractLockManagerImpl.java:47) at io.cattle.platform.engine.server.impl.ProcessInstanceDispatcherImpl.execute(ProcessInstanceDispatcherImpl.java:120) at io.cattle.platform.engine.server.ProcessInstanceReference.doRun(ProcessInstanceReference.java:59) at org.apache.cloudstack.managed.context.NoExceptionRunnable.runInContext(NoExceptionRunnable.java:15) at org.apache.cloudstack.managed.context.ManagedContextRunnable$1.run(ManagedContextRunnable.java:49) at org.apache.cloudstack.managed.context.impl.DefaultManagedContext$1.call(DefaultManagedContext.java:55) at org.apache.cloudstack.managed.context.impl.DefaultManagedContext.callWithContext(DefaultManagedContext.java:108) at org.apache.cloudstack.managed.context.impl.DefaultManagedContext.runWithContext(DefaultManagedContext.java:52) at org.apache.cloudstack.managed.context.ManagedContextRunnable.run(ManagedContextRunnable.java:46) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745) time="2019-09-26T21:18:00Z" level=info msg="Stack Create Event Received" eventId=7cf2d829-9ac5-4a2d-accd-832f0dbad89d resourceId=1st1328 time="2019-09-26T21:18:00Z" level=info msg="Stack Create Event Done" eventId=7cf2d829-9ac5-4a2d-accd-832f0dbad89d resourceId=1st1328 time="2019-09-26T21:19:12Z" level=info msg="Shutting down backend 55a0180f-0643-48f5-5fc0-daccd241a882. Connection closed because: read tcp 172.17.0.2:8080->18.224.61.180:36880: read: connection reset by peer." time="2019-09-26T21:19:12Z" level=info msg="Removed backend. Key: 55a0180f-0643-48f5-5fc0-daccd241a882. Session ID 900ada27-b56b-4b07-ae83-9cd0b96a0266 ." 2019-09-26 21:19:12,430 ERROR [:] [] [] [] [ecutorService-1] [o.a.c.m.context.NoExceptionRunnable ] Container should have been running but is in stopped state. Check logs for more information. 2019-09-26 21:19:27,954 ERROR [5d98a1e9-05a4-470e-9ee3-c745c3b0af51:226365] [instance:10056->instanceHostMap:7251] [instance.start->(InstanceStart)->instancehostmap.activate] [] [cutorService-42] [c.p.e.p.i.DefaultProcessInstanceImpl] Agent error for [compute.instance.activate.reply;agent=27]: Error response from daemon: OCI runtime create failed: container_linux.go:345: starting container process caused "process_linux.go:430: container init caused \"process_linux.go:396: setting cgroup config for procHooks process caused \\\"kernel memory accounting disabled in this runc build\\\"\"": unknown 2019-09-26 21:19:27,954 ERROR [5d98a1e9-05a4-470e-9ee3-c745c3b0af51:226365] [instance:10056] [instance.start->(InstanceStart)] [] [cutorService-42] [i.c.p.process.instance.InstanceStart] Failed [1/5] to Starting for instance [10056] 2019-09-26 21:19:29,346 ERROR [5d98a1e9-05a4-470e-9ee3-c745c3b0af51:226365] [instance:10056->instanceHostMap:7251] [instance.start->(InstanceStart)->instancehostmap.activate] [] [cutorService-42] [c.p.e.p.i.DefaultProcessInstanceImpl] Agent error for [compute.instance.activate.reply;agent=27]: Error response from daemon: OCI runtime create failed: container_linux.go:345: starting container process caused "process_linux.go:430: container init caused \"process_linux.go:396: setting cgroup config for procHooks process caused \\\"kernel memory accounting disabled in this runc build\\\"\"": unknown 2019-09-26 21:19:29,346 ERROR [5d98a1e9-05a4-470e-9ee3-c745c3b0af51:226365] [instance:10056] [instance.start->(InstanceStart)] [] [cutorService-42] [i.c.p.process.instance.InstanceStart] Failed [2/5] to Starting for instance [10056] 2019-09-26 21:19:30,657 ERROR [5d98a1e9-05a4-470e-9ee3-c745c3b0af51:226365] [instance:10056->instanceHostMap:7251] [in ```
test
rhel hosts go to disconnected reconnecting state version steps create a cattle setup with rhel nodes native docker and selinux enabled run automation tests the hosts go to disconnected and reconnecting states frequently rancher server logs below error unknown exception running process on java lang illegalstateexception more than one volume exists for volume uri or name externalid and storage pool at io cattle platform docker process dao impl dockercomputedaoimpl getdockervolumeinpool dockercomputedaoimpl java at sun reflect invoke unknown source at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org springframework aop support aoputils invokejoinpointusingreflection aoputils java at org springframework aop framework reflectivemethodinvocation invokejoinpoint reflectivemethodinvocation java at org springframework aop framework reflectivemethodinvocation proceed reflectivemethodinvocation java at org springframework transaction interceptor transactioninterceptor proceedwithinvocation transactioninterceptor java at org springframework transaction interceptor transactionaspectsupport invokewithintransaction transactionaspectsupport java at org springframework transaction interceptor transactioninterceptor invoke transactioninterceptor java at org springframework aop framework reflectivemethodinvocation proceed reflectivemethodinvocation java at org springframework aop interceptor exposeinvocationinterceptor invoke exposeinvocationinterceptor java at org springframework aop framework reflectivemethodinvocation proceed reflectivemethodinvocation java at org springframework aop framework jdkdynamicaopproxy invoke jdkdynamicaopproxy java at com sun proxy getdockervolumeinpool unknown source at io cattle platform docker process instancehostmap dockerpostinstancehostmapactivate createvolumeinstoragepool dockerpostinstancehostmapactivate java at io cattle platform docker process instancehostmap dockerpostinstancehostmapactivate processvolumes dockerpostinstancehostmapactivate java at io cattle platform docker process instancehostmap dockerpostinstancehostmapactivate handle dockerpostinstancehostmapactivate java at io cattle platform engine process impl defaultprocessinstanceimpl runhandler defaultprocessinstanceimpl java at io cattle platform engine process impl defaultprocessinstanceimpl execute defaultprocessinstanceimpl java at io cattle platform engine process impl defaultprocessinstanceimpl execute defaultprocessinstanceimpl java at io cattle platform engine idempotent idempotent execute idempotent java at io cattle platform engine process impl defaultprocessinstanceimpl runhandlers defaultprocessinstanceimpl java at io cattle platform engine process impl defaultprocessinstanceimpl runlogic defaultprocessinstanceimpl java at io cattle platform engine process impl defaultprocessinstanceimpl runwithprocesslock defaultprocessinstanceimpl java at io cattle platform engine process impl defaultprocessinstanceimpl dowithlocknoresult defaultprocessinstanceimpl java at io cattle platform lock lockcallbacknoreturn dowithlock lockcallbacknoreturn java at io cattle platform lock lockcallbacknoreturn dowithlock lockcallbacknoreturn java at io cattle platform lock impl abstractlockmanagerimpl dowithlock abstractlockmanagerimpl java at io cattle platform lock impl lockmanagerimpl dolock lockmanagerimpl java at io cattle platform lock impl abstractlockmanagerimpl lock abstractlockmanagerimpl java at io cattle platform lock impl abstractlockmanagerimpl lock abstractlockmanagerimpl java at io cattle platform engine process impl defaultprocessinstanceimpl acquirelockandrun defaultprocessinstanceimpl java at io cattle platform engine process impl defaultprocessinstanceimpl rundelegateloop defaultprocessinstanceimpl java at io cattle platform engine process impl defaultprocessinstanceimpl executewithprocessinstancelock defaultprocessinstanceimpl java at io cattle platform engine process impl defaultprocessinstanceimpl dowithlock defaultprocessinstanceimpl java at io cattle platform engine process impl defaultprocessinstanceimpl dowithlock defaultprocessinstanceimpl java at io cattle platform lock impl abstractlockmanagerimpl dowithlock abstractlockmanagerimpl java at io cattle platform lock impl lockmanagerimpl dolock lockmanagerimpl java at io cattle platform lock impl abstractlockmanagerimpl lock abstractlockmanagerimpl java at io cattle platform lock impl abstractlockmanagerimpl lock abstractlockmanagerimpl java at io cattle platform engine process impl defaultprocessinstanceimpl execute defaultprocessinstanceimpl java at io cattle platform object process impl defaultobjectprocessmanager executestandardprocess defaultobjectprocessmanager java at io cattle platform process common handler abstractobjectprocesslogic activate abstractobjectprocesslogic java at io cattle platform process instance instancestart compute instancestart java at io cattle platform process instance instancestart handle instancestart java at io cattle platform engine process impl defaultprocessinstanceimpl runhandler defaultprocessinstanceimpl java at io cattle platform engine process impl defaultprocessinstanceimpl execute defaultprocessinstanceimpl java at io cattle platform engine process impl defaultprocessinstanceimpl execute defaultprocessinstanceimpl java at io cattle platform engine idempotent idempotent execute idempotent java at io cattle platform engine process impl defaultprocessinstanceimpl runhandlers defaultprocessinstanceimpl java at io cattle platform engine process impl defaultprocessinstanceimpl runlogic defaultprocessinstanceimpl java at io cattle platform engine process impl defaultprocessinstanceimpl runwithprocesslock defaultprocessinstanceimpl java at io cattle platform engine process impl defaultprocessinstanceimpl dowithlocknoresult defaultprocessinstanceimpl java at io cattle platform lock lockcallbacknoreturn dowithlock lockcallbacknoreturn java at io cattle platform lock lockcallbacknoreturn dowithlock lockcallbacknoreturn java at io cattle platform lock impl abstractlockmanagerimpl dowithlock abstractlockmanagerimpl java at io cattle platform lock impl lockmanagerimpl dolock lockmanagerimpl java at io cattle platform lock impl abstractlockmanagerimpl lock abstractlockmanagerimpl java at io cattle platform lock impl abstractlockmanagerimpl lock abstractlockmanagerimpl java at io cattle platform engine process impl defaultprocessinstanceimpl acquirelockandrun defaultprocessinstanceimpl java at io cattle platform engine process impl defaultprocessinstanceimpl rundelegateloop defaultprocessinstanceimpl java at io cattle platform engine process impl defaultprocessinstanceimpl executewithprocessinstancelock defaultprocessinstanceimpl java at io cattle platform engine process impl defaultprocessinstanceimpl dowithlock defaultprocessinstanceimpl java at io cattle platform engine process impl defaultprocessinstanceimpl dowithlock defaultprocessinstanceimpl java at io cattle platform lock impl abstractlockmanagerimpl dowithlock abstractlockmanagerimpl java at io cattle platform lock impl lockmanagerimpl dolock lockmanagerimpl java at io cattle platform lock impl abstractlockmanagerimpl lock abstractlockmanagerimpl java at io cattle platform lock impl abstractlockmanagerimpl lock abstractlockmanagerimpl java at io cattle platform engine process impl defaultprocessinstanceimpl execute defaultprocessinstanceimpl java at io cattle platform engine server impl processinstancedispatcherimpl processexecutewithlock processinstancedispatcherimpl java at io cattle platform engine server impl processinstancedispatcherimpl dowithlock processinstancedispatcherimpl java at io cattle platform engine server impl processinstancedispatcherimpl dowithlock processinstancedispatcherimpl java at io cattle platform lock impl abstractlockmanagerimpl dowithlock abstractlockmanagerimpl java at io cattle platform lock impl lockmanagerimpl dolock lockmanagerimpl java at io cattle platform lock impl abstractlockmanagerimpl trylock abstractlockmanagerimpl java at io cattle platform lock impl abstractlockmanagerimpl trylock abstractlockmanagerimpl java at io cattle platform engine server impl processinstancedispatcherimpl execute processinstancedispatcherimpl java at io cattle platform engine server processinstancereference dorun processinstancereference java at org apache cloudstack managed context noexceptionrunnable runincontext noexceptionrunnable java at org apache cloudstack managed context managedcontextrunnable run managedcontextrunnable java at org apache cloudstack managed context impl defaultmanagedcontext call defaultmanagedcontext java at org apache cloudstack managed context impl defaultmanagedcontext callwithcontext defaultmanagedcontext java at org apache cloudstack managed context impl defaultmanagedcontext runwithcontext defaultmanagedcontext java at org apache cloudstack managed context managedcontextrunnable run managedcontextrunnable java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java time level info msg stack create event received eventid accd resourceid time level info msg stack create event done eventid accd resourceid time level info msg shutting down backend connection closed because read tcp read connection reset by peer time level info msg removed backend key session id error container should have been running but is in stopped state check logs for more information error agent error for error response from daemon oci runtime create failed container linux go starting container process caused process linux go container init caused process linux go setting cgroup config for prochooks process caused kernel memory accounting disabled in this runc build unknown error failed to starting for instance error agent error for error response from daemon oci runtime create failed container linux go starting container process caused process linux go container init caused process linux go setting cgroup config for prochooks process caused kernel memory accounting disabled in this runc build unknown error failed to starting for instance error in
1
172,725
6,515,786,489
IssuesEvent
2017-08-26 20:31:51
ianroberts131/visual-algorithms
https://api.github.com/repos/ianroberts131/visual-algorithms
opened
Add pause/play/reset buttons
Priority: Low Type: Feature
# Feature ## Why is this feature being added? When an algorithm is running, there is no way to stop an algorithm and inspect the current state. ## What should your feature do? Allow the user to stop an algorithm while it is running, and then also to restart it. Analyzing a single iteration in detail can really help understanding how an algorithm works.
1.0
Add pause/play/reset buttons - # Feature ## Why is this feature being added? When an algorithm is running, there is no way to stop an algorithm and inspect the current state. ## What should your feature do? Allow the user to stop an algorithm while it is running, and then also to restart it. Analyzing a single iteration in detail can really help understanding how an algorithm works.
non_test
add pause play reset buttons feature why is this feature being added when an algorithm is running there is no way to stop an algorithm and inspect the current state what should your feature do allow the user to stop an algorithm while it is running and then also to restart it analyzing a single iteration in detail can really help understanding how an algorithm works
0
227,950
18,126,953,397
IssuesEvent
2021-09-24 00:09:32
dotnet/source-build
https://api.github.com/repos/dotnet/source-build
closed
[ArPow] Add smoke tests
area-ci-testing
The smoke tests need to be added into ArPow so that they can be run after performing a source-build. The smoke tests should also be included in the tarball build CI.
1.0
[ArPow] Add smoke tests - The smoke tests need to be added into ArPow so that they can be run after performing a source-build. The smoke tests should also be included in the tarball build CI.
test
add smoke tests the smoke tests need to be added into arpow so that they can be run after performing a source build the smoke tests should also be included in the tarball build ci
1
214,733
16,608,668,686
IssuesEvent
2021-06-02 08:17:05
openenclave/openenclave
https://api.github.com/repos/openenclave/openenclave
closed
Child_process tests are failing for ctest on SGX1 LLC
SGX testing triaged
Hi, I'm running Build&Test on latest master of OE. OS version: Ubuntu 18.04 CPU used: Intel(R) Core(TM) i7-10710U CPU @ 1.10GHz (Supporting LLC only) ``` SGX: Software Guard Extensions supported = true SGX_LC: SGX launch config supported = false SGX capability (0x12/0): SGX1 supported = true SGX2 supported = false ``` **All tests except child_process are passing.** ``` 99% tests passed, 4 tests failed out of 945 Total Test time (real) = 370.62 sec The following tests did not run: 78 - tests/mbedtls_tls_e2e (Skipped) 79 - tests/mbedtls_tls_e2e-lvi-cfg (Skipped) 80 - tests/openssl_tls_e2e (Skipped) 81 - tests/openssl_tls_e2e-lvi-cfg (Skipped) 94 - tests/attestation_cert_api_mbedtls (Skipped) 95 - tests/attestation_cert_api_mbedtls-lvi-cfg (Skipped) 96 - tests/attestation_cert_api_openssl (Skipped) 97 - tests/attestation_cert_api_openssl-lvi-cfg (Skipped) 98 - tests/attestation_plugin_mbedtls (Skipped) 99 - tests/attestation_plugin_mbedtls-lvi-cfg (Skipped) 100 - tests/attestation_plugin_openssl (Skipped) 101 - tests/attestation_plugin_openssl-lvi-cfg (Skipped) 102 - tests/attestation_plugin_cert_mbedtls (Skipped) 103 - tests/attestation_plugin_cert_mbedtls-lvi-cfg (Skipped) 104 - tests/attestation_plugin_cert_openssl (Skipped) 105 - tests/attestation_plugin_cert_openssl-lvi-cfg (Skipped) 108 - tests/bigmalloc (Skipped) 109 - tests/bigmalloc-lvi-cfg (Skipped) 410 - tests/report_attestation_without_enclave (Skipped) 411 - tests/report_attestation_without_enclave-lvi-cfg (Skipped) The following tests FAILED: 935 - tests/child_process_ecall (Child aborted) 936 - tests/child_process_ecall-lvi-cfg (Child aborted) 937 - tests/child_process_destroy (Child aborted) 938 - tests/child_process_destroy-lvi-cfg (Child aborted) Errors while running CTest ``` Strace Tail(50) ``` pipe([9, 10]) = 0 epoll_ctl(8, EPOLL_CTL_ADD, 9, {EPOLLIN|EPOLLET, {u32=9, u64=9}}) = 0 socket(AF_UNIX, SOCK_STREAM, 0) = 11 connect(11, {sa_family=AF_UNIX, sun_path="/var/run/aesmd/aesm.socket"}, 110) = 0 epoll_ctl(8, EPOLL_CTL_ADD, 11, {EPOLLET, {u32=11, u64=11}}) = 0 fcntl(11, F_GETFL) = 0x2 (flags O_RDWR) fcntl(11, F_SETFL, O_RDWR|O_NONBLOCK) = 0 write(11, "\275\1\0\0", 4) = 4 epoll_ctl(8, EPOLL_CTL_MOD, 11, {EPOLLET, {u32=11, u64=11}}) = 0 epoll_ctl(8, EPOLL_CTL_MOD, 11, {EPOLLET, {u32=11, u64=11}}) = 0 write(11, "\32\272\3\n E\337\377\205\317+\204d\333@\v\343\362\204^\352j.\262\364\347hPG>\240\214"..., 445) = 445 epoll_ctl(8, EPOLL_CTL_MOD, 11, {EPOLLET, {u32=11, u64=11}}) = 0 epoll_ctl(8, EPOLL_CTL_MOD, 11, {EPOLLET, {u32=11, u64=11}}) = 0 epoll_ctl(8, EPOLL_CTL_MOD, 11, {EPOLLIN|EPOLLET, {u32=11, u64=11}}) = 0 read(11, 0x1800050, 4) = -1 EAGAIN (Resource temporarily unavailable) epoll_wait(8, [{EPOLLIN, {u32=11, u64=11}}], 12, 15000) = 1 read(11, "8\1\0\0", 4) = 4 epoll_ctl(8, EPOLL_CTL_MOD, 11, {EPOLLET, {u32=11, u64=11}}) = 0 epoll_ctl(8, EPOLL_CTL_MOD, 11, {EPOLLIN|EPOLLET, {u32=11, u64=11}}) = 0 read(11, "\32\265\2\10\0\22\260\2\1\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0"..., 312) = 312 epoll_ctl(8, EPOLL_CTL_MOD, 11, {EPOLLET, {u32=11, u64=11}}) = 0 close(8) = 0 close(9) = 0 close(10) = 0 close(11) = 0 ioctl(5, _IOC(_IOC_WRITE, 0xa4, 0x2, 0x18), 0x7ffea0f0d790) = 0 getcwd("/jenkinsdata/workspace/Pipelines/OpenEnclave-TestInfra/openenclave/build/tests/child_process", 4096) = 93 lstat("/jenkinsdata/workspace/Pipelines/OpenEnclave-TestInfra/openenclave/build/tests/child_process/enc", {st_mode=S_IFDIR|0755, st_size=4096, ...}) = 0 lstat("/jenkinsdata/workspace/Pipelines/OpenEnclave-TestInfra/openenclave/build/tests/child_process/enc/child_process_enc", {st_mode=S_IFREG|0755, st_size=2795208, ...}) = 0 munmap(0x7f1aafcad000, 2797568) = 0 munmap(0x7f1aaf9eb000, 2891776) = 0 futex(0x65c5ec, FUTEX_WAKE_PRIVATE, 2147483647) = 0 clone(child_stack=NULL, flags=CLONE_CHILD_CLEARTID|CLONE_CHILD_SETTID|SIGCHLD, child_tidptr=0x7f1ab178ea10) = 4366 getpid() = 4365 fstat(1, {st_mode=S_IFREG|0664, st_size=161380, ...}) = 0 munmap(0x7f1aae000000, 16777216) = 0 wait4(-1, Test failed: ../tests/child_process/host/host.cpp(136): main result == OE_OK [{WIFSIGNALED(s) && WTERMSIG(s) == SIGABRT && WCOREDUMP(s)}], 0, NULL) = 4366 --- SIGCHLD {si_signo=SIGCHLD, si_code=CLD_DUMPED, si_pid=4366, si_uid=1000, si_status=SIGABRT, si_utime=2, si_stime=1} --- wait4(-1, 0x7ffea0f0edc8, 0, NULL) = -1 ECHILD (No child processes) write(2, "Test failed: ../tests/child_proc"..., 79Test failed: ../tests/child_process/host/host.cpp(149): main child_status == 0 ) = 79 rt_sigprocmask(SIG_UNBLOCK, [ABRT], NULL, 8) = 0 rt_sigprocmask(SIG_BLOCK, ~[RTMIN RT_1], [], 8) = 0 getpid() = 4365 gettid() = 4365 tgkill(4365, 4365, SIGABRT) = 0 rt_sigprocmask(SIG_SETMASK, [], NULL, 8) = 0 --- SIGABRT {si_signo=SIGABRT, si_code=SI_TKILL, si_pid=4365, si_uid=1000} --- +++ killed by SIGABRT (core dumped) +++ ``` What is the purpose of these tests? Is it expected to pass on SGX1 non FLC platforms. If so, how to fix them? +@MWShan @sfblackl
1.0
Child_process tests are failing for ctest on SGX1 LLC - Hi, I'm running Build&Test on latest master of OE. OS version: Ubuntu 18.04 CPU used: Intel(R) Core(TM) i7-10710U CPU @ 1.10GHz (Supporting LLC only) ``` SGX: Software Guard Extensions supported = true SGX_LC: SGX launch config supported = false SGX capability (0x12/0): SGX1 supported = true SGX2 supported = false ``` **All tests except child_process are passing.** ``` 99% tests passed, 4 tests failed out of 945 Total Test time (real) = 370.62 sec The following tests did not run: 78 - tests/mbedtls_tls_e2e (Skipped) 79 - tests/mbedtls_tls_e2e-lvi-cfg (Skipped) 80 - tests/openssl_tls_e2e (Skipped) 81 - tests/openssl_tls_e2e-lvi-cfg (Skipped) 94 - tests/attestation_cert_api_mbedtls (Skipped) 95 - tests/attestation_cert_api_mbedtls-lvi-cfg (Skipped) 96 - tests/attestation_cert_api_openssl (Skipped) 97 - tests/attestation_cert_api_openssl-lvi-cfg (Skipped) 98 - tests/attestation_plugin_mbedtls (Skipped) 99 - tests/attestation_plugin_mbedtls-lvi-cfg (Skipped) 100 - tests/attestation_plugin_openssl (Skipped) 101 - tests/attestation_plugin_openssl-lvi-cfg (Skipped) 102 - tests/attestation_plugin_cert_mbedtls (Skipped) 103 - tests/attestation_plugin_cert_mbedtls-lvi-cfg (Skipped) 104 - tests/attestation_plugin_cert_openssl (Skipped) 105 - tests/attestation_plugin_cert_openssl-lvi-cfg (Skipped) 108 - tests/bigmalloc (Skipped) 109 - tests/bigmalloc-lvi-cfg (Skipped) 410 - tests/report_attestation_without_enclave (Skipped) 411 - tests/report_attestation_without_enclave-lvi-cfg (Skipped) The following tests FAILED: 935 - tests/child_process_ecall (Child aborted) 936 - tests/child_process_ecall-lvi-cfg (Child aborted) 937 - tests/child_process_destroy (Child aborted) 938 - tests/child_process_destroy-lvi-cfg (Child aborted) Errors while running CTest ``` Strace Tail(50) ``` pipe([9, 10]) = 0 epoll_ctl(8, EPOLL_CTL_ADD, 9, {EPOLLIN|EPOLLET, {u32=9, u64=9}}) = 0 socket(AF_UNIX, SOCK_STREAM, 0) = 11 connect(11, {sa_family=AF_UNIX, sun_path="/var/run/aesmd/aesm.socket"}, 110) = 0 epoll_ctl(8, EPOLL_CTL_ADD, 11, {EPOLLET, {u32=11, u64=11}}) = 0 fcntl(11, F_GETFL) = 0x2 (flags O_RDWR) fcntl(11, F_SETFL, O_RDWR|O_NONBLOCK) = 0 write(11, "\275\1\0\0", 4) = 4 epoll_ctl(8, EPOLL_CTL_MOD, 11, {EPOLLET, {u32=11, u64=11}}) = 0 epoll_ctl(8, EPOLL_CTL_MOD, 11, {EPOLLET, {u32=11, u64=11}}) = 0 write(11, "\32\272\3\n E\337\377\205\317+\204d\333@\v\343\362\204^\352j.\262\364\347hPG>\240\214"..., 445) = 445 epoll_ctl(8, EPOLL_CTL_MOD, 11, {EPOLLET, {u32=11, u64=11}}) = 0 epoll_ctl(8, EPOLL_CTL_MOD, 11, {EPOLLET, {u32=11, u64=11}}) = 0 epoll_ctl(8, EPOLL_CTL_MOD, 11, {EPOLLIN|EPOLLET, {u32=11, u64=11}}) = 0 read(11, 0x1800050, 4) = -1 EAGAIN (Resource temporarily unavailable) epoll_wait(8, [{EPOLLIN, {u32=11, u64=11}}], 12, 15000) = 1 read(11, "8\1\0\0", 4) = 4 epoll_ctl(8, EPOLL_CTL_MOD, 11, {EPOLLET, {u32=11, u64=11}}) = 0 epoll_ctl(8, EPOLL_CTL_MOD, 11, {EPOLLIN|EPOLLET, {u32=11, u64=11}}) = 0 read(11, "\32\265\2\10\0\22\260\2\1\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0"..., 312) = 312 epoll_ctl(8, EPOLL_CTL_MOD, 11, {EPOLLET, {u32=11, u64=11}}) = 0 close(8) = 0 close(9) = 0 close(10) = 0 close(11) = 0 ioctl(5, _IOC(_IOC_WRITE, 0xa4, 0x2, 0x18), 0x7ffea0f0d790) = 0 getcwd("/jenkinsdata/workspace/Pipelines/OpenEnclave-TestInfra/openenclave/build/tests/child_process", 4096) = 93 lstat("/jenkinsdata/workspace/Pipelines/OpenEnclave-TestInfra/openenclave/build/tests/child_process/enc", {st_mode=S_IFDIR|0755, st_size=4096, ...}) = 0 lstat("/jenkinsdata/workspace/Pipelines/OpenEnclave-TestInfra/openenclave/build/tests/child_process/enc/child_process_enc", {st_mode=S_IFREG|0755, st_size=2795208, ...}) = 0 munmap(0x7f1aafcad000, 2797568) = 0 munmap(0x7f1aaf9eb000, 2891776) = 0 futex(0x65c5ec, FUTEX_WAKE_PRIVATE, 2147483647) = 0 clone(child_stack=NULL, flags=CLONE_CHILD_CLEARTID|CLONE_CHILD_SETTID|SIGCHLD, child_tidptr=0x7f1ab178ea10) = 4366 getpid() = 4365 fstat(1, {st_mode=S_IFREG|0664, st_size=161380, ...}) = 0 munmap(0x7f1aae000000, 16777216) = 0 wait4(-1, Test failed: ../tests/child_process/host/host.cpp(136): main result == OE_OK [{WIFSIGNALED(s) && WTERMSIG(s) == SIGABRT && WCOREDUMP(s)}], 0, NULL) = 4366 --- SIGCHLD {si_signo=SIGCHLD, si_code=CLD_DUMPED, si_pid=4366, si_uid=1000, si_status=SIGABRT, si_utime=2, si_stime=1} --- wait4(-1, 0x7ffea0f0edc8, 0, NULL) = -1 ECHILD (No child processes) write(2, "Test failed: ../tests/child_proc"..., 79Test failed: ../tests/child_process/host/host.cpp(149): main child_status == 0 ) = 79 rt_sigprocmask(SIG_UNBLOCK, [ABRT], NULL, 8) = 0 rt_sigprocmask(SIG_BLOCK, ~[RTMIN RT_1], [], 8) = 0 getpid() = 4365 gettid() = 4365 tgkill(4365, 4365, SIGABRT) = 0 rt_sigprocmask(SIG_SETMASK, [], NULL, 8) = 0 --- SIGABRT {si_signo=SIGABRT, si_code=SI_TKILL, si_pid=4365, si_uid=1000} --- +++ killed by SIGABRT (core dumped) +++ ``` What is the purpose of these tests? Is it expected to pass on SGX1 non FLC platforms. If so, how to fix them? +@MWShan @sfblackl
test
child process tests are failing for ctest on llc hi i m running build test on latest master of oe os version ubuntu cpu used intel r core tm cpu supporting llc only sgx software guard extensions supported true sgx lc sgx launch config supported false sgx capability supported true supported false all tests except child process are passing tests passed tests failed out of total test time real sec the following tests did not run tests mbedtls tls skipped tests mbedtls tls lvi cfg skipped tests openssl tls skipped tests openssl tls lvi cfg skipped tests attestation cert api mbedtls skipped tests attestation cert api mbedtls lvi cfg skipped tests attestation cert api openssl skipped tests attestation cert api openssl lvi cfg skipped tests attestation plugin mbedtls skipped tests attestation plugin mbedtls lvi cfg skipped tests attestation plugin openssl skipped tests attestation plugin openssl lvi cfg skipped tests attestation plugin cert mbedtls skipped tests attestation plugin cert mbedtls lvi cfg skipped tests attestation plugin cert openssl skipped tests attestation plugin cert openssl lvi cfg skipped tests bigmalloc skipped tests bigmalloc lvi cfg skipped tests report attestation without enclave skipped tests report attestation without enclave lvi cfg skipped the following tests failed tests child process ecall child aborted tests child process ecall lvi cfg child aborted tests child process destroy child aborted tests child process destroy lvi cfg child aborted errors while running ctest strace tail pipe epoll ctl epoll ctl add epollin epollet socket af unix sock stream connect sa family af unix sun path var run aesmd aesm socket epoll ctl epoll ctl add epollet fcntl f getfl flags o rdwr fcntl f setfl o rdwr o nonblock write epoll ctl epoll ctl mod epollet epoll ctl epoll ctl mod epollet write n e v epoll ctl epoll ctl mod epollet epoll ctl epoll ctl mod epollet epoll ctl epoll ctl mod epollin epollet read eagain resource temporarily unavailable epoll wait read epoll ctl epoll ctl mod epollet epoll ctl epoll ctl mod epollin epollet read epoll ctl epoll ctl mod epollet close close close close ioctl ioc ioc write getcwd jenkinsdata workspace pipelines openenclave testinfra openenclave build tests child process lstat jenkinsdata workspace pipelines openenclave testinfra openenclave build tests child process enc st mode s ifdir st size lstat jenkinsdata workspace pipelines openenclave testinfra openenclave build tests child process enc child process enc st mode s ifreg st size munmap munmap futex futex wake private clone child stack null flags clone child cleartid clone child settid sigchld child tidptr getpid fstat st mode s ifreg st size munmap test failed tests child process host host cpp main result oe ok null sigchld si signo sigchld si code cld dumped si pid si uid si status sigabrt si utime si stime null echild no child processes write test failed tests child proc failed tests child process host host cpp main child status rt sigprocmask sig unblock null rt sigprocmask sig block getpid gettid tgkill sigabrt rt sigprocmask sig setmask null sigabrt si signo sigabrt si code si tkill si pid si uid killed by sigabrt core dumped what is the purpose of these tests is it expected to pass on non flc platforms if so how to fix them mwshan sfblackl
1
349,370
10,468,084,609
IssuesEvent
2019-09-22 11:06:17
k8smeetup/website-tasks
https://api.github.com/repos/k8smeetup/website-tasks
closed
/docs/reference/setup-tools/kubeadm/generated/kubeadm_init_phase_control-plane_apiserver.md
finished lang/zh priority/P0 sync/new version/1.14
Source File: [/docs/reference/setup-tools/kubeadm/generated/kubeadm_init_phase_control-plane_apiserver.md](https://github.com/kubernetes/website/blob/release-1.14/content/en/docs/reference/setup-tools/kubeadm/generated/kubeadm_init_phase_control-plane_apiserver.md)
1.0
/docs/reference/setup-tools/kubeadm/generated/kubeadm_init_phase_control-plane_apiserver.md - Source File: [/docs/reference/setup-tools/kubeadm/generated/kubeadm_init_phase_control-plane_apiserver.md](https://github.com/kubernetes/website/blob/release-1.14/content/en/docs/reference/setup-tools/kubeadm/generated/kubeadm_init_phase_control-plane_apiserver.md)
non_test
docs reference setup tools kubeadm generated kubeadm init phase control plane apiserver md source file
0
281,696
24,413,131,200
IssuesEvent
2022-10-05 13:57:15
apache/beam
https://api.github.com/repos/apache/beam
closed
beam_PostCommit_Java_DataflowV2 failures parent bug
stale java dataflow P1 bug test-failures flake
Multi-bug for all the Java v2 PostCommit failures:    SpannerChangeStreamIT.testReadSpannerChangeStream SpannerChangeStreamTransactionBoundariesIT.testTransactionBoundaries BigQueryIOStorageWriteIT.testBigQueryStorageWrite30MProto pubsublite.ReadWriteIT.testReadWrite Imported from Jira [BEAM-14276](https://issues.apache.org/jira/browse/BEAM-14276). Original Jira may contain additional context. Reported by: emilymye.
1.0
beam_PostCommit_Java_DataflowV2 failures parent bug - Multi-bug for all the Java v2 PostCommit failures:    SpannerChangeStreamIT.testReadSpannerChangeStream SpannerChangeStreamTransactionBoundariesIT.testTransactionBoundaries BigQueryIOStorageWriteIT.testBigQueryStorageWrite30MProto pubsublite.ReadWriteIT.testReadWrite Imported from Jira [BEAM-14276](https://issues.apache.org/jira/browse/BEAM-14276). Original Jira may contain additional context. Reported by: emilymye.
test
beam postcommit java failures parent bug multi bug for all the java postcommit failures     spannerchangestreamit testreadspannerchangestream spannerchangestreamtransactionboundariesit testtransactionboundaries bigqueryiostoragewriteit pubsublite readwriteit testreadwrite imported from jira original jira may contain additional context reported by emilymye
1
128,701
17,601,166,625
IssuesEvent
2021-08-17 12:03:24
department-of-veterans-affairs/va.gov-team
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
closed
Displaying the Call Center phone number (Vet Center)
design vsa vsa-facilities vet-center
## Issue Description As a Veteran, I need to know how to contact someone representing my local Vet Center. **Current protocol** - As of 12/21, the Vet Center call center only does warm handoffs to the crisis line. - The Call center gives the local Vet Center contact info to the Veteran and sends an email to the relevant local Vet Center about the Veteran's call (generally within a few hours) - If a Veteran calls the Vet Center directly after hours, he/she is patched through to the Call Center. - There is a new director for the Call Center so this process is subject to change. --- ## Tasks - [ ] Consider the best display for the Call Center considering no warm handoff is given ## Acceptance Criteria - [ ] _What will be created or happen as a result of this story?_ --- ## How to configure this issue - [ ] **Attached to a Milestone** (when will this be completed?) - [ ] **Attached to an Epic** (what body of work is this a part of?) - [ ] **Labeled with Team** (`product support`, `analytics-insights`, `operations`, `service-design`, `tools-be`, `tools-fe`) - [ ] **Labeled with Practice Area** (`backend`, `frontend`, `devops`, `design`, `research`, `product`, `ia`, `qa`, `analytics`, `contact center`, `research`, `accessibility`, `content`) - [ ] **Labeled with Type** (`bug`, `request`, `discovery`, `documentation`, etc.)
1.0
Displaying the Call Center phone number (Vet Center) - ## Issue Description As a Veteran, I need to know how to contact someone representing my local Vet Center. **Current protocol** - As of 12/21, the Vet Center call center only does warm handoffs to the crisis line. - The Call center gives the local Vet Center contact info to the Veteran and sends an email to the relevant local Vet Center about the Veteran's call (generally within a few hours) - If a Veteran calls the Vet Center directly after hours, he/she is patched through to the Call Center. - There is a new director for the Call Center so this process is subject to change. --- ## Tasks - [ ] Consider the best display for the Call Center considering no warm handoff is given ## Acceptance Criteria - [ ] _What will be created or happen as a result of this story?_ --- ## How to configure this issue - [ ] **Attached to a Milestone** (when will this be completed?) - [ ] **Attached to an Epic** (what body of work is this a part of?) - [ ] **Labeled with Team** (`product support`, `analytics-insights`, `operations`, `service-design`, `tools-be`, `tools-fe`) - [ ] **Labeled with Practice Area** (`backend`, `frontend`, `devops`, `design`, `research`, `product`, `ia`, `qa`, `analytics`, `contact center`, `research`, `accessibility`, `content`) - [ ] **Labeled with Type** (`bug`, `request`, `discovery`, `documentation`, etc.)
non_test
displaying the call center phone number vet center issue description as a veteran i need to know how to contact someone representing my local vet center current protocol as of the vet center call center only does warm handoffs to the crisis line the call center gives the local vet center contact info to the veteran and sends an email to the relevant local vet center about the veteran s call generally within a few hours if a veteran calls the vet center directly after hours he she is patched through to the call center there is a new director for the call center so this process is subject to change tasks consider the best display for the call center considering no warm handoff is given acceptance criteria what will be created or happen as a result of this story how to configure this issue attached to a milestone when will this be completed attached to an epic what body of work is this a part of labeled with team product support analytics insights operations service design tools be tools fe labeled with practice area backend frontend devops design research product ia qa analytics contact center research accessibility content labeled with type bug request discovery documentation etc
0
67,479
7,048,554,865
IssuesEvent
2018-01-02 18:10:45
rubberduck-vba/Rubberduck
https://api.github.com/repos/rubberduck-vba/Rubberduck
closed
Get Coverage results from AppVeyor and make use of them
enhancement meta rubberduck-test-project
See #3454 and #3448 After the migration to NUnit was successful we can now get AppVeyor to create coverage reports for us and do something with them
1.0
Get Coverage results from AppVeyor and make use of them - See #3454 and #3448 After the migration to NUnit was successful we can now get AppVeyor to create coverage reports for us and do something with them
test
get coverage results from appveyor and make use of them see and after the migration to nunit was successful we can now get appveyor to create coverage reports for us and do something with them
1
287,036
24,804,155,539
IssuesEvent
2022-10-25 01:55:29
DMTF/libspdm
https://api.github.com/repos/DMTF/libspdm
closed
LLVM warning in test_spdm_requester/get_certificate.c
enhancement test
When `test_spdm_requester/get_certificate.c` is compiled with LLVM it produces lots of warnings like ``` get_certificate.c:1196:13: warning: variable 'hash_size' is used uninitialized whenever 'if' condition is false [-Wsometimes-uninitialized] ```
1.0
LLVM warning in test_spdm_requester/get_certificate.c - When `test_spdm_requester/get_certificate.c` is compiled with LLVM it produces lots of warnings like ``` get_certificate.c:1196:13: warning: variable 'hash_size' is used uninitialized whenever 'if' condition is false [-Wsometimes-uninitialized] ```
test
llvm warning in test spdm requester get certificate c when test spdm requester get certificate c is compiled with llvm it produces lots of warnings like get certificate c warning variable hash size is used uninitialized whenever if condition is false
1
392,626
26,951,204,470
IssuesEvent
2023-02-08 11:50:13
Gleethos/Tribalism
https://api.github.com/repos/Gleethos/Tribalism
closed
Improve API of custom property based ORM/DAL
documentation enhancement
Before further changes I should improve the ORM some more with respect to its documentation and method names...
1.0
Improve API of custom property based ORM/DAL - Before further changes I should improve the ORM some more with respect to its documentation and method names...
non_test
improve api of custom property based orm dal before further changes i should improve the orm some more with respect to its documentation and method names
0
7,831
7,112,283,488
IssuesEvent
2018-01-17 16:34:26
dzhw/metadatamanagement
https://api.github.com/repos/dzhw/metadatamanagement
opened
linkage MDM - DLP
category:crosscutting prio:? scope:project infrastructure status:discussion
We need to define the linkage between the MDM and the DLP In the specification of the DLP it says > Ein Kunde wählt auf einer Website (Metadatensuchsystem) verschiedene Produkte aus, diese werden über eine Reminder-Funktion auf einen Merkzettel gesetzt. > Hat der Kunde seine Auswahl der Produkte abgeschlossen, kann er einen Button betätigen (z.B. "Produktauswahl abgeschlossen"), wodurch die Produkte über eine Schnittstelle an das CRM-System übergeben werden. > Anschließend muss der Kunde sich über das Kundenportal anmelden oder neu registrieren. >
1.0
linkage MDM - DLP - We need to define the linkage between the MDM and the DLP In the specification of the DLP it says > Ein Kunde wählt auf einer Website (Metadatensuchsystem) verschiedene Produkte aus, diese werden über eine Reminder-Funktion auf einen Merkzettel gesetzt. > Hat der Kunde seine Auswahl der Produkte abgeschlossen, kann er einen Button betätigen (z.B. "Produktauswahl abgeschlossen"), wodurch die Produkte über eine Schnittstelle an das CRM-System übergeben werden. > Anschließend muss der Kunde sich über das Kundenportal anmelden oder neu registrieren. >
non_test
linkage mdm dlp we need to define the linkage between the mdm and the dlp in the specification of the dlp it says ein kunde wählt auf einer website metadatensuchsystem verschiedene produkte aus diese werden über eine reminder funktion auf einen merkzettel gesetzt hat der kunde seine auswahl der produkte abgeschlossen kann er einen button betätigen z b produktauswahl abgeschlossen wodurch die produkte über eine schnittstelle an das crm system übergeben werden anschließend muss der kunde sich über das kundenportal anmelden oder neu registrieren
0
330,844
10,056,388,980
IssuesEvent
2019-07-22 09:02:35
laterpay/laterpay-wordpress-plugin
https://api.github.com/repos/laterpay/laterpay-wordpress-plugin
closed
Fix Technical Arrow on Advanced Tab
Low Priority
**Steps to Reproduce** - On a laptop, navigate to the LaterPay > Advanced Tab - Scroll to the bottom of the page **Observed Result** - The arrow for the Technical section is displayed at the bottom of the page **Desired Result** - It should be displayed to the right, like the other arrows ![image](https://user-images.githubusercontent.com/41650164/59714076-6250c480-91d6-11e9-91ee-32a2292c8894.png)
1.0
Fix Technical Arrow on Advanced Tab - **Steps to Reproduce** - On a laptop, navigate to the LaterPay > Advanced Tab - Scroll to the bottom of the page **Observed Result** - The arrow for the Technical section is displayed at the bottom of the page **Desired Result** - It should be displayed to the right, like the other arrows ![image](https://user-images.githubusercontent.com/41650164/59714076-6250c480-91d6-11e9-91ee-32a2292c8894.png)
non_test
fix technical arrow on advanced tab steps to reproduce on a laptop navigate to the laterpay advanced tab scroll to the bottom of the page observed result the arrow for the technical section is displayed at the bottom of the page desired result it should be displayed to the right like the other arrows
0
53,727
6,342,828,562
IssuesEvent
2017-07-27 16:15:11
tgstation/tgstation
https://api.github.com/repos/tgstation/tgstation
closed
Nuke explosion did not actually detonate or display round end text.
Bug Needs Reproducing/Testing
During nuke ops, the ops successfully blew up the station, but the explosion cutscene never happened, instead staying at the zoomed out view of the station screen and staying there until it vanished, no boom. Everyone still died and the station was still wrecked. Greentext was not displayed at all, though "Rebooting World in 60 seconds. Station destroyed by Nuclear Device." did show. Not sure what the cause was.
1.0
Nuke explosion did not actually detonate or display round end text. - During nuke ops, the ops successfully blew up the station, but the explosion cutscene never happened, instead staying at the zoomed out view of the station screen and staying there until it vanished, no boom. Everyone still died and the station was still wrecked. Greentext was not displayed at all, though "Rebooting World in 60 seconds. Station destroyed by Nuclear Device." did show. Not sure what the cause was.
test
nuke explosion did not actually detonate or display round end text during nuke ops the ops successfully blew up the station but the explosion cutscene never happened instead staying at the zoomed out view of the station screen and staying there until it vanished no boom everyone still died and the station was still wrecked greentext was not displayed at all though rebooting world in seconds station destroyed by nuclear device did show not sure what the cause was
1
312,490
26,868,471,805
IssuesEvent
2023-02-04 06:11:02
opentibiabr/canary
https://api.github.com/repos/opentibiabr/canary
closed
Monster pushItems() bug
Type: Bug Priority: High Status: Pending Test
### Priority High ### Area - [ ] Datapack - [X] Source - [ ] Map - [ ] Other ### What happened? The monster only moves objects in front of it, even if they are not moving in that direction. If it goes in a different direction than it is facing, it will not move the object, but it should. I think problem is here(monster.cpp): ``` if (result && (canPushItems() || canPushCreatures())) { const Position& pos = Spells::getCasterPosition(this, direction); Tile* posTile = g_game.map.getTile(pos); if (posTile) { if (canPushItems()) { Monster::pushItems(posTile); } ``` `Spells::getCasterPosition(this, direction);` returns the side the monster is facing, not which side it is going This is very disturbing in PVP https://user-images.githubusercontent.com/63909484/192546452-59a4e2ef-cbb9-4697-bb9f-6920de0c1e26.mp4 ### What OS are you seeing the problem on? Linux ### Code of Conduct - [X] I agree to follow this project's Code of Conduct
1.0
Monster pushItems() bug - ### Priority High ### Area - [ ] Datapack - [X] Source - [ ] Map - [ ] Other ### What happened? The monster only moves objects in front of it, even if they are not moving in that direction. If it goes in a different direction than it is facing, it will not move the object, but it should. I think problem is here(monster.cpp): ``` if (result && (canPushItems() || canPushCreatures())) { const Position& pos = Spells::getCasterPosition(this, direction); Tile* posTile = g_game.map.getTile(pos); if (posTile) { if (canPushItems()) { Monster::pushItems(posTile); } ``` `Spells::getCasterPosition(this, direction);` returns the side the monster is facing, not which side it is going This is very disturbing in PVP https://user-images.githubusercontent.com/63909484/192546452-59a4e2ef-cbb9-4697-bb9f-6920de0c1e26.mp4 ### What OS are you seeing the problem on? Linux ### Code of Conduct - [X] I agree to follow this project's Code of Conduct
test
monster pushitems bug priority high area datapack source map other what happened the monster only moves objects in front of it even if they are not moving in that direction if it goes in a different direction than it is facing it will not move the object but it should i think problem is here monster cpp if result canpushitems canpushcreatures const position pos spells getcasterposition this direction tile postile g game map gettile pos if postile if canpushitems monster pushitems postile spells getcasterposition this direction returns the side the monster is facing not which side it is going this is very disturbing in pvp what os are you seeing the problem on linux code of conduct i agree to follow this project s code of conduct
1
412,356
12,041,647,515
IssuesEvent
2020-04-14 09:11:49
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.ebay.co.uk - site is not usable
browser-firefox-tablet engine-gecko priority-important
<!-- @browser: Firefox Mobile (Tablet) 68.0 --> <!-- @ua_header: Mozilla/5.0 (Android 8.1.0; Tablet; rv:68.0) Gecko/68.0 Firefox/68.0 --> <!-- @reported_with: mobile-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/51485 --> **URL**: https://www.ebay.co.uk/sch/i.html?_from=R40&_sacat=0&_requireRaptorProxy=false%7Cfalse%7Cfalse%7Cfalse%7Cfalse&_nkw=hornby&_pgn=5&_skc=800&rt=nc&_requireRaptorProxy=false&_requireRaptorProxy=false&_requireRaptorProxy=false&_requireRaptorProxy=false&_requireRaptorProxy=false&_requireRaptorProxy=false&_requireRaptorProxy=false&_requireRaptorProxy=false&_requireRaptorProxy=false&_requireRaptorProxy=false&_requireRaptorProxy=false&_requireRaptorProxy=false **Browser / Version**: Firefox Mobile (Tablet) 68.0 **Operating System**: Android 8.1.0 **Tested Another Browser**: Yes Chrome **Problem type**: Site is not usable **Description**: Page not loading correctly **Steps to Reproduce**: Page keeps saying service unavailable <details><summary>View the screenshot</summary><img alt='Screenshot' src='https://webcompat.com/uploads/2020/4/407225e4-d2d2-4af1-b6eb-b4b2e18f18ae.jpeg'></details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200316183117</li><li>channel: beta</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2020/4/40f2a00a-fd21-4622-93f7-1d923bd74547) Submitted in the name of `@Mshone` _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.ebay.co.uk - site is not usable - <!-- @browser: Firefox Mobile (Tablet) 68.0 --> <!-- @ua_header: Mozilla/5.0 (Android 8.1.0; Tablet; rv:68.0) Gecko/68.0 Firefox/68.0 --> <!-- @reported_with: mobile-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/51485 --> **URL**: https://www.ebay.co.uk/sch/i.html?_from=R40&_sacat=0&_requireRaptorProxy=false%7Cfalse%7Cfalse%7Cfalse%7Cfalse&_nkw=hornby&_pgn=5&_skc=800&rt=nc&_requireRaptorProxy=false&_requireRaptorProxy=false&_requireRaptorProxy=false&_requireRaptorProxy=false&_requireRaptorProxy=false&_requireRaptorProxy=false&_requireRaptorProxy=false&_requireRaptorProxy=false&_requireRaptorProxy=false&_requireRaptorProxy=false&_requireRaptorProxy=false&_requireRaptorProxy=false **Browser / Version**: Firefox Mobile (Tablet) 68.0 **Operating System**: Android 8.1.0 **Tested Another Browser**: Yes Chrome **Problem type**: Site is not usable **Description**: Page not loading correctly **Steps to Reproduce**: Page keeps saying service unavailable <details><summary>View the screenshot</summary><img alt='Screenshot' src='https://webcompat.com/uploads/2020/4/407225e4-d2d2-4af1-b6eb-b4b2e18f18ae.jpeg'></details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200316183117</li><li>channel: beta</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2020/4/40f2a00a-fd21-4622-93f7-1d923bd74547) Submitted in the name of `@Mshone` _From [webcompat.com](https://webcompat.com/) with ❤️_
non_test
site is not usable url browser version firefox mobile tablet operating system android tested another browser yes chrome problem type site is not usable description page not loading correctly steps to reproduce page keeps saying service unavailable view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel beta hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false submitted in the name of mshone from with ❤️
0
29,023
5,487,423,139
IssuesEvent
2017-03-14 04:30:29
gridlab-d/gridlab-d
https://api.github.com/repos/gridlab-d/gridlab-d
reopened
Players that start with delta timesteps cause unhandled exceptions
bug defect
If a player starts directly with a delta timestep, it appears to generate a unhandled exception within GridLAB-D.. [exception_with_delta_player.zip](https://github.com/gridlab-d/gridlab-d/files/656587/exception_with_delta_player.zip)
1.0
Players that start with delta timesteps cause unhandled exceptions - If a player starts directly with a delta timestep, it appears to generate a unhandled exception within GridLAB-D.. [exception_with_delta_player.zip](https://github.com/gridlab-d/gridlab-d/files/656587/exception_with_delta_player.zip)
non_test
players that start with delta timesteps cause unhandled exceptions if a player starts directly with a delta timestep it appears to generate a unhandled exception within gridlab d
0
276,865
24,027,490,115
IssuesEvent
2022-09-15 12:43:52
brave/brave-browser
https://api.github.com/repos/brave/brave-browser
closed
Need more spacing between `Submit` and `Cancel` dialogs for Tor bridges
bug feature/tor feature/user-interface priority/P3 QA/Yes QA/Test-Plan-Specified OS/Desktop
<!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue. PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE. INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED--> ## Description <!--Provide a brief description of the issue--> Need more spacing between `Submit` and `Cancel` dialogs for Tor bridges ## Steps to Reproduce <!--Please add a series of steps to reproduce the issue--> 1. install `1.44.78` 2. launch Brave 3. open `brave://settings/privacy` 4. scroll to `Use Bridges` 5. click on the `Request a New Bridge...` button 6. look at the spacing between the `Submit` and `Cancel` dialogs ## Actual result: <!--Please add screenshots if needed--> example | example ---------|--------- <img width="1300" alt="Screenshot 2022-09-07 at 2 31 20 PM" src="https://user-images.githubusercontent.com/387249/188987869-b3f94484-e4fd-4d4d-8c61-6b5f32a91baf.png"> | <img width="191" alt="Screenshot 2022-09-07 at 2 31 37 PM" src="https://user-images.githubusercontent.com/387249/188987876-00d54829-805e-4c87-b175-758d2259ef4e.png"> ## Expected result: Compare spacing with `brave://settings/clearBrowserData` example | example ---------|--------- <img width="1300" alt="Screenshot 2022-09-07 at 2 34 33 PM" src="https://user-images.githubusercontent.com/387249/188988138-d749c30f-c61f-4b85-b03c-1c1173c39f48.png"> | <img width="246" alt="Screenshot 2022-09-07 at 2 35 11 PM" src="https://user-images.githubusercontent.com/387249/188988139-6e76cf39-1cb4-4059-b1b3-21210556a084.png"> ## Reproduces how often: <!--[Easily reproduced/Intermittent issue/No steps to reproduce]--> 100% ## Brave version (brave://version info) <!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details--> Brave | 1.44.78 Chromium: 105.0.5195.102 (Official Build) beta (x86_64) -- | -- Revision | 4c16f5ffcc2da70ee2600d5db77bed423ac03a5a-refs/branch-heads/5195_55@{#4} OS | macOS Version 13.0 (Build 22A5331f) ## Version/Channel Information: <!--Does this issue happen on any other channels? Or is it specific to a certain channel?--> - Can you reproduce this issue with the current release? - Can you reproduce this issue with the beta channel? - Can you reproduce this issue with the nightly channel? ## Other Additional Information: - Does the issue resolve itself when disabling Brave Shields? - Does the issue resolve itself when disabling Brave Rewards? - Is the issue reproducible on the latest version of Chrome? ## Miscellaneous Information: <!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue--> cc @boocmp
1.0
Need more spacing between `Submit` and `Cancel` dialogs for Tor bridges - <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue. PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE. INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED--> ## Description <!--Provide a brief description of the issue--> Need more spacing between `Submit` and `Cancel` dialogs for Tor bridges ## Steps to Reproduce <!--Please add a series of steps to reproduce the issue--> 1. install `1.44.78` 2. launch Brave 3. open `brave://settings/privacy` 4. scroll to `Use Bridges` 5. click on the `Request a New Bridge...` button 6. look at the spacing between the `Submit` and `Cancel` dialogs ## Actual result: <!--Please add screenshots if needed--> example | example ---------|--------- <img width="1300" alt="Screenshot 2022-09-07 at 2 31 20 PM" src="https://user-images.githubusercontent.com/387249/188987869-b3f94484-e4fd-4d4d-8c61-6b5f32a91baf.png"> | <img width="191" alt="Screenshot 2022-09-07 at 2 31 37 PM" src="https://user-images.githubusercontent.com/387249/188987876-00d54829-805e-4c87-b175-758d2259ef4e.png"> ## Expected result: Compare spacing with `brave://settings/clearBrowserData` example | example ---------|--------- <img width="1300" alt="Screenshot 2022-09-07 at 2 34 33 PM" src="https://user-images.githubusercontent.com/387249/188988138-d749c30f-c61f-4b85-b03c-1c1173c39f48.png"> | <img width="246" alt="Screenshot 2022-09-07 at 2 35 11 PM" src="https://user-images.githubusercontent.com/387249/188988139-6e76cf39-1cb4-4059-b1b3-21210556a084.png"> ## Reproduces how often: <!--[Easily reproduced/Intermittent issue/No steps to reproduce]--> 100% ## Brave version (brave://version info) <!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details--> Brave | 1.44.78 Chromium: 105.0.5195.102 (Official Build) beta (x86_64) -- | -- Revision | 4c16f5ffcc2da70ee2600d5db77bed423ac03a5a-refs/branch-heads/5195_55@{#4} OS | macOS Version 13.0 (Build 22A5331f) ## Version/Channel Information: <!--Does this issue happen on any other channels? Or is it specific to a certain channel?--> - Can you reproduce this issue with the current release? - Can you reproduce this issue with the beta channel? - Can you reproduce this issue with the nightly channel? ## Other Additional Information: - Does the issue resolve itself when disabling Brave Shields? - Does the issue resolve itself when disabling Brave Rewards? - Is the issue reproducible on the latest version of Chrome? ## Miscellaneous Information: <!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue--> cc @boocmp
test
need more spacing between submit and cancel dialogs for tor bridges have you searched for similar issues before submitting this issue please check the open issues and add a note before logging a new issue please use the template below to provide information about the issue insufficient info will get the issue closed it will only be reopened after sufficient info is provided description need more spacing between submit and cancel dialogs for tor bridges steps to reproduce install launch brave open brave settings privacy scroll to use bridges click on the request a new bridge button look at the spacing between the submit and cancel dialogs actual result example example img width alt screenshot at pm src img width alt screenshot at pm src expected result compare spacing with brave settings clearbrowserdata example example img width alt screenshot at pm src img width alt screenshot at pm src reproduces how often brave version brave version info brave chromium   official build  beta  revision refs branch heads os macos version build version channel information can you reproduce this issue with the current release can you reproduce this issue with the beta channel can you reproduce this issue with the nightly channel other additional information does the issue resolve itself when disabling brave shields does the issue resolve itself when disabling brave rewards is the issue reproducible on the latest version of chrome miscellaneous information cc boocmp
1
120,920
10,141,828,307
IssuesEvent
2019-08-03 17:45:27
dexpenses/dexpenses-extract
https://api.github.com/repos/dexpenses/dexpenses-extract
closed
Implement test receipt normal/bs-aldi-kastanienallee-cash
enhancement test-data
Receipt to implement: ![normal/bs-aldi-kastanienallee-cash](https://firebasestorage.googleapis.com/v0/b/dexpenses-207219-test-images/o/normal%2Fbs-aldi-kastanienallee-cash.JPEG?alt=media "normal/bs-aldi-kastanienallee-cash")
1.0
Implement test receipt normal/bs-aldi-kastanienallee-cash - Receipt to implement: ![normal/bs-aldi-kastanienallee-cash](https://firebasestorage.googleapis.com/v0/b/dexpenses-207219-test-images/o/normal%2Fbs-aldi-kastanienallee-cash.JPEG?alt=media "normal/bs-aldi-kastanienallee-cash")
test
implement test receipt normal bs aldi kastanienallee cash receipt to implement normal bs aldi kastanienallee cash
1
590,675
17,784,338,046
IssuesEvent
2021-08-31 09:14:06
EscolaLMS/Cart
https://api.github.com/repos/EscolaLMS/Cart
closed
admin GET `/order` missing `user_id` field
priority high
admin GET `/api/admin/orders` missing `user_id` field
1.0
admin GET `/order` missing `user_id` field - admin GET `/api/admin/orders` missing `user_id` field
non_test
admin get order missing user id field admin get api admin orders missing user id field
0
566,810
16,831,139,711
IssuesEvent
2021-06-18 05:08:37
fifasports/UAT
https://api.github.com/repos/fifasports/UAT
opened
#71 - UI improvement
Priority - Normal Type - Feature
Notice the commentary page is mixed between BG and RB. Highlighted are from RB Red box are from BG Maybe can have a buttons to switch between providers. A checkbox to on/off auto refresh will be good also, since match has ended, no need to spam office-api. ![image](https://user-images.githubusercontent.com/14333265/122509328-cad51500-d02d-11eb-9820-34c9c8468c62.png)
1.0
#71 - UI improvement - Notice the commentary page is mixed between BG and RB. Highlighted are from RB Red box are from BG Maybe can have a buttons to switch between providers. A checkbox to on/off auto refresh will be good also, since match has ended, no need to spam office-api. ![image](https://user-images.githubusercontent.com/14333265/122509328-cad51500-d02d-11eb-9820-34c9c8468c62.png)
non_test
ui improvement notice the commentary page is mixed between bg and rb highlighted are from rb red box are from bg maybe can have a buttons to switch between providers a checkbox to on off auto refresh will be good also since match has ended no need to spam office api
0