Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
757
labels
stringlengths
4
664
body
stringlengths
3
261k
index
stringclasses
10 values
text_combine
stringlengths
96
261k
label
stringclasses
2 values
text
stringlengths
96
232k
binary_label
int64
0
1
6,806
2,860,765,472
IssuesEvent
2015-06-03 17:21:56
dart-lang/sdk
https://api.github.com/repos/dart-lang/sdk
closed
[unittest] replace expectAsync* with runZonedExperimental?
Area-Test Priority-Unassigned Triaged Type-Enhancement
runZonedExperimental would allow test(...) to intercept "runAsync" calls in the test body and wait for them to complete. That way, you never need to worry about forgetting an "expectAsync" call somewhere.
1.0
[unittest] replace expectAsync* with runZonedExperimental? - runZonedExperimental would allow test(...) to intercept "runAsync" calls in the test body and wait for them to complete. That way, you never need to worry about forgetting an "expectAsync" call somewhere.
non_defect
replace expectasync with runzonedexperimental runzonedexperimental would allow test to intercept quot runasync quot calls in the test body and wait for them to complete that way you never need to worry about forgetting an quot expectasync quot call somewhere
0
81,968
31,836,873,032
IssuesEvent
2023-09-14 14:01:03
SeleniumHQ/selenium
https://api.github.com/repos/SeleniumHQ/selenium
closed
[🐛 Bug]: Headless mode is not working properly after google chrome update.
I-defect I-issue-template needs-triaging G-chromedriver
### What happened? Headless mode is not working properly after google chrome update. is.Displayed() and click() command is not working after google update python. Please Solve it . ### How can we reproduce the issue? ```shell You need to update selenium. ``` ### Relevant log output ```shell Error ``` ### Operating System Windows 10 ### Selenium version Python 4.12.0 (August 31, 2023) ### What are the browser(s) and version(s) where you see this issue? Version 117.0.5938.63 (Official Build) (64-bit) ### What are the browser driver(s) and version(s) where you see this issue? 117.0.5938.62 ### Are you using Selenium Grid? _No response_
1.0
[🐛 Bug]: Headless mode is not working properly after google chrome update. - ### What happened? Headless mode is not working properly after google chrome update. is.Displayed() and click() command is not working after google update python. Please Solve it . ### How can we reproduce the issue? ```shell You need to update selenium. ``` ### Relevant log output ```shell Error ``` ### Operating System Windows 10 ### Selenium version Python 4.12.0 (August 31, 2023) ### What are the browser(s) and version(s) where you see this issue? Version 117.0.5938.63 (Official Build) (64-bit) ### What are the browser driver(s) and version(s) where you see this issue? 117.0.5938.62 ### Are you using Selenium Grid? _No response_
defect
headless mode is not working properly after google chrome update what happened headless mode is not working properly after google chrome update is displayed and click command is not working after google update python please solve it how can we reproduce the issue shell you need to update selenium relevant log output shell error operating system windows selenium version python august what are the browser s and version s where you see this issue version official build bit what are the browser driver s and version s where you see this issue are you using selenium grid no response
1
80,297
15,380,995,620
IssuesEvent
2021-03-02 21:57:02
fossasia/open-event-frontend
https://api.github.com/repos/fossasia/open-event-frontend
opened
Make email in order form uneditable
Priority: High bug codeheat
Email is disabled in order form if user email is set, however, if you autofill in chrome, it is still overridden and when user tries to submit the order, it saves user's email as the one in form, which may result in an error. 1. If user email exists, the form field should be disabled and uneditable 2. Even if it somehow gets filled, it should not be saved as user's email if user's email already exists
1.0
Make email in order form uneditable - Email is disabled in order form if user email is set, however, if you autofill in chrome, it is still overridden and when user tries to submit the order, it saves user's email as the one in form, which may result in an error. 1. If user email exists, the form field should be disabled and uneditable 2. Even if it somehow gets filled, it should not be saved as user's email if user's email already exists
non_defect
make email in order form uneditable email is disabled in order form if user email is set however if you autofill in chrome it is still overridden and when user tries to submit the order it saves user s email as the one in form which may result in an error if user email exists the form field should be disabled and uneditable even if it somehow gets filled it should not be saved as user s email if user s email already exists
0
331,158
28,564,561,735
IssuesEvent
2023-04-21 00:02:26
facebook/react-native
https://api.github.com/repos/facebook/react-native
closed
Network error: the network request failed on some phone brand and POST type requests
Stale Platform: Android Platform: Linux Needs: Author Feedback Needs: Verify on Latest Version
## Description Hello Our application is in production and has problems for POST type requests. On some phone brands and all android versions (5, 6, 7,8,9,10). and on the other hand the Post type requests work very well on other phones and all android version (5,6,7,8,9,10) The exact error produced is: Error · Error: Network error: Network request failed. React native command result: Different solutions already tested: - By default, access to our backend is via https using cloudflare. I even thought that it could be an https certificate problem, I changed cloudflare to use a certificate from Let's Encryptletsencrypt, until then we have the same problem. We figured this was a problem with the https protocol and decided on http and added to AndroidManifest.xml the directive: ``` <android application: usesCleartextTraffic = "true"> </application> ``` despite all this we have the same problem namely the POST type request are not functional on some brands of phone and any android version (5,6,7,8,9,10) Thank you for any help in relation to this problem because at this time we already no longer block more than 1200 Registrations on our Mobile application. ## React Native version: ``` System: OS: Linux 4.19 Debian GNU/Linux 10 (buster) 10 (buster) CPU: (4) x64 Intel(R) Core(TM) i3-5005U CPU @ 2.00GHz Memory: 761.15 MB / 11.65 GB Shell: 5.0.3 - /bin/bash Binaries: Node: 10.15.2 - /usr/bin/node Yarn: 1.17.3 - /usr/bin/yarn npm: 6.14.5 - /usr/bin/npm IDEs: Android Studio: 4.0 AI-193.6911.18.40.6514223 npmPackages: react: 16.9.0 => 16.9.0 react-native: 0.61.5 => 0.61.5 npmGlobalPackages: react-native-cli: 2.0.1 react-native-rename: 2.4.1 ```
1.0
Network error: the network request failed on some phone brand and POST type requests - ## Description Hello Our application is in production and has problems for POST type requests. On some phone brands and all android versions (5, 6, 7,8,9,10). and on the other hand the Post type requests work very well on other phones and all android version (5,6,7,8,9,10) The exact error produced is: Error · Error: Network error: Network request failed. React native command result: Different solutions already tested: - By default, access to our backend is via https using cloudflare. I even thought that it could be an https certificate problem, I changed cloudflare to use a certificate from Let's Encryptletsencrypt, until then we have the same problem. We figured this was a problem with the https protocol and decided on http and added to AndroidManifest.xml the directive: ``` <android application: usesCleartextTraffic = "true"> </application> ``` despite all this we have the same problem namely the POST type request are not functional on some brands of phone and any android version (5,6,7,8,9,10) Thank you for any help in relation to this problem because at this time we already no longer block more than 1200 Registrations on our Mobile application. ## React Native version: ``` System: OS: Linux 4.19 Debian GNU/Linux 10 (buster) 10 (buster) CPU: (4) x64 Intel(R) Core(TM) i3-5005U CPU @ 2.00GHz Memory: 761.15 MB / 11.65 GB Shell: 5.0.3 - /bin/bash Binaries: Node: 10.15.2 - /usr/bin/node Yarn: 1.17.3 - /usr/bin/yarn npm: 6.14.5 - /usr/bin/npm IDEs: Android Studio: 4.0 AI-193.6911.18.40.6514223 npmPackages: react: 16.9.0 => 16.9.0 react-native: 0.61.5 => 0.61.5 npmGlobalPackages: react-native-cli: 2.0.1 react-native-rename: 2.4.1 ```
non_defect
network error the network request failed on some phone brand and post type requests description hello our application is in production and has problems for post type requests on some phone brands and all android versions and on the other hand the post type requests work very well on other phones and all android version the exact error produced is error · error network error network request failed react native command result different solutions already tested by default access to our backend is via https using cloudflare i even thought that it could be an https certificate problem i changed cloudflare to use a certificate from let s encryptletsencrypt until then we have the same problem we figured this was a problem with the https protocol and decided on http and added to androidmanifest xml the directive despite all this we have the same problem namely the post type request are not functional on some brands of phone and any android version thank you for any help in relation to this problem because at this time we already no longer block more than registrations on our mobile application react native version system os linux debian gnu linux buster buster cpu intel r core tm cpu memory mb gb shell bin bash binaries node usr bin node yarn usr bin yarn npm usr bin npm ides android studio ai npmpackages react react native npmglobalpackages react native cli react native rename
0
653,812
21,627,100,700
IssuesEvent
2022-05-05 04:42:00
googleapis/java-spanner
https://api.github.com/repos/googleapis/java-spanner
closed
spanner.it.ITDatabaseTest: databaseDeletedTest failed
priority: p2 type: bug api: spanner flakybot: issue flakybot: flaky
Note: #1772 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky. ---- commit: 0be09d5f91521b0d23086a99f2ae46c2dd99b64a buildURL: [Build Status](https://source.cloud.google.com/results/invocations/393f0d5f-eb50-4c38-8584-39995aee3236), [Sponge](http://sponge2/393f0d5f-eb50-4c38-8584-39995aee3236) status: failed <details><summary>Test output</summary><br><pre>com.google.cloud.spanner.SpannerException: RESOURCE_EXHAUSTED: com.google.cloud.spanner.SpannerException: RESOURCE_EXHAUSTED: io.grpc.StatusRuntimeException: RESOURCE_EXHAUSTED: Unable to create database 'testdb_1595150865_0007' because the instance 'projects/gcloud-devel/instances/spanner-testing-east1' has already reached the maximum database limit (100). Please delete a database in the instance and try again, or choose a different instance. at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerExceptionPreformatted(SpannerExceptionFactory.java:284) at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerException(SpannerExceptionFactory.java:61) at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerException(SpannerExceptionFactory.java:181) at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerException(SpannerExceptionFactory.java:110) at com.google.cloud.spanner.testing.RemoteSpannerHelper.createTestDatabase(RemoteSpannerHelper.java:148) at com.google.cloud.spanner.testing.RemoteSpannerHelper.createTestDatabase(RemoteSpannerHelper.java:100) at com.google.cloud.spanner.it.ITDatabaseTest.databaseDeletedTest(ITDatabaseTest.java:80) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54) at org.junit.rules.RunRules.evaluate(RunRules.java:20) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.apache.maven.surefire.junitcore.JUnitCore.run(JUnitCore.java:55) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.createRequestAndRun(JUnitCoreWrapper.java:137) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.executeLazy(JUnitCoreWrapper.java:119) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:87) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:75) at org.apache.maven.surefire.junitcore.JUnitCoreProvider.invoke(JUnitCoreProvider.java:158) at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:377) at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:138) at org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:465) at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:451) Caused by: java.util.concurrent.ExecutionException: com.google.cloud.spanner.SpannerException: RESOURCE_EXHAUSTED: io.grpc.StatusRuntimeException: RESOURCE_EXHAUSTED: Unable to create database 'testdb_1595150865_0007' because the instance 'projects/gcloud-devel/instances/spanner-testing-east1' has already reached the maximum database limit (100). Please delete a database in the instance and try again, or choose a different instance. at com.google.common.util.concurrent.AbstractFuture.getDoneValue(AbstractFuture.java:588) at com.google.common.util.concurrent.AbstractFuture.get(AbstractFuture.java:567) at com.google.common.util.concurrent.FluentFuture$TrustedFuture.get(FluentFuture.java:91) at com.google.common.util.concurrent.ForwardingFuture.get(ForwardingFuture.java:66) at com.google.api.gax.longrunning.OperationFutureImpl.get(OperationFutureImpl.java:125) at com.google.cloud.spanner.testing.RemoteSpannerHelper.createTestDatabase(RemoteSpannerHelper.java:137) ... 43 more Caused by: com.google.cloud.spanner.SpannerException: RESOURCE_EXHAUSTED: io.grpc.StatusRuntimeException: RESOURCE_EXHAUSTED: Unable to create database 'testdb_1595150865_0007' because the instance 'projects/gcloud-devel/instances/spanner-testing-east1' has already reached the maximum database limit (100). Please delete a database in the instance and try again, or choose a different instance. at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerExceptionPreformatted(SpannerExceptionFactory.java:284) at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerException(SpannerExceptionFactory.java:61) at com.google.cloud.spanner.SpannerExceptionFactory.fromApiException(SpannerExceptionFactory.java:299) at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerException(SpannerExceptionFactory.java:174) at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerException(SpannerExceptionFactory.java:110) at com.google.cloud.spanner.DatabaseAdminClientImpl.lambda$createDatabase$7(DatabaseAdminClientImpl.java:350) at com.google.api.core.ApiFutures$ApiFunctionToGuavaFunction.apply(ApiFutures.java:239) at com.google.common.util.concurrent.AbstractCatchingFuture$CatchingFuture.doFallback(AbstractCatchingFuture.java:234) at com.google.common.util.concurrent.AbstractCatchingFuture$CatchingFuture.doFallback(AbstractCatchingFuture.java:222) at com.google.common.util.concurrent.AbstractCatchingFuture.run(AbstractCatchingFuture.java:133) at com.google.common.util.concurrent.DirectExecutor.execute(DirectExecutor.java:31) at com.google.common.util.concurrent.AbstractFuture.executeListener(AbstractFuture.java:1270) at com.google.common.util.concurrent.AbstractFuture.complete(AbstractFuture.java:1038) at com.google.common.util.concurrent.AbstractFuture.setException(AbstractFuture.java:808) at com.google.common.util.concurrent.AbstractTransformFuture.run(AbstractTransformFuture.java:104) at com.google.common.util.concurrent.DirectExecutor.execute(DirectExecutor.java:31) at com.google.common.util.concurrent.AbstractFuture.executeListener(AbstractFuture.java:1270) at com.google.common.util.concurrent.AbstractFuture.complete(AbstractFuture.java:1038) at com.google.common.util.concurrent.AbstractFuture.setException(AbstractFuture.java:808) at com.google.api.gax.retrying.BasicRetryingFuture.handleAttempt(BasicRetryingFuture.java:200) at com.google.api.gax.retrying.CallbackChainRetryingFuture$AttemptCompletionListener.handle(CallbackChainRetryingFuture.java:135) at com.google.api.gax.retrying.CallbackChainRetryingFuture$AttemptCompletionListener.run(CallbackChainRetryingFuture.java:117) at com.google.common.util.concurrent.DirectExecutor.execute(DirectExecutor.java:31) at com.google.common.util.concurrent.AbstractFuture.executeListener(AbstractFuture.java:1270) at com.google.common.util.concurrent.AbstractFuture.complete(AbstractFuture.java:1038) at com.google.common.util.concurrent.AbstractFuture.setException(AbstractFuture.java:808) at com.google.common.util.concurrent.AbstractTransformFuture.run(AbstractTransformFuture.java:104) at com.google.common.util.concurrent.DirectExecutor.execute(DirectExecutor.java:31) at com.google.common.util.concurrent.AbstractFuture.executeListener(AbstractFuture.java:1270) at com.google.common.util.concurrent.AbstractFuture.complete(AbstractFuture.java:1038) at com.google.common.util.concurrent.AbstractFuture.setException(AbstractFuture.java:808) at com.google.api.gax.retrying.BasicRetryingFuture.handleAttempt(BasicRetryingFuture.java:200) at com.google.api.gax.retrying.CallbackChainRetryingFuture$AttemptCompletionListener.handle(CallbackChainRetryingFuture.java:135) at com.google.api.gax.retrying.CallbackChainRetryingFuture$AttemptCompletionListener.run(CallbackChainRetryingFuture.java:117) at com.google.common.util.concurrent.DirectExecutor.execute(DirectExecutor.java:31) at com.google.common.util.concurrent.AbstractFuture.executeListener(AbstractFuture.java:1270) at com.google.common.util.concurrent.AbstractFuture.complete(AbstractFuture.java:1038) at com.google.common.util.concurrent.AbstractFuture.setException(AbstractFuture.java:808) at com.google.api.core.AbstractApiFuture$InternalSettableFuture.setException(AbstractApiFuture.java:94) at com.google.api.core.AbstractApiFuture.setException(AbstractApiFuture.java:76) at com.google.api.gax.grpc.GrpcExceptionCallable$ExceptionTransformingFuture.onFailure(GrpcExceptionCallable.java:97) at com.google.api.core.ApiFutures$1.onFailure(ApiFutures.java:67) at com.google.common.util.concurrent.Futures$CallbackListener.run(Futures.java:1132) at com.google.common.util.concurrent.DirectExecutor.execute(DirectExecutor.java:31) at com.google.common.util.concurrent.AbstractFuture.executeListener(AbstractFuture.java:1270) at com.google.common.util.concurrent.AbstractFuture.complete(AbstractFuture.java:1038) at com.google.common.util.concurrent.AbstractFuture.setException(AbstractFuture.java:808) at io.grpc.stub.ClientCalls$GrpcFuture.setException(ClientCalls.java:563) at io.grpc.stub.ClientCalls$UnaryStreamToFuture.onClose(ClientCalls.java:533) at io.grpc.PartialForwardingClientCallListener.onClose(PartialForwardingClientCallListener.java:39) at io.grpc.ForwardingClientCallListener.onClose(ForwardingClientCallListener.java:23) at io.grpc.ForwardingClientCallListener$SimpleForwardingClientCallListener.onClose(ForwardingClientCallListener.java:40) at com.google.api.gax.grpc.ChannelPool$ReleasingClientCall$1.onClose(ChannelPool.java:535) at io.grpc.PartialForwardingClientCallListener.onClose(PartialForwardingClientCallListener.java:39) at io.grpc.ForwardingClientCallListener.onClose(ForwardingClientCallListener.java:23) at io.grpc.ForwardingClientCallListener$SimpleForwardingClientCallListener.onClose(ForwardingClientCallListener.java:40) at io.grpc.PartialForwardingClientCallListener.onClose(PartialForwardingClientCallListener.java:39) at io.grpc.ForwardingClientCallListener.onClose(ForwardingClientCallListener.java:23) at io.grpc.ForwardingClientCallListener$SimpleForwardingClientCallListener.onClose(ForwardingClientCallListener.java:40) at com.google.cloud.spanner.spi.v1.SpannerErrorInterceptor$1$1.onClose(SpannerErrorInterceptor.java:100) at io.grpc.internal.DelayedClientCall$DelayedListener$3.run(DelayedClientCall.java:463) at io.grpc.internal.DelayedClientCall$DelayedListener.delayOrExecute(DelayedClientCall.java:427) at io.grpc.internal.DelayedClientCall$DelayedListener.onClose(DelayedClientCall.java:460) at io.grpc.internal.ClientCallImpl.closeObserver(ClientCallImpl.java:562) at io.grpc.internal.ClientCallImpl.access$300(ClientCallImpl.java:70) at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInternal(ClientCallImpl.java:743) at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInContext(ClientCallImpl.java:722) at io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37) at io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:133) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:750) Caused by: io.grpc.StatusRuntimeException: RESOURCE_EXHAUSTED: Unable to create database 'testdb_1595150865_0007' because the instance 'projects/gcloud-devel/instances/spanner-testing-east1' has already reached the maximum database limit (100). Please delete a database in the instance and try again, or choose a different instance. at io.grpc.Status.asRuntimeException(Status.java:535) ... 28 more </pre></details>
1.0
spanner.it.ITDatabaseTest: databaseDeletedTest failed - Note: #1772 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky. ---- commit: 0be09d5f91521b0d23086a99f2ae46c2dd99b64a buildURL: [Build Status](https://source.cloud.google.com/results/invocations/393f0d5f-eb50-4c38-8584-39995aee3236), [Sponge](http://sponge2/393f0d5f-eb50-4c38-8584-39995aee3236) status: failed <details><summary>Test output</summary><br><pre>com.google.cloud.spanner.SpannerException: RESOURCE_EXHAUSTED: com.google.cloud.spanner.SpannerException: RESOURCE_EXHAUSTED: io.grpc.StatusRuntimeException: RESOURCE_EXHAUSTED: Unable to create database 'testdb_1595150865_0007' because the instance 'projects/gcloud-devel/instances/spanner-testing-east1' has already reached the maximum database limit (100). Please delete a database in the instance and try again, or choose a different instance. at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerExceptionPreformatted(SpannerExceptionFactory.java:284) at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerException(SpannerExceptionFactory.java:61) at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerException(SpannerExceptionFactory.java:181) at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerException(SpannerExceptionFactory.java:110) at com.google.cloud.spanner.testing.RemoteSpannerHelper.createTestDatabase(RemoteSpannerHelper.java:148) at com.google.cloud.spanner.testing.RemoteSpannerHelper.createTestDatabase(RemoteSpannerHelper.java:100) at com.google.cloud.spanner.it.ITDatabaseTest.databaseDeletedTest(ITDatabaseTest.java:80) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54) at org.junit.rules.RunRules.evaluate(RunRules.java:20) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.junit.runners.Suite.runChild(Suite.java:128) at org.junit.runners.Suite.runChild(Suite.java:27) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.apache.maven.surefire.junitcore.JUnitCore.run(JUnitCore.java:55) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.createRequestAndRun(JUnitCoreWrapper.java:137) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.executeLazy(JUnitCoreWrapper.java:119) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:87) at org.apache.maven.surefire.junitcore.JUnitCoreWrapper.execute(JUnitCoreWrapper.java:75) at org.apache.maven.surefire.junitcore.JUnitCoreProvider.invoke(JUnitCoreProvider.java:158) at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:377) at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:138) at org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:465) at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:451) Caused by: java.util.concurrent.ExecutionException: com.google.cloud.spanner.SpannerException: RESOURCE_EXHAUSTED: io.grpc.StatusRuntimeException: RESOURCE_EXHAUSTED: Unable to create database 'testdb_1595150865_0007' because the instance 'projects/gcloud-devel/instances/spanner-testing-east1' has already reached the maximum database limit (100). Please delete a database in the instance and try again, or choose a different instance. at com.google.common.util.concurrent.AbstractFuture.getDoneValue(AbstractFuture.java:588) at com.google.common.util.concurrent.AbstractFuture.get(AbstractFuture.java:567) at com.google.common.util.concurrent.FluentFuture$TrustedFuture.get(FluentFuture.java:91) at com.google.common.util.concurrent.ForwardingFuture.get(ForwardingFuture.java:66) at com.google.api.gax.longrunning.OperationFutureImpl.get(OperationFutureImpl.java:125) at com.google.cloud.spanner.testing.RemoteSpannerHelper.createTestDatabase(RemoteSpannerHelper.java:137) ... 43 more Caused by: com.google.cloud.spanner.SpannerException: RESOURCE_EXHAUSTED: io.grpc.StatusRuntimeException: RESOURCE_EXHAUSTED: Unable to create database 'testdb_1595150865_0007' because the instance 'projects/gcloud-devel/instances/spanner-testing-east1' has already reached the maximum database limit (100). Please delete a database in the instance and try again, or choose a different instance. at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerExceptionPreformatted(SpannerExceptionFactory.java:284) at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerException(SpannerExceptionFactory.java:61) at com.google.cloud.spanner.SpannerExceptionFactory.fromApiException(SpannerExceptionFactory.java:299) at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerException(SpannerExceptionFactory.java:174) at com.google.cloud.spanner.SpannerExceptionFactory.newSpannerException(SpannerExceptionFactory.java:110) at com.google.cloud.spanner.DatabaseAdminClientImpl.lambda$createDatabase$7(DatabaseAdminClientImpl.java:350) at com.google.api.core.ApiFutures$ApiFunctionToGuavaFunction.apply(ApiFutures.java:239) at com.google.common.util.concurrent.AbstractCatchingFuture$CatchingFuture.doFallback(AbstractCatchingFuture.java:234) at com.google.common.util.concurrent.AbstractCatchingFuture$CatchingFuture.doFallback(AbstractCatchingFuture.java:222) at com.google.common.util.concurrent.AbstractCatchingFuture.run(AbstractCatchingFuture.java:133) at com.google.common.util.concurrent.DirectExecutor.execute(DirectExecutor.java:31) at com.google.common.util.concurrent.AbstractFuture.executeListener(AbstractFuture.java:1270) at com.google.common.util.concurrent.AbstractFuture.complete(AbstractFuture.java:1038) at com.google.common.util.concurrent.AbstractFuture.setException(AbstractFuture.java:808) at com.google.common.util.concurrent.AbstractTransformFuture.run(AbstractTransformFuture.java:104) at com.google.common.util.concurrent.DirectExecutor.execute(DirectExecutor.java:31) at com.google.common.util.concurrent.AbstractFuture.executeListener(AbstractFuture.java:1270) at com.google.common.util.concurrent.AbstractFuture.complete(AbstractFuture.java:1038) at com.google.common.util.concurrent.AbstractFuture.setException(AbstractFuture.java:808) at com.google.api.gax.retrying.BasicRetryingFuture.handleAttempt(BasicRetryingFuture.java:200) at com.google.api.gax.retrying.CallbackChainRetryingFuture$AttemptCompletionListener.handle(CallbackChainRetryingFuture.java:135) at com.google.api.gax.retrying.CallbackChainRetryingFuture$AttemptCompletionListener.run(CallbackChainRetryingFuture.java:117) at com.google.common.util.concurrent.DirectExecutor.execute(DirectExecutor.java:31) at com.google.common.util.concurrent.AbstractFuture.executeListener(AbstractFuture.java:1270) at com.google.common.util.concurrent.AbstractFuture.complete(AbstractFuture.java:1038) at com.google.common.util.concurrent.AbstractFuture.setException(AbstractFuture.java:808) at com.google.common.util.concurrent.AbstractTransformFuture.run(AbstractTransformFuture.java:104) at com.google.common.util.concurrent.DirectExecutor.execute(DirectExecutor.java:31) at com.google.common.util.concurrent.AbstractFuture.executeListener(AbstractFuture.java:1270) at com.google.common.util.concurrent.AbstractFuture.complete(AbstractFuture.java:1038) at com.google.common.util.concurrent.AbstractFuture.setException(AbstractFuture.java:808) at com.google.api.gax.retrying.BasicRetryingFuture.handleAttempt(BasicRetryingFuture.java:200) at com.google.api.gax.retrying.CallbackChainRetryingFuture$AttemptCompletionListener.handle(CallbackChainRetryingFuture.java:135) at com.google.api.gax.retrying.CallbackChainRetryingFuture$AttemptCompletionListener.run(CallbackChainRetryingFuture.java:117) at com.google.common.util.concurrent.DirectExecutor.execute(DirectExecutor.java:31) at com.google.common.util.concurrent.AbstractFuture.executeListener(AbstractFuture.java:1270) at com.google.common.util.concurrent.AbstractFuture.complete(AbstractFuture.java:1038) at com.google.common.util.concurrent.AbstractFuture.setException(AbstractFuture.java:808) at com.google.api.core.AbstractApiFuture$InternalSettableFuture.setException(AbstractApiFuture.java:94) at com.google.api.core.AbstractApiFuture.setException(AbstractApiFuture.java:76) at com.google.api.gax.grpc.GrpcExceptionCallable$ExceptionTransformingFuture.onFailure(GrpcExceptionCallable.java:97) at com.google.api.core.ApiFutures$1.onFailure(ApiFutures.java:67) at com.google.common.util.concurrent.Futures$CallbackListener.run(Futures.java:1132) at com.google.common.util.concurrent.DirectExecutor.execute(DirectExecutor.java:31) at com.google.common.util.concurrent.AbstractFuture.executeListener(AbstractFuture.java:1270) at com.google.common.util.concurrent.AbstractFuture.complete(AbstractFuture.java:1038) at com.google.common.util.concurrent.AbstractFuture.setException(AbstractFuture.java:808) at io.grpc.stub.ClientCalls$GrpcFuture.setException(ClientCalls.java:563) at io.grpc.stub.ClientCalls$UnaryStreamToFuture.onClose(ClientCalls.java:533) at io.grpc.PartialForwardingClientCallListener.onClose(PartialForwardingClientCallListener.java:39) at io.grpc.ForwardingClientCallListener.onClose(ForwardingClientCallListener.java:23) at io.grpc.ForwardingClientCallListener$SimpleForwardingClientCallListener.onClose(ForwardingClientCallListener.java:40) at com.google.api.gax.grpc.ChannelPool$ReleasingClientCall$1.onClose(ChannelPool.java:535) at io.grpc.PartialForwardingClientCallListener.onClose(PartialForwardingClientCallListener.java:39) at io.grpc.ForwardingClientCallListener.onClose(ForwardingClientCallListener.java:23) at io.grpc.ForwardingClientCallListener$SimpleForwardingClientCallListener.onClose(ForwardingClientCallListener.java:40) at io.grpc.PartialForwardingClientCallListener.onClose(PartialForwardingClientCallListener.java:39) at io.grpc.ForwardingClientCallListener.onClose(ForwardingClientCallListener.java:23) at io.grpc.ForwardingClientCallListener$SimpleForwardingClientCallListener.onClose(ForwardingClientCallListener.java:40) at com.google.cloud.spanner.spi.v1.SpannerErrorInterceptor$1$1.onClose(SpannerErrorInterceptor.java:100) at io.grpc.internal.DelayedClientCall$DelayedListener$3.run(DelayedClientCall.java:463) at io.grpc.internal.DelayedClientCall$DelayedListener.delayOrExecute(DelayedClientCall.java:427) at io.grpc.internal.DelayedClientCall$DelayedListener.onClose(DelayedClientCall.java:460) at io.grpc.internal.ClientCallImpl.closeObserver(ClientCallImpl.java:562) at io.grpc.internal.ClientCallImpl.access$300(ClientCallImpl.java:70) at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInternal(ClientCallImpl.java:743) at io.grpc.internal.ClientCallImpl$ClientStreamListenerImpl$1StreamClosed.runInContext(ClientCallImpl.java:722) at io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37) at io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:133) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:750) Caused by: io.grpc.StatusRuntimeException: RESOURCE_EXHAUSTED: Unable to create database 'testdb_1595150865_0007' because the instance 'projects/gcloud-devel/instances/spanner-testing-east1' has already reached the maximum database limit (100). Please delete a database in the instance and try again, or choose a different instance. at io.grpc.Status.asRuntimeException(Status.java:535) ... 28 more </pre></details>
non_defect
spanner it itdatabasetest databasedeletedtest failed note was also for this test but it was closed more than days ago so i didn t mark it flaky commit buildurl status failed test output com google cloud spanner spannerexception resource exhausted com google cloud spanner spannerexception resource exhausted io grpc statusruntimeexception resource exhausted unable to create database testdb because the instance projects gcloud devel instances spanner testing has already reached the maximum database limit please delete a database in the instance and try again or choose a different instance at com google cloud spanner spannerexceptionfactory newspannerexceptionpreformatted spannerexceptionfactory java at com google cloud spanner spannerexceptionfactory newspannerexception spannerexceptionfactory java at com google cloud spanner spannerexceptionfactory newspannerexception spannerexceptionfactory java at com google cloud spanner spannerexceptionfactory newspannerexception spannerexceptionfactory java at com google cloud spanner testing remotespannerhelper createtestdatabase remotespannerhelper java at com google cloud spanner testing remotespannerhelper createtestdatabase remotespannerhelper java at com google cloud spanner it itdatabasetest databasedeletedtest itdatabasetest java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at org junit runners parentrunner evaluate parentrunner java at org junit runners evaluate java at org junit runners parentrunner runleaf parentrunner java at org junit runners runchild java at org junit runners runchild java at org junit runners parentrunner run parentrunner java at org junit runners parentrunner schedule parentrunner java at org junit runners parentrunner runchildren parentrunner java at org junit runners parentrunner access parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit rules externalresource evaluate externalresource java at org junit rules runrules evaluate runrules java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner run parentrunner java at org junit runners suite runchild suite java at org junit runners suite runchild suite java at org junit runners parentrunner run parentrunner java at org junit runners parentrunner schedule parentrunner java at org junit runners parentrunner runchildren parentrunner java at org junit runners parentrunner access parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner run parentrunner java at org apache maven surefire junitcore junitcore run junitcore java at org apache maven surefire junitcore junitcorewrapper createrequestandrun junitcorewrapper java at org apache maven surefire junitcore junitcorewrapper executelazy junitcorewrapper java at org apache maven surefire junitcore junitcorewrapper execute junitcorewrapper java at org apache maven surefire junitcore junitcorewrapper execute junitcorewrapper java at org apache maven surefire junitcore junitcoreprovider invoke junitcoreprovider java at org apache maven surefire booter forkedbooter runsuitesinprocess forkedbooter java at org apache maven surefire booter forkedbooter execute forkedbooter java at org apache maven surefire booter forkedbooter run forkedbooter java at org apache maven surefire booter forkedbooter main forkedbooter java caused by java util concurrent executionexception com google cloud spanner spannerexception resource exhausted io grpc statusruntimeexception resource exhausted unable to create database testdb because the instance projects gcloud devel instances spanner testing has already reached the maximum database limit please delete a database in the instance and try again or choose a different instance at com google common util concurrent abstractfuture getdonevalue abstractfuture java at com google common util concurrent abstractfuture get abstractfuture java at com google common util concurrent fluentfuture trustedfuture get fluentfuture java at com google common util concurrent forwardingfuture get forwardingfuture java at com google api gax longrunning operationfutureimpl get operationfutureimpl java at com google cloud spanner testing remotespannerhelper createtestdatabase remotespannerhelper java more caused by com google cloud spanner spannerexception resource exhausted io grpc statusruntimeexception resource exhausted unable to create database testdb because the instance projects gcloud devel instances spanner testing has already reached the maximum database limit please delete a database in the instance and try again or choose a different instance at com google cloud spanner spannerexceptionfactory newspannerexceptionpreformatted spannerexceptionfactory java at com google cloud spanner spannerexceptionfactory newspannerexception spannerexceptionfactory java at com google cloud spanner spannerexceptionfactory fromapiexception spannerexceptionfactory java at com google cloud spanner spannerexceptionfactory newspannerexception spannerexceptionfactory java at com google cloud spanner spannerexceptionfactory newspannerexception spannerexceptionfactory java at com google cloud spanner databaseadminclientimpl lambda createdatabase databaseadminclientimpl java at com google api core apifutures apifunctiontoguavafunction apply apifutures java at com google common util concurrent abstractcatchingfuture catchingfuture dofallback abstractcatchingfuture java at com google common util concurrent abstractcatchingfuture catchingfuture dofallback abstractcatchingfuture java at com google common util concurrent abstractcatchingfuture run abstractcatchingfuture java at com google common util concurrent directexecutor execute directexecutor java at com google common util concurrent abstractfuture executelistener abstractfuture java at com google common util concurrent abstractfuture complete abstractfuture java at com google common util concurrent abstractfuture setexception abstractfuture java at com google common util concurrent abstracttransformfuture run abstracttransformfuture java at com google common util concurrent directexecutor execute directexecutor java at com google common util concurrent abstractfuture executelistener abstractfuture java at com google common util concurrent abstractfuture complete abstractfuture java at com google common util concurrent abstractfuture setexception abstractfuture java at com google api gax retrying basicretryingfuture handleattempt basicretryingfuture java at com google api gax retrying callbackchainretryingfuture attemptcompletionlistener handle callbackchainretryingfuture java at com google api gax retrying callbackchainretryingfuture attemptcompletionlistener run callbackchainretryingfuture java at com google common util concurrent directexecutor execute directexecutor java at com google common util concurrent abstractfuture executelistener abstractfuture java at com google common util concurrent abstractfuture complete abstractfuture java at com google common util concurrent abstractfuture setexception abstractfuture java at com google common util concurrent abstracttransformfuture run abstracttransformfuture java at com google common util concurrent directexecutor execute directexecutor java at com google common util concurrent abstractfuture executelistener abstractfuture java at com google common util concurrent abstractfuture complete abstractfuture java at com google common util concurrent abstractfuture setexception abstractfuture java at com google api gax retrying basicretryingfuture handleattempt basicretryingfuture java at com google api gax retrying callbackchainretryingfuture attemptcompletionlistener handle callbackchainretryingfuture java at com google api gax retrying callbackchainretryingfuture attemptcompletionlistener run callbackchainretryingfuture java at com google common util concurrent directexecutor execute directexecutor java at com google common util concurrent abstractfuture executelistener abstractfuture java at com google common util concurrent abstractfuture complete abstractfuture java at com google common util concurrent abstractfuture setexception abstractfuture java at com google api core abstractapifuture internalsettablefuture setexception abstractapifuture java at com google api core abstractapifuture setexception abstractapifuture java at com google api gax grpc grpcexceptioncallable exceptiontransformingfuture onfailure grpcexceptioncallable java at com google api core apifutures onfailure apifutures java at com google common util concurrent futures callbacklistener run futures java at com google common util concurrent directexecutor execute directexecutor java at com google common util concurrent abstractfuture executelistener abstractfuture java at com google common util concurrent abstractfuture complete abstractfuture java at com google common util concurrent abstractfuture setexception abstractfuture java at io grpc stub clientcalls grpcfuture setexception clientcalls java at io grpc stub clientcalls unarystreamtofuture onclose clientcalls java at io grpc partialforwardingclientcalllistener onclose partialforwardingclientcalllistener java at io grpc forwardingclientcalllistener onclose forwardingclientcalllistener java at io grpc forwardingclientcalllistener simpleforwardingclientcalllistener onclose forwardingclientcalllistener java at com google api gax grpc channelpool releasingclientcall onclose channelpool java at io grpc partialforwardingclientcalllistener onclose partialforwardingclientcalllistener java at io grpc forwardingclientcalllistener onclose forwardingclientcalllistener java at io grpc forwardingclientcalllistener simpleforwardingclientcalllistener onclose forwardingclientcalllistener java at io grpc partialforwardingclientcalllistener onclose partialforwardingclientcalllistener java at io grpc forwardingclientcalllistener onclose forwardingclientcalllistener java at io grpc forwardingclientcalllistener simpleforwardingclientcalllistener onclose forwardingclientcalllistener java at com google cloud spanner spi spannererrorinterceptor onclose spannererrorinterceptor java at io grpc internal delayedclientcall delayedlistener run delayedclientcall java at io grpc internal delayedclientcall delayedlistener delayorexecute delayedclientcall java at io grpc internal delayedclientcall delayedlistener onclose delayedclientcall java at io grpc internal clientcallimpl closeobserver clientcallimpl java at io grpc internal clientcallimpl access clientcallimpl java at io grpc internal clientcallimpl clientstreamlistenerimpl runinternal clientcallimpl java at io grpc internal clientcallimpl clientstreamlistenerimpl runincontext clientcallimpl java at io grpc internal contextrunnable run contextrunnable java at io grpc internal serializingexecutor run serializingexecutor java at java util concurrent executors runnableadapter call executors java at java util concurrent futuretask run futuretask java at java util concurrent scheduledthreadpoolexecutor scheduledfuturetask access scheduledthreadpoolexecutor java at java util concurrent scheduledthreadpoolexecutor scheduledfuturetask run scheduledthreadpoolexecutor java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java caused by io grpc statusruntimeexception resource exhausted unable to create database testdb because the instance projects gcloud devel instances spanner testing has already reached the maximum database limit please delete a database in the instance and try again or choose a different instance at io grpc status asruntimeexception status java more
0
56,084
14,924,461,137
IssuesEvent
2021-01-24 00:01:21
BOINC/boinc
https://api.github.com/repos/BOINC/boinc
closed
[Android] Log output has extra EOLs before and after every message
C: Android - Manager E: 1 day Newbie P: Trivial R: fixed T: Defect
**System Information (please complete the following information):** - OS: Android - BOINC Version: 7.16.3
1.0
[Android] Log output has extra EOLs before and after every message - **System Information (please complete the following information):** - OS: Android - BOINC Version: 7.16.3
defect
log output has extra eols before and after every message system information please complete the following information os android boinc version
1
10,031
2,618,932,237
IssuesEvent
2015-03-03 00:00:43
chrsmith/open-ig
https://api.github.com/repos/chrsmith/open-ig
closed
NPE when attempting to load saved game after update
auto-migrated Priority-Medium Technology Type-Defect
``` Game version: 0.95.124 Operating System: Windows XP (32 bit) Java runtime version: 1.7.0_10-b18 Installed using the Launcher? yes Game language (en, hu, de): hu What steps will reproduce the problem? 1. Start and play a campaign using version 0.95.123 2. Update to 0.95.124 using the update button 3. Try to resume your campaign What is the expected output? What do you see instead? The expected output: no exception :) See the attachment. Please provide any additional information below. Starting a new game works fine but loading the saved campaign fails. The campaign was played a fair amount of time (Garthogs conquered, met the Dargslans). Is there an easy way to revert the upgrade? Please upload any save before and/or after the problem happened in zip compressed format. Please attach the open-ig.log file found in the application's directory. ``` Original issue reported on code.google.com by `pistu...@gmail.com` on 2 Jan 2013 at 9:07 Attachments: * [openig-crash.zip](https://storage.googleapis.com/google-code-attachments/open-ig/issue-709/comment-0/openig-crash.zip)
1.0
NPE when attempting to load saved game after update - ``` Game version: 0.95.124 Operating System: Windows XP (32 bit) Java runtime version: 1.7.0_10-b18 Installed using the Launcher? yes Game language (en, hu, de): hu What steps will reproduce the problem? 1. Start and play a campaign using version 0.95.123 2. Update to 0.95.124 using the update button 3. Try to resume your campaign What is the expected output? What do you see instead? The expected output: no exception :) See the attachment. Please provide any additional information below. Starting a new game works fine but loading the saved campaign fails. The campaign was played a fair amount of time (Garthogs conquered, met the Dargslans). Is there an easy way to revert the upgrade? Please upload any save before and/or after the problem happened in zip compressed format. Please attach the open-ig.log file found in the application's directory. ``` Original issue reported on code.google.com by `pistu...@gmail.com` on 2 Jan 2013 at 9:07 Attachments: * [openig-crash.zip](https://storage.googleapis.com/google-code-attachments/open-ig/issue-709/comment-0/openig-crash.zip)
defect
npe when attempting to load saved game after update game version operating system windows xp bit java runtime version installed using the launcher yes game language en hu de hu what steps will reproduce the problem start and play a campaign using version update to using the update button try to resume your campaign what is the expected output what do you see instead the expected output no exception see the attachment please provide any additional information below starting a new game works fine but loading the saved campaign fails the campaign was played a fair amount of time garthogs conquered met the dargslans is there an easy way to revert the upgrade please upload any save before and or after the problem happened in zip compressed format please attach the open ig log file found in the application s directory original issue reported on code google com by pistu gmail com on jan at attachments
1
70,354
9,411,902,055
IssuesEvent
2019-04-10 01:37:03
kubernetes-sigs/federation-v2
https://api.github.com/repos/kubernetes-sigs/federation-v2
closed
Objects stuck with finalizer "delete-from-underlying-clusters"
kind/bug kind/documentation
<!-- Please use this template while reporting a bug and provide as much info as possible. Not doing so may result in your bug not being addressed in a timely manner. Thanks!--> **What happened**: Federated environment with 5 clusters, 2 of them are down. We tried to remove a namespace that was federated across the 5 clusters, objects got stuck by finalizer `federation.kubernetes.io/delete-from-underlying-clusters`. We couldn't remove non-available clusters using kubefed2 tool (before deleting the namespace). When removing the finalizer manually, then the objects got stuck by finalizer `kubernetes` because the CRD instances in the project weren't deleted. **What you expected to happen**: Having some sort of configuration on the control plane where you can define a grace period for non-available clusters, after this grace period, those clusters wouldn't be taken into account for the finalizer `federation.kubernetes.io/delete-from-underlying-clusters`. **How to reproduce it (as minimally and precisely as possible)**: 1. Federate 2 clusters. 2. Deploy a federated workload. 3. Powerdown one of the clusters. 4. Try to delete the federated namespace from the host cluster. **Anything else we need to know?**: **Environment**: - Kubernetes version (use `kubectl version`) ``` openshift v3.11.43 kubernetes v1.11.0+d4cacc0 ``` - Federation version ``` v0.0.4 ``` - Install method (helm vs script) ``` script ``` - Install configuration (namespaced vs cluster-scoped) ``` cluster-scoped ``` - Others <!-- DO NOT EDIT BELOW THIS LINE --> /kind bug
1.0
Objects stuck with finalizer "delete-from-underlying-clusters" - <!-- Please use this template while reporting a bug and provide as much info as possible. Not doing so may result in your bug not being addressed in a timely manner. Thanks!--> **What happened**: Federated environment with 5 clusters, 2 of them are down. We tried to remove a namespace that was federated across the 5 clusters, objects got stuck by finalizer `federation.kubernetes.io/delete-from-underlying-clusters`. We couldn't remove non-available clusters using kubefed2 tool (before deleting the namespace). When removing the finalizer manually, then the objects got stuck by finalizer `kubernetes` because the CRD instances in the project weren't deleted. **What you expected to happen**: Having some sort of configuration on the control plane where you can define a grace period for non-available clusters, after this grace period, those clusters wouldn't be taken into account for the finalizer `federation.kubernetes.io/delete-from-underlying-clusters`. **How to reproduce it (as minimally and precisely as possible)**: 1. Federate 2 clusters. 2. Deploy a federated workload. 3. Powerdown one of the clusters. 4. Try to delete the federated namespace from the host cluster. **Anything else we need to know?**: **Environment**: - Kubernetes version (use `kubectl version`) ``` openshift v3.11.43 kubernetes v1.11.0+d4cacc0 ``` - Federation version ``` v0.0.4 ``` - Install method (helm vs script) ``` script ``` - Install configuration (namespaced vs cluster-scoped) ``` cluster-scoped ``` - Others <!-- DO NOT EDIT BELOW THIS LINE --> /kind bug
non_defect
objects stuck with finalizer delete from underlying clusters what happened federated environment with clusters of them are down we tried to remove a namespace that was federated across the clusters objects got stuck by finalizer federation kubernetes io delete from underlying clusters we couldn t remove non available clusters using tool before deleting the namespace when removing the finalizer manually then the objects got stuck by finalizer kubernetes because the crd instances in the project weren t deleted what you expected to happen having some sort of configuration on the control plane where you can define a grace period for non available clusters after this grace period those clusters wouldn t be taken into account for the finalizer federation kubernetes io delete from underlying clusters how to reproduce it as minimally and precisely as possible federate clusters deploy a federated workload powerdown one of the clusters try to delete the federated namespace from the host cluster anything else we need to know environment kubernetes version use kubectl version openshift kubernetes federation version install method helm vs script script install configuration namespaced vs cluster scoped cluster scoped others kind bug
0
60,366
17,023,407,076
IssuesEvent
2021-07-03 01:52:11
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
Potlatch again didn't save
Component: potlatch (flash editor) Priority: major Resolution: duplicate Type: defect
**[Submitted to the original trac issue database at 9.34pm, Saturday, 23rd May 2009]** Working offline with the new Potlatch seemed to work saver than before. Now I have to discover it actually isnt. Today I entered 35 tracks converted and unlocked them connected them to other streets and defined them. Several times I saved my work. One time it didnt finish saving. I aborted saving and repeated it. This save then was prompted correctly within seconds. The last save before finishing also seemed to be correct. Nothing is correct. Great part of my work is corrupted ! Instead of finished streets I have a lot of green points. I dont know where they come from. Obviously this terrible tool handles aborted saves like correct saves and doesnt care about the actually unsaved objects when repeating saving.
1.0
Potlatch again didn't save - **[Submitted to the original trac issue database at 9.34pm, Saturday, 23rd May 2009]** Working offline with the new Potlatch seemed to work saver than before. Now I have to discover it actually isnt. Today I entered 35 tracks converted and unlocked them connected them to other streets and defined them. Several times I saved my work. One time it didnt finish saving. I aborted saving and repeated it. This save then was prompted correctly within seconds. The last save before finishing also seemed to be correct. Nothing is correct. Great part of my work is corrupted ! Instead of finished streets I have a lot of green points. I dont know where they come from. Obviously this terrible tool handles aborted saves like correct saves and doesnt care about the actually unsaved objects when repeating saving.
defect
potlatch again didn t save working offline with the new potlatch seemed to work saver than before now i have to discover it actually isnt today i entered tracks converted and unlocked them connected them to other streets and defined them several times i saved my work one time it didnt finish saving i aborted saving and repeated it this save then was prompted correctly within seconds the last save before finishing also seemed to be correct nothing is correct great part of my work is corrupted instead of finished streets i have a lot of green points i dont know where they come from obviously this terrible tool handles aborted saves like correct saves and doesnt care about the actually unsaved objects when repeating saving
1
34,504
7,452,466,535
IssuesEvent
2018-03-29 08:30:00
kerdokullamae/test_koik_issued
https://api.github.com/repos/kerdokullamae/test_koik_issued
closed
Leidandmed tasanditel allarhiiv,sari,allsari,arhivaal
P: normal R: duplicate T: defect
**Reported by katrin vesterblom on 7 Jun 2013 09:52 UTC** rahvusarhiiv.tietotest.ee Sarja (ning allsarja ja allarhiivi) tasandil peaks leidandmetena näitama ülemise tasandi (ehk siis arhiivi, kuna see on nende puhul ilmselt ainus ülem, kus leidandmed on) leidandmeid. Praegu ei näita. Proovipäringu tegin arhiiviüksus=TLA, siis on parasjagu väike loetelu, kus aga on kohe võimalik tuvastada, et sarjal ei näidata leidandmeid. Arhivaali tasandil tundub et näitab ainult seda osa, mis on konkreetselt arhivaali tasandil (juurde) sisestatud. Aga peaks näitama tervikut-säiliku nr+arhivaali tähis. http://rahvusarhiiv.tietotest.ee/et/description_unit/view/?id=29397 peaks leidandmed olema TLA.9999.21-C.2.200.10 , aga on lihtsalt 200.10. Mõlema puuduse kohta peaks olema ka kunagi enne tehtud ticketid.
1.0
Leidandmed tasanditel allarhiiv,sari,allsari,arhivaal - **Reported by katrin vesterblom on 7 Jun 2013 09:52 UTC** rahvusarhiiv.tietotest.ee Sarja (ning allsarja ja allarhiivi) tasandil peaks leidandmetena näitama ülemise tasandi (ehk siis arhiivi, kuna see on nende puhul ilmselt ainus ülem, kus leidandmed on) leidandmeid. Praegu ei näita. Proovipäringu tegin arhiiviüksus=TLA, siis on parasjagu väike loetelu, kus aga on kohe võimalik tuvastada, et sarjal ei näidata leidandmeid. Arhivaali tasandil tundub et näitab ainult seda osa, mis on konkreetselt arhivaali tasandil (juurde) sisestatud. Aga peaks näitama tervikut-säiliku nr+arhivaali tähis. http://rahvusarhiiv.tietotest.ee/et/description_unit/view/?id=29397 peaks leidandmed olema TLA.9999.21-C.2.200.10 , aga on lihtsalt 200.10. Mõlema puuduse kohta peaks olema ka kunagi enne tehtud ticketid.
defect
leidandmed tasanditel allarhiiv sari allsari arhivaal reported by katrin vesterblom on jun utc rahvusarhiiv tietotest ee sarja ning allsarja ja allarhiivi tasandil peaks leidandmetena näitama ülemise tasandi ehk siis arhiivi kuna see on nende puhul ilmselt ainus ülem kus leidandmed on leidandmeid praegu ei näita proovipäringu tegin arhiiviüksus tla siis on parasjagu väike loetelu kus aga on kohe võimalik tuvastada et sarjal ei näidata leidandmeid arhivaali tasandil tundub et näitab ainult seda osa mis on konkreetselt arhivaali tasandil juurde sisestatud aga peaks näitama tervikut säiliku nr arhivaali tähis peaks leidandmed olema tla c aga on lihtsalt mõlema puuduse kohta peaks olema ka kunagi enne tehtud ticketid
1
8,180
2,611,469,688
IssuesEvent
2015-02-27 05:14:54
chrsmith/hedgewars
https://api.github.com/repos/chrsmith/hedgewars
closed
Hedgewars closes when i press online
auto-migrated OpSys-OSX Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. Open Hedgewars 2. press online 3. it will close What is the expected output? What do you see instead? it closes and says send to apple What version of the product are you using? On what operating system? newest version mac os x 10.6.6 Please provide any additional information below. ``` Original issue reported on code.google.com by `downdann...@gmail.com` on 2 Feb 2011 at 3:27 * Merged into: #180
1.0
Hedgewars closes when i press online - ``` What steps will reproduce the problem? 1. Open Hedgewars 2. press online 3. it will close What is the expected output? What do you see instead? it closes and says send to apple What version of the product are you using? On what operating system? newest version mac os x 10.6.6 Please provide any additional information below. ``` Original issue reported on code.google.com by `downdann...@gmail.com` on 2 Feb 2011 at 3:27 * Merged into: #180
defect
hedgewars closes when i press online what steps will reproduce the problem open hedgewars press online it will close what is the expected output what do you see instead it closes and says send to apple what version of the product are you using on what operating system newest version mac os x please provide any additional information below original issue reported on code google com by downdann gmail com on feb at merged into
1
452,030
13,045,100,779
IssuesEvent
2020-07-29 06:30:51
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.bose.com - see bug description
browser-firefox-mobile engine-gecko priority-normal
<!-- @browser: Firefox Mobile 79.0 --> <!-- @ua_header: Mozilla/5.0 (Android 11; Mobile; rv:79.0) Gecko/79.0 Firefox/79.0 --> <!-- @reported_with: desktop-reporter --> **URL**: https://www.bose.com/en_us/products/frames/bose-frames-alto.html **Browser / Version**: Firefox Mobile 79.0 **Operating System**: Android **Tested Another Browser**: Yes Chrome **Problem type**: Something else **Description**: crashes Firefox **Steps to Reproduce**: Site starts loading and then Firefox crashes (before any pictures load) <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200713203149</li><li>channel: beta</li><li>hasTouchScreen: true</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.bose.com - see bug description - <!-- @browser: Firefox Mobile 79.0 --> <!-- @ua_header: Mozilla/5.0 (Android 11; Mobile; rv:79.0) Gecko/79.0 Firefox/79.0 --> <!-- @reported_with: desktop-reporter --> **URL**: https://www.bose.com/en_us/products/frames/bose-frames-alto.html **Browser / Version**: Firefox Mobile 79.0 **Operating System**: Android **Tested Another Browser**: Yes Chrome **Problem type**: Something else **Description**: crashes Firefox **Steps to Reproduce**: Site starts loading and then Firefox crashes (before any pictures load) <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200713203149</li><li>channel: beta</li><li>hasTouchScreen: true</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_defect
see bug description url browser version firefox mobile operating system android tested another browser yes chrome problem type something else description crashes firefox steps to reproduce site starts loading and then firefox crashes before any pictures load browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel beta hastouchscreen true from with ❤️
0
17,826
3,013,058,714
IssuesEvent
2015-07-29 05:53:33
yawlfoundation/yawl
https://api.github.com/repos/yawlfoundation/yawl
closed
Editor issues removing task decomposition variables in offline mode
auto-migrated Priority-Low Type-Defect
``` If the Editor is in "offline" mode (no Engine and Resource Service connections), strange behaviour occurs when trying to remove task decomposition variables. The following are observed use cases (just open the Editor without the YAWL backend having been started, create a new specification, add a new atomic task and create a task decomposition for it): 1. -- Add any two variables -- Remove the first; both get removed instead of only the first This still occurs if you add the two, click Done, go back into the task decomposition and then try to remove the first 2. -- Add any two variables -- Save the spec. and close it -- Reopen the spec, edit the task decomposition and remove the first -- Display doesn't update and can't click on the variable lines properly, but clicking Done and going back in shows that the removal was actually done properly In response to previous bugs I logged, Michael changed the Editor for 2.0.1 to check for Engine/RS connections in certain situations. I suspect that this code has introduced these problems when in offline mode (might be wrong of course! May still have existed in v2.0). Putting as priority Low since appears not to occur (at least not consistently) when the Editor has Engine and RS connections. What steps will reproduce the problem? As above What is the expected output? What do you see instead? As above ``` Original issue reported on code.google.com by `monsieur...@gmail.com` on 31 Mar 2010 at 11:12
1.0
Editor issues removing task decomposition variables in offline mode - ``` If the Editor is in "offline" mode (no Engine and Resource Service connections), strange behaviour occurs when trying to remove task decomposition variables. The following are observed use cases (just open the Editor without the YAWL backend having been started, create a new specification, add a new atomic task and create a task decomposition for it): 1. -- Add any two variables -- Remove the first; both get removed instead of only the first This still occurs if you add the two, click Done, go back into the task decomposition and then try to remove the first 2. -- Add any two variables -- Save the spec. and close it -- Reopen the spec, edit the task decomposition and remove the first -- Display doesn't update and can't click on the variable lines properly, but clicking Done and going back in shows that the removal was actually done properly In response to previous bugs I logged, Michael changed the Editor for 2.0.1 to check for Engine/RS connections in certain situations. I suspect that this code has introduced these problems when in offline mode (might be wrong of course! May still have existed in v2.0). Putting as priority Low since appears not to occur (at least not consistently) when the Editor has Engine and RS connections. What steps will reproduce the problem? As above What is the expected output? What do you see instead? As above ``` Original issue reported on code.google.com by `monsieur...@gmail.com` on 31 Mar 2010 at 11:12
defect
editor issues removing task decomposition variables in offline mode if the editor is in offline mode no engine and resource service connections strange behaviour occurs when trying to remove task decomposition variables the following are observed use cases just open the editor without the yawl backend having been started create a new specification add a new atomic task and create a task decomposition for it add any two variables remove the first both get removed instead of only the first this still occurs if you add the two click done go back into the task decomposition and then try to remove the first add any two variables save the spec and close it reopen the spec edit the task decomposition and remove the first display doesn t update and can t click on the variable lines properly but clicking done and going back in shows that the removal was actually done properly in response to previous bugs i logged michael changed the editor for to check for engine rs connections in certain situations i suspect that this code has introduced these problems when in offline mode might be wrong of course may still have existed in putting as priority low since appears not to occur at least not consistently when the editor has engine and rs connections what steps will reproduce the problem as above what is the expected output what do you see instead as above original issue reported on code google com by monsieur gmail com on mar at
1
40,517
10,028,056,881
IssuesEvent
2019-07-17 10:40:24
MDAnalysis/mdanalysis
https://api.github.com/repos/MDAnalysis/mdanalysis
closed
MDAnalysis.transformations not automatically imported
Component-Docs Component-Transformations Difficulty-easy defect usability
The documentation on the transformations is not clear: I tried all of these: ```python MDAnalysis.transformations.center(ag,center='mass')(ts) MDAnalysis.lib.transformations.center(ag,center='mass')(ts) MDAnalysis.lib.transformations.center_in_box(ag,center='mass')(ts) MDAnalysis.transformations.center_in_box(ag,center='mass')(ts) ``` and they all throw: ``` AttributeError: module 'MDAnalysis' has no attribute ....? ``` and then got crazy and tried ```python MDAnalysis.translate.center_in_box(ag,center='mass')(ts) MDAnalysis.translate.center(ag,center='mass')(ts) ```
1.0
MDAnalysis.transformations not automatically imported - The documentation on the transformations is not clear: I tried all of these: ```python MDAnalysis.transformations.center(ag,center='mass')(ts) MDAnalysis.lib.transformations.center(ag,center='mass')(ts) MDAnalysis.lib.transformations.center_in_box(ag,center='mass')(ts) MDAnalysis.transformations.center_in_box(ag,center='mass')(ts) ``` and they all throw: ``` AttributeError: module 'MDAnalysis' has no attribute ....? ``` and then got crazy and tried ```python MDAnalysis.translate.center_in_box(ag,center='mass')(ts) MDAnalysis.translate.center(ag,center='mass')(ts) ```
defect
mdanalysis transformations not automatically imported the documentation on the transformations is not clear i tried all of these python mdanalysis transformations center ag center mass ts mdanalysis lib transformations center ag center mass ts mdanalysis lib transformations center in box ag center mass ts mdanalysis transformations center in box ag center mass ts and they all throw attributeerror module mdanalysis has no attribute and then got crazy and tried python mdanalysis translate center in box ag center mass ts mdanalysis translate center ag center mass ts
1
277,709
24,097,557,446
IssuesEvent
2022-09-19 20:15:02
microsoft/vscode
https://api.github.com/repos/microsoft/vscode
closed
Testing: test hierarchy is broken with C++ TestMate extension
bug author-verification-requested testing
<!-- 🪓 If you answered No above, use 'Help: Start Extension Bisect' from Command Palette to try to identify the cause. --> <!-- 📣 Issues caused by an extension need to be reported directly to the extension publisher. The 'Help > Report Issue' dialog can assist with this. --> - VS Code Version: 1.72.0-insider (0f2de31e485688) With the C++ TestMate extension, the hierarchy of test executables and contained tests is no longer displayed correctly. This is a regression that was introduced with c11dabf9ce669. Here's an example. Create a cpp file with the following content: ```c++ #define CATCH_CONFIG_MAIN #include "catch.hpp" TEST_CASE("T1") {} TEST_CASE("T2") {} TEST_CASE("T3") {} ``` and compile it using `catch.hpp` taken from [here](https://github.com/catchorg/Catch2/blob/v2.x/single_include/catch2/catch.hpp). Then, set the `testMate.cpp.test.executables` config to point to the compiled executable. With VS Code v1.71.0 you get the expected browser content: <img width="581" alt="image" src="https://user-images.githubusercontent.com/1225667/188911598-281f62c2-6e77-405b-94ca-9195b8234b69.png"> With the latest Insiders build, or any build from c11dabf9ce669 on, you get this instead: <img width="581" alt="image" src="https://user-images.githubusercontent.com/1225667/188912204-e1b6af79-8f55-4b04-9a88-20351c3a7e61.png"> Note how T1 is now a sibling of the executable, not a child. Also, the executable no longer has a play button. @connor4312 Cc: to @matepek since I'm not entirely sure if this is VS Code's fault, or if there's maybe something wrong with how the extension sets the URIs.
1.0
Testing: test hierarchy is broken with C++ TestMate extension - <!-- 🪓 If you answered No above, use 'Help: Start Extension Bisect' from Command Palette to try to identify the cause. --> <!-- 📣 Issues caused by an extension need to be reported directly to the extension publisher. The 'Help > Report Issue' dialog can assist with this. --> - VS Code Version: 1.72.0-insider (0f2de31e485688) With the C++ TestMate extension, the hierarchy of test executables and contained tests is no longer displayed correctly. This is a regression that was introduced with c11dabf9ce669. Here's an example. Create a cpp file with the following content: ```c++ #define CATCH_CONFIG_MAIN #include "catch.hpp" TEST_CASE("T1") {} TEST_CASE("T2") {} TEST_CASE("T3") {} ``` and compile it using `catch.hpp` taken from [here](https://github.com/catchorg/Catch2/blob/v2.x/single_include/catch2/catch.hpp). Then, set the `testMate.cpp.test.executables` config to point to the compiled executable. With VS Code v1.71.0 you get the expected browser content: <img width="581" alt="image" src="https://user-images.githubusercontent.com/1225667/188911598-281f62c2-6e77-405b-94ca-9195b8234b69.png"> With the latest Insiders build, or any build from c11dabf9ce669 on, you get this instead: <img width="581" alt="image" src="https://user-images.githubusercontent.com/1225667/188912204-e1b6af79-8f55-4b04-9a88-20351c3a7e61.png"> Note how T1 is now a sibling of the executable, not a child. Also, the executable no longer has a play button. @connor4312 Cc: to @matepek since I'm not entirely sure if this is VS Code's fault, or if there's maybe something wrong with how the extension sets the URIs.
non_defect
testing test hierarchy is broken with c testmate extension report issue dialog can assist with this vs code version insider with the c testmate extension the hierarchy of test executables and contained tests is no longer displayed correctly this is a regression that was introduced with here s an example create a cpp file with the following content c define catch config main include catch hpp test case test case test case and compile it using catch hpp taken from then set the testmate cpp test executables config to point to the compiled executable with vs code you get the expected browser content img width alt image src with the latest insiders build or any build from on you get this instead img width alt image src note how is now a sibling of the executable not a child also the executable no longer has a play button cc to matepek since i m not entirely sure if this is vs code s fault or if there s maybe something wrong with how the extension sets the uris
0
103,166
16,602,008,598
IssuesEvent
2021-06-01 20:54:46
samq-ghdemo/SEARCH-NCJIS-nibrs
https://api.github.com/repos/samq-ghdemo/SEARCH-NCJIS-nibrs
opened
CVE-2020-11620 (High) detected in multiple libraries
security vulnerability
## CVE-2020-11620 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-databind-2.9.8.jar</b>, <b>jackson-databind-2.8.10.jar</b>, <b>jackson-databind-2.9.5.jar</b>, <b>jackson-databind-2.9.6.jar</b></p></summary> <p> <details><summary><b>jackson-databind-2.9.8.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: SEARCH-NCJIS-nibrs/tools/nibrs-summary-report-common/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.1.5.RELEASE.jar (Root Library) - spring-boot-starter-json-2.1.5.RELEASE.jar - :x: **jackson-databind-2.9.8.jar** (Vulnerable Library) </details> <details><summary><b>jackson-databind-2.8.10.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: SEARCH-NCJIS-nibrs/tools/nibrs-fbi-service/pom.xml</p> <p>Path to vulnerable library: SEARCH-NCJIS-nibrs/tools/nibrs-fbi-service/target/nibrs-fbi-service-1.0.0/WEB-INF/lib/jackson-databind-2.8.10.jar,canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.10/jackson-databind-2.8.10.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.8.10.jar** (Vulnerable Library) </details> <details><summary><b>jackson-databind-2.9.5.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: SEARCH-NCJIS-nibrs/tools/nibrs-flatfile/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar</p> <p> Dependency Hierarchy: - tika-parsers-1.18.jar (Root Library) - :x: **jackson-databind-2.9.5.jar** (Vulnerable Library) </details> <details><summary><b>jackson-databind-2.9.6.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: SEARCH-NCJIS-nibrs/tools/nibrs-validation/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,SEARCH-NCJIS-nibrs/web/nibrs-web/target/nibrs-web/WEB-INF/lib/jackson-databind-2.9.6.jar,canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.9.6.jar** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/samq-ghdemo/SEARCH-NCJIS-nibrs/commit/2643373aa9a184ff4ea81e98caf4009bf2ee8e91">2643373aa9a184ff4ea81e98caf4009bf2ee8e91</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to org.apache.commons.jelly.impl.Embedded (aka commons-jelly). <p>Publish Date: 2020-04-07 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11620>CVE-2020-11620</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11620">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11620</a></p> <p>Release Date: 2020-04-07</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.4</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.8","packageFilePaths":["/tools/nibrs-summary-report-common/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:2.1.5.RELEASE;org.springframework.boot:spring-boot-starter-json:2.1.5.RELEASE;com.fasterxml.jackson.core:jackson-databind:2.9.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.4"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.10","packageFilePaths":["/tools/nibrs-fbi-service/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.8.10","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.4"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.5","packageFilePaths":["/tools/nibrs-flatfile/pom.xml","/tools/nibrs-validate-common/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.apache.tika:tika-parsers:1.18;com.fasterxml.jackson.core:jackson-databind:2.9.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.4"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.6","packageFilePaths":["/tools/nibrs-validation/pom.xml","/tools/nibrs-staging-data/pom.xml","/tools/nibrs-staging-data-common/pom.xml","/tools/nibrs-xmlfile/pom.xml","/tools/nibrs-route/pom.xml","/tools/nibrs-summary-report/pom.xml","/web/nibrs-web/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.4"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-11620","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to org.apache.commons.jelly.impl.Embedded (aka commons-jelly).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11620","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-11620 (High) detected in multiple libraries - ## CVE-2020-11620 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-databind-2.9.8.jar</b>, <b>jackson-databind-2.8.10.jar</b>, <b>jackson-databind-2.9.5.jar</b>, <b>jackson-databind-2.9.6.jar</b></p></summary> <p> <details><summary><b>jackson-databind-2.9.8.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: SEARCH-NCJIS-nibrs/tools/nibrs-summary-report-common/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.1.5.RELEASE.jar (Root Library) - spring-boot-starter-json-2.1.5.RELEASE.jar - :x: **jackson-databind-2.9.8.jar** (Vulnerable Library) </details> <details><summary><b>jackson-databind-2.8.10.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: SEARCH-NCJIS-nibrs/tools/nibrs-fbi-service/pom.xml</p> <p>Path to vulnerable library: SEARCH-NCJIS-nibrs/tools/nibrs-fbi-service/target/nibrs-fbi-service-1.0.0/WEB-INF/lib/jackson-databind-2.8.10.jar,canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.10/jackson-databind-2.8.10.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.8.10.jar** (Vulnerable Library) </details> <details><summary><b>jackson-databind-2.9.5.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: SEARCH-NCJIS-nibrs/tools/nibrs-flatfile/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar</p> <p> Dependency Hierarchy: - tika-parsers-1.18.jar (Root Library) - :x: **jackson-databind-2.9.5.jar** (Vulnerable Library) </details> <details><summary><b>jackson-databind-2.9.6.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: SEARCH-NCJIS-nibrs/tools/nibrs-validation/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,SEARCH-NCJIS-nibrs/web/nibrs-web/target/nibrs-web/WEB-INF/lib/jackson-databind-2.9.6.jar,canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.9.6.jar** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/samq-ghdemo/SEARCH-NCJIS-nibrs/commit/2643373aa9a184ff4ea81e98caf4009bf2ee8e91">2643373aa9a184ff4ea81e98caf4009bf2ee8e91</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to org.apache.commons.jelly.impl.Embedded (aka commons-jelly). <p>Publish Date: 2020-04-07 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11620>CVE-2020-11620</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11620">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11620</a></p> <p>Release Date: 2020-04-07</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.4</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.8","packageFilePaths":["/tools/nibrs-summary-report-common/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:2.1.5.RELEASE;org.springframework.boot:spring-boot-starter-json:2.1.5.RELEASE;com.fasterxml.jackson.core:jackson-databind:2.9.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.4"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.10","packageFilePaths":["/tools/nibrs-fbi-service/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.8.10","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.4"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.5","packageFilePaths":["/tools/nibrs-flatfile/pom.xml","/tools/nibrs-validate-common/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.apache.tika:tika-parsers:1.18;com.fasterxml.jackson.core:jackson-databind:2.9.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.4"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.6","packageFilePaths":["/tools/nibrs-validation/pom.xml","/tools/nibrs-staging-data/pom.xml","/tools/nibrs-staging-data-common/pom.xml","/tools/nibrs-xmlfile/pom.xml","/tools/nibrs-route/pom.xml","/tools/nibrs-summary-report/pom.xml","/web/nibrs-web/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.4"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-11620","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to org.apache.commons.jelly.impl.Embedded (aka commons-jelly).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11620","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_defect
cve high detected in multiple libraries cve high severity vulnerability vulnerable libraries jackson databind jar jackson databind jar jackson databind jar jackson databind jar jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file search ncjis nibrs tools nibrs summary report common pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring boot starter web release jar root library spring boot starter json release jar x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file search ncjis nibrs tools nibrs fbi service pom xml path to vulnerable library search ncjis nibrs tools nibrs fbi service target nibrs fbi service web inf lib jackson databind jar canner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file search ncjis nibrs tools nibrs flatfile pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy tika parsers jar root library x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file search ncjis nibrs tools nibrs validation pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar search ncjis nibrs web nibrs web target nibrs web web inf lib jackson databind jar canner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache commons jelly impl embedded aka commons jelly publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org springframework boot spring boot starter web release org springframework boot spring boot starter json release com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency false dependencytree com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency true dependencytree org apache tika tika parsers com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency false dependencytree com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind basebranches vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache commons jelly impl embedded aka commons jelly vulnerabilityurl
0
49,034
13,185,202,237
IssuesEvent
2020-08-12 20:55:41
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
opened
fix installation issues of cmake manpages (Trac #635)
Incomplete Migration Migrated from Trac defect tools/ports
<details> <summary><em>Migrated from https://code.icecube.wisc.edu/ticket/635 , reported by nega and owned by nega</em></summary> <p> ```json { "status": "closed", "changetime": "2014-11-23T03:37:57", "description": "port complains of permission issues when installing the cmake man pages.", "reporter": "nega", "cc": "", "resolution": "wontfix", "_ts": "1416713877165085", "component": "tools/ports", "summary": "fix installation issues of cmake manpages", "priority": "normal", "keywords": "cmake man docs", "time": "2011-05-17T15:18:36", "milestone": "", "owner": "nega", "type": "defect" } ``` </p> </details>
1.0
fix installation issues of cmake manpages (Trac #635) - <details> <summary><em>Migrated from https://code.icecube.wisc.edu/ticket/635 , reported by nega and owned by nega</em></summary> <p> ```json { "status": "closed", "changetime": "2014-11-23T03:37:57", "description": "port complains of permission issues when installing the cmake man pages.", "reporter": "nega", "cc": "", "resolution": "wontfix", "_ts": "1416713877165085", "component": "tools/ports", "summary": "fix installation issues of cmake manpages", "priority": "normal", "keywords": "cmake man docs", "time": "2011-05-17T15:18:36", "milestone": "", "owner": "nega", "type": "defect" } ``` </p> </details>
defect
fix installation issues of cmake manpages trac migrated from reported by nega and owned by nega json status closed changetime description port complains of permission issues when installing the cmake man pages reporter nega cc resolution wontfix ts component tools ports summary fix installation issues of cmake manpages priority normal keywords cmake man docs time milestone owner nega type defect
1
55,194
14,265,608,007
IssuesEvent
2020-11-20 17:22:18
idaholab/moose
https://api.github.com/repos/idaholab/moose
opened
NodalKernel computation duplicated on nodes shared between subdomains
T: defect
## Bug Description This is reported at https://github.com/idaholab/moose/discussions/16227#discussioncomment-134829. ## Steps to Reproduce You can just look at the current code: ```c++ const std::set<SubdomainID> & block_ids = _aux_sys.mesh().getNodeBlockIds(*node); for (const auto & block : block_ids) if (_nkernel_warehouse->hasActiveBlockObjects(block, _tid)) { std::set<TagID> needed_fe_var_vector_tags; _nkernel_warehouse->updateBlockFEVariableCoupledVectorTagDependency( block, needed_fe_var_vector_tags, _tid); _fe_problem.setActiveFEVariableCoupleableVectorTags(needed_fe_var_vector_tags, _tid); const auto & objects = _nkernel_warehouse->getActiveBlockObjects(block, _tid); for (const auto & nodal_kernel : objects) nodal_kernel->computeResidual(); } ``` ## Impact This produces fundamentally wrong results. I guess you could work around this by restricting every `NodalKernel` to a single subdomain, but a user will not intuitively think to do that, and will most likely silently get wrong results.
1.0
NodalKernel computation duplicated on nodes shared between subdomains - ## Bug Description This is reported at https://github.com/idaholab/moose/discussions/16227#discussioncomment-134829. ## Steps to Reproduce You can just look at the current code: ```c++ const std::set<SubdomainID> & block_ids = _aux_sys.mesh().getNodeBlockIds(*node); for (const auto & block : block_ids) if (_nkernel_warehouse->hasActiveBlockObjects(block, _tid)) { std::set<TagID> needed_fe_var_vector_tags; _nkernel_warehouse->updateBlockFEVariableCoupledVectorTagDependency( block, needed_fe_var_vector_tags, _tid); _fe_problem.setActiveFEVariableCoupleableVectorTags(needed_fe_var_vector_tags, _tid); const auto & objects = _nkernel_warehouse->getActiveBlockObjects(block, _tid); for (const auto & nodal_kernel : objects) nodal_kernel->computeResidual(); } ``` ## Impact This produces fundamentally wrong results. I guess you could work around this by restricting every `NodalKernel` to a single subdomain, but a user will not intuitively think to do that, and will most likely silently get wrong results.
defect
nodalkernel computation duplicated on nodes shared between subdomains bug description this is reported at steps to reproduce you can just look at the current code c const std set block ids aux sys mesh getnodeblockids node for const auto block block ids if nkernel warehouse hasactiveblockobjects block tid std set needed fe var vector tags nkernel warehouse updateblockfevariablecoupledvectortagdependency block needed fe var vector tags tid fe problem setactivefevariablecoupleablevectortags needed fe var vector tags tid const auto objects nkernel warehouse getactiveblockobjects block tid for const auto nodal kernel objects nodal kernel computeresidual impact this produces fundamentally wrong results i guess you could work around this by restricting every nodalkernel to a single subdomain but a user will not intuitively think to do that and will most likely silently get wrong results
1
17,097
5,330,138,783
IssuesEvent
2017-02-15 16:24:38
eclipse/che
https://api.github.com/repos/eclipse/che
closed
The "Projects" info page does not refresh after deleting a project.
kind/bug severity/P2 status/code-review
**Reproduction Steps:** 1. **"Workspaces"** -> **"Configure workspace"** -> select tab **"Projects"**. 2. Create any project. 3. Select the project and click "Delete" button . 4. Confirm deleting. **Expected behavior:** The project is removed and not visible in the "Projects" tab. **Observed behavior:** The project is removed but the projects list isn't refreshed. The deleted project name is still visible. **Che version:** 5.3.0-SNAPSHOT **OS and version:** Ubuntu 16.04 LTS **Docker version:** 1.12.3 **Che install:** Docker container **Additional information:** Problem can be reliably reproduced, doesn't happen randomly: [Yes] ![project_refresh](https://cloud.githubusercontent.com/assets/7760565/22877426/06165544-f1de-11e6-81b6-19158c2b1e19.gif)
1.0
The "Projects" info page does not refresh after deleting a project. - **Reproduction Steps:** 1. **"Workspaces"** -> **"Configure workspace"** -> select tab **"Projects"**. 2. Create any project. 3. Select the project and click "Delete" button . 4. Confirm deleting. **Expected behavior:** The project is removed and not visible in the "Projects" tab. **Observed behavior:** The project is removed but the projects list isn't refreshed. The deleted project name is still visible. **Che version:** 5.3.0-SNAPSHOT **OS and version:** Ubuntu 16.04 LTS **Docker version:** 1.12.3 **Che install:** Docker container **Additional information:** Problem can be reliably reproduced, doesn't happen randomly: [Yes] ![project_refresh](https://cloud.githubusercontent.com/assets/7760565/22877426/06165544-f1de-11e6-81b6-19158c2b1e19.gif)
non_defect
the projects info page does not refresh after deleting a project reproduction steps workspaces configure workspace select tab projects create any project select the project and click delete button confirm deleting expected behavior the project is removed and not visible in the projects tab observed behavior the project is removed but the projects list isn t refreshed the deleted project name is still visible che version snapshot os and version ubuntu lts docker version che install docker container additional information problem can be reliably reproduced doesn t happen randomly
0
100,652
8,752,544,022
IssuesEvent
2018-12-14 03:40:03
wacossusca34/glava
https://api.github.com/repos/wacossusca34/glava
closed
[Ubuntu 18.04 LTS] `Circle` visualizer displays grey screen
duplicate needs testing
Its basically all in the title. I tried verbose output, no errors there. Any reason why circle doesnt work but every other visualizer works?
1.0
[Ubuntu 18.04 LTS] `Circle` visualizer displays grey screen - Its basically all in the title. I tried verbose output, no errors there. Any reason why circle doesnt work but every other visualizer works?
non_defect
circle visualizer displays grey screen its basically all in the title i tried verbose output no errors there any reason why circle doesnt work but every other visualizer works
0
125,973
12,279,609,626
IssuesEvent
2020-05-08 12:34:27
Jimmi08/e107-free-themes-by-e107sk
https://api.github.com/repos/Jimmi08/e107-free-themes-by-e107sk
opened
How to use inline css code in theme prefs
documentation
Just insert css rule in theme prefs: ![image](https://user-images.githubusercontent.com/5429548/81405973-fe5e2400-9138-11ea-853d-a70db7da76a9.png) ![image](https://user-images.githubusercontent.com/5429548/81405965-faca9d00-9138-11ea-9766-ae9d3a090ea2.png)
1.0
How to use inline css code in theme prefs - Just insert css rule in theme prefs: ![image](https://user-images.githubusercontent.com/5429548/81405973-fe5e2400-9138-11ea-853d-a70db7da76a9.png) ![image](https://user-images.githubusercontent.com/5429548/81405965-faca9d00-9138-11ea-9766-ae9d3a090ea2.png)
non_defect
how to use inline css code in theme prefs just insert css rule in theme prefs
0
79,493
28,311,200,140
IssuesEvent
2023-04-10 15:32:52
dotCMS/core
https://api.github.com/repos/dotCMS/core
reopened
Show archived not working - at least when I archive a blog and check "show archived", nothing shows up.
Type : Defect Team : Scout Triage
### Parent Issue https://github.com/dotCMS/core/issues/24326 ### Problem Statement Show archived not working - at least when I archive a blog and check "show archived", nothing shows up. ### Steps to Reproduce ![image](https://user-images.githubusercontent.com/37185433/229943007-d242ffbb-4bcc-4b07-b923-4e763f8e2760.png) ### Acceptance Criteria 1) archive a page 2) click on "Archive" checkbox and you shoul see it Show archived should display archived pages
1.0
Show archived not working - at least when I archive a blog and check "show archived", nothing shows up. - ### Parent Issue https://github.com/dotCMS/core/issues/24326 ### Problem Statement Show archived not working - at least when I archive a blog and check "show archived", nothing shows up. ### Steps to Reproduce ![image](https://user-images.githubusercontent.com/37185433/229943007-d242ffbb-4bcc-4b07-b923-4e763f8e2760.png) ### Acceptance Criteria 1) archive a page 2) click on "Archive" checkbox and you shoul see it Show archived should display archived pages
defect
show archived not working at least when i archive a blog and check show archived nothing shows up parent issue problem statement show archived not working at least when i archive a blog and check show archived nothing shows up steps to reproduce acceptance criteria archive a page click on archive checkbox and you shoul see it show archived should display archived pages
1
255,129
27,484,758,212
IssuesEvent
2023-03-04 01:16:02
panasalap/linux-4.1.15
https://api.github.com/repos/panasalap/linux-4.1.15
opened
CVE-2017-10663 (High) detected in linux-yocto-devv4.2.8
security vulnerability
## CVE-2017-10663 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yocto-devv4.2.8</b></p></summary> <p> <p>Linux Embedded Kernel - tracks the next mainline release</p> <p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto-dev>https://git.yoctoproject.org/git/linux-yocto-dev</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The sanity_check_ckpt function in fs/f2fs/super.c in the Linux kernel before 4.12.4 does not validate the blkoff and segno arrays, which allows local users to gain privileges via unspecified vectors. <p>Publish Date: 2017-08-19 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2017-10663>CVE-2017-10663</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2017-10663">https://nvd.nist.gov/vuln/detail/CVE-2017-10663</a></p> <p>Release Date: 2017-08-19</p> <p>Fix Resolution: 4.12.4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2017-10663 (High) detected in linux-yocto-devv4.2.8 - ## CVE-2017-10663 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yocto-devv4.2.8</b></p></summary> <p> <p>Linux Embedded Kernel - tracks the next mainline release</p> <p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto-dev>https://git.yoctoproject.org/git/linux-yocto-dev</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The sanity_check_ckpt function in fs/f2fs/super.c in the Linux kernel before 4.12.4 does not validate the blkoff and segno arrays, which allows local users to gain privileges via unspecified vectors. <p>Publish Date: 2017-08-19 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2017-10663>CVE-2017-10663</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2017-10663">https://nvd.nist.gov/vuln/detail/CVE-2017-10663</a></p> <p>Release Date: 2017-08-19</p> <p>Fix Resolution: 4.12.4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve high detected in linux yocto cve high severity vulnerability vulnerable library linux yocto linux embedded kernel tracks the next mainline release library home page a href found in base branch master vulnerable source files vulnerability details the sanity check ckpt function in fs super c in the linux kernel before does not validate the blkoff and segno arrays which allows local users to gain privileges via unspecified vectors publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
167,495
14,113,447,702
IssuesEvent
2020-11-07 11:18:35
Project-Books/book-project
https://api.github.com/repos/Project-Books/book-project
closed
Make a vision document
documentation
Make a document that explains the vision of the Book Project. It should explain why we exist and what we set out to do.
1.0
Make a vision document - Make a document that explains the vision of the Book Project. It should explain why we exist and what we set out to do.
non_defect
make a vision document make a document that explains the vision of the book project it should explain why we exist and what we set out to do
0
58,349
16,495,285,663
IssuesEvent
2021-05-25 09:42:20
galasa-dev/projectmanagement
https://api.github.com/repos/galasa-dev/projectmanagement
closed
Issue with retrieveJar failing to properly inspect versions
Manager: Artifact defect
The retrieveJar method within BundleResourcesImp.java is currently failing to compare versions and instead treats files as though there is no attached version.
1.0
Issue with retrieveJar failing to properly inspect versions - The retrieveJar method within BundleResourcesImp.java is currently failing to compare versions and instead treats files as though there is no attached version.
defect
issue with retrievejar failing to properly inspect versions the retrievejar method within bundleresourcesimp java is currently failing to compare versions and instead treats files as though there is no attached version
1
100,891
11,206,776,364
IssuesEvent
2020-01-05 23:55:04
RandomFractals/geo-data-viewer
https://api.github.com/repos/RandomFractals/geo-data-viewer
opened
add recommended extensions section to docs
documentation
include Data Preivew for grid/charts view and GistPad for now gists loading for now
1.0
add recommended extensions section to docs - include Data Preivew for grid/charts view and GistPad for now gists loading for now
non_defect
add recommended extensions section to docs include data preivew for grid charts view and gistpad for now gists loading for now
0
47,437
13,056,185,324
IssuesEvent
2020-07-30 03:55:19
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
closed
enable negative times in glshovel (Trac #557)
Migrated from Trac defect glshovel
Migrated from https://code.icecube.wisc.edu/ticket/557 ```json { "status": "closed", "changetime": "2009-09-30T15:00:23", "description": "", "reporter": "troy", "cc": "", "resolution": "duplicate", "_ts": "1254322823000000", "component": "glshovel", "summary": "enable negative times in glshovel", "priority": "normal", "keywords": "", "time": "2009-06-12T13:51:36", "milestone": "", "owner": "troy", "type": "defect" } ```
1.0
enable negative times in glshovel (Trac #557) - Migrated from https://code.icecube.wisc.edu/ticket/557 ```json { "status": "closed", "changetime": "2009-09-30T15:00:23", "description": "", "reporter": "troy", "cc": "", "resolution": "duplicate", "_ts": "1254322823000000", "component": "glshovel", "summary": "enable negative times in glshovel", "priority": "normal", "keywords": "", "time": "2009-06-12T13:51:36", "milestone": "", "owner": "troy", "type": "defect" } ```
defect
enable negative times in glshovel trac migrated from json status closed changetime description reporter troy cc resolution duplicate ts component glshovel summary enable negative times in glshovel priority normal keywords time milestone owner troy type defect
1
41,785
10,606,661,405
IssuesEvent
2019-10-11 00:19:23
idaholab/moose
https://api.github.com/repos/idaholab/moose
opened
Website menu broken in some cases
T: defect
## Bug Description I don't know if it is just me, but I see the following sometimes: <img width="1330" alt="moosedocs" src="https://user-images.githubusercontent.com/7664823/66615311-e280b700-ec17-11e9-985d-955a064f6304.png"> ## Steps to Reproduce I can reproduce this on a Mac (both High Sierra and Mojave) with both Safari and Chrome in the following order: 1) Go to https://mooseframework.inl.gov/python/MooseDocs/specification.html 2) Then hover the mouse over `Documentation` to try and navigate somewhere else ## Impact Minor usability issue that might be affecting others and not just me
1.0
Website menu broken in some cases - ## Bug Description I don't know if it is just me, but I see the following sometimes: <img width="1330" alt="moosedocs" src="https://user-images.githubusercontent.com/7664823/66615311-e280b700-ec17-11e9-985d-955a064f6304.png"> ## Steps to Reproduce I can reproduce this on a Mac (both High Sierra and Mojave) with both Safari and Chrome in the following order: 1) Go to https://mooseframework.inl.gov/python/MooseDocs/specification.html 2) Then hover the mouse over `Documentation` to try and navigate somewhere else ## Impact Minor usability issue that might be affecting others and not just me
defect
website menu broken in some cases bug description i don t know if it is just me but i see the following sometimes img width alt moosedocs src steps to reproduce i can reproduce this on a mac both high sierra and mojave with both safari and chrome in the following order go to then hover the mouse over documentation to try and navigate somewhere else impact minor usability issue that might be affecting others and not just me
1
17,857
10,820,650,368
IssuesEvent
2019-11-08 16:49:40
planningcenter/developers
https://api.github.com/repos/planningcenter/developers
closed
Best way to Get Last Scheduled Date
Services
**Related Applications** Which application(s) is this question related to? Services <!-- Pick one: Check-Ins, Giving, People, Services, Resources, Webhooks --> <!-- PLEASE DO NOT INCLUDE YOUR CLIENT ID, APPLICATION ID, OR SECRET IN THIS ISSUE. WE WILL REVOKE YOUR TOKEN IF YOU DO. --> **Describe the question** We've written an integration between Rock and Planning Center that syncs demographic data. One of our users has asked if we could pull the date that person was last scheduled when we sync their information from Planning Center. I've looked through the Services API, but it's not obvious to me what endpoint I should use for that (schedule vs plan). Could you point me in the right direction for the best way to get this date for a person? Thanks. **What have you tried that worked?** **What have you tried that didn't work?** **Additional context** I did find this issue: https://github.com/planningcenter/developers/issues/114, but it references a People/NNN/Activities endpoint that I can't seem to find in the documentation. ## I have.. - [x] Reviewed the documentation found at https://developer.planning.center/docs - [x] Searched for previous issues answering this question - [x] Removed all private information from this issue (credentials, tokens, emails, phone numbers, etc.) - [x] Reviewed my issue for completeness
1.0
Best way to Get Last Scheduled Date - **Related Applications** Which application(s) is this question related to? Services <!-- Pick one: Check-Ins, Giving, People, Services, Resources, Webhooks --> <!-- PLEASE DO NOT INCLUDE YOUR CLIENT ID, APPLICATION ID, OR SECRET IN THIS ISSUE. WE WILL REVOKE YOUR TOKEN IF YOU DO. --> **Describe the question** We've written an integration between Rock and Planning Center that syncs demographic data. One of our users has asked if we could pull the date that person was last scheduled when we sync their information from Planning Center. I've looked through the Services API, but it's not obvious to me what endpoint I should use for that (schedule vs plan). Could you point me in the right direction for the best way to get this date for a person? Thanks. **What have you tried that worked?** **What have you tried that didn't work?** **Additional context** I did find this issue: https://github.com/planningcenter/developers/issues/114, but it references a People/NNN/Activities endpoint that I can't seem to find in the documentation. ## I have.. - [x] Reviewed the documentation found at https://developer.planning.center/docs - [x] Searched for previous issues answering this question - [x] Removed all private information from this issue (credentials, tokens, emails, phone numbers, etc.) - [x] Reviewed my issue for completeness
non_defect
best way to get last scheduled date related applications which application s is this question related to services describe the question we ve written an integration between rock and planning center that syncs demographic data one of our users has asked if we could pull the date that person was last scheduled when we sync their information from planning center i ve looked through the services api but it s not obvious to me what endpoint i should use for that schedule vs plan could you point me in the right direction for the best way to get this date for a person thanks what have you tried that worked what have you tried that didn t work additional context i did find this issue but it references a people nnn activities endpoint that i can t seem to find in the documentation i have reviewed the documentation found at searched for previous issues answering this question removed all private information from this issue credentials tokens emails phone numbers etc reviewed my issue for completeness
0
44,379
12,123,919,507
IssuesEvent
2020-04-22 13:28:10
ShaikASK/Testing
https://api.github.com/repos/ShaikASK/Testing
closed
Production : Candidate personal information like SSN is exposed on the internet in the pdf file
Candidate Module Defect P1 Production Issue
Issues in detail : "I got alert from an online security protection company that my personal information like SSN is exposed on the internet by AskStaffing.. Anyone on the internet can see my SSN in the pdf file that you kept in public domain at https://docs.ova.work/docs/askstaffing/Echosign/248/1042/1583722906347.pdf Can you immediately please stop putting this on public internet for anyone to see my SSN. Thank you and please help protect my privacy."
1.0
Production : Candidate personal information like SSN is exposed on the internet in the pdf file - Issues in detail : "I got alert from an online security protection company that my personal information like SSN is exposed on the internet by AskStaffing.. Anyone on the internet can see my SSN in the pdf file that you kept in public domain at https://docs.ova.work/docs/askstaffing/Echosign/248/1042/1583722906347.pdf Can you immediately please stop putting this on public internet for anyone to see my SSN. Thank you and please help protect my privacy."
defect
production candidate personal information like ssn is exposed on the internet in the pdf file issues in detail i got alert from an online security protection company that my personal information like ssn is exposed on the internet by askstaffing anyone on the internet can see my ssn in the pdf file that you kept in public domain at can you immediately please stop putting this on public internet for anyone to see my ssn thank you and please help protect my privacy
1
72,491
24,141,685,642
IssuesEvent
2022-09-21 15:16:39
FreeRADIUS/freeradius-server
https://api.github.com/repos/FreeRADIUS/freeradius-server
closed
[defect]: starting freeradius in full debug mode -XXX can cause Segmentaion fault
defect
### What type of defect/bug is this? Crash or memory corruption (segv, abort, etc...) ### How can the issue be reproduced? When in check-eap-tls config file in the section ``` server check-eap-tls { authorize { ``` contains an if-clause is with a length > 1024 chars then freeradius crashes with a Segmentation fault when started with - freeradius -XXX -d /etc/raddb/ The Segfault happens exactly 4 lines after the ~ 2000 char long if-clause is printed out in the terminal - interesting is that the if-clause debug output is shorted to ~1024 chars leading to my assumption that the debug output is copied to a char field which can only contain 1024 chars without boundary check The if clause contains conditional args like `if ( (&TLS-Client-Cert-Common-Name == "TESTING" && &NAS-IP-Address == "10.10.1.2" && &NAS-Port == 11) || (&TLS-Client-Cert-Common-Name == "TESTING" && &NAS-IP-Address == "10.10.2.2" && &NAS-Port == 11) || ........` - Started normally or with only -X there is no segfault as the Debug output does not contain the authorize section - shorting the if clause to <1024 chars freeradius also works with -XXX debug level This happens with freeradius 3.0.17 ### Log output from the FreeRADIUS daemon ```shell freeradius -XXX -d /etc/raddb Tue Aug 2 14:51:35 2022 : Debug: server check-eap-tls { # from file /etc/raddb//sites-enabled/check-eap-tls Tue Aug 2 14:51:35 2022 : Debug: authorize { Tue Aug 2 14:51:35 2022 : Debug: if ((xxxxxxxxxxxxxxxxxxxxxxxroughtly 1024 charsxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx ) { Tue Aug 2 14:51:35 2022 : Debug: update { Tue Aug 2 14:51:35 2022 : Debug: &control:Auth-Type = Accept Tue Aug 2 14:51:35 2022 : Debug: } Tue Aug 2 14:51:35 2022 : Debug: } Segmentation fault ``` ### Relevant log output from client utilities _No response_ ### Backtrace from LLDB or GDB _No response_
1.0
[defect]: starting freeradius in full debug mode -XXX can cause Segmentaion fault - ### What type of defect/bug is this? Crash or memory corruption (segv, abort, etc...) ### How can the issue be reproduced? When in check-eap-tls config file in the section ``` server check-eap-tls { authorize { ``` contains an if-clause is with a length > 1024 chars then freeradius crashes with a Segmentation fault when started with - freeradius -XXX -d /etc/raddb/ The Segfault happens exactly 4 lines after the ~ 2000 char long if-clause is printed out in the terminal - interesting is that the if-clause debug output is shorted to ~1024 chars leading to my assumption that the debug output is copied to a char field which can only contain 1024 chars without boundary check The if clause contains conditional args like `if ( (&TLS-Client-Cert-Common-Name == "TESTING" && &NAS-IP-Address == "10.10.1.2" && &NAS-Port == 11) || (&TLS-Client-Cert-Common-Name == "TESTING" && &NAS-IP-Address == "10.10.2.2" && &NAS-Port == 11) || ........` - Started normally or with only -X there is no segfault as the Debug output does not contain the authorize section - shorting the if clause to <1024 chars freeradius also works with -XXX debug level This happens with freeradius 3.0.17 ### Log output from the FreeRADIUS daemon ```shell freeradius -XXX -d /etc/raddb Tue Aug 2 14:51:35 2022 : Debug: server check-eap-tls { # from file /etc/raddb//sites-enabled/check-eap-tls Tue Aug 2 14:51:35 2022 : Debug: authorize { Tue Aug 2 14:51:35 2022 : Debug: if ((xxxxxxxxxxxxxxxxxxxxxxxroughtly 1024 charsxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx ) { Tue Aug 2 14:51:35 2022 : Debug: update { Tue Aug 2 14:51:35 2022 : Debug: &control:Auth-Type = Accept Tue Aug 2 14:51:35 2022 : Debug: } Tue Aug 2 14:51:35 2022 : Debug: } Segmentation fault ``` ### Relevant log output from client utilities _No response_ ### Backtrace from LLDB or GDB _No response_
defect
starting freeradius in full debug mode xxx can cause segmentaion fault what type of defect bug is this crash or memory corruption segv abort etc how can the issue be reproduced when in check eap tls config file in the section server check eap tls authorize contains an if clause is with a length chars then freeradius crashes with a segmentation fault when started with freeradius xxx d etc raddb the segfault happens exactly lines after the char long if clause is printed out in the terminal interesting is that the if clause debug output is shorted to chars leading to my assumption that the debug output is copied to a char field which can only contain chars without boundary check the if clause contains conditional args like if tls client cert common name testing nas ip address nas port tls client cert common name testing nas ip address nas port started normally or with only x there is no segfault as the debug output does not contain the authorize section shorting the if clause to chars freeradius also works with xxx debug level this happens with freeradius log output from the freeradius daemon shell freeradius xxx d etc raddb tue aug debug server check eap tls from file etc raddb sites enabled check eap tls tue aug debug authorize tue aug debug if xxxxxxxxxxxxxxxxxxxxxxxroughtly charsxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx tue aug debug update tue aug debug control auth type accept tue aug debug tue aug debug segmentation fault relevant log output from client utilities no response backtrace from lldb or gdb no response
1
37,186
8,289,058,786
IssuesEvent
2018-09-19 13:46:34
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
closed
Client is not able to change to another owner member on disconnection
Source: Jet Team: Client Type: Defect
Client sometimes fails to switch owner with this exception: ``` java.util.concurrent.ExecutionException: com.hazelcast.client.AuthenticationException: Server already processed a newer authentication from client with UUID 99e543e6-1157-43f8-aa8c-794c029c8714. Not applying requested ownership change to 5dd380ed-3b5d-4c56-8478-28297e35f23c at com.hazelcast.client.spi.impl.ClientInvocationFuture.resolve(ClientInvocationFuture.java:73) ~[hazelcast-jet-enterprise-0.7-20180913.092154-29.jar:na] at com.hazelcast.spi.impl.AbstractInvocationFuture$1.run(AbstractInvocationFuture.java:250) ~[hazelcast-jet-enterprise-0.7-20180913.092154-29.jar:na] at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) [na:1.8.0_131] at java.util.concurrent.FutureTask.run(FutureTask.java:266) [na:1.8.0_131] at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180) [na:1.8.0_131] at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293) [na:1.8.0_131] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) [na:1.8.0_131] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) [na:1.8.0_131] at java.lang.Thread.run(Thread.java:748) [na:1.8.0_131] at com.hazelcast.util.executor.HazelcastManagedThread.executeRun(HazelcastManagedThread.java:64) [hazelcast-jet-enterprise-0.7-20180913.092154-29.jar:na] at com.hazelcast.util.executor.HazelcastManagedThread.run(HazelcastManagedThread.java:80) [hazelcast-jet-enterprise-0.7-20180913.092154-29.jar:na] Caused by: com.hazelcast.client.AuthenticationException: Server already processed a newer authentication from client with UUID 99e543e6-1157-43f8-aa8c-794c029c8714. Not applying requested ownership change to 5dd380ed-3b5d-4c56-8478-28297e35f23c at com.hazelcast.client.impl.operations.ClientReAuthOperation.doRun(ClientReAuthOperation.java:69) ~[hazelcast-jet-enterprise-0.7-20180913.092154-29.jar:na] at com.hazelcast.client.impl.operations.ClientReAuthOperation.access$000(ClientReAuthOperation.java:38) ~[hazelcast-jet-enterprise-0.7-20180913.092154-29.jar:na] at com.hazelcast.client.impl.operations.ClientReAuthOperation$ClientReauthTask.run(ClientReAuthOperation.java:135) ~[hazelcast-jet-enterprise-0.7-20180913.092154-29.jar:na] at com.hazelcast.util.executor.CachedExecutorServiceDelegate$Worker.run(CachedExecutorServiceDelegate.java:227) ~[hazelcast-jet-enterprise-0.7-20180913.092154-29.jar:na] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) [na:1.8.0_131] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) [na:1.8.0_131] at java.lang.Thread.run(Thread.java:748) [na:1.8.0_131] at com.hazelcast.util.executor.HazelcastManagedThread.executeRun(HazelcastManagedThread.java:64) [hazelcast-jet-enterprise-0.7-20180913.092154-29.jar:na] at com.hazelcast.util.executor.HazelcastManagedThread.run(HazelcastManagedThread.java:80) [hazelcast-jet-enterprise-0.7-20180913.092154-29.jar:na] at ------ submitted from ------.(Unknown Source) ~[na:na] at com.hazelcast.spi.impl.operationservice.impl.InvocationFuture.resolve(InvocationFuture.java:127) ~[hazelcast-jet-enterprise-0.7-20180913.092154-29.jar:na] at com.hazelcast.spi.impl.AbstractInvocationFuture$1.run(AbstractInvocationFuture.java:250) ~[hazelcast-jet-enterprise-0.7-20180913.092154-29.jar:na] ... 5 common frames omitted ``` Test to reproduce, start cluster first and then client: ``` public class ClientTest { @Test public void cluster() throws InterruptedException { Config config = new Config(); config.setLicenseKey("UNLIMITED_LICENSE#99Nodes#VuE0OIH7TbfKwAUNmSj1JlyFkr6a53911000199920009119011112151009"); HazelcastInstance instance = Hazelcast.newHazelcastInstance(config); HazelcastInstance instance2 = Hazelcast.newHazelcastInstance(config); while (true) { Thread.sleep(20000); instance.shutdown(); instance = Hazelcast.newHazelcastInstance(config); } } @Test public void client() throws InterruptedException { ClientConfig clientConfig = new ClientConfig(); clientConfig.getConnectionStrategyConfig().setAsyncStart(true); clientConfig.getConnectionStrategyConfig().setReconnectMode(ClientConnectionStrategyConfig.ReconnectMode.ASYNC); clientConfig.getNetworkConfig().setConnectionAttemptLimit(0); clientConfig.getNetworkConfig().setSmartRouting(true); HazelcastClient.newHazelcastClient(clientConfig); Thread.currentThread().join(); } } ```
1.0
Client is not able to change to another owner member on disconnection - Client sometimes fails to switch owner with this exception: ``` java.util.concurrent.ExecutionException: com.hazelcast.client.AuthenticationException: Server already processed a newer authentication from client with UUID 99e543e6-1157-43f8-aa8c-794c029c8714. Not applying requested ownership change to 5dd380ed-3b5d-4c56-8478-28297e35f23c at com.hazelcast.client.spi.impl.ClientInvocationFuture.resolve(ClientInvocationFuture.java:73) ~[hazelcast-jet-enterprise-0.7-20180913.092154-29.jar:na] at com.hazelcast.spi.impl.AbstractInvocationFuture$1.run(AbstractInvocationFuture.java:250) ~[hazelcast-jet-enterprise-0.7-20180913.092154-29.jar:na] at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) [na:1.8.0_131] at java.util.concurrent.FutureTask.run(FutureTask.java:266) [na:1.8.0_131] at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180) [na:1.8.0_131] at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293) [na:1.8.0_131] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) [na:1.8.0_131] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) [na:1.8.0_131] at java.lang.Thread.run(Thread.java:748) [na:1.8.0_131] at com.hazelcast.util.executor.HazelcastManagedThread.executeRun(HazelcastManagedThread.java:64) [hazelcast-jet-enterprise-0.7-20180913.092154-29.jar:na] at com.hazelcast.util.executor.HazelcastManagedThread.run(HazelcastManagedThread.java:80) [hazelcast-jet-enterprise-0.7-20180913.092154-29.jar:na] Caused by: com.hazelcast.client.AuthenticationException: Server already processed a newer authentication from client with UUID 99e543e6-1157-43f8-aa8c-794c029c8714. Not applying requested ownership change to 5dd380ed-3b5d-4c56-8478-28297e35f23c at com.hazelcast.client.impl.operations.ClientReAuthOperation.doRun(ClientReAuthOperation.java:69) ~[hazelcast-jet-enterprise-0.7-20180913.092154-29.jar:na] at com.hazelcast.client.impl.operations.ClientReAuthOperation.access$000(ClientReAuthOperation.java:38) ~[hazelcast-jet-enterprise-0.7-20180913.092154-29.jar:na] at com.hazelcast.client.impl.operations.ClientReAuthOperation$ClientReauthTask.run(ClientReAuthOperation.java:135) ~[hazelcast-jet-enterprise-0.7-20180913.092154-29.jar:na] at com.hazelcast.util.executor.CachedExecutorServiceDelegate$Worker.run(CachedExecutorServiceDelegate.java:227) ~[hazelcast-jet-enterprise-0.7-20180913.092154-29.jar:na] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) [na:1.8.0_131] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) [na:1.8.0_131] at java.lang.Thread.run(Thread.java:748) [na:1.8.0_131] at com.hazelcast.util.executor.HazelcastManagedThread.executeRun(HazelcastManagedThread.java:64) [hazelcast-jet-enterprise-0.7-20180913.092154-29.jar:na] at com.hazelcast.util.executor.HazelcastManagedThread.run(HazelcastManagedThread.java:80) [hazelcast-jet-enterprise-0.7-20180913.092154-29.jar:na] at ------ submitted from ------.(Unknown Source) ~[na:na] at com.hazelcast.spi.impl.operationservice.impl.InvocationFuture.resolve(InvocationFuture.java:127) ~[hazelcast-jet-enterprise-0.7-20180913.092154-29.jar:na] at com.hazelcast.spi.impl.AbstractInvocationFuture$1.run(AbstractInvocationFuture.java:250) ~[hazelcast-jet-enterprise-0.7-20180913.092154-29.jar:na] ... 5 common frames omitted ``` Test to reproduce, start cluster first and then client: ``` public class ClientTest { @Test public void cluster() throws InterruptedException { Config config = new Config(); config.setLicenseKey("UNLIMITED_LICENSE#99Nodes#VuE0OIH7TbfKwAUNmSj1JlyFkr6a53911000199920009119011112151009"); HazelcastInstance instance = Hazelcast.newHazelcastInstance(config); HazelcastInstance instance2 = Hazelcast.newHazelcastInstance(config); while (true) { Thread.sleep(20000); instance.shutdown(); instance = Hazelcast.newHazelcastInstance(config); } } @Test public void client() throws InterruptedException { ClientConfig clientConfig = new ClientConfig(); clientConfig.getConnectionStrategyConfig().setAsyncStart(true); clientConfig.getConnectionStrategyConfig().setReconnectMode(ClientConnectionStrategyConfig.ReconnectMode.ASYNC); clientConfig.getNetworkConfig().setConnectionAttemptLimit(0); clientConfig.getNetworkConfig().setSmartRouting(true); HazelcastClient.newHazelcastClient(clientConfig); Thread.currentThread().join(); } } ```
defect
client is not able to change to another owner member on disconnection client sometimes fails to switch owner with this exception java util concurrent executionexception com hazelcast client authenticationexception server already processed a newer authentication from client with uuid not applying requested ownership change to at com hazelcast client spi impl clientinvocationfuture resolve clientinvocationfuture java at com hazelcast spi impl abstractinvocationfuture run abstractinvocationfuture java at java util concurrent executors runnableadapter call executors java at java util concurrent futuretask run futuretask java at java util concurrent scheduledthreadpoolexecutor scheduledfuturetask access scheduledthreadpoolexecutor java at java util concurrent scheduledthreadpoolexecutor scheduledfuturetask run scheduledthreadpoolexecutor java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java at com hazelcast util executor hazelcastmanagedthread executerun hazelcastmanagedthread java at com hazelcast util executor hazelcastmanagedthread run hazelcastmanagedthread java caused by com hazelcast client authenticationexception server already processed a newer authentication from client with uuid not applying requested ownership change to at com hazelcast client impl operations clientreauthoperation dorun clientreauthoperation java at com hazelcast client impl operations clientreauthoperation access clientreauthoperation java at com hazelcast client impl operations clientreauthoperation clientreauthtask run clientreauthoperation java at com hazelcast util executor cachedexecutorservicedelegate worker run cachedexecutorservicedelegate java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java at com hazelcast util executor hazelcastmanagedthread executerun hazelcastmanagedthread java at com hazelcast util executor hazelcastmanagedthread run hazelcastmanagedthread java at submitted from unknown source at com hazelcast spi impl operationservice impl invocationfuture resolve invocationfuture java at com hazelcast spi impl abstractinvocationfuture run abstractinvocationfuture java common frames omitted test to reproduce start cluster first and then client public class clienttest test public void cluster throws interruptedexception config config new config config setlicensekey unlimited license hazelcastinstance instance hazelcast newhazelcastinstance config hazelcastinstance hazelcast newhazelcastinstance config while true thread sleep instance shutdown instance hazelcast newhazelcastinstance config test public void client throws interruptedexception clientconfig clientconfig new clientconfig clientconfig getconnectionstrategyconfig setasyncstart true clientconfig getconnectionstrategyconfig setreconnectmode clientconnectionstrategyconfig reconnectmode async clientconfig getnetworkconfig setconnectionattemptlimit clientconfig getnetworkconfig setsmartrouting true hazelcastclient newhazelcastclient clientconfig thread currentthread join
1
35,735
7,800,196,734
IssuesEvent
2018-06-09 06:15:17
StrikeNP/trac_test
https://api.github.com/repos/StrikeNP/trac_test
closed
Rename Ncm_initial to Nc_in_cloud0 and Ncm_in_cloud to Nc_in_cloud (Trac #513)
Migrated from Trac clubb_src defect raut@uwm.edu
We've used the name Ncm_initial for years, but the "m" denotes layer average, whereas in fact Ncm_initial is an overall within-cloud value. This caused confusion for years (see ticket:490). Therefore, let's rename Ncm_initial everywhere to Nc_in_cloud0. The zero denotes that it is an initial value that's a scalar, not a profile vector. Ncm_initial will need to be renamed everywhere in the source code and all input/case_setups/...model.in files. Second, CLUBB uses the variable name "Ncm_in_cloud", but this may seem contradictory, because "m" denotes layer average, but "in_cloud" denotes a within-cloud average. Therefore let's change it to Nc_in_cloud. Both these changes should leave the output of the standard CLUBB simulations bit-for-bit identical with the old results. Let's check this before committing the new code. Attachments: Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/513 ```json { "status": "closed", "changetime": "2013-08-28T17:15:43", "description": "\nWe've used the name Ncm_initial for years, but the \"m\" denotes layer average, whereas in fact Ncm_initial is an overall within-cloud value. This caused confusion for years (see ticket:490). Therefore, let's rename Ncm_initial everywhere to Nc_in_cloud0. The zero denotes that it is an initial value that's a scalar, not a profile vector. Ncm_initial will need to be renamed everywhere in the source code and all input/case_setups/...model.in files.\n\nSecond, CLUBB uses the variable name \"Ncm_in_cloud\", but this may seem contradictory, because \"m\" denotes layer average, but \"in_cloud\" denotes a within-cloud average. Therefore let's change it to Nc_in_cloud.\n\nBoth these changes should leave the output of the standard CLUBB simulations bit-for-bit identical with the old results. Let's check this before committing the new code.", "reporter": "vlarson@uwm.edu", "cc": "vlarson@uwm.edu, bmg2@uwm.edu", "resolution": "fixed", "_ts": "1377710143501379", "component": "clubb_src", "summary": "Rename Ncm_initial to Nc_in_cloud0 and Ncm_in_cloud to Nc_in_cloud", "priority": "minor", "keywords": "", "time": "2012-07-02T21:15:45", "milestone": "3. Refactor CLUBB", "owner": "raut@uwm.edu", "type": "defect" } ```
1.0
Rename Ncm_initial to Nc_in_cloud0 and Ncm_in_cloud to Nc_in_cloud (Trac #513) - We've used the name Ncm_initial for years, but the "m" denotes layer average, whereas in fact Ncm_initial is an overall within-cloud value. This caused confusion for years (see ticket:490). Therefore, let's rename Ncm_initial everywhere to Nc_in_cloud0. The zero denotes that it is an initial value that's a scalar, not a profile vector. Ncm_initial will need to be renamed everywhere in the source code and all input/case_setups/...model.in files. Second, CLUBB uses the variable name "Ncm_in_cloud", but this may seem contradictory, because "m" denotes layer average, but "in_cloud" denotes a within-cloud average. Therefore let's change it to Nc_in_cloud. Both these changes should leave the output of the standard CLUBB simulations bit-for-bit identical with the old results. Let's check this before committing the new code. Attachments: Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/513 ```json { "status": "closed", "changetime": "2013-08-28T17:15:43", "description": "\nWe've used the name Ncm_initial for years, but the \"m\" denotes layer average, whereas in fact Ncm_initial is an overall within-cloud value. This caused confusion for years (see ticket:490). Therefore, let's rename Ncm_initial everywhere to Nc_in_cloud0. The zero denotes that it is an initial value that's a scalar, not a profile vector. Ncm_initial will need to be renamed everywhere in the source code and all input/case_setups/...model.in files.\n\nSecond, CLUBB uses the variable name \"Ncm_in_cloud\", but this may seem contradictory, because \"m\" denotes layer average, but \"in_cloud\" denotes a within-cloud average. Therefore let's change it to Nc_in_cloud.\n\nBoth these changes should leave the output of the standard CLUBB simulations bit-for-bit identical with the old results. Let's check this before committing the new code.", "reporter": "vlarson@uwm.edu", "cc": "vlarson@uwm.edu, bmg2@uwm.edu", "resolution": "fixed", "_ts": "1377710143501379", "component": "clubb_src", "summary": "Rename Ncm_initial to Nc_in_cloud0 and Ncm_in_cloud to Nc_in_cloud", "priority": "minor", "keywords": "", "time": "2012-07-02T21:15:45", "milestone": "3. Refactor CLUBB", "owner": "raut@uwm.edu", "type": "defect" } ```
defect
rename ncm initial to nc in and ncm in cloud to nc in cloud trac we ve used the name ncm initial for years but the m denotes layer average whereas in fact ncm initial is an overall within cloud value this caused confusion for years see ticket therefore let s rename ncm initial everywhere to nc in the zero denotes that it is an initial value that s a scalar not a profile vector ncm initial will need to be renamed everywhere in the source code and all input case setups model in files second clubb uses the variable name ncm in cloud but this may seem contradictory because m denotes layer average but in cloud denotes a within cloud average therefore let s change it to nc in cloud both these changes should leave the output of the standard clubb simulations bit for bit identical with the old results let s check this before committing the new code attachments migrated from json status closed changetime description nwe ve used the name ncm initial for years but the m denotes layer average whereas in fact ncm initial is an overall within cloud value this caused confusion for years see ticket therefore let s rename ncm initial everywhere to nc in the zero denotes that it is an initial value that s a scalar not a profile vector ncm initial will need to be renamed everywhere in the source code and all input case setups model in files n nsecond clubb uses the variable name ncm in cloud but this may seem contradictory because m denotes layer average but in cloud denotes a within cloud average therefore let s change it to nc in cloud n nboth these changes should leave the output of the standard clubb simulations bit for bit identical with the old results let s check this before committing the new code reporter vlarson uwm edu cc vlarson uwm edu uwm edu resolution fixed ts component clubb src summary rename ncm initial to nc in and ncm in cloud to nc in cloud priority minor keywords time milestone refactor clubb owner raut uwm edu type defect
1
622,380
19,633,707,107
IssuesEvent
2022-01-08 00:17:10
bounswe/2021SpringGroup6
https://api.github.com/repos/bounswe/2021SpringGroup6
opened
Delete operations: account and event
Type: Feature Status: In Progress Platform: Front-end Priority: Medium
Users may delete their accounts or event they created. Implement it using 2-layer correction. You may use a modal to get approval of the action.
1.0
Delete operations: account and event - Users may delete their accounts or event they created. Implement it using 2-layer correction. You may use a modal to get approval of the action.
non_defect
delete operations account and event users may delete their accounts or event they created implement it using layer correction you may use a modal to get approval of the action
0
44,423
12,149,378,792
IssuesEvent
2020-04-24 16:03:21
department-of-veterans-affairs/va.gov-team
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
closed
[KEYBOARD]: VAOS - Upcoming and past appointment tabs MUST handle keyboard navigation and data properly
508-defect-1 508-issue-focus-mgmt 508/Accessibility vaos
**Feedback framework** - **❗️ Must** for if the feedback must be applied - **⚠️Should** if the feedback is best practice - **✔️ Consider** for suggestions/enhancements ## Description <!-- This is a detailed description of the issue. It should include a restatement of the title, and provide more background information. --> The upcoming and past appointments links are marked up as ARIA tabs, which is a strong interface selection for this type of presentation. They need to handle keyboard focus and arrow key events as laid out in the WCAG WAI-ARIA authoring practices spec. I've found a good article on Inclusive Components that outlines the spec minus a couple of details, which I'll add in the success criteria. [Inclusive Components: True Tabbed Interfaces](https://inclusive-components.design/tabbed-interfaces/#truetabbedinterfaces) [Stefan Judis: What's roving tabindex?](https://www.stefanjudis.com/today-i-learned/roving-tabindex/) [WCAG 2.0 Success Criteria 2.1.1: Keyboard](https://www.w3.org/TR/UNDERSTANDING-WCAG20/keyboard-operation-keyboard-operable.html) ## Point of Contact <!-- If this issue is being opened by a VFS team member, please add a point of contact. Usually this is the same person who enters the issue ticket. --> **VFS Point of Contact:** _Trevor_ ## Acceptance Criteria <!-- As a keyboard user, I want to open the Level of Coverage widget by pressing Spacebar or pressing Enter. These keypress actions should not interfere with the mouse click event also opening the widget. --> - [x] Confirm the Past appointments tab shows when users copy/paste `/past` onto the existing URL. Talked with @jbalboni this is how the application behaves now, and is a good behavior for large data fetches. - [x] The left and right arrow keys should toggle the selected tab. Tabs are treated like radio buttons in the spec, using a roving tab index. - [x] Pressing the down arrow should set focus on the selected tab panel - [x] Tab content should be in a div with `role="tabpanel"` for screen readers, per the Inclusive Components guidance - ~~The past appointments should be loaded async in the background as soon as possible, so when users click or arrow to the second tab, the data is ready instantly.~~
1.0
[KEYBOARD]: VAOS - Upcoming and past appointment tabs MUST handle keyboard navigation and data properly - **Feedback framework** - **❗️ Must** for if the feedback must be applied - **⚠️Should** if the feedback is best practice - **✔️ Consider** for suggestions/enhancements ## Description <!-- This is a detailed description of the issue. It should include a restatement of the title, and provide more background information. --> The upcoming and past appointments links are marked up as ARIA tabs, which is a strong interface selection for this type of presentation. They need to handle keyboard focus and arrow key events as laid out in the WCAG WAI-ARIA authoring practices spec. I've found a good article on Inclusive Components that outlines the spec minus a couple of details, which I'll add in the success criteria. [Inclusive Components: True Tabbed Interfaces](https://inclusive-components.design/tabbed-interfaces/#truetabbedinterfaces) [Stefan Judis: What's roving tabindex?](https://www.stefanjudis.com/today-i-learned/roving-tabindex/) [WCAG 2.0 Success Criteria 2.1.1: Keyboard](https://www.w3.org/TR/UNDERSTANDING-WCAG20/keyboard-operation-keyboard-operable.html) ## Point of Contact <!-- If this issue is being opened by a VFS team member, please add a point of contact. Usually this is the same person who enters the issue ticket. --> **VFS Point of Contact:** _Trevor_ ## Acceptance Criteria <!-- As a keyboard user, I want to open the Level of Coverage widget by pressing Spacebar or pressing Enter. These keypress actions should not interfere with the mouse click event also opening the widget. --> - [x] Confirm the Past appointments tab shows when users copy/paste `/past` onto the existing URL. Talked with @jbalboni this is how the application behaves now, and is a good behavior for large data fetches. - [x] The left and right arrow keys should toggle the selected tab. Tabs are treated like radio buttons in the spec, using a roving tab index. - [x] Pressing the down arrow should set focus on the selected tab panel - [x] Tab content should be in a div with `role="tabpanel"` for screen readers, per the Inclusive Components guidance - ~~The past appointments should be loaded async in the background as soon as possible, so when users click or arrow to the second tab, the data is ready instantly.~~
defect
vaos upcoming and past appointment tabs must handle keyboard navigation and data properly feedback framework ❗️ must for if the feedback must be applied ⚠️should if the feedback is best practice ✔️ consider for suggestions enhancements description the upcoming and past appointments links are marked up as aria tabs which is a strong interface selection for this type of presentation they need to handle keyboard focus and arrow key events as laid out in the wcag wai aria authoring practices spec i ve found a good article on inclusive components that outlines the spec minus a couple of details which i ll add in the success criteria point of contact if this issue is being opened by a vfs team member please add a point of contact usually this is the same person who enters the issue ticket vfs point of contact trevor acceptance criteria confirm the past appointments tab shows when users copy paste past onto the existing url talked with jbalboni this is how the application behaves now and is a good behavior for large data fetches the left and right arrow keys should toggle the selected tab tabs are treated like radio buttons in the spec using a roving tab index pressing the down arrow should set focus on the selected tab panel tab content should be in a div with role tabpanel for screen readers per the inclusive components guidance the past appointments should be loaded async in the background as soon as possible so when users click or arrow to the second tab the data is ready instantly
1
250,392
7,976,271,005
IssuesEvent
2018-07-17 12:09:21
mozilla/addons-frontend
https://api.github.com/repos/mozilla/addons-frontend
closed
Remove video from top of about:addons (disco pane)
priority: p3 project: disco pane
As a first step in making the content at the top of about:addons better, remove the video from the top of about:addons. The screenshot below shows content and a link to the video. Remove the video. ![disco pane screenshot](https://user-images.githubusercontent.com/911251/41627175-04a396bc-73d4-11e8-9051-a5fbd2e43351.png)
1.0
Remove video from top of about:addons (disco pane) - As a first step in making the content at the top of about:addons better, remove the video from the top of about:addons. The screenshot below shows content and a link to the video. Remove the video. ![disco pane screenshot](https://user-images.githubusercontent.com/911251/41627175-04a396bc-73d4-11e8-9051-a5fbd2e43351.png)
non_defect
remove video from top of about addons disco pane as a first step in making the content at the top of about addons better remove the video from the top of about addons the screenshot below shows content and a link to the video remove the video
0
27,882
4,344,436,843
IssuesEvent
2016-07-29 08:23:17
bitcoin/bitcoin
https://api.github.com/repos/bitcoin/bitcoin
opened
Rare walletbackup.py failure: cannot reproduce
Tests
I had this problem once while I was running the `walletbackup.py` test on 0.13: ``` walletbackup.py: Initializing test directory /tmp/testo3khmc00/37 INFO: Generating initial blockchain Assertion failed: 76.69962300 != 50 Stopping nodes Not cleaning up dir /tmp/testo3khmc00/37 Failed ``` I'm unfortunately unable to reproduce it again, looks like a race condition.
1.0
Rare walletbackup.py failure: cannot reproduce - I had this problem once while I was running the `walletbackup.py` test on 0.13: ``` walletbackup.py: Initializing test directory /tmp/testo3khmc00/37 INFO: Generating initial blockchain Assertion failed: 76.69962300 != 50 Stopping nodes Not cleaning up dir /tmp/testo3khmc00/37 Failed ``` I'm unfortunately unable to reproduce it again, looks like a race condition.
non_defect
rare walletbackup py failure cannot reproduce i had this problem once while i was running the walletbackup py test on walletbackup py initializing test directory tmp info generating initial blockchain assertion failed stopping nodes not cleaning up dir tmp failed i m unfortunately unable to reproduce it again looks like a race condition
0
70,981
23,393,903,095
IssuesEvent
2022-08-11 20:46:49
department-of-veterans-affairs/va.gov-team
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
closed
Screenreader announces the full day name (VAMC)
frontend 508/Accessibility vsa 508-defect-3 Facilities 508-issue-cognition frontend-vamc vamc-pages
## Issue Description This issue was originally reported in #6025. The shorthand pattern for expressing hours of operation like **Mon:** 24/7 to represent Monday, 24 hours a day, 7 days a week is not made clear to assistive technology. The markup for both day and time should be improved but this issue will address only the day of the week from the template perspective. Hours are sent via API and need to be handled differently. ## Task: - [ ] Update markup to better serve screen readers. A code snippet has been provided by Trevor below. ``` <span aria-hidden="true">Mon:</span> <span class="vads-u-visibility--screen-reader">Monday</span> ``` ## Type of Issue template ## Acceptance Criteria * [ ] Screen reader announces the full day name ## Screenshots or Trace Logs <!-- Drop any screenshots or error logs that might be useful for debugging --> <img width="657" alt="Screen Shot 2020-02-19 at 4 31 54 PM" src="https://user-images.githubusercontent.com/934879/74883339-5e151200-5336-11ea-8723-743807338d1e.png">
1.0
Screenreader announces the full day name (VAMC) - ## Issue Description This issue was originally reported in #6025. The shorthand pattern for expressing hours of operation like **Mon:** 24/7 to represent Monday, 24 hours a day, 7 days a week is not made clear to assistive technology. The markup for both day and time should be improved but this issue will address only the day of the week from the template perspective. Hours are sent via API and need to be handled differently. ## Task: - [ ] Update markup to better serve screen readers. A code snippet has been provided by Trevor below. ``` <span aria-hidden="true">Mon:</span> <span class="vads-u-visibility--screen-reader">Monday</span> ``` ## Type of Issue template ## Acceptance Criteria * [ ] Screen reader announces the full day name ## Screenshots or Trace Logs <!-- Drop any screenshots or error logs that might be useful for debugging --> <img width="657" alt="Screen Shot 2020-02-19 at 4 31 54 PM" src="https://user-images.githubusercontent.com/934879/74883339-5e151200-5336-11ea-8723-743807338d1e.png">
defect
screenreader announces the full day name vamc issue description this issue was originally reported in the shorthand pattern for expressing hours of operation like mon to represent monday hours a day days a week is not made clear to assistive technology the markup for both day and time should be improved but this issue will address only the day of the week from the template perspective hours are sent via api and need to be handled differently task update markup to better serve screen readers a code snippet has been provided by trevor below mon monday type of issue template acceptance criteria screen reader announces the full day name screenshots or trace logs img width alt screen shot at pm src
1
65,529
3,231,821,781
IssuesEvent
2015-10-13 01:00:15
aic-collections/aicdams-lakeshore
https://api.github.com/repos/aic-collections/aicdams-lakeshore
closed
Create conservation metadata class
LOW priority
Belongs to 1 and only 1 asset class and contains permission restrictions so that it is only editable by members of the conservation department.
1.0
Create conservation metadata class - Belongs to 1 and only 1 asset class and contains permission restrictions so that it is only editable by members of the conservation department.
non_defect
create conservation metadata class belongs to and only asset class and contains permission restrictions so that it is only editable by members of the conservation department
0
27,376
4,981,712,505
IssuesEvent
2016-12-07 08:59:20
Cockatrice/Cockatrice
https://api.github.com/repos/Cockatrice/Cockatrice
closed
OSX Sierra, Load Deck loads empty list.. Not previous folder.
Defect - Basic OS - macOS UI / UX
<b>OS:</b> OSX 10.12.1 (Sierra) <b>Cockatrice version:</b> Version 3498b16 (2016-10-30) ___ <br> I recently upgraded from El Capitan to Sierra on my mac. Previously, opening a deck opened up the graphical window where you select the deck. When I did this, it loaded the last folder I opened a deck from. This made selecting decks easy, both in Local Game, and in Deck Editor. However now in Sierra, I've noticed that when I want to open a deck, the list of file is empty. This is essentially what it looks like: https://www.dropbox.com/s/onq2jw0sx0jp5pa/Screenshot%202016-12-03%2014.37.54.png?dl=0 I can workaround this by clicking the top part and selecting a location from the last used locations, so the bug is rather minor, but impacts usability. https://www.dropbox.com/s/as923vd2od25sc4/Screenshot%202016-12-03%2014.39.19.png?dl=0
1.0
OSX Sierra, Load Deck loads empty list.. Not previous folder. - <b>OS:</b> OSX 10.12.1 (Sierra) <b>Cockatrice version:</b> Version 3498b16 (2016-10-30) ___ <br> I recently upgraded from El Capitan to Sierra on my mac. Previously, opening a deck opened up the graphical window where you select the deck. When I did this, it loaded the last folder I opened a deck from. This made selecting decks easy, both in Local Game, and in Deck Editor. However now in Sierra, I've noticed that when I want to open a deck, the list of file is empty. This is essentially what it looks like: https://www.dropbox.com/s/onq2jw0sx0jp5pa/Screenshot%202016-12-03%2014.37.54.png?dl=0 I can workaround this by clicking the top part and selecting a location from the last used locations, so the bug is rather minor, but impacts usability. https://www.dropbox.com/s/as923vd2od25sc4/Screenshot%202016-12-03%2014.39.19.png?dl=0
defect
osx sierra load deck loads empty list not previous folder os osx sierra cockatrice version version i recently upgraded from el capitan to sierra on my mac previously opening a deck opened up the graphical window where you select the deck when i did this it loaded the last folder i opened a deck from this made selecting decks easy both in local game and in deck editor however now in sierra i ve noticed that when i want to open a deck the list of file is empty this is essentially what it looks like i can workaround this by clicking the top part and selecting a location from the last used locations so the bug is rather minor but impacts usability
1
79,351
28,120,546,591
IssuesEvent
2023-03-31 13:57:16
cf-convention/cf-conventions
https://api.github.com/repos/cf-convention/cf-conventions
closed
Should cf_role be deprecated in favor of standard_name?
defect change agreed
Dear CF community, Based on recent conversation and a number of experiences where this caused confusion, I wonder if it would be wise to deprecate the `cf_role` attribute in favor of extended `standard_name` attributes? Scanning the spec for instances of `cf_role` I don't see any cases where a `standard_name` couldn't be used instead. I ask because I have been confused about the purpose (role) of `standard_name` and the `cf_role` attribute. It seems that someone introduced `cf_role` with a separation of concerns between functional and quantity type in mind. Other people who have contributed did not continue using that separation of concerns for other parts of the specification. Maybe I'm missing something that forces the inclusion of `cf_role` but none of the examples show use of both `cf_role` and `standard_name` so the reason for the additional functional descriptor is not clear. If we are to keep `cf_role` and `standard_name`, it would be useful to document the need for both more clearly in the specification and examples? Regards -- Dave @JonathanGregory writes: Following discussion, Dave and I have [proposed a change](https://github.com/cf-convention/cf-conventions/issues/430#issuecomment-1464389052) in order to clarify the purpose of `cf_role`. We think this change would correct a defect in the convention text but would not be a material change to the convention. @JonathanGregory writes: [Pull request 434](https://github.com/cf-convention/cf-conventions/pull/434) implements this change.
1.0
Should cf_role be deprecated in favor of standard_name? - Dear CF community, Based on recent conversation and a number of experiences where this caused confusion, I wonder if it would be wise to deprecate the `cf_role` attribute in favor of extended `standard_name` attributes? Scanning the spec for instances of `cf_role` I don't see any cases where a `standard_name` couldn't be used instead. I ask because I have been confused about the purpose (role) of `standard_name` and the `cf_role` attribute. It seems that someone introduced `cf_role` with a separation of concerns between functional and quantity type in mind. Other people who have contributed did not continue using that separation of concerns for other parts of the specification. Maybe I'm missing something that forces the inclusion of `cf_role` but none of the examples show use of both `cf_role` and `standard_name` so the reason for the additional functional descriptor is not clear. If we are to keep `cf_role` and `standard_name`, it would be useful to document the need for both more clearly in the specification and examples? Regards -- Dave @JonathanGregory writes: Following discussion, Dave and I have [proposed a change](https://github.com/cf-convention/cf-conventions/issues/430#issuecomment-1464389052) in order to clarify the purpose of `cf_role`. We think this change would correct a defect in the convention text but would not be a material change to the convention. @JonathanGregory writes: [Pull request 434](https://github.com/cf-convention/cf-conventions/pull/434) implements this change.
defect
should cf role be deprecated in favor of standard name dear cf community based on recent conversation and a number of experiences where this caused confusion i wonder if it would be wise to deprecate the cf role attribute in favor of extended standard name attributes scanning the spec for instances of cf role i don t see any cases where a standard name couldn t be used instead i ask because i have been confused about the purpose role of standard name and the cf role attribute it seems that someone introduced cf role with a separation of concerns between functional and quantity type in mind other people who have contributed did not continue using that separation of concerns for other parts of the specification maybe i m missing something that forces the inclusion of cf role but none of the examples show use of both cf role and standard name so the reason for the additional functional descriptor is not clear if we are to keep cf role and standard name it would be useful to document the need for both more clearly in the specification and examples regards dave jonathangregory writes following discussion dave and i have in order to clarify the purpose of cf role we think this change would correct a defect in the convention text but would not be a material change to the convention jonathangregory writes implements this change
1
1,299
2,603,750,652
IssuesEvent
2015-02-24 17:43:59
chrsmith/bwapi
https://api.github.com/repos/chrsmith/bwapi
closed
High apm, when there are no selected units
auto-migrated Performance Type-Defect
``` What steps will reproduce the problem? When a BWAPI game is started and there is no units selected, my (my AI's) APM is near 1400. This means 1 action per frame. But when I select something, the APM falls to normal values (depending on that, what is the AI doing). When the selected unit dies/goes out of vision (if enemy) APM goes high again. Another effect of this is, when there is nothing selected and a building rally position is set at this time, the rally position doesn't work. I turned all BWAPI functions (including map analyzer, read map), but this actions again exist. It's not so big problem, but every time i need to select something, in order to see the actual AI's APM. What version of the product are you using? On what operating system? BWAPI_Beta_2.6.1 on Windows XP Please provide any additional information below. When I open the replay with BWChart, i see actions "Deselect" with parameter "lost vision on selected unit". It's 1 action every frame. And they exist there, where no units are selected through the GUI. ``` ----- Original issue reported on code.google.com by `Nuclear...@gmail.com` on 15 Feb 2010 at 9:25
1.0
High apm, when there are no selected units - ``` What steps will reproduce the problem? When a BWAPI game is started and there is no units selected, my (my AI's) APM is near 1400. This means 1 action per frame. But when I select something, the APM falls to normal values (depending on that, what is the AI doing). When the selected unit dies/goes out of vision (if enemy) APM goes high again. Another effect of this is, when there is nothing selected and a building rally position is set at this time, the rally position doesn't work. I turned all BWAPI functions (including map analyzer, read map), but this actions again exist. It's not so big problem, but every time i need to select something, in order to see the actual AI's APM. What version of the product are you using? On what operating system? BWAPI_Beta_2.6.1 on Windows XP Please provide any additional information below. When I open the replay with BWChart, i see actions "Deselect" with parameter "lost vision on selected unit". It's 1 action every frame. And they exist there, where no units are selected through the GUI. ``` ----- Original issue reported on code.google.com by `Nuclear...@gmail.com` on 15 Feb 2010 at 9:25
defect
high apm when there are no selected units what steps will reproduce the problem when a bwapi game is started and there is no units selected my my ai s apm is near this means action per frame but when i select something the apm falls to normal values depending on that what is the ai doing when the selected unit dies goes out of vision if enemy apm goes high again another effect of this is when there is nothing selected and a building rally position is set at this time the rally position doesn t work i turned all bwapi functions including map analyzer read map but this actions again exist it s not so big problem but every time i need to select something in order to see the actual ai s apm what version of the product are you using on what operating system bwapi beta on windows xp please provide any additional information below when i open the replay with bwchart i see actions deselect with parameter lost vision on selected unit it s action every frame and they exist there where no units are selected through the gui original issue reported on code google com by nuclear gmail com on feb at
1
26,031
4,545,454,712
IssuesEvent
2016-09-11 08:21:08
colour-science/colour
https://api.github.com/repos/colour-science/colour
opened
"LLAB(l:c)" colour appearance model unit tests are not being run.
API Defect Major
I noticed that decorating two methods in `colour.appearance.tests.tests_llab.TestLLABColourAppearanceModel` class with `mock.patch` prevents the unit tests to run properly. I commented those and discovered that the first fixture test case is broken: hue angle *h_L* should be equal to 229.5 [1] but we compute 270. References ----------- [1] Fairchild, M. D. (2004). Color Appearance Models (2nd ed.). book, Wiley. ISBN:978-0470012161
1.0
"LLAB(l:c)" colour appearance model unit tests are not being run. - I noticed that decorating two methods in `colour.appearance.tests.tests_llab.TestLLABColourAppearanceModel` class with `mock.patch` prevents the unit tests to run properly. I commented those and discovered that the first fixture test case is broken: hue angle *h_L* should be equal to 229.5 [1] but we compute 270. References ----------- [1] Fairchild, M. D. (2004). Color Appearance Models (2nd ed.). book, Wiley. ISBN:978-0470012161
defect
llab l c colour appearance model unit tests are not being run i noticed that decorating two methods in colour appearance tests tests llab testllabcolourappearancemodel class with mock patch prevents the unit tests to run properly i commented those and discovered that the first fixture test case is broken hue angle h l should be equal to but we compute references fairchild m d color appearance models ed book wiley isbn
1
29,486
5,638,326,405
IssuesEvent
2017-04-06 11:38:27
StackExchange/StackExchange.Redis
https://api.github.com/repos/StackExchange/StackExchange.Redis
closed
[Question] 1.1.603 release notes, or at least tag for source code?
documentation
Hello. Can we see release notes for 1.1.603 nuget package? Tag on source code will work for me, thanks. Regards, Anatoly.
1.0
[Question] 1.1.603 release notes, or at least tag for source code? - Hello. Can we see release notes for 1.1.603 nuget package? Tag on source code will work for me, thanks. Regards, Anatoly.
non_defect
release notes or at least tag for source code hello can we see release notes for nuget package tag on source code will work for me thanks regards anatoly
0
7,950
11,167,822,962
IssuesEvent
2019-12-27 18:50:07
openopps/openopps-platform
https://api.github.com/repos/openopps/openopps-platform
closed
Add security clearance to internship view
Approved Opportunity Page Requirements Ready State Dept.
Who: Internship applicants What: Ability to see what type of security clearance they will be getting Why: in order to help the process Acceptance Criteria: - add security clearance to the applicant view of the internship - Field label should be "Required security level" - Field should display last in the blue box on the right of the page. (if this is too much for this box please advise) -Display the value from the security clearance field on the created internship. If value is none display ,"None (Low Risk Public Trust Certificate) "
1.0
Add security clearance to internship view - Who: Internship applicants What: Ability to see what type of security clearance they will be getting Why: in order to help the process Acceptance Criteria: - add security clearance to the applicant view of the internship - Field label should be "Required security level" - Field should display last in the blue box on the right of the page. (if this is too much for this box please advise) -Display the value from the security clearance field on the created internship. If value is none display ,"None (Low Risk Public Trust Certificate) "
non_defect
add security clearance to internship view who internship applicants what ability to see what type of security clearance they will be getting why in order to help the process acceptance criteria add security clearance to the applicant view of the internship field label should be required security level field should display last in the blue box on the right of the page if this is too much for this box please advise display the value from the security clearance field on the created internship if value is none display none low risk public trust certificate
0
63,355
7,719,115,455
IssuesEvent
2018-05-23 18:21:42
Opentrons/opentrons
https://api.github.com/repos/Opentrons/opentrons
opened
Ingredient setup well selection wierdness
feature protocol designer
## Acceptance criteria: -When you select well(s) with all the same ingredient, it should open the form with that info in the fields -When you select a set of wells where some ingredients are different than others, it should open the form with no fields filled out -When you select a set of wells where some ingredients are different than others, fill out the blank form, and hit save, it should just overwrite the pre-existing well contents. Currently it creates a mixed grey well. ## Designs: n/a
1.0
Ingredient setup well selection wierdness - ## Acceptance criteria: -When you select well(s) with all the same ingredient, it should open the form with that info in the fields -When you select a set of wells where some ingredients are different than others, it should open the form with no fields filled out -When you select a set of wells where some ingredients are different than others, fill out the blank form, and hit save, it should just overwrite the pre-existing well contents. Currently it creates a mixed grey well. ## Designs: n/a
non_defect
ingredient setup well selection wierdness acceptance criteria when you select well s with all the same ingredient it should open the form with that info in the fields when you select a set of wells where some ingredients are different than others it should open the form with no fields filled out when you select a set of wells where some ingredients are different than others fill out the blank form and hit save it should just overwrite the pre existing well contents currently it creates a mixed grey well designs n a
0
314,120
26,976,825,126
IssuesEvent
2023-02-09 10:09:47
wazuh/wazuh-qa
https://api.github.com/repos/wazuh/wazuh-qa
closed
Improve `test_assign_groups_guess`
team/qa test/system type/fix subteam/qa-main
## Description After the research done in #3776, we could see that `test_assign_groups_guess` is stable. However, it takes 21 minutes to finish the test execution. We need to reduce this time. In addition, after the research done in #3876, we saw that we could improve the test by checking the following logs: - Worker ``` 2023/02/01 13:40:28 wazuh-remoted[1718] manager.c:556 at assign_group_to_agent_worker(): DEBUG: Sending message to master node: '{"daemon_name":"remoted","message":{"command":"assigngroup","parameters":{"agent":"060","md5":"x"}}}' 2023/02/01 13:40:28 wazuh-remoted[1718] manager.c:560 at assign_group_to_agent_worker(): DEBUG: Message received from master node: '{"error":0,"message":"ok","data":{"group":"default"}}' 2023/02/01 13:40:28 wazuh-remoted[1718] manager.c:1598 at wait_for_msgs(): DEBUG: Sending file 'default/merged.mg' to agent '060'. 2023/02/01 13:40:28 wazuh-remoted[1718] manager.c:1614 at wait_for_msgs(): DEBUG: End sending file 'default/merged.mg' to agent '060'. ``` - Master ``` 2023/02/01 13:40:28 wazuh-remoted[1877] manager.c:508 at assign_group_to_agent(): DEBUG: Agent '060' with file 'merged.mg' MD5 'x' 2023/02/01 13:40:28 wazuh-remoted[1877] manager.c:527 at assign_group_to_agent(): DEBUG: Group assigned: 'default' ```
1.0
Improve `test_assign_groups_guess` - ## Description After the research done in #3776, we could see that `test_assign_groups_guess` is stable. However, it takes 21 minutes to finish the test execution. We need to reduce this time. In addition, after the research done in #3876, we saw that we could improve the test by checking the following logs: - Worker ``` 2023/02/01 13:40:28 wazuh-remoted[1718] manager.c:556 at assign_group_to_agent_worker(): DEBUG: Sending message to master node: '{"daemon_name":"remoted","message":{"command":"assigngroup","parameters":{"agent":"060","md5":"x"}}}' 2023/02/01 13:40:28 wazuh-remoted[1718] manager.c:560 at assign_group_to_agent_worker(): DEBUG: Message received from master node: '{"error":0,"message":"ok","data":{"group":"default"}}' 2023/02/01 13:40:28 wazuh-remoted[1718] manager.c:1598 at wait_for_msgs(): DEBUG: Sending file 'default/merged.mg' to agent '060'. 2023/02/01 13:40:28 wazuh-remoted[1718] manager.c:1614 at wait_for_msgs(): DEBUG: End sending file 'default/merged.mg' to agent '060'. ``` - Master ``` 2023/02/01 13:40:28 wazuh-remoted[1877] manager.c:508 at assign_group_to_agent(): DEBUG: Agent '060' with file 'merged.mg' MD5 'x' 2023/02/01 13:40:28 wazuh-remoted[1877] manager.c:527 at assign_group_to_agent(): DEBUG: Group assigned: 'default' ```
non_defect
improve test assign groups guess description after the research done in we could see that test assign groups guess is stable however it takes minutes to finish the test execution we need to reduce this time in addition after the research done in we saw that we could improve the test by checking the following logs worker wazuh remoted manager c at assign group to agent worker debug sending message to master node daemon name remoted message command assigngroup parameters agent x wazuh remoted manager c at assign group to agent worker debug message received from master node error message ok data group default wazuh remoted manager c at wait for msgs debug sending file default merged mg to agent wazuh remoted manager c at wait for msgs debug end sending file default merged mg to agent master wazuh remoted manager c at assign group to agent debug agent with file merged mg x wazuh remoted manager c at assign group to agent debug group assigned default
0
74,334
9,780,301,330
IssuesEvent
2019-06-07 16:31:09
fastify/fastify
https://api.github.com/repos/fastify/fastify
closed
Requests body is not logged in POST request
bug documentation good first issue
## 🐛 Bug Report Hello, I cannot see the request body in my logs when I do POST requests to my test server. ## To Reproduce ```js 'use strict' // Read the .env file. require('dotenv').config() // Require the framework const Fastify = require('fastify') // Instantiate Fastify with some config // const app = Fastify({ // logger: true, // pluginTimeout: 10000 // }) const app = Fastify({ logger: { prettyPrint: true, serializers: { res (res) { // the default return { statusCode: res.statusCode } }, req (req) { return { method: req.method, url: req.url, path: req.path, // parameters: req.parameters, // Including the body and headers in the log could be in violation // of privacy laws, e.g. GDPR. You should use the "redact" option to // remove sensitive fields. It could also leak authentication data in // the logs. body: req.body, headers: req.headers } } } } }) // Register your application as a normal plugin. app.register(require('./app.js')) // Start listening. app.listen(process.env.PORT || 5123, (err) => { if (err) { app.log.error(err) process.exit(1) } }) ``` ## Expected behavior The request body in the logs: ```js { "gender": [ "f", "m" ], "country": "ITA", "isTeam": false, "isSearch": true, "name": "xxxx", "fetchPage": 1 } ``` ## What I get ```js [1559408233939] INFO (15868 on sauron): Server listening at http://127.0.0.1:5123 [1559408239792] INFO (15868 on sauron): incoming request reqId: 1 req: { "method": "POST", "url": "/athletes/search", "headers": { "content-type": "application/json", "user-agent": "PostmanRuntime/7.13.0", "accept": "*/*", "cache-control": "no-cache", "postman-token": "f4002054-80c5-4b01-8fa0-0a173f12e73b", "host": "localhost:5123", "accept-encoding": "gzip, deflate", "content-length": "98", "connection": "keep-alive" } } ``` ## Your Environment - *node version*: 11 - *fastify version*: = 2.4.1 - *os*: Windows
1.0
Requests body is not logged in POST request - ## 🐛 Bug Report Hello, I cannot see the request body in my logs when I do POST requests to my test server. ## To Reproduce ```js 'use strict' // Read the .env file. require('dotenv').config() // Require the framework const Fastify = require('fastify') // Instantiate Fastify with some config // const app = Fastify({ // logger: true, // pluginTimeout: 10000 // }) const app = Fastify({ logger: { prettyPrint: true, serializers: { res (res) { // the default return { statusCode: res.statusCode } }, req (req) { return { method: req.method, url: req.url, path: req.path, // parameters: req.parameters, // Including the body and headers in the log could be in violation // of privacy laws, e.g. GDPR. You should use the "redact" option to // remove sensitive fields. It could also leak authentication data in // the logs. body: req.body, headers: req.headers } } } } }) // Register your application as a normal plugin. app.register(require('./app.js')) // Start listening. app.listen(process.env.PORT || 5123, (err) => { if (err) { app.log.error(err) process.exit(1) } }) ``` ## Expected behavior The request body in the logs: ```js { "gender": [ "f", "m" ], "country": "ITA", "isTeam": false, "isSearch": true, "name": "xxxx", "fetchPage": 1 } ``` ## What I get ```js [1559408233939] INFO (15868 on sauron): Server listening at http://127.0.0.1:5123 [1559408239792] INFO (15868 on sauron): incoming request reqId: 1 req: { "method": "POST", "url": "/athletes/search", "headers": { "content-type": "application/json", "user-agent": "PostmanRuntime/7.13.0", "accept": "*/*", "cache-control": "no-cache", "postman-token": "f4002054-80c5-4b01-8fa0-0a173f12e73b", "host": "localhost:5123", "accept-encoding": "gzip, deflate", "content-length": "98", "connection": "keep-alive" } } ``` ## Your Environment - *node version*: 11 - *fastify version*: = 2.4.1 - *os*: Windows
non_defect
requests body is not logged in post request 🐛 bug report hello i cannot see the request body in my logs when i do post requests to my test server to reproduce js use strict read the env file require dotenv config require the framework const fastify require fastify instantiate fastify with some config const app fastify logger true plugintimeout const app fastify logger prettyprint true serializers res res the default return statuscode res statuscode req req return method req method url req url path req path parameters req parameters including the body and headers in the log could be in violation of privacy laws e g gdpr you should use the redact option to remove sensitive fields it could also leak authentication data in the logs body req body headers req headers register your application as a normal plugin app register require app js start listening app listen process env port err if err app log error err process exit expected behavior the request body in the logs js gender f m country ita isteam false issearch true name xxxx fetchpage what i get js info on sauron server listening at info on sauron incoming request reqid req method post url athletes search headers content type application json user agent postmanruntime accept cache control no cache postman token host localhost accept encoding gzip deflate content length connection keep alive your environment node version fastify version os windows
0
245,033
7,880,740,888
IssuesEvent
2018-06-26 16:48:04
aowen87/FOO
https://api.github.com/repos/aowen87/FOO
closed
Initial gui window size needs tweaking with QT 5
Likelihood: 3 - Occasional OS: All Priority: Normal Severity: 2 - Minor Irritation Support Group: Any Target Version: 2.11.0 bug version: trunk
When VisIt is built with QT 5, initial gui window size is smaller, doesn't fill the height of the screen. Same with viewer window.
1.0
Initial gui window size needs tweaking with QT 5 - When VisIt is built with QT 5, initial gui window size is smaller, doesn't fill the height of the screen. Same with viewer window.
non_defect
initial gui window size needs tweaking with qt when visit is built with qt initial gui window size is smaller doesn t fill the height of the screen same with viewer window
0
49,417
7,503,830,870
IssuesEvent
2018-04-10 00:03:47
mono/mono
https://api.github.com/repos/mono/mono
closed
Dead link (404) in a runtime execution engine exception
documentation good first issue
Got this error message from a FullAOT test where I have some native code that wants to invoke a managed delegate. Following the link [`https://developer.xamarin.com/guides/ios/advanced_topics/limitations/`](https://developer.xamarin.com/guides/ios/advanced_topics/limitations/) gives a 404. ``` System.Reflection.TargetInvocationException: Exception has been thrown by the target of an invocation. ---> System.ExecutionEngineException: Attempting to JIT compile method '(wrapper native-to-managed) Tests/<>c__DisplayClass5_0:<test_0_setjmp_exn_handler>b__0 ()' while running in aot-only mode. See https://developer.xamarin.com/guides/ios/advanced_topics/limitations/ for more information. at (wrapper managed-to-native) System.Object.__icall_wrapper_mono_delegate_to_ftnptr(object) at (wrapper managed-to-native) Tests.mono_test_setjmp_and_call(Tests/VoidVoidDelegate,intptr&) at Tests.test_0_setjmp_exn_handler () [0x00023] in <16258fbbedad473992f15c8212374900>:0 at (wrapper managed-to-native) System.Reflection.MonoMethod.InternalInvoke(System.Reflection.MonoMethod,object,object[],System.Exception&) at System.Reflection.MonoMethod.Invoke (System.Object obj, System.Reflection.BindingFlags invokeAttr, System.Reflection.Binder binder, System.Object[] parameters, System.Globalization.CultureInfo culture) [0x00032] in <e1ccb1c8ff6b4a10b926c2520e3aa15d>:0 --- End of inner exception stack trace --- at System.Reflection.MonoMethod.Invoke (System.Object obj, System.Reflection.BindingFlags invokeAttr, System.Reflection.Binder binder, System.Object[] parameters, System.Globalization.CultureInfo culture) [0x0004b] in <e1ccb1c8ff6b4a10b926c2520e3aa15d>:0 at System.Reflection.MethodBase.Invoke (System.Object obj, System.Object[] parameters) [0x00000] in <e1ccb1c8ff6b4a10b926c2520e3aa15d>:0 at TestDriver.RunTests (System.Type type, System.String[] args, TestDriverReporter reporter) [0x003b9] in <e541aee421a44561b3fa0f2b98acb890>:0 at TestDriver.RunTests (System.Type type) [0x00001] in <e541aee421a44561b3fa0f2b98acb890>:0 at Tests.Main () [0x00001] in <16258fbbedad473992f15c8212374900>:0 ```
1.0
Dead link (404) in a runtime execution engine exception - Got this error message from a FullAOT test where I have some native code that wants to invoke a managed delegate. Following the link [`https://developer.xamarin.com/guides/ios/advanced_topics/limitations/`](https://developer.xamarin.com/guides/ios/advanced_topics/limitations/) gives a 404. ``` System.Reflection.TargetInvocationException: Exception has been thrown by the target of an invocation. ---> System.ExecutionEngineException: Attempting to JIT compile method '(wrapper native-to-managed) Tests/<>c__DisplayClass5_0:<test_0_setjmp_exn_handler>b__0 ()' while running in aot-only mode. See https://developer.xamarin.com/guides/ios/advanced_topics/limitations/ for more information. at (wrapper managed-to-native) System.Object.__icall_wrapper_mono_delegate_to_ftnptr(object) at (wrapper managed-to-native) Tests.mono_test_setjmp_and_call(Tests/VoidVoidDelegate,intptr&) at Tests.test_0_setjmp_exn_handler () [0x00023] in <16258fbbedad473992f15c8212374900>:0 at (wrapper managed-to-native) System.Reflection.MonoMethod.InternalInvoke(System.Reflection.MonoMethod,object,object[],System.Exception&) at System.Reflection.MonoMethod.Invoke (System.Object obj, System.Reflection.BindingFlags invokeAttr, System.Reflection.Binder binder, System.Object[] parameters, System.Globalization.CultureInfo culture) [0x00032] in <e1ccb1c8ff6b4a10b926c2520e3aa15d>:0 --- End of inner exception stack trace --- at System.Reflection.MonoMethod.Invoke (System.Object obj, System.Reflection.BindingFlags invokeAttr, System.Reflection.Binder binder, System.Object[] parameters, System.Globalization.CultureInfo culture) [0x0004b] in <e1ccb1c8ff6b4a10b926c2520e3aa15d>:0 at System.Reflection.MethodBase.Invoke (System.Object obj, System.Object[] parameters) [0x00000] in <e1ccb1c8ff6b4a10b926c2520e3aa15d>:0 at TestDriver.RunTests (System.Type type, System.String[] args, TestDriverReporter reporter) [0x003b9] in <e541aee421a44561b3fa0f2b98acb890>:0 at TestDriver.RunTests (System.Type type) [0x00001] in <e541aee421a44561b3fa0f2b98acb890>:0 at Tests.Main () [0x00001] in <16258fbbedad473992f15c8212374900>:0 ```
non_defect
dead link in a runtime execution engine exception got this error message from a fullaot test where i have some native code that wants to invoke a managed delegate following the link gives a system reflection targetinvocationexception exception has been thrown by the target of an invocation system executionengineexception attempting to jit compile method wrapper native to managed tests c b while running in aot only mode see for more information at wrapper managed to native system object icall wrapper mono delegate to ftnptr object at wrapper managed to native tests mono test setjmp and call tests voidvoiddelegate intptr at tests test setjmp exn handler in at wrapper managed to native system reflection monomethod internalinvoke system reflection monomethod object object system exception at system reflection monomethod invoke system object obj system reflection bindingflags invokeattr system reflection binder binder system object parameters system globalization cultureinfo culture in end of inner exception stack trace at system reflection monomethod invoke system object obj system reflection bindingflags invokeattr system reflection binder binder system object parameters system globalization cultureinfo culture in at system reflection methodbase invoke system object obj system object parameters in at testdriver runtests system type type system string args testdriverreporter reporter in at testdriver runtests system type type in at tests main in
0
57,056
15,614,529,339
IssuesEvent
2021-03-19 17:53:59
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
opened
Vertical video calls from some element-android devices are distorted
T-Defect
During video calls between element-web and element-android users, the element-web user will see an extremely distorted picture if the element-android's device is oriented vertically. If the android user rotates their phone horizontally, the distortion goes away. This behavior has been tested on five phones: Pixel 3a with stock Android 11: DISTORTED Pixel 3 with CalyxOS Android 11: DISTORTED Pixel 4a with stock Android 11: Not distorted Google Pixel (gen1) with stock Android 10: Not distorted Unihertz Jelly2 with stock Android 10: Not distorted Testing was performed with a combination of accounts, homed to matrix.org and mssj.me. mssj.me is currently running Synapse 1.29.0. Android apps were installed from either the Google Play store or F-Droid, and are up to date with their respective repositories. an example of the distortion when video is in vertical mode: ![Screenshot_20210319102148](https://user-images.githubusercontent.com/8409433/111821479-f108de80-889f-11eb-81c1-377233fef4c0.png) This is a duplicate of [this issue for element-android](https://github.com/vector-im/element-android/issues/3033) as this has crossover requirements in order to replicate.
1.0
Vertical video calls from some element-android devices are distorted - During video calls between element-web and element-android users, the element-web user will see an extremely distorted picture if the element-android's device is oriented vertically. If the android user rotates their phone horizontally, the distortion goes away. This behavior has been tested on five phones: Pixel 3a with stock Android 11: DISTORTED Pixel 3 with CalyxOS Android 11: DISTORTED Pixel 4a with stock Android 11: Not distorted Google Pixel (gen1) with stock Android 10: Not distorted Unihertz Jelly2 with stock Android 10: Not distorted Testing was performed with a combination of accounts, homed to matrix.org and mssj.me. mssj.me is currently running Synapse 1.29.0. Android apps were installed from either the Google Play store or F-Droid, and are up to date with their respective repositories. an example of the distortion when video is in vertical mode: ![Screenshot_20210319102148](https://user-images.githubusercontent.com/8409433/111821479-f108de80-889f-11eb-81c1-377233fef4c0.png) This is a duplicate of [this issue for element-android](https://github.com/vector-im/element-android/issues/3033) as this has crossover requirements in order to replicate.
defect
vertical video calls from some element android devices are distorted during video calls between element web and element android users the element web user will see an extremely distorted picture if the element android s device is oriented vertically if the android user rotates their phone horizontally the distortion goes away this behavior has been tested on five phones pixel with stock android distorted pixel with calyxos android distorted pixel with stock android not distorted google pixel with stock android not distorted unihertz with stock android not distorted testing was performed with a combination of accounts homed to matrix org and mssj me mssj me is currently running synapse android apps were installed from either the google play store or f droid and are up to date with their respective repositories an example of the distortion when video is in vertical mode this is a duplicate of as this has crossover requirements in order to replicate
1
78,710
27,726,693,177
IssuesEvent
2023-03-15 03:11:12
NREL/EnergyPlus
https://api.github.com/repos/NREL/EnergyPlus
closed
Kiva: Index out of bounds possible for certain offset floor polygons
Defect
Issue overview -------------- When using the detailed exposed perimeter method on concave foundation floor shapes, the offset is not guaranteed to have the same number of vertices as the original polygon. As a result, one loop in the source code may attempt to access memory beyond the size of the polygon with undefined results. Updating to Kiva 0.6.6 will address this issue. ### Details Some additional details for this issue (if relevant): - Platform (Operating system, version) - Version of EnergyPlus (if using an intermediate build, include SHA) - Unmethours link or helpdesk ticket number ### Checklist Add to this list or remove from it as applicable. This is a simple templated set of guidelines. - [ ] Defect file added (list location of defect file here) - [ ] Ticket added to Pivotal for defect (development team task) - [ ] Pull request created (the pull request will have additional tasks related to reviewing changes that fix this defect)
1.0
Kiva: Index out of bounds possible for certain offset floor polygons - Issue overview -------------- When using the detailed exposed perimeter method on concave foundation floor shapes, the offset is not guaranteed to have the same number of vertices as the original polygon. As a result, one loop in the source code may attempt to access memory beyond the size of the polygon with undefined results. Updating to Kiva 0.6.6 will address this issue. ### Details Some additional details for this issue (if relevant): - Platform (Operating system, version) - Version of EnergyPlus (if using an intermediate build, include SHA) - Unmethours link or helpdesk ticket number ### Checklist Add to this list or remove from it as applicable. This is a simple templated set of guidelines. - [ ] Defect file added (list location of defect file here) - [ ] Ticket added to Pivotal for defect (development team task) - [ ] Pull request created (the pull request will have additional tasks related to reviewing changes that fix this defect)
defect
kiva index out of bounds possible for certain offset floor polygons issue overview when using the detailed exposed perimeter method on concave foundation floor shapes the offset is not guaranteed to have the same number of vertices as the original polygon as a result one loop in the source code may attempt to access memory beyond the size of the polygon with undefined results updating to kiva will address this issue details some additional details for this issue if relevant platform operating system version version of energyplus if using an intermediate build include sha unmethours link or helpdesk ticket number checklist add to this list or remove from it as applicable this is a simple templated set of guidelines defect file added list location of defect file here ticket added to pivotal for defect development team task pull request created the pull request will have additional tasks related to reviewing changes that fix this defect
1
65,815
19,703,060,553
IssuesEvent
2022-01-12 18:38:35
department-of-veterans-affairs/va.gov-cms
https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms
closed
Do not allow empty media references
Defect Drupal engineering Content governance Media Content model Needs refining Sitewide CMS Team
## User Story or Problem Statement As an editor of a page with content blocks, i shouldn't be able to add empty media references. Related FE ticket to be more defensive: [forthcoming] <details><summary>FE when form is not filled out</summary> <img width="1580" alt="Cursor_and_West_Los_Angeles_Healthcare_Center___VA_Greater_Los_Angeles_Health_Care___Veterans_Affairs" src="https://user-images.githubusercontent.com/643678/145836911-667ee70b-3ff3-4490-9f56-a03093c77330.png"> </details> <details><summary>Form showing field that should be required</summary> <img width="844" alt="Edit_VAMC_Detail_Page_West_Los_Angeles_Healthcare_Center___VA_gov_CMS" src="https://user-images.githubusercontent.com/643678/145837209-c4aafada-f4ee-41be-a192-19f12ea70270.png"> </details> ## Acceptance Criteria - [ ] Media field is required on content blocks ## Implementation steps <optional> ## Design principles Veteran-centered - [ ] `Single source of truth`: Increase reliability and consistency of content on VA.gov by providing a single source of truth. - [ ] `Accessible, plain language`: Provide guardrails and guidelines to ensure content quality. - [x] `Purposely structured content`: Ensure Content API can deliver content whose meaning matches its structure. - [x] `Content lifecycle governance`: Produce tools, processes and policies to maintain content quality throughout its lifecycle. Editor-centered - [ ] `Purpose-driven`: Create an opportunity to involve the editor community in VA’s mission and content strategy goals. - [ ] `Efficient`: Remove distractions and create clear, straightforward paths to get the job done. - [ ] `Approachable`: Offer friendly guidance over authoritative instruction. - [x] `Consistent`: Reduce user’s mental load by allowing them to fall back on pattern recognition to complete tasks. - [ ] `Empowering`: Provide clear information to help editors make decisions about their work. ### CMS Team Please leave only the team that will do this work selected. If you're not sure, it's fine to leave both selected. - [ ] `Platform CMS Team` - [x] `Sitewide CMS Team`
1.0
Do not allow empty media references - ## User Story or Problem Statement As an editor of a page with content blocks, i shouldn't be able to add empty media references. Related FE ticket to be more defensive: [forthcoming] <details><summary>FE when form is not filled out</summary> <img width="1580" alt="Cursor_and_West_Los_Angeles_Healthcare_Center___VA_Greater_Los_Angeles_Health_Care___Veterans_Affairs" src="https://user-images.githubusercontent.com/643678/145836911-667ee70b-3ff3-4490-9f56-a03093c77330.png"> </details> <details><summary>Form showing field that should be required</summary> <img width="844" alt="Edit_VAMC_Detail_Page_West_Los_Angeles_Healthcare_Center___VA_gov_CMS" src="https://user-images.githubusercontent.com/643678/145837209-c4aafada-f4ee-41be-a192-19f12ea70270.png"> </details> ## Acceptance Criteria - [ ] Media field is required on content blocks ## Implementation steps <optional> ## Design principles Veteran-centered - [ ] `Single source of truth`: Increase reliability and consistency of content on VA.gov by providing a single source of truth. - [ ] `Accessible, plain language`: Provide guardrails and guidelines to ensure content quality. - [x] `Purposely structured content`: Ensure Content API can deliver content whose meaning matches its structure. - [x] `Content lifecycle governance`: Produce tools, processes and policies to maintain content quality throughout its lifecycle. Editor-centered - [ ] `Purpose-driven`: Create an opportunity to involve the editor community in VA’s mission and content strategy goals. - [ ] `Efficient`: Remove distractions and create clear, straightforward paths to get the job done. - [ ] `Approachable`: Offer friendly guidance over authoritative instruction. - [x] `Consistent`: Reduce user’s mental load by allowing them to fall back on pattern recognition to complete tasks. - [ ] `Empowering`: Provide clear information to help editors make decisions about their work. ### CMS Team Please leave only the team that will do this work selected. If you're not sure, it's fine to leave both selected. - [ ] `Platform CMS Team` - [x] `Sitewide CMS Team`
defect
do not allow empty media references user story or problem statement as an editor of a page with content blocks i shouldn t be able to add empty media references related fe ticket to be more defensive fe when form is not filled out img width alt cursor and west los angeles healthcare center va greater los angeles health care veterans affairs src form showing field that should be required img width alt edit vamc detail page west los angeles healthcare center va gov cms src acceptance criteria media field is required on content blocks implementation steps design principles veteran centered single source of truth increase reliability and consistency of content on va gov by providing a single source of truth accessible plain language provide guardrails and guidelines to ensure content quality purposely structured content ensure content api can deliver content whose meaning matches its structure content lifecycle governance produce tools processes and policies to maintain content quality throughout its lifecycle editor centered purpose driven create an opportunity to involve the editor community in va’s mission and content strategy goals efficient remove distractions and create clear straightforward paths to get the job done approachable offer friendly guidance over authoritative instruction consistent reduce user’s mental load by allowing them to fall back on pattern recognition to complete tasks empowering provide clear information to help editors make decisions about their work cms team please leave only the team that will do this work selected if you re not sure it s fine to leave both selected platform cms team sitewide cms team
1
188,739
14,451,304,006
IssuesEvent
2020-12-08 10:47:05
open-traffic-generator/ixnetwork
https://api.github.com/repos/open-traffic-generator/ixnetwork
closed
Add test for UDP header with Random Pattern for UDP src,dst port
test
- Add new test script inside tests/udp/ - For each pattern type in Random - Provide non-default counter pattern values for src port, dst port - Validate that src port, dst port are as expected from captured frame
1.0
Add test for UDP header with Random Pattern for UDP src,dst port - - Add new test script inside tests/udp/ - For each pattern type in Random - Provide non-default counter pattern values for src port, dst port - Validate that src port, dst port are as expected from captured frame
non_defect
add test for udp header with random pattern for udp src dst port add new test script inside tests udp for each pattern type in random provide non default counter pattern values for src port dst port validate that src port dst port are as expected from captured frame
0
49,719
26,274,334,429
IssuesEvent
2023-01-06 20:16:55
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
kv: apply log entries outside of raft state machine loop
C-performance A-kv-replication T-kv-replication
Extracted from #17500. After #94165, raft log entry disk writes are asynchronous with respect to the raft state machine loop. However, the (non-durable) engine access for state machine application is still performed inline. The [async storage writes interface](https://github.com/etcd-io/raft/pull/8) permits us to extract all of this work onto a separate goroutine. This would provide three benefits: 1. faster state machine loop iteration => less interference between entries => lower latency 2. larger apply batches => more efficient state machine application => higher throughput 3. flexible scheduling permits deferred application on followers => bigger batches, see benefit 2
True
kv: apply log entries outside of raft state machine loop - Extracted from #17500. After #94165, raft log entry disk writes are asynchronous with respect to the raft state machine loop. However, the (non-durable) engine access for state machine application is still performed inline. The [async storage writes interface](https://github.com/etcd-io/raft/pull/8) permits us to extract all of this work onto a separate goroutine. This would provide three benefits: 1. faster state machine loop iteration => less interference between entries => lower latency 2. larger apply batches => more efficient state machine application => higher throughput 3. flexible scheduling permits deferred application on followers => bigger batches, see benefit 2
non_defect
kv apply log entries outside of raft state machine loop extracted from after raft log entry disk writes are asynchronous with respect to the raft state machine loop however the non durable engine access for state machine application is still performed inline the permits us to extract all of this work onto a separate goroutine this would provide three benefits faster state machine loop iteration less interference between entries lower latency larger apply batches more efficient state machine application higher throughput flexible scheduling permits deferred application on followers bigger batches see benefit
0
56,037
14,910,199,138
IssuesEvent
2021-01-22 09:14:24
jOOQ/jOOQ
https://api.github.com/repos/jOOQ/jOOQ
closed
Add native support for CREATE INDEX .. INCLUDE in Db2
C: DB: DB2 C: Functionality E: Enterprise Edition P: Medium T: Defect
It appears that Db2 has always supported the `INCLUDE` keyword on indexes: https://www.ibm.com/support/knowledgecenter/SSEPGG_9.7.0/com.ibm.db2.luw.sql.ref.doc/doc/r0000919.html We're currently emulating it for all Db2 versions.
1.0
Add native support for CREATE INDEX .. INCLUDE in Db2 - It appears that Db2 has always supported the `INCLUDE` keyword on indexes: https://www.ibm.com/support/knowledgecenter/SSEPGG_9.7.0/com.ibm.db2.luw.sql.ref.doc/doc/r0000919.html We're currently emulating it for all Db2 versions.
defect
add native support for create index include in it appears that has always supported the include keyword on indexes we re currently emulating it for all versions
1
27,112
4,875,506,328
IssuesEvent
2016-11-16 09:48:21
TNGSB/eWallet
https://api.github.com/repos/TNGSB/eWallet
closed
eWallet_MobileApp_On the Home page there are no games available for download yet after selecting the Airtime & Game module # 61
Defect - Medium (Sev-3)
On the home page after clicking the “Airtime & Game” tab it was observed that the Airtime & Game module does not have any games available in there for download
1.0
eWallet_MobileApp_On the Home page there are no games available for download yet after selecting the Airtime & Game module # 61 - On the home page after clicking the “Airtime & Game” tab it was observed that the Airtime & Game module does not have any games available in there for download
defect
ewallet mobileapp on the home page there are no games available for download yet after selecting the airtime game module on the home page after clicking the “airtime game” tab it was observed that the airtime game module does not have any games available in there for download
1
81,496
30,880,125,145
IssuesEvent
2023-08-03 16:55:33
department-of-veterans-affairs/va.gov-team
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
opened
[Assistive tech and device support] Alert and status messages aren't announced without receiving focus. (11.22.1)
accessibility a11y-defect-3 collab-cycle-feedback Staging CCIssue11.22 CC-Dashboard forms-audit-digitize 21P-0847
### General Information #### VFS team name Veteran Facing Forms #### VFS product name 21P-0847 #### VFS feature name #### Point of Contact/Reviewers Brian DeConinck - @briandeconinck - Accessibility *For more information on how to interpret this ticket, please refer to the [Anatomy of a Staging Review issue ticket](https://depo-platform-documentation.scrollhelp.site/collaboration-cycle/Anatomy-of-a-Staging-Review-Issue-ticket.2060320997.html) guidance on Platform Website. --- ### Platform Issue Alert and status messages aren't announced without receiving focus. ### Issue Details When using the new memorable date component (eg. [/deceased-claimant-personal-information](https://staging.va.gov/supporting-forms-for-claims/substitute-claimant-form-21P-0847/deceased-claimant-personal-information)), the day and year inputs have hint text beneath the input indicating the maximum number of characters for each field. The hint text isn't announced when I test with a screen reader. ### Link, screenshot or steps to recreate Steps to reproduce: 1. Navigate through the form flow to a page using the memorable date component, eg. [/deceased-claimant-personal-information](https://staging.va.gov/supporting-forms-for-claims/substitute-claimant-form-21P-0847/deceased-claimant-personal-information). 2. Activate a screen reader for testing (I used NVDA on Windows). 3. Tab to the day input. 4. Type in a number. 5. Note that although "Max. 2 characters" appears on screen, the screen reader does not announce that text. ### VA.gov Experience Standard [Category Number 11, Issue Number 22](https://depo-platform-documentation.scrollhelp.site/collaboration-cycle/VA.gov-experience-standards.1683980311.html) ### Other References WCAG SC 4.1.3_AA ### Platform Recommendation I think the implementation here is almost perfect, with one small issue. The max character text appears in a span with `aria-live="polite"`, which tells screen readers to watch that span for any changes and to announce the change. But the element doesn't exist on page load, it's added as you type your response. Screen readers are only watching for *changes to the contents of the element*, and the addition of the span itself doesn't trigger anything --- even with the `aria-live` attribute. Recommended fix is for the span to be present but empty on page load, then the text gets added dynamically. As an `aria-live` region that should trigger screen readers to read out the max character text. Alternatively, is there any reason why that text isn't just there on page load? That could save some JS and avoid any ARIA issues entirely. --- ### VFS Guidance - Close the ticket when the issue has been resolved or validated by your Product Owner - If your team has additional questions or needs Platform help validating the issue, please comment on the ticket - Some feedback provided may be out of scope for your iteration of the product, however, Platform's OCTO leadership has stated that all identified issues need to be documented and it is still your responsibility to resolve the issue. - If you do not believe that this Staging Review issue ticket is the responsibility of your team, comment below providing an explanation and who you believe is responsible. Please tag the Point of Contact/Reviewers. Governance team will research and will follow up.
1.0
[Assistive tech and device support] Alert and status messages aren't announced without receiving focus. (11.22.1) - ### General Information #### VFS team name Veteran Facing Forms #### VFS product name 21P-0847 #### VFS feature name #### Point of Contact/Reviewers Brian DeConinck - @briandeconinck - Accessibility *For more information on how to interpret this ticket, please refer to the [Anatomy of a Staging Review issue ticket](https://depo-platform-documentation.scrollhelp.site/collaboration-cycle/Anatomy-of-a-Staging-Review-Issue-ticket.2060320997.html) guidance on Platform Website. --- ### Platform Issue Alert and status messages aren't announced without receiving focus. ### Issue Details When using the new memorable date component (eg. [/deceased-claimant-personal-information](https://staging.va.gov/supporting-forms-for-claims/substitute-claimant-form-21P-0847/deceased-claimant-personal-information)), the day and year inputs have hint text beneath the input indicating the maximum number of characters for each field. The hint text isn't announced when I test with a screen reader. ### Link, screenshot or steps to recreate Steps to reproduce: 1. Navigate through the form flow to a page using the memorable date component, eg. [/deceased-claimant-personal-information](https://staging.va.gov/supporting-forms-for-claims/substitute-claimant-form-21P-0847/deceased-claimant-personal-information). 2. Activate a screen reader for testing (I used NVDA on Windows). 3. Tab to the day input. 4. Type in a number. 5. Note that although "Max. 2 characters" appears on screen, the screen reader does not announce that text. ### VA.gov Experience Standard [Category Number 11, Issue Number 22](https://depo-platform-documentation.scrollhelp.site/collaboration-cycle/VA.gov-experience-standards.1683980311.html) ### Other References WCAG SC 4.1.3_AA ### Platform Recommendation I think the implementation here is almost perfect, with one small issue. The max character text appears in a span with `aria-live="polite"`, which tells screen readers to watch that span for any changes and to announce the change. But the element doesn't exist on page load, it's added as you type your response. Screen readers are only watching for *changes to the contents of the element*, and the addition of the span itself doesn't trigger anything --- even with the `aria-live` attribute. Recommended fix is for the span to be present but empty on page load, then the text gets added dynamically. As an `aria-live` region that should trigger screen readers to read out the max character text. Alternatively, is there any reason why that text isn't just there on page load? That could save some JS and avoid any ARIA issues entirely. --- ### VFS Guidance - Close the ticket when the issue has been resolved or validated by your Product Owner - If your team has additional questions or needs Platform help validating the issue, please comment on the ticket - Some feedback provided may be out of scope for your iteration of the product, however, Platform's OCTO leadership has stated that all identified issues need to be documented and it is still your responsibility to resolve the issue. - If you do not believe that this Staging Review issue ticket is the responsibility of your team, comment below providing an explanation and who you believe is responsible. Please tag the Point of Contact/Reviewers. Governance team will research and will follow up.
defect
alert and status messages aren t announced without receiving focus general information vfs team name veteran facing forms vfs product name vfs feature name point of contact reviewers brian deconinck briandeconinck accessibility for more information on how to interpret this ticket please refer to the guidance on platform website platform issue alert and status messages aren t announced without receiving focus issue details when using the new memorable date component eg the day and year inputs have hint text beneath the input indicating the maximum number of characters for each field the hint text isn t announced when i test with a screen reader link screenshot or steps to recreate steps to reproduce navigate through the form flow to a page using the memorable date component eg activate a screen reader for testing i used nvda on windows tab to the day input type in a number note that although max characters appears on screen the screen reader does not announce that text va gov experience standard other references wcag sc aa platform recommendation i think the implementation here is almost perfect with one small issue the max character text appears in a span with aria live polite which tells screen readers to watch that span for any changes and to announce the change but the element doesn t exist on page load it s added as you type your response screen readers are only watching for changes to the contents of the element and the addition of the span itself doesn t trigger anything even with the aria live attribute recommended fix is for the span to be present but empty on page load then the text gets added dynamically as an aria live region that should trigger screen readers to read out the max character text alternatively is there any reason why that text isn t just there on page load that could save some js and avoid any aria issues entirely vfs guidance close the ticket when the issue has been resolved or validated by your product owner if your team has additional questions or needs platform help validating the issue please comment on the ticket some feedback provided may be out of scope for your iteration of the product however platform s octo leadership has stated that all identified issues need to be documented and it is still your responsibility to resolve the issue if you do not believe that this staging review issue ticket is the responsibility of your team comment below providing an explanation and who you believe is responsible please tag the point of contact reviewers governance team will research and will follow up
1
30,030
5,971,372,744
IssuesEvent
2017-05-31 02:14:00
kaneless/mybatisnet
https://api.github.com/repos/kaneless/mybatisnet
closed
Support for inclusion of <result> fragments into <resultMap> elements
auto-migrated Priority-Low Type-Defect
``` What version of the MyBatis.NET are you using? IBatis.DataMapper.1.6.2 Please describe the problem. Unit tests are best! This is an enhancement request. I would like iBATIS to support <include/> kind of functionality for the <resultMap/> and <parameterMap/> elements -- Something similar to what iBATIS already provides for <statement/> tags family via <sql/> and <include/> tags. What is the expected output? What do you see instead? Please provide any additional information below. There are use cases where I have to map the results of different SELECT queries into objects which do not form any inheritance hierarchy, though these objects do share some properties (and unfortunately, their inheritance line cannot be changed), and this leads to a lot of repetition of <result/> elements in different <resultMap/>s. If I could have a feature similar to <sql/> and <include/> working for result/parameter maps as well, then it will really help in keeping the XML concise. ``` Original issue reported on code.google.com by `sodhi.b@gmail.com` on 7 Aug 2010 at 2:34
1.0
Support for inclusion of <result> fragments into <resultMap> elements - ``` What version of the MyBatis.NET are you using? IBatis.DataMapper.1.6.2 Please describe the problem. Unit tests are best! This is an enhancement request. I would like iBATIS to support <include/> kind of functionality for the <resultMap/> and <parameterMap/> elements -- Something similar to what iBATIS already provides for <statement/> tags family via <sql/> and <include/> tags. What is the expected output? What do you see instead? Please provide any additional information below. There are use cases where I have to map the results of different SELECT queries into objects which do not form any inheritance hierarchy, though these objects do share some properties (and unfortunately, their inheritance line cannot be changed), and this leads to a lot of repetition of <result/> elements in different <resultMap/>s. If I could have a feature similar to <sql/> and <include/> working for result/parameter maps as well, then it will really help in keeping the XML concise. ``` Original issue reported on code.google.com by `sodhi.b@gmail.com` on 7 Aug 2010 at 2:34
defect
support for inclusion of fragments into elements what version of the mybatis net are you using ibatis datamapper please describe the problem unit tests are best this is an enhancement request i would like ibatis to support kind of functionality for the and elements something similar to what ibatis already provides for tags family via and tags what is the expected output what do you see instead please provide any additional information below there are use cases where i have to map the results of different select queries into objects which do not form any inheritance hierarchy though these objects do share some properties and unfortunately their inheritance line cannot be changed and this leads to a lot of repetition of elements in different s if i could have a feature similar to and working for result parameter maps as well then it will really help in keeping the xml concise original issue reported on code google com by sodhi b gmail com on aug at
1
1,031
2,594,453,965
IssuesEvent
2015-02-20 03:38:25
BALL-Project/ball
https://api.github.com/repos/BALL-Project/ball
opened
BALLView crashes under Linux on Intel hardware
C: VIEW P: blocker T: defect
**Reported by dstoeckel on 17 Jan 43139740 19:24 UTC** BALLView crashes under Ubuntu 12.04 on Intel hardware. This seems to be a race condition in the OpenGL initialization as sometimes it will start, but produce the output: ``` Mesa 8.0.4 implementation error: meta program compile failed: source: #version 130 attribute vec4 position; void main() { gl_Position = position; } Please report at bugs.freedesktop.org Mesa 8.0.4 implementation error: meta program compile failed: source: #version 130 uniform ivec4 color; out ivec4 out_color; void main() { out_color = color; } Please report at bugs.freedesktop.org ``` Notice the encoding error at the very top. Here is the ouput of lspci: ``` 00:02.0 VGA compatible controller: Intel Corporation Ivy Bridge Graphics Controller (rev 09) (prog-if 00 [controller](VGA)) Subsystem: Dell Device 052c Flags: bus master, fast devsel, latency 0, IRQ 45 Memory at f7800000 (64-bit, non-prefetchable) [ Memory at e0000000 (64-bit, prefetchable) [size=256M](size=4M] ) I/O ports at f000 [ Expansion ROM at <unassigned> [disabled](size=64] ) Capabilities: [MSI: Enable+ Count=1/1 Maskable- 64bit- Capabilities: [d0](90]) Power Management version 2 Capabilities: [a4] PCI Advanced Features Kernel driver in use: i915 Kernel modules: i915 ``` The graphics card/cpu is an Intel(R) Core(TM) i7-3770. The attachment is the output of glxinfo.
1.0
BALLView crashes under Linux on Intel hardware - **Reported by dstoeckel on 17 Jan 43139740 19:24 UTC** BALLView crashes under Ubuntu 12.04 on Intel hardware. This seems to be a race condition in the OpenGL initialization as sometimes it will start, but produce the output: ``` Mesa 8.0.4 implementation error: meta program compile failed: source: #version 130 attribute vec4 position; void main() { gl_Position = position; } Please report at bugs.freedesktop.org Mesa 8.0.4 implementation error: meta program compile failed: source: #version 130 uniform ivec4 color; out ivec4 out_color; void main() { out_color = color; } Please report at bugs.freedesktop.org ``` Notice the encoding error at the very top. Here is the ouput of lspci: ``` 00:02.0 VGA compatible controller: Intel Corporation Ivy Bridge Graphics Controller (rev 09) (prog-if 00 [controller](VGA)) Subsystem: Dell Device 052c Flags: bus master, fast devsel, latency 0, IRQ 45 Memory at f7800000 (64-bit, non-prefetchable) [ Memory at e0000000 (64-bit, prefetchable) [size=256M](size=4M] ) I/O ports at f000 [ Expansion ROM at <unassigned> [disabled](size=64] ) Capabilities: [MSI: Enable+ Count=1/1 Maskable- 64bit- Capabilities: [d0](90]) Power Management version 2 Capabilities: [a4] PCI Advanced Features Kernel driver in use: i915 Kernel modules: i915 ``` The graphics card/cpu is an Intel(R) Core(TM) i7-3770. The attachment is the output of glxinfo.
defect
ballview crashes under linux on intel hardware reported by dstoeckel on jan utc ballview crashes under ubuntu on intel hardware this seems to be a race condition in the opengl initialization as sometimes it will start but produce the output mesa implementation error meta program compile failed source version attribute position void main gl position position please report at bugs freedesktop org mesa implementation error meta program compile failed source version uniform color out out color void main out color color please report at bugs freedesktop org notice the encoding error at the very top here is the ouput of lspci vga compatible controller intel corporation ivy bridge graphics controller rev prog if vga subsystem dell device flags bus master fast devsel latency irq memory at bit non prefetchable size i o ports at size capabilities msi enable count maskable capabilities power management version capabilities pci advanced features kernel driver in use kernel modules the graphics card cpu is an intel r core tm the attachment is the output of glxinfo
1
46,841
2,965,111,038
IssuesEvent
2015-07-10 20:39:31
IQSS/dataverse
https://api.github.com/repos/IQSS/dataverse
opened
My Data: When granted access to restricted file, the file does not appear in My Data as file downloader.
Component: UX & Upgrade Priority: High Status: Dev Type: Bug
When granted access to restricted file, the file does not appear in My Data as file downloader. 1. Create restricted file 2. Request access by another user 3. Grant access 4. View My Data, notice no files are present.
1.0
My Data: When granted access to restricted file, the file does not appear in My Data as file downloader. - When granted access to restricted file, the file does not appear in My Data as file downloader. 1. Create restricted file 2. Request access by another user 3. Grant access 4. View My Data, notice no files are present.
non_defect
my data when granted access to restricted file the file does not appear in my data as file downloader when granted access to restricted file the file does not appear in my data as file downloader create restricted file request access by another user grant access view my data notice no files are present
0
48,765
20,255,133,394
IssuesEvent
2022-02-14 22:11:52
Azure/azure-sdk-for-python
https://api.github.com/repos/Azure/azure-sdk-for-python
closed
Python SDK circular import error
bug Storage Service Attention Client customer-reported needs-team-attention
I am trying to make a connection to the azure queue in Python but I am getting a circular import error. Here's my code: `from azure.storage.queue import (QueueClient) connection_str = "", q_name = "test2" queue_client = QueueClient.from_connection_string(connection_str, q_name) queue_client.create_queue()` Error I am getting: `from azure.storage.queue import (QueueClient) ImportError: cannot import name 'QueueClient' from partially initialized module 'azure.storage.queue' (most likely due to a circular import) (/home/deepak/azure-env/lib/python3.8/site-packages/azure/storage/queue/__init__.py)`
1.0
Python SDK circular import error - I am trying to make a connection to the azure queue in Python but I am getting a circular import error. Here's my code: `from azure.storage.queue import (QueueClient) connection_str = "", q_name = "test2" queue_client = QueueClient.from_connection_string(connection_str, q_name) queue_client.create_queue()` Error I am getting: `from azure.storage.queue import (QueueClient) ImportError: cannot import name 'QueueClient' from partially initialized module 'azure.storage.queue' (most likely due to a circular import) (/home/deepak/azure-env/lib/python3.8/site-packages/azure/storage/queue/__init__.py)`
non_defect
python sdk circular import error i am trying to make a connection to the azure queue in python but i am getting a circular import error here s my code from azure storage queue import queueclient connection str q name queue client queueclient from connection string connection str q name queue client create queue error i am getting from azure storage queue import queueclient importerror cannot import name queueclient from partially initialized module azure storage queue most likely due to a circular import home deepak azure env lib site packages azure storage queue init py
0
744,683
25,951,513,162
IssuesEvent
2022-12-17 17:12:00
themotte/rDrama
https://api.github.com/repos/themotte/rDrama
reopened
Solve the Blocking Weaponization Issue
P2 priority
Right now the block page says "They cannot reply to your content or notify you with a username mention". I don't know if this is accurate. Check to see if it's accurate! If it is, I think we should change it to eliminate notifications, both username mentions and reply mentions, and maybe to hide the blockee from the blocker's view. This makes it less of a "block" and more of a "hide" option, but that's, y'know, kind of what we want. But don't actually restrict posting - you shouldn't be able to prevent someone from posting. If you have a better idea, though, let me know. I think it is unlikely that we run into trouble with this immediately but it should probably be fixed at some point.
1.0
Solve the Blocking Weaponization Issue - Right now the block page says "They cannot reply to your content or notify you with a username mention". I don't know if this is accurate. Check to see if it's accurate! If it is, I think we should change it to eliminate notifications, both username mentions and reply mentions, and maybe to hide the blockee from the blocker's view. This makes it less of a "block" and more of a "hide" option, but that's, y'know, kind of what we want. But don't actually restrict posting - you shouldn't be able to prevent someone from posting. If you have a better idea, though, let me know. I think it is unlikely that we run into trouble with this immediately but it should probably be fixed at some point.
non_defect
solve the blocking weaponization issue right now the block page says they cannot reply to your content or notify you with a username mention i don t know if this is accurate check to see if it s accurate if it is i think we should change it to eliminate notifications both username mentions and reply mentions and maybe to hide the blockee from the blocker s view this makes it less of a block and more of a hide option but that s y know kind of what we want but don t actually restrict posting you shouldn t be able to prevent someone from posting if you have a better idea though let me know i think it is unlikely that we run into trouble with this immediately but it should probably be fixed at some point
0
57,734
8,203,904,552
IssuesEvent
2018-09-03 02:40:17
solbu/hldig
https://api.github.com/repos/solbu/hldig
closed
htdig: rename to hldig
documentation pri: 3 (normal) wip
- [x] htdig/ needs to be renamed - [x] text inside source code needs to be change - [x] htdig.* files need to be renamed - [x] update docs - [x] change references to "htdig" on the [web site](https://solbu.github.io/hldig/) (#50)
1.0
htdig: rename to hldig - - [x] htdig/ needs to be renamed - [x] text inside source code needs to be change - [x] htdig.* files need to be renamed - [x] update docs - [x] change references to "htdig" on the [web site](https://solbu.github.io/hldig/) (#50)
non_defect
htdig rename to hldig htdig needs to be renamed text inside source code needs to be change htdig files need to be renamed update docs change references to htdig on the
0
64,906
18,959,061,384
IssuesEvent
2021-11-19 00:52:24
jccastillo0007/eFacturaT
https://api.github.com/repos/jccastillo0007/eFacturaT
opened
Conector Carta Porte 2.0 - Figura Transporte
bug defect
No está reconociendo esta sección de **Partes transporte**: //FiguraTransporte={TipoFigura|RFCFigura|NumLicencia|NombreFigura|NumRegIdTribFigura|ResidenciaFiscalFigura} //PartesTransporte={parteTransporte|calle|noExt|noInt|colonia|localidad|referencia|municipio|estado|pais|CP},{},...{} FiguraTransporte=01|CAAJ7506072P2|FED1234|ADRIAN MORENO GARCIA|| **PartesTransporte={|Insurgentes|1000|Depto C-503|0388|03||039|JAL|MEX|44770}** Partes transporte, es un subnodo (opcional) que puede llevar un solo atributo y a la vez, puede o no incluir el subnodo domicilio. Hay que considerar la etiqueta, solo como Figura, en lugar de FiguraTransporte
1.0
Conector Carta Porte 2.0 - Figura Transporte - No está reconociendo esta sección de **Partes transporte**: //FiguraTransporte={TipoFigura|RFCFigura|NumLicencia|NombreFigura|NumRegIdTribFigura|ResidenciaFiscalFigura} //PartesTransporte={parteTransporte|calle|noExt|noInt|colonia|localidad|referencia|municipio|estado|pais|CP},{},...{} FiguraTransporte=01|CAAJ7506072P2|FED1234|ADRIAN MORENO GARCIA|| **PartesTransporte={|Insurgentes|1000|Depto C-503|0388|03||039|JAL|MEX|44770}** Partes transporte, es un subnodo (opcional) que puede llevar un solo atributo y a la vez, puede o no incluir el subnodo domicilio. Hay que considerar la etiqueta, solo como Figura, en lugar de FiguraTransporte
defect
conector carta porte figura transporte no está reconociendo esta sección de partes transporte figuratransporte tipofigura rfcfigura numlicencia nombrefigura numregidtribfigura residenciafiscalfigura partestransporte partetransporte calle noext noint colonia localidad referencia municipio estado pais cp figuratransporte adrian moreno garcia partestransporte insurgentes depto c jal mex partes transporte es un subnodo opcional que puede llevar un solo atributo y a la vez puede o no incluir el subnodo domicilio hay que considerar la etiqueta solo como figura en lugar de figuratransporte
1
70,871
23,349,418,181
IssuesEvent
2022-08-09 21:31:06
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
Element keeps crashing when I click "I am the host" in Jitsi widget
T-Defect
### Steps to reproduce We have self-hosted Jitsi server. This server is specified in .well-known/matrix/client. So, when I create a video-chat with phone, element uses our self hosted link. On jitsi server we have enabled auth. When you create a video-chat you recieve a message like "we need to wait a host". You can press "I am the host". You just press this button, enter password and that's it. It used to work. Now, when you press "I am the host" element stops working 1. Create a video conference in element from Android phone. Self hosted jitsi with enabled auth. 2. Enter to this conference. 3. Press "I am the host" 4. Element stops working **_When I press "I am the host" from PC, everything works fine._** ### Outcome #### What did you expect? Enter to video conference as a host #### What happened instead? Element stops working ### Operating system Android ### Application version Element - 1.4.31, OLM - 3.2.12 ### How did you install the app? Google Play ### Homeserver Self-hosted server. server_version "1.64.0" ### Will you send logs? No
1.0
Element keeps crashing when I click "I am the host" in Jitsi widget - ### Steps to reproduce We have self-hosted Jitsi server. This server is specified in .well-known/matrix/client. So, when I create a video-chat with phone, element uses our self hosted link. On jitsi server we have enabled auth. When you create a video-chat you recieve a message like "we need to wait a host". You can press "I am the host". You just press this button, enter password and that's it. It used to work. Now, when you press "I am the host" element stops working 1. Create a video conference in element from Android phone. Self hosted jitsi with enabled auth. 2. Enter to this conference. 3. Press "I am the host" 4. Element stops working **_When I press "I am the host" from PC, everything works fine._** ### Outcome #### What did you expect? Enter to video conference as a host #### What happened instead? Element stops working ### Operating system Android ### Application version Element - 1.4.31, OLM - 3.2.12 ### How did you install the app? Google Play ### Homeserver Self-hosted server. server_version "1.64.0" ### Will you send logs? No
defect
element keeps crashing when i click i am the host in jitsi widget steps to reproduce we have self hosted jitsi server this server is specified in well known matrix client so when i create a video chat with phone element uses our self hosted link on jitsi server we have enabled auth when you create a video chat you recieve a message like we need to wait a host you can press i am the host you just press this button enter password and that s it it used to work now when you press i am the host element stops working create a video conference in element from android phone self hosted jitsi with enabled auth enter to this conference press i am the host element stops working when i press i am the host from pc everything works fine outcome what did you expect enter to video conference as a host what happened instead element stops working operating system android application version element olm how did you install the app google play homeserver self hosted server server version will you send logs no
1
576,044
17,069,665,824
IssuesEvent
2021-07-07 11:47:36
weaveworks/eksctl
https://api.github.com/repos/weaveworks/eksctl
closed
Cluster API controller
area/gitops kind/feature needs-investigation priority/important-longterm
We would like to implement [Cluster API](https://github.com/kubernetes-sigs/cluster-api). One of the modes in which one can use Cluster API is by having a management cluster which runs a controller (actuator), which creates new clusters and nodegroups when it receives Cluster API objects. Another mode of using it would be via CLI (#19).
1.0
Cluster API controller - We would like to implement [Cluster API](https://github.com/kubernetes-sigs/cluster-api). One of the modes in which one can use Cluster API is by having a management cluster which runs a controller (actuator), which creates new clusters and nodegroups when it receives Cluster API objects. Another mode of using it would be via CLI (#19).
non_defect
cluster api controller we would like to implement one of the modes in which one can use cluster api is by having a management cluster which runs a controller actuator which creates new clusters and nodegroups when it receives cluster api objects another mode of using it would be via cli
0
76,493
14,629,812,988
IssuesEvent
2020-12-23 16:30:38
4GeeksAcademy/About-4Geeks-Academy
https://api.github.com/repos/4GeeksAcademy/About-4Geeks-Academy
closed
Website chat
:computer: bc-coder :star: P1 enhancement
Implement the active campaign chat on the website. https://www.activecampaign.com/customer-messaging/conversations The implementation must be done thru google tag manager and the chat script must trigger only when the visitor has English language on the browser.
1.0
Website chat - Implement the active campaign chat on the website. https://www.activecampaign.com/customer-messaging/conversations The implementation must be done thru google tag manager and the chat script must trigger only when the visitor has English language on the browser.
non_defect
website chat implement the active campaign chat on the website the implementation must be done thru google tag manager and the chat script must trigger only when the visitor has english language on the browser
0
39,284
9,376,127,178
IssuesEvent
2019-04-04 07:03:14
martinrotter/rssguard
https://api.github.com/repos/martinrotter/rssguard
closed
[3.5.5] Bug: Can't open context menu in feed list
Component-Feed-List Component-GUI Status-Not-Enough-Data Type-Defect
The context menu (mouse right click) doesn't open in the feed list. It opened before, but now has stopped working. Can't reproduce it for now. #### Brief description of the issue. <!--- Write your description here. Remove this line, pls. --> #### How to reproduce the bug? 1. Don't really know for now... #### What is the expected result? Should open the context menu. #### What actually happened? Nothing happens.
1.0
[3.5.5] Bug: Can't open context menu in feed list - The context menu (mouse right click) doesn't open in the feed list. It opened before, but now has stopped working. Can't reproduce it for now. #### Brief description of the issue. <!--- Write your description here. Remove this line, pls. --> #### How to reproduce the bug? 1. Don't really know for now... #### What is the expected result? Should open the context menu. #### What actually happened? Nothing happens.
defect
bug can t open context menu in feed list the context menu mouse right click doesn t open in the feed list it opened before but now has stopped working can t reproduce it for now brief description of the issue how to reproduce the bug don t really know for now what is the expected result should open the context menu what actually happened nothing happens
1
6,824
2,610,282,765
IssuesEvent
2015-02-26 19:30:20
chrsmith/scribefire-chrome
https://api.github.com/repos/chrsmith/scribefire-chrome
closed
Doesn't work!
auto-migrated Priority-Medium Type-Defect
``` What's the problem? Despite getting the confirmation pop-up that my post had been published, it never appeared at Blogger. What browser are you using? Chrome v 14.0.835.186 What version of ScribeFire are you running? 1.9 ``` ----- Original issue reported on code.google.com by `gsshar...@gmail.com` on 24 Sep 2011 at 10:46
1.0
Doesn't work! - ``` What's the problem? Despite getting the confirmation pop-up that my post had been published, it never appeared at Blogger. What browser are you using? Chrome v 14.0.835.186 What version of ScribeFire are you running? 1.9 ``` ----- Original issue reported on code.google.com by `gsshar...@gmail.com` on 24 Sep 2011 at 10:46
defect
doesn t work what s the problem despite getting the confirmation pop up that my post had been published it never appeared at blogger what browser are you using chrome v what version of scribefire are you running original issue reported on code google com by gsshar gmail com on sep at
1
448,652
12,955,140,085
IssuesEvent
2020-07-20 05:38:23
eAbsentee/eAbsentee
https://api.github.com/repos/eAbsentee/eAbsentee
closed
Disable Flask debug in production
priority
https://github.com/rau/eAbsentee/blob/2a1b5e223284aac3c667ad4edbcd26487fca88dc/app.py#L19 We should use an environment variable to keep track of whether debug mode should be enabled.
1.0
Disable Flask debug in production - https://github.com/rau/eAbsentee/blob/2a1b5e223284aac3c667ad4edbcd26487fca88dc/app.py#L19 We should use an environment variable to keep track of whether debug mode should be enabled.
non_defect
disable flask debug in production we should use an environment variable to keep track of whether debug mode should be enabled
0
13,467
15,951,651,264
IssuesEvent
2021-04-15 10:05:55
unicode-org/icu4x
https://api.github.com/repos/unicode-org/icu4x
opened
Decide on strategy for updating pinned Rust versions
C-process discuss
We now have pinned Stable and Nightly rust versions (#618, #374, etc). However, we don't have a strategy in place to update them and keep them in sync. For the nightly version in particular, we currently have two versions in tree, `nightly-2021-02-28` (required for WebAssembly) and `nightly-2021-03-15` (a more up-to-date version for the coverage and memory benchmarking tools). CC @gregtatum @dminor
1.0
Decide on strategy for updating pinned Rust versions - We now have pinned Stable and Nightly rust versions (#618, #374, etc). However, we don't have a strategy in place to update them and keep them in sync. For the nightly version in particular, we currently have two versions in tree, `nightly-2021-02-28` (required for WebAssembly) and `nightly-2021-03-15` (a more up-to-date version for the coverage and memory benchmarking tools). CC @gregtatum @dminor
non_defect
decide on strategy for updating pinned rust versions we now have pinned stable and nightly rust versions etc however we don t have a strategy in place to update them and keep them in sync for the nightly version in particular we currently have two versions in tree nightly required for webassembly and nightly a more up to date version for the coverage and memory benchmarking tools cc gregtatum dminor
0
72,281
24,033,602,049
IssuesEvent
2022-09-15 17:01:04
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
Icon in systray is labelled «electron»
T-Defect
### Steps to reproduce 1. Launch Element 2. Check Element icon in systray ### Outcome #### What did you expect? The icon in the systray to be labeled Element. #### What happened instead? The icon in the systray is labeled Element. Label when hovering the systray icon: ![image](https://user-images.githubusercontent.com/2446451/190462488-a7f669df-d447-4521-9c4a-530a3fcee0b4.png) Label when the icon is hidden part of the systray: ![image](https://user-images.githubusercontent.com/2446451/190461728-218aa514-dd50-44b7-8b87-9bb654980b7d.png) ### Operating system Arch Linux ### Application version Element version 1.11.4, olm version 3.2.12 ### How did you install the app? https://archlinux.org/packages/community/x86_64/element-desktop/ ### Homeserver _No response_ ### Will you send logs? No
1.0
Icon in systray is labelled «electron» - ### Steps to reproduce 1. Launch Element 2. Check Element icon in systray ### Outcome #### What did you expect? The icon in the systray to be labeled Element. #### What happened instead? The icon in the systray is labeled Element. Label when hovering the systray icon: ![image](https://user-images.githubusercontent.com/2446451/190462488-a7f669df-d447-4521-9c4a-530a3fcee0b4.png) Label when the icon is hidden part of the systray: ![image](https://user-images.githubusercontent.com/2446451/190461728-218aa514-dd50-44b7-8b87-9bb654980b7d.png) ### Operating system Arch Linux ### Application version Element version 1.11.4, olm version 3.2.12 ### How did you install the app? https://archlinux.org/packages/community/x86_64/element-desktop/ ### Homeserver _No response_ ### Will you send logs? No
defect
icon in systray is labelled «electron» steps to reproduce launch element check element icon in systray outcome what did you expect the icon in the systray to be labeled element what happened instead the icon in the systray is labeled element label when hovering the systray icon label when the icon is hidden part of the systray operating system arch linux application version element version olm version how did you install the app homeserver no response will you send logs no
1
365,692
25,548,300,393
IssuesEvent
2022-11-29 20:55:39
celestiaorg/celestia-core
https://api.github.com/repos/celestiaorg/celestia-core
closed
Document release process
C:documentation
## Summary This repo could benefit from a guide on how to cut a new release ## Problem Definition Current release process is to ask @evan-forbes or @liamsi to cut a new release. However this process has a low bus factor. It would be nice if the steps were documented so that any team member with permission is able to self-service cut a new release. ## Proposal Steps documented in a Markdown file (e.g. `README.md`) that describe the release process, who has permission to cut a release, and any non standard release steps. Something like: 1. Create a release in Github UI 1. Upgrade to the new celestia-core version in cosmos-sdk and celestia-app
1.0
Document release process - ## Summary This repo could benefit from a guide on how to cut a new release ## Problem Definition Current release process is to ask @evan-forbes or @liamsi to cut a new release. However this process has a low bus factor. It would be nice if the steps were documented so that any team member with permission is able to self-service cut a new release. ## Proposal Steps documented in a Markdown file (e.g. `README.md`) that describe the release process, who has permission to cut a release, and any non standard release steps. Something like: 1. Create a release in Github UI 1. Upgrade to the new celestia-core version in cosmos-sdk and celestia-app
non_defect
document release process summary this repo could benefit from a guide on how to cut a new release problem definition current release process is to ask evan forbes or liamsi to cut a new release however this process has a low bus factor it would be nice if the steps were documented so that any team member with permission is able to self service cut a new release proposal steps documented in a markdown file e g readme md that describe the release process who has permission to cut a release and any non standard release steps something like create a release in github ui upgrade to the new celestia core version in cosmos sdk and celestia app
0
1,119
27,123,958,625
IssuesEvent
2023-02-16 02:34:46
restincode/restincode
https://api.github.com/repos/restincode/restincode
opened
Bruce Esquibel (Dr. Ripco)
People Add Person Needs Review
Please fill out as much information as you can. No fields are required, but the more you can provide the better. **General Info** * First name: Bruce * Last name: Esquibel * Handle: Dr. Ripco * Birth Year: * Death Year: 2023 * Link to Obituary: * Group Affiliations: * URL to main photo (or attach to Issue): * Description of person and/or activities: * Facebook memorial group URL: **Social Media Links** * Twitter: * Github: * LinkedIn: * Facebook: * Other: **Contributions** Repeat as many times as needed * Project name: * Project URL: * Project Description: **Photo Gallery** * URL(s) to additional photos: https://twitter.com/textfiles/status/1625878219114127362 / https://archive.ph/cfrlm https://twitter.com/DethVeggie/status/1626016692613812224 / https://archive.ph/n1rxl
1.0
Bruce Esquibel (Dr. Ripco) - Please fill out as much information as you can. No fields are required, but the more you can provide the better. **General Info** * First name: Bruce * Last name: Esquibel * Handle: Dr. Ripco * Birth Year: * Death Year: 2023 * Link to Obituary: * Group Affiliations: * URL to main photo (or attach to Issue): * Description of person and/or activities: * Facebook memorial group URL: **Social Media Links** * Twitter: * Github: * LinkedIn: * Facebook: * Other: **Contributions** Repeat as many times as needed * Project name: * Project URL: * Project Description: **Photo Gallery** * URL(s) to additional photos: https://twitter.com/textfiles/status/1625878219114127362 / https://archive.ph/cfrlm https://twitter.com/DethVeggie/status/1626016692613812224 / https://archive.ph/n1rxl
non_defect
bruce esquibel dr ripco please fill out as much information as you can no fields are required but the more you can provide the better general info first name bruce last name esquibel handle dr ripco birth year death year link to obituary group affiliations url to main photo or attach to issue description of person and or activities facebook memorial group url social media links twitter github linkedin facebook other contributions repeat as many times as needed project name project url project description photo gallery url s to additional photos
0
663,345
22,173,920,591
IssuesEvent
2022-06-06 06:02:22
wso2/api-manager
https://api.github.com/repos/wso2/api-manager
opened
Error occurs when loading API Policies view for Read-Only user
Type/Bug Priority/Normal
### Description Following error appears when logged in to the publisher portal as a read only (internal/observer) user and click on the 'Policies' on left menu. ![Uploading Screenshot 2022-06-06 at 11.17.36.png…]() ### Steps to Reproduce - Sign in to the WSO2 Management Console (https://<APIM_host>:<APIM_port>/carbon) as the admin - Create a REST API - Create a user and assign the observer default role. - Login to the Publisher portal as the newly created observer user - Select the previously created API -> Policies ### Affected Component APIM ### Version 4.1.0 ### Environment Details (with versions) Mac Book Pro M1 chip OS version 12.4 ### Relevant Log Output _No response_ ### Related Issues _No response_ ### Suggested Labels react
1.0
Error occurs when loading API Policies view for Read-Only user - ### Description Following error appears when logged in to the publisher portal as a read only (internal/observer) user and click on the 'Policies' on left menu. ![Uploading Screenshot 2022-06-06 at 11.17.36.png…]() ### Steps to Reproduce - Sign in to the WSO2 Management Console (https://<APIM_host>:<APIM_port>/carbon) as the admin - Create a REST API - Create a user and assign the observer default role. - Login to the Publisher portal as the newly created observer user - Select the previously created API -> Policies ### Affected Component APIM ### Version 4.1.0 ### Environment Details (with versions) Mac Book Pro M1 chip OS version 12.4 ### Relevant Log Output _No response_ ### Related Issues _No response_ ### Suggested Labels react
non_defect
error occurs when loading api policies view for read only user description following error appears when logged in to the publisher portal as a read only internal observer user and click on the policies on left menu steps to reproduce sign in to the management console as the admin create a rest api create a user and assign the observer default role login to the publisher portal as the newly created observer user select the previously created api policies affected component apim version environment details with versions mac book pro chip os version relevant log output no response related issues no response suggested labels react
0
68,214
21,558,283,087
IssuesEvent
2022-04-30 19:57:49
martinrotter/rssguard
https://api.github.com/repos/martinrotter/rssguard
closed
[BUG]: Missing option to hide read items in articles list
Type-Defect
### Brief description of the issue I just updated to v 4.2.2, and the option to control display of already read items in articles list is gone. I tried adding the 'Show unread items only' button back to the 'Toolbar for articles list', but it changes the display of the feed list, not the articles list. ### How to reproduce the bug? 1. Open RSS Guard 2. Select feed 3. The articles list shows both read and unread items 4. Open the 'Articles' menu. There is no option to 'Show unread items only' ### What was the expected result? There should be an option to 'Show unread items only' for the Articles list ### What actually happened? There is no option to 'Show unread items only' ### Debug log I fail to see how a debug log will help with this. I will happily provide if it's actually useful. ### Operating system and version * OS: Windows 10 21H2 (19044.1682) * RSS Guard version: 4.2.2
1.0
[BUG]: Missing option to hide read items in articles list - ### Brief description of the issue I just updated to v 4.2.2, and the option to control display of already read items in articles list is gone. I tried adding the 'Show unread items only' button back to the 'Toolbar for articles list', but it changes the display of the feed list, not the articles list. ### How to reproduce the bug? 1. Open RSS Guard 2. Select feed 3. The articles list shows both read and unread items 4. Open the 'Articles' menu. There is no option to 'Show unread items only' ### What was the expected result? There should be an option to 'Show unread items only' for the Articles list ### What actually happened? There is no option to 'Show unread items only' ### Debug log I fail to see how a debug log will help with this. I will happily provide if it's actually useful. ### Operating system and version * OS: Windows 10 21H2 (19044.1682) * RSS Guard version: 4.2.2
defect
missing option to hide read items in articles list brief description of the issue i just updated to v and the option to control display of already read items in articles list is gone i tried adding the show unread items only button back to the toolbar for articles list but it changes the display of the feed list not the articles list how to reproduce the bug open rss guard select feed the articles list shows both read and unread items open the articles menu there is no option to show unread items only what was the expected result there should be an option to show unread items only for the articles list what actually happened there is no option to show unread items only debug log i fail to see how a debug log will help with this i will happily provide if it s actually useful operating system and version os windows rss guard version
1
239,164
19,825,463,516
IssuesEvent
2022-01-20 05:40:08
apache/pulsar
https://api.github.com/repos/apache/pulsar
opened
Flaky-test: C++ ProducerTest.testSendAsyncCloseAsyncConcurrentlyWithLazyProducers
component/c++ flaky-tests
ProducerTest.testSendAsyncCloseAsyncConcurrentlyWithLazyProducers fails sporadically. [example failure](https://github.com/apache/pulsar/runs/4877678504?check_suite_focus=true#step:11:748) ``` 2022-01-20 05:08:14.347 INFO [139723758618368] ProducerTest:208 | Closing 2022-01-20 05:08:14.347 INFO [139723758618368] ProducerImpl:621 | [persistent://public/default/testProducerIsConnectedPartitioned-1642655293-partition-7, standalone-0-401] Closing producer for topic persistent://public/default/testProducerIsConnectedPartitioned-1642655293-partition-7 2022-01-20 05:08:14.347 INFO [139723758618368] ProducerTest:210 | Close called 2022-01-20 05:08:14.347 INFO [139724043835136] ProducerImpl:664 | [persistent://public/default/testProducerIsConnectedPartitioned-1642655293-partition-7, standalone-0-401] Closed producer 2022-01-20 05:08:14.347 INFO [139723758618368] ProducerTest:215 | Closed /pulsar/pulsar-client-cpp/tests/ProducerTest.cc:236: Failure Value of: res Actual: Ok Expected: ResultAlreadyClosed Which is: AlreadyClosed [ FAILED ] ProducerTest.testSendAsyncCloseAsyncConcurrentlyWithLazyProducers (3431 ms) [----------] 1 test from ProducerTest (3431 ms total) [----------] Global test environment tear-down [==========] 1 test from 1 test case ran. (3435 ms total) [ PASSED ] 0 tests. [ FAILED ] 1 test, listed below: [ FAILED ] ProducerTest.testSendAsyncCloseAsyncConcurrentlyWithLazyProducers ``` full logs at https://gist.github.com/lhotari/f37aefd7919f030399f3a5b7020baa23
1.0
Flaky-test: C++ ProducerTest.testSendAsyncCloseAsyncConcurrentlyWithLazyProducers - ProducerTest.testSendAsyncCloseAsyncConcurrentlyWithLazyProducers fails sporadically. [example failure](https://github.com/apache/pulsar/runs/4877678504?check_suite_focus=true#step:11:748) ``` 2022-01-20 05:08:14.347 INFO [139723758618368] ProducerTest:208 | Closing 2022-01-20 05:08:14.347 INFO [139723758618368] ProducerImpl:621 | [persistent://public/default/testProducerIsConnectedPartitioned-1642655293-partition-7, standalone-0-401] Closing producer for topic persistent://public/default/testProducerIsConnectedPartitioned-1642655293-partition-7 2022-01-20 05:08:14.347 INFO [139723758618368] ProducerTest:210 | Close called 2022-01-20 05:08:14.347 INFO [139724043835136] ProducerImpl:664 | [persistent://public/default/testProducerIsConnectedPartitioned-1642655293-partition-7, standalone-0-401] Closed producer 2022-01-20 05:08:14.347 INFO [139723758618368] ProducerTest:215 | Closed /pulsar/pulsar-client-cpp/tests/ProducerTest.cc:236: Failure Value of: res Actual: Ok Expected: ResultAlreadyClosed Which is: AlreadyClosed [ FAILED ] ProducerTest.testSendAsyncCloseAsyncConcurrentlyWithLazyProducers (3431 ms) [----------] 1 test from ProducerTest (3431 ms total) [----------] Global test environment tear-down [==========] 1 test from 1 test case ran. (3435 ms total) [ PASSED ] 0 tests. [ FAILED ] 1 test, listed below: [ FAILED ] ProducerTest.testSendAsyncCloseAsyncConcurrentlyWithLazyProducers ``` full logs at https://gist.github.com/lhotari/f37aefd7919f030399f3a5b7020baa23
non_defect
flaky test c producertest testsendasynccloseasyncconcurrentlywithlazyproducers producertest testsendasynccloseasyncconcurrentlywithlazyproducers fails sporadically info producertest closing info producerimpl closing producer for topic persistent public default testproducerisconnectedpartitioned partition info producertest close called info producerimpl closed producer info producertest closed pulsar pulsar client cpp tests producertest cc failure value of res actual ok expected resultalreadyclosed which is alreadyclosed producertest testsendasynccloseasyncconcurrentlywithlazyproducers ms test from producertest ms total global test environment tear down test from test case ran ms total tests test listed below producertest testsendasynccloseasyncconcurrentlywithlazyproducers full logs at
0
20,626
3,389,359,087
IssuesEvent
2015-11-30 01:03:11
realisim/realisim
https://api.github.com/repos/realisim/realisim
closed
ornement is compromise
auto-migrated Priority-Medium Sargam Type-Defect
``` What steps will reproduce the problem? 1. have an ornement across 2 bar 2. insert note at beginning of bar 3. ornement will be broken The ornement should have been extended. ``` Original issue reported on code.google.com by `pierreolivier.beaudoin@gmail.com` on 1 May 2015 at 11:52
1.0
ornement is compromise - ``` What steps will reproduce the problem? 1. have an ornement across 2 bar 2. insert note at beginning of bar 3. ornement will be broken The ornement should have been extended. ``` Original issue reported on code.google.com by `pierreolivier.beaudoin@gmail.com` on 1 May 2015 at 11:52
defect
ornement is compromise what steps will reproduce the problem have an ornement across bar insert note at beginning of bar ornement will be broken the ornement should have been extended original issue reported on code google com by pierreolivier beaudoin gmail com on may at
1
74,751
25,300,788,723
IssuesEvent
2022-11-17 10:31:39
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
closed
Missing join rule [HZ-1735]
Type: Defect Priority: Medium Source: Internal Team: SQL to-jira
``` sql> explain select 1 from table(generate_stream(1)) > join table(generate_series(1, 3)) on 1=1; There are not enough rules to produce a node with desired properties: convention=PHYSICAL, sort=[], DistributionTraitDef=ANY. Missing conversion is JoinLogicalRel[convention: LOGICAL -> PHYSICAL] There is 1 empty subset: rel#101:RelSubset#6.PHYSICAL.[].ANY, the relevant part of the original plan is as follows 91:JoinLogicalRel(condition=[true], joinType=[inner]) 85:FullScanLogicalRel(subset=[rel#89:RelSubset#4.LOGICAL.[].ANY], table=[[hazelcast, stream, stream_d4ca0cbe_fcb2_47fd_9dd4_6ce317fbc4d4[projects=[$0]]]]) 86:FullScanLogicalRel(subset=[rel#90:RelSubset#5.LOGICAL.[].ANY], table=[[hazelcast, series, series_83e98708_3c72_4eba_bef2_199e93496138[projects=[$0]]]]) ```
1.0
Missing join rule [HZ-1735] - ``` sql> explain select 1 from table(generate_stream(1)) > join table(generate_series(1, 3)) on 1=1; There are not enough rules to produce a node with desired properties: convention=PHYSICAL, sort=[], DistributionTraitDef=ANY. Missing conversion is JoinLogicalRel[convention: LOGICAL -> PHYSICAL] There is 1 empty subset: rel#101:RelSubset#6.PHYSICAL.[].ANY, the relevant part of the original plan is as follows 91:JoinLogicalRel(condition=[true], joinType=[inner]) 85:FullScanLogicalRel(subset=[rel#89:RelSubset#4.LOGICAL.[].ANY], table=[[hazelcast, stream, stream_d4ca0cbe_fcb2_47fd_9dd4_6ce317fbc4d4[projects=[$0]]]]) 86:FullScanLogicalRel(subset=[rel#90:RelSubset#5.LOGICAL.[].ANY], table=[[hazelcast, series, series_83e98708_3c72_4eba_bef2_199e93496138[projects=[$0]]]]) ```
defect
missing join rule sql explain select from table generate stream join table generate series on there are not enough rules to produce a node with desired properties convention physical sort distributiontraitdef any missing conversion is joinlogicalrel there is empty subset rel relsubset physical any the relevant part of the original plan is as follows joinlogicalrel condition jointype fullscanlogicalrel subset any table fullscanlogicalrel subset any table
1
15,218
2,850,331,095
IssuesEvent
2015-05-31 13:49:17
damonkohler/android-scripting
https://api.github.com/repos/damonkohler/android-scripting
closed
eventGetBrodcastCategories is misspelled
auto-migrated Priority-Medium Type-Defect
``` SL4A r6 Option menu -> API Browser eventGetBrodcastCategories() is misspelled. eventGetBroadcastCategories() is what it should be. Broadcast, not Brodcast. ``` Original issue reported on code.google.com by `eldman...@gmail.com` on 3 Oct 2012 at 7:26
1.0
eventGetBrodcastCategories is misspelled - ``` SL4A r6 Option menu -> API Browser eventGetBrodcastCategories() is misspelled. eventGetBroadcastCategories() is what it should be. Broadcast, not Brodcast. ``` Original issue reported on code.google.com by `eldman...@gmail.com` on 3 Oct 2012 at 7:26
defect
eventgetbrodcastcategories is misspelled option menu api browser eventgetbrodcastcategories is misspelled eventgetbroadcastcategories is what it should be broadcast not brodcast original issue reported on code google com by eldman gmail com on oct at
1
9,920
2,616,010,465
IssuesEvent
2015-03-02 00:53:57
jasonhall/bwapi
https://api.github.com/repos/jasonhall/bwapi
closed
isBuildable and isWalkable ignore the map bottom
auto-migrated Component-Logic Priority-Low Type-Defect Usability
``` What steps will reproduce the problem? 1. Edit the ExampleAIModule's show_visibility_data section to display isBuildable or isWalkable data. 2. Scroll to the bottom of a map and look at the last row and second to last row. What is the expected output? What do you see instead? - It appears (based on trying to build structures in-game and also trying to place them in the campaign editor) that the bottom row of the map plus 5 tiles on the left and right on the 2nd to last row are not buildable no matter what the underlying terrain is. - BWAPI reports these tiles as buildable and walkable if the underlying terrain is even though they are not. - These are the tiles that can't ever be fully seen in-game due to the UI at the bottom of the screen (which explains why they aren't just based on the underlying terrain like everywhere else). What version of the product are you using? On what operating system? - BWAPI Beta 2.6.1 Please provide any additional information below. - It's fairly easy to work around this issue once you know what's going on, but it would still be nice if this special case was just integrated into the base API calls. PS: I just found this project a couple days ago (a friend showed me the AIIDE contest) and I'm having a blast playing around with it. Keep up the great work! ``` Original issue reported on code.google.com by `cfbradf...@gmail.com` on 23 Mar 2010 at 12:32
1.0
isBuildable and isWalkable ignore the map bottom - ``` What steps will reproduce the problem? 1. Edit the ExampleAIModule's show_visibility_data section to display isBuildable or isWalkable data. 2. Scroll to the bottom of a map and look at the last row and second to last row. What is the expected output? What do you see instead? - It appears (based on trying to build structures in-game and also trying to place them in the campaign editor) that the bottom row of the map plus 5 tiles on the left and right on the 2nd to last row are not buildable no matter what the underlying terrain is. - BWAPI reports these tiles as buildable and walkable if the underlying terrain is even though they are not. - These are the tiles that can't ever be fully seen in-game due to the UI at the bottom of the screen (which explains why they aren't just based on the underlying terrain like everywhere else). What version of the product are you using? On what operating system? - BWAPI Beta 2.6.1 Please provide any additional information below. - It's fairly easy to work around this issue once you know what's going on, but it would still be nice if this special case was just integrated into the base API calls. PS: I just found this project a couple days ago (a friend showed me the AIIDE contest) and I'm having a blast playing around with it. Keep up the great work! ``` Original issue reported on code.google.com by `cfbradf...@gmail.com` on 23 Mar 2010 at 12:32
defect
isbuildable and iswalkable ignore the map bottom what steps will reproduce the problem edit the exampleaimodule s show visibility data section to display isbuildable or iswalkable data scroll to the bottom of a map and look at the last row and second to last row what is the expected output what do you see instead it appears based on trying to build structures in game and also trying to place them in the campaign editor that the bottom row of the map plus tiles on the left and right on the to last row are not buildable no matter what the underlying terrain is bwapi reports these tiles as buildable and walkable if the underlying terrain is even though they are not these are the tiles that can t ever be fully seen in game due to the ui at the bottom of the screen which explains why they aren t just based on the underlying terrain like everywhere else what version of the product are you using on what operating system bwapi beta please provide any additional information below it s fairly easy to work around this issue once you know what s going on but it would still be nice if this special case was just integrated into the base api calls ps i just found this project a couple days ago a friend showed me the aiide contest and i m having a blast playing around with it keep up the great work original issue reported on code google com by cfbradf gmail com on mar at
1
37,665
5,135,007,939
IssuesEvent
2017-01-11 10:51:31
IDgis/geoportaal-test
https://api.github.com/repos/IDgis/geoportaal-test
closed
filenaam xml van geosticker zichtbaar in metadata stylesheet
gebruikerstest impact laag wens
Wens is om de volledige naam van de xml terug te laten komen in de metadata van de stylesheet. Blijkt nu al wel in rapportage te staan, zie #331
1.0
filenaam xml van geosticker zichtbaar in metadata stylesheet - Wens is om de volledige naam van de xml terug te laten komen in de metadata van de stylesheet. Blijkt nu al wel in rapportage te staan, zie #331
non_defect
filenaam xml van geosticker zichtbaar in metadata stylesheet wens is om de volledige naam van de xml terug te laten komen in de metadata van de stylesheet blijkt nu al wel in rapportage te staan zie
0
4,818
2,610,157,274
IssuesEvent
2015-02-26 18:49:59
chrsmith/republic-at-war
https://api.github.com/repos/chrsmith/republic-at-war
opened
Text
auto-migrated Priority-Medium Type-Defect
``` FIx new planet descriptions ``` ----- Original issue reported on code.google.com by `z3r0...@gmail.com` on 30 Jan 2011 at 2:42
1.0
Text - ``` FIx new planet descriptions ``` ----- Original issue reported on code.google.com by `z3r0...@gmail.com` on 30 Jan 2011 at 2:42
defect
text fix new planet descriptions original issue reported on code google com by gmail com on jan at
1
35,876
7,817,027,387
IssuesEvent
2018-06-13 07:42:05
line/armeria
https://api.github.com/repos/line/armeria
opened
Armeria server does not handle the `Expect: 100-continue` header correctly.
defect
`Http1ObjectDecoder` and `Http2ObjectDecoder` does not send `100 Continue` informational header when a client sends a request with the `Expect: 100-continue` header, which is obviously a violation of the HTTP specification, unless we want to reject the request. Reported by @jwills
1.0
Armeria server does not handle the `Expect: 100-continue` header correctly. - `Http1ObjectDecoder` and `Http2ObjectDecoder` does not send `100 Continue` informational header when a client sends a request with the `Expect: 100-continue` header, which is obviously a violation of the HTTP specification, unless we want to reject the request. Reported by @jwills
defect
armeria server does not handle the expect continue header correctly and does not send continue informational header when a client sends a request with the expect continue header which is obviously a violation of the http specification unless we want to reject the request reported by jwills
1
9,813
2,615,175,345
IssuesEvent
2015-03-01 06:58:37
chrsmith/reaver-wps
https://api.github.com/repos/chrsmith/reaver-wps
opened
Hacking a hidden wireless
auto-migrated Priority-Triage Type-Defect
``` I found the name of the AP : @@CR7)) when i try to hack it using reaver i got an error which said syntaxe error : unexpetected ")" what's the solution for this problem? ``` Original issue reported on code.google.com by `marouane...@gmail.com` on 17 Jul 2014 at 6:49
1.0
Hacking a hidden wireless - ``` I found the name of the AP : @@CR7)) when i try to hack it using reaver i got an error which said syntaxe error : unexpetected ")" what's the solution for this problem? ``` Original issue reported on code.google.com by `marouane...@gmail.com` on 17 Jul 2014 at 6:49
defect
hacking a hidden wireless i found the name of the ap when i try to hack it using reaver i got an error which said syntaxe error unexpetected what s the solution for this problem original issue reported on code google com by marouane gmail com on jul at
1
49,180
13,185,281,976
IssuesEvent
2020-08-12 21:05:02
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
opened
[photospline] - divzero - loop pre-conditions aren't checked (Trac #921)
Incomplete Migration Migrated from Trac combo reconstruction defect
<details> <summary><em>Migrated from https://code.icecube.wisc.edu/ticket/921 , reported by nega and owned by jvansanten</em></summary> <p> ```json { "status": "closed", "changetime": "2015-04-12T17:41:42", "description": "http://goo.gl/s1mdUk\n\nloop preconditions aren't checked allowing a potential divide-by-zero error to occur in 8 steps.\n\nfix: pre-check and hard fail if loop pre-conditions suck", "reporter": "nega", "cc": "", "resolution": "fixed", "_ts": "1428860502423619", "component": "combo reconstruction", "summary": "[photospline] - divzero - loop pre-conditions aren't checked", "priority": "normal", "keywords": "photospline divzero", "time": "2015-04-10T04:13:30", "milestone": "", "owner": "jvansanten", "type": "defect" } ``` </p> </details>
1.0
[photospline] - divzero - loop pre-conditions aren't checked (Trac #921) - <details> <summary><em>Migrated from https://code.icecube.wisc.edu/ticket/921 , reported by nega and owned by jvansanten</em></summary> <p> ```json { "status": "closed", "changetime": "2015-04-12T17:41:42", "description": "http://goo.gl/s1mdUk\n\nloop preconditions aren't checked allowing a potential divide-by-zero error to occur in 8 steps.\n\nfix: pre-check and hard fail if loop pre-conditions suck", "reporter": "nega", "cc": "", "resolution": "fixed", "_ts": "1428860502423619", "component": "combo reconstruction", "summary": "[photospline] - divzero - loop pre-conditions aren't checked", "priority": "normal", "keywords": "photospline divzero", "time": "2015-04-10T04:13:30", "milestone": "", "owner": "jvansanten", "type": "defect" } ``` </p> </details>
defect
divzero loop pre conditions aren t checked trac migrated from reported by nega and owned by jvansanten json status closed changetime description preconditions aren t checked allowing a potential divide by zero error to occur in steps n nfix pre check and hard fail if loop pre conditions suck reporter nega cc resolution fixed ts component combo reconstruction summary divzero loop pre conditions aren t checked priority normal keywords photospline divzero time milestone owner jvansanten type defect
1
71,169
23,480,444,768
IssuesEvent
2022-08-17 10:04:20
zed-industries/feedback
https://api.github.com/repos/zed-industries/feedback
closed
Can't open certain files during collaborative session
defect discussed collaboration
### Check for existing issues - [X] Completed ### Describe the bug As a guest in a collaborative session, I wasn't able to open all files in the workspace. ### To reproduce Not sure how to reproduce it... I didn't have a way to fully test it but: 1. Join as guest 2. Open a file 3. CMD+Click to go to definition (won't work) 4. Try to navigate to the definition file (through command palette or tree view) ### Expected behavior The file is open in the buffer ### Environment Zed 0.49.1 – /Applications/Zed.app macOS 12.1 architecture arm64 ### If applicable, add mockups / screenshots to help explain present your vision of the feature https://user-images.githubusercontent.com/8927326/182865997-96403def-c5e9-4917-a2eb-56e7f50c0000.mp4 ### If applicable, attach your `~/Library/Logs/Zed/Zed.log` file to this issue Caused by: 0: IO error: Connection reset by peer (os error 54) 1: Connection reset by peer (os error 54) 06:03:23 [INFO] set status on client 0: ConnectionLost 06:03:23 [INFO] set status on client 0: Reauthenticating 06:03:23 [INFO] set status on client 0: Reconnecting 06:03:23 [INFO] connected to rpc endpoint https://collab.zed.dev/rpc 06:03:24 [INFO] add connection to peer 06:03:24 [INFO] add_connection; 06:03:24 [INFO] set status to connected 98 06:03:24 [INFO] set status on client 0: Connected { connection_id: ConnectionId(98) } 06:18:59 [ERROR] connection error: error reading rpc message from socket Caused by: 0: IO error: Connection reset by peer (os error 54) 1: Connection reset by peer (os error 54) 06:18:59 [INFO] set status on client 0: ConnectionLost 06:18:59 [INFO] set status on client 0: Reauthenticating 06:18:59 [INFO] set status on client 0: Reconnecting 06:19:00 [INFO] connected to rpc endpoint https://collab.zed.dev/rpc 06:19:01 [INFO] add connection to peer 06:19:01 [INFO] add_connection; 06:19:01 [INFO] set status to connected 99 06:19:01 [INFO] set status on client 0: Connected { connection_id: ConnectionId(99) } 06:36:28 [ERROR] connection error: error reading rpc message from socket Caused by: 0: IO error: Connection reset by peer (os error 54) 1: Connection reset by peer (os error 54) 06:36:28 [INFO] set status on client 0: ConnectionLost 06:36:28 [INFO] set status on client 0: Reauthenticating 06:36:28 [INFO] set status on client 0: Reconnecting 06:36:29 [INFO] connected to rpc endpoint https://collab.zed.dev/rpc 06:36:30 [INFO] add connection to peer 06:36:30 [INFO] add_connection; 06:36:30 [INFO] set status to connected 100 06:36:30 [INFO] set status on client 0: Connected { connection_id: ConnectionId(100) } 06:47:11 [ERROR] connection error: delay between messages too long 06:47:11 [INFO] set status on client 0: ConnectionLost 06:47:11 [INFO] set status on client 0: Reauthenticating 06:47:11 [INFO] set status on client 0: Reconnecting 06:51:04 [INFO] connected to rpc endpoint https://collab.zed.dev/rpc 06:51:05 [INFO] add connection to peer 06:51:05 [INFO] add_connection; 06:51:05 [INFO] set status to connected 101 06:51:05 [INFO] set status on client 0: Connected { connection_id: ConnectionId(101) } 07:07:00 [ERROR] connection error: error reading rpc message from socket Caused by: 0: IO error: Connection reset by peer (os error 54) 1: Connection reset by peer (os error 54) 07:07:00 [INFO] set status on client 0: ConnectionLost 07:07:00 [INFO] set status on client 0: Reauthenticating 07:07:00 [INFO] set status on client 0: Reconnecting 07:07:01 [INFO] connected to rpc endpoint https://collab.zed.dev/rpc 07:07:01 [INFO] add connection to peer 07:07:01 [INFO] add_connection; 07:07:01 [INFO] set status to connected 102 07:07:01 [INFO] set status on client 0: Connected { connection_id: ConnectionId(102) } 07:18:44 [ERROR] connection error: error reading rpc message from socket Caused by: 0: IO error: Connection reset by peer (os error 54) 1: Connection reset by peer (os error 54) 07:18:45 [INFO] set status on client 0: ConnectionLost 07:18:45 [INFO] set status on client 0: Reauthenticating 07:18:45 [INFO] set status on client 0: Reconnecting 07:18:45 [INFO] connected to rpc endpoint https://collab.zed.dev/rpc 07:18:48 [INFO] add connection to peer 07:18:48 [INFO] add_connection; 07:18:48 [INFO] set status to connected 103 07:18:48 [INFO] set status on client 0: Connected { connection_id: ConnectionId(103) } 08:40:54 [INFO] open paths ["/Users/federico/Code/github.com/Think-iT-Labs/think-it-website"] 11:05:21 [INFO] open paths ["/Users/federico/Code/github.com/Think-iT-Labs/internal-edc-ui"] 11:05:26 [WARN] incoming response: unknown request connection_id=103 message_id=22 responding_to=536 11:06:31 [WARN] incoming response: unknown request connection_id=103 message_id=243 responding_to=573 11:10:53 [WARN] incoming response: unknown request connection_id=103 message_id=1038 responding_to=731 11:10:58 [WARN] incoming response: unknown request connection_id=103 message_id=1073 responding_to=744 11:24:35 [ERROR] error handling message. client_id:0, message_id:1032, sender_id:Some(6808), type:UpdateBufferFile, error:no such buffer 11:24:38 [ERROR] error handling message. client_id:0, message_id:1049, sender_id:Some(6808), type:UpdateBufferFile, error:no such buffer 11:24:38 [ERROR] error handling message. client_id:0, message_id:1050, sender_id:Some(6808), type:UpdateBufferFile, error:no such buffer 11:24:52 [ERROR] error handling message. client_id:0, message_id:1067, sender_id:Some(6808), type:UpdateBufferFile, error:no such buffer 11:24:52 [ERROR] error handling message. client_id:0, message_id:1068, sender_id:Some(6808), type:UpdateBufferFile, error:no such buffer 11:24:52 [ERROR] error handling message. client_id:0, message_id:1069, sender_id:Some(6808), type:UpdateBufferFile, error:no such buffer 11:24:52 [ERROR] error handling message. client_id:0, message_id:1078, sender_id:Some(6808), type:UpdateBufferFile, error:no such buffer 11:24:52 [ERROR] error handling message. client_id:0, message_id:1079, sender_id:Some(6808), type:UpdateBufferFile, error:no such buffer 11:24:52 [ERROR] error handling message. client_id:0, message_id:1080, sender_id:Some(6808), type:UpdateBufferFile, error:no such buffer 11:25:04 [ERROR] error handling message. client_id:0, message_id:1088, sender_id:Some(6808), type:UpdateBufferFile, error:no such buffer 11:25:04 [ERROR] error handling message. client_id:0, message_id:1089, sender_id:Some(6808), type:UpdateBufferFile, error:no such buffer 11:25:04 [ERROR] error handling message. client_id:0, message_id:1090, sender_id:Some(6808), type:UpdateBufferFile, error:no such buffer 11:25:39 [ERROR] error handling message. client_id:0, message_id:1126, sender_id:Some(6808), type:UpdateBufferFile, error:no such buffer 11:25:39 [ERROR] error handling message. client_id:0, message_id:1128, sender_id:Some(6808), type:UpdateBufferFile, error:no such buffer 11:25:39 [ERROR] error handling message. client_id:0, message_id:1129, sender_id:Some(6808), type:UpdateBufferFile, error:no such buffer 11:26:47 [ERROR] error handling message. client_id:0, message_id:1154, sender_id:Some(6808), type:UpdateBufferFile, error:no such buffer 11:26:47 [ERROR] error handling message. client_id:0, message_id:1157, sender_id:Some(6808), type:UpdateBufferFile, error:no such buffer 11:26:47 [ERROR] error handling message. client_id:0, message_id:1158, sender_id:Some(6808), type:UpdateBufferFile, error:no such buffer 11:26:49 [ERROR] error handling message. client_id:0, message_id:1167, sender_id:Some(6808), type:UpdateBufferFile, error:no such buffer 11:26:49 [ERROR] error handling message. client_id:0, message_id:1170, sender_id:Some(6808), type:UpdateBufferFile, error:no such buffer 11:26:49 [ERROR] error handling message. client_id:0, message_id:1171, sender_id:Some(6808), type:UpdateBufferFile, error:no such buffer
1.0
Can't open certain files during collaborative session - ### Check for existing issues - [X] Completed ### Describe the bug As a guest in a collaborative session, I wasn't able to open all files in the workspace. ### To reproduce Not sure how to reproduce it... I didn't have a way to fully test it but: 1. Join as guest 2. Open a file 3. CMD+Click to go to definition (won't work) 4. Try to navigate to the definition file (through command palette or tree view) ### Expected behavior The file is open in the buffer ### Environment Zed 0.49.1 – /Applications/Zed.app macOS 12.1 architecture arm64 ### If applicable, add mockups / screenshots to help explain present your vision of the feature https://user-images.githubusercontent.com/8927326/182865997-96403def-c5e9-4917-a2eb-56e7f50c0000.mp4 ### If applicable, attach your `~/Library/Logs/Zed/Zed.log` file to this issue Caused by: 0: IO error: Connection reset by peer (os error 54) 1: Connection reset by peer (os error 54) 06:03:23 [INFO] set status on client 0: ConnectionLost 06:03:23 [INFO] set status on client 0: Reauthenticating 06:03:23 [INFO] set status on client 0: Reconnecting 06:03:23 [INFO] connected to rpc endpoint https://collab.zed.dev/rpc 06:03:24 [INFO] add connection to peer 06:03:24 [INFO] add_connection; 06:03:24 [INFO] set status to connected 98 06:03:24 [INFO] set status on client 0: Connected { connection_id: ConnectionId(98) } 06:18:59 [ERROR] connection error: error reading rpc message from socket Caused by: 0: IO error: Connection reset by peer (os error 54) 1: Connection reset by peer (os error 54) 06:18:59 [INFO] set status on client 0: ConnectionLost 06:18:59 [INFO] set status on client 0: Reauthenticating 06:18:59 [INFO] set status on client 0: Reconnecting 06:19:00 [INFO] connected to rpc endpoint https://collab.zed.dev/rpc 06:19:01 [INFO] add connection to peer 06:19:01 [INFO] add_connection; 06:19:01 [INFO] set status to connected 99 06:19:01 [INFO] set status on client 0: Connected { connection_id: ConnectionId(99) } 06:36:28 [ERROR] connection error: error reading rpc message from socket Caused by: 0: IO error: Connection reset by peer (os error 54) 1: Connection reset by peer (os error 54) 06:36:28 [INFO] set status on client 0: ConnectionLost 06:36:28 [INFO] set status on client 0: Reauthenticating 06:36:28 [INFO] set status on client 0: Reconnecting 06:36:29 [INFO] connected to rpc endpoint https://collab.zed.dev/rpc 06:36:30 [INFO] add connection to peer 06:36:30 [INFO] add_connection; 06:36:30 [INFO] set status to connected 100 06:36:30 [INFO] set status on client 0: Connected { connection_id: ConnectionId(100) } 06:47:11 [ERROR] connection error: delay between messages too long 06:47:11 [INFO] set status on client 0: ConnectionLost 06:47:11 [INFO] set status on client 0: Reauthenticating 06:47:11 [INFO] set status on client 0: Reconnecting 06:51:04 [INFO] connected to rpc endpoint https://collab.zed.dev/rpc 06:51:05 [INFO] add connection to peer 06:51:05 [INFO] add_connection; 06:51:05 [INFO] set status to connected 101 06:51:05 [INFO] set status on client 0: Connected { connection_id: ConnectionId(101) } 07:07:00 [ERROR] connection error: error reading rpc message from socket Caused by: 0: IO error: Connection reset by peer (os error 54) 1: Connection reset by peer (os error 54) 07:07:00 [INFO] set status on client 0: ConnectionLost 07:07:00 [INFO] set status on client 0: Reauthenticating 07:07:00 [INFO] set status on client 0: Reconnecting 07:07:01 [INFO] connected to rpc endpoint https://collab.zed.dev/rpc 07:07:01 [INFO] add connection to peer 07:07:01 [INFO] add_connection; 07:07:01 [INFO] set status to connected 102 07:07:01 [INFO] set status on client 0: Connected { connection_id: ConnectionId(102) } 07:18:44 [ERROR] connection error: error reading rpc message from socket Caused by: 0: IO error: Connection reset by peer (os error 54) 1: Connection reset by peer (os error 54) 07:18:45 [INFO] set status on client 0: ConnectionLost 07:18:45 [INFO] set status on client 0: Reauthenticating 07:18:45 [INFO] set status on client 0: Reconnecting 07:18:45 [INFO] connected to rpc endpoint https://collab.zed.dev/rpc 07:18:48 [INFO] add connection to peer 07:18:48 [INFO] add_connection; 07:18:48 [INFO] set status to connected 103 07:18:48 [INFO] set status on client 0: Connected { connection_id: ConnectionId(103) } 08:40:54 [INFO] open paths ["/Users/federico/Code/github.com/Think-iT-Labs/think-it-website"] 11:05:21 [INFO] open paths ["/Users/federico/Code/github.com/Think-iT-Labs/internal-edc-ui"] 11:05:26 [WARN] incoming response: unknown request connection_id=103 message_id=22 responding_to=536 11:06:31 [WARN] incoming response: unknown request connection_id=103 message_id=243 responding_to=573 11:10:53 [WARN] incoming response: unknown request connection_id=103 message_id=1038 responding_to=731 11:10:58 [WARN] incoming response: unknown request connection_id=103 message_id=1073 responding_to=744 11:24:35 [ERROR] error handling message. client_id:0, message_id:1032, sender_id:Some(6808), type:UpdateBufferFile, error:no such buffer 11:24:38 [ERROR] error handling message. client_id:0, message_id:1049, sender_id:Some(6808), type:UpdateBufferFile, error:no such buffer 11:24:38 [ERROR] error handling message. client_id:0, message_id:1050, sender_id:Some(6808), type:UpdateBufferFile, error:no such buffer 11:24:52 [ERROR] error handling message. client_id:0, message_id:1067, sender_id:Some(6808), type:UpdateBufferFile, error:no such buffer 11:24:52 [ERROR] error handling message. client_id:0, message_id:1068, sender_id:Some(6808), type:UpdateBufferFile, error:no such buffer 11:24:52 [ERROR] error handling message. client_id:0, message_id:1069, sender_id:Some(6808), type:UpdateBufferFile, error:no such buffer 11:24:52 [ERROR] error handling message. client_id:0, message_id:1078, sender_id:Some(6808), type:UpdateBufferFile, error:no such buffer 11:24:52 [ERROR] error handling message. client_id:0, message_id:1079, sender_id:Some(6808), type:UpdateBufferFile, error:no such buffer 11:24:52 [ERROR] error handling message. client_id:0, message_id:1080, sender_id:Some(6808), type:UpdateBufferFile, error:no such buffer 11:25:04 [ERROR] error handling message. client_id:0, message_id:1088, sender_id:Some(6808), type:UpdateBufferFile, error:no such buffer 11:25:04 [ERROR] error handling message. client_id:0, message_id:1089, sender_id:Some(6808), type:UpdateBufferFile, error:no such buffer 11:25:04 [ERROR] error handling message. client_id:0, message_id:1090, sender_id:Some(6808), type:UpdateBufferFile, error:no such buffer 11:25:39 [ERROR] error handling message. client_id:0, message_id:1126, sender_id:Some(6808), type:UpdateBufferFile, error:no such buffer 11:25:39 [ERROR] error handling message. client_id:0, message_id:1128, sender_id:Some(6808), type:UpdateBufferFile, error:no such buffer 11:25:39 [ERROR] error handling message. client_id:0, message_id:1129, sender_id:Some(6808), type:UpdateBufferFile, error:no such buffer 11:26:47 [ERROR] error handling message. client_id:0, message_id:1154, sender_id:Some(6808), type:UpdateBufferFile, error:no such buffer 11:26:47 [ERROR] error handling message. client_id:0, message_id:1157, sender_id:Some(6808), type:UpdateBufferFile, error:no such buffer 11:26:47 [ERROR] error handling message. client_id:0, message_id:1158, sender_id:Some(6808), type:UpdateBufferFile, error:no such buffer 11:26:49 [ERROR] error handling message. client_id:0, message_id:1167, sender_id:Some(6808), type:UpdateBufferFile, error:no such buffer 11:26:49 [ERROR] error handling message. client_id:0, message_id:1170, sender_id:Some(6808), type:UpdateBufferFile, error:no such buffer 11:26:49 [ERROR] error handling message. client_id:0, message_id:1171, sender_id:Some(6808), type:UpdateBufferFile, error:no such buffer
defect
can t open certain files during collaborative session check for existing issues completed describe the bug as a guest in a collaborative session i wasn t able to open all files in the workspace to reproduce not sure how to reproduce it i didn t have a way to fully test it but join as guest open a file cmd click to go to definition won t work try to navigate to the definition file through command palette or tree view expected behavior the file is open in the buffer environment zed – applications zed app macos architecture if applicable add mockups screenshots to help explain present your vision of the feature if applicable attach your library logs zed zed log file to this issue caused by io error connection reset by peer os error connection reset by peer os error set status on client connectionlost set status on client reauthenticating set status on client reconnecting connected to rpc endpoint add connection to peer add connection set status to connected set status on client connected connection id connectionid connection error error reading rpc message from socket caused by io error connection reset by peer os error connection reset by peer os error set status on client connectionlost set status on client reauthenticating set status on client reconnecting connected to rpc endpoint add connection to peer add connection set status to connected set status on client connected connection id connectionid connection error error reading rpc message from socket caused by io error connection reset by peer os error connection reset by peer os error set status on client connectionlost set status on client reauthenticating set status on client reconnecting connected to rpc endpoint add connection to peer add connection set status to connected set status on client connected connection id connectionid connection error delay between messages too long set status on client connectionlost set status on client reauthenticating set status on client reconnecting connected to rpc endpoint add connection to peer add connection set status to connected set status on client connected connection id connectionid connection error error reading rpc message from socket caused by io error connection reset by peer os error connection reset by peer os error set status on client connectionlost set status on client reauthenticating set status on client reconnecting connected to rpc endpoint add connection to peer add connection set status to connected set status on client connected connection id connectionid connection error error reading rpc message from socket caused by io error connection reset by peer os error connection reset by peer os error set status on client connectionlost set status on client reauthenticating set status on client reconnecting connected to rpc endpoint add connection to peer add connection set status to connected set status on client connected connection id connectionid open paths open paths incoming response unknown request connection id message id responding to incoming response unknown request connection id message id responding to incoming response unknown request connection id message id responding to incoming response unknown request connection id message id responding to error handling message client id message id sender id some type updatebufferfile error no such buffer error handling message client id message id sender id some type updatebufferfile error no such buffer error handling message client id message id sender id some type updatebufferfile error no such buffer error handling message client id message id sender id some type updatebufferfile error no such buffer error handling message client id message id sender id some type updatebufferfile error no such buffer error handling message client id message id sender id some type updatebufferfile error no such buffer error handling message client id message id sender id some type updatebufferfile error no such buffer error handling message client id message id sender id some type updatebufferfile error no such buffer error handling message client id message id sender id some type updatebufferfile error no such buffer error handling message client id message id sender id some type updatebufferfile error no such buffer error handling message client id message id sender id some type updatebufferfile error no such buffer error handling message client id message id sender id some type updatebufferfile error no such buffer error handling message client id message id sender id some type updatebufferfile error no such buffer error handling message client id message id sender id some type updatebufferfile error no such buffer error handling message client id message id sender id some type updatebufferfile error no such buffer error handling message client id message id sender id some type updatebufferfile error no such buffer error handling message client id message id sender id some type updatebufferfile error no such buffer error handling message client id message id sender id some type updatebufferfile error no such buffer error handling message client id message id sender id some type updatebufferfile error no such buffer error handling message client id message id sender id some type updatebufferfile error no such buffer error handling message client id message id sender id some type updatebufferfile error no such buffer
1
78,574
9,764,381,190
IssuesEvent
2019-06-05 15:39:20
cityofaustin/techstack
https://api.github.com/repos/cityofaustin/techstack
closed
Representing guides on topic & other pages
Content Type: Guides Epic Team: Design + Research
### Residents should be able to distinguish and view when guides are available for the content that they are viewing, so the can navigate to it. Guides are comprised of many pages, and may also be represented on one or more topics, and possibly topic collections. How do users distinguish a guide from other topic content? What form visually does a link to a guide take? This issue covers how residents can see when guide are available when they are looking at related content on TOPIC COLLECTIONS. Use "Get a mobile food vendor permit" as your example. Low fidelity is fine.
1.0
Representing guides on topic & other pages - ### Residents should be able to distinguish and view when guides are available for the content that they are viewing, so the can navigate to it. Guides are comprised of many pages, and may also be represented on one or more topics, and possibly topic collections. How do users distinguish a guide from other topic content? What form visually does a link to a guide take? This issue covers how residents can see when guide are available when they are looking at related content on TOPIC COLLECTIONS. Use "Get a mobile food vendor permit" as your example. Low fidelity is fine.
non_defect
representing guides on topic other pages residents should be able to distinguish and view when guides are available for the content that they are viewing so the can navigate to it guides are comprised of many pages and may also be represented on one or more topics and possibly topic collections how do users distinguish a guide from other topic content what form visually does a link to a guide take this issue covers how residents can see when guide are available when they are looking at related content on topic collections use get a mobile food vendor permit as your example low fidelity is fine
0
178,949
14,685,866,939
IssuesEvent
2021-01-01 11:48:15
https-github-com-JuriBurakov/bitcoin
https://api.github.com/repos/https-github-com-JuriBurakov/bitcoin
opened
Good first issue
DNS TXT Payeertoken bug documentation duplicate enhancement good first issue help wanted invalid main question security findings wontfix
<!-- Needs the label "good first issue" assigned manually before or after opening --> <!-- A good first issue is an uncontroversial issue, that has a relatively unique and obvious solution --> <!-- Motivate the issue and explain the solution briefly --> #### Useful skills: <!-- (For example, “C++11 std::thread”, “Qt5 GUI and async GUI design” or “basic understanding of Bitcoin mining and the Bitcoin Core RPC interface”.) --> #### Want to work on this issue? For guidance on contributing, please read [CONTRIBUTING.md](https://github.com/bitcoin/bitcoin/blob/master/CONTRIBUTING.md) before opening your pull request.
1.0
Good first issue - <!-- Needs the label "good first issue" assigned manually before or after opening --> <!-- A good first issue is an uncontroversial issue, that has a relatively unique and obvious solution --> <!-- Motivate the issue and explain the solution briefly --> #### Useful skills: <!-- (For example, “C++11 std::thread”, “Qt5 GUI and async GUI design” or “basic understanding of Bitcoin mining and the Bitcoin Core RPC interface”.) --> #### Want to work on this issue? For guidance on contributing, please read [CONTRIBUTING.md](https://github.com/bitcoin/bitcoin/blob/master/CONTRIBUTING.md) before opening your pull request.
non_defect
good first issue useful skills want to work on this issue for guidance on contributing please read before opening your pull request
0
8,565
2,611,517,952
IssuesEvent
2015-02-27 05:52:28
chrsmith/hedgewars
https://api.github.com/repos/chrsmith/hedgewars
closed
Engine crash
auto-migrated Component-Engine Priority-Medium Type-Defect
``` An unhandled exception occurred at $00496113 : EAccessViolation : Access violation $00496113 DOSTEPHEDGEHOGFREE, line 1171 of uGearsHedgehog.pas $0049663C DOSTEPHEDGEHOG, line 1274 of uGearsHedgehog.pas $0044526F PROCESSGEARS, line 242 of uGears.pas $00411BD1 DOGAMETICK, line 92 of uGame.pas $00403661 DOTIMER, line 99 of hwengine.pas $00403D94 MAINLOOP, line 289 of hwengine.pas $004042F8 GAME, line 420 of hwengine.pas $004045E6 main, line 550 of hwengine.pas log attached. Seems to be related to the switcher being active at the moment. ``` Original issue reported on code.google.com by `unC0Rr` on 30 Apr 2013 at 6:59 Attachments: * [game0.log.gz](https://storage.googleapis.com/google-code-attachments/hedgewars/issue-620/comment-0/game0.log.gz)
1.0
Engine crash - ``` An unhandled exception occurred at $00496113 : EAccessViolation : Access violation $00496113 DOSTEPHEDGEHOGFREE, line 1171 of uGearsHedgehog.pas $0049663C DOSTEPHEDGEHOG, line 1274 of uGearsHedgehog.pas $0044526F PROCESSGEARS, line 242 of uGears.pas $00411BD1 DOGAMETICK, line 92 of uGame.pas $00403661 DOTIMER, line 99 of hwengine.pas $00403D94 MAINLOOP, line 289 of hwengine.pas $004042F8 GAME, line 420 of hwengine.pas $004045E6 main, line 550 of hwengine.pas log attached. Seems to be related to the switcher being active at the moment. ``` Original issue reported on code.google.com by `unC0Rr` on 30 Apr 2013 at 6:59 Attachments: * [game0.log.gz](https://storage.googleapis.com/google-code-attachments/hedgewars/issue-620/comment-0/game0.log.gz)
defect
engine crash an unhandled exception occurred at eaccessviolation access violation dostephedgehogfree line of ugearshedgehog pas dostephedgehog line of ugearshedgehog pas processgears line of ugears pas dogametick line of ugame pas dotimer line of hwengine pas mainloop line of hwengine pas game line of hwengine pas main line of hwengine pas log attached seems to be related to the switcher being active at the moment original issue reported on code google com by on apr at attachments
1
133,544
12,543,862,708
IssuesEvent
2020-06-05 16:15:26
Azure/azure-cosmos-dotnet-v3
https://api.github.com/repos/Azure/azure-cosmos-dotnet-v3
closed
Doc samples don't compile
VNext bug documentation
This doc: https://docs.microsoft.com/en-us/dotnet/api/azure.cosmos.cosmoscontainer.getitemquerystreamiterator?view=azure-dotnet-preview Uses this code: ```csharp public class ToDoActivity{ public string id {get; set;} public string status {get; set;} public int cost {get; set;} } QueryDefinition queryDefinition = new QueryDefinition("select * from ToDos t where t.cost > @expensive") .WithParameter("@expensive", 9000); await foreach(Response response in this.Container.GetItemQueryStreamIterator( queryDefinition, null, new QueryRequestOptions() { PartitionKey = new PartitionKey("Error")})) { using (StreamReader sr = new StreamReader(response.Content)) using (JsonTextReader jtr = new JsonTextReader(sr)) { JObject result = JObject.Load(jtr); } } ``` Which will not compile because the `.Content` property no longer exists.
1.0
Doc samples don't compile - This doc: https://docs.microsoft.com/en-us/dotnet/api/azure.cosmos.cosmoscontainer.getitemquerystreamiterator?view=azure-dotnet-preview Uses this code: ```csharp public class ToDoActivity{ public string id {get; set;} public string status {get; set;} public int cost {get; set;} } QueryDefinition queryDefinition = new QueryDefinition("select * from ToDos t where t.cost > @expensive") .WithParameter("@expensive", 9000); await foreach(Response response in this.Container.GetItemQueryStreamIterator( queryDefinition, null, new QueryRequestOptions() { PartitionKey = new PartitionKey("Error")})) { using (StreamReader sr = new StreamReader(response.Content)) using (JsonTextReader jtr = new JsonTextReader(sr)) { JObject result = JObject.Load(jtr); } } ``` Which will not compile because the `.Content` property no longer exists.
non_defect
doc samples don t compile this doc uses this code csharp public class todoactivity public string id get set public string status get set public int cost get set querydefinition querydefinition new querydefinition select from todos t where t cost expensive withparameter expensive await foreach response response in this container getitemquerystreamiterator querydefinition null new queryrequestoptions partitionkey new partitionkey error using streamreader sr new streamreader response content using jsontextreader jtr new jsontextreader sr jobject result jobject load jtr which will not compile because the content property no longer exists
0
68,415
21,664,305,797
IssuesEvent
2022-05-07 00:57:09
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
cachebust ILAG home pages correctly
T-Defect P1 A-ILAG
the lack of cache busting on the homepage resources is going to bite us monumentally
1.0
cachebust ILAG home pages correctly - the lack of cache busting on the homepage resources is going to bite us monumentally
defect
cachebust ilag home pages correctly the lack of cache busting on the homepage resources is going to bite us monumentally
1
80,331
30,237,082,425
IssuesEvent
2023-07-06 10:59:56
vector-im/element-x-ios
https://api.github.com/repos/vector-im/element-x-ios
opened
No sync spinner while catching up on to-device msgs
T-Defect
### Steps to reproduce 1. Launch app for first time in a few weeks 2. UISIs, UISIs, everywhere 3. Over the next minute or so, they slowly decrypt. 4. No spinner at all to let you know that the app is working through a massive todevice encryption backlog. ### Outcome #### What did you expect? Spinner, if we're still frantically syncing data. Should SS perhaps just have a flag in its responses to let the client know when it's churned through the backlog, as the metric for when to hide the syncing spinner? #### What happened instead? No spinner, much flakiness, much sad. ### Your phone model _No response_ ### Operating system version _No response_ ### Application version 276 ### Homeserver _No response_ ### Will you send logs? Yes
1.0
No sync spinner while catching up on to-device msgs - ### Steps to reproduce 1. Launch app for first time in a few weeks 2. UISIs, UISIs, everywhere 3. Over the next minute or so, they slowly decrypt. 4. No spinner at all to let you know that the app is working through a massive todevice encryption backlog. ### Outcome #### What did you expect? Spinner, if we're still frantically syncing data. Should SS perhaps just have a flag in its responses to let the client know when it's churned through the backlog, as the metric for when to hide the syncing spinner? #### What happened instead? No spinner, much flakiness, much sad. ### Your phone model _No response_ ### Operating system version _No response_ ### Application version 276 ### Homeserver _No response_ ### Will you send logs? Yes
defect
no sync spinner while catching up on to device msgs steps to reproduce launch app for first time in a few weeks uisis uisis everywhere over the next minute or so they slowly decrypt no spinner at all to let you know that the app is working through a massive todevice encryption backlog outcome what did you expect spinner if we re still frantically syncing data should ss perhaps just have a flag in its responses to let the client know when it s churned through the backlog as the metric for when to hide the syncing spinner what happened instead no spinner much flakiness much sad your phone model no response operating system version no response application version homeserver no response will you send logs yes
1
47,815
13,066,249,512
IssuesEvent
2020-07-30 21:18:04
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
closed
[genie-icetray] needs updated RELEASE_NOTES (Trac #1213)
Migrated from Trac combo simulation defect
There are several commits on trunk that need to be added to the RELEASE_NOTES. Migrated from https://code.icecube.wisc.edu/ticket/1213 ```json { "status": "closed", "changetime": "2016-03-18T21:14:03", "description": "There are several commits on trunk that need to be added to the RELEASE_NOTES.", "reporter": "david.schultz", "cc": "melanie.day", "resolution": "fixed", "_ts": "1458335643235016", "component": "combo simulation", "summary": "[genie-icetray] needs updated RELEASE_NOTES", "priority": "blocker", "keywords": "", "time": "2015-08-19T19:39:26", "milestone": "", "owner": "kclark", "type": "defect" } ```
1.0
[genie-icetray] needs updated RELEASE_NOTES (Trac #1213) - There are several commits on trunk that need to be added to the RELEASE_NOTES. Migrated from https://code.icecube.wisc.edu/ticket/1213 ```json { "status": "closed", "changetime": "2016-03-18T21:14:03", "description": "There are several commits on trunk that need to be added to the RELEASE_NOTES.", "reporter": "david.schultz", "cc": "melanie.day", "resolution": "fixed", "_ts": "1458335643235016", "component": "combo simulation", "summary": "[genie-icetray] needs updated RELEASE_NOTES", "priority": "blocker", "keywords": "", "time": "2015-08-19T19:39:26", "milestone": "", "owner": "kclark", "type": "defect" } ```
defect
needs updated release notes trac there are several commits on trunk that need to be added to the release notes migrated from json status closed changetime description there are several commits on trunk that need to be added to the release notes reporter david schultz cc melanie day resolution fixed ts component combo simulation summary needs updated release notes priority blocker keywords time milestone owner kclark type defect
1
50,955
13,187,999,634
IssuesEvent
2020-08-13 05:15:49
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
closed
HiveSplitter segfault with IceAct OMKey(0, 1) (Trac #1732)
Migrated from Trac combo reconstruction defect
Hi everyone, HiveSplitter segfaults in HiveSplitter/private/HiveSplitter/HiveSplitter.cxx:522 The reason behind it is HiveSplitter/public/HiveSplitter/OMKeyHash.h:60 where a negative value is assigned to an unsigned int leading to a very large positive int. In line 522 it is tried to access uninitialised memory. Probably it is best to tell HiveSplitter in the BOOST_FOR_EACH loop to ignore special DOMs like IceAct and the Scintillators checking omkey->IsScintillator() and omkey->IsIceAct() (the IceAct one does not exist yet) Without a fix the muon level3 cannot be processed. Cheers, Leif <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1732">https://code.icecube.wisc.edu/ticket/1732</a>, reported by lraedel and owned by mzoll</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:12:47", "description": "Hi everyone,\n\nHiveSplitter segfaults in HiveSplitter/private/HiveSplitter/HiveSplitter.cxx:522\nThe reason behind it is HiveSplitter/public/HiveSplitter/OMKeyHash.h:60 where a negative value is assigned to an unsigned int leading to a very large positive int. In line 522 it is tried to access uninitialised memory.\n\nProbably it is best to tell HiveSplitter in the BOOST_FOR_EACH loop to ignore special DOMs like IceAct and the Scintillators checking omkey->IsScintillator() and omkey->IsIceAct() (the IceAct one does not exist yet)\n\nWithout a fix the muon level3 cannot be processed.\n\nCheers,\nLeif", "reporter": "lraedel", "cc": "joertlin", "resolution": "fixed", "_ts": "1550067167842669", "component": "combo reconstruction", "summary": "HiveSplitter segfault with IceAct OMKey(0, 1)", "priority": "blocker", "keywords": "", "time": "2016-06-08T19:45:45", "milestone": "", "owner": "mzoll", "type": "defect" } ``` </p> </details>
1.0
HiveSplitter segfault with IceAct OMKey(0, 1) (Trac #1732) - Hi everyone, HiveSplitter segfaults in HiveSplitter/private/HiveSplitter/HiveSplitter.cxx:522 The reason behind it is HiveSplitter/public/HiveSplitter/OMKeyHash.h:60 where a negative value is assigned to an unsigned int leading to a very large positive int. In line 522 it is tried to access uninitialised memory. Probably it is best to tell HiveSplitter in the BOOST_FOR_EACH loop to ignore special DOMs like IceAct and the Scintillators checking omkey->IsScintillator() and omkey->IsIceAct() (the IceAct one does not exist yet) Without a fix the muon level3 cannot be processed. Cheers, Leif <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1732">https://code.icecube.wisc.edu/ticket/1732</a>, reported by lraedel and owned by mzoll</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:12:47", "description": "Hi everyone,\n\nHiveSplitter segfaults in HiveSplitter/private/HiveSplitter/HiveSplitter.cxx:522\nThe reason behind it is HiveSplitter/public/HiveSplitter/OMKeyHash.h:60 where a negative value is assigned to an unsigned int leading to a very large positive int. In line 522 it is tried to access uninitialised memory.\n\nProbably it is best to tell HiveSplitter in the BOOST_FOR_EACH loop to ignore special DOMs like IceAct and the Scintillators checking omkey->IsScintillator() and omkey->IsIceAct() (the IceAct one does not exist yet)\n\nWithout a fix the muon level3 cannot be processed.\n\nCheers,\nLeif", "reporter": "lraedel", "cc": "joertlin", "resolution": "fixed", "_ts": "1550067167842669", "component": "combo reconstruction", "summary": "HiveSplitter segfault with IceAct OMKey(0, 1)", "priority": "blocker", "keywords": "", "time": "2016-06-08T19:45:45", "milestone": "", "owner": "mzoll", "type": "defect" } ``` </p> </details>
defect
hivesplitter segfault with iceact omkey trac hi everyone hivesplitter segfaults in hivesplitter private hivesplitter hivesplitter cxx the reason behind it is hivesplitter public hivesplitter omkeyhash h where a negative value is assigned to an unsigned int leading to a very large positive int in line it is tried to access uninitialised memory probably it is best to tell hivesplitter in the boost for each loop to ignore special doms like iceact and the scintillators checking omkey isscintillator and omkey isiceact the iceact one does not exist yet without a fix the muon cannot be processed cheers leif migrated from json status closed changetime description hi everyone n nhivesplitter segfaults in hivesplitter private hivesplitter hivesplitter cxx nthe reason behind it is hivesplitter public hivesplitter omkeyhash h where a negative value is assigned to an unsigned int leading to a very large positive int in line it is tried to access uninitialised memory n nprobably it is best to tell hivesplitter in the boost for each loop to ignore special doms like iceact and the scintillators checking omkey isscintillator and omkey isiceact the iceact one does not exist yet n nwithout a fix the muon cannot be processed n ncheers nleif reporter lraedel cc joertlin resolution fixed ts component combo reconstruction summary hivesplitter segfault with iceact omkey priority blocker keywords time milestone owner mzoll type defect
1
776,729
27,264,601,370
IssuesEvent
2023-02-22 17:04:46
ascheid/itsg33-pbmm-issue-gen
https://api.github.com/repos/ascheid/itsg33-pbmm-issue-gen
opened
IA-4(4): Identifier Management | Identify User Status
Priority: P2 Class: Technical ITSG-33 Suggested Assignment: IT Operations Group Control: IA-4
# Control Definition IDENTIFIER MANAGEMENT | IDENTIFY USER STATUS The organization manages individual identifiers by uniquely identifying each individual as [Assignment: organization-defined characteristic identifying individual status]. # Class Technical # Supplemental Guidance Characteristics identifying the status of individuals include, for example, contractors and foreign nationals. Identifying the status of individuals by specific characteristics provides additional information about the people with whom organizational personnel are communicating. For example, it might be useful for a government employee to know that one of the individuals on an email message is a contractor. Related control: AT-2. # Suggested Assignment IT Operations Group # Support Teams IT Security Function, IT Projects
1.0
IA-4(4): Identifier Management | Identify User Status - # Control Definition IDENTIFIER MANAGEMENT | IDENTIFY USER STATUS The organization manages individual identifiers by uniquely identifying each individual as [Assignment: organization-defined characteristic identifying individual status]. # Class Technical # Supplemental Guidance Characteristics identifying the status of individuals include, for example, contractors and foreign nationals. Identifying the status of individuals by specific characteristics provides additional information about the people with whom organizational personnel are communicating. For example, it might be useful for a government employee to know that one of the individuals on an email message is a contractor. Related control: AT-2. # Suggested Assignment IT Operations Group # Support Teams IT Security Function, IT Projects
non_defect
ia identifier management identify user status control definition identifier management identify user status the organization manages individual identifiers by uniquely identifying each individual as class technical supplemental guidance characteristics identifying the status of individuals include for example contractors and foreign nationals identifying the status of individuals by specific characteristics provides additional information about the people with whom organizational personnel are communicating for example it might be useful for a government employee to know that one of the individuals on an email message is a contractor related control at suggested assignment it operations group support teams it security function it projects
0
29,113
5,539,703,958
IssuesEvent
2017-03-22 07:42:55
primefaces/primefaces
https://api.github.com/repos/primefaces/primefaces
closed
dropMode doesn't work correctly with dragdrop event on Tree
6.0.18 defect
Reported By PRO User; > When we dragged node has a different nodeType than the node on which it is being dropped it throws an error. In earlier version I could change the nodeType in Backing bean’s dragdrop method and update the tree. That would work.
1.0
dropMode doesn't work correctly with dragdrop event on Tree - Reported By PRO User; > When we dragged node has a different nodeType than the node on which it is being dropped it throws an error. In earlier version I could change the nodeType in Backing bean’s dragdrop method and update the tree. That would work.
defect
dropmode doesn t work correctly with dragdrop event on tree reported by pro user when we dragged node has a different nodetype than the node on which it is being dropped it throws an error in earlier version i could change the nodetype in backing bean’s dragdrop method and update the tree that would work
1
86,953
17,108,289,047
IssuesEvent
2021-07-09 22:04:39
phetsims/circuit-construction-kit-common
https://api.github.com/repos/phetsims/circuit-construction-kit-common
closed
What is the convention for naming imported sounds?
dev:code-review
In code review https://github.com/phetsims/circuit-construction-kit-common/issues/672 @jonathanolson asked: ```js //REVIEW: What is the convention for naming of these? import dogBarkSound from '../../sounds/dog-bark_mp3.js'; ```
1.0
What is the convention for naming imported sounds? - In code review https://github.com/phetsims/circuit-construction-kit-common/issues/672 @jonathanolson asked: ```js //REVIEW: What is the convention for naming of these? import dogBarkSound from '../../sounds/dog-bark_mp3.js'; ```
non_defect
what is the convention for naming imported sounds in code review jonathanolson asked js review what is the convention for naming of these import dogbarksound from sounds dog bark js
0
81,886
31,798,434,995
IssuesEvent
2023-09-13 09:30:43
vector-im/element-x-ios
https://api.github.com/repos/vector-im/element-x-ios
closed
Push notifications leak HTML + Markdown in rich messages
T-Defect S-Major O-Occasional A-Notifications Z-Schedule
### Steps to reproduce 1. Receive a push with a rich message 2. Observe that it shows the raw body of the message, which includes MD (and by extension HTML) formatting, which looks ugly and unprofessional. The recipient doesn't care what markup the sender used to write the message; they care about reading its contents instead. e.g. ![IMG_2858](https://github.com/vector-im/element-x-ios/assets/1294269/0a8b7412-8335-41ce-97cb-8c29cd47580d) ![IMG_2857](https://github.com/vector-im/element-x-ios/assets/1294269/b275bdca-028e-463f-8316-20ad9ccdb53c) ### Outcome #### What did you expect? HTML of rich msgs should be 'flattened' to plaintext to be rendered in notifications, stripping out the tags. (It's interesting that in this example with bnjbvr, the sense of the message would have been lost, however, by stripping out the <del> tags. Perhaps we leave semantically important tags in place.) #### What happened instead? Ugly mix of MD & tags in notifs. ### Your phone model _No response_ ### Operating system version _No response_ ### Application version 380 ### Homeserver _No response_ ### Will you send logs? No
1.0
Push notifications leak HTML + Markdown in rich messages - ### Steps to reproduce 1. Receive a push with a rich message 2. Observe that it shows the raw body of the message, which includes MD (and by extension HTML) formatting, which looks ugly and unprofessional. The recipient doesn't care what markup the sender used to write the message; they care about reading its contents instead. e.g. ![IMG_2858](https://github.com/vector-im/element-x-ios/assets/1294269/0a8b7412-8335-41ce-97cb-8c29cd47580d) ![IMG_2857](https://github.com/vector-im/element-x-ios/assets/1294269/b275bdca-028e-463f-8316-20ad9ccdb53c) ### Outcome #### What did you expect? HTML of rich msgs should be 'flattened' to plaintext to be rendered in notifications, stripping out the tags. (It's interesting that in this example with bnjbvr, the sense of the message would have been lost, however, by stripping out the <del> tags. Perhaps we leave semantically important tags in place.) #### What happened instead? Ugly mix of MD & tags in notifs. ### Your phone model _No response_ ### Operating system version _No response_ ### Application version 380 ### Homeserver _No response_ ### Will you send logs? No
defect
push notifications leak html markdown in rich messages steps to reproduce receive a push with a rich message observe that it shows the raw body of the message which includes md and by extension html formatting which looks ugly and unprofessional the recipient doesn t care what markup the sender used to write the message they care about reading its contents instead e g outcome what did you expect html of rich msgs should be flattened to plaintext to be rendered in notifications stripping out the tags it s interesting that in this example with bnjbvr the sense of the message would have been lost however by stripping out the tags perhaps we leave semantically important tags in place what happened instead ugly mix of md tags in notifs your phone model no response operating system version no response application version homeserver no response will you send logs no
1